Next Article in Journal
A Computational Approach to Identification of Candidate Biomarkers in High-Dimensional Molecular Data
Previous Article in Journal
Artificial Intelligence in Colon Capsule Endoscopy—A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection and Characterization of Gastric Cancer Using Cascade Deep Learning Model in Endoscopic Images

1
School of Medical Sciences, Fujita Health University, Toyoake 470-1192, Japan
2
Department of Gastroenterology and Hepatology, Fujita Health University, Toyoake 470-1192, Japan
3
Faculty of Engineering, Gifu University, Gifu 501-1194, Japan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2022, 12(8), 1996; https://doi.org/10.3390/diagnostics12081996
Submission received: 26 June 2022 / Revised: 31 July 2022 / Accepted: 17 August 2022 / Published: 18 August 2022
(This article belongs to the Special Issue Advances in Gastric Cancer Imaging)

Abstract

:
Endoscopy is widely applied in the examination of gastric cancer. However, extensive knowledge and experience are required, owing to the need to examine the lesion while manipulating the endoscope. Various diagnostic support techniques have been reported for this examination. In our previous study, segmentation of invasive areas of gastric cancer was performed directly from endoscopic images and the detection sensitivity per case was 0.98. This method has challenges of false positives and computational costs because segmentation was applied to all healthy images that were captured during the examination. In this study, we propose a cascaded deep learning model to perform categorization of endoscopic images and identification of the invasive region to solve the above challenges. Endoscopic images are first classified as normal, showing early gastric cancer and showing advanced gastric cancer using a convolutional neural network. Segmentation on the extent of gastric cancer invasion is performed for the images classified as showing cancer using two separate U-Net models. In an experiment, 1208 endoscopic images collected from healthy subjects, 533 images collected from patients with early stage gastric cancer, and 637 images from patients with advanced gastric cancer were used for evaluation. The sensitivity and specificity of the proposed approach in the detection of gastric cancer via image classification were 97.0% and 99.4%, respectively. Furthermore, both detection sensitivity and specificity reached 100% in a case-based evaluation. The extent of invasion was also identified at an acceptable level, suggesting that the proposed method may be considered useful for the classification of endoscopic images and identification of the extent of cancer invasion.

1. Introduction

1.1. Background

Gastric cancer is a malignant tumor that mainly affects the gastric mucosa and is the second-most-common cause of death among all cancers after lung cancer [1,2]. It is a highly regional disease, with more than 50% of cases occurring in East Asia. Diagnostic and treatment techniques for gastric cancer continue to improve and early detection has been shown to reduce mortality rates among patients with gastric cancer.
Gastric examination is performed by endoscopy and gastrofluoroscopy using barium. Because endoscopy has superior sensitivity in detecting early stage gastric cancer and allows for tissue collection and treatment under observation, it has been widely adopted for gastric cancer screening and detailed diagnosis. In endoscopic diagnosis, a physician checks the endoscopic images while operating the endoscope and records still images at key points when abnormalities are detected. In addition, magnification and staining are often performed to classify diseases and diagnose the extent and depth of lesions [3].
However, the procedure is highly complex, requiring multiple tasks to be performed during the examination. Hence, there is some concern that lesions may be missed. According to the results of a survey, the probability of missing a lesion during endoscopy was 22.2% and the accuracy of diagnosis depends, to a considerable extent, on the experience and skill of the physician. Therefore, technologies to assist examinations are strongly desired to reduce the burden on the physician and to improve diagnostic accuracy.
In recent years, artificial intelligence technologies have made remarkable progress and deep learning technology has shown excellent performance in the field of image recognition [4,5]. Deep learning techniques have been proposed for various applications involving different types of medical images [6,7,8]. In this study, we propose a deep-learning-based method for the detection and identification of the depth and extent of cancer invasion to support endoscopy.

1.2. Related Works

Various studies have been conducted on the automated detection of gastric cancer lesions in endoscopic images, the extraction of the extent of invasion, and the evaluation of cancer depth [9,10,11,12,13,14,15]. Among early works on the subject, Hirasawa et al. developed a method for automated detection of early gastric cancer lesions and extraction of the extent of invasion using a single-shot detector (SSD), an object detection model [10]. The sensitivity of detection using 15,880 original endoscopic images was 92.2% and the positive predictive value was 30.6%. Sakai et al. developed a method to detect gastric cancer by feeding a convolutional neural network (CNN) with finely cut patches of endoscopic images to classify areas showing gastric cancer and normal areas [11] and achieved a detection sensitivity and specificity using 926 original images of 80.0% and 94.8%, respectively. Shibata et al. proposed a method for detecting the presence of early gastric cancer and extracting invasive regions using Mask R-CNN, which was designed to perform both object detection and segmentation [12]. They achieved a sensitivity of 96.0% and a segmentation concordance of 71% using 1741 original endoscopic images in the automated detection of early gastric cancer. Teramoto et al. proposed a U-Net R-CNN model to detect early gastric cancer using two CNNs to perform segmentation and classification [13]. First, they used U-Net, a CNN model for segmentation, to detect regions of early gastric cancer. Then, the detected regions were classified as normal or containing gastric cancer by a CNN model. They obtained a detection sensitivity of 98% and reduced false positives by 70% compared to their previous study on the same image database.
To evaluate the depth of gastric cancer, Zhu et al. used a CNN to classify cancer as remaining within the mucosa or as having invaded the submucosa, the latter indicating an advanced cancer. The evaluation resulted in a sensitivity of 76.47% and a specificity of 95.56% using the original 993 endoscopic images [14]. Hamada et al. also proposed a method to classify early gastric cancer as mucosal or submucosal using a CNN and obtained a classification accuracy of 78.9% using 3508 original images [15]. Endoscopic ultrasonography (EUS) is often used to assess the depth and internal structure of upper gastrointestinal lesions, including gastric cancer, and diagnostic assistive technologies are required [16,17]. Hirai et al. used an AI system to classify subepithelial lesions (SEL) using EUS images. The results showed that the correct classification rate of the images was 86.1% using 16,110 original images, which was much higher than that of endoscopists [17].
These techniques present independent methods to detect the presence and evaluate the depth of gastric carcinoma. In our previous study, semantic segmentation was applied directly to endoscopic images; however, the many healthy images collected during the examination were subject to processing, resulting in false positives and high computational cost. The actual detailed diagnoses were performed by first detecting abnormal areas during the endoscopic procedure and approaching the lesion. Hence, a more automated detection and classification scheme in line with the diagnostic flow is desirable. This cascaded processing has been applied to other medical images. For example, Krzysztof et al. proposed a method for segmenting brain tumors from head MR images that uses two types of U-Nets to perform tumor location identification and detailed multiclass segmentation [18]. Jie et al. proposed a cascaded network model for the segmentation of brain regions in 3T head MR images in combination with 7T head MR images and obtained better results than conventional methods [19]. In addition, Nina et al. analyzed the aortic root using a cascade of two types of CNNs [20]. These methods do not use a single network model to perform a single task but divide the roles among multiple networks. Similar to these methods, a performance improvement can be expected in the detection of gastric cancer using endoscopic images by introducing a cascade structure of two different tasks. In this study, we propose a cascade deep learning model designed to perform automatic detection and classification, which combines image classification and region extraction to detect normal and gastric cancer, evaluate the depth of the lesion, and identify the extent of invasion.
The organization of this paper is as follows: In the Methods, we propose a cascade-based gastric cancer detection and classification method. In the Results, we evaluate the detection characteristics by calculating the detection performance and activation maps. In the Discussion section, the detection performance and its comparison with previous studies are discussed. Finally, in the Conclusions, we describe the effectiveness and contribution of this study.

2. Methods

2.1. Proposed Method

Figure 1 shows an outline of the proposed method. Still images taken during the endoscopy procedure are fed to a CNN model trained to perform image classification and are classified as healthy, indicating early gastric cancer, and indicating advanced gastric cancer. Then, the images classified as showing gastric cancer are input into a CNN model trained to perform segmentation to extract the regions of early and advanced gastric cancer and the extent of invasion is identified.

2.2. Image Dataset

For this study, endoscopic images were collected from healthy subjects, patients with early stage gastric cancer and patients with advanced gastric cancer. All images were collected at the endoscopy center of Fujita Health University Hospital during an examination. They comprised 1208 endoscopic images collected from 42 healthy subjects, 533 images from 93 patients with early gastric cancer, and 637 images from 50 patients with advanced gastric cancer. Images were obtained using upper endoscopes (GIF-290Z, GIF-HQ290, GIF-XP290N, GIF-260Z; Olympus Medical Systems, Co., Ltd., Tokyo, Japan; and EG-L600ZW7; Fujifilm Corp., Tokyo, Japan) and standard endoscopic video systems (EVIS LUCERA CV-260/CLV-260, EVIS LUCERA ELITE CV-290/CLV-290SL; Olympus Medical Systems; and VP-4450HD/LL-4450; Fujifilm Corp.). The image matrix sizes ranged from 640 × 480 to 1440 × 1080 pixels; these were standard white-light images stored in JPEG format. This study was approved by an institutional review board and patients’ informed consent was obtained under the condition that all data were anonymized (No. HM17-226).
Here, early gastric cancer was defined as lesions that remained in the mucosal or submucosal layer of the stomach, whereas advanced gastric cancer was defined as lesions that reached to the intrinsic muscular layer or deeper than that layer. For healthy subjects, we reassessed the cases diagnosed as normal by the endoscopists and characterized a case as “healthy” when there was no specific lesion, such as a polyp, tumor, or gastritis.
Images stored by endoscope systems may contain text information, such as patient names, and the shape of the field of view differs depending on the model of the endoscope. Therefore, as shown in Figure 2, a perfect circle was inscribed in the effective field of view of the endoscope, the area outside of the circle was filled in black, and a square area was cropped with the bounding rectangle of the circle. The circular field of view was implemented to avoid the bias caused by different endoscope models and maintain image uniformity when data augmentation was performed as described below. Sample images of the dataset are shown in Figure 3.

2.3. Annotation of Gastric Cancer Region

For images diagnosed as gastric cancer, labeled images of gastric cancer areas were created for the segmentation of gastric cancer areas. Using an in-house annotation tool, labeled images with the invasion areas of gastric cancer marked were created and confirmed by a board-certified endoscopist (T. S.).

2.4. Data Augmentation

The number of images in the three categories differed. In addition, training a network using a small number of images taken at a certain angle of view may result in overfitting. In endoscopic examination, images of a single lesion are taken from various positions and distances and there is wide variation in the position and size of the lesion in the image, with sufficient data available. On the other hand, there is not much variation in the rotation of images due to the limitations of examination techniques. Therefore, to avoid bias and overfitting caused by the number of images and viewing angle, data augmentation was performed by image rotation [12]. The rate of augmentation was varied according to the number of original images belonging to the class. Images of healthy subjects were augmented by a factor of 4 at a pitch of 90 degrees, images of early gastric cancer were augmented by a factor of 9 at a pitch of 40 degrees, and images of advanced gastric cancer were augmented by a factor of 7 at a pitch of 50 degrees.

2.5. Network Architecture for Image Classification

The proposed method involves image classification and segmentation by two CNNs, as shown in Figure 1. First, the given images are classified as healthy, showing early gastric cancer or showing advanced gastric cancer by a CNN model. In this study, VGG-16/19 [21], InceptionV3 [22], ResNet-50/-101/-152 [23], and DenseNet-121/-169/-201 [24] models were constructed and compared in terms of their classification performance.
We introduced transfer learning to the training of the CNN described above, which involves transferring the processing ability acquired in solving one task to learning to solve another task. If the CNN of the original diversion is a well-trained model with a huge amount of data, so it may provide a higher capacity [25]. In this study, a CNN model trained on the ImageNet database consisting of more than 10 million images was adapted for classification of endoscopic images. To conduct the transfer learning, the fully connected layers of the original CNNs were removed, 1024 and 2 units of fully connected layers were newly connected, and the weights of the fully connected layers were adjusted by training using endoscopic images. We used the Adam optimizer with a learning rate of 0.0001 and 50 training epochs.

2.6. Network Architecture for Image Segmentation

Images classified by the first-stage CNN were transferred to the next-stage CNN for segmentation according to their classification results. Images classified as healthy were judged as not including cancer areas and, thus, segmentation was not performed. Images classified as showing early and advanced gastric cancer were fed to a dedicated CNN for segmentation processing.
We implemented U-Net as a CNN model to perform classification, given its excellent performance in medical image segmentation [26]. The structure of the network is shown in Figure 4 and included five layers of encoders and decoders. The encoders extract the features of the image. The decoders upscale the extracted feature maps and generate a pattern similar to that of the label image given as the teaching data. Furthermore, between the same layer of encoders and decoders, spatial information reduced by the encoder layer is given to the decoders by a mechanism called skip connections, which transfer information between the encoder and decoder layers.
Here, the morphology shown in images of early and advanced gastric cancer differs considerably. Therefore, two independent U-Nets were created to segment early and advanced gastric cancer. The Adam optimizer was used for training with a learning coefficient of 0.0001 and 50 training epochs.

2.7. Evaluation Metrics

To confirm the effectiveness of the proposed method, we evaluated the output of the CNN models for image classification and segmentation. First, to evaluate the performance of the CNN for image classification in the first stage, a confusion matrix was created based on the CNN classification results. Based on the matrix, we calculated the accuracy, sensitivity, and specificity of the models. NHealthy, NEGC, and NAGC indicate the numbers of images for healthy, early gastric cancer, and advanced gastric cancer, respectively, and CHealty, CEGC, and CAGC are the numbers of images successfully classified as healthy, early gastric cancer, and advanced gastric cancer, respectively.
AccuracyOverall = (CHealty + CEGC + CAGC)/(NHealthy + NEGC + NAGC)
AccuracyHealth = CHealty/NHealthy
AccuracyEGC = CEGC/NEGC
AccuracyAGC = CAGC/NAGC
AccuracyBalanced = (AccuracyHealth + AccuracyEGC + AccuracyAGC)/3
Sensitivity = (CEGC + CAGC)/(NEGC + NAGC)
Specificity = CHealty/NHealthy
The above indices were evaluated on an image-by-image basis (image-based evaluation) and on a case-by-case basis (case-based evaluation). For the former, the results were tabulated when each image was classified into the class with the largest CNN output value. For the case-by-case evaluation, the output values of images collected from the same case were averaged for each class and the class with the highest average value was considered as the classification result.
As for the visualization method of CNN output, class activation mapping (CAM) calculates what parts of an image influence the prediction based on the feature map during inference. Among various CAM methods, Grad-CAM calculates the activation map by calculating the gradient of the CNN feature map and can provide a stable activation map regardless of the model [27]. In this study, we calculated activation maps for healthy cases, early gastric cancer, and advanced gastric cancer using Grad-CAM to visualize the basis for classification.
In the second step of the segmentation process, the Dice coefficient (Di) and Jaccard coefficient (Ji) were used to evaluate the extent to which the invasive area output by U-Net was correctly extracted. Di and Ji were defined by the following equation to evaluate the degree of agreement between the label image output by U-Net and the correct label (Ground Truth) created by the endoscopist.
Di = 2|A∩B|/(|A| + |B|) × 100 [%]
Ji = |A∩B|/|A∪B| × 100 [%]
where A is the ground truth prepared by the gastroenterologist and B represents the region of gastric cancer obtained by the CNN. Cross-validation was used to evaluate the classification and segmentation process [28]. In cross-validation, the image dataset was divided into K subsets to avoid fragmenting the cases. The CNN was trained on K-1 subsets and the image data belonging to the remaining one subset was defined as testing data; then, the classification results were evaluated. In the cross-validation method, the test results for all data were obtained by training and testing K times while changing the subset used as the testing data. The data were divided into 5 subsets (5-fold cross-validation) and the classification performance was evaluated.
The calculations of two CNNs were performed using software we developed in the Python programming language with an AMD Ryzen 9 3950X processor (16 CPU cores, 4.7 GHz) with 128 GB of DDR4 memory. The training processes of the CNNs were accelerated by an NVIDIA Quadro RTX 8000 GPU (48 GB memory).

3. Results

Image Classification Result

Table 1 shows the results of the comparison of the classification performance of the CNN models for image classification, which performed in the first stage of this study. DenseNet-121 had the highest classification performance for both image-based and case-based classification, with a detection sensitivity and specificity of 0.970 and 0.994, respectively, in the image-based evaluation. In the case-based evaluation, all of the values were 1.0. A confusion matrix of the classification results from DenseNet-121 is shown in Table 2. Next, the correctly and incorrectly classified images are shown in Figure 5, along with the Grad-CAM output.
The results of U-Net segmentation are shown in Figure 6 and the Dice and Jaccard coefficients are listed in Table 3. Note that the segmentation of healthy images often extracted false-positive regions. The evaluation results showed that 283 false-positive regions were detected in 1208 healthy images and the number of false positives per image (FPI) was 0.234 (283/1208). Because the proposed method performs image classification in the first stage, images classified as healthy need not be segmented. The false positives from the segmentation results were investigated by excluding the images correctly classified as healthy in the first stage and false-positive regions were extracted from six images, yielding an FPI of 0.005 (6/1208).

4. Discussion

In this study, we proposed a cascaded deep learning model for endoscopic images, designed to classify the depth of normal and gastric cancer using a CNN model to perform image classification and then performed segmentation on the extent of gastric cancer invasion using U-Net.
In the first stage of the image classification task, nine different CNN models were employed and their classification performances were compared. The results showed that the performance differences among the models were not large, being within 5% for most of the evaluation indices. Among the CNN models, DenseNet-121 achieved the highest classification performance for all items, with a detection sensitivity of 97% and a specificity of 99.4% for gastric cancer. A slight misclassification occurred in the image-based evaluation. In cases of early gastric cancer, images with extensive areas of invasion were classified as advanced gastric cancer. In healthy cases, some bubbles adhering to the gastric mucosa and some hyperemic areas produced by endoscopy were misclassified as advanced gastric cancer. Endoscopy involves collecting multiple images for a single patient. The classification results of the individual images were tabulated and a decision was made for each patient. Consequently, all cases were classified into the correct category, as shown in Table 2. These results indicate that the CNN established in the first stage was able to accurately classify not only normal samples and those exhibiting gastric cancer but also the depth of the cancer.
As a result of acquiring the regions of interest for image classification by Grad-CAM, the lesion and its surrounding area tended to be focused on for early and advanced gastric cancer. In healthy subjects, the results showed that the entire gastric mucosa was uniformly focused on. These results are similar to the process of diagnosis using images, which can be considered reasonable.
In the second stage of the cascade model, the invasive extent was segmented by U-Net for images classified as early or advanced gastric cancer and the agreement between the invasive extent output by U-Net and the gold standard specified by the endoscopist was evaluated using the Dice and Jaccard coefficients. The Dice coefficient was 0.555 for early gastric cancer and 0.716 for advanced gastric cancer. This was due to the lack of morphological changes in the images of early gastric cancer and the fact that many of the images were taken from the side of the gastric wall rather than from the front, so the contours of the invasive area were not completely defined. Even with this underestimation, the extent of the invasion output, like the Grad-CAM output, provides a basis for image classification and can provide important information during endoscopy.
To compare the detection performance of gastric cancer, Table 4 shows the performances of previous studies and the proposed method. In our previous study [13], we presented a method that obtained a case-based detection sensitivity of 0.989, and the number of false positives (FPI) per image was 0.011. When compared under the same conditions, those obtained in the present work were 1.000 and 0.005, respectively.
Our previous study faced challenges about false positives and high computational cost because semantic segmentation was applied to all healthy images collected during the examination. The detection sensitivity and the number of false positives in the proposed cascade model were superior to those of our previous study and the computational cost of the segmentation process was eliminated because healthy images were cut off by the classification CNN. Although an accurate comparison is not possible because a different database was used, the proposed method with cascaded model had a better detection sensitivity than those in previous studies using an SSD (sensitivity of 99.2%) [10]. The proposed method can classify images with high sensitivity and specificity and can also provide the extent of invasion, suggesting that it may be considered a highly effective technique for assisting in endoscopic screening of gastric cancer.
A limitation of this study is that the endoscopic images were collected at a single facility. In the future, an evaluation of usefulness of the proposed approach should be conducted using images from multiple facilities. Regarding the evaluation of the extraction accuracy of the invasive area, the Dice and Jaccard coefficients were introduced. However, a subjective evaluation of whether the clinically important invasive area was extracted is required. Furthermore, we intend to develop endoscopy support software that incorporates the constructed model and confirm its usefulness in practical medicine.

5. Conclusions

In this study, we proposed a cascade deep learning model to support the automated detection and classification of gastric cancer during endoscopy by combining image classification and segmentation to detect normal samples and gastric cancer, to evaluate the degree of depth, and to identify the extent of invasion. The evaluation results showed that the accuracy of first-stage image classification was high and the proposed method accurately classified normal images and gastric cancer and classified the depth of invasion. In addition, the extent of gastric cancer invasion was correctly extracted from several images. The number of false positives generated by the proposed method was significantly low, indicating that the method is effective for endoscopic screening.

Author Contributions

Conceptualization, A.T., T.S., Y.H., K.S. and H.F.; data curation, T.S. and H.Y.; methodology, A.T.; software, A.T.; validation, T.S., A.T. and H.Y.; investigation, A.T., T.S. and H.Y.; writing—original draft preparation, A.T. and T.S.; writing—review and editing, Y.H., K.S. and H.F.; visualization, A.T.; project administration, A.T. and T.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was approved by the Ethical Review Committee of Fujita Health University (HM20-021, 3 June 2020) and carried out in accordance with the World Medical Association’s Declaration of Helsinki.

Informed Consent Statement

Patients’ informed consent was obtained via opt-out in the endoscopic center of Fujita Health University Hospital and all data were anonymized.

Data Availability Statement

The source code and additional information used to support the findings of this study will be made available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fitzmaurice, C.; Akinyemiju, T.F.; Al Lami, F.H.; Alam, T.; Alizadeh-Navaei, R.; Allen, C.; Alsharif, U.; Alvis-Guzman, N.; Amini, E.; Anderson, B.O.; et al. Global, regional, and national cancer incidence, mortality, years of life lost, years lived with disability, and disability-adjusted life-years for 29 cancer groups, 1990 to 2016: A systematic analysis for the global burden of disease study global burden of disease cancer collaboration. JAMA Oncol. 2018, 4, 1553–1568. [Google Scholar] [PubMed]
  2. Karger Publishers. GLOBOCAN 2012: Estimated Cancer Incidence, Mortality, and Prevalence Worldwide in 2012. Available online: http://globocan.iarc.fr/Pages/fact_sheets_cancer.aspx (accessed on 31 October 2021).
  3. Tashiro, A.; Sano, M.; Kinameri, K.; Fujita, K.; Takeuchi, Y. Comparing mass screening techniques for gastric cancer in Japan. World J. Gastroenterol. 2006, 12, 4873–4874. [Google Scholar]
  4. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1106–1114. [Google Scholar] [CrossRef]
  5. LeCun, Y.; Bengio, Y.; Hinton, G.E. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  6. Teramoto, A.; Fujita, H.; Yamamuro, O.; Tamaki, T. Automated detection of pulmonary nodules in PET/CT images: Ensemble false-positive reduction using a convolutional neural network technique. Med. Phys. 2016, 43, 2821–2827. [Google Scholar] [CrossRef] [PubMed]
  7. Yan, K.; Cai, J.; Zheng, Y.; Harrison, A.P.; Jin, D.; Tang, Y.B.; Tang, Y.X.; Huang, L.; Xiao, J.; Lu, L. Learning from Multiple Datasets with Heterogeneous and Partial Labels for Universal Lesion Detection in CT. arXiv 2020, arXiv:2009.02577. [Google Scholar] [CrossRef] [PubMed]
  8. Teramoto, A.; Kiriyama, Y.; Tsukamoto, T.; Sakurai, E.; Michiba, E.; Imaizumi, K.; Saito, K.; Fujita, H. Weakly Supervised Learning for Classification of Lung Cytological Images Using Attention-Based Multiple Instance Learning. Sci. Rep. 2011, 11, 20317. [Google Scholar] [CrossRef]
  9. Jin, Z.; Gan, T.; Wang, P.; Fu, Z.; Zhang, C.; Yan, Q.; Zheng, X.; Liang, X.; Ye, X. Deep learning for gastroscopic images: Computer-aided techniques for clinicians. BioMed. Eng. OnLine 2022, 21, 12. [Google Scholar] [CrossRef]
  10. Hirasawa, T.; Aoyama, K.; Tanimoto, T.; Ishihara, S.; Shichijo, S.; Ozawa, T.; Ohnishi, T.; Fujishiro, M.; Matsuo, K.; Fujisaki, J.; et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer 2018, 21, 653–660. [Google Scholar] [CrossRef]
  11. Sakai, Y.; Takemoto, S.; Hori, K.; Nishimura, M.; Ikematsu, H.; Yano, T.; Yokota, H. Automatic detection of early gastric cancer in endoscopic images using a transferring convolutional neural network. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 4138–4141. [Google Scholar]
  12. Shibata, T.; Teramoto, A.; Yamada, H.; Ohmiya, N.; Saito, K.; Fujita, H. Automated detection and segmentation of early gastric cancer from endoscopic images using mask R-CNN. Appl. Sci. 2020, 10, 3842. [Google Scholar] [CrossRef]
  13. Teramoto, A.; Shibata, T.; Yamada, H.; Hirooka, Y.; Saito, K.; Fujita, H. Automated Detection of Gastric Cancer by Retrospective Endoscopic Image Dataset Using U-Net R-CNN. Appl. Sci. 2021, 11, 11275. [Google Scholar] [CrossRef]
  14. Zhu, Y.; Wang, Q.C.; Xu, M.D.; Zhang, Z.; Cheng, J.; Zhong, Y.S.; Zhang, Y.-Q.; Chen, W.-F.; Yao, L.-Q.; Zhou, P.-H.; et al. Application of convolutional neural networks in the diagnosis of the invasion depth of gastric cancer based on conventional endoscopy. Gastrointest. Endosc. 2019, 89, 806–815. [Google Scholar] [CrossRef]
  15. Hamada, K.; Kawahara, Y.; Tanimoto, T.; Ohto, A.; Aida, T.; Yamasaki, Y.; Gotoda, T.; Ogawa, T.; Abe, M.; Okanoue, S.; et al. Application of convolutional neural networks for evaluating the depth of invasion of early gastric cancer based on endoscopic images. Gastroenterol. Hepatol. 2022, 37, 352–357. [Google Scholar] [CrossRef]
  16. Ungureanu, B.; Sacerdotianu, V.; Turcu-Stiolica, A.; Cazacu, I.; Saftoiu, A. Endoscopic Ultrasound vs. Computed Tomography for Gastric Cancer Staging: A Network Meta-Analysis. Diagnostics 2021, 11, 134. [Google Scholar] [CrossRef]
  17. Hirai, K.; Kuwahara, T.; Furukawa, K.; Kakushima, N.; Furune, S.; Yamamoto, H.; Marukawa, T.; Asai, H.; Matsui, K.; Sasaki, Y.; et al. Artificial intelligence-based diagnosis of upper gastrointestinal subepithelial lesions on endoscopic ultrasonography images. Gastric Cancer 2022, 25, 382–391. [Google Scholar] [CrossRef] [PubMed]
  18. Kotowski, K.; Adamski, S.; Malara, W.; Machura, B.; Zarudzki, L.; Nalepa, J. Segmenting Brain Tumors from MRI Using Cascaded 3D U-Nets. Lect. Notes. Comput. Sci. 2021, 12659, 265–277. [Google Scholar]
  19. Wei, J.; Wu, Z.; Wang, L.; Toan, D.; Qu, L.; Yap, P.-T.; Xia, Y.; Li, G.; Shen, D. A cascaded nested network for 3T brain MR image segmentation guided by 7T labeling. Pattern Recognit. 2022, 124, 108420. [Google Scholar] [CrossRef]
  20. Krüger, N.; Meyer, A.; Tautz, L.; Hüllebrand, M.; Wamala, I.; Pullig, M.; Kofler, M.; Kempfert, J.; Sündermann, S.; Falk, V.; et al. Cascaded neural network-based CT image processing for aortic root analysis. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 507–519. [Google Scholar] [CrossRef]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  22. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  25. Zamir, A.R.; Sax, A.; Shen, W.; Guibas, L.J.; Malik, J.; Savarese, S. Taskonomy: Disentangling Task Transfer Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 3712–3722. [Google Scholar]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. Lect. Notes. Comput. Sci. 2015, 9351, 234–241. [Google Scholar]
  27. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  28. Efron, B. Estimating the error rate of a prediction rule: Improvement on cross-validation. J. Am. Stat. Assoc. 1983, 78, 316–331. [Google Scholar] [CrossRef]
Figure 1. Outline of the proposed method.
Figure 1. Outline of the proposed method.
Diagnostics 12 01996 g001
Figure 2. Image cropping for regularization.
Figure 2. Image cropping for regularization.
Diagnostics 12 01996 g002
Figure 3. Sample images of the image dataset. (a) healthy subjects; (b) early gastric cancer; (c) advanced gastric cancer.
Figure 3. Sample images of the image dataset. (a) healthy subjects; (b) early gastric cancer; (c) advanced gastric cancer.
Diagnostics 12 01996 g003
Figure 4. U-Net architecture for gastric cancer segmentation.
Figure 4. U-Net architecture for gastric cancer segmentation.
Diagnostics 12 01996 g004
Figure 5. Correctly and incorrectly classified images: (a) classification results of early gastric cancer; (b) classification results of advanced gastric cancer; (c) classification results of healthy subjects.
Figure 5. Correctly and incorrectly classified images: (a) classification results of early gastric cancer; (b) classification results of advanced gastric cancer; (c) classification results of healthy subjects.
Diagnostics 12 01996 g005
Figure 6. Lesions detected and missed in the segmentation phase: (a,b) correctly detected region of advanced gastric cancer; (c) missed region of advanced gastric cancer; (d,e) correctly detected region of early gastric cancer; (f) missed region of early gastric cancer.
Figure 6. Lesions detected and missed in the segmentation phase: (a,b) correctly detected region of advanced gastric cancer; (c) missed region of advanced gastric cancer; (d,e) correctly detected region of early gastric cancer; (f) missed region of early gastric cancer.
Diagnostics 12 01996 g006
Table 1. Comparison of CNN models for image classification. Values in bold indicate the results of the CNN model with the highest performance.
Table 1. Comparison of CNN models for image classification. Values in bold indicate the results of the CNN model with the highest performance.
(a) Image-Based Classification
AccuracyHealthyAccuracyEGCAccuracyAGCAccuracyoverallAccuracyBalancedSensitivitySpecificity
VGG160.9640.9190.9220.9430.9350.9510.964
VGG190.9940.9470.9040.9600.9490.9390.994
InceptionV30.9880.9210.9060.9510.9380.9380.988
DenseNet1210.9940.9980.9450.9820.9790.9700.994
DenseNet1690.9950.9980.9030.9710.9650.9480.995
DenseNet2010.9960.9890.9430.9800.9760.9650.996
ResNet500.9910.9910.9220.9720.9680.9560.991
ResNet1010.9900.9810.9200.9690.9640.9490.990
ResNet1520.9960.9940.9180.9750.9700.9540.996
(b) Case-based classification
AccuracyHealthyAccuracyEGCAccuracyAGCAccuracyoverallAccuracyBalancedSensitivitySpecificity
VGG160.9760.9370.9600.9520.9580.9790.976
VGG191.0000.9470.9400.9570.9620.9721.000
InceptionV31.0000.9580.9800.9730.9790.9861.000
DenseNet1211.0001.0001.0001.0001.0001.0001.000
DenseNet1691.0001.0000.9600.9890.9870.9861.000
DenseNet2011.0001.0000.9800.9950.9930.9931.000
ResNet501.0000.9891.0000.9950.9961.0001.000
ResNet1011.0001.0000.9800.9950.9930.9931.000
ResNet1521.0001.0000.9800.9950.9930.9931.000
Table 2. Confusion matrices of classification using DenseNet121.
Table 2. Confusion matrices of classification using DenseNet121.
(a) Image-Based Classification
Predicted
HealthyEarly Gastric CancerAdvanced Gastric Cancer
ActualHealthy120107
Early gastric cancer05311
Advanced gastric cancer350602
(b) Case-based classification
Predicted
HealthyEarly Gastric CancerAdvanced Gastric Cancer
ActualHealthy4200
Early gastric cancer0950
Advanced gastric cancer0050
Table 3. Evaluation results of cancer segmentation.
Table 3. Evaluation results of cancer segmentation.
DiJi
Early gastric cancer0.5550.427
Advanced gastric cancer0.7160.611
Table 4. Performance comparison of gastric cancer detection and segmentation.
Table 4. Performance comparison of gastric cancer detection and segmentation.
AuthorMethodImage DatasetDetection PerformanceSegmentation Performance
Hirasawa et al. [10]SSDOriginal
15,880 images
Sensitivity = 0.922
Positive predictive value = 0.306
-
Sakai et al. [11]CNNOriginal
926 images
Sensitivity = 0.800
Specificity = 0.948
-
Shibata et al. [12]Mask R-CNNOriginal
1741 images
Sensitivity = 0.96
False positives per image = 0.105
Dice index = 0.54
Teramoto et al. [13]U-Net + CNNOriginal
1741 images
Sensitivity = 0.98
False positives per image = 0.011
Dice index = 0.56
Proposed methodCascade CNNOriginal
2378 images
Sensitivity = 1.00
False positives per image = 0.005
Dice index = 0.56
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Teramoto, A.; Shibata, T.; Yamada, H.; Hirooka, Y.; Saito, K.; Fujita, H. Detection and Characterization of Gastric Cancer Using Cascade Deep Learning Model in Endoscopic Images. Diagnostics 2022, 12, 1996. https://doi.org/10.3390/diagnostics12081996

AMA Style

Teramoto A, Shibata T, Yamada H, Hirooka Y, Saito K, Fujita H. Detection and Characterization of Gastric Cancer Using Cascade Deep Learning Model in Endoscopic Images. Diagnostics. 2022; 12(8):1996. https://doi.org/10.3390/diagnostics12081996

Chicago/Turabian Style

Teramoto, Atsushi, Tomoyuki Shibata, Hyuga Yamada, Yoshiki Hirooka, Kuniaki Saito, and Hiroshi Fujita. 2022. "Detection and Characterization of Gastric Cancer Using Cascade Deep Learning Model in Endoscopic Images" Diagnostics 12, no. 8: 1996. https://doi.org/10.3390/diagnostics12081996

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop