Next Article in Journal
Imaging Hallmarks of Sarcoma Progression Via X-ray Computed Tomography: Beholding the Flower of Evil
Previous Article in Journal
Factors That Affect the Formation of Chromosomal Translocations in Cells
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Segmentation of Pancreatic Cancer in Endoscopic Ultrasound Images Using Deep Learning Approach

1
Department of Electrical and Electronics Engineering, Chung-Ang University, Seoul 06974, Korea
2
Division of Gastroenterology, Department of Internal Medicine, Inha University School of Medicine, Incheon 22332, Korea
3
Department of Electronic and Electrical Engineering, Ewha Womans University, Seoul 03760, Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2022, 14(20), 5111; https://doi.org/10.3390/cancers14205111
Submission received: 30 August 2022 / Revised: 12 October 2022 / Accepted: 18 October 2022 / Published: 18 October 2022

Abstract

:

Simple Summary

Surgical therapy is critical to pancreatic cancer survival. The segmentation of pancreatic cancer in endoscopic ultrasonography (EUS) images can provide critical characteristics of the pancreatic cancer for surgical therapy. However, EUS has high operator dependency, and it requires a considerable level of experience and competency to stage pancreatic cancer. Deep learning approaches have been used on EUS images, but there have been no studies on the segmentation of pancreatic cancer. The purpose of this study is to segment pancreatic cancer from EUS images using a neural network model with deep attention features, called DAF-Net. The significance of this study lies in the successful segmentation performance of the pancreatic cancer using the DAF-Net, regardless of the size and location of the cancer, and its usefulness in preoperative planning.

Abstract

Endoscopic ultrasonography (EUS) plays an important role in diagnosing pancreatic cancer. Surgical therapy is critical to pancreatic cancer survival and can be planned properly, with the characteristics of the target cancer determined. The physical characteristics of the pancreatic cancer, such as size, location, and shape, can be determined by semantic segmentation of EUS images. This study proposes a deep learning approach for the segmentation of pancreatic cancer in EUS images. EUS images were acquired from 150 patients diagnosed with pancreatic cancer. A network with deep attention features (DAF-Net) is proposed for pancreatic cancer segmentation using EUS images. The performance of the deep learning models (U-Net, Attention U-Net, and DAF-Net) was evaluated by 5-fold cross-validation. For the evaluation metrics, the Dice similarity coefficient (DSC), intersection over union (IoU), receiver operating characteristic (ROC) curve, and area under the curve (AUC) were chosen. Statistical analysis was performed for different stages and locations of the cancer. DAF-Net demonstrated superior segmentation performance for the DSC, IoU, AUC, sensitivity, specificity, and precision with scores of 82.8%, 72.3%, 92.7%, 89.0%, 98.1%, and 85.1%, respectively. The proposed deep learning approach can provide accurate segmentation of pancreatic cancer in EUS images and can effectively assist in the planning of surgical therapies.

1. Introduction

The incidence and mortality rates of pancreatic cancer are growing quickly (doubling in 10 years from 1997) due to the aging of the global population [1], as 90% of the population diagnosed with pancreatic cancer are older than 55. The prognosis of pancreatic cancer is critical given the 5-year survival rate for pancreatic cancer is approximately 6% [2]. Pancreatic cancer is a highly invasive malignancy as it can quickly invade surrounding tissues and organs [1,3]. Due to the location and nonspecific symptoms of pancreatic cancer for diagnosis, there are difficulties in detecting pancreatic cancer. To diagnose pancreatic cancer, multi-detector computed tomography (MDCT), magnetic resonance imaging (MRI), and endoscopic ultrasonography (EUS) are widely used imaging methods [2,3,4,5,6]. EUS is an ultrasound imaging technology equipped with a high-frequency transducer at the end of the endoscope. The endoscopic transducer is placed in the proximal stomach to obtain high-resolution images of the pancreas [5,7]. Endoscopic ultrasonography-guided fine-needle aspiration (EUS-FNA) is a reliable diagnostic tool for pancreatic cancer [3,5]. Surgical therapy is a critical factor in surviving pancreatic cancer. EUS can stage pancreatic cancer and plays an important role in finding surgically resectable lesions. Semantic segmentation of the cancer area from EUS images is expected to provide the physical characteristics of the cancer, such as size, location, and shape, to help determine the plan for surgical therapies.
Deep learning models are being developed to extract complex features from image data and are widely used in the field of medical imaging [8]. Previous studies have used deep learning models for the classification and segmentation of pancreatic EUS images [9,10,11,12,13]. Kuwahara et al. [9] proved that a deep learning model (ResNet50) can diagnose the malignancy of Intraductal Papillary Mucinous Neoplasm (IPMN). Our previous study classified two types of pancreatic cyst lesions, mucinous cystic neoplasm (MCN) and serous cystic neoplasm (SCN), through Resnet50 [10]. Zhang et al. [11] conducted classification and segmentation to locate the regions of the pancreas through deep learning models (Resnet50 and U-Net++, respectively). Iwasa et al. [12] performed pancreatic cancer segmentation using contrast-enhanced EUS (CE-EUS) images through a deep learning model (U-Net). Additionally, Oh et al. [13] evaluated the segmentation performance of pancreatic cyst lesions through the application of various deep learning models (U-Net, Attention U-Net (Att-U-Net), U-Net++, and Residual U-Net) to EUS images.
In this study, we aim to suggest an efficient deep learning model for pancreatic cancer segmentation by comparing several deep learning models trained by EUS images acquired from patients diagnosed with pancreatic cancer. While recent studies [9,10,11,12,13] have employed deep learning approaches and demonstrated good performance for various other purposes using pancreatic EUS images, to the best of our knowledge, this is the first study on pancreatic cancer segmentation using EUS images. The segmentation performance of the suggested network was evaluated for different T stages [14] and locations of the pancreatic cancer.

2. Materials and Methods

2.1. Data Acquisition

EUS data were collected between January 2010 and July 2021 at the endoscopic center of the Inha University Hospital. This study was reviewed and approved by the Institutional Review Boards of the Inha University Hospital (2020-12-005). Patients were sedated intravenously with propofol (20–120 mg) and midazolam (2.0–5.0 mg) after 8 hours of fasting. EUS was performed using linear (GF-UCT 240; Olympus Optical corp. Ltd., Tokyo, Japan) echoendoscopes with an ultrasound scanning system (SSD 5500, 5 and 10; Aloka, Tokyo, Japan). Video recording started from the point when the echoendoscope entered the stomach and reached the pancreas, and the recording was terminated when all procedures were completed. The procedure was conducted by endosonographers with more than 5 years of experience. EUS images from patients diagnosed with pancreatic cancer were captured from the video recordings. For this study, 330 EUS images were acquired from 150 patients. A summary of the characteristics of the patients (age and gender) and the cancer (size, T-stage, and location) is shown in Table 1.
The operator performed the procedure after thoroughly reviewing clinical data such as MDCT, positron emission tomography (PET), or tumor markers (CA19-9). For the patients with a suspected pancreatic cancer diagnosis, EUS-guided fine-needle aspiration (FNA) was performed when there were signs of pancreatic cancer observed in the EUS images, such as a heterogeneous hypoechoic mass with irregular margins, tumorous pseudopodia, or the abrupt narrowing of the upstream pancreatic duct. In the case in which the results of the EUS-FNA were atypical or negative for malignancy, the patient underwent surgical treatment with further confirmation of a cancer diagnosis provided by a biopsy.

2.2. Method Overview

Figure 1 shows a flowchart of the proposed deep learning approach for pancreatic cancer segmentation. The acquired EUS data were subjected to five-fold cross-validation.

2.3. Data Preparation and Preprocessing

The ground truth (GT) for the segmentation of pancreatic cancer in the EUS images was generated by the endosonographers who determined the cancer area and non-cancer area. For the augmentation of the training data, images were processed by rotating −5° to 5° in steps of 2.5°, scaling by 0.9 and 1.1 times, horizontally flipping, and elastic deformation (by applying random deformation between −2.048 and 2.048 and Gaussian filtering with a standard deviation of 20.48) [15]. Training and test data were resized to 256 × 256 pixels. For 5-fold cross-validation, each fold was split with 66 images from 30 patients.

2.4. Deep Learning Approach

The deep learning model used in this study is a network with deep attentional features (DAF), which has been proposed for prostate segmentation of transrectal ultrasound images [16] as transrectal ultrasound images are geometrically similar to the EUS images. In this paper, this network is denoted as DAF-Net. DAF-Net contains encoder (Figure 2a) and decoder (Figure 2b) parts. The encoder part consists of ResNext101 [17] and the decoder part consists of single-layer features (SLFs), multi-layer features (MLF), DAF module, attentional features (AFs), and single images (SIs). SLFs are obtained by the linear interpolation of blocks from the encoder part (Figure 2a). An MLF is created by the convolution of a concatenated SLF. Attentional features (AFs) are obtained by feeding SLFs and MLF into the DAF module (Figure 3). SIs are obtained by the linear interpolation of AFs. The output is the mean of the SIs. As shown in Figure 2, the DAF module was used to obtain the AFs. In the DAF module, Fx is obtained by concatenating the SLF and MLF. Then, Fx goes through the convolution block to obtain Wx. Through softmax, the attention map (Ax) is calculated. To acquire the AFs, pixel-wise multiplication of Ax and the MLF is concatenated with the SLF and is followed by a 1 × 1 convolution layer.
The encoder part of DAF-Net was initialized by pretrained weights using the ImageNet [18] dataset, and the decoder part of the network was initialized by random values in the range of [0, 1]. A stochastic gradient descent (SGD) with a momentum of 0.9 and weight decay of 0.01 was used to train the network. The initial learning rate was set to 0.007. The learning rate was reduced by 0.02 times at the 5th epoch and by 0.02 times at the 20th epoch. The network was trained by a batch size of 32 and epoch of 50.
The Dice loss (L) is defined as follows:
L = 1 2 n ( X Y ) + smooth   factor n ( X ) + n ( Y ) + smooth   factor
where n() is the total pixels in the area, X is the GT area, Y is the segmentation prediction area, and the smooth factor was 1. The total loss Ltotal was determined as the sum of the Dice losses over all the predicted SLFs and SIs:
L total = i   = 1 k 1 L i + j   = 1 k 2 L j
where Li represents the Dice loss of the i-th layer of the SLFs, Lj represents the Dice loss of the j-th layer of the SIs, and k1 and k2 are the numbers of SLFs and SIs, respectively. In this study, k1 and k2 are 4. The network was trained and tested using an Intel(R) Core (TM) i9-10900X CPU @ 3.70GHz processor, a CUDA-enabled Nvidia RTX 3090 graphical processing unit (GPU), Python 3.8, and PyTorch 1.10.

2.5. Performance Evaluation

The performance of the segmentation results was evaluated using 10 evaluation indices: Dice similarity coefficient (DSC) [19,20], intersection over union (IoU), sensitivity (SEN), specificity (SP), precision (PC), average distance (AVD) [21], mean absolute distance (MAD) [22], Hausdorff distance (HD) [23], receiver operating characteristic (ROC) curve, and area under the curve (AUC). All the values were averaged for 5-fold.
The DSC and IoU are metrics that determine how much the GT and prediction overlap. The two metrics range from 0 to 1, and are defined as follows:
DSC = 2 n ( X Y ) n ( X ) + n ( Y ) ,   IoU =   n ( X Y ) n ( X Y )
where n() is the total pixels in the area, X is the GT area, and Y is the segmentation prediction area. The AVD, HD, and MAD are metrics that measure the distance between a point on the contour of the GT and a point on the contour of the prediction. They are calculated as follows:
AVD = max ( 1 N a a A min b B d ( a , b ) , 1 N b b B min a A d ( b , a ) )
HD = max ( max a A min b B d ( a , b ) , max b B min a A d ( b , a ) )
MAD = 1 2 ( 1 N a a A min b B d ( a , b ) , 1 N b b B min a A d ( b , a ) )
where A and B represent the contour regions of the GT and prediction, respectively, a and b represent points in the region A and B, d(a, b) represents the distance from a to b, and Na and Nb represent the numbers of a and b in the regions A and B, respectively. SEN, SP, and PC are calculated as follows:
SEN = TP TP + FN ,   SP =   TN FP + TN ,   PC = TP TP + FP
which are associated with the true positive (TP), true negative (TN), false positive (FP), and false negative (FN).

2.6. Statistical Analysis

To determine if the data were normally distributed, a test was performed using the Shapiro–Wilk test [24]. The Kruskal–Wallis test [25] was used to analyze the results with regard to the T-stage and cancer location. A p value greater than 0.05 was considered statistically nonsignificant.

3. Results

3.1. Deep Learning Network

Figure 4 shows the learning curves for the 5-fold cross-validation during training of the DAF-Net using pancreas EUS images. It is observed that the learning curves of the 5 folds show similar trends and are flattened at the loss of 0.90 after 20 epochs. Thus, the learning curves show that the network was trained successfully for the segmentation of pancreas cancer using EUS images. The training loss of the DAF-Net is relatively large (>0.9) because the loss (Equation (2)) was computed by comparing the SLFs (Figure 2b) and GTs. SLFs correspond to the upper attention part of the network which is far from the output of the network.

3.2. Performance Evaluation

The evaluation results of U-Net, Att-U-Net, and DAF-Net for the segmentation of pancreatic cancer are shown in Table 2. Overall, DAF-Net achieved a higher DSC, IoU, and SEN than U-Net and Att-U-Net. Regarding the distance-based metrics (i.e., AVD, MAD, and HD), DAF-Net performs the best compared to the other networks.
The ROC curves of U-Net, Att-U-Net, and DAF-Net are shown in Figure 5. The ROC curve of DAF-Net is located more towards the upper-left corner compared with other networks. Also, the AUC for DAF-Net is 10.1% and 11.3% greater than that for U-Net and Att-U-Net, respectively. This implies that DAF-Net is superior to other networks in distinguishing cancerous and non-cancerous regions.
In Figure 6, examples of original EUS images (Figure 6a), ground truth (Figure 6b), and segmentation results from DAF-Net, U-Net and Att-U-Net (Figure 6c–e, respectively) are shown. While the segmented lesions from DAF-Net are very close to those of the ground truth, U-Net and Att-U-Net cannot predict the shapes of the lesions correctly. In addition, U-Net and Att-U-Net resulted in multiple segmented lesions, although the training was performed with EUS images with a single lesion. Thus, it is clearly observed that, in terms of the quality of the segmentation maps of the DAF-Net, they contain fewer errors and produce more precise segmentation compared to other models. The results match well with the performance evaluation presented in Table 2.
The segmentation performance in terms of the size and position of the pancreatic cancer were compared by Kruskal–Wallis test and visualized by a box plot. The Kruskal–Wallis tests show that no significant differences (p value = 0.98 and 0.9 for DSC and IoU, respectively) were found among the three different stages (i.e., T1, T2, and T3), and that no significant differences (p value = 0.13 and 0.13 for DSC and IoU respectively) were found among the five different parts (i.e., body, head, neck, tail, and uncinate). However, the p value is relatively low, and the most significant difference was observed between the cancers located in the tail and uncinate where the p value was 0.01 for both the DSC and IoU. Figure 7a,b show box plots with the median (red line inside the box), upper and lower quartiles (top and bottom edges of the box), and min and max of nonoutliers (whiskers) of the DSC (Figure 7a) and IoU (Figure 7b) of DAF-Net from three different cancer stages (T1–3), respectively. Figure 7c,d show the median, upper and lower quartiles, and min and max of nonoutliers of the DSC (Figure 7c) and IoU (Figure 7d) of DAF-Net for five different locations of the pancreatic cancer, respectively. Similar to the result from the Kruskal–Wallis test, the box plots show that there is no significant difference in terms of the tumor stage and the position of the pancreatic cancer.

4. Discussion

In this study, we conducted pancreatic cancer segmentation using EUS images through a deep learning model. The occurrence of pancreatic cancer is relatively low, and it is more challenging to collect EUS images with the diagnosis of pancreatic cancer than endoscopic images [26]. Thus, the number of studies using EUS images of pancreatic cancer is fewer than those using endoscopy images. Although this study was performed with a limited number of images, overall, the proposed deep learning approach using DAF-Net achieved good performance (>82% DSC and >72% IoU) for pancreatic cancer segmentation using EUS images. Although the performance may vary slightly with the selection of hyperparameters, each network was optimized with the hyperparameters resulting in the best possible performance.
Most of the effort before the surgical therapy of pancreatic cancer is devoted to the staging of the cancer, which is based on the cancer size and location, to avoid misclassification of the resectable area [27]. Due to the surrounding non-cancerous lesions, the boundary of a pancreatic cancer is obscure. Additionally, the presence of arteries and ducts near the pancreatic cancer is the main factor contributing to the uncertainty when segmenting the cancerous region. The pancreas is situated in a very complex position near the liver, kidney, duodenum, and various vascular structures, affecting the ability to clearly define the boundary of the pancreatic cancer. Considering the aforementioned difficulties, there is no single imaging modality that can clearly determine the stage of pancreatic cancer [28]. Because EUS has high operator dependency, it requires a considerable level of experience and competency to stage pancreatic cancer [29]. Hence, the significance of this study lies in the successful segmentation performance of the pancreatic cancer, regardless of the size and location of the cancer, and its usefulness in preoperative planning.
Before the segmentation process, determining whether there is the presence of cancer in the EUS image or not is an important process. Figure 8 shows consecutive EUS images acquired from a recorded EUS video file. Initially, the proposed approach forms no segmentation mask from the EUS images in Figure 8a,b, which show common bile duct without relevance to cancer; then, it forms the segmentation mask to locate the cancerous region in Figure 8c–e. Although the proposed deep learning approach is not developed for the purpose of cancer detection, it can further be extended to determine the presence of the cancer. When the dataset acquired from various situations (e.g., varying scanning locations and angles) is sufficient, our study will be further extended to determine the presence of cancer in the EUS images.
To determine the success of the segmentation results, the IoU is commonly used by checking whether the IoU between the segmented area and the ground truth is greater than a predefined threshold value (0.5) [30]. U-Net, Att-U-Net, and DAF-Net achieved an IoU greater than 0.5 in 75.2%, 70.9%, and 90.9%, respectively. While this result shows that the suggested approach can overcome many difficulties in the segmentation of the pancreatic cancer, Figure 9 shows the cases where the IoU is less than 0.5 from the test using our DAF-Net. Figure 9a shows the case in which the EUS probe was not properly in contact with the tissue; thus, the EUS image does not show all the 180 angles of the pancreas, resulting in inaccurate segmentation results. Figure 9b shows a case with pancreatic duct dilation, which was misrecognized as a pancreatic cancer by DAF-Net. In Figure 9c, due to the similarity of the surrounding blood vessels and organs to the pancreatic cancer in the EUS images, other parts were also segmented as pancreatic cancer. As shown in Figure 9d, due to high invasiveness or the accompanying parenchymal atrophy caused by chronic pancreatitis, it is difficult to define clear boundaries of the tumor. In Figure 9e, inaccurate prediction due to intratumoral necrosis associated with chronic pancreatitis is observed. Figure 9f shows that when the size of the tumor is large (5.5 cm in this example), ultrasound penetration can be limited, and a heterogeneous echo is generated due to necrosis, calcified lesions, and degenerative changes inside the mass. Thus, it is difficult to clearly distinguish the main lesion from the marginal area. In the statistical analysis (Figure 7), it was observed that there was no significant difference in the segmentation performance in terms of the size of the tumor and the position of the pancreatic cancer. However, the amount of patient data with cancers in stage T3 or cancers located in the uncinate is relatively lower than those with other conditions (Table 1). Thus, further investigation with more data is necessary. In a future study, more training images from cases that are challenging and include various stages of pancreatic cancer will be acquired to further improve the performance of the current network and confirm the statistical analysis.
The process of pancreatic cancer segmentation using EUS images has several limitations to overcome. First, EUS images can be degraded due to inherent characteristics such as speckles, shadows, and missing boundaries [31]. Thus, contrast-enhanced EUS was utilized to enhance the image quality in other studies [12,32]. By injecting the contrast agent (e.g., microbubbles), one could overcome the weakness of EUS with higher contrast and spatial resolution. However, there still remains the issue of safety when using contrast agents. Despite the inherent characteristics of the EUS images without the contrast agent, the deep learning approach proposed in this study performed well in the segmentation of the pancreatic cancer. Second, there are two types of EUS, which are radial EUS and linear EUS. In this study, images from linear EUS were used only to train and test the deep learning network. For future work, images from radial EUS will also be utilized. The accurate and precise segmentation of the cancer area is critical for the surgical therapy of pancreatic cancer. In addition, vascular invasion is an important factor to diagnose the stage of cancer and plan for the therapeutic strategy [33,34]. Thus, our future work will include an investigation of vascular invasion in pancreatic cancer using a deep learning approach. Therefore, the proposed method in this study can be improved to help the diagnosis as well as the decision for the therapy to treat pancreatic cancer.

5. Conclusions

In summary, the proposed deep learning approach using DAF-Net provides a superior segmentation performance for pancreatic cancer using EUS images. It is expected that the proposed accurate and precise segmentation approach can provide effective and efficient assistance in planning for the surgical therapy of pancreatic cancer.

Author Contributions

Conceptualization, J.-S.P. and S.P.; data curation, J.-S.P., J.-H.L., K.S., J.S. and L.S.N.; software, methodology, and investigation, K.S., J.-H.L., J.S., L.S.N., H.Y. and S.P.; writing-original draft, review & editing, K.S., J.-H.L., H.Y. and S.P.; supervision, J.-S.P. and S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Korea Medical Device Development Fund grant funded by the Korean government (the Ministry of Science and ICT, the Ministry of Trade, Industry and Energy, the Ministry of Health & Welfare, and the Ministry of Food and Drug Safety) (Project Number: 202016B01).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Boards of the Inha University Hospital (2020-12-005).

Informed Consent Statement

Patient consent was waived because of the retrospective nature of the study and the anonymous analysis.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

This research was supported by the Chung-Ang University Graduate Research Scholarship in 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pourshams, A.; Sepanlou, S.G.; Ikuta, K.S.; Bisignano, C.; Safiri, S.; Roshandel, G.; Sharif, M.; Khatibian, M.; Fitzmaurice, C.; Nixon, M.R.; et al. The global, regional, and national burden of pancreatic cancer and its attributable risk factors in 195 countries and territories, 1990–2017: A systematic analysis for the Global Burden of Disease Study 2017. Lancet Gastroenterol. Hepatol. 2019, 4, 934–947. [Google Scholar] [CrossRef] [Green Version]
  2. Hu, J.-X.; Zhao, C.-F.; Chen, W.-B.; Liu, Q.-C.; Li, Q.-W.; Lin, Y.-Y.; Gao, F. Pancreatic cancer: A review of epidemiology, trend, and risk factors. World J. Gastroenterol. 2021, 27, 4298–4321. [Google Scholar] [CrossRef] [PubMed]
  3. Zhao, Z.; Liu, W. Pancreatic Cancer: A Review of Risk Factors, Diagnosis, and Treatment. Technol. Cancer Res. Treat. 2020, 19, 1533033820962117. [Google Scholar] [CrossRef] [PubMed]
  4. Kamisawa, T.; Wood, L.D.; Itoi, T.; Takaori, K. Pancreatic cancer. Lancet 2016, 388, 73–85. [Google Scholar] [CrossRef]
  5. Zhang, L.; Sanagapalli, S.; Stoita, A. Challenges in diagnosis of pancreatic cancer. World J. Gastroenterol. 2018, 24, 2047–2060. [Google Scholar] [CrossRef] [PubMed]
  6. Park, H.S.; Lee, J.M.; Choi, H.K.; Hong, S.H.; Han, J.K.; Choi, B.I. Preoperative evaluation of pancreatic cancer: Comparison of gadolinium-enhanced dynamic MRI with MR cholangiopancreatography versus MDCT. J. Magn. Reson. Imaging 2009, 30, 586–595. [Google Scholar] [CrossRef] [PubMed]
  7. Kitano, M.; Yoshida, T.; Itonaga, M.; Tamura, T.; Hatamaru, K.; Yamashita, Y. Impact of endoscopic ultrasonography on diagnosis of pancreatic cancer. J. Gastroenterol. 2019, 54, 19–32. [Google Scholar] [CrossRef] [Green Version]
  8. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  9. Kuwahara, T.; Hara, K.; Mizuno, N.; Okuno, N.; Matsumoto, S.; Obata, M.; Kurita, Y.; Koda, H.; Toriyama, K.; Onishi, S.; et al. Usefulness of Deep Learning Analysis for the Diagnosis of Malignancy in Intraductal Papillary Mucinous Neoplasms of the Pancreas. Clin. Transl. Gastroenterol. 2019, 10, e00045. [Google Scholar] [CrossRef]
  10. Nguon, L.S.; Seo, K.; Lim, J.-H.; Song, T.-J.; Cho, S.-H.; Park, J.-S.; Park, S. Deep Learning-Based Differentiation between Mucinous Cystic Neoplasm and Serous Cystic Neoplasm in the Pancreas Using Endoscopic Ultrasonography. Diagnostics 2021, 11, 1052. [Google Scholar] [CrossRef]
  11. Zhang, J.; Zhu, L.; Yao, L.; Ding, X.; Chen, D.; Wu, H.; Lu, Z.; Zhou, W.; Zhang, L.; An, P.; et al. Deep learning–based pancreas segmentation and station recognition system in EUS: Development and validation of a useful training tool (with video). Gastrointest. Endosc. 2020, 92, 874–885.e873. [Google Scholar] [CrossRef]
  12. Iwasa, Y.; Iwashita, T.; Takeuchi, Y.; Ichikawa, H.; Mita, N.; Uemura, S.; Shimizu, M.; Kuo, Y.-T.; Wang, H.-P.; Hara, T. Automatic Segmentation of Pancreatic Tumors Using Deep Learning on a Video Image of Contrast-Enhanced Endoscopic Ultrasound. J. Clin. Med. 2021, 10, 3589. [Google Scholar] [CrossRef]
  13. Oh, S.; Kim, Y.-J.; Park, Y.-T.; Kim, K.-G. Automatic Pancreatic Cyst Lesion Segmentation on EUS Images Using a Deep-Learning Approach. Sensors 2021, 22, 245. [Google Scholar] [CrossRef]
  14. Amin, M.B.; Edge, S.B.; Greene, F.L.; Byrd, D.R.; Brookland, R.K.; Washington, M.K.; Gershenwald, J.E.; Compton, C.C.; Hess, K.R.; Sullivan, D.C.; et al. AJCC Cancer Staging Manual, 8th ed.; Springer: New York, NY, USA, 2017. [Google Scholar]
  15. Simard, P.Y.; Steinkraus, D.; Platt, J.C. Best practices for convolutional neural networks applied to visual document analysis. In Proceedings of the Seventh International Conference on Document Analysis and Recognition, Edinburgh, UK, 6 August 2003; pp. 958–963. [Google Scholar]
  16. Wang, Y.; Deng, Z.; Hu, X.; Zhu, L.; Yang, X.; Xu, X.; Heng, P.-A.; Ni, D. Deep Attentional Features for Prostate Segmentation in Ultrasound. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2018, Granada, Spain, 16–20 September 2018; pp. 523–530. [Google Scholar]
  17. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  18. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Kai, L.; Li, F.-F. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  19. Dice, L.R. Measures of the Amount of Ecologic Association Between Species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  20. Sørensen, T.J. A method of establishing groups of equal amplitude in plant sociology based on similarity of species content and its application to analyses of the vegetation on Danish commons. K. Dan. Vidensk. Selsk. 1948, 5, 1–34. [Google Scholar]
  21. Taha, A.A.; Hanbury, A. Metrics for evaluating 3D medical image segmentation: Analysis, selection, and tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef] [Green Version]
  22. Shahedi, M.; Cool, D.W.; Bauman, G.S.; Bastian-Jordan, M.; Fenster, A.; Ward, A.D. Accuracy Validation of an Automated Method for Prostate Segmentation in Magnetic Resonance Imaging. J. Digit. Imaging 2017, 30, 782–795. [Google Scholar] [CrossRef]
  23. Huttenlocher, D.P.; Klanderman, G.A.; Rucklidge, W.J. Comparing images using the Hausdorff distance. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 850–863. [Google Scholar] [CrossRef] [Green Version]
  24. Shapiro, S.S.; Wilk, M.B. An Analysis of Variance Test for Normality (Complete Samples). Biometrika 1965, 52, 591–611. [Google Scholar] [CrossRef]
  25. Kruskal, W.H.; Wallis, W.A. Use of Ranks in One-Criterion Variance Analysis. J. Am. Stat. Assoc. 1952, 47, 583–621. [Google Scholar] [CrossRef]
  26. Kuwahara, T.; Hara, K.; Mizuno, N.; Haba, S.; Okuno, N.; Koda, H.; Miyano, A.; Fumihara, D. Current status of artificial intelligence analysis for endoscopic ultrasonography. Dig. Endosc. 2021, 33, 298–305. [Google Scholar] [CrossRef]
  27. Soriano, A.; Castells, A.; Ayuso, C.; Ayuso, J.R.; de Caralt, M.T.; Ginès, M.À.; Real, M.I.; Gilabert, R.; Quintó, L.; Trilla, A.; et al. Preoperative Staging and Tumor Resectability Assessment of Pancreatic Cancer: Prospective Study Comparing Endoscopic Ultrasonography, Helical Computed Tomography, Magnetic Resonance Imaging, and Angiography. Off. J. Am. Coll. Gastroenterol. ACG 2004, 99, 492–501. [Google Scholar] [CrossRef]
  28. Li, J.; Feng, C.; Shen, Q.; Lin, X.; Qian, X. Pancreatic cancer segmentation in unregistered multi-parametric MRI with adversarial learning and multi-scale supervision. Neurocomputing 2022, 467, 310–322. [Google Scholar] [CrossRef]
  29. Eisen, G.M.; Dominitz, J.A.; Faigel, D.O.; Goldstein, J.A.; Petersen, B.T.; Raddawi, H.M.; Ryan, M.E.; Vargo, J.J.; Young, H.S.; Wheeler-Harbaugh, J.; et al. Guidelines for credentialing and granting privileges for endoscopic ultrasound. Gastrointest. Endosc. 2001, 54, 811–814. [Google Scholar] [CrossRef]
  30. Zou, Z.; Shi, Z.; Guo, Y.; Ye, J. Object detection in 20 years: A survey. arXiv 2019, arXiv:1905.05055. [Google Scholar]
  31. Van Dam, J.; Brady, P.G.; Freeman, M.; Gress, F.; Gross, G.W.; Hassall, E.; Hawes, R.; Jacobsen, N.A.; Liddle, R.A.; Ligresti, R.J. Guidelines for training in endoscopic ultrasound. Gastrointest. Endosc. 1999, 49, 829–833. [Google Scholar]
  32. Iwashita, T.; Uemura, S.; Mita, N.; Iwasa, Y.; Ichikawa, H.; Senju, A.; Yasuda, I.; Shimizu, M. Utility of endoscopic ultrasound and endoscopic ultrasound-guided fine-needle aspiration for the diagnosis and management of pancreatic cystic lesions: Differences between the guidelines. Dig. Endosc. 2020, 32, 251–262. [Google Scholar] [CrossRef]
  33. Kim, S.-H. Endoscopic Ultrasonography for Vascular Invasion in Pancreatic Cancer. Clin. Endosc. 2019, 52, 397–398. [Google Scholar] [CrossRef] [Green Version]
  34. Yamada, K.; Kawashima, H.; Ohno, E.; Ishikawa, T.; Tanaka, H.; Nakamura, M.; Miyahara, R.; Ishigami, M.; Hirooka, Y.; Fujishiro, M. Diagnosis of vascular invasion in pancreatic ductal adenocarcinoma using endoscopic ultrasound elastography. BMC Gastroenterol. 2020, 20, 81. [Google Scholar] [CrossRef]
Figure 1. Overall flowchart of the proposed deep learning approach.
Figure 1. Overall flowchart of the proposed deep learning approach.
Cancers 14 05111 g001
Figure 2. Block diagram of (a) encoder and (b) decoder parts of DAF-Net.
Figure 2. Block diagram of (a) encoder and (b) decoder parts of DAF-Net.
Cancers 14 05111 g002
Figure 3. Block diagram of the DAF module.
Figure 3. Block diagram of the DAF module.
Cancers 14 05111 g003
Figure 4. Learning curves of DAF-Net for the 5-fold cross-validation.
Figure 4. Learning curves of DAF-Net for the 5-fold cross-validation.
Cancers 14 05111 g004
Figure 5. ROC curves for the three different networks (U-Net, Att-U-Net, and DAF-Net).
Figure 5. ROC curves for the three different networks (U-Net, Att-U-Net, and DAF-Net).
Cancers 14 05111 g005
Figure 6. Examples of (a) original EUS images with pancreatic cancer, (b) GT, and segmentation results from (c) U-Net, (d) Att-U-Net, and (e) DAF-Net.
Figure 6. Examples of (a) original EUS images with pancreatic cancer, (b) GT, and segmentation results from (c) U-Net, (d) Att-U-Net, and (e) DAF-Net.
Cancers 14 05111 g006
Figure 7. Statistical analysis for the (a) DSC and (b) IoU for three different cancer stages and the (c) DSC and (d) IoU for five different locations of the pancreatic cancer. (A red line inside the box represents the median value, the top and bottom whiskers represent the min and max of nonoutliers, respectively, and a red star indicates outliers.).
Figure 7. Statistical analysis for the (a) DSC and (b) IoU for three different cancer stages and the (c) DSC and (d) IoU for five different locations of the pancreatic cancer. (A red line inside the box represents the median value, the top and bottom whiskers represent the min and max of nonoutliers, respectively, and a red star indicates outliers.).
Cancers 14 05111 g007
Figure 8. Consecutive EUS images in 0.5 s intervals from a recorded EUS video. (a,b) show common bile duct with no visible cancerous region, (ce) show pancreatic cancer precisely segmented using DAF-Net (Red line: GT boundary, blue line: predicted boundary).
Figure 8. Consecutive EUS images in 0.5 s intervals from a recorded EUS video. (a,b) show common bile duct with no visible cancerous region, (ce) show pancreatic cancer precisely segmented using DAF-Net (Red line: GT boundary, blue line: predicted boundary).
Cancers 14 05111 g008
Figure 9. Cases where the IoU is less than 0.5 in the segmentation using DAF-Net (Red line: GT boundary, blue line: predicted boundary). (a) EUS probe not properly contacted with the tissue, (b) cancer along with pancreatic duct dilation, (c) similar texture of the surrounding blood vessels and organs, (d) parenchymal atrophy by chronic pancreatitis (e) intratumoral necrosis associated with chronic pancreatitis, and (f) large tumor size.
Figure 9. Cases where the IoU is less than 0.5 in the segmentation using DAF-Net (Red line: GT boundary, blue line: predicted boundary). (a) EUS probe not properly contacted with the tissue, (b) cancer along with pancreatic duct dilation, (c) similar texture of the surrounding blood vessels and organs, (d) parenchymal atrophy by chronic pancreatitis (e) intratumoral necrosis associated with chronic pancreatitis, and (f) large tumor size.
Cancers 14 05111 g009
Table 1. Characteristics of the patients and cancer.
Table 1. Characteristics of the patients and cancer.
CharacteristicsCount
Age, years, mean (range)71.2 (40–94)
Gender (male/female)96/54
Cancer size, mm, mean (range)31.0 (10–130)
Cancer T-stage (T1, T2, T3) [9]33/99/18
Cancer location (body, head, neck, tail, uncinate)37/60/18/33/6
Table 2. Summary of network performance evaluation (Bold font indicates best result obtained).
Table 2. Summary of network performance evaluation (Bold font indicates best result obtained).
MethodDSCIoUAVDHDMADSENSPPC
U-Net0.740.6213.8444.2610.470.800.980.83
Att-U-Net0.720.6014.6943.4711.280.780.980.81
DAF-Net0.830.729.0427.357.530.840.980.85
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Seo, K.; Lim, J.-H.; Seo, J.; Nguon, L.S.; Yoon, H.; Park, J.-S.; Park, S. Semantic Segmentation of Pancreatic Cancer in Endoscopic Ultrasound Images Using Deep Learning Approach. Cancers 2022, 14, 5111. https://doi.org/10.3390/cancers14205111

AMA Style

Seo K, Lim J-H, Seo J, Nguon LS, Yoon H, Park J-S, Park S. Semantic Segmentation of Pancreatic Cancer in Endoscopic Ultrasound Images Using Deep Learning Approach. Cancers. 2022; 14(20):5111. https://doi.org/10.3390/cancers14205111

Chicago/Turabian Style

Seo, Kangwon, Jung-Hyun Lim, Jeongwung Seo, Leang Sim Nguon, Hongeun Yoon, Jin-Seok Park, and Suhyun Park. 2022. "Semantic Segmentation of Pancreatic Cancer in Endoscopic Ultrasound Images Using Deep Learning Approach" Cancers 14, no. 20: 5111. https://doi.org/10.3390/cancers14205111

APA Style

Seo, K., Lim, J. -H., Seo, J., Nguon, L. S., Yoon, H., Park, J. -S., & Park, S. (2022). Semantic Segmentation of Pancreatic Cancer in Endoscopic Ultrasound Images Using Deep Learning Approach. Cancers, 14(20), 5111. https://doi.org/10.3390/cancers14205111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop