Next Article in Journal
Performance Comparison of Object Detection Networks for Shrapnel Identification in Ultrasound Images
Next Article in Special Issue
Semi-Supervised Medical Image Segmentation with Co-Distribution Alignment
Previous Article in Journal
Bioactive Aspergteroids G–J from Soft-Coral-Associated Symbiotic and Epiphytic Fungus Aspergillus terreus EGF7-0-1
Previous Article in Special Issue
Radiomics-Based Quality Control System for Automatic Cardiac Segmentation: A Feasibility Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Segmentation of Gastric Polyps in Endoscopic Images Based on Convolutional Neural Networks and an Integrated Evaluation Approach

1
School of Mechanical Engineering, Hubei University of Arts and Science, Xiangyang 441053, China
2
Department of Electromechanical Engineering, University of Macau, Taipa, Macau 999078, China
3
Xiangyang Central Hospital, Affiliated Hospital of Hubei University of Arts and Science, Xiangyang 441021, China
4
Faculty of Medicine, Macau University of Science and Technology, Taipa, Macau 999078, China
5
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
6
School of Medicine, Shanghai Jiao Tong University, Shanghai 200240, China
7
Department of General Surgery, Centro Hospitalar Conde de São Januário, Macau 999078, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work and should be considered as co-first authors.
Bioengineering 2023, 10(7), 806; https://doi.org/10.3390/bioengineering10070806
Submission received: 13 June 2023 / Revised: 27 June 2023 / Accepted: 3 July 2023 / Published: 5 July 2023
(This article belongs to the Special Issue Recent Advance of Machine Learning in Biomedical Image Analysis)

Abstract

:
Convolutional neural networks (CNNs) have received increased attention in endoscopic images due to their outstanding advantages. Clinically, some gastric polyps are related to gastric cancer, and accurate identification and timely removal are critical. CNN-based semantic segmentation can delineate each polyp region precisely, which is beneficial to endoscopists in the diagnosis and treatment of gastric polyps. At present, just a few studies have used CNN to automatically diagnose gastric polyps, and studies on their semantic segmentation are lacking. Therefore, we contribute pioneering research on gastric polyp segmentation in endoscopic images based on CNN. Seven classical semantic segmentation models, including U-Net, UNet++, DeepLabv3, DeepLabv3+, Pyramid Attention Network (PAN), LinkNet, and Muti-scale Attention Net (MA-Net), with the encoders of ResNet50, MobineNetV2, or EfficientNet-B1, are constructed and compared based on the collected dataset. The integrated evaluation approach to ascertaining the optimal CNN model combining both subjective considerations and objective information is proposed since the selection from several CNN models is difficult in a complex problem with conflicting multiple criteria. UNet++ with the MobineNet v2 encoder obtains the best scores in the proposed integrated evaluation method and is selected to build the automated polyp-segmentation system. This study discovered that the semantic segmentation model has a high clinical value in the diagnosis of gastric polyps, and the integrated evaluation approach can provide an impartial and objective tool for the selection of numerous models. Our study can further advance the development of endoscopic gastrointestinal disease identification techniques, and the proposed evaluation technique has implications for mathematical model-based selection methods for clinical technologies.

Graphical Abstract

1. Introduction

Gastric cancer is the fifth most common cancer globally and the third most prominent cause of cancer-related mortality, with more than 782,000 deaths annually [1]. The 5-year survival rate for gastric cancer exceeds 90% if diagnosed at an early stage, but it will decrease to 10–20% if diagnosed at later stages [2]. Gastric polyps show a wide variety of malignant potential and are harbingers of gastric adenocarcinoma. The appearance of gastric polyps may potentially signify a higher risk of intestinal or extra-intestinal malignancy. Depending on histological type, the majority of gastric polyps are categorized as fundic gland, hyperplastic, or adenomatous polyps [3]. In a large pathology study in the United States, the percentage of patients diagnosed with gastric polyps among patients who underwent Esophagogastroduodenoscopy (EGD) and gastric biopsies was 6.35%; 77% were fundic gland polyps, 17% were hyperplastic polyps, and 0.69% were adenomas [4]. Among them, adenomatous polyps have been considered true neoplasms, with malignant transformation occurring in 6% to 47% of cases [5]. Therefore, timely diagnosis, biopsy, and removal of gastric polyps are necessary, which help eliminate the hidden dangers of adenoma early and lower the risk of gastric cancer [6].
Gastric polyps are protuberant lesions originating from the epithelium or submucosa that project into the lumen. Since most gastric polyps are asymptomatic, it is challenging to discover them before a physical examination. EGD has been an effective and direct way to diagnose polyps and perform polypectomy in clinical practice [7]. Endoscopists can examine the situation of the gastric epithelium and treat the polyps through EGD. Nevertheless, the diagnostic accuracy of gastric polyps depends highly on endoscopists and environmental factors. On the one hand, endoscopic characteristics, pathogenesis, histopathological prediction, and management of polyps rely on endoscopists’ experience [7]. Even if experienced endoscopists are involved, operation level and diagnostic rate will be influenced by high workload and fatigue [8]. On the other hand, due to the stomach’s complicated structure and inconsistency in image quality, diagnosis outcomes show large variations in terms of inter- and intra-observer agreement [9].
Computer-aided diagnosis (CAD) systems have been adopted recently to overcome the above challenges [10]. In recent years, CNN-based CAD systems have shown great potential for assisting in the diagnosis of endoscopic polyps [11]. CNN is composed of hierarchical processing, including convolution layers, pooling layers, fully connected layers, et cetera, and can learn the most representation of data with multiple abstractions [12]. Compared with traditional machine learning techniques, CNN can automatically capture spatial and temporal dependencies in images without manual intervention [13]. Image classification, object detection, and semantic segmentation are the essential tasks of using CNN. Differing from image classification and object detection, semantic segmentation can assign each pixel a predefined category label and segment each object region precisely from its background region. Additionally, semantic segmentation can split input images into mutually exclusive subsets and delineate their boundaries accurately, each of which corresponds to a meaningful section of the original image [14]. However, the bounding boxes in object detection are unable to portray the boundaries of polyps precisely. On the contrary, semantic segmentation can accurately delineate the boundaries of polyps and elicit information about their morphology. It is beneficial to clinical diagnosis and treatment because endoscopists must consider polyp topography, size, and endoscopic appearance to decide whether and how to perform polypectomy or whether to enroll in an endoscopic surveillance program when operating EGD [15]. Therefore, an automated CNN-based polyp-segmentation system is developed in this study.
Many semantic segmentation models have been proposed and extensively applied [14]. The selection of the best models for the automated polyp-segmentation system is a complex problem that mainly refers to the performance of the models concerning multiple criteria. Most existing research focuses on improving single-criteria performance without considering integral performance or realistic application. To evaluate the models holistically and select the models with higher performance for the automated polyp-segmentation system appropriately, decision makers (i.e., clinical endoscopists and artificial intelligence researchers) could utilize methodological tools combined with quantitative and qualitative analyses to evaluate the performance of the models. The selection is regarded as a Multiple Criteria Decision Making (MCDM) problem, often arranged as a decision matrix in which alternatives are evaluated with corresponding criteria [16]. After computation, the weights of alternatives that represent the importance of each criterion to each other are obtained and used to calculate the overall weighted score for selection. The subjective method, derived from the decision maker’s judgment, and the objective method, derived by quantifying the intrinsic information of each metric to help eliminate bias and improve objectivity, are the two main MCDM methods [17]. In the selection of clinical assistant techniques, they both are of great importance and should be comprehensively considered to determine the weights and overall score. MCDM methods are capable of dealing with complex situations considering multiple criteria, particularly conflicting criteria, and can facilitate the merger of quantitative and qualitative analyses in a scientific way.
In this study, the diagnosis of gastric polyps is the goal, semantic segmentation is the assisted technology, and MCDM is the evaluation approach, all combining to build an automated polyp-segmentation system to assist endoscopists in the diagnosis of gastric polyps. The main contributions of this study are described as follows:
(1)
A high-quality gastric polyp dataset for training, validation, and testing of the semantic segmentation models is built, and the dataset will be publicly available for further research.
(2)
This is pioneering research on gastric polyp segmentation. Additionally, seven semantic segmentation models, including U-Net, UNet++, DeepLabv3, DeepLabv3+, PAN, LinkNet, and MA-Net, with the encoders of ResNet50, MobineNetV2, or EfficientNet-B1, are constructed and compared.
(3)
The objective and subjective evaluation methods are combined to propose a novel integrated evaluation approach to evaluate the experimental results, aiming at the determination of the best CNN model for the automated polyp-segmentation system.
The remaining sections are organized as follows: Section 2 reviews several existing studies about CNN-based gastric polyp diagnosis and some MCDM methods. Section 3 introduces the collected gastric polyp dataset, the semantic segmentation models used in this paper, and the novel integrated evaluation approach. Section 4 introduces the implementation details and test results of segmentation models with different encoders. An in-depth discussion of the result is also presented. Section 5 concludes and provides our future research plan.

2. Related Work

2.1. CNN-Based Gastric Polyp Diagnosis

In this section, a brief review of CNN-based gastric polyp diagnosis is summarized. Zhang et al. [18] developed a CNN for gastric polyp detection based on the Sigle Shot MultiBox Detector (SSD). It can achieve real-time detection speed with 50 frames per second (FPS), and the mean average precision (mAP) is increased by 90.4%. Not only did they pioneer research on gastric polyp detection, but they also published their dataset of 404 endoscopic images with gastric images. Laddha et al. [19] applied YOLOv3 and YOLOv3-tiny for gastric polyp detection. They obtained a mAP of 0.91 and a mAP of 0.82. Wang et al. [20] used an improved Faster R-CNN network to detect gastric polyps and obtained a precision of 78.96%, a recall rate of 76.07%, and an F1-score of 77.49%. Cao et al. [21] developed the YOLOv3 network, which combined a feature extraction module and a self-developed fusion module, to help with small gastric polyp detection. The improved network obtained a decent performance with an F1 score of 88.8%. Durak et al. [22] collected 2195 endoscopic images consisting of 3031 polyp labels retrospectively and implemented them on YOLOv4, CenterNet, EfficientNet, and the other five existing models. The YOLOv4 model showed the best performance, with a mAP of 87.95%.
As summarized in Table 1, up to now, the existing papers on the diagnosis of gastric polyps are all based on CNN, but only object detection is performed; the function of semantic segmentation has not been available yet. Moreover, the evaluation metrics are solitary without adequate consideration of clinical requirements. In addition to accuracy, metrics of other aspects (e.g., speed and number of model parameters) should also be considered to evaluate the performance of the model.
Likewise, Zhang et al. [23] introduce a lightweight transformer detection head and a plain FPN into the original YOLOX, which obtain a precision of 62% for erosion and a precision of 77% for ulcers in their own dataset that contains 2074 images. But their model obtains a precision of 99.72% in colon polyp detection in the public dataset. YOLOX is robust in multiple-object-tracking tasks and has a very competitive inference speed. YOLOX is a new generation of high performance and speed in the field of object detection, which shows great potential in gastric polyp detection and other lesion detection in endoscopy.

2.2. MCDM Methods

Numerous MDCM methods have been proposed in the literature and widely applied in different fields since 1970 [24]. MCDM methods are usually designed for specific cases and have their pros and cons. Some typical MCDM methods are reviewed for deeper comprehension. The Analytic Hierarchy Process (AHP) is well-known and widely used [25]. It is a structured technique for organizing and analyzing complex decisions based on mathematics and psychology. Another technique called the analytic network process method is available. This method is generalized from AHP and allows modeling the interaction and dependencies between criteria [26]. In [27], a decision-making trial and evaluation laboratory approach was discussed, which analyzes the interdependent relationships among factors in a complex system and ranks them for long-term strategic decisions made by considering experts’ judgments. In addition, a method called Technique for Order Preference by Similarity (TOPSIS) is also available [28]. TOPSIS uses the Euclidean distance to determine the relative proximity of an alternative to the optimal solution. An alternative priority order can be determined by comparing relative proximity. In the above MDCM methods, every method shows a unique specialty and different advantages. This article utilizes one of them to evaluate the CNN models.

3. Materials and Methods

3.1. Dataset

Public datasets for gastric polyps are lacking. The sole public dataset offered by Zhang et al. [18] only contains 354 gastric polyp images. The number is insufficient, so we create our gastric polyp dataset to train and test the semantic segmentation models. This single-center research was approved by the ethics committees of the Xiangyang Center Hospital (XCH) and followed the Declaration of Helsinki [29]. As a retrospective study, written informed consent was waived. The patient’s information (e.g., name, I.D. number, etc.) was removed from the original EGD images to maintain confidentiality. The endoscopists searched the cases containing the diagnosis of gastric polyps in the medical record database of the endoscopy center from January 2017 to January 2021. From 1428 patients, 10,596 endoscopic images were initially collected and screened by a junior endoscopist. The original images underwent further screening to obtain high-quality images. The exclusion criteria are as follows:
(1)
Images that used endoscopic optics other than standard white-light endoscopy.
(2)
The anatomical position of the image is not in the stomach (like the esophagus).
(3)
Images that contain no polyps.
(4)
Images that are damaged and low-quality due to halation, mucous, blurring, lack of focus, low insufflation of air, et cetera.
After two rounds of selection and rechecking, 487 images from 148 patients were acquired. Their sizes vary from 1296 × 1084 pixels to 356 × 302 pixels since they were taken on different devices. To ensure the labels of polyps were correct, annotation was performed by two endoscopists using Lebelme software. Endoscopists outline the edges of the gastric polyps in the endoscopic images. After processing, the corresponding ground-truth masks are generated, which are the visual representation of manual segmentation. In the masks, all pixels are classified into two categories. As shown in Figure 1, the labels of all pixels in the red part of the masks are the gastric polyps, and the labels of the pixels in the remaining part are the background. The dataset is divided into three sections with a ratio of 8:1:1, which are used for training, validation, and testing.
It is noteworthy that the endoscopic images in the dataset all contain polyps. Normal stomach endoscopic frames are the main part of most live endoscopic videos, which are excluded, leading to a sharp plummet in the number of images in the dataset. The reason why normal stomach endoscopic images are excluded is due to the performance of the semantic segmentation models and clinical consideration. For model performance, worse generalization will occur if the training images contain normal stomach endoscopic images [30]. For clinical considerations, a higher false positive (FP) rate is more desirable than a high false negative (FN) rate since the system’s sensitivity to polyps can help the endoscopists focus more on the suspicious lesions. In terms of the CAD system, the benefit of a misdiagnosis outweighs the benefit of a missed diagnosis [30].

3.2. Existing Semantic Segmentation Methods

Various semantic segmentation algorithms have proven excellent results using CNN in natural images [14]. Generally, CNN can be trained to learn a mapping from original images to ground-truth masks through successive operations, such as convolution, pooling, and upsampling. In this paper, some classical semantic segmentation models, including U-Net, UNet++, DeepLabv3, DeepLabv3+, PAN, LinkNet, and MA-Net, are applied with different encoders to accomplish end-to-end semantic segmentation of gastric polyps [31,32,33,34,35,36,37]. To avoid a dull description, a general U-Net is utilized to explain the idea of most semantic segmentation models, and its architecture is shown in Figure 2. The encoder will usually decrease the spatial resolution and extract the image features, in which convolution and max-pooling are essential to achieve feature extraction, while the symmetric decoder will upsample to the original input resolution and result in low-dimension predictions. In the symmetric decoder, the feature maps are concatenated from the upsampling and the skip connection. The skip connections in U-Net are valid and innovative, increasing and compensating for the semantic information from upsampling.
In other models, the essence of U-Net remains, but with varying constructions, convolution operations, upsampling methods, or combinations with others, such as deep supervision and attention mechanisms. Among the various models adopted in this study, as summarized in Table 2, each has its own strengths and weaknesses. U-Net takes a simple structure and consumes a small amount of training data, but an unchanged structure makes it difficult to find the optimal structure. UNet++ takes advantage of redesigned skip pathways and deep supervision, which improve the segmentation quality of varying-size objects, but it is time-consuming and easily prone to error. Both Deeplabv3 and Deeplabv3+ suffer from the gridding effect problem, though Deeplabv3+ can capture smaller objects and has a more defined boundary. PAN is light and has a low computational cost, but it is difficult to locate small objects precisely. LinkNet is real-time-oriented but requires a large amount of training data and produces inaccurate results. Although MA-Net can obtain good performance in many cases, it is too complex and time-consuming [31,32,33,34,35,36,37]. Given the different strengths and weaknesses of each model, extensive experiments are necessarily required to select the optimal model.

3.3. Integrated Evaluation Approach

Each model has its own potential and drawbacks. It is arduous to select the most optimal model for application in the context of multiple criteria, especially when some are mutually exclusive. Thus, it is necessary to propose a specific evaluation approach to evaluate the performance of semantic segmentation comprehensively.
Complex systematic metrics based on the MCDM method are newly proposed. Figure 3 shows the dendrogram of systematic metrics. The dimensions of first-level metrics are segmentation accuracy and computational efficiency. For segmentation accuracy, Intersection over Union (IoU), accuracy, recall (i.e., sensitivity), precision, and F1-score were utilized as the specific metrics following the recommendation from [38]. The thresholds are all set to 0.5. The experiments belong to pixel–binary classification tasks. According to the realistic category and prediction results, pixel results are divided into true positive (TP), FP, true negative (TN), and FN. IoU calculates the degree of similarity between the predicted masks and the ground-truth masks. IoU varies from 0 to 1 (0–100%), with 0 or 0% indicating no overlap (i.e., the worst) and 1 or 100% indicating perfect segmentation. Accuracy is the percentage of pixels categorized correctly in the endoscopic image. Recall measures the proportion of the pixels labeled as polyps correctly recognized. Precision measures the proportion of the pixels correctly labeled as polyps. The F1-score is the harmonic mean of accuracy and recall. Table 3 shows the calculation equations for the above metrics.
Computational efficiency is a property of an algorithm that refers to the number of computational resources and running time used by the algorithm, divided into the third-level metrics of model complexity and detection speed [39]. To measure the complexity of a model, the number of model parameters and multiply–accumulate operations (MACs) are used. The model parameters in Figure 3 are the sum of the numbers of weights and biases on CNN, which stands for the number of parameters that will be learned during forward inference and backward learning. MACs compute the product of two numbers and add that product to an accumulator physically [40]. A MAC is one multiplication and one addition, which can each be a floating point operation (FLOP), i.e., 1 MAC counts as roughly 2 FLOPs. Most modern hardware architectures use the fused multiply–add instruction for operations with tensors, and MACs follow the instruction directly [40]. As a result, MACs are preferred over FLOPs in this paper. To evaluate detection speed, frames per second (FPS) are calculated by using the mean inference time per image in the test dataset [41].
In MCDM, the rank is determined by the overall weighted score, which will help determine the most optimal model. Hence, an integrated evaluation approach is presented that takes the quantitative and qualitative analyses in the context of MCDM into account. The integrated evaluation approach combines an objective method with a subjective method. Diakoulaki et al. [42] proposed the method entitled Criteria Importance Through Intercriteria Correlation (CRITIC) for the verification of objective weights of comparative importance in financial analysis. In comparison to other objective methods, the CRITIC method, by incorporating both contrast and intensity conflict, provides perception into the quiddity of the quandary generated in the structure of the MCDM problem. It gives the ideal values of cost and benefit criteria concurrently to normalize the decision matrix, but the other methods handle them individually. Moreover, it is the first method that calculates the similarity of the criteria with the correlation coefficient among the criteria. After applying the CRITIC method to the metrics, the objective weights and corresponding objective scores of the models are obtained. The CRITIC method consists of the following steps:
There are n models to be evaluated and p metrics to form the original index matrix:
X = x 11 x 1 p x n 1 x n p
where x i j means the value of the j th metric of the i th model. Dimensionless processing should be used for all metrics to reduce the influence of different dimensions on evaluation outcomes.
b e n e f i t m e t r i c : x i j = x j x m i n x m a x x m i n c o s t m e t r i c : x i j = = x m a x x j x m a x x m i n
The contrast intensity of the corresponding criterion can be expressed as standard deviation σ j . Conflict between factors can be expressed as a correlation coefficient, which is calculated using Equation (3).
R j = i = 1 P 1 r i j
where r i j is the Spearman rank correlation coefficient. The amount of information C j emitted by the j th metric can be determined by composing the measures that quantify the two notions using Equation (4).
C j = σ j i = 1 P 1 r i j = σ j R j
Objective weight is calculated by normalizing these values to unity using Equation (5).
ω j = C j j = 1 n C j
So, the CRITIC score is obtained using Equation (6).
S i = j = 1 n ω j x i j
The objective method is not totally perfect because the expert’s experience always plays a crucial role in clinical treatment. Experienced experts tend to be instinctive in problem-solving and more efficient at dealing with complex emergencies, and patients intend to trust experienced experts more as a result. The subjective weighting of experts should be used in the selection of models. The weights of the subject method are derived from experienced endoscopists from XCH and some medical doctors in Mainland China and Macau, as shown in Table 4. The coefficient of each metric means the relative importance of all metrics, which depends mostly on clinical needs. IoU and FPS occupy the largest proportion because making a clinical diagnosis accurately and quickly is a fundamental requirement for endoscopists, as is the automated polyp-segmentation system.
Therfore, the scores of the subjective method are calculated as follows:
S i , s u b = x i , I o U × 0.3 + x i , A C C × 0.05 + x i , R E × 0.05 + x i , P R × 0.05 + x i , F 1 × 0.05 + x i , P a r a m s × 0.1 + x i , M A C × 0.1 + x i , F P S × 0.3
where S i , s u b denotes the subjective score of the models. x i , I o U , x i , A C C , x i , R E , x i , P R , x i , F 1 , x i , P a r a m s , x i , M A C and x i , F P S denotes normalized results of IoU, accuracy, recall, precision, F1-score, number of parameters, number of MACs, and FPS, respectively.
The final quantitative score is obtained by combining the subjective and objective evaluation methods. It is calculated as follows:
S f i n a l = δ S i + μ S i , s u b ; δ + μ = 1
where S i stands for the CRITIC score. S f i n a l denotes the final score. The coefficients δ and μ are used to determine how much importance should be assigned to objective and subjective methods. δ = μ = 0.5 is employed in our research for comprehensive consideration. Therefore, the best semantic segmentation model can be settled by rank, and the “best score” model is finally selected as the core algorithm of the proposed automated polyp-segmentation system.

4. Experiments and Results

4.1. Experimental Configuration

Transfer learning is flexible and robust; a pre-trained model can be extracted as the weights for a model on a new task. All semantic segmentation models were pre-trained on ImageNet to improve their prior knowledge [43]. The hyperparameters were defined empirically. The Adam optimization algorithm was employed for the optimization of the network with a learning rate of 0.0008 and a batch of 4 images with 80 epochs [44]. Random rotation, random horizontal flipping, and random vertical flipping were utilized for data augmentation, while the input resolution was resized to 256 × 256 using the bicubic interpolation method [45]. The loss function is of paramount importance because it triggers the backward learning process in model training. Some loss functions are designed to alleviate the problem of data imbalance, of which the Dice loss function is a typical one. Dice loss is a kind of region-based loss, which means that the loss value and gradient of a pixel are not only related to the label and predicted value of this pixel but also related to the labels and predicted values of other pixels. As the Dice loss function tends to dig for polyp areas, it is selected as the loss function in the training process [46]. ResNet50, MobileNetV2, and Efficient-B1 were used as the encoders due to their powerful feature extraction ability and inference speed [47,48,49]. Table 5 lists the specifications of the workstation and environmental configurations. To ensure fair-minded experiments, all of the models were trained on the same workstation with the same hyperparameters.

4.2. Results

Feature extraction networks are decisive for pixel classification in semantic segmentation [50]. Therefore, three different advanced CNNs (i.e., ResNet50, MobileNetV2, and EfficieNet-B1) were used as the encoders. Among the ResNet series algorithms, ResNet50 is a fast and accurate approach that is frequently employed as the backbone of deep neural networks. Compared with the deeper networks (e.g., ResNet101 and ResNet152), ResNet50 has a speed advantage. Compared with the shallower networks (e.g., ResNet18 and ResNet34), which consume less storage and computational power, ResNet50 can achieve better training results due to its better feature extraction ability. Therefore, the ResNet50 algorithm was chosen as the encoder based on the requirements of the polyp segmentation task, taking into account both speed and precision [51]. Table 6 shows the quantitative results and evaluation scores for the test dataset in the experiments.
In the class of U-Net in Table 6, the U-Net with the encoder of EfficientNet-B1 obtains the best performance in segmentation accuracy, with an IoU of 96.53%, an accuracy of 98.14%, a recall of 98.16%, a precision of 98.12%, and an F1-score of 98.14%, respectively. However, its detection speed is 22 FPS, which is the slowest in the U-Net experiment. When using ResNet50 as the encoder, the number of model parameters and the number of MACs of U-Net are several times that of the others, which requires more storage and computational power. Moreover, its segmentation accuracy is the worst in this class. The number of model parameters in UNet++ is greater than that of U-Net, but UNet++ models can almost improve segmentation accuracy as compared with U-Net. In the class of UNet++, UNet++ with the encoder of EfficientNet-B1 obtains the best segmentation accuracy, but its detection speed is 21 FPS, which is the slowest. Although UNet++ with the encoder of MobileNetV2 achieves unremarkable performance, it creates a trade-off between segmentation accuracy and computational efficiency. It obtains the highest score in both the CRITIC method and the subjective method. Moreover, its final score not only far outperforms other U-Net++ models but also achieves the best score of all semantic segmentation models in the overall experiment.
DeepLabv3 encoded with ResNet50 obtains the best performance in segmentation accuracy in the class of DeepLabv3. The IoU, accuracy, recall, precision, and F1-score are 96.23%, 97.95%, 97.98%, 97.93%, and 97.96%, respectively. However, it has a large model complexity. DeepLabv3 with the encoder of MobileNetV2 obtains the highest final score of all DeepLabv3 models, which implies that the overall performance is more satisfactory. DeepLabv3+ with the encoder of MobileNetV2 obtains an outstanding result in terms of detection speed, with 27 FPS. However, the IoU, accuracy, recall, precision, and F1-score are 95.22%, 97.36%, 97.42%, 97.32%, and 97.37%, respectively. DeepLabv3+ with the encoder of ResNet50 obtains a competitive score in both the CRITIC method and the subjective method, but its model complexity is a big concern.
In the overall result of Table 6, PAN obtains the best performance in model complexity when using Mobilenet-v2 as the encoder. However, it obtains poor results in segmentation accuracy, with an IoU of 91.83%, an accuracy of 95.09%, a recall of 98.86%, a precision of 92.71%, and an F1-score of 95.52%, respectively. PAN with the encoder of ResNet50 has obvious advantages in segmentation accuracy and detection speed as compared with the other two semantic segmentation models in this class. Although it has a large model complexity, the final score is still the highest out of the three models.
When using EfficientNet-B1 as the encoder, LinkNet obtains the best performance in segmentation accuracy, with an IoU of 96.32%, an accuracy of 98.04%, a recall of 98.08%, a precision of 98.00%, and an F1-score of 98.04%. Although it has a small model complexity, its detection speed is slower than the other two models in this class. LinkNet with the encoder of ResNet50 obtains the highest final score in this class because of its outstanding segmentation accuracy and detection speed.
In the class of MA-Net in Table 6, MA-Net obtains the best performance in segmentation accuracy and model complexity when using EfficineNet-b1 as the encoder. Specifically, the IoU, accuracy, recall, precision, F1-score, number of model parameters, and GMACs are 96.57%, 98.15%, 98.16%, 98.14%, 98.15%, 11.6 M, and 2.41, respectively. However, its detection speed is too slow, resulting in an unremarkable final score for the model. MA-Net with the encoder of ResNet50 and MA-Net with the encoder of EfficineNet-b1 obtain outstanding results in segmentation accuracy, but they show obvious flaws in detection speed.
The semantic segmentation models show their own pros and cons from a qualitative perspective when compared with each other. Due to the simplicity of U-Net’s topology, its GMACs are small but nevertheless produce a decent result in segmentation accuracy. UNet++ strikes a balance between segmentation accuracy and computation efficiency and obtains excellent results in both metrics. DeepLabv3 and DeepLabv3+ yield results that are nearly identical, although DeepLabv3+ has a slightly faster detection speed. Although PAN and LinNet are both lightweight and necessitate fewer computational resources, it seems difficult to locate the polyps precisely, yielding subpar segmentation accuracy. MA-Net is very heavy and expensive in computational resources, although it obtains the best result in segmentation accuracy.
Each model has its own characteristics based on its quantitative metrics. In general, models with high segmentation accuracy are usually complicated to achieve high computational efficiency. Correspondingly, models with high computational efficiency usually tend to have low segmentation accuracy. Model comparison is extremely challenging when there are conflicts between the metrics. The proposed integrated evaluation approach provides a measurable score for comparison. Table 6 shows that UNet++ with the encoder of MobineNet v2 obtains the highest points in the CRITIC score, subjective score, and final score. It means that no matter the aspect of quality, quantity, or a combination of both, the comprehensive performance of the model is the best. In other words, a trade-off between segmentation accuracy and computational efficiency has been achieved. Moreover, the detection speed of 26 FPS shows that the goal of real-time segmentation has been achieved. Figure 4 shows that the Dice loss curve converges quickly and smoothly, which implies that the training process can converge to the global minima with no sign of overfitting or underfitting. The complexity of this model is appropriate for the dataset and is neither overly complex nor overly simplistic. After all considerations, UNet++ with the encoder of MobineNet v2 is chosen as the final model for the automated poly-segmentation system.
Figure 5 shows some segmentation results from the “best score” model. The predicted masks have excellent locations and a fine-grained, well-segmented boundary regardless of small polyps, multiple polyps, or polyps in a dim background. It means that the pixels labeled as gastric polyps can be correctly classified. However, some pixels are misclassified. Figure 6 shows polyps are misclassified (i.e., the red area framed by the yellow box). As previously mentioned, the FP rate increases when the training data sets all contain gastric polyps. The reflective spots and gastric mucosal folds in the endoscopic images can easily be misdiagnosed as gastric polyps. Due to lighting or angles, these regions can indeed be easily misdiagnosed, even with the naked eye. Figure 7 shows that some polyps are missed by the system (i.e., the red area framed by the green box). These missed polyps are small polyps located at the edges of the images. Owing to hardware limitations, all endoscopic images are resized to 256 × 256, which affects the segmentation accuracy of small polyps. Even though there is room for improvement, the accuracy of the selected model is still very satisfactory.
All in all, the automated polyp-segmentation system may be beneficial clinically in three aspects, given its overall excellent performance. The system can play the role of a second observer by segmenting the detected polyps in real time on the monitor adjacent to the main monitor. It could be a useful tool for reducing skill differences among endoscopists and enhancing the quality of routine EGD. However, the final decision still relies on the endoscopist. Nevertheless, it can prompt the endoscopists for further processing, reducing the possibility of missing some polyps with the naked eye. For patients, the proposed system can improve the efficiency of physical examinations and shorten the time required for waiting and examination. Moreover, it can provide expert-level diagnosis, improving patients’ trust, cooperation, and satisfaction. Soon, this system will be open-source and free as compared to commercial software, making it very friendly for medical trainees and junior doctors to train and improve their clinical skills.
However, this system has a variety of limitations. First, it is only capable of segmenting gastric polyps and is not able to identify other gastric diseases simultaneously under EGD. Second, lesions are categorized as benign or malignant most of the time, which is difficult to distinguish under conventional white-light imaging endoscopy, and histopathological prognosis after identification can be more clinically valuable when combined with other advanced endoscopes. Thirdly, this system has not been subjected to multicenter clinical trials with large data sets, and it is still dubious whether it is robust enough to handle a large range of clinical uses.

5. Conclusions

Given the clinical impact of an increased gastric polyp detection rate on the incidence of gastric cancer, an automated polyp-segmentation system based on CNN is developed in this study to assist endoscopists in the diagnosis of gastric polyps, and a novel integrated evaluation approach is also proposed to select a CNN model with the best overall performance. In addition, some state-of-the-art semantic segmentation models are applied and compared, including U-Net, UNet++, DeepLabv3, DeepLabv3+, PAN, LinkNet, and MA-Net. Additionally, several advanced deep network architectures (i.e., ResNet50, MobileNetV2, and EfficientNet-B1) are used as encoders in these models. To evaluate the overall performance of semantic segmentation models, an integrated evaluation approach that combines the objective and subjective methods of MCDM is proposed. Eight metrics (i.e., IoU, accuracy, recall, precision, F1-score, model parameters, MACs, and FPS) are expanded from two dimensions (i.e., segmentation accuracy and computational efficiency). The metrics are condensed into the final scores to rank the candidate models by the CRITIC method and experts’ weight. The evaluation results show that UNet++ with the encoder of MobineNet v2 achieves the best scores in both subjective and objective as well as integrated evaluation approaches, although it is not the best on some single metrics. The proposed system achieves excellent performance in both segmentation accuracy and computational efficiency, contributing to the diagnosis and identification of gastric polyps. The automated polyp-segmentation system and the integrated evaluation approach are enlightening and show great potential in clinical applications. The originality of this study is summarized as follows:
(1)
This study is pioneering research on gastric polyp segmentation. A high-quality gastric polyp dataset is generated. Seven semantic segmentation models are constructed and evaluated to determine the core model for the automated polyp-segmentation system.
(2)
To comprehensively evaluate the results, the integrated evaluation approach combined with the CRITIC method and experts’ weight are combined to rank the candidate models, which is the first attempt at a polyp-segmentation task.
Although the above research results are remarkable, prospective research possibilities and potential enhancements are still needed to have better accuracy and efficiency in the detection of gastric lesions, which are listed as follows: (1) Clinically, the patient’s lesions are often more complex than a single lesion, such as chronic atrophic gastritis with intestinal metaplasia. For EGD, it will be more practical for the CAD systems to detect multiple lesions simultaneously. (2) Not only white-light imaging endoscopy but other more advanced endoscopes, such as narrow-band imaging and chromoendoscopy, should also be considered for the pathological prediction of gastric polyps. Histopathological prediction of small polypoid lesions based on identification allows the endoscopists to decide on a treatment plan once and for all to avoid progressive enlargement of polypoid lesions [52]. (3) Although the complex models show better performance in many cases, the high consumption of memory space and computational resources is an important reason that makes it difficult to implement them on various hardware platforms effectively. Thus, model compression and acceleration are the research directions for the deployment of the system; (4) The proposed integrated evaluation approach should be applied to more semantic segmentation tasks to evaluate its effectiveness, but not only for a one-time-used tool; and (5) The system should then be integrated into endoscopy equipment and used in clinical practice as the next step. The performance of the proposed system should be examined using multi-center and large-scale clinical data.

Author Contributions

Conceptualization, T.Y., P.K.W., Y.H. and S.G.; methodology, T.Y.,Y.Y.Q. and H.R.; software, T.Y. and Y.Y.Q.; validation, H.R., C.H.W., C.I.C., S.G. and P.P.C.; formal analysis, Y.Y.Q., P.K.W., H.R., C.H.W., L.Y., C.I.C., S.G. and P.P.C.; investigation, Y.Y.Q., C.H.W., L.Y. and C.I.C.; resources, H.R., Y.H., S.G. and P.P.C.; data curation, T.Y., H.R.,C.H.W., L.Y. and C.I.C.; writing—original draft preparation, Y.Y.Q.; writing—review and editing, T.Y., C.H.W., L.Y. and C.I.C.; visualization, T.Y. and Y.Y.Q.; supervision, P.K.W.; project administration, P.K.W.; funding acquisition, T.Y, Y.H. and P.K.W.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guangdong Basic and Applied Basic Research Fund, Shenzhen Joint Fund (Guangdong-Shenzhen Joint Fund), Guangdong-Hong Kong-Macau Research Team Project (No. 2021B1515130003), the Science and Technology Development Fund of Macau (No. 0026/2022/A), the Key Research and Development Plan of Hubei Province (No. 2022BCE034), and the Project of Xiangyang Science and Technology on Medical and Health Field (No. 2022YL05A).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of Hubei University of Arts and Science (protocol code: 2021-001; date of approval: 8 March 2021).

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets are available on reasonable requests from corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Fock, K.M.; Talley, N.; Moayyedi, P.; Hunt, R.; Azuma, T.; Sugano, K.; Xiao, S.D.; Lam, S.K.; Goh, K.L.; Chiba, T.; et al. Asia–Pacific consensus guidelines on gastric cancer prevention. J. Gastroenterol. Hepatol. 2008, 23, 351–365. [Google Scholar] [CrossRef] [PubMed]
  3. Goddard, A.F.; Badreldin, R.; Pritchard, D.M.; Walker, M.M.; Warren, B. The Management of gastric polyps. Gut 2010, 59, 1270–1276. [Google Scholar] [CrossRef] [Green Version]
  4. Carmack, S.W.; Genta, R.M.; Schuler, C.M.; Saboorian, M.H. The current spectrum of gastric polyps: A 1-year national study of over 120,000 patients. Am. J. Gastroenterol. 2009, 104, 1524–1532. [Google Scholar] [CrossRef]
  5. Jung, M.K.; Jeon, S.W.; Park, S.Y.; Cho, C.M.; Tak, W.Y.; Kweon, Y.O.; Kim, S.K.; Choi, Y.H.; Bae, H.I. Endoscopic characteristics of gastric adenomas suggesting carcinomatous transformation. Surg. Endosc. 2008, 22, 2705–2711. [Google Scholar] [CrossRef]
  6. Carmack, S.W.; Genta, R.M.; Graham, D.Y.; Lauwers, G.Y. Management of gastric polyps: A pathology-based guide for gastroenterologists. Nat. Rev. Gastroenterol. Hepatol. 2009, 6, 331–341. [Google Scholar] [CrossRef]
  7. Shaib, Y.H.; Rugge, M.; Graham, D.Y.; Genta, R.M. Management of gastric polyps: An endoscopy-based approach. Clin. Gastroenterol. Hepatol. 2013, 11, 1374–1384. [Google Scholar] [CrossRef] [Green Version]
  8. Zheng, B.; Rieder, E.; Cassera, M.A.; Martinec, D.V.; Lee, G.; Panton, O.N.; Park, A.; Swanström, L.L. Quantifying mental workloads of surgeons performing natural orifice transluminal endoscopic surgery (NOTES) procedures. Surg. Endosc. 2012, 26, 1352–1358. [Google Scholar] [CrossRef] [PubMed]
  9. Asfeldt, A.M.; Straume, B.; Paulssen, E.J. Impact of observer variability on the usefulness of endoscopic images for the documentation of upper gastrointestinal endoscopy. Scand. J. Gastroenterol. 2007, 42, 1106–1112. [Google Scholar] [CrossRef]
  10. Wong, P.K.; Chan, I.N.; Yan, H.M.; Gao, S.; Wong, C.H.; Yan, T.; Yao, L.; Hu, Y.; Wang, Z.R.; Yu, H.H. Deep learning based radiomics for gastrointestinal cancer diagnosis and treatment: A minireview. World J. Gastroenterol. 2022, 28, 6363–6379. [Google Scholar] [CrossRef] [PubMed]
  11. Rees, C.J.; Koo, S. Artificial Intelligence—Upping the game in gastrointestinal endoscopy? Nat. Rev. Gastroenterol. Hepatol. 2019, 16, 584–585. [Google Scholar] [CrossRef] [PubMed]
  12. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  13. Wang, J.; Liu, X. Medical image recognition and segmentation of pathological slices of gastric cancer based on Deeplab v3+neural network. Comput. Methods Programs Biomed. 2021, 207, 106210. [Google Scholar] [CrossRef] [PubMed]
  14. Hao, S.J.; Zhou, Y.; Guo, Y.R. A brief survey on semantic segmentation with deep learning. Neurocomputing 2020, 406, 302–321. [Google Scholar] [CrossRef]
  15. Corral, J.E.; Keihanian, T.; Diaz, L.I.; Morgan, D.R.; Sussman, D.A. Management patterns of gastric polyps in the United States. Frontline Gastroenterol. 2019, 10, 16–23. [Google Scholar] [CrossRef]
  16. Zionts, S. MCDM—If not a roman numeral, then what? Interfaces 1979, 9, 94–101. [Google Scholar] [CrossRef]
  17. Sitorus, F.; Brito-Parada, P.R. The selection of renewable energy technologies using a hybrid subjective and objective multiple criteria decision making method. Expert Syst. Appl. 2022, 206, 117839. [Google Scholar] [CrossRef]
  18. Zhang, X.; Chen, F.; Yu, T.; An, J.; Huang, Z.; Liu, J.; Hu, W.; Wang, L.; Duan, H.; Si, J. Real-time gastric polyp detection using convolutional neural networks. PLoS ONE 2019, 14, e0214133. [Google Scholar] [CrossRef] [Green Version]
  19. Laddha, M.; Jindal, S.; Wojciechowski, J. Gastric polyp detection using deep convolutional neural network. In Proceedings of the 2019 4th International Conference on Biomedical Imaging, Nagoya, Japan, 17–19 October 2019; pp. 55–59. [Google Scholar] [CrossRef]
  20. Wang, R.; Zhang, W.; Nie, W.; Yu, Y. Gastric polyps detection by improved faster R-CNN. In Proceedings of the 2019 8th International Conference on Computing and Pattern Recognition, Beijing, China, 23–25 October 2019; pp. 128–133. [Google Scholar] [CrossRef]
  21. Cao, C.; Wang, R.; Yu, Y.; Zhang, H.; Yu, Y.; Sun, C. Gastric polyp detection in gastroscopic images using deep neural network. PLoS ONE 2021, 16, e0250632. [Google Scholar] [CrossRef]
  22. Durak, S.; Bayram, B.; Bakırman, T.; Erkut, M.; Doğan, M.; Gürtürk, M.; Akpınar, B. Deep neural network approaches for detecting gastric polyps in endoscopic images. Med. Biol. Eng. Comput. 2021, 59, 1563–1574. [Google Scholar] [CrossRef]
  23. Zhang, Z.; Chen, Q.; Chen, S.; Liu, X.; Cao, Y.; Liu, B.; Zhang, H. Automatic disease detection in endoscopy with light weight transformer. Smart Health 2023, 28, 100393. [Google Scholar] [CrossRef]
  24. Triantaphyllou, E. Multi-criteria decision making methods. In Multi-Criteria Decision Making Methods: A Comparative Study: Applied Optimization; Springer: Boston, MA, USA, 2000; pp. 5–21. [Google Scholar]
  25. Saaty, R.W. The analytic hierarchy process—What it is and how it is used. Math. Model. 1987, 9, 161–176. [Google Scholar] [CrossRef] [Green Version]
  26. Shafaghat, A.; Keyvanfar, A.; Ket, C.W. A decision support tool for evaluating the wildlife corridor design and conservation performance using analytic network process (ANP). J. Nat. Conserv. 2022, 70, 126280. [Google Scholar] [CrossRef]
  27. Akter, S.; Debnath, B.; Bari, A.B.M.M. A grey decision-making trial and evaluation laboratory approach for evaluating the disruption risk factors in the emergency life-saving drugs supply chains. Healthc. Anal. 2022, 2, 100120. [Google Scholar] [CrossRef]
  28. Behzadian, M.; Otaghsara, S.K.; Yazdani, M.; Ignatiusc, J. A state-of-the-art survey of TOPSIS applications. Expert Syst. Appl. 2012, 39, 13051–13069. [Google Scholar] [CrossRef]
  29. Goodyear, M.D.; Krleza-Jeric, K.; Lemmens, T. The declaration of Helsinki. BMJ 2007, 335, 624–625. [Google Scholar] [CrossRef]
  30. Yeung, M.; Sala, E.; Schönlieb, C.B.; Rundo, L. Focus U-Net: A novel dual attention-gated CNN for polyp segmentation during colonoscopy. Comput. Biol. Med. 2021, 137, 104815. [Google Scholar] [CrossRef]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar] [CrossRef]
  32. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 2020, 39, 1856–1867. [Google Scholar] [CrossRef]
  33. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  34. Chen, L.C.; Zhu, Y.K.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with Atrous separable convolution for semantic image segmentation. arXiv 2018, arXiv:1802.02611. [Google Scholar] [CrossRef]
  35. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. arXiv 2018, arXiv:1805.10180. [Google Scholar] [CrossRef]
  36. Chaurasia, A.; Culurciello, E. LinkNet: Exploiting encoder representations for efficient semantic segmentation. arXiv 2017, arXiv:1707.03718. [Google Scholar] [CrossRef]
  37. Fan, T.L.; Wang, G.L.; Li, Y.; Wang, H.R. MA-Net: A multi-scale attention network for liver and tumor segmentation. IEEE Access 2020, 8, 179656–179665. [Google Scholar] [CrossRef]
  38. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Halvorsen, P.; Lange, T.; Johansen, D.; Johansen, H.D. Kvasir-seg: A segmented polyp dataset. arXiv 2014, arXiv:1911.07069. [Google Scholar] [CrossRef]
  39. Livni, R.; Shalev-Shwartz, S.; Shamir, O. On the computational efficiency of training neural networks. arXiv 2014, arXiv:1410.1141. [Google Scholar] [CrossRef]
  40. Lyakhov, P.; Valueva, M.; Valuev, G.; Nagornov, N. A method of increasing digital filter performance based on truncated multiply-accumulate units. Appl. Sci. 2020, 10, 9052. [Google Scholar] [CrossRef]
  41. Molchanov, P.; Tyree, S.; Karras, T.; Aila, T.; Kautz, J. Pruning convolutional neural networks for resource efficient inference. arXiv 2016, arXiv:1611.06440. [Google Scholar] [CrossRef]
  42. Diakoulaki, D.; Mavrotas, G.; Papayannakis, L. Determining objective weights in multiple criteria problems—The CRITIC method. Comput. Oper. Res. 1995, 22, 763–770. [Google Scholar] [CrossRef]
  43. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  44. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  45. Keys, R.G. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef] [Green Version]
  46. Jadon, S. A survey of Loss functions for semantic segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology, Via del Mar, Chile, 27–29 October 2020; pp. 115–121. [Google Scholar] [CrossRef]
  47. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  48. Sandler, M.; Howard, A.; Zhu, M.L.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  49. Tan, M.X.; Le, Q.V. EfficientNet: Rethinking model scaling for convolutional neural networks. arXiv 2019, arXiv:1905.11946. [Google Scholar] [CrossRef]
  50. Chowdhary, C.L.; Acharjya, D.P. Segmentation and feature extraction in medical imaging: A systematic review. Procedia Comput. Sci. 2020, 167, 26–36. [Google Scholar] [CrossRef]
  51. Sarwinda, D.; Paradisa, R.H.; Bustamam, A.; Anggia, P. Deep Learning in Image Classification using Residual Network (ResNet) Variants for Detection of Colorectal Cancer. Procedia Comput. Sci. 2021, 179, 423–431. [Google Scholar] [CrossRef]
  52. Mihara, M.; Yasuo, T.; Kitaya, K. Precision Medicine for Chronic Endometritis: Computer-Aided Diagnosis Using Deep Learning Model. Diagnostics 2023, 13, 936. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Original images of gastric polyps (i.e., (ah)) and corresponding ground-truth masks (i.e., (AH)) in the dataset.
Figure 1. Original images of gastric polyps (i.e., (ah)) and corresponding ground-truth masks (i.e., (AH)) in the dataset.
Bioengineering 10 00806 g001
Figure 2. Architecture of U-Net.
Figure 2. Architecture of U-Net.
Bioengineering 10 00806 g002
Figure 3. Systematic design of evaluation metrics.
Figure 3. Systematic design of evaluation metrics.
Bioengineering 10 00806 g003
Figure 4. Dice loss curve of U-Net++ with different encoders. (a) Dice loss curve of U-Net++ with the encoder of ResNet50; (b) Dice loss curve of U-Net++ with the encoder of MobineNet v2; (c) Dice loss curve of U-Net++ with the encoder of EfficientNet-B1.
Figure 4. Dice loss curve of U-Net++ with different encoders. (a) Dice loss curve of U-Net++ with the encoder of ResNet50; (b) Dice loss curve of U-Net++ with the encoder of MobineNet v2; (c) Dice loss curve of U-Net++ with the encoder of EfficientNet-B1.
Bioengineering 10 00806 g004
Figure 5. Sample results of the proposed automated poly-segmentation system.
Figure 5. Sample results of the proposed automated poly-segmentation system.
Bioengineering 10 00806 g005
Figure 6. FP examples of the proposed automated poly-segmentation system.
Figure 6. FP examples of the proposed automated poly-segmentation system.
Bioengineering 10 00806 g006
Figure 7. FN examples of the proposed automated poly-segmentation system.
Figure 7. FN examples of the proposed automated poly-segmentation system.
Bioengineering 10 00806 g007
Table 1. Constraints of the extant literature on gastric polyp segmentation employing CNN.
Table 1. Constraints of the extant literature on gastric polyp segmentation employing CNN.
ReferenceDatasetObjectiveBaselineConstraint
Zhang et al. [18]404 imagesReal-time detection of gastric polypsSSD1. Most of the research focused on object detection, which is unable to portray the boundaries of polyps precisely.
2. The evaluation metrics are solitary without adequate consideration of clinical requirements.
Laddha et al. [19]654 imagesDetection of gastric polypsYOLOv3 and YOLOv3-tiny
Wang et al. [20]1941 imagesDetection of gastric polypsFaster R-CNN
Cao et al. [21] 2270 imagesStomach classification and detection of gastric polypsYOLOv3
Durak et al. [22]2195 imageDetection of gastric polypsYOLOv4, CenterNet,
EfficientNet, Cross Stage ResNext50-SPP, YOLOv3, YOLOv3-SPP, Single Shot Detection, and Faster Regional CNN
Table 2. Strengths and weaknesses of each model.
Table 2. Strengths and weaknesses of each model.
ModelStrengthWeakness
U-Net
  • Simple structure
  • Less training data needed
  • Unknown depth of optimal architecture.
  • Unnecessarily restrictive fusion scheme at the same-scale feature maps.
UNet++
  • A highly flexible feature fusion scheme
  • Enhancement of segmentation quality of varying-size objects
  • Long time-consuming
  • Easily prone to errors
DeepLabv3
  • A more general framework
  • It can segment objects effectively at muti-scale.
  • Gridding Effect Problem
  • Long-ranged information might be not relevant.
DeepLabv3+
  • Multiple effective field-of-view
  • Results of shaper object boundary
  • Gridding Effect Problem
  • Bilinear upsample cannot restore lost semantic information
PAN
  • It exploits global contextual information well
  • Light and small computation
  • It needs numerous training data
  • Difficult in locating small objects precisely.
LinkNet
  • Real-time oriented
  • Simple structure and small computation
  • It needs numerous training data
  • Lack of accuracy
MA-Net
  • It can capture rich contextual dependencies.
  • Good performance
  • Heavy and expensive in computational resources
  • Time-consuming
Table 3. Equations for metrics of segmentation accuracy.
Table 3. Equations for metrics of segmentation accuracy.
MetricsEquations
IoU I o U = | A B | | A B | = | A B | A + B | A B | = T P T P + F P + F N
ACC T P + T N T P + T N + F P + F N
RE T P T P + F N
PR T P T P + F P
F1 2 × P R × R E P R + R E
Remark: ACC stands for accuracy, RE for recall, PR for precision, and F1 stands for F1-score.
Table 4. Weights of metrics.
Table 4. Weights of metrics.
First-Level MetricsSecond-Level MetricsWeights
Segmentation accuracyIoU0.3
Accuracy0.05
Recall0.05
Precision0.05
F1-Score0.05
Computational efficiencyNumber of parameters0.1
Number of MACs0.1
FPS0.3
Sum of weight1
Table 5. Workstation hardware and environmental configurations.
Table 5. Workstation hardware and environmental configurations.
ConfigurationVersion
CPU11th Gen Intel(R) Core (TM) i9-11900 @ 2.50 GHz
GPUNVIDIA GeForce RTX 3080
RAM64.0 GB
Operating SystemWindows 10
Programing LanguagePython 3.9
FramePytorch-1.10.0
CUDA11.4.1
cuDNN11.4
Table 6. Test results based on the test dataset.
Table 6. Test results based on the test dataset.
ModelEncoderIoU
(%)
ACC
(%)
RE
(%)
PR
(%)
F1
(%)
No. of Parameters
(M)
GMACsFPSCRITIC
Score
Subjective ScoreFinal
Score
U-NetResNet5094.9697.2997.3197.2797.2932.5210.7240.590.630.61
MobileNet v295.5697.5697.5797.5597.566.633.39260.750.810.78
EfficientNet-B196.5398.1498.1698.1298.148.762.53220.770.740.75
UNet++ResNet5096.5798.1898.2098.1698.1848.9957.54200.530.530.53
MobileNet v296.2798.0098.0397.9898.006.824.5260.840.880.86
EfficientNet-B196.7998.1198.1498.0998.119.085.1210.730.700.72
DeepLabv3ResNet5096.2397.9597.9897.9397.9639.6340.99240.650.700.67
MobileNet v295.7997.6997.7397.6697.6912.6512.74260.750.810.78
EfficientNet-B195.4697.3797.4097.3497.379.813.37230.630.650.64
DeepLabv3+ResNet5096.2597.9798.0297.9397.9726.689.2260.810.860.83
MobileNet v295.2297.3697.4297.3297.374.381.52270.750.820.78
EfficientNet-B195.9397.7597.8397.7097.767.410.56230.720.720.72
PANResNet5096.0697.8798.0097.7897.898.7124.26260.770.820.80
MobileNet v295.1797.2797.5097.1497.312.420.79260.720.770.75
EfficientNet-B191.8395.0998.8692.7195.526.60.09220.520.330.43
LinkNetResNet5096.3198.0398.0398.0398.0331.1810.77260.810.860.83
MobileNet v295.1197.3297.3397.3297.324.320.94260.710.770.74
EfficientNet-B196.3298.0498.0898.0098.043.670.19220.760.720.74
MA-NetResNet5096.3498.0398.0698.0198.03147.4418.64210.540.550.55
MobileNet v296.2397.9898.0097.9697.9848.895.27240.740.760.75
EfficientNet-B196.5798.1598.1698.1498.1511.62.41210.740.700.72
Remark: Boldface number means the best for each metric or score. ACC indicates accuracy, RE indicates recall, PR indicates precision, F1 indicates F1-score; Params indicates the number of model parameters in M. (1 M = 1 × 106); GMACs indicates 1 × 109 MACs; Subjective score indicates that the score is calculated by the subjective evaluation method.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, T.; Qin, Y.Y.; Wong, P.K.; Ren, H.; Wong, C.H.; Yao, L.; Hu, Y.; Chan, C.I.; Gao, S.; Chan, P.P. Semantic Segmentation of Gastric Polyps in Endoscopic Images Based on Convolutional Neural Networks and an Integrated Evaluation Approach. Bioengineering 2023, 10, 806. https://doi.org/10.3390/bioengineering10070806

AMA Style

Yan T, Qin YY, Wong PK, Ren H, Wong CH, Yao L, Hu Y, Chan CI, Gao S, Chan PP. Semantic Segmentation of Gastric Polyps in Endoscopic Images Based on Convolutional Neural Networks and an Integrated Evaluation Approach. Bioengineering. 2023; 10(7):806. https://doi.org/10.3390/bioengineering10070806

Chicago/Turabian Style

Yan, Tao, Ye Ying Qin, Pak Kin Wong, Hao Ren, Chi Hong Wong, Liang Yao, Ying Hu, Cheok I Chan, Shan Gao, and Pui Pun Chan. 2023. "Semantic Segmentation of Gastric Polyps in Endoscopic Images Based on Convolutional Neural Networks and an Integrated Evaluation Approach" Bioengineering 10, no. 7: 806. https://doi.org/10.3390/bioengineering10070806

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop