Next Article in Journal
Analysis of Trends in Mega-Sized Container Ships Using the K-Means Clustering Algorithm
Previous Article in Journal
Effect of Temperatures on Polyphenols during Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

COMMA: Propagating Complementary Multi-Level Aggregation Network for Polyp Segmentation

School of Industrial and Management Engineering, Korea University, 145, Anam-Ro, Seongbuk-Gu, Seoul 02841, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(4), 2114; https://doi.org/10.3390/app12042114
Submission received: 6 January 2022 / Revised: 13 February 2022 / Accepted: 16 February 2022 / Published: 17 February 2022
(This article belongs to the Topic Medical Image Analysis)

Abstract

:
Colonoscopy is an effective method for detecting polyps to prevent colon cancer. Existing studies have achieved satisfactory polyp detection performance by aggregating low-level boundary and high-level region information in convolutional neural networks (CNNs) for precise polyp segmentation in colonoscopy images. However, multi-level aggregation provides limited polyp segmentation owing to the distribution discrepancy that occurs when integrating different layer representations. To address this problem, previous studies have employed complementary low- and high- level representations. In contrast to existing methods, we focus on propagating complementary information such that the complementary low-level explicit boundary with abstracted high-level representations diminishes the discrepancy. This study proposes COMMA, which propagates complementary multi-level aggregation to reduce distribution discrepancies. COMMA comprises a complementary masking module (CMM) and a boundary propagation module (BPM) as a multi-decoder. The CMM masks the low-level boundary noises through the abstracted high-level representation and leverages the masked information at both levels. Similarly, the BPM incorporates the lowest- and highest-level representations to obtain explicit boundary information and propagates the boundary to the CMMs to improve polyp detection. CMMs can discriminate polyps more elaborately than prior CMMs based on boundary and complementary representations. Moreover, we propose a hybrid loss function to mitigate class imbalance and noisy annotations in polyp segmentation. To evaluate the COMMA performance, we conducted experiments on five benchmark datasets using five metrics. The results proved that the proposed network outperforms state-of-the-art methods in terms of all datasets. Specifically, COMMA improved mIoU performance by 0.043 on average for all datasets compared to the existing state-of-the-art methods.

1. Introduction

Colorectal cancer (CRC), which is one of the most common cancers globally, usually begins as a polyp in the colon mucosa, and approximately one-quarter of untreated polyps can develop into colon cancer [1]. Early polyp detection is a significant task in preventing CRC, and colonoscopy is used extensively as a standard polyp detection method [2,3,4]. Although colonoscopy is an effective method for detecting polyps at the early stages, polyp detection using colonoscopy images is a challenging task owing to the ambiguous image context. As polyps are usually small and their boundaries are low in contrast to their surroundings, polyps can easily be mistaken for wrinkles or other intestinal structures, leading to inaccurate segmentation and over-segmentation. Therefore, discrimination of the precise polyp region from an ambiguous context is critical for improving early polyp detection and preventing CRC.
Based on the need for elaborate segmentation, early studies utilized handcrafted features with a classifier [5,6]. However, the handcrafted approaches suffer from unsatisfactory performance because they cannot cover both intra- and inter-class variations [7]. Recent studies have proposed deep learning-based approaches, including fully convolutional networks (FCNs) [8] and U-Net [9]. Moreover, as alternative approach, Mask R-CNN [10] based models have also been proposed for precise polyp segmentation [11]. In contrast to existing CNN-based networks, PNS-Net [12] introduced a normalized self-attention network that dynamically operated the receptive field of the network. Along with the network architecture, a recent approach [13] considered excluding the effect of colors because polyp images are gathered under varying conditions. By alleviating discrepancies in color distribution, they improved the generalization performance of the network.
The U-Net comprises encoder-decoder structures that aggregate multiple encoder representations at the decoder to overcome the insufficient representations of the handcrafted approaches. Although the encoder-decoder structure improves the polyp segmentation performance, this structure exhibits a distribution discrepancy between the low- and high-level representations when aggregating multi-level features. Because the low-level representation contains sufficient boundary information with background noise, whereas the high-level feature presents abstracted region information [14,15,16]. To overcome the drawbacks of both representations, each level requires the elimination of noise and an emphasis on boundary information. Without the denoising and refining fine information, the multi-level aggregation discrepancy occurs and it generates rough boundaries with noise in the prediction map, which limits the model performance. To reduce this discrepancy, U-Net++ [17] proposed multiple convolution layers on the skip pathways between the encoder and decoder features. A previous study [18] improved on existing U-Net++-based architecture by applying conditional random field and test-time augmentation to the ResUNet++ [19] for precise polyp segmentation. Focusing more on the multi-level aggregation problem, previous studies [20,21] organized the network including selective feature extraction and aggregation using multiple kernel sizes. In addition, Enhanced U-Net [22] introduced attention modules designed to extract distinct features from the highest-level representation with different patch sizes and to refine the encoder features by utilizing the distinct features. Similarly, SANet [13] proposed fusing both low- and high-level representations to detect fine polyps by excluding background noise in low-level representations with abstracted high-level information. These existing studies contributed to diminishing the multi-level distribution discrepancy; however, they did not propagate complementary low- and high-level aggregation to decoder structure to clarify polyp representation. Based on the lack of research on the propagation of complementary multi-level aggregated information, we studied the relationship between low- and high-level representations and focused on the propagation of complementary information.
Another approach is the use of boundary information to compensate for insufficient polyp representation. Psi-Net [23] addressed a joint training strategy using polyp region and boundary detection tasks. Moreover, SFANet [20] incorporated boundary-sensitive loss with boundary deep supervision maps [24] to detect polyp boundaries more elaborately. In terms of boundary refinement, MSBNet [25] utilized a low-level representation with a Gaussian kernel to enhance boundary information at the highest representation. Another study [26] organized multiple boundary attention modules designed to discriminate boundary information using encoder and decoder representations adjacent to each other at each decoder. In PraNet [27], which improved the model efficiency and outperformed existing studies, a parallel reverse attention method with partial decoders was employed to incorporate the polyp area and boundary features [15,28]. Although PraNet considered the region and boundaries, the reverse attention method could not sufficiently discriminate the boundary from the background. Because the boundary is obtained by aggregating the high-level encoder outputs, which present the abstracted region information, the reversed region contains insufficient boundaries compared to low-level outputs. In previous studies, the explicit boundary information can be propagated across the decoders to detect the polyp boundaries more elaborately, although the boundary was either used independently in a single decoder or not proliferated. Therefore, we focused on propagating the explicit boundary by employing a multi-decoder structure.
In medical image segmentation tasks, binary cross entropy (BCE) and IoU losses are commonly used to deal with local and global structural awareness. The BCE and IoU losses cause a class imbalance problem because they treat all pixels with equal importance. To reduce this problem, weighted BCE (wBCE) and IoU (wIoU) losses have been proposed in previous studies on salient object detection tasks [29]. In polyp segmentation, PraNet [27] demonstrated state-of-the-art performance by adopting wBCE and wIoU losses. In addition, we consider an L1 distance loss, which can handle noisy annotations [30,31] often encountered in segmentation tasks. Thus, we combine the wBCE, wIoU, and L1 loss functions to treat class imbalance and noisy label problems.
This study proposes a propagating complementary multi-level aggregation network (COMMA), which comprises a complementary masking module (CMM) and boundary propagation module (BPM) as the multi-decoder structure. The CMM clarifies the boundary noise in the low-level through the abstracted high-level representation and propagates the refined information to another decoder. The BPM generates an explicit boundary by masking the lowest-level representation through the highest-level outputs. The boundary is propagated to the CMMs in the next decoder to enhance the segmentation performance. We also propose a hybrid loss function that allows the network to learn robustly on noisy labels and assigns different importance to each pixel.
The main contributions of this study are as follows: First, we propose complementary multi-level aggregation, which contributes to reducing the multi-level distribution discrepancy by applying the abstracted high-level representation as a mask to the low-level boundary noises and propagating the complementary information. Second, the explicit boundary propagation for the multi-decoder discriminates polyps in an ambiguous context and enhances the segmentation performance. Third, we design a hybrid loss function consisting of weighted BCE, weighted IoU, and L1 distance loss. The hybrid loss function enables the model to focus on relatively important pixels and to be robust against noisy annotations. Fourth, as a novel network, COMMA achieves state-of-the-art performance with a significant improvement in the generalization performance.
The remainder of this study is as follows. Section 2 explains the proposed COMMA architecture for polyp segmentation. Section 3 describes the experimental setup and obtained results. Section 4 presents a discussion of the proposed method. Finally, Section 5 concludes this study and outlines future works.

2. Materials and Methods

In this section, we present COMMA, which is designed to reduce the multi-level distribution discrepancy by propagating both refined levels and explicit boundary information. Figure 1 represents the proposed COMMA architecture, consisting of an encoder block and a decoder block containing the CMM and BPM. The encoder blocks consist of four stages based on ResNet [32] and Res2Net [33], and the CMM extracts a complementary representation by aggregating the encoder output and the previous decoder output. The BPM in the decoder combines the highest- and lowest-level features to extract explicit boundary information. Furthermore, to proliferate distinct information, we employ multi-decoder structures consisting of CMMs and BPM.

2.1. Complementary Masking Module

Based on the relationship between the relatively low- and high-level representations, the CMM masks the background noise at the low-level representation through the abstracted high-level representation. After obtaining the denoised representation, we propagate the complementary information to the decoder representation and the next decoder. Here, we denote the i-th encoder block representation E i ( j ) in the j-th skip connection path, where i { 1 , 2 , 3 , 4 } and j { 1 , 2 } . D i ( j ) indicates the i-th decoder block representation in the j-th decoder, where i { 3 , 4 } and j { 1 , 2 } . In the CMM, the encoder output E i ( j ) R C i × H i × W i is masked by the decoder output D i + 1 ( j ) R C i + 1 × H i + 1 × W i + 1 , as follows:
m a s k = σ ( F ( U p ( D i + 1 ( j ) ) ) ) R 1 × H i × W i C = F ( E i ( j ) ) m a s k
Here, U p ( · ) and F ( · ) indicate the bilinear upsampling and post-activated convolutional operations (i.e., Conv2D-Batch normalization-ReLU) [34], respectively. We apply a dual post-activation convolutional operation to E 4 ( 1 ) for the first decoder representation D 4 ( 1 ) . To exclude the background noise, we apply F ( · ) to the upsampled decoder representation U p ( D i + 1 ( j ) ) to obtain the m a s k that high-level representation is channel-wise aggregated. Subsequently, we employ a sigmoid function σ to discriminate the mask information and eliminate the background noise in E i ( j ) through m a s k . Following the masking, a post-activated convolutional operation is applied to complementary representation C to refine the features in each path of low- and high-level. These features are fused with the original CMM input features ( E i ( j ) and D i + 1 ( j ) ) to propagate the next CMM and decoder as follows:
E i ( j ) = F ( C ) + E i ( j ) , D i ( j ) = F ( C ) + U p ( D i + 1 ( j ) )
For an efficient computation, we include two CMMs in a single decoder. Then, to progressively improve and propagate the complementarity of the output propagated to the next decoder CMM, we designed a cascading multi-decoder structure. We describe an experiment in Section 3.4.1 in which we investigated the effects of progressive improvement in complementarity. Finally, we obtain five deep supervision maps [24]: two decoder path outputs ( D S 1 and D S 2 ) and three skip-connected path outputs ( D S 3 , D S 4 , and D S 5 ). Each deep supervision map is upsampled to the ground truth size to compute the model loss.

2.2. Boundary Propagation Module

Explicit boundary information contributes to a more elaborate discrimination of polyps in ambiguous contexts. Existing studies [25,26] generated edge information by ensembling both low- and relatively high-level representations. They only utilized this information to refine the highest-level feature or adjacent pair of encoder-decoder outputs, rather than the entire network. In contrast to existing methods, the BPM is designed not only to detect explicit boundary information but also to propagate the boundary to complementary multi-level features by incorporating the lowest- and highest-level representations. Although both levels exhibit a large distribution discrepancy, we overcome the discrepancy with a property of each level to obtain a detailed boundary. The low-level encoder output contains sufficient boundary information including background noise; in contrast, the high-level output contains abstracted position information such that it exhibits ambiguous boundaries [14,15,16]. Based on these properties, we consider both the lowest and highest levels to maximize boundary detection performance. To verify the effectiveness of this adoption, we conducted an ablation study related on the representation level for boundary generation in Section 4.2.
We denoise the boundary noise at the lowest-level E 1 ( 1 ) through the highest-level representation D 4 ( 1 ) to obtain a high-quality boundary, as follows:
X = F ( E 1 ( 1 ) ) U p ( D 4 ( 1 ) ) B = σ ( F ( X ) )
A convolutional operation F ( · ) removing background noise [35,36] is applied to E 1 ( 1 ) . D 4 ( 1 ) for the effective aggregation of two different levels is directly rescaled to the same size as E 1 ( 1 ) by the bilinear upsampling U p ( · ) . Subsequently, we refine the fused feature X using F ( · ) and employ a sigmoid function to discriminate the boundary representation B. After generating the explicit boundary B, the BPM propagates downsampled B to complementary representations E i ( 2 ) , which are the CMM outputs of the first decoder, to clarify the fine information, as follows:
E i ( 2 ) = E i ( 2 ) D o w n ( B ) + E i ( 2 ) , w h e r e i = 2 , 3 , 4
Here, D o w n ( · ) indicates bilinear downsampling to the same size as E i ( 2 ) . We emphasize boundary information in E i ( 2 ) using resized B to each E i ( 2 ) . By enhancing the fine features at each E i ( 2 ) , which is the input of the second decoder, the second decoder can more easily detect ambiguous polyps compared to the first. That is, boundary propagation is capable of learning a robust network against ambiguous edge information. To verify the effectiveness of boundary propagation, we present experimental results depending on the BPM application in Section 4.1. Finally, as demonstrated in existing studies [20,23,37], we generate a boundary ground truth G T B to improve the boundary detection performance. To support the explainability of the module, we visualize feature maps corresponding to the BPM application state in Section 4.3.

2.3. Hybrid Loss Function

We employ hybrid loss to constrain the area, boundary, and noisy annotations in the loss function as follows: L = α * L B C E w + β * L I o U w + γ * L L 1 . In medical image segmentation tasks, binary cross entropy (BCE) and IoU losses are adopted extensively to impose local and global constraints. The BCE and IoU losses consider all pixels to make equal contributions [29]. That is, because colonoscopy images contain more background than polyp regions, a class imbalance problem can occur. To prevent this imbalance, we utilize the weighted BCE (wBCE) and IoU (wIoU) losses, which focus on hard pixels, as demonstrated in previous study [29]. The weighted loss uses λ to calculate the importance of each pixel, and λ i j is calculated as follows:
λ i j = h , w A i j y h , w h , w A i j 1 y i j
Here, A i j refers to the area surrounding the target pixel ( i , j ) . The value of λ i j is larger when it is placed at the boundaries of the polyp compared to the center of the polyp. Thus, the weighted loss allows the network to focus on the boundary regions. For example, wBCE loss is given by Equation (6). Here, y and y ^ indicate the ground truth and prediction of each pixel, respectively. The notation c { 0 , 1 } refers to binary classes. γ is a hyper-parameter and is set to 5 as in previous study [29]. In wBCE, the importance of each pixel is assigned using weight λ i j . As a result, the network can focus more on the local structures at the boundaries than using the BCE.
L B C E w = i H j W ( 1 + γ · λ i j ) c = 0 1 ( y c l o g ( y ^ c ) + ( 1 y c ) l o g ( 1 y ^ c ) ) i H j W ( 1 + γ · λ i j )
In contrast, the wIoU loss detects the global structure of the polyp rather than individual pixels based on the pixel importance λ i j .
L I o U w = 1 i H j W ( y i j y ^ i j ) ( 1 + γ · λ i j ) i H j W ( y i j + y ^ i j y i j y ^ i j ) ( 1 + γ · λ i j )
Moreover, we consider the L1 distance loss to be robust to noisy labels [30,31], which often occurs when annotating in segmentation tasks, so that the model converges robustly and reliably on noisy annotations.
L L 1 = 1 H × W i H j W | y i j y ^ i j |
We provide the ablation studies exploring different combinations by calibrating the loss weights (i.e., α , β , and γ ) in Section 4.4.2. Moreover, existing study [27] revealed that explicitly using the dependency of both region and boundary causes an overfitting problem; however, we avoid this issue by applying independent boundary loss and data augmentations. Based on the hybrid loss, the final loss is calculated using the deep supervision maps as follows:
L f i n a l = i = 1 5 L ( G T , D S i ) + L ( G T B , B )
As we combine three different loss functions, we investigate the effectiveness of each and their contributions to the performance improvement in Section 4.4.

3. Results

In this section, we demonstrate the five benchmark datasets, five evaluation metrics, and experimental setups used for as well as the results of evaluating COMMA performance compared to existing methods.

3.1. Dataset

We validated the proposed COMMA using five datasets. Kvasir-SEG [38] is the largest dataset, containing 1000 challenging images for polyp segmentation. We divided Kvasir into 80%, 10%, and 10% for the training, validation, and testing images, respectively, following the experimental settings of the existing study [27]. CVC-ClinicDB (CVC-612) [39] contains 612 images in which the boundaries are low in contrast to their surroundings. In the same manner, we used 550 images (90%) for training and validation and 62 images (10%) for testing. To enable a fair comparison, all test sets were equivalent to those in previous study [27]. CVC-ColonDB [40], ETIS [41], and EndoScene (CVC-T, CVC-300) [42] were used to evaluate the generalization performance, because these small datasets contain 380, 196, and 60 images, respectively.

3.2. Experimental Setup

3.2.1. Evaluation Metrics

To validate the performance of the proposed model, we employed the mean Dice and mean IoU metrics, which are widely used in medical image segmentation tasks. Furthermore, we evaluated three additional metrics (i.e., S m , E m , and MAE) that are commonly used in salient object detection tasks [15,37,43]. The S-measure [44], which evaluates structural similarities, is calculated as follows: S m = α × S o + ( 1 α ) × S r . S o and S r indicate the object- and region-aware structural similarity, respectively, and we set α = 0.5 . The E-measure [45] considers the difference between the prediction and ground truth in terms of both the global and pixel levels. The MAE is computed by the average of the pixel-wise absolute values.

3.2.2. Implementation Details

For a fair comparison, we followed the experimental settings of the existing study [27]. The study concatenated both Kvasir and CVC-ClinicDB datasets. Afterward, they separated training, validation, and test sets to 80%:10%:10%, respectively. We applied a flip, blur, brightness, and distortion series for data augmentation to improve the model generalization effect. An Adam optimizer with a learning rate of 1 ×   10 4 and weight decay of 1 ×   10 4 was employed. We calibrated the learning rate with an increment of 0.1, by monitoring the validation loss if the validation loss did not reduce after 10 epochs. Moreover, an early stopping strategy was applied if the validation loss did not decrease for 20 epochs. We set the batch size to 32 and the maximum epochs to 200, and the input images were resized to 384 × 384. ResNet-50 [32] and Res2Net-50 [33] were employed as backbone networks by initializing the pre-trained on ImageNet datasets. We used a AMD-RYZEN R9 5900X CPU and single RTX 3090 GPU in this experiment, and COMMA was implemented using the PyTorch framework.

3.3. Experimental Results

3.3.1. Comparison with State-of-the-Art Methods

We compared the proposed network with eight existing methods [9,17,19,20,22,25,27,46]. For unbiased comparison, we compared the proposed method to the state-of-the- arts [9,17,19,20,22,25,27,46] using the prediction maps pre-computed by the existing study [27] and the scores obtained from the published papers [19,22,25]. As demonstrated in Table 1, COMMA achieved state-of-the-art performance on five evaluation metrics compared to previous methods. ResNet-50, as a backbone encoder, showed outstanding performance on the CVC-ClinicDB, whereas Res2Net50 outperformed the existing methods on the Kvasir dataset. In terms of the generalization performance, as shown in Table 2, COMMA exhibited significant improvement on the three unseen datasets (i.e., CVC-ColonDB, ETIS, and CVC-T). Compared to the state-of-the-art method, PraNet (32.55 M), which is the previous outstanding approach, COMMA required fewer learning parameters (31.1 M), but the segmentation performance was improved. As we obtained network robustness by leveraging the explicit boundary propagation and large samples with data augmentation, COMMA could detect unknown polyps more elaborately.

3.3.2. Qualitative Comparison

As illustrated in Figure 2, we randomly sampled polyp images from each dataset to validate the COMMA performance through the prediction maps. The images in the first and third rows were necessarily difficult cases because they included small polyps in the low-brightness areas. Although U-Net, U-Net++, and PraNet did not predict polyp areas, the proposed method detected polyp areas with some noise. This may be helpful to clinicians performing colonoscopy by improving polyp segmentation, where reducing false negatives is important. The images in the second and fifth rows show a low color contrast with the surrounding tissue. As as result of similar issues, previous methods tend to overestimate the surrounding tissue as a polyp, whereas the proposed method accurately predicted polyps. In the fourth-row image, which contains a large-scale polyp, the existing methods could not detect polyp regions with boundaries elaborately; in contrast, the proposed method was able to discriminate the polyp regions and boundaries. As a result, COMMA segregated the polyps in low-contrast and varied-size contexts by leveraging explicit boundaries and complementary representation.

3.3.3. Inference Analysis

As listed in Table 3, we compared the proposed method with the PraNet with respect to inference times. All execution times were measured under the model declaration and data I/O times. To provide an unbiased comparison, we used the code and hyper-parameters released by the authors [27]. For the same batch size, COMMA required fewer parameters, was 1.39× faster, and performed significantly better than PraNet. Furthermore, we changed the batch size to eight and found that the average FPS was 74.70, which is sufficient for real-time operation.

3.4. Further Experiments

3.4.1. Effectiveness of Multi-Decoder Structure

We also conducted experiments on the effectiveness of the number of decoders in multi-decoder structures. As listed in Table 4, we observed that the multi-decoder structures are more effective in obtaining detailed polyp regions than the single-decoder structure. In the multi-decoder structures, while all the number of decoders from two to five are effective, two decoders achieve slightly better performance with fewer learning parameters, justifying this selection as the basis of the number of decoders.

3.4.2. Individual Learning

In the experiments, we trained a total dataset that merged CVC-ClinicDB and Kvasir datasets to obtain the leverage effect of a large distribution. We compared the results of training using CVC-ClinicDB and Kvasir as training datasets to investigate whether the benefits of the total dataset help to improve performance. As listed in Table 5, we observed that using the total dataset training strategy outperformed individual dataset training on three datasets (i.e., CVC-ClinicDB, ETIS, and CVC-T). The mean dice scores were 1.4–5.6% higher than when individually trained. However, using the total dataset training strategy slightly worsens the performance on the Kvasir dataset, whereas it performed similarly on the ColonDB dataset. This suggests that the data distribution of the CVC-ClinicDB dataset included in the total dataset is different from that of the Kvasir dataset. Therefore, noise was added to the distribution of the Kvasir dataset, resulting in a slight performance degradation. Nevertheless, we note that using a joint dataset training strategy is robust in terms of the generalization performance.

3.4.3. Comparison of Interpolation Method of Up/Down-Sampling

In semantic segmentation, pixel interpolation is widely used when performing up/down-sampling of images. Several interpolation methods have been developed, including nearest, bilinear, and bicubic interpolation. We conducted an experiments to evaluate performance of various interpolation methods. In Table 6, it is observed that the bilinear method significantly outperformed the nearest and bicubic methods. Although the nearest method is computationally faster than the other methods, it causes significant performance degradation because the boundaries are not preserved. However, the cubic method, which interpolates using the product of 16 adjacent pixel values and weights according to distance, exhibited less image distortion than the nearest and bilinear methods. However, contrary to expectations, bicubic interpolation performed better than the nearest, but overall worse than bilinear in our experiments. Therefore, we used the bilinear method as the default method based on the result of this experiment and previous studies [9,17,20,27].

4. Discussion

As demonstrated in Section 2, we proposed three components, including CMM, BPM, and a hybrid loss function, to improve the performance of polyp segmentation. To validate the effectiveness of each component and show the model explainability, we conducted an additional analysis related to the components. The remainder of the analysis considered the effectiveness of each module and BPM combinations, visualization of the proposed module operation, and comparison of the proposed loss function and the existing single loss functions.

4.1. Effectiveness of Proposed Modules

To validate the effectiveness of the proposed modules, we compared the performance gain of the proposed modules. As listed in Table 7, we found that the performance gain mainly originated from the CMM. Although the performance gain was observed when the BPM was applied, in the second row of each dataset, the CMM significantly improved the polyp detection performance. That is, the CMM contributed more to the performance improvement than the BPM when applying an individual module. In terms of information propagation, employing both modules could achieve an outstanding performance compared to using an individual combination. This is because the proposed network (Base + CMM + BPM) propagated the boundary information obtained from the BPM to the second decoder structure (CMM). This structure could contribute to obtaining more refined polyp regions than the individual combinations (Base + CMM and Base + BPM).

4.2. Effectiveness of BPM Combinations

In BPM, explicit boundary information is obtained by combining the low-level feature E 1 ( 1 ) , which has rich boundary information, and the highest-level feature D 4 ( 1 ) of the first decoder. As presented in Table 8, experiments were conducted to investigate the effects of different level features of the decoder ( D 1 4 ( 1 ) ). Based on mDice and mIoU, the combination with D 4 ( 1 ) , which is the highest-level representation, yielded much better performance on the CVC-ClinicDB and CVC-T datasets than D 2 ( 1 ) and D 3 ( 1 ) , respectively. However, on the Kvasir and ColonDB datasets, the combination with D 2 ( 1 ) and D 3 ( 1 ) performed best, respectively. In MAE, we observe that the combination with D 4 ( 1 ) outperformed all the datasets. This experiment suggests that D 4 ( 1 ) , which is the representation with the most abstraction, extracts clean boundary information by removing the background noise of E 1 ( 1 ) . Moreover, as in the previous methods [20,23], an experiment was performed to extract and propagate boundary information using an independent boundary decoder (row 4). We observe that the independent boundary decoder performs similarly well on CVC-ClincDB and Kvasir datasets, while its usage leads to a drop in the generalization performance (i.e., three unseen datasets). This suggests that using the independent boundary decoder with more parameters may better represent the distribution of the training dataset (i.e., CVC-ClinicDB, Kvasir) but is not good for generalization.

4.3. CMM and BPM Visualization

We visualized feature maps obtained from the CMM and BPM operations to explain complementary propagation.

4.3.1. Complementary Masking Visualization

To verify the effectiveness of the complementary multi-level aggregation, we sampled random channels from the low- and high-level representations and compared them after applying complementary information. In Figure 3-CMM, (a) and (b) denote relatively low-level and high-level features, respectively, whereas (c) is the complementary information masked by the high-level representation. We observed that the polyp regions were enhanced, and the discrepancy was diminished between (a)–(b) and (d)–(e) when applying the complementary information. The CMM in the first decoder propagated the refined representations to the next CMM and the second decoder. Following the complementary propagation, the second decoder could present the polyp regions under the discrepancy reduced representations.

4.3.2. Explicit Boundary Visualization

We have addressed the BPM generating explicit boundary information, and we visualized the lowest- and highest-level representations to observe the boundary. In Figure 3-BPM, (a), (d) and (b), (e) are the lowest- and highest-levels (i.e., E 1 and D 4 ), respectively. Each feature was randomly sampled from the channel-wise representations. Based on the properties of both levels, because the highest-level presented abstracted region information, it could be employed as a mask, which eliminated the background noise in the lowest-level representation. As a result, we obtained the explicit boundary features, as indicated in (c) and (f), following which the boundaries were propagated to the CMMs in the second decoder to clarify the polyp borderlines. By leveraging the explicit boundary and computing the boundary loss, COMMA could discriminate ambiguous polyp contexts more precisely.

4.4. Analysis of Loss Functions

4.4.1. Comparison of Loss Functions

To investigate the effectiveness of each loss function in the proposed hybrid loss function, a comparative experiment was performed between the hybrid loss function and other single loss functions in Table 9. First, we compared the BCE, IoU, and L1 loss functions and observed that IoU loss performed better than BCE loss on mean dice and mean IoU, whereas BCE loss performed better than IoU loss on S m and E m ; however, L1 loss outperformed in most evaluation metrics. Second, to verify the effectiveness of the weighted loss function, BCE and weighted BCE were comapred along with IoU and weighted IoU. The weighted loss functions significantly improved the overall performance. This suggests that the weighted loss function allows the network to focus on the boundary information, leading to improved performance. Third, we compared the weighted BCE and the weighted IoU loss functions and found similar trends to the comparison of the BCE and IoU loss functions. Based on the results of this experiment, we concluded that the weighted IoU affected mDice and mIOU more than weighted BCE and L1, whereas weighted BCE affected S m and E m more. In addition, L1 loss, which exhibited the best MAE, helped the model to make pixel-by-pixel predictions with high confidence. Although some single loss functions outperformed the hybrid loss function on certain metrics, the proposed hybrid loss function significantly outperformed the alternatives on most evaluation metrics.

4.4.2. Weighted Loss Combination

We examine the combinations of the proposed hybrid loss function to observe the influence on local-global structure awareness and robustness against noisy annotations. As shown in Table 10, we excluded the L1 loss (ver. 1), the overall performance decreases. In ver. 2, which imposed more weights on the weighted BCE, mDice and mIoU increased because the wBCE enabled the network to learn the local structure, but S m and MAE were unsatisfactory. In contrast, higher weights for the wIoU (ver. 3) showed better local-global structure awareness ( S m ) compared to other combinations, but this combination was unable to handle noisy labels for stable convergence. Although some combinations outperformed the equally weighted combinations in certain metrics (ver. 5), we adopted an equally weighted combination because it showed a satisfactory generalization performance.

5. Conclusions

This study has focused on the propagation of complementary multi-level aggregation to overcome multi-level distribution discrepancy. The proposed method, COMMA, leverages refined low- and high-level representations in a multi-decoder structure to discriminate polyps in ambiguous contexts. The decoder, which comprises a complementary masking module (CMM) and boundary propagation module (BPM), refines the complementary features by masking the low-level representation through the high-level representation, and propagates the complementary information to the next decoder. We also introduce a hybrid loss function combining weighted BCE, weighted IoU, and L1 distance loss to address class imbalance and noisy label problems. To verify the effectiveness of the proposed approach, we evaluated COMMA compared to existing segmentation methods on five benchmark datasets. COMMA achieved state-of-the-art performance, as well as significant generalization performance. In future work, we intend to study complementary multi-level aggregation to achieve greater memory efficiency and improve segmentation performance.

Author Contributions

Conceptualization, W.S.; methodology, W.S.; data curation, W.S.; formal analysis, W.S.; writing—original draft, W.S.; visualization, W.S.; methodology, M.S.L.; writing—original draft & editing, M.S.L.; data curation, M.S.L. and S.W.H.; formal analysis, M.S.L.; supervision, S.W.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Brain Korea 21 FOUR. This research was also supported by National Research Foundation of Korea (NRF-2019R1F1A1060250).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no potential conflict of interest.

References

  1. Granados-Romero, J.J.; Valderrama-Treviño, A.I.; Contreras-Flores, E.H.; Barrera-Mera, B.; Herrera Enríquez, M.; Uriarte-Ruíz, K.; Ceballos-Villalba, J.; Estrada-Mata, A.G.; Alvarado Rodríguez, C.; Arauz-Peña, G. Colorectal cancer: A review. Int. J. Res. Med. Sci. 2017, 5, 4667–4676. [Google Scholar] [CrossRef] [Green Version]
  2. Schreuders, E.H.; Ruco, A.; Rabeneck, L.; Schoen, R.E.; Sung, J.J.; Young, G.P.; Kuipers, E.J. Colorectal cancer screening: A global overview of existing programmes. Gut 2015, 64, 1637–1649. [Google Scholar] [CrossRef]
  3. Mármol, I.; Sánchez-de Diego, C.; Pradilla Dieste, A.; Cerrada, E.; Rodriguez Yoldi, M.J. Colorectal carcinoma: A general overview and future perspectives in colorectal cancer. Int. J. Mol. Sci. 2017, 18, 197. [Google Scholar] [CrossRef] [Green Version]
  4. Siegel, R.L.; Miller, K.D.; Goding Sauer, A.; Fedewa, S.A.; Butterly, L.F.; Anderson, J.C.; Cercek, A.; Smith, R.A.; Jemal, A. Colorectal cancer statistics, 2020. CA Cancer J. Clin. 2020, 70, 145–164. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Mamonov, A.V.; Figueiredo, I.N.; Figueiredo, P.N.; Tsai, Y.H.R. Automated polyp detection in colon capsule endoscopy. IEEE Trans. Med. Imaging 2014, 33, 1488–1502. [Google Scholar] [CrossRef] [Green Version]
  6. Tajbakhsh, N.; Gurudu, S.R.; Liang, J. Automated polyp detection in colonoscopy videos using shape and context information. IEEE Trans. Med. Imaging 2015, 35, 630–644. [Google Scholar] [CrossRef] [PubMed]
  7. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Integrating online and offline three-dimensional deep learning for automated polyp detection in colonoscopy videos. IEEE J. Biomed. Health Inform. 2016, 21, 65–75. [Google Scholar] [CrossRef] [PubMed]
  8. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  9. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  10. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  11. Brandao, P.; Mazomenos, E.; Ciuti, G.; Caliò, R.; Bianchi, F.; Menciassi, A.; Dario, P.; Koulaouzidis, A.; Arezzo, A.; Stoyanov, D. Fully convolutional neural networks for polyp segmentation in colonoscopy. In Medical Imaging 2017: Computer-Aided Diagnosis; SPIE: Orlando, FL, USA, 2017; Volume 10134, pp. 101–107. [Google Scholar]
  12. Ji, G.P.; Chou, Y.C.; Fan, D.P.; Chen, G.; Fu, H.; Jha, D.; Shao, L. Progressively normalized self-attention network for video polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 142–152. [Google Scholar]
  13. Wei, J.; Hu, Y.; Zhang, R.; Li, Z.; Zhou, S.K.; Cui, S. Shallow attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 699–708. [Google Scholar]
  14. Wang, T.; Zhang, L.; Wang, S.; Lu, H.; Yang, G.; Ruan, X.; Borji, A. Detect globally, refine locally: A novel approach to saliency detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3127–3135. [Google Scholar]
  15. Wu, Z.; Su, L.; Huang, Q. Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 12–20 June 2019; pp. 3907–3916. [Google Scholar]
  16. Zhang, P.; Wang, D.; Lu, H.; Wang, H.; Ruan, X. Amulet: Aggregating multi-level convolutional features for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 202–211. [Google Scholar]
  17. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  18. Jha, D.; Smedsrud, P.H.; Johansen, D.; de Lange, T.; Johansen, H.D.; Halvorsen, P.; Riegler, M.A. A comprehensive study on colorectal polyp segmentation with ResUNet++, conditional random field and test-time augmentation. IEEE J. Biomed. Health Inform. 2021, 25, 2029–2040. [Google Scholar] [CrossRef]
  19. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; De Lange, T.; Halvorsen, P.; Johansen, H.D. Resunet++: An advanced architecture for medical image segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar]
  20. Fang, Y.; Chen, C.; Yuan, Y.; Tong, K.y. Selective feature aggregation network with area-boundary constraints for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 302–310. [Google Scholar]
  21. Zhang, R.; Li, G.; Li, Z.; Cui, S.; Qian, D.; Yu, Y. Adaptive context selection for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 253–262. [Google Scholar]
  22. Patel, K.; Bur, A.M.; Wang, G. Enhanced u-net: A feature enhancement network for polyp segmentation. In Proceedings of the 2021 18th Conference on Robots and Vision (CRV), Burnaby, BC, Canada, 26–28 May 2021; pp. 181–188. [Google Scholar]
  23. Murugesan, B.; Sarveswaran, K.; Shankaranarayana, S.M.; Ram, K.; Joseph, J.; Sivaprakasam, M. Psi-Net: Shape and boundary aware joint multi-task deep network for medical image segmentation. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 7223–7226. [Google Scholar]
  24. Lee, C.Y.; Xie, S.; Gallagher, P.; Zhang, Z.; Tu, Z. Deeply-supervised nets. In Proceedings of the Artificial Intelligence and Statistics, San Diego, CA, USA, 9–12 May 2015; pp. 562–570. [Google Scholar]
  25. Wang, D.; Hao, M.; Xia, R.; Zhu, J.; Li, S.; He, X. MSB-Net: Multi-Scale Boundary Net for Polyp Segmentation. In Proceedings of the 2021 IEEE 10th Data Driven Control and Learning Systems Conference (DDCLS), Suzhou, China, 14–16 May 2021; pp. 88–93. [Google Scholar]
  26. Cao, F.; Gao, C.; Ye, H. A novel method for image segmentation: Two-stage decoding network with boundary attention. Int. J. Mach. Learn. Cybern. 2021, 1–13. [Google Scholar] [CrossRef]
  27. Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 263–273. [Google Scholar]
  28. Chen, S.; Tan, X.; Wang, B.; Hu, X. Reverse attention for salient object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 234–250. [Google Scholar]
  29. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, Feedback and Focus for Salient Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Hilton, NY, USA, 7–12 February 2020; Volume 34, pp. 12321–12328. [Google Scholar]
  30. Ghosh, A.; Kumar, H.; Sastry, P. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  31. Wang, G.; Liu, X.; Li, C.; Xu, Z.; Ruan, J.; Zhu, H.; Meng, T.; Li, K.; Huang, N.; Zhang, S. A noise-robust framework for automatic segmentation of COVID-19 pneumonia lesions from CT images. IEEE Trans. Med. Imaging 2020, 39, 2653–2663. [Google Scholar] [CrossRef]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Gao, S.; Cheng, M.M.; Zhao, K.; Zhang, X.Y.; Yang, M.H.; Torr, P.H. Res2net: A new multi-scale backbone architecture. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 652–662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 630–645. [Google Scholar]
  35. Jain, V.; Seung, S. Natural image denoising with convolutional networks. Adv. Neural Inf. Process. Syst. 2008, 21, 769–776. [Google Scholar]
  36. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [Green Version]
  37. Zhao, J.X.; Liu, J.J.; Fan, D.P.; Cao, Y.; Yang, J.; Cheng, M.M. EGNet: Edge guidance network for salient object detection. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 8779–8788. [Google Scholar]
  38. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Halvorsen, P.; Lange, T.d.; Johansen, D.; Johansen, H.D. Kvasir-seg: A segmented polyp dataset. In Proceedings of the International Conference on Multimedia Modeling, Daejeon, Korea, 5–8 January 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 451–462. [Google Scholar]
  39. Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; Gil, D.; Rodríguez, C.; Vilariño, F. WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Comput. Med. Imaging Graph. 2015, 43, 99–111. [Google Scholar] [CrossRef]
  40. Bernal, J.; Sánchez, J.; Vilarino, F. Towards automatic polyp detection with a polyp appearance model. Pattern Recognit. 2012, 45, 3166–3182. [Google Scholar] [CrossRef]
  41. Silva, J.; Histace, A.; Romain, O.; Dray, X.; Granado, B. Toward embedded detection of polyps in wce images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 283–293. [Google Scholar] [CrossRef] [PubMed]
  42. Vázquez, D.; Bernal, J.; Sánchez, F.J.; Fernández-Esparrach, G.; López, A.M.; Romero, A.; Drozdzal, M.; Courville, A. A benchmark for endoluminal scene segmentation of colonoscopy images. J. Healthc. Eng. 2017, 2017, 4037190. [Google Scholar] [CrossRef] [PubMed]
  43. Chen, Z.; Xu, Q.; Cong, R.; Huang, Q. Global context-aware progressive aggregation network for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Hilton, NY, USA, 7–12 February 2020; Volume 34, pp. 10599–10606. [Google Scholar]
  44. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4548–4557. [Google Scholar]
  45. Fan, D.P.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
  46. Zhang, Z.; Liu, Q.; Wang, Y. Road extraction by deep residual u-net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Overall architecture of proposed COMMA.
Figure 1. Overall architecture of proposed COMMA.
Applsci 12 02114 g001
Figure 2. Qualitative comparison with existing methods.
Figure 2. Qualitative comparison with existing methods.
Applsci 12 02114 g002
Figure 3. Visualization of complementary masking and explicit boundary generation.
Figure 3. Visualization of complementary masking and explicit boundary generation.
Applsci 12 02114 g003
Table 1. Comparison of COMMA performance with existing state-of-the-art methods on CVC-ClinicDB and Kvasir datasets, where n/a indicates an inaccessible prediction map, and * and † indicate that the ResNet and Res2Net backbone encoders were employed, respectively.
Table 1. Comparison of COMMA performance with existing state-of-the-art methods on CVC-ClinicDB and Kvasir datasets, where n/a indicates an inaccessible prediction map, and * and † indicate that the ResNet and Res2Net backbone encoders were employed, respectively.
ModelCVC-ClinicDBKvasir
mDicemIoU S m E m MAEmDicemIoU S m E m MAE
U-Net [9]0.8230.7600.8900.9530.0190.8180.7500.8580.8930.055
U-Net++ [17]0.7940.7330.8730.9310.0220.8210.7470.8620.9090.048
ResUNet-mod [46]0.7790.455n/an/an/a0.7910.429n/an/an/a
ResUNet++ [19]0.7960.796n/an/an/a0.8130.793n/an/an/a
SFA [20]0.7020.6110.7930.8850.0420.7250.6140.7810.8490.075
PraNet [27]0.8990.8530.9370.9790.0090.8970.8440.9150.9480.030
COMMA *0.9160.8710.9470.9790.0080.9040.8600.9250.9630.024
COMMA 0.9330.8910.9560.9850.0070.9010.8520.9190.9510.027
Table 2. Comparison of COMMA generalization performance with state-of-the-art methods on CVC-ColonDB, ETIS, and CVC-T datasets. * and † indicate that the ResNet and Res2Net backbone encoders were employed, respectively.
Table 2. Comparison of COMMA generalization performance with state-of-the-art methods on CVC-ColonDB, ETIS, and CVC-T datasets. * and † indicate that the ResNet and Res2Net backbone encoders were employed, respectively.
DatasetModelmDicemIoU S m E m MAE
ColonDBU-Net [9]0.5040.4400.7100.7810.059
U-Net++ [17]0.4820.4120.6930.7640.061
SFA [20]0.4570.3410.6280.7530.094
PraNet [27]0.7110.6440.8200.8720.043
E: U-Net [22]0.7400.663---
MSBNet [25]0.741-0.8260.8750.040
COMMA *0.7120.6450.8230.8640.045
COMMA 0.7540.6890.8490.8970.037
ETISU-Net [9]0.3990.3400.6840.7400.036
U-Net++ [17]0.4010.3480.6830.7760.035
SFA [20]0.2980.2210.5570.6320.109
PraNet [27]0.6280.5710.7940.8410.031
E: U-Net [22]0.6510.582---
MSBNet [25]0.606-0.7720.8410.023
COMMA *0.7090.6430.8450.8870.018
COMMA 0.7110.6480.8440.8870.015
CVC-TU-Net [9]0.7110.6310.8430.8750.022
U-Net++ [17]0.7080.6290.8390.8980.018
SFA [20]0.4680.3340.6410.8170.065
PraNet [27]0.8710.8010.9250.9720.010
E: U-Net [22]0.8860.813---
MSBNet [25]0.866-0.9170.9660.010
COMMA *0.8710.8010.9240.9800.011
COMMA 0.9060.8430.9450.9880.006
Table 3. Inference analysis of four datasets in the same environment. FPS refers to the number of frames processed per second.
Table 3. Inference analysis of four datasets in the same environment. FPS refers to the number of frames processed per second.
ModelsBatch.#ParamsCVC-ClinicDBKvasirColonDBETISMean FPS
mDiceMAEFPSmDiceMAEFPSmDiceMAEFPSmDiceMAEFPS
PraNet132.55 M0.8990.00932.990.8970.03031.710.7110.04337.100.6280.03119.5630.34
COMMA131.10 M0.9330.00742.320.9010.02731.900.7540.03751.960.7110.01542.3142.12
COMMA831.10 M0.9330.00771.300.9010.02746.470.7540.037110.000.7110.01571.0774.70
Table 4. Comparison of COMMA performance based on the decoder structures.
Table 4. Comparison of COMMA performance based on the decoder structures.
Dataset#Decoder (d)#ParamsmDicemIoU S m E m MAE
CVC-ClinicDB128.43 M0.9190.8750.9440.9780.0072
231.10 M0.9330.8910.9560.9850.0066
333.76 M0.9250.8820.9450.9810.0074
436.42 M0.9310.8890.9530.9820.0068
539.09 M0.9210.8780.9500.9790.0074
Kvasir128.43 M0.8700.8230.8980.9200.032
231.10 M0.9010.8520.9190.9510.027
333.76 M0.8980.8490.9190.9530.028
436.42 M0.8970.8480.9180.9570.028
539.09 M0.9010.8510.9190.9500.027
ColonDB128.43 M0.7010.6650.8170.8440.051
231.10 M0.7540.6890.8490.8970.037
333.76 M0.7620.6970.8520.8740.039
436.42 M0.7560.6890.8500.8750.035
539.09 M0.7530.6890.8460.8760.039
ETIS128.43 M0.6770.6210.8310.8800.0160
231.10 M0.7110.6480.8440.8870.0151
333.76 M0.7080.6440.8440.8780.0176
436.42 M0.7110.6490.8440.8930.0164
539.09 M0.6940.6330.8360.8740.0167
CVC-T128.43 M0.8500.7930.8940.9570.012
231.10 M0.9060.8430.9450.9880.006
333.76 M0.8920.8260.9350.9870.007
436.42 M0.8700.8030.9270.9640.008
539.09 M0.8880.8220.9360.9770.009
Table 5. Leverage effectiveness in large samples. † and * indicate that CVC-ClinicDB and Kvasir were used as training datasets, respectively.
Table 5. Leverage effectiveness in large samples. † and * indicate that CVC-ClinicDB and Kvasir were used as training datasets, respectively.
DatasetModelmDicemIoU S m E m MAE
CVC-ClinicDBPraNet0.8990.8530.9370.9790.009
COMMA 0.9190.8770.9480.9840.007
COMMA0.9330.8910.9560.9850.007
KvasirPraNet0.8970.8440.9150.9480.030
COMMA *0.9130.8670.9290.9650.024
COMMA0.9010.8520.9190.9510.027
ColonDBPraNet0.7110.6440.8200.8720.043
COMMA 0.7570.6900.8490.8930.038
COMMA *0.6760.6090.8030.8580.045
COMMA0.7540.6890.8490.8970.037
ETISPraNet0.6280.5710.7940.8410.031
COMMA 0.6710.5930.8200.8300.036
COMMA *0.6890.6160.8290.8630.023
COMMA0.7110.6480.8440.8870.015
CVC-TPraNet0.8710.8010.9250.9720.010
COMMA 0.8500.7850.9140.9600.014
COMMA *0.8440.7650.9090.9430.010
COMMA0.9060.8430.9450.9880.006
Table 6. Comparison of COMMA performance based on the method of up/down-sampling interpolation method.
Table 6. Comparison of COMMA performance based on the method of up/down-sampling interpolation method.
DatasetModelmDicemIoU S m E m MAE
CVC-ClinicDBNearest0.9010.8420.9290.9780.009
Bilinear0.9330.8910.9560.9850.007
Bicubic0.9200.8760.9430.9790.008
KvasirNearest0.8810.8210.9040.9420.031
Bilinear0.9010.8520.9190.9510.027
Bicubic0.8900.8410.9140.9410.028
ColonDBNearest0.7500.6720.8420.8830.037
Bilinear0.7540.6890.8490.8970.037
Bicubic0.7530.6870.8460.8850.038
ETISNearest0.6810.6040.8210.8460.014
Bilinear0.7110.6480.8440.8870.015
Bicubic0.6990.6370.8370.8460.013
CVC-TNearest0.8650.7820.9160.9800.009
Bilinear0.9060.8430.9450.9880.006
Bicubic0.8910.8260.9360.9780.007
Table 7. Ablation studies for the combination of modules.
Table 7. Ablation studies for the combination of modules.
DatasetModelmDicemIoU S m E m MAE
CVC-ClinicDBBase0.8980.8500.9330.9670.010
Base + CMM0.9260.8820.9450.9810.008
Base + BPM0.9110.8630.9400.9720.008
Base + CMM + BPM0.9330.8910.9560.9850.007
KvasirBase0.8820.8310.9060.9450.032
Base + CMM0.8970.8470.9150.9490.029
Base + BPM0.8900.8360.9120.9500.034
Base + CMM + BPM0.9010.8520.9190.9510.027
ColonDBBase0.6760.6150.8060.8100.043
Base + CMM0.7200.6520.8220.8550.040
Base + BPM0.6710.6090.8030.8130.044
Base + CMM + BPM0.7540.6890.8490.8970.037
ETISBase0.6280.5670.7990.7690.028
Base + CMM0.6750.6090.8320.8720.021
Base + BPM0.6640.5970.8160.8100.018
Base + CMM + BPM0.7110.6480.8440.8870.015
CVC-TBase0.8300.7520.8900.9310.015
Base + CMM0.8870.8190.9360.9850.008
Base + BPM0.8590.7840.9150.9610.009
Base + CMM + BPM0.9060.8430.9450.9880.006
Table 8. Comparison of COMMA performance with regard to BPM combinations. † indicates that the boundary information was extracted through an independent boundary decoder, as in previous methods [20,23].
Table 8. Comparison of COMMA performance with regard to BPM combinations. † indicates that the boundary information was extracted through an independent boundary decoder, as in previous methods [20,23].
DatasetBPM Combination#ParamsmDicemIoU S m E m MAE
CVC-ClinicDB E 1 ( 1 ) , D 2 ( 1 ) 31.10 M0.9270.8840.9440.9830.0070
E 1 ( 1 ) , D 3 ( 1 ) 31.10 M0.9300.8890.9490.9840.0068
E 1 ( 1 ) , D 4 ( 1 ) 31.10 M0.9330.8910.9560.9850.0066
E 1 ( 1 ) , D 33.32 M0.9280.8850.9540.9820.0070
Kvasir E 1 ( 1 ) , D 2 ( 1 ) 31.10 M0.9050.8550.9240.9550.027
E 1 ( 1 ) , D 3 ( 1 ) 31.10 M0.8910.8420.9150.9450.029
E 1 ( 1 ) , D 4 ( 1 ) 31.10 M0.9010.8520.9190.9510.027
E 1 ( 1 ) , D 33.32 M0.9030.8520.9200.9540.028
ColonDB E 1 ( 1 ) , D 2 ( 1 ) 31.10 M0.7620.6970.8520.8830.037
E 1 ( 1 ) , D 3 ( 1 ) 31.10 M0.7380.6770.8400.8640.039
E 1 ( 1 ) , D 4 ( 1 ) 31.10 M0.7540.6890.8490.8970.037
E 1 ( 1 ) , D 33.32 M0.7530.6840.8490.8870.037
ETIS E 1 ( 1 ) , D 2 ( 1 ) 31.10 M0.6790.6190.8300.8300.016
E 1 ( 1 ) , D 3 ( 1 ) 31.10 M0.7140.6560.8450.8580.015
E 1 ( 1 ) , D 4 ( 1 ) 31.10 M0.7110.6480.8440.8870.015
E 1 ( 1 ) , D 33.32 M0.6970.6290.8400.8510.024
CVC-T E 1 ( 1 ) , D 2 ( 1 ) 31.10 M0.8800.8140.9330.9790.007
E 1 ( 1 ) , D 3 ( 1 ) 31.10 M0.8980.8350.9390.9810.007
E 1 ( 1 ) , D 4 ( 1 ) 31.10 M0.9060.8430.9450.9880.006
E 1 ( 1 ) , D 33.32 M0.8690.8000.9260.9840.011
Table 9. Comparison of the hybrid loss function and single loss functions.
Table 9. Comparison of the hybrid loss function and single loss functions.
Loss FunctionKvasirColonDB
mDicemIoU S m E m MAEmDicemIoU S m E m MAE
BCE0.8680.8050.9130.9480.0360.7030.6320.8410.8620.042
IoU0.8860.8380.9020.9370.0380.7290.6640.8290.8480.045
L10.8860.8350.9030.9410.0320.7510.6810.8430.8690.038
wBCE0.8760.8150.9150.9450.0330.7310.6580.8520.8780.042
wIoU0.8920.8420.9050.9400.0370.7620.6990.8450.8700.041
Hybrid (wBCE + wIoU + L1)0.9010.8520.9190.9510.0270.7540.6890.8490.8970.037
Table 10. Comparison of hybrid loss function combinations. α , β , and γ refer to the weights of L B C E w , L I o U w , and L L 1 , respectively.
Table 10. Comparison of hybrid loss function combinations. α , β , and γ refer to the weights of L B C E w , L I o U w , and L L 1 , respectively.
α / β / γ Ver.CVC-ClinicDBETIS
mDicemIoU S m E m MAEmDicemIoU S m E m MAE
1.0 / 1.0 / 0.0(1)0.9290.8860.9560.9840.0080.7000.6390.8390.8800.020
1.0 / 0.5 / 0.5(2)0.9330.8920.9450.9810.0080.7120.6490.8330.8630.018
0.5 / 1.0 / 0.5(3)0.9260.8840.9590.9860.0090.6890.6300.8440.8860.021
0.5 / 0.5 / 1.0(4)0.9250.8830.9470.9800.0070.6820.6240.8350.8600.017
1.0 / 1.0 / 1.0(5)0.9330.8910.9560.9850.0070.7110.6480.8440.8870.015
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shin, W.; Lee, M.S.; Han, S.W. COMMA: Propagating Complementary Multi-Level Aggregation Network for Polyp Segmentation. Appl. Sci. 2022, 12, 2114. https://doi.org/10.3390/app12042114

AMA Style

Shin W, Lee MS, Han SW. COMMA: Propagating Complementary Multi-Level Aggregation Network for Polyp Segmentation. Applied Sciences. 2022; 12(4):2114. https://doi.org/10.3390/app12042114

Chicago/Turabian Style

Shin, Wooseok, Min Seok Lee, and Sung Won Han. 2022. "COMMA: Propagating Complementary Multi-Level Aggregation Network for Polyp Segmentation" Applied Sciences 12, no. 4: 2114. https://doi.org/10.3390/app12042114

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop