Next Article in Journal
Validation of Automated Chromosome Recovery in the Reconstruction of Ancestral Gene Order
Previous Article in Journal
A New Cascade-Correlation Growing Deep Learning Neural Network Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Circle-U-Net: An Efficient Architecture for Semantic Segmentation

1
Experimental Teaching Center for Liberal Arts, Zhejiang Normal University, Jinhua 321004, China
2
The School of AI, Bangalore 560002, India
3
Lotte Data Communication R & D Center India LLP, Chennai 600113, India
4
Key Laboratory of Advanced Manufacturing Technology of Ministry of Education, Guizhou University, Guiyang 550025, China
5
School of Mechanical Engineering, Guizhou University, Guiyang 550025, China
6
Engineering College, Zhejiang Normal University, Jinhua 321004, China
*
Author to whom correspondence should be addressed.
These two authors contributed equally.
Algorithms 2021, 14(6), 159; https://doi.org/10.3390/a14060159
Submission received: 10 April 2021 / Revised: 15 May 2021 / Accepted: 19 May 2021 / Published: 21 May 2021
(This article belongs to the Topic Medical Image Analysis)

Abstract

:
State-of-the-art semantic segmentation methods rely too much on complicated deep networks and thus cannot train efficiently. This paper introduces a novel Circle-U-Net architecture that exceeds the original U-Net on several standards. The proposed model includes circle connect layers, which is the backbone of ResUNet-a architecture. The model possesses a contracting part with residual bottleneck and circle connect layers that capture context and expanding paths, with sampling layers and merging layers for a pixel-wise localization. The results of the experiment show that the proposed Circle-U-Net achieves an improved accuracy of 5.6676%, 2.1587% IoU (Intersection of union, IoU) and can detect 67% classes greater than U-Net, which is better than current results.

1. Introduction

Convolutional neural networks (CNN) have achieved state-of-the-art experiments in many computer vision tasks, such as attribute recognition, object detection, semantic segmentation, labels optimizing and so on [1,2,3,4]. Semantic segmentation is of paramount importance in the computer vision field because of its applications in box-supervised and texture features. Mask R-CNN [5] and U-Net [6] are used in semantic segmentation.
Mask R-CNN was developed by adding a fork, which aimed at predicting and generating a high-quality segmentation mask. Bhuiyan et al. [7] used Mask R-CNN to automatically detect and sort IWPs in the North Slope of Alaska. Mask R-CNN is widely applied in the remote sensing domain. Mahmoud et al. [8] proposed an adaptive Mask R-CNN in multi-class objects, detected in a remote sensing domain. Zhao et al. [9] created polygons using segmentation algorithms based on Mask R-CNN. Li et al. [10] proposed HTMask R-CNN, which is based on Mask R-CNN. HTMask R-CNN could adopt the features of single-object segmentation from Mask R-CNN. Mask R-CNN and U-Net [6] are widely applied in object segmentation. However, there are some limitations to U-Net. For example. the mIoU (mean Intersection of union, mIoU) is low and is unable to detect 2, 4 and 8 classes.
In this paper, we propose a new type of U-Net [6], called Circle-U-Net. Circle-U-Net has 101 layers and is inspired by a residual net from deep learning and a circle from geometry. We prove that Circle-U-Net can segment objects better than other networks, due to both its depth and residual layer; the more profound the network model, the more incredible because it can segment. Moreover, only in a deeper network can we say that additional functionalities, such as residual layers, attention layers, are helpful. The residual layers that form a circular pattern suggest that the recurring patterns vividly help the network understand specific parts. Our contributions can be summarized below.
(1)
We put forward a Circle-U-Net network with a circle connect model, which exceeds the performance of the attention mechanism. Our network improves 0.78 mIoU than adding the attention mechanism to our model. The circle connects model is robust and capable in object segment, which performs better than most state-of-the-art experiments.
(2)
Circle-U-Net cannot detect 2 classes, while some networks cannot see 2, 4 and 8 classes. In other words, Circle-U-Net has high power to detect than other networks.
(3)
We prove that the proposed method has a better performance in comparison with the state-of-the-art networks. Furthermore, we organize the rest of this paper as follows: Section 2 describes related works and Section 3 reviews the proposed Circle-U-Net structure in detail. Experimental results and comparisons are described in Section 4, followed by conclusions in Section 5.

2. Related Work

Olaf et al. [5] present U-Net, which consists of a contracting path to capture context and a symmetric expanding path that enables precise localization. U-Net is fast, and the segmentation of a 512 × 512 image takes less than a second on a recent GPU. Bhakti et al. [11] present Eff-UNet, which combines the effectiveness of compound scaled Efficient Nets and the encoder for feature extraction with U-Net decoder for reconstructing the fine-grained segmentation map. The Eff-UNet combines high-level feature information, as well as low-level spatial valuable information, for precise segmentation. Nazanin et al. [12] propose a Squeeze U-Net inspired version of U-Net for image segmentation. The Squeeze U-Net is efficient in both low MACs and memory use. Edgar et al. [13] propose that U-Net-based architecture provides detailed per-pixel feedback to the generator, while maintaining the global coherence of synthesized images by providing global image feedback. Eisuke et al. [14] propose Feedback U-Net using Convolutional LSTM, the segmentation method using the Convolutional LSTM and feedback process. Wei et al. [15] introduce a novel recurrent U-Net architecture that preserves the compactness of the original U-Net. Benjamin et al. [16] propose an enhanced Rotation-Equivariant U-Net for Nuclear Segmentation. RUNet [17] can learn the relationship between a set of degraded low-resolution images and their corresponding original high-resolution images. Reza et al. [18] put forward a bi-directional Co-nvLSTM (BConvLSTM) U-Net with densely connected convolutions for medical image segmentation, taking full advantage of U-Net, bi-directional ConvLSTM, and the mechanism of dense convolutions. Additionally, BConvLSTM combines the feature maps extracted from the corresponding encoding path and the previous decoding up-convolutional layer in a non-linear way. W-Net [19] is a reinforced U-Net for density map estimation. Retina U-Net [20] is the one-stage detector, which combines detection with an auxiliary segmentation. The adaptive triple [21] U-Net with a test-time augmentation can segment common and internal carotid arteries more efficiently and accurately. R2U-Net [22] realizes the power of Residual Network, U-Net, as well as RCNN, which are several advantages for segmentation tasks. Ozan et al. [23] propose Attention U-Net, which is assessed on CT datasets and is for multi-class image segmentation. UNet++ [24] is a nested U-Net architecture which gets an average IoU gain of 3.9 points over U-Net.

3. Method

3.1. Circle-U-Net

There is ten-block which starts with a 7 × 7 block followed by a pooling layer called C1, then followed by C2 with three bottleneck block, C3 with four bottleneck block with one circle connect, C4 with 23 bottleneck blocks with two circles connects, and C5 with three bottleneck blocks and one circle connect. C1 to C5 forms the contracting layers that capture the context of the image, and C6 to C10 forms the expanding layers to achieve pixel-wise localization and finally forming a segmented image. C6 comprises a merging layer from C4, C7 comprises a merging layer from C3, C8 comprises a merging layer from C2, C9 gets from C1. We have C10 forming the final expanding layer, which produces the output by pixel-wise classification. Figure 1 denotes the Circle-U-Net architecture as described above.
In Figure 2, the Concat-Conv Structure suggests that an up-convolution layer is merged with a layer from the contracting path, which is done to add high-resolution features from the contracting path to the expanding path for better localization, and convolution blocks are added successfully to learn to produce precise output.
This is the schematic of our framework, which is described in Figure 3. An input image of 256 × 256 size is pre-processed by resizing, normalization, and then parsing labels. It is then fed to CircleUnet for training, where the contracting path extracts the features and expands the way it localizes. In image segmentation, pixel-wise classification takes place, and we use the CCE loss function. After training with the trained model, we can predict the output.

3.2. Circle-U-Net with Attention

Figure 4 describes Circle-U-Net with attention network. There are ten blocks, starting with a 7 × 7 block, followed by a pooling layer called C1, then followed by C2 with three bottleneck block, C3 with four bottleneck block with one circle connect, C4 with 23 bottleneck blocks with two circle connects and C5 with three bottleneck blocks, and one circle connect. C1 to C5 forms the contracting layers, and C6 to C10 forms the expanding layers to achieve a segmented image. C6 comprises gated Attention with a merging layer from C4, C7 comprises gated Attention with a merging layer from C3, C8 comprises gated Attention with a merging layer from C2, C9 gets from C1.We have C10 forming the final expanding layer, which produces the output by pixel-wise classification.
Bottleneck layers. Residual bottleneck layers are used in the neural network to encourage compressed feature representations. They are leading to the reduction in the number of parameters and matrix multiplication with an increased depth. Residual blocks [25] act as a critical ingredient by acting as a backbone for Circle-U-Net. Figure 5 describes a bottleneck layer that has 3-Convolutional blocks. Table 1 gives the structure of each block in the contracting block.
Circle Connect. Circle connect is the new concept we propose through this paper. We take Cx−1 first layer (i.e., a residual bottleneck block) and connect with the end of Layer x as described in Figure 6. By Table 1 we have proved that using such types of connecting layers for every periodic layer has increased accuracy. When a layer has the information of its far previous it can associate and learn current features with the farther ones. We have four circles connect layers in our architecture, as described in Table 2.
Gated signals. The gated signals include resizing the layer feature map to the up-layer feature map, using 1 × 1 convolution. The gating signal used with skip connection aggregates image features from multiple image scales and is fed for attention blocks.
Attention gates. Attention gates (AG) is proposed by Schlemper et al. [26]. AG could learn to automatically focus on the sizes and shapes of target structures. AGs are used when the model implicitly learns to suppress irrelevant features and highlights salient features. When AG is used for CNN models such as U-Net, it could increase prediction accuracy and sensitivity. We are using attention gates in the expansion path at C6, C7, C8, and C9 layers in Circle-U-Net with Gated Signal Attention (GSA) network. AG self- learns to focus on the target structure of different shapes and sizes. Where AGs are used, the model implicitly learns to suppress irrelevant features and highlights salient features. Figure 7 describes attention gates, where gating signal and layer x are concatenated and passed on with Relu activation, which when further resampled, produces attention layer.
q a t t , i l = φ T ( σ 1 ( W x T x i l + W g T g + b x g ) ) + b φ
α l = σ 2 ( q a t t l   ( x l , g ; a t t   )
where σ 1 ( x ) ➔ element-wise non-linearity, σ 2 (x) ➔ normalization function; x i l ➔ Activation map of a chosen layer i; g -> global feature vector provides information to AGs to point out irrelevant features in x i l ; W x , W g ➔ Linear transformations; b x g , b φ ➔ Bias terms.
Loss function. We have tested our model with categorical cross-entropy loss function and Focal Tversky loss function, and it works better with both the type of loss functions.
Categorical cross-function is described as below.
Loss = 1 N i = 1 N c = 1 C 1 y i   C y i l o g   P m o d e l [ y i C e ]
N is the total number of observations, i denotes specific observation, C denotes the total number of categories, and c denotes a specific category. i = 1 N c = 1 C ➔ Observations of ‘I’, till N and categories ‘c’ till ‘C’; 1 y i C y i ith observation for cth category; P m o d e l [ y i   C c ] ➔ probability predicted by the model for ith observation which belongs to cth category
Focal Tversky loss function is described as below.
T I c = i = 1 N p i c g i c + i = 1 N p i c g i c + α i = 1 N p i c ¯ g i c + β i = 1 N p i c g i c ¯ +
where TI is the Tversky index, α + β = 1 ;   p i c ➔ probability that the pixel i is of the lesion class c; p i c ¯ ➔ probability that the pixel i is of the non-lesion class c; g i c ➔ ground truth that the pixel i is of the lesion class c; g i c ¯ ➔ ground truth that the pixel i is of the non-lesion class c;
F T L c = c ( 1 T I c ) 1 γ
where γ controls the non-linearity loss.We use the focal Tversky loss function. It solves class imbalance as FTL is robust to class imbalances.

4. Experiments

4.1. Datasets

ICG Drone dataset contains 400 images with 20 label classes prepared for semantic understanding of urban scenes for increasing the safety of autonomous drone flight and landing procedures. The dataset is acquired at an altitude of 5 to 30 m above ground, and imagery shows more than 20 houses in a bird’s eye view. This dataset is similar to biomedical segmentation, where there would be the top view of cells or bacteria. Still, it is very different from datasets, such as CamVid or Citryscape, where the datasets would be in a frontal or side view. We split the train and test set randomly to maintain uniformity in class distribution while training. We show the spread of data in the ICG Drone dataset in Figure 8, and we can see that paved areas (37.7%) and grass (19.8%) is occupying most of the dataset.

4.2. Experimental Setup We Use the Machine with RTX 2080 Ti GPU and 256G RAM for Our Experiments

To develop all the models and legally testify them, we use a similar setup. The tensorflow-gpu version is 2.1.4. We train Circle-U-Net for 60 epochs. We set the output height as 256. The batch size is 5. The output width is 256 and the optimizer is Adam for training, and we do not use any other data augmentation techniques. Initially, we faced an overfitting issue when we split the dataset sequentially. Later, when we identified the root cause that the entire dataset is like a video sequence that contains particular objects in one part and inevitable thing in another aspect, we split the train and tested the set randomly. We were able to overcome the overfitting issue. You could visit our website for more commands and usage. Both the dataset and code will be available at https://github.com/sunfeng90/Circle-U-Net, accessed on 20 May 2021.

4.3. Comparison to the State of the Art

As shown in Table 3, we can find the sample source and the number of training sample and testing sample.
As shown in Table 4, we can observe from Table 4 that Circle-U-Net (without GSA and Attention) is one of the architectures that outperformed all of the other architectures, irrespective of the loss of function or inaccuracy. The GSA is the abbreviation of Gated Signal Attention. The CCE is the abbreviation of Categorical Cross-Entropy loss. Circle-U-Net (without GSA and Attention) improves 2.1587% IoU and 5.6676% accuracy than U-Net.

4.4. Comparing Top 8 Classes

U-net is the base architecture, squeeze U-net is developed after Attentation Unet and ResUnet, which is a lightweight architecture. ResUnet (Resnet101) and CircleUnet are heavy architecture. In Attentation Unet, attention layers are used with gating signal on top of Unet base architecture, In Squeeze Unet, squeeze block are used by modifying Unet, In ResUnet we have used residual blocks and in Circle net we have used residual block and circle connect by modifying Unet. In Circlenet101 with GSA we have used residual block, circle connect and attention layers.
Since it is an encoder-decoder type of architecture without any detectors in the first stage, U-Net cannot outperform small objects’ criteria. Moreover, as the number of classes increases in many datasets, class imbalance takes place. The imbalance makes the models’ accuracy reduce, but a highly available object in every image will be higher. In Table 5, we can find that U-Net and Attention U-Net cannot detect eight classes, whereas Circle-U-Net could not detect only two classes and has considerable accuracy. This experiment proves the Circle-U-Net can segment more objects and achieve considerable mIoU while segmenting.

4.5. Ablation Study

To compare the results of Circle-U-Net with attention, and Circle-U-Net without attention. Table 6 tells that Circle-U-Net has performed better than attention architectures.
As shown in Table 4, we can find Circle-U-Net (without Attention) improve 0.78mIoU and 0.82% accuracy than Circle-U-Net (with Attention).
Figure 9 are the predictions of Circle-U-Net and corresponding ground truth.

5. Conclusions

In this paper, we propose a Circle-U-Net which helps to understand the difference between other layers. These deeper layers can recognize more objects and residual connections with circle connect and help to understand the specific part vividly. We can find Circle-U-Net surpasses the attention model through experimental results. In the future, we will apply Circle-U-Net in the services robot domain and put forward a new network about object segmentation.

Author Contributions

Conceived of and designed the study, F.S. Implementing the algorithms and analysis of the data, F.S. and A.K.V. Writing, G.Y. Checked the syntax errors of the manuscript and provided the ex-periment machine, A.Z. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No.61863005), the Science and Technology Support Plan of Guizhou Province (PTRC[2018]5702, QKHZC[2019]2814, [2020]4Y056,PTRC[2020]6007, [2021]439), the Experimental Technology and Development Project of Zhejiang Normal University( SJ202123).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of NAME OF INSTITUTE (protocol code XXX and date of approval).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data was obtained from https://github.com/sunfeng90/Circle-U-Net (accessed on 20 May 2021).

Acknowledgments

We thank anonymous reviewers for valuable comments that help improve the paper during revision.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, G.; Chen, Z.; Li, Y.; Su, Z. Rapid Relocation Method for Mobile Robot Based on Improved ORB-SLAM2 Algorithm. Remote Sens. 2019, 11, 149. [Google Scholar] [CrossRef] [Green Version]
  2. Su, Z.; Li, Y.; Yang, G. Dietary Composition Perception Algorithm Using Social Robot Audition for Mandarin Chinese. IEEE Access 2020, 8, 8768–8782. [Google Scholar] [CrossRef]
  3. Lin, J.; Li, Y.; Yang, G. FPGAN: Face de-identification method with generative adversarial networks for social robots. Neural Netw. 2021, 133, 132–147. [Google Scholar] [CrossRef] [PubMed]
  4. Yang, G.; Yang, J.; Sheng, W.; Junior, F.E.F.; Li, S. Convolutional Neural Network-based Embarrassing Situation Detection under Camera for Social Robot in Smart Homes. Sensors 2018, 18, 1530. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef] [PubMed]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  7. Bhuiyan, M.A.E.; Witharana, C.; Liljedahl, A.K. Use of Very High Spatial Resolution Commercial Satellite Imagery and Deep Learning to Automatically Map Ice-Wedge Polygons across Tundra Vegetation Types. J. Imaging 2020, 6, 137. [Google Scholar] [CrossRef]
  8. Mahmoud, A.; Mohamed, S.; El-Khoribi, R.; Abdelsalam, H. Object Detection Using Adaptive Mask RCNN in Optical Remote Sensing Images. Int. J. Intell. Eng. Syst. 2020, 13, 65–76. [Google Scholar] [CrossRef]
  9. Zhao, K.; Kang, J.; Jung, J.; Sohn, G. Building Extraction from Satellite Images Using Mask R-CNN with Building Boundary Regularization. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  10. Li, Y.; Xu, W.; Chen, H.; Jiang, J.; Li, X. A Novel Framework Based on Mask R-CNN and Histogram Thresholding for Scalable Segmentation of New and Old Rural Buildings. Remote Sens. 2021, 13, 1070. [Google Scholar] [CrossRef]
  11. Bhakti, B.; Innani, S.; Gajre, S.; Talbar, S. Eff-UNet: A Novel Architecture for Semantic Segmentation in Unstructured Environment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition WorkShop (CVPRW), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  12. Beheshti, N.; Johnsson, L. Squeeze U-Net: A Memory and Energy Efficient Image Segmentation Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshop (CVPRW), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  13. Schönfeld, E.; Schiele, B.; Khoreva, A. A U-Net Based Discriminator for Generative Adversarial Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  14. Shibuya, E.; Hotta, K. Feedback U-net for Cell Image Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  15. Wang, W.; Yu, K.; Hugonot, J.; Fua, P.; Salzmann, M. Recurrent U-Net for Resource-Constrained Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  16. Chidester, B.; Ton, T.; Tran, M.; Ma, J.; Do, M.N. Enhanced Rotation-Equivariant U-Net for Nuclear Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshop (CVPRW), Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  17. Hu, X.; Naiel, M.A.; Wong, A.; Lamm, M.; Fieguth, P. RUNet: A Robust UNet Architecture for Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  18. Azad, R.; Asadi-Aghbolaghi, M.; Fathy, M.; Escalera, S. Bi-Directional ConvLSTM U-Net with Densley Connected Convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019. [Google Scholar]
  19. Valloli, V.K.; Mehta, K. W-Net: Reinforced U-Net for Density Map Estimation. arXiv 2019, arXiv:1903.11249. [Google Scholar]
  20. Jaeger, P.F.; Kohl, S.A.A.; Bickelhaupt, S.; Isensee, F.; Kuder, T.A.; Schlemmer, H.P.; Maier-Hein, K.H. Retina U-Net: Embarrassingly Simple Exploitation of Segmentation Supervision for Medical Object Detection. arXiv 2018, arXiv:1811.08661v1. [Google Scholar]
  21. Zhao, B.; Chen, X.; Li, Z.; Yu, Z.; Yao, S.; Yan, L.; Wang, Y.; Liu, Z.; Liang, C.; Han, C. Triple U-net: Hematoxylin-aware nuclei segmentation with progressive dense feature aggregation. Med. Image Anal. 2020, 65, 101786. [Google Scholar] [CrossRef] [PubMed]
  22. Alom, M.Z.; Hasan, M.; Yakopcic, C.; Taha, T.M.; Asari, V.K. Recurrent Residual Convolutional Neural Network based on U-Net (R2U-Net) for Medical Image Segmentation. arXiv 2018, arXiv:1802.06955. [Google Scholar]
  23. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  24. Zhou, Z.; Siddiquee, M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the 4th Deep Learning in Medical Image Analysis Workshop, Granada, Spain, 9 September 2018; Springer: Cham, Switzerland, 2018; Volume 11045, pp. 3–11. [Google Scholar]
  25. Brostow, G.J.; Fauqueur, J.; Cipolla, R. Semantic object classes in video: A high-definition ground truth database. Pattern Recognit. Lett. 2009, 30, 88–97. [Google Scholar] [CrossRef]
  26. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef] [PubMed]
  27. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Circle-U-Net structure.
Figure 1. Circle-U-Net structure.
Algorithms 14 00159 g001
Figure 2. Concat-Conv structure.
Figure 2. Concat-Conv structure.
Algorithms 14 00159 g002
Figure 3. The schematic of our framework.
Figure 3. The schematic of our framework.
Algorithms 14 00159 g003
Figure 4. Circle-U-Net with attention structure.
Figure 4. Circle-U-Net with attention structure.
Algorithms 14 00159 g004
Figure 5. Bottleneck Block Structure.
Figure 5. Bottleneck Block Structure.
Algorithms 14 00159 g005
Figure 6. Circle connect Structure.
Figure 6. Circle connect Structure.
Algorithms 14 00159 g006
Figure 7. Attention gates structure.
Figure 7. Attention gates structure.
Algorithms 14 00159 g007
Figure 8. The constitute of ICG Drone dataset.
Figure 8. The constitute of ICG Drone dataset.
Algorithms 14 00159 g008
Figure 9. The predictions of Circle-U-Net.
Figure 9. The predictions of Circle-U-Net.
Algorithms 14 00159 g009
Table 1. The Structure of every block.
Table 1. The Structure of every block.
Block NameNumber of Bottleneck LayersStructure
C23Contains 7 × 7 with max-pooling at the first layer
C34Bottleneck with residual at the first layer
C423Bottleneck with residual at the first layer and circle connect at last
C53Bottleneck with residual at first the layer and circle connect at last
Table 2. Circle connect name and layers.
Table 2. Circle connect name and layers.
Circle Connect NameConnected Layers
cc1Conv2_1 with conv 3_4
cc2Conv 3_1 with conv 4_11
cc3Conv 4_11 with conv 4_23
cc4Conv 4_1 with conv 5_3
Table 3. Samples Number.
Table 3. Samples Number.
Source: https://www.tugraz.at/index.php?id=22387 (Accessed on 20 May 2021)
Total SamplesTraining SampleTesting Sample
40036040
Table 4. Comparing against the state of the art (unit: %).
Table 4. Comparing against the state of the art (unit: %).
ModelLoss FunctionAccuracyIoU
U-Net [6]CCE69.9088318.462254
Attention U-Net [23]CCE69.012216.37
Squeeze U-Net [12]CCE72.05303319.439353
ResUNet-a [27]CCE71.90650715.959859
Circle-U-Net (without GSA and Attention)CCE75.57636420.620922
Circle-U-Net (without GSA and Attention)Trversky73.96130419.437675
Circle-U-Net (with Attention)CCE72.7548919.843579
Circle-U-Net (with Attention)Trversky70.4017.076342
Table 5. Comparing against the mIoU of the state of the art.
Table 5. Comparing against the mIoU of the state of the art.
Network NameLoss
Function
Accuracy
(Top 8 Classes)
mIoU
(8 Classes)
Undetected Classes
U-Net [6]CCE55.160.41328
Attention U-Net [23]CCE58.260.41098
Squeeze U-Net [12]CCE59.840.44074
ResUNet-a [27]CCE57.170.40872
Circle-U-Net (without GSA)CCE59.290.41322
Circle-U-Net (without GSA)Tversky42.070.27204
Circle-U-NetTrversky59.300.46472
Circle-U-NetCCE56.290.58912
Table 6. Circle-U-Net with or without Attention (unit: %).
Table 6. Circle-U-Net with or without Attention (unit: %).
ModelLoss
Function
mIoUAccuracyLayers
Circle-U-Net (without Attention)CCE19.6274.57101+
Circle-U-Net (with Attention)CCE18.8473.75101+
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sun, F.; V, A.K.; Yang, G.; Zhang, A.; Zhang, Y. Circle-U-Net: An Efficient Architecture for Semantic Segmentation. Algorithms 2021, 14, 159. https://doi.org/10.3390/a14060159

AMA Style

Sun F, V AK, Yang G, Zhang A, Zhang Y. Circle-U-Net: An Efficient Architecture for Semantic Segmentation. Algorithms. 2021; 14(6):159. https://doi.org/10.3390/a14060159

Chicago/Turabian Style

Sun, Feng, Ajith Kumar V, Guanci Yang, Ansi Zhang, and Yiyun Zhang. 2021. "Circle-U-Net: An Efficient Architecture for Semantic Segmentation" Algorithms 14, no. 6: 159. https://doi.org/10.3390/a14060159

APA Style

Sun, F., V, A. K., Yang, G., Zhang, A., & Zhang, Y. (2021). Circle-U-Net: An Efficient Architecture for Semantic Segmentation. Algorithms, 14(6), 159. https://doi.org/10.3390/a14060159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop