You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

9 September 2020

FAU-Net: Fixup Initialization Channel Attention Neural Network for Complex Blood Vessel Segmentation

,
,
,
and
1
Shanghai Film Academy, Shanghai University, Shanghai 200072, China
2
The Faculty of Science, Design and Technology, Pool House, University of Bournemouth, Poole BH12 5BB, UK
3
The Faculty of Engineering and Informatics, University of Bradford, Bradford BD7 1DP, UK
*
Author to whom correspondence should be addressed.
This article belongs to the Section Applied Biosciences and Bioengineering

Abstract

Medical image segmentation based on deep learning is a central research issue in the field of computer vision. Many existing segmentation networks can achieve accurate segmentation using fewer data sets. However, they have disadvantages such as poor network flexibility and do not adequately consider the interdependence between feature channels. In response to these problems, this paper proposes a new de-normalized channel attention network, which uses an improved de-normalized residual block structure and a new channel attention module in the network for the segmentation of sophisticated vessels. The de-normalized network sends the extracted rough features to the channel attention network. The channel attention module can explicitly model the interdependence between channels and pay attention to the correlation with crucial information in multiple feature channels. It can focus on the channels with the most association with vital information among multiple feature channels, and get more detailed feature results. Experimental results show that the network proposed in this paper is feasible, is robust, can accurately segment blood vessels, and is particularly suitable for complex blood vessel structures. Finally, we compared and verified the network proposed in this paper with the state-of-the-art network and obtained better experimental results.

1. Introduction

Medical images have multiple modalities, such as Magnetic Resonance Angiography (MRA), Computed Tomography Angiography (CTA), Positron Emission Tomography (PET), ultrasound imaging, and more. In clinical diagnosis and treatment, the segmentation technology of medical images affects the reliability of diagnosis results to a great extent. Moreover, medical image segmentation technology is the first step of many medical image processing technologies, such as visualization, 3d reconstruction. Therefore, its development will affect the evolution of other related technologies in medical image processing.
With the rapid development of deep learning, a large number of intelligent methods based on neural networks for medical images segmentation have emerged in recent years. Nevertheless, medical image segmentation has some limitations, including data scarcity and class imbalance [1]. To get the results we want through deep learning training, we often need many medical images as a data set. For solving the problem of sparse data sets, different data transformation or enhancement techniques are sometimes used to increase the number of available labeled samples, such as the automatic data augmentation method proposed by Google in 2018 [2], or use a patch-based approach to resolve class imbalances.
To reduce large data sets required for network training and improve the efficiency of the work, U-Net [3] was proposed, so that we can complete the network training with very few data sets and accurately segment the region of interest. However, U-Net has some defects, such as the accuracy of the model is limited by the batch size, and the correlation between the channels is not fully considered.
To address these issues, we propose an improved de-normalized residual block. It uses a new initialization method, which solves the problem of gradient explosion and gradient disappearance by properly adjusting initialization parameters at the beginning of training. During the training process, it can converge faster. The attention mechanism, because of its low model complexity, can be calculated in parallel, and the advantage of being able to obtain critical information is also increasingly used in medical image segmentation tasks. This paper proposes a new channel attention module called SCA, which can calculate the feature mapping value in each channel, to explore the correlation of channel.
The contributions of our work can be summarized as below:
(1)
We propose a de-normalized channel attention network, which consists of a new channel attention module SCA and an improved de-normalized block.
(2)
We construct two brain vascular data sets, which uses multiple angles of CTA and MRA images from various patients, called the ELE and CORO data sets, respectively. The labeling process consults brain vascular experts.
(3)
We perform experiments on the public retinal vascular data set DRIVE [4] and brain vascular data sets ELE and CORO, compare the results with the state-of-the-art methods, and obtain high accuracy. The effects of de-normalized structure on vascular segmentation were tested on DRIVE, ELE, and CORO data sets, respectively.

3. Methods

3.1. Fixup Initialization Channel Attention Neural Network

Our de-normalized channel attention network follows the basic encoder-decoder structure of U-net. Its feature fusion method is to stitch the features together in the channel dimension to form a more “thick” feature. We want to be able to extract more elaborated features, so we adopt a dual U-Net structure with multiple pairs of skip connections. Its complete structure is shown in Figure 1.
Figure 1. Structure of FAU-net.
The first de-normalized network uses an improved de-normalized residual block structure, which we call Fixupblock (Figure 2 right). Since it does not use batch normalization, the efficiency of the model is significantly improved. The second channel attention network adds the SCA module to Basicblock (Figure 3), called Attentionblock. We first input the patches of the original picture and the mask picture into the de-normalized network, which can quickly extract rough features of the blood vessel and send the feature values to the second channel attention network (Hereafter called CAU-Net). CAU-Net can model the interdependence between feature channels so that the weight of useful feature maps in the channel is higher than useless ones, and the weight of invalid or less useful feature maps is smaller so that features can be refined. The last layer of CAU-Net is the SoftMax layer, so the loss function we adopt is the negative log-likelihood loss function NLLLoss. Comparison experiments on three data sets with multiple state-of-the-art networks prove that our network has achieved the highest AUC value currently.
Figure 2. Resnet’s Basiclock structure on the left, and an improved Fixupblock structure on the right.
Figure 3. Left is the overall structure of the Attentionblock, right is the structure of the channel attention module SCA.

3.2. Improved De-Normalized Residuals Block

ResNet [27] was proposed in 2015, and its basic block is called the residual block. Residual block there are generally two kinds of forms, BasicBlock and Bottleneck, respectively. BasicBlock structure mainly uses two convolution sums of 3 × 3 . The branch of residual is:
OUT = CONV 3 × 3 + B N + R E L U + CONV 3 × 3 + B N
Then add it to the identity map and go through a RELU layer:
OUT = OUT + x + RELU
The structure of BasicBlock is shown on the left in Figure 2. The reason BasicBlock can work has a significant relationship with the BN layer it adds. Many residual networks cannot correctly account for the effects of residual connections without normalization techniques such as BatchNorm, resulting in gradient explosions. Lim et al. [28] experimentally verified that normalizing features using BN would eliminate the range flexibility of the network. Moreover, in the segmentation task of blood vessels, in order to obtain higher experimental accuracy, we often use larger batch size, which will significantly increase the memory usage of the GPU. A new initialization method [29] that does not require any normalization method was proposed. Based on this method, we propose an improved de-normalization block called Fixupblock. It appropriately adjusts the initial offset parameter at the beginning of training to ensure that the update of the network function (Gradient) stays in the proper range. It is independent of the network depth so that it can solve the problem of gradient explosion and gradient disappearance. At the same time, we add the two drop-out layers in Fixupblock to prevent some features from overfitting. It makes the weight update is no longer dependent on the combination of a stable relationship between hidden nodes. It avoids situations where certain features only work for certain other features, forcing the network to learn more robust features.
Improved Fixupblock structure shown in the right of Figure 2, where the bias term bias1a, bias2a, bias2a, bias2b all initialized to zero, the scale is initialized to one.

3.3. Channel Attention Module

Similar attention networks used for segmentation, such as [25,30], have bigger regions of interest and are more suitable for pathological segmentation. Our channel attention network can allocate the weight after each convolution and enhance the useful features, which is suitable for the segmentation of small blood vessels.
Generally, the output of the convolutional layer does not consider the channel dependence. We adopt the structure in SEblock [19], which is called Se calculation. It can enhance the effective features and suppress ineffective features between channels. We combine it with the channel attention module and propose a new channel attention module called SCA (Figure 3). Then add the SCA module to the second residual block of u-net and call it an Attentionblock (Figure 3). The following is the detailed calculation process of the channel attention module.
We assume that the original feature is F R C × H × W , and the original feature goes through three pipelines 2, 3, and 4 as shown in Figure 3. After passing through these three pipelines, the original feature first undergoes Squeeze-and-Excitation calculation (Hereinafter referred to as Se).
In Se calculation, the first step is the Squeeze operation F s q , which is the operation of global average pooling:
Z c = F s q f c = 1 H × W i = 1 H j = 1 W f c ( i , j )
Among them, f c represents the initial feature (i.e., the three-dimensional matrix F R C × H × W ), and the subscript c represents the c-th two-dimensional matrix in F, i.e., the channel. Equation (3) converts the input of C × H × W into 1 × 1 × C , which indicates the numerical distribution of the c feature maps of the layer, i.e., the global information. The Squeeze operation in Equation (3) operated in the feature map of a channel, and the operation in Equation (4) is to fuse the weight of the feature map of each channel, i.e., the Excitation operation F e x :
s = F e x ( z , W ) = σ ( g ( z , W ) ) = σ W 2 δ W 1 z
where z is the result of Equation (3), W 1 multiplied by z is the first fully connected layer operation, and the dimension of W 1 is C r × C , where r is a scaling parameter, to reduce the number of channels and thus reduce the amount of calculation. In this paper, r sets to 2. The dimension of z is 1 × 1 × C , then the dimension of W 1 z is 1 × 1 × C r . Then it goes through a RELU layer with the same dimensions, multiply it with W 2 , which is the operation of the second fully connected layer. The dimension of W 2 is c × c r , and the dimension of output is 1 × 1 × C . Finally, the dimension is unchanged after passing the sigmoid function. The obtained s (Equation (4)) is the feature map weight of the channel learned through the previous fully connected and non-linear layers. It is the result of pipeline 2, 3, 4 after Se operation.
In pipeline 4, we transform the initial characteristics F R C × H × W to F R C × N , where N = H × W represents the pixels of the picture. Pipeline 3 was reshaped after SE processed the original features. Then perform matrix multiplication on the reshaped F and the transpose of F passing through pipeline 4, and finally use a SoftMax layer to get the channel attention map X R α :
x i j = s o f t m a x ( F i · F j ) = e x p ( F i · F j ) i = 1 C e x p ( F i · F j )
where x i j represents the influence of the i-th channel on the j-th channel.
In addition to the above operation, we perform matrix multiplication on the transpose of X and F obtained through pipeline 2 and reshape the result to R C × H × W . Then multiply this result by a scale parameter β and perform a pixel-level addition with F of pipeline 1 to get E R C × H × W :
E j = β i = 1 C x j i F i + F j
Finally, the result E passes through a RELU and softmax layer, and the final feature is P R C × H × W .

4. Quantitative Assessment Method

The segmentation of blood vessels is a binary classification problem. In classification problems, we usually use some indicators to evaluate the pros and cons of the algorithm. In this article, we use the values of Accuracy (AC), Precision, Recall, and Area under the ROC curve (AUC) and F β s c o r e ( F β ) evaluated the performance of FAU-net. We will use True Positive (TP), True Negative (TN), False Positive (FP), recall rate and False Negative (FN) to calculate the above indicators. Performance metrics calculation is as follows:
A C = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
We assume that the positive samples are pixels marked as blood vessels and the negative samples are background. We assign different weights to recall and precision to indicate the preference for classification models. The calculation equation is as follows:
F β = 1 + β 2 · P r e c i s i o n · R e c a l l β 2 · P r e c i s i o n + R e c a l l
It can be seen that when β = 1, then F β changes back to F1; when β > 1 , the model prefers to improve recall, which means that the model pays more attention to the ability to recognize positive samples. when β < 1 , the model pays more attention to the ability to distinguish negative samples. In this article, we do not want the model to produce the wrong results to cover more samples. It will affect the doctor’s diagnosis of the patient’s condition. Therefore, we will adjust the parameter β to 0.75.
Considering the class imbalance of the data set, we also introduce Receiver Operating Characteristic (ROC) and Precision-Recall (PR) curves. The ROC curve is drawn with True Positive Rate (TPR) as the y-axis and False Positive Rate (FPR) as the x-axis. With the accuracy as the y-axis and the recall as the x-axis, we get the PR curve. The calculation equations of TPR and FPR are as follows:
T P R = T P T P + F N
F P R = F P F P + T N

5. Experiments

5.1. Training Parameters

Our network is an open-source framework based on Python 3.7, PyTorch 1.1.0. The experimental platform system is a Linux operating system (Ubuntu18.04), Intel Xeon E5-2620CPU@2.1 GHZ processor, 64 GB memory, Nvidia Titan Xp graphics card.
FAU-Net is trained for a total of 200 epochs. The learning efficiency set in the first 150 epochs is 0.001, and the learning efficiency of the last 50 epochs is 0.0001. This method of learning rate decay makes the model easier to approach the optimal solution. It consists of two 5-layer U-shaped networks, using Adam optimizer and cross-entropy loss function. In Fixupblock, we set initialization parameters bias term bias1a, bias2a, bias2a, bias2b all as 0, the scale as 1. The rate of the two drop-out layers is 0.25. We use a 3 × 3 convolution with padding equals 1. In Attentionblock, the number of intermediate channels is 512, and the drop-out rate is 0.1.

5.2. Database Preparation

To evaluate the quality of the network, in addition to use the common retinal blood vessel dataset DRIVE, we also constructed two cerebral blood vessel datasets, called ELE and CORO datasets. For the selection of brain blood vessel data set, we mainly consider the problem of class imbalance. Our region of interest is blood vessels, so when selecting pictures, we should choose the more evenly distributed blood vessels. Otherwise, there will be many patches with only background but no blood vessels, which will reduce the accuracy of network training. The brain vascular data set is different from the retinal vascular data set in that it may have a more significant imbalance problem. The difficulty of retinal vascular segmentation is that there is no significant difference between the retinal vascular segmentation and the background. Compared with the data set of retinal vascular, ELE, and CORO brain blood vessel data set have a significant gap between the background and blood vessels, but there is a lot of noise. Moreover, the structure of the brain blood vessels is not coherent. It is easy to break during the segmentation, which can be a good evaluation of the performance of the segmentation network. How to make the overall structure of blood vessels clearer is the main problem we need to solve.
DRIVE data set [4]: We use the retinal vascular public data set DRIVE, and the brain blood vessel 3D CTA image data set ELE, and the brain blood vessel MRA data set. Each of their original images is 565 × 584 pixels in size.
ELE data set: The ELE data set consists of 40 colorful CTA images of the brain of different patients at different angles. Each image corresponds to an artificially labeled vascular mask picture, 20 pictures for training, and 20 samples for testing. The size of the image is 565 × 584 pixels.
CORO data set: The CORO data set consists of 40 brain MRA images, 20 pictures for training, and 20 samples for testing, which are images of multiple patients and multiple different angles. Each image corresponds to an artificially labeled vascular mask map, and the image size is 565 × 584 pixels.
Each picture in these three data sets will generate a square data set during training (see Figure 4). In order to improve the reliability of training results, 190,000 patches were randomly selected, among which 10 % was used as the verification data, of which 171,000 patches were used for training, and the remaining 19,000 patches were used for verification. The size of each patch was 48 × 48 pixels.
Figure 4. From left to right are the data sets DRIVE,ELE,CORO. Above are patches of the original image, and below are patches of the output results.

5.3. Segmentation Results and Discussion

Figure 5 shows the results of segmenting using FAU-Net. The first line is the grayscale image of the input image, the second line is the ground truth, and the third line is the experimental output. The first column is the result of the segmentation of the blood vessels of the eyeball. Compared with the marked pictures, our network can segment the marked vessels very well, even smaller vessels that are not marked.
Figure 5. Complex vessel segmentation results. From left to right are the DRIVE, CORO, and ELE datasets.
The second column is the CORO segmentation result of the brain MRA vascular data set. From the segmentation results, we can distinguish between the blood vessel and the background noise, and there is less artifact information. Our segmentation makes the brain blood vessel the overall structure is clearer.
The third column is the segmentation result of our brain CTA image. The blood vessel structure of the CTA image is relatively simple, but it is difficult to distinguish the thinner and darker blood vessels from the image. Our network can distinguish these thinner and darker blood vessels well, enabling doctors to perform better medical analysis.

5.4. Comparative Experiments

We compare our network with the state-of-the-art models, such as U-Net, LadderNet, Dynamic Deep Networks, and IterNet, respectively, on three data sets, and use the standard evaluation method to evaluate.
Figure 6 shows the segmentation details of U-Net [3], LadderNet [16], Dynamic Deep Network [17], and our network on the three data sets of DRIVE, ELE, and CORO. See from a partially enlarged view of the image, our network can distinguish the retinal blood vessels from the background noise, and there is less noise in the boundary part of the eyeball. From the results of the segmentation of the ELE data set, our network can distinguish between thinner and darker one’s vascular structure. For the CORO data set, our segmentation made the vascular structure clearer and has fewer artifacts. Compared with other networks, the blood vessel structure segmented by our network is more complete and there are fewer broken blood vessels. In summary, we can say that our network has a more precise segmentation of small blood vessels than other networks, and has less background noise.
Figure 6. Visualization of the segmentation results on DRIVE, ELE, and CORO datasets.
Figure 7 shows the precision-recall curve and ROC curve obtained by running our network on the DRIVE data set. From the figure, we can see that the area under the precision-recall curve has reached 0.9100, and the area under the ROC curve reached 0.9853, indicating that our model is very robust. Figure 8 and Figure 9 show the precision-recall curve and ROC curve obtained by running our network on the ELE data set and the CORO data set, respectively, and their area under the curve has reached above 0.98.
Figure 7. Precision-recall curve and ROC curve experimentally obtained from FAU-Net on the DRIVE data set from left to right.
Figure 8. Precision-recall curve and ROC curve experimentally obtained from FAU-Net on the ELE data set from left to right.
Figure 9. From left to right, the Precision-recall curve and ROC curve experimentally obtained on the CORO data set by FAU-Net.
We compared the three data sets to verify the effect of our improved de-normalization structure on vascular segmentation. The approach with fixup means that we replace the original Basicblock with Fixupblock and do not use the channel attention mechanism. The data with the bold font in the table is the better data result. The last digit is the result of rounding. All data are the result of training 200 epochs.
Table 1 compares the results of the original U-net, LadderNet, and U-net using the de-normalized structure on the DRIVE data set. AUC has improved. LadderNet, which uses a de-normalized structure, has achieved better results in both Precision and F β -score. Table 2 and Table 3 compare the test results of U-net and LadderNet on the ELE and CORO data sets, respectively. Under the de-normalized structure, U-net’s various indicators under the ELE data set have improved and achieved better performance. On the CORO data set, the results of using the de-normalized u-net and the original u-net are not large, and the original LadderNet achieves better performance on F β -score and AUC. According to the results of these experiments, better segmentation accuracy can be obtained using the improved de-normalization method.
Table 1. Results on DRIVE data set.
Table 2. Results on ELE data set.
Table 3. Results on CORO data set.
Table 4 compares the experimental results of our network with the current state-of-the-art network for vascular image segmentation on three data sets. On two data sets, DRIVE and ELE, our network has made significant improvements to Accuracy, F β -score, and AUC. These indicators are more critical for vascular segmentation. On the CORO data set, our network also has significant improvements in the two indicators, Accuracy and AUC. The area under our network-trained ROC curve and the precision-recall curve has achieved Accuracy above 0.9 on all three data sets, which shows that our model is more efficient and robust.
Table 4. Experimental results of FAU-net compared with other networks on DRIVE, ELE, CORO data sets.
Table 5 shows the calculation time for the test phase of each sample. We compared the U-Net, Recurrent U-Net, Residual U-Net, LadderNet, Dynamic Deep Networks, IterNet(Patched), R2U-Net and FAU-Net on the three blood vessel data set DRIVE and the brain blood vessel data sets ELE and CORO, respectively. Results proved that our model could achieve high accuracy under high efficiency.
Table 5. Computational time for testing(Time (s)/sample).
Taking a comprehensive look at Table 1, Table 2 and Table 3 and Table 5, we can see that using this de-normalized structure not only can get good experimental results but also can significantly reduce the computational time for testing. It is significant for clinical applications.

6. Conclusions and Future Work

This paper proposes a de-normalized channel attention network based on a dual U-Net structure. De-normalized blocks and channel attention modules were added to accurately segment sophisticated blood vessels such as retinal blood vessels and brain blood vessels, and our experiments have achieved superior results. To test the impact of the de-normalized block on the segmentation network, we examined the effects using batch normalization and de-normalization on three data sets and compared them. Experiments show that our de-normalized block can improve network performance and robustness while achieving higher accuracy. In the next step, we will continue to refine our segmentation network and apply the segmentation results to the vascular navigation path planning.

Author Contributions

Methodology, D.H., L.Y., H.G., W.T. and T.R.W.; software, D.H., L.Y. and H.G.; writing—original draft preparation, D.H., L.Y., H.G., W.T. and T.R.W.; writing—review and editing, D.H., L.Y., T.R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shanghai Natural Science Foundation of China grant number 19ZR1419100, the National Natural Science Foundation of China grant number 61402278.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Japkowicz, N.; Stephen, S. The class imbalance problem: A systematic study. Intell. Data Anal. 2002, 6, 429–449. [Google Scholar] [CrossRef]
  2. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation policies from data. arXiv 2018, arXiv:1805.09501. [Google Scholar]
  3. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  4. Staal, J.; Abramoff, M.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef] [PubMed]
  5. Al-Amri, S.S.; Kalyankar, N.V. Image segmentation by using threshold techniques. arXiv 2010, arXiv:1005.4020. [Google Scholar]
  6. Gould, S.; Gao, T.; Koller, D. Region-based segmentation and object detection. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 7–10 December 2009; pp. 655–663. [Google Scholar]
  7. Sumengen, B.; Manjunath, B. Multi-scale edge detection and image segmentation. In Proceedings of the 2005 13th IEEE European Signal Processing Conference, Antalya, Turkey, 4–8 September 2005; pp. 1–4. [Google Scholar]
  8. Zhang, G.M.; Chen, S.P.; Liao, J.N. Otsu image segmentation algorithm based on morphology and wavelet transformation. In Proceedings of the 2011 3rd IEEE International Conference on Computer Research and Development, Shanghai, China, 11–13 March 2011; Volume 1, pp. 279–283. [Google Scholar]
  9. Maulik, U. Medical image segmentation using genetic algorithms. IEEE Trans. Inf. Technol. Biomed. 2009, 13, 166–173. [Google Scholar] [CrossRef] [PubMed]
  10. Fan, Y.; Jiang, T.; Evans, D.J. Volumetric segmentation of brain images using parallel genetic algorithms. IEEE Trans. Med. Imaging 2002, 21, 904–909. [Google Scholar] [PubMed]
  11. Wang, L.; Li, C.; Sun, Q.; Xia, D.; Kao, C.Y. Active contours driven by local and global intensity fitting energy with application to brain MR image segmentation. Comput. Med. Imaging Graph. 2009, 33, 520–531. [Google Scholar] [CrossRef] [PubMed]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  13. Korez, R.; Likar, B.; Pernuš, F.; Vrtovec, T. Model-based segmentation of vertebral bodies from MR images with 3D CNNs. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 433–441. [Google Scholar]
  14. Zhou, X.; Ito, T.; Takayama, R.; Wang, S.; Hara, T.; Fujita, H. Three-dimensional CT image segmentation by combining 2D fully convolutional network with 3D majority voting. In Deep Learning and Data Labeling for Medical Applications; Springer: Berlin/Heidelberg, Germany, 2016; pp. 111–120. [Google Scholar]
  15. Alom, M.Z.; Hasan, M.; Yakopcic, C.; Taha, T.M.; Asari, V.K. Recurrent residual convolutional neural network based on u-net (R2U-net) for medical image segmentation. arXiv 2018, arXiv:1802.06955. [Google Scholar]
  16. Zhuang, J. LadderNet: Multi-path networks based on U-Net for medical image segmentation. arXiv 2018, arXiv:1810.07810. [Google Scholar]
  17. Khanal, A.; Estrada, R. Dynamic Deep Networks for Retinal Vessel Segmentation. arXiv 2019, arXiv:1903.07803. [Google Scholar] [CrossRef]
  18. Li, L.; Verma, M.; Nakashima, Y.; Nagahara, H.; Kawasaki, R. IterNet: Retinal Image Segmentation Utilizing Structural Redundancy in Vessel Networks. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020. [Google Scholar]
  19. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  20. Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context encoding for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7151–7160. [Google Scholar]
  21. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. arXiv 2018, arXiv:1805.10180. [Google Scholar]
  22. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
  23. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  24. Sinha, A.; Dolz, J. Multi-scale guided attention for medical image segmentation. arXiv 2019, arXiv:1906.02849. [Google Scholar] [CrossRef] [PubMed]
  25. Bastidas, A.A.; Tang, H. Channel Attention Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–20 June 2019; pp. 881–888. [Google Scholar]
  26. Ghani, A.; See, C.; Sudhakaran, V.; Ahmad, J.; Abd-Alhameed, R. Accelerating retinal fundus image classification using artificial neural networks (ANNs) and reconfigurable hardware (FPGA). Electronics 2019, 8, 1522. [Google Scholar] [CrossRef]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  29. Zhang, H.; Dauphin, Y.N.; Ma, T. Fixup initialization: Residual learning without normalization. arXiv 2019, arXiv:1901.09321. [Google Scholar]
  30. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual Attention Network for Image Classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.