Next Article in Journal
Quantum Confinement Effect and Photoenhancement of Photoluminescence of PbS and PbS/MnS Quantum Dots
Previous Article in Journal
Structural and Magnetic Properties of NiZn Ferrite Nanoparticles Synthesized by a Thermal Decomposition Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FAU-Net: Fixup Initialization Channel Attention Neural Network for Complex Blood Vessel Segmentation

1
Shanghai Film Academy, Shanghai University, Shanghai 200072, China
2
The Faculty of Science, Design and Technology, Pool House, University of Bournemouth, Poole BH12 5BB, UK
3
The Faculty of Engineering and Informatics, University of Bradford, Bradford BD7 1DP, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(18), 6280; https://doi.org/10.3390/app10186280
Submission received: 29 July 2020 / Revised: 24 August 2020 / Accepted: 4 September 2020 / Published: 9 September 2020
(This article belongs to the Section Applied Biosciences and Bioengineering)

Abstract

:
Medical image segmentation based on deep learning is a central research issue in the field of computer vision. Many existing segmentation networks can achieve accurate segmentation using fewer data sets. However, they have disadvantages such as poor network flexibility and do not adequately consider the interdependence between feature channels. In response to these problems, this paper proposes a new de-normalized channel attention network, which uses an improved de-normalized residual block structure and a new channel attention module in the network for the segmentation of sophisticated vessels. The de-normalized network sends the extracted rough features to the channel attention network. The channel attention module can explicitly model the interdependence between channels and pay attention to the correlation with crucial information in multiple feature channels. It can focus on the channels with the most association with vital information among multiple feature channels, and get more detailed feature results. Experimental results show that the network proposed in this paper is feasible, is robust, can accurately segment blood vessels, and is particularly suitable for complex blood vessel structures. Finally, we compared and verified the network proposed in this paper with the state-of-the-art network and obtained better experimental results.

1. Introduction

Medical images have multiple modalities, such as Magnetic Resonance Angiography (MRA), Computed Tomography Angiography (CTA), Positron Emission Tomography (PET), ultrasound imaging, and more. In clinical diagnosis and treatment, the segmentation technology of medical images affects the reliability of diagnosis results to a great extent. Moreover, medical image segmentation technology is the first step of many medical image processing technologies, such as visualization, 3d reconstruction. Therefore, its development will affect the evolution of other related technologies in medical image processing.
With the rapid development of deep learning, a large number of intelligent methods based on neural networks for medical images segmentation have emerged in recent years. Nevertheless, medical image segmentation has some limitations, including data scarcity and class imbalance [1]. To get the results we want through deep learning training, we often need many medical images as a data set. For solving the problem of sparse data sets, different data transformation or enhancement techniques are sometimes used to increase the number of available labeled samples, such as the automatic data augmentation method proposed by Google in 2018 [2], or use a patch-based approach to resolve class imbalances.
To reduce large data sets required for network training and improve the efficiency of the work, U-Net [3] was proposed, so that we can complete the network training with very few data sets and accurately segment the region of interest. However, U-Net has some defects, such as the accuracy of the model is limited by the batch size, and the correlation between the channels is not fully considered.
To address these issues, we propose an improved de-normalized residual block. It uses a new initialization method, which solves the problem of gradient explosion and gradient disappearance by properly adjusting initialization parameters at the beginning of training. During the training process, it can converge faster. The attention mechanism, because of its low model complexity, can be calculated in parallel, and the advantage of being able to obtain critical information is also increasingly used in medical image segmentation tasks. This paper proposes a new channel attention module called SCA, which can calculate the feature mapping value in each channel, to explore the correlation of channel.
The contributions of our work can be summarized as below:
(1)
We propose a de-normalized channel attention network, which consists of a new channel attention module SCA and an improved de-normalized block.
(2)
We construct two brain vascular data sets, which uses multiple angles of CTA and MRA images from various patients, called the ELE and CORO data sets, respectively. The labeling process consults brain vascular experts.
(3)
We perform experiments on the public retinal vascular data set DRIVE [4] and brain vascular data sets ELE and CORO, compare the results with the state-of-the-art methods, and obtain high accuracy. The effects of de-normalized structure on vascular segmentation were tested on DRIVE, ELE, and CORO data sets, respectively.

2. Related Work

In the past, many traditional algorithms can be used for medical image segmentation, including threshold-based segmentation [5], region-based image segmentation [6], edge detection-based segmentation [7], wavelet analysis and wavelet transform image segmentation method [8], genetic algorithm-based image segmentation [9,10], active contour model-based segmentation method [11]. However, traditional segmentation methods have some problems, such as the need for human intervention in the segmentation process, prior knowledge before segmentation, the existence of over-segmentation in segmentation, and low segmentation accuracy.
Deep learning-based methods for image segmentation have made significant achievements. Their accuracy has surpassed that of traditional methods. Jonathan Long et al. [12] proposed a fully convolutional network for semantic segmentation to classify images at the pixel level, solving the problem of semantic level image segmentation method. Based on it, Korez et al. [13] proposed a 3DFCN network structure. They optimized the spine structure segmented by the 3D FCN network using a deformation model algorithm, which further improved the accuracy of the segmentation of spinal MR images. Zhou et al. [14] combined the FCN algorithm and the majority voting algorithm to segment 19 targets in the human torso CT image. However, the results obtained by FCN are still not subtle enough. It classifies each pixel, does not adequately consider the relationship between pixels, and is not sensitive to the details in the image, so it is not suitable for the segmentation of blood vessels.
U-Net [3] is undoubtedly one of the most successful medical image segmentation methods. After the emergence of U-net, many network variants appeared. For example, Alom et al. [15] proposed a recursive residual convolutional neural network (R2U-net) based on U-net. This residual network can achieve deeper Network, R2U-net can make the network extract more features than U-Net. LadderNet [16], proposed by Zhuang, uses an improved residual block, where two convolutional layers in one block share the same weight, and LadderNet has more information flow paths than U-Net. This allows it to get more accurate segmentation results. Similar to Ladder, Khanal et al. [17] proposed a dynamic deep network, which adds a small U-net to classify the blurred areas of the image in a more refined manner, thereby achieving higher accuracy. Currently, the network with the most top performance on the DRIVE data set is IterNet [18] proposed by Li et al. Its structure is very similar to the dynamic deep network, has a u-net, and a mini u-net. And the blurred details of the blood vessel to be divided can be found from the segmented image itself. It has obtained the highest AUC value of 0.9816 on the DRIVE data set [4]. However, these networks do not fully consider the correlation of each channel, and a lot of effective information will be lost during the training process, so this paper introduces a channel attention mechanism.
In many image segmentation tasks, the attention module is introduced. For example, SeNet proposed by He et al. [19], it automatically learns the importance of each feature channel through learning, then according to the importance to enhance useful features and suppress useless for the current task. Zhang et al. [20] proposed the EncNet structure, in which the Context Encoding Module is used to capture the semantic context of the scene and selectively emphasize the category-related feature maps. Li et al. [21] proposed the Feature Pyramid Attention module (FPA) and Global Attention Up-sample module (GAU) and introduced an attention mechanism for semantic segmentation. Fu et al. [22] proposed DANet to capture frame feature dependencies in space and channel dimensions. The position attention module can learn the correlation of spatial features, and the channel attention module can model the correlation of channels.
At present, there are many attention-related state-of-the-art models for medical image segmentation, but they are quite different from the implementation of our network. For example, a medical imaging attention gate (AG) model was designed in Attention U-Net proposed by Oktay et al. [23], which can automatically learn to focus on different shapes and sizes of target structures, learn to find the pancreas. Sinha et al. [24] proposed multi-scale guided attention for medical image segmentation. It obtains global features through a multi-size strategy, and then introduces the learned global features into the attention module. Bastidas et al. [25] proposed the channel attention network, which uses two u-nets to extract the features of the visible stream and infrared stream of the input image respectively, and then extracts the features to obtain the final segmentation results through the attention mechanism. The channel attention module proposed in this paper will enhance the useful features after each up-sampling and sub-sampling.
Yet, there is a common problem with the above networks, i.e., to obtain accurate segmentation results, it often requires a lot of training time, and the segmentation efficiency of the network is very low. Arfan et al. [26] proposed a method based on artificial neural networks and fully parallel field-programmable gate arrays (FPGAs). The hardware implementation proposed by this method can effectively improve the segmentation efficiency. In our paper, in order to solve the performance problem, we use a de-normalization method, which can greatly reduce training time.

3. Methods

3.1. Fixup Initialization Channel Attention Neural Network

Our de-normalized channel attention network follows the basic encoder-decoder structure of U-net. Its feature fusion method is to stitch the features together in the channel dimension to form a more “thick” feature. We want to be able to extract more elaborated features, so we adopt a dual U-Net structure with multiple pairs of skip connections. Its complete structure is shown in Figure 1.
The first de-normalized network uses an improved de-normalized residual block structure, which we call Fixupblock (Figure 2 right). Since it does not use batch normalization, the efficiency of the model is significantly improved. The second channel attention network adds the SCA module to Basicblock (Figure 3), called Attentionblock. We first input the patches of the original picture and the mask picture into the de-normalized network, which can quickly extract rough features of the blood vessel and send the feature values to the second channel attention network (Hereafter called CAU-Net). CAU-Net can model the interdependence between feature channels so that the weight of useful feature maps in the channel is higher than useless ones, and the weight of invalid or less useful feature maps is smaller so that features can be refined. The last layer of CAU-Net is the SoftMax layer, so the loss function we adopt is the negative log-likelihood loss function NLLLoss. Comparison experiments on three data sets with multiple state-of-the-art networks prove that our network has achieved the highest AUC value currently.

3.2. Improved De-Normalized Residuals Block

ResNet [27] was proposed in 2015, and its basic block is called the residual block. Residual block there are generally two kinds of forms, BasicBlock and Bottleneck, respectively. BasicBlock structure mainly uses two convolution sums of 3 × 3 . The branch of residual is:
OUT = CONV 3 × 3 + B N + R E L U + CONV 3 × 3 + B N
Then add it to the identity map and go through a RELU layer:
OUT = OUT + x + RELU
The structure of BasicBlock is shown on the left in Figure 2. The reason BasicBlock can work has a significant relationship with the BN layer it adds. Many residual networks cannot correctly account for the effects of residual connections without normalization techniques such as BatchNorm, resulting in gradient explosions. Lim et al. [28] experimentally verified that normalizing features using BN would eliminate the range flexibility of the network. Moreover, in the segmentation task of blood vessels, in order to obtain higher experimental accuracy, we often use larger batch size, which will significantly increase the memory usage of the GPU. A new initialization method [29] that does not require any normalization method was proposed. Based on this method, we propose an improved de-normalization block called Fixupblock. It appropriately adjusts the initial offset parameter at the beginning of training to ensure that the update of the network function (Gradient) stays in the proper range. It is independent of the network depth so that it can solve the problem of gradient explosion and gradient disappearance. At the same time, we add the two drop-out layers in Fixupblock to prevent some features from overfitting. It makes the weight update is no longer dependent on the combination of a stable relationship between hidden nodes. It avoids situations where certain features only work for certain other features, forcing the network to learn more robust features.
Improved Fixupblock structure shown in the right of Figure 2, where the bias term bias1a, bias2a, bias2a, bias2b all initialized to zero, the scale is initialized to one.

3.3. Channel Attention Module

Similar attention networks used for segmentation, such as [25,30], have bigger regions of interest and are more suitable for pathological segmentation. Our channel attention network can allocate the weight after each convolution and enhance the useful features, which is suitable for the segmentation of small blood vessels.
Generally, the output of the convolutional layer does not consider the channel dependence. We adopt the structure in SEblock [19], which is called Se calculation. It can enhance the effective features and suppress ineffective features between channels. We combine it with the channel attention module and propose a new channel attention module called SCA (Figure 3). Then add the SCA module to the second residual block of u-net and call it an Attentionblock (Figure 3). The following is the detailed calculation process of the channel attention module.
We assume that the original feature is F R C × H × W , and the original feature goes through three pipelines 2, 3, and 4 as shown in Figure 3. After passing through these three pipelines, the original feature first undergoes Squeeze-and-Excitation calculation (Hereinafter referred to as Se).
In Se calculation, the first step is the Squeeze operation F s q , which is the operation of global average pooling:
Z c = F s q f c = 1 H × W i = 1 H j = 1 W f c ( i , j )
Among them, f c represents the initial feature (i.e., the three-dimensional matrix F R C × H × W ), and the subscript c represents the c-th two-dimensional matrix in F, i.e., the channel. Equation (3) converts the input of C × H × W into 1 × 1 × C , which indicates the numerical distribution of the c feature maps of the layer, i.e., the global information. The Squeeze operation in Equation (3) operated in the feature map of a channel, and the operation in Equation (4) is to fuse the weight of the feature map of each channel, i.e., the Excitation operation F e x :
s = F e x ( z , W ) = σ ( g ( z , W ) ) = σ W 2 δ W 1 z
where z is the result of Equation (3), W 1 multiplied by z is the first fully connected layer operation, and the dimension of W 1 is C r × C , where r is a scaling parameter, to reduce the number of channels and thus reduce the amount of calculation. In this paper, r sets to 2. The dimension of z is 1 × 1 × C , then the dimension of W 1 z is 1 × 1 × C r . Then it goes through a RELU layer with the same dimensions, multiply it with W 2 , which is the operation of the second fully connected layer. The dimension of W 2 is c × c r , and the dimension of output is 1 × 1 × C . Finally, the dimension is unchanged after passing the sigmoid function. The obtained s (Equation (4)) is the feature map weight of the channel learned through the previous fully connected and non-linear layers. It is the result of pipeline 2, 3, 4 after Se operation.
In pipeline 4, we transform the initial characteristics F R C × H × W to F R C × N , where N = H × W represents the pixels of the picture. Pipeline 3 was reshaped after SE processed the original features. Then perform matrix multiplication on the reshaped F and the transpose of F passing through pipeline 4, and finally use a SoftMax layer to get the channel attention map X R α :
x i j = s o f t m a x ( F i · F j ) = e x p ( F i · F j ) i = 1 C e x p ( F i · F j )
where x i j represents the influence of the i-th channel on the j-th channel.
In addition to the above operation, we perform matrix multiplication on the transpose of X and F obtained through pipeline 2 and reshape the result to R C × H × W . Then multiply this result by a scale parameter β and perform a pixel-level addition with F of pipeline 1 to get E R C × H × W :
E j = β i = 1 C x j i F i + F j
Finally, the result E passes through a RELU and softmax layer, and the final feature is P R C × H × W .

4. Quantitative Assessment Method

The segmentation of blood vessels is a binary classification problem. In classification problems, we usually use some indicators to evaluate the pros and cons of the algorithm. In this article, we use the values of Accuracy (AC), Precision, Recall, and Area under the ROC curve (AUC) and F β s c o r e ( F β ) evaluated the performance of FAU-net. We will use True Positive (TP), True Negative (TN), False Positive (FP), recall rate and False Negative (FN) to calculate the above indicators. Performance metrics calculation is as follows:
A C = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
We assume that the positive samples are pixels marked as blood vessels and the negative samples are background. We assign different weights to recall and precision to indicate the preference for classification models. The calculation equation is as follows:
F β = 1 + β 2 · P r e c i s i o n · R e c a l l β 2 · P r e c i s i o n + R e c a l l
It can be seen that when β = 1, then F β changes back to F1; when β > 1 , the model prefers to improve recall, which means that the model pays more attention to the ability to recognize positive samples. when β < 1 , the model pays more attention to the ability to distinguish negative samples. In this article, we do not want the model to produce the wrong results to cover more samples. It will affect the doctor’s diagnosis of the patient’s condition. Therefore, we will adjust the parameter β to 0.75.
Considering the class imbalance of the data set, we also introduce Receiver Operating Characteristic (ROC) and Precision-Recall (PR) curves. The ROC curve is drawn with True Positive Rate (TPR) as the y-axis and False Positive Rate (FPR) as the x-axis. With the accuracy as the y-axis and the recall as the x-axis, we get the PR curve. The calculation equations of TPR and FPR are as follows:
T P R = T P T P + F N
F P R = F P F P + T N

5. Experiments

5.1. Training Parameters

Our network is an open-source framework based on Python 3.7, PyTorch 1.1.0. The experimental platform system is a Linux operating system (Ubuntu18.04), Intel Xeon [email protected] GHZ processor, 64 GB memory, Nvidia Titan Xp graphics card.
FAU-Net is trained for a total of 200 epochs. The learning efficiency set in the first 150 epochs is 0.001, and the learning efficiency of the last 50 epochs is 0.0001. This method of learning rate decay makes the model easier to approach the optimal solution. It consists of two 5-layer U-shaped networks, using Adam optimizer and cross-entropy loss function. In Fixupblock, we set initialization parameters bias term bias1a, bias2a, bias2a, bias2b all as 0, the scale as 1. The rate of the two drop-out layers is 0.25. We use a 3 × 3 convolution with padding equals 1. In Attentionblock, the number of intermediate channels is 512, and the drop-out rate is 0.1.

5.2. Database Preparation

To evaluate the quality of the network, in addition to use the common retinal blood vessel dataset DRIVE, we also constructed two cerebral blood vessel datasets, called ELE and CORO datasets. For the selection of brain blood vessel data set, we mainly consider the problem of class imbalance. Our region of interest is blood vessels, so when selecting pictures, we should choose the more evenly distributed blood vessels. Otherwise, there will be many patches with only background but no blood vessels, which will reduce the accuracy of network training. The brain vascular data set is different from the retinal vascular data set in that it may have a more significant imbalance problem. The difficulty of retinal vascular segmentation is that there is no significant difference between the retinal vascular segmentation and the background. Compared with the data set of retinal vascular, ELE, and CORO brain blood vessel data set have a significant gap between the background and blood vessels, but there is a lot of noise. Moreover, the structure of the brain blood vessels is not coherent. It is easy to break during the segmentation, which can be a good evaluation of the performance of the segmentation network. How to make the overall structure of blood vessels clearer is the main problem we need to solve.
DRIVE data set [4]: We use the retinal vascular public data set DRIVE, and the brain blood vessel 3D CTA image data set ELE, and the brain blood vessel MRA data set. Each of their original images is 565 × 584 pixels in size.
ELE data set: The ELE data set consists of 40 colorful CTA images of the brain of different patients at different angles. Each image corresponds to an artificially labeled vascular mask picture, 20 pictures for training, and 20 samples for testing. The size of the image is 565 × 584 pixels.
CORO data set: The CORO data set consists of 40 brain MRA images, 20 pictures for training, and 20 samples for testing, which are images of multiple patients and multiple different angles. Each image corresponds to an artificially labeled vascular mask map, and the image size is 565 × 584 pixels.
Each picture in these three data sets will generate a square data set during training (see Figure 4). In order to improve the reliability of training results, 190,000 patches were randomly selected, among which 10 % was used as the verification data, of which 171,000 patches were used for training, and the remaining 19,000 patches were used for verification. The size of each patch was 48 × 48 pixels.

5.3. Segmentation Results and Discussion

Figure 5 shows the results of segmenting using FAU-Net. The first line is the grayscale image of the input image, the second line is the ground truth, and the third line is the experimental output. The first column is the result of the segmentation of the blood vessels of the eyeball. Compared with the marked pictures, our network can segment the marked vessels very well, even smaller vessels that are not marked.
The second column is the CORO segmentation result of the brain MRA vascular data set. From the segmentation results, we can distinguish between the blood vessel and the background noise, and there is less artifact information. Our segmentation makes the brain blood vessel the overall structure is clearer.
The third column is the segmentation result of our brain CTA image. The blood vessel structure of the CTA image is relatively simple, but it is difficult to distinguish the thinner and darker blood vessels from the image. Our network can distinguish these thinner and darker blood vessels well, enabling doctors to perform better medical analysis.

5.4. Comparative Experiments

We compare our network with the state-of-the-art models, such as U-Net, LadderNet, Dynamic Deep Networks, and IterNet, respectively, on three data sets, and use the standard evaluation method to evaluate.
Figure 6 shows the segmentation details of U-Net [3], LadderNet [16], Dynamic Deep Network [17], and our network on the three data sets of DRIVE, ELE, and CORO. See from a partially enlarged view of the image, our network can distinguish the retinal blood vessels from the background noise, and there is less noise in the boundary part of the eyeball. From the results of the segmentation of the ELE data set, our network can distinguish between thinner and darker one’s vascular structure. For the CORO data set, our segmentation made the vascular structure clearer and has fewer artifacts. Compared with other networks, the blood vessel structure segmented by our network is more complete and there are fewer broken blood vessels. In summary, we can say that our network has a more precise segmentation of small blood vessels than other networks, and has less background noise.
Figure 7 shows the precision-recall curve and ROC curve obtained by running our network on the DRIVE data set. From the figure, we can see that the area under the precision-recall curve has reached 0.9100, and the area under the ROC curve reached 0.9853, indicating that our model is very robust. Figure 8 and Figure 9 show the precision-recall curve and ROC curve obtained by running our network on the ELE data set and the CORO data set, respectively, and their area under the curve has reached above 0.98.
We compared the three data sets to verify the effect of our improved de-normalization structure on vascular segmentation. The approach with fixup means that we replace the original Basicblock with Fixupblock and do not use the channel attention mechanism. The data with the bold font in the table is the better data result. The last digit is the result of rounding. All data are the result of training 200 epochs.
Table 1 compares the results of the original U-net, LadderNet, and U-net using the de-normalized structure on the DRIVE data set. AUC has improved. LadderNet, which uses a de-normalized structure, has achieved better results in both Precision and F β -score. Table 2 and Table 3 compare the test results of U-net and LadderNet on the ELE and CORO data sets, respectively. Under the de-normalized structure, U-net’s various indicators under the ELE data set have improved and achieved better performance. On the CORO data set, the results of using the de-normalized u-net and the original u-net are not large, and the original LadderNet achieves better performance on F β -score and AUC. According to the results of these experiments, better segmentation accuracy can be obtained using the improved de-normalization method.
Table 4 compares the experimental results of our network with the current state-of-the-art network for vascular image segmentation on three data sets. On two data sets, DRIVE and ELE, our network has made significant improvements to Accuracy, F β -score, and AUC. These indicators are more critical for vascular segmentation. On the CORO data set, our network also has significant improvements in the two indicators, Accuracy and AUC. The area under our network-trained ROC curve and the precision-recall curve has achieved Accuracy above 0.9 on all three data sets, which shows that our model is more efficient and robust.
Table 5 shows the calculation time for the test phase of each sample. We compared the U-Net, Recurrent U-Net, Residual U-Net, LadderNet, Dynamic Deep Networks, IterNet(Patched), R2U-Net and FAU-Net on the three blood vessel data set DRIVE and the brain blood vessel data sets ELE and CORO, respectively. Results proved that our model could achieve high accuracy under high efficiency.
Taking a comprehensive look at Table 1, Table 2 and Table 3 and Table 5, we can see that using this de-normalized structure not only can get good experimental results but also can significantly reduce the computational time for testing. It is significant for clinical applications.

6. Conclusions and Future Work

This paper proposes a de-normalized channel attention network based on a dual U-Net structure. De-normalized blocks and channel attention modules were added to accurately segment sophisticated blood vessels such as retinal blood vessels and brain blood vessels, and our experiments have achieved superior results. To test the impact of the de-normalized block on the segmentation network, we examined the effects using batch normalization and de-normalization on three data sets and compared them. Experiments show that our de-normalized block can improve network performance and robustness while achieving higher accuracy. In the next step, we will continue to refine our segmentation network and apply the segmentation results to the vascular navigation path planning.

Author Contributions

Methodology, D.H., L.Y., H.G., W.T. and T.R.W.; software, D.H., L.Y. and H.G.; writing—original draft preparation, D.H., L.Y., H.G., W.T. and T.R.W.; writing—review and editing, D.H., L.Y., T.R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shanghai Natural Science Foundation of China grant number 19ZR1419100, the National Natural Science Foundation of China grant number 61402278.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Japkowicz, N.; Stephen, S. The class imbalance problem: A systematic study. Intell. Data Anal. 2002, 6, 429–449. [Google Scholar] [CrossRef]
  2. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation policies from data. arXiv 2018, arXiv:1805.09501. [Google Scholar]
  3. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  4. Staal, J.; Abramoff, M.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef] [PubMed]
  5. Al-Amri, S.S.; Kalyankar, N.V. Image segmentation by using threshold techniques. arXiv 2010, arXiv:1005.4020. [Google Scholar]
  6. Gould, S.; Gao, T.; Koller, D. Region-based segmentation and object detection. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 7–10 December 2009; pp. 655–663. [Google Scholar]
  7. Sumengen, B.; Manjunath, B. Multi-scale edge detection and image segmentation. In Proceedings of the 2005 13th IEEE European Signal Processing Conference, Antalya, Turkey, 4–8 September 2005; pp. 1–4. [Google Scholar]
  8. Zhang, G.M.; Chen, S.P.; Liao, J.N. Otsu image segmentation algorithm based on morphology and wavelet transformation. In Proceedings of the 2011 3rd IEEE International Conference on Computer Research and Development, Shanghai, China, 11–13 March 2011; Volume 1, pp. 279–283. [Google Scholar]
  9. Maulik, U. Medical image segmentation using genetic algorithms. IEEE Trans. Inf. Technol. Biomed. 2009, 13, 166–173. [Google Scholar] [CrossRef] [PubMed]
  10. Fan, Y.; Jiang, T.; Evans, D.J. Volumetric segmentation of brain images using parallel genetic algorithms. IEEE Trans. Med. Imaging 2002, 21, 904–909. [Google Scholar] [PubMed]
  11. Wang, L.; Li, C.; Sun, Q.; Xia, D.; Kao, C.Y. Active contours driven by local and global intensity fitting energy with application to brain MR image segmentation. Comput. Med. Imaging Graph. 2009, 33, 520–531. [Google Scholar] [CrossRef] [PubMed]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  13. Korez, R.; Likar, B.; Pernuš, F.; Vrtovec, T. Model-based segmentation of vertebral bodies from MR images with 3D CNNs. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 433–441. [Google Scholar]
  14. Zhou, X.; Ito, T.; Takayama, R.; Wang, S.; Hara, T.; Fujita, H. Three-dimensional CT image segmentation by combining 2D fully convolutional network with 3D majority voting. In Deep Learning and Data Labeling for Medical Applications; Springer: Berlin/Heidelberg, Germany, 2016; pp. 111–120. [Google Scholar]
  15. Alom, M.Z.; Hasan, M.; Yakopcic, C.; Taha, T.M.; Asari, V.K. Recurrent residual convolutional neural network based on u-net (R2U-net) for medical image segmentation. arXiv 2018, arXiv:1802.06955. [Google Scholar]
  16. Zhuang, J. LadderNet: Multi-path networks based on U-Net for medical image segmentation. arXiv 2018, arXiv:1810.07810. [Google Scholar]
  17. Khanal, A.; Estrada, R. Dynamic Deep Networks for Retinal Vessel Segmentation. arXiv 2019, arXiv:1903.07803. [Google Scholar] [CrossRef]
  18. Li, L.; Verma, M.; Nakashima, Y.; Nagahara, H.; Kawasaki, R. IterNet: Retinal Image Segmentation Utilizing Structural Redundancy in Vessel Networks. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020. [Google Scholar]
  19. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  20. Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context encoding for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7151–7160. [Google Scholar]
  21. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. arXiv 2018, arXiv:1805.10180. [Google Scholar]
  22. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
  23. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  24. Sinha, A.; Dolz, J. Multi-scale guided attention for medical image segmentation. arXiv 2019, arXiv:1906.02849. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Bastidas, A.A.; Tang, H. Channel Attention Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–20 June 2019; pp. 881–888. [Google Scholar]
  26. Ghani, A.; See, C.; Sudhakaran, V.; Ahmad, J.; Abd-Alhameed, R. Accelerating retinal fundus image classification using artificial neural networks (ANNs) and reconfigurable hardware (FPGA). Electronics 2019, 8, 1522. [Google Scholar] [CrossRef] [Green Version]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  29. Zhang, H.; Dauphin, Y.N.; Ma, T. Fixup initialization: Residual learning without normalization. arXiv 2019, arXiv:1901.09321. [Google Scholar]
  30. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual Attention Network for Image Classification. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar]
Figure 1. Structure of FAU-net.
Figure 1. Structure of FAU-net.
Applsci 10 06280 g001
Figure 2. Resnet’s Basiclock structure on the left, and an improved Fixupblock structure on the right.
Figure 2. Resnet’s Basiclock structure on the left, and an improved Fixupblock structure on the right.
Applsci 10 06280 g002
Figure 3. Left is the overall structure of the Attentionblock, right is the structure of the channel attention module SCA.
Figure 3. Left is the overall structure of the Attentionblock, right is the structure of the channel attention module SCA.
Applsci 10 06280 g003
Figure 4. From left to right are the data sets DRIVE,ELE,CORO. Above are patches of the original image, and below are patches of the output results.
Figure 4. From left to right are the data sets DRIVE,ELE,CORO. Above are patches of the original image, and below are patches of the output results.
Applsci 10 06280 g004
Figure 5. Complex vessel segmentation results. From left to right are the DRIVE, CORO, and ELE datasets.
Figure 5. Complex vessel segmentation results. From left to right are the DRIVE, CORO, and ELE datasets.
Applsci 10 06280 g005
Figure 6. Visualization of the segmentation results on DRIVE, ELE, and CORO datasets.
Figure 6. Visualization of the segmentation results on DRIVE, ELE, and CORO datasets.
Applsci 10 06280 g006
Figure 7. Precision-recall curve and ROC curve experimentally obtained from FAU-Net on the DRIVE data set from left to right.
Figure 7. Precision-recall curve and ROC curve experimentally obtained from FAU-Net on the DRIVE data set from left to right.
Applsci 10 06280 g007
Figure 8. Precision-recall curve and ROC curve experimentally obtained from FAU-Net on the ELE data set from left to right.
Figure 8. Precision-recall curve and ROC curve experimentally obtained from FAU-Net on the ELE data set from left to right.
Applsci 10 06280 g008
Figure 9. From left to right, the Precision-recall curve and ROC curve experimentally obtained on the CORO data set by FAU-Net.
Figure 9. From left to right, the Precision-recall curve and ROC curve experimentally obtained on the CORO data set by FAU-Net.
Applsci 10 06280 g009
Table 1. Results on DRIVE data set.
Table 1. Results on DRIVE data set.
MethodsAccuracyPrecision F β -ScoreAUC
U-net0.95310.88520.81420.9755
U-net with fixup0.95610.86120.81960.9783
LadderNet0.95610.85930.82020.9793
LadderNet with fixup0.95720.86160.83050.9779
Table 2. Results on ELE data set.
Table 2. Results on ELE data set.
MethodsAccuracyPrecision F β -ScoreAUC
U-net0.98010.96310.94860.9875
U-net with fixup0.98910.97350.96850.9982
LadderNet0.98900.97090.96880.9983
LadderNet with fixup0.98960.97170.96920.9982
Table 3. Results on CORO data set.
Table 3. Results on CORO data set.
MethodsAccuracyPrecision F β -ScoreAUC
U-net0.98010.97840.96330.9808
U-net with fixup0.98470.99140.97210.9960
LadderNet0.98420.99060.97230.9962
LadderNet with fixup0.98430.99180.96950.9957
Table 4. Experimental results of FAU-net compared with other networks on DRIVE, ELE, CORO data sets.
Table 4. Experimental results of FAU-net compared with other networks on DRIVE, ELE, CORO data sets.
DatasetYearMethodsAccuracyPrecision F β -ScoreAUC
DRIVE2015U-net [3]0.95310.88520.81420.9755
2018Recurrent U-Net [15]0.95560.86030.81550.9782
2018Residual U-Net [15]0.95530.86140.81490.9779
2018R2U-Net [15]0.95560.85890.81710.9784
2018LadderNet [16]0.95610.85930.82020.9793
2019Dynamic Deep Networks [17]0.96930.82840.82590.9775
2020IterNet(Patched) [18]0.95730.85340.82050.9816
2020FAU-net0.96980.86510.83200.9853
ELE2015U-net0.98010.96310.94860.9875
2018Recurrent U-Net0.97760.97360.94650.9930
2018Residual U-Net0.98370.96700.95820.9902
2018R2U-Net0.98060.95010.96840.9955
2018LadderNet0.98890.97090.96850.9980
2019Dynamic Deep Networks0.98160.94350.96840.9942
2020IterNet(Patched)0.98210.96640.95680.9961
2020FAU-net0.98910.96950.96880.9982
CORO2015U-net0.98010.97840.96330.9808
2018Recurrent U-Net0.97390.98060.95410.9912
2018Residual U-Net0.97160.97920.94980.9755
2018R2U-Net0.97860.98920.96420.9893
2018LadderNet0.98430.99060.97230.9962
2019Dynamic Deep Networks0.97980.97580.96490.9897
2020IterNet(Patched)0.97390.98250.96540.9953
2020FAU-net0.98340.99320.97300.9928
Table 5. Computational time for testing(Time (s)/sample).
Table 5. Computational time for testing(Time (s)/sample).
DatasetDRIVEELECORO
U-Net2.16 s2.35 s2.05 s
Recurrent U-Net5.23 s4.56 s4.23 s
Residual U-Net3.57 s4.16 s3.13 s
R2U-Net5.75 s2.54 s5.34 s
IterNet (Patched)16.53 s14.45 s15.86 s
LadderNet1.78 s2.01 s1.41 s
Dynamic Deep Networks4.72 s4.23 s3.88 s
FAU-Net1.43 s1.27 s1.21 s

Share and Cite

MDPI and ACS Style

Huang, D.; Yin, L.; Guo, H.; Tang, W.; Wan, T.R. FAU-Net: Fixup Initialization Channel Attention Neural Network for Complex Blood Vessel Segmentation. Appl. Sci. 2020, 10, 6280. https://doi.org/10.3390/app10186280

AMA Style

Huang D, Yin L, Guo H, Tang W, Wan TR. FAU-Net: Fixup Initialization Channel Attention Neural Network for Complex Blood Vessel Segmentation. Applied Sciences. 2020; 10(18):6280. https://doi.org/10.3390/app10186280

Chicago/Turabian Style

Huang, Dongjin, Liwen Yin, Hao Guo, Wen Tang, and Tao Ruan Wan. 2020. "FAU-Net: Fixup Initialization Channel Attention Neural Network for Complex Blood Vessel Segmentation" Applied Sciences 10, no. 18: 6280. https://doi.org/10.3390/app10186280

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop