Next Article in Journal
A Granulation Tissue Detection Model to Track Chronic Wound Healing in DM Foot Ulcers
Next Article in Special Issue
Improvement of Image Stitching Using Binocular Camera Calibration Model
Previous Article in Journal
Multilevel Pyramid Network for Monocular Depth Estimation Based on Feature Refinement and Adaptive Fusion
Previous Article in Special Issue
An Improved YOLOv5s Algorithm for Object Detection with an Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LSW-Net: A Learning Scattering Wavelet Network for Brain Tumor and Retinal Image Segmentation

1
School of Artificial Intelligence, Chongqing University of Technology, Chongqing 401135, China
2
College of Science, Chongqing University of Technology, Chongqing 400054, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(16), 2616; https://doi.org/10.3390/electronics11162616
Submission received: 15 July 2022 / Revised: 16 August 2022 / Accepted: 18 August 2022 / Published: 20 August 2022
(This article belongs to the Special Issue Advances in Image Enhancement)

Abstract

:
Convolutional network models have been widely used in image segmentation. However, there are many types of boundary contour features in medical images which seriously affect the stability and accuracy of image segmentation models, such as the ambiguity of tumors, the variability of lesions, and the weak boundaries of fine blood vessels. In this paper, in order to solve these problems we first introduce the dual-tree complex wavelet scattering transform module, and then innovatively propose a learning scattering wavelet network model. In addition, a new improved active contour loss function is further constructed to deal with complex segmentation. Finally, the equilibrium coefficient of our model is discussed. Experiments on the BraTS2020 dataset show that the LSW-Net model has improved the Dice coefficient, accuracy, and sensitivity of the classic FCN, SegNet, and At-Unet models by at least 3.51%, 2.11%, and 0.46%, respectively. In addition, the LSW-Net model still has an advantage in the average measure of Dice coefficients compared with some advanced segmentation models. Experiments on the DRIVE dataset prove that our model outperforms the other 14 algorithms in both Dice coefficient and specificity measures. In particular, the sensitivity of our model provides a 3.39% improvement when compared with the Unet model, and the model’s effect is obvious.

1. Introduction

Image segmentation is a class of image processing problems, and its task is to divide an image into two or more meaningful regions. The accuracy of image segmentation is particularly important in practical applications. In particular, biomedical image segmentation is prominent in clinical analysis, diagnosis, treatment planning, and the measurement of disease progression. Traditional image segmentation methods, such as the threshold method [1], region growing method [2], level set method [3,4,5], etc., have struggled to meet the need for accurate image segmentation in the context of big data.
In recent years, deep neural networks have made great progress in various artificial intelligence tasks including image recognition and image segmentation. A convolutional neural network (CNN) [6] introduces semantic information when segmenting objects; thereby, injecting new vitality into semantic segmentation research. Fully convolutional network models [7,8,9] based on CNN architecture have achieved excellent performance in automatic medical image segmentation, which further promotes the application of deep learning in image segmentation for applications such as brain tumor segmentation [10]. SegNet [11] adopts the encoder–decoder structure and transfers the pixel index value of the maximum pooling operation in the encoding process into the decoder, which not only retains the detailed information of the pixels but also improves the accuracy of semantic segmentation. The Attention Unet network (At-Unet) [12] adds an attention gating unit to the Unet model to provide pixel level attention for the feature map. The network tends to focus on feature points with more information and improves the feature extraction ability of the model. The Deeplab network [13,14] obtains multiscale context information by cascading atrous convolutions with different atrous rates, and then introduces a conditional random field to enhance the relevance of contextual semantic information, which in turn improves the segmentation accuracy. Although the above-mentioned network models have improved the image segmentation accuracy of some datasets, they are still unable to accurately extract the boundary features of brain tumors in images. This is due to the invasiveness of the imaging process and the ambiguity between biological forms, as is the case between tumors and adjacent organs or changes in lesions over different periods. This invasiveness and ambiguity can lead to the discontinuity of some segmentation boundaries, as shown in Figure 1a. When comparing Figure 1a,b, there are many discontinuous segmentations in Figure 1a.
At the same time, many researchers are devoted to establishing the minimal loss energy function model. A level set function is proposed, that is, the region term of the CV energy function is used as the loss function, from which the CNN model can learn the spatial information of the image, which can improve the accuracy of image segmentation [14]. In order to solve the boundary error segmentation problem, an active contour loss function (AC-Loss) is constructed [15]. The AC-Loss function fully considers the internal and external areas of the segmented object, and the perimeter of the boundary. Unfortunately, some experiments have shown that when dealing with biomedical images with complex boundaries, such as retinal vessel images, because the AC-Loss function constrains the perimeter of the segmentation object boundary, it also limits the model’s ability to segment small boundaries. The under-segmentation phenomenon of fine blood vessels is avoided in Figure 1c. By comparing Figure 1c,d, there are many small blood vessels that can be seen in Figure 1d that are not segmented in Figure 1c.
The problem of complex boundary contour features in medical images, also increases the difficulty of image boundary feature extraction and characterization in deep neural network learning. Inspired by the dual-tree wavelet scattering transform, we propose a boundary feature extraction module which can improve the network’s ability to extract image boundary features. Specifically, the process can be described as follows: First, the dual-tree complex wavelet scattering transform is used to separate the high-frequency and low-frequency features of the feature map. Second, a convolution operator is adopted to extract low-frequency body features and high-frequency boundary features. Finally, the dual-tree complex wavelet scattering transform is then built into a fully convolutional network model, and a new learning scattering wavelet network (LSW-Net) semantic segmentation model is designed through end-to-end data-driven scattering learning transform features. In order to enhance its ability to extract image boundary contour information, the network utilizes the Unet network [8] as the backbone network and introduces the dual-tree complex wavelet scattering transform (DTCWT-Scat) during downsampling for boundary feature extraction. In order to further improve the network model’s ability to extract complex boundary contours, an improved active contour loss function (IAC-Loss) is further constructed on the basis of the LSW-Net network. This loss function not only improves the network’s sensitivity to small boundaries, it also better solves the problem of the under-segmentation of boundary contours.
Our main contributions are summarized as follows:
  • In order to separate the high-frequency and low-frequency features of the feature map during downsampling, we introduce the DTCWT-Scat module into the Unet and innovatively propose the LSW-Net model.
  • We design an improved active contour loss function, which can improve sensitivity to small boundaries and can better solve the problem of boundary under-segmentation.
  • Through BraTS brain tumor segmentation experiments, our LSW-Net network has advantages when compared with traditional FCN, SegNet, At-Unet, and some advanced segmentation algorithms in terms of Dice coefficient, accuracy, sensitivity, and other indicators.
  • Through the DRIVE retinal vessel segmentation experiments, the effectiveness and robustness of the LSW-Net + IAC-Loss model are illustrated.

2. Related Work

2.1. Dual-Tree Complex Scattering Wavelet Transform

Wavelet transform is a local waveform transform that can provide local representation of multiscale signals in both time and frequency domains. S. Mallat first proposed a wavelet scattering network with a non-feedback structure [16]. This network can not only present the image energy distribution in the frequency domain, but also maintain stability against small deformations. This partially makes up for the shortcomings of the CNN model, including small object segmentation and image boundary extraction capabilities. Some scholars have also actively tried to combine the wavelet algorithm with the CNN model. Oyallon [17] used a wavelet scattering network to replace the first layer of a residual network. The modified residual network produces roughly the same performance as the original residual network, but the training parameters are greatly reduced. Rodriguez [18] proposed a deep adaptive wavelet network to capture basic information from the input data for image classification. Through experiments on three image classification datasets, it was found that the model achieved high accuracy and also reduced training parameters. Recently, Cotter [19] proposed a dual-tree complex wavelet scattering network. After being combined with a CNN model, it achieves high accuracy in image classification tasks as well as fast inference ability. Figure 2 shows the output results of the first-order dual-tree complex wavelet scattering of brain tumor MRI images, including one low-frequency signal and high-frequency signals in six directions. The low-frequency signal is the main feature of the image, and the six high-frequency signals are the boundary feature of the image.

2.2. Related Loss Functions

While the widely used cross-entropy loss function (CE-Loss) is not sensitive to the segmentation of small object boundaries, when the existing model is trained, the network model will optimize its parameters using a gradient descent method according to the loss function error. Figure 3 shows that the CE-Loss function does not perform very well for cases with small boundaries or a small number of misclassified boundaries. To solve this problem, Williams [15] et al. proposed the AC-Loss loss energy function, which can be described as; where the Region item is the area of the segmentation region, the AC item is the boundary length of the segmentation object, the item is the area of the segmentation region, and the item is the boundary length of the segmentation object. In order to reduce false boundary segmentation, this energy function is expected to minimize the area energy of the segmentation region and the energy of the segmentation target boundary length during model training.
Unfortunately, it can be observed that the AC-Loss function is more sensitive than the CE-Loss term when there is a small boundary in complex medical images, such as the retinal vessel shown in Figure 3. However, when the target segmentation is completely correct, the AC-Loss term still maintains a high error value. As the model is further trained, this will further reduce the length of the segmentation target boundary resulting in under-segmentation for some boundaries.

3. Proposed Method

In this section, we first construct the DTCWT-Scat module and then propose a novel LSW-Net network model after introducing the DTCWT-Scat module into the Unet network. Furthermore, in order to solve the small target segmentation task, a new IAC-Loss function is designed. Finally, we document the LSW-Net algorithm and the IAC-Loss function calculation algorithm.

3.1. Learning Scattering Wavelet Network

Wavelet scattering can extract image texture features and boundary information but cannot make full use of contextual semantic information for image segmentation. FCN integrates multiscale contextual information through multilayer pooling and subsampling; however, it is still unable to distinguish boundary information from overall information. The natural solution of combining the two functional modules can not only enhance the complementarity between the boundary information and the global information but also improve the classification accuracy of image boundaries. Therefore, we designed a novel LSW-Net model that combines a wavelet scattering network and a fully convolutional network, which is based on the encoder–decoder structure of the fully convolutional network [20]. The LSW-Net framework can be described in detail as follows: First, the dual-tree complex wavelet scattering transform [19] is added during downsampling in order to effectively separate the high-frequency features and low-frequency features of the feature map. Second, the convolution operator is used to select the low-frequency main features and high-frequency boundary features of the feature map, respectively. Finally, we concentrate these features. The decoder is a process that uses a multilayer upsampling method to gradually restore its original resolution. The algorithm is shown in Algorithm 1.
Algorithm 1: Learning Scattering Wavelet Network
Input: Preprocess image, x ;
Num of encoder–decoder layers, m = 4 ;
Kernel size, k = 3 ;
Num of encoder kernels, n i = 64 × 2 i ;
Num of decoder kernels, n j = 64 × 2 j 1 ;
Output: Predictive segmentation map, u ;
initialization;
x 1 = F ( x , k , n i ) , ( i = 0 )
Encoder:
for i = 1 to m do
z i + 1 = DTCWT _ S c a t ( x i )
x i + 1 = F ( z i + 1 , k , n × 2 i )
end
Decoder:
d m + 1 = x m + 1
for j = m + 1 to 2 do
p j 1 = concate ( x j 1 , upsample ( d j ) )
d j 1 = F ( p j 1 , k , n × 2 j )
end
u = softmax ( c o n v ( d 1 , 1 ) )
return u
The LSW-Net framework contains a convolutional feature extraction module that is followed by batch normalization [21] after each convolution. The purpose is to accelerate the convergence speed of the LSW-Net framework and reduce the correlation between layers, see Figure 4 for details. The details can be described as follows: First, we use a 3 × 3 size kernel for convolution and batch normalization. Then, we use the ReLU function to activate and to achieve the purpose of nonlinear transformation. Finally, the above process is repeated once. The mathematical expression is F ( , k , n ) = [ ReLU ( n o r m ( c o n v ( , k , n ) ) ) ] 2 , where , k , n , [ ] 2 indicates, respectively, the input map, the size of the convolution kernel, the number of convolution kernels, and the convolution feature extraction module which is executed twice.

3.2. DTCWT-Scat Module

The DTCWT-Scat module can be described in detail as follows. First, the dual-tree complex wavelet transform is performed on the feature map. Then the low-frequency information is processed using average pooling low-pass filtering, that is, S = A v g p o o l ( y l , 2 ) , and the magnitude ( m a g ) of the high-frequency real and imaginary are also calculated, i.e., U = m a g ( Re ( y h ) , Im ( y h ) ) = ( Re ( y h ) ) 2 + ( Im ( y h ) ) 2 . Then, we merge S and U , as shown in Figure 4b, where y l , y h , Re ( ) , Im ( ) represents low-frequency information, high-frequency information, and real and imaginary operators, respectively.
The DTCWT-Scat module has two significant advantages. The first advantage is that it is able to perform a dual-tree complex wavelet transform on the input image. This transform supports the backpropagation of errors and can update the parameters so that the parameters of the previous convolution layer can be learned. Afterwards, the frequency domain features can be extracted. The second advantage is that the wavelet function has local waveform characteristics and is stable to local deformation. As a result, the LSW-Net model will be more stable and sensitive to small deformations in medical images such as tumors and will be more accurate for small feature extraction.

3.3. IAC-Loss Function

The flaws of CE-Loss and AC-Loss are acknowledged in Section 2.2. After absorbing the advantages of the AC-Loss function, we designed a contour segmentation minimum energy function, which can be written as follows,
min c 1 , c 2 R e g i o n
s.t. C | u | d s = C | v | d s .
where R e g i o n = Ω ( ( c 1 v ) 2 ( c 2 v ) 2 ) u d x , u , v , s , C , Ω represents the predicted image, segmentation image, curve arc length, segmentation contour curve, and image area, respectively. The variables c 1 , c 2 are constant variables. Since | | u | | v | | | ( u v ) | , there is,
C | | u | | v | | d s C | ( u v ) | d s .
Using the Lagrangian multiplier method, we construct a new contour segmentation energy function.
min c 1 , c 2 , C L o s s I A C = R e g i o n + α I A C ( L e n g t h ) ,
I A C ( L e n g t h ) = C | ( u v ) | d s ,
where α is the equilibrium coefficient. The first item is the area of the segmentation target area, and the second item is the difference between the target boundary length of the predicted image and the ground truth.
Figure 3 verifies the image segmentation advantages of the IAC-Loss energy function in complex backgrounds. It can be observed that when the segmentation target has no small target, I A C ( L e n g t h ) = C E ( L o s s ) = 0 , but A C ( L e n g t h ) = 314 , which indicates that IAC-Loss and CE-Loss will stop during minimization, but AC-Loss will continue to decrease. When the segmentation has small targets, I A C ( L e n g t h ) = 12 , C E ( L o s s ) = 1.109 , A C ( L e n g t h ) = 320 , which shows that IAC-Loss not only improves the sensitivity to small boundaries but also better solves the problem of under-segmentation, see Figure 3.
In order to facilitate numerical calculation, the specific calculation discrete are also written,
R e g i o n = Ω i = 1 , j = 1 u i , j ( c 1 v i , j ) 2 + Ω i = 1 , j = 1 ( 1 u i , j ) ( c 2 v i , j ) 2 ,
I A C ( L e n g t h ) = Ω i = 1 , j = 1 ( | u x i , j v x i , j | + | u y i , j v y i , j | ) ,
where α is the balance coefficient, u i j [ 0 , 1 ] is the predicted probability map, v i j { 0 , 1 } is the binary code of the ground truth, and c 1 , c 2 can be defined as a constant of 1 or 0. u x i , j , u y i , j , v x i , j , v y i , j are the differences of u i , j and v i , j in the horizontal and vertical directions, respectively. The algorithm flow is shown in Algorithm 2.
Algorithm 2: Improved AC-Loss function
Input: Predictive segmentation map, u ;
    Binary ground truth map, v ;
    Equilibrium coefficient, α ;
    Batch size, B ; Channels, C ;
    Image width, W ; Image height, H ;
Output: IAC-Loss Error, L o s s I A C ;
 initialization;
c i n = [ 1 ] B × C × W × H , c o u t = [ 0 ] B × C × W × H
R e g i o n i n = u × ( v c i n ) 2
R e g i o n o u t = ( 1 u ) × ( v c o u t ) 2
R e g i o n = R e g i o n i n + R e g i o n o u t
h x = h [ : , : , 1 : , : ] h [ : , : , : 1 , : ] ,   h = u , v
h y = h [ : , : , : , 1 : ] h [ : , : , : , : 1 ] ,   h = u , v
I A C ( L e n g t h ) = | u x v x | + | u y v y |
L o s s I A C = R e g i o n + α I A C ( L e n g t h )
return L o s s I A C

4. Experiments

In the following experiments, we use the DRIVE [22] and MICCAI-BraTS2020 [23] datasets. In the experimental results, the BraTS brain tumor segmentation evaluation metrics were recorded when e p o c h = 200 , and the DRIVE retinal blood vessel segmentation evaluation metrics were recorded when e p o c h = 10 .
All models are trained on an i7-10750H, NVIDIA RTX 2070 GPU with 8G RAM. The Python language is used for programming and the deep learning framework used is Pytorch.

4.1. Data Preprocessing and Evaluation Metrics

The BraTS2020 brain tumor dataset has 369 patient samples, and each patient contains 4 modalities of MRI image data. After splicing and slicing the four-modal data, slices are obtained. In this experiment, 297 samples are randomly selected as the training set and validation set, and the remaining 72 samples are reserved as the test set. After removing the slices without lesions there are still 19,874 slices, of which 80% of the slices are randomly selected as the training set and 20% of the slices are selected as the validation set. In the DRIVE retinal dataset, the first 20 images are selected as the training set and validation set and the last 20 images are used as the test set.
In the DRIVE retinal dataset, the first 20 images are selected as the training set and validation set, and the last 20 images are used as the test set. In this experiment, since there are only a small number of sample sets in the DRIVE retinal dataset, we preprocess the images of the training set according to image rotation, horizontal flip, vertical flip, translated, and random cropping, in order to expand the sample size of the training set.
In this paper, our model quality is evaluated in terms of the standard evaluation metrics such as precision, Dice coefficient, sensitivity, specificity, and accuracy, which are shown in Table 1. T P , F P , F N , T N represent true positives, false positives, false negatives, and true negatives, respectively.

4.2. Experiment 1: BraTS Brain Tumor Segmentation

In this subsection, we will compare the evaluation metrics of the LSW-Net model with FCN, SegNet, and At-Unet models, as well as the 3D visualization. We use a combination of binary cross entropy (BCE) and Dice loss to train the LSW-Net. The loss is formulated as:
l o s s B r a T s = l o s s D i c e + 0.5 l o s s B C E ,
where   l o s s D i c e = 1 2 y i y ^ i y i + y ^ i , l o s s B C E = 1 N [ y i · log ( y ^ i ) + ( 1 y i ) log ( y ^ i ) ] , y i { 0 , 1 } is the binary-coded value of the ground truth, and y ^ i [ 0 , 1 ] is the predicted value.
In Table 2, the evaluation metrics of the LSW-Net model are recorded, where ET, TC, WT, and AVG represent the enhanced tumor area, tumor core, the entire tumor area, and the average metric, respectively. After comparison to the classic FCN, SegNet, and At-Unet models, it can be observed that the Dice coefficient, accuracy, and sensitivity of the LSW-Net model are all excellent. The LSW-Net model improved the Dice coefficient, accuracy, and sensitivity by at least 3.51%, 2.11%, and 0.46%, respectively.
Figure 5 shows the 2D visualization comparison of the segmentation results of four brain tumor samples between the LSW-Net model and the classic FCN, SegNet, At-Unet models. After comparison with the ground truth, it can be observed that the LSW-Net model performs better than the three classical models in terms of segmentation and is more suitable for BraTS brain tumor dataset image segmentation. In the segmentation results in line one of Figure 5, it can be observed that the other three classic models have misclassified in the enhanced tumor area and the edema area. Conversely, the LSW-Net model has a clear and complete outline, which also shows the validity of the LSW-Net model.
The LSW-Net model segmentation results have fewer outliers and mis-segmented blocks, so they are closer to ground truth when compared to the 3D visualization of the classical FCN, SegNet, and At-Unet segmentation results. The 3D visualization in Figure 6 shows that the LSW-Net model has achieved a good overall segmentation effect. This contributes to a clearer understanding and judgment of tumor size, boundary, and location.
In addition to the experimental comparison of the LSW-Net with the classical segmentation algorithms, the performance of the LSW-Net model, using the BraTS brain tumor dataset, is assessed here against several advanced segmentation algorithms developed in recent years by researchers such as Zhang et al. [24], Li et al. [25], Feng et al. [26], Latif et al. [27] and Hao et al. [28], see Table 3. The comparison assesses performance in terms of the Dice coefficients for ET, TC, WT, and average (AVG). The evaluation metrics in Table 3 show that the LSW-Net model and these advanced segmentation algorithms have advantages and disadvantages in ET, TC, and WT indicators. However, the most important evaluation indicator is the average indicator of the Dice coefficient, that is, the AVG indicator. The AVG index of the LSW-Net model is the highest, which also shows that our model has better segmentation performance on the BraTS2020 dataset.

4.3. Experiment 2: DRIVE Retinal Segmentation

In this subsection, we will verify the effectiveness of the IAC-Loss function for segmentation on the DRIVE retina dataset. In the experiment, the LSW-Net model will be used as the backbone network and the loss function will be the IAC-Loss function, denoted as the LSW-Net + IAC-Loss model. Finally, the LSW-Net + IAC-Loss model segmentation results are compared with other 14 models, including Cheng et al. [29], Azzopardi et al. [30], Roychowdhury et al. [31], DRIU [32], HED [33], Unet [34], Recurrent Unet [34], R2Unet [34], Guo et al. [35], Du et al. [36], Arias et al. [37], Zou et al. [38], and MD-Net [39] models. In addition, two examples of segmentation effects are shown in terms of overall and local details. Compared with the other 14 models in Table 4, it can be seen that the LSW-Net + IAC-Loss model is higher than the other 14 algorithms in terms of Dice coefficient and specificity; it is second only to the MD-Net [39] model in the accuracy index. Compared with the segmentation results of the classic Unet model, the sensitivity of the LSW-Net model offers an improvement of 3.39%, which is an obvious improvement and has a significant effect. These advantages indicate that our model performs well.
The comparison experiment, using the segmentation results from the DRIVE retinal blood vessel dataset, is shown in Figure 7. In the first and third rows of Figure 7 it can be observed that the segmentation results of DRIU [32] and HED [33] have obvious over-segmentation. In lines two and four of Figure 7, it can be seen that the segmentation results of the LSW-Net + IAC-Loss model have less noise and clearer contours. Through this experimental comparison, it can be shown that the LSW-Net + IAC-Loss model has better segmentation effectiveness.

4.4. Experiment 3: Discussion on Equilibrium Coefficient α

To evaluate the effect of the balance coefficient α in the LSW-Net + IAC-Loss model, ablation experiments are performed on the DRIVE dataset in this subsection.
In the experiment, the evaluation metrics of the LSW-Net + IAC-Loss model are recorded in Table 5, where α [ 0.1 , 0.5 ] . Table 5 shows that the differences in each metric are not obvious; however, they all reach a high level, which also shows that the LSW-Net + IAC-Loss model has better robustness to the balance coefficient α . When α = 0.3 , the specificity is the highest while the Dice coefficient and sensitivity are relatively low. Alternatively, when α = 0.1 or α = 0.5 , the Dice coefficient and sensitivity index values increase. Therefore, we suggest that the metrics can be fine-tuned by controlling the balance coefficient α according to actual needs.

4.5. Experiment 4: IAC-Loss Effectiveness Evaluation

In this subsection, we further evaluate the advantages of the IAC-Loss function. For the segmentation of the DRIVE retina dataset, the LSW-Net is used as the backbone network and the loss functions are the CE-Loss, AC-Loss, and IAC-Loss functions, denoted as +CE-Loss, +AC-Loss, and +IAC-Loss models, respectively. The segmentation results are shown in Table 6.
Compared with the +AC-Loss model, the +IAC-Loss model improves on the Dice coefficient and sensitivity by 3.56% and 8.1%, respectively. The accuracy is also increased by 0.56%. This illustrates the effectiveness of the IAC-Loss function for image segmentation, see Table 6.
After comparing the enlarged details of lines two and four in Figure 8, it can be seen that the boundary contour segmentation of the +IAC-Loss model is the best. Through the comparative experiments above, it can be determined that the IAC-Loss function has greater advantages for complex image boundary contours.

5. Conclusions

In this research, we have proposed the LSW-Net model for the BraTS2020 dataset, which achieved good experimental simulation results on the segmentation discontinuity problem. We have constructed an LSW-Net + IAC-Loss model in order to solve the weak boundary problem of small blood vessels in the DRIVE retinal vessel dataset. After introducing the dual-tree complex wavelet transform, the experimental results show that the LSW-Net has the ability to extract features and achieve better segmentation results. In the future we will further integrate the attention mechanism and the transformer method to design a better image segmentation network model.

Author Contributions

Methodology, R.L.; software, H.N.; validation, H.N. and Y.Z.; formal analysis, R.L.; resources, R.L., T.X. and Z.Y.; project administration, R.L.; funding acquisition, R.L. and Z.Y.; data curation, T.X. and Z.Y.; writing—review and editing, R.L.; visualization, H.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by: Chongqing Natural Science Foundation under Grants No. cstc2019jcyj-msxmX0500 and No. cstc2019jcyj-msxmX0240; Technology Research Program of Chongqing Education Commission under Grand No. KJQN202001129.

Data Availability Statement

The DRIVE and BraTS2020 datasets are publicly available [22,23], https://drive.grand-challenge.org/. https://www.med.upenn.edu/cbica/brats2020/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kaur, T.; Saini, B.S.; Gupta, S. A novel fully automatic multilevel thresholding technique based on optimized intuitionistic fuzzy sets and tsallis entropy for MR brain tumor image segmentation. Australas. Phys. Eng. Sci. Med. 2018, 41, 41–58. [Google Scholar] [CrossRef]
  2. Sukanya, A.; Rajeswari, R.; Murugan, K.S. Region based coronary artery segmentation using modified frangi’s vesselness measure. Int. J. Imaging Syst. Technol. 2020, 30, 716–730. [Google Scholar] [CrossRef]
  3. Chen, Y.; Chen, G.; Wang, Y.; Dey, N.; Sherratt, R.S.; Shi, F. A distance regularized level-set evolution model based MRI dataset segmentation of brain’s caudate nucleus. IEEE Access 2019, 7, 124128–124140. [Google Scholar] [CrossRef]
  4. Chan, T.F.; Vese, L.A. Active contours without edges. IEEE Trans. Image Process. 2001, 10, 266–277. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Lie, J.; Lysaker, M.; Tai, X.-C. A binary level set model and some applications to mumford-shah image segmentation. IEEE Trans. Image Process. 2006, 15, 1171–1181. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef]
  7. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  8. Ronneberger, O.; Fischer, P.; Brox, T. Unet: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
  9. Cicek, O.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3d Unet: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin, Germany, 2016; pp. 424–432. [Google Scholar]
  10. Shrestha, H.; Dhasarathan, C.; Kumar, M.; Nidhya, R.; Shankar, A.; Kumar, M. A deep learning based convolution neural network-DCNN approach to detect brain tumor. In Proceedings of the Academia-Industry Consortium for Data Science, Wenzhou, China, 19–20 December 2020; Springer: Singapore, 2022; pp. 115–127. [Google Scholar]
  11. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  12. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef]
  13. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  14. Kim, Y.; Kim, S.; Kim, T.; Kim, C. CNN-based semantic segmentation using level set loss. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; IEEE: New York, NY, USA, 2019; pp. 1752–1760. [Google Scholar]
  15. Chen, X.; Williams, B.M.; Vallabhaneni, S.R.; Czanner, G.; Williams, R.; Zheng, Y. Learning active contour models for medical image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11632–11640. [Google Scholar]
  16. Bruna, J.; Mallat, S. Invariant scattering convolution networks. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1872–1886. [Google Scholar] [CrossRef] [Green Version]
  17. Oyallon, E.; Belilovsky, E.; Zagoruyko, S. Scaling the scattering transform: Deep hybrid networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5618–5627. [Google Scholar]
  18. Rodriguez, M.X.B.; Gruson, A.; Polania, L.; Fujieda, S.; Prieto, F.; Takayama, K.; Hachisuka, T. Deep adaptive wavelet network. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 3111–3119. [Google Scholar]
  19. Cotter, F. Uses of Complex Wavelets in Deep Convolutional Neural Networks. Ph.D. Thesis, University of Cambridge, Cambridge, UK, 2020. [Google Scholar]
  20. Pathak, D.; Krahenbuhl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context encoders: Feature learning by inpainting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2536–2544. [Google Scholar]
  21. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  22. Staal, J.; Abramoff, M.D.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef] [PubMed]
  23. Menze, B.H.; Jakab, A.; Bauer, S.; Cramer, J.K.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (brats). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef] [PubMed]
  24. Zhang, C.; Shen, X.; Cheng, H.; Qian, Q. Brain tumor segmentation based on hybrid clustering and morphological operations. Int. J. Biomed. Imaging 2019, 1–11. [Google Scholar] [CrossRef]
  25. Li, H.; Li, A.; Wang, M. A novel end-to-end brain tumor segmentation method using improved fully convolutional networks. Comput. Biol. Med. 2019, 108, 150–160. [Google Scholar] [CrossRef]
  26. Bowen, F.; Qi, L.X.; Yu, G.; Qing, L.; Yang, L. Three-dimensional parallel convolution neural network brain tumor segmentation based on dilated convolution. Laser Optoelectron. Prog. 2020, 57, 141009. [Google Scholar] [CrossRef]
  27. Latif, U.; Shahid, A.R.; Raza, B.; Ziauddin, S.; Khan, M.A. An end-to-end brain tumor segmentation system using multi-inception-Unet. Int. J. Imaging Syst. Technol. 2021, 31, 1803–1816. [Google Scholar] [CrossRef]
  28. Hao, K.; Lin, S.; Qiao, J.; Tu, Y. A generalized pooling for brain tumor segmentation. IEEE Access 2021, 9, 159283–159290. [Google Scholar] [CrossRef]
  29. Cheng, E.; Du, L.; Wu, Y.; Zhu, Y.J.; Megalooikonomou, V.; Ling, H. Discriminative vessel segmentation in retinal images by fusing contextaware hybrid features. Mach. Vis. Appl. 2014, 25, 1779–1792. [Google Scholar] [CrossRef]
  30. Azzopardi, G.; Strisciuglio, N.; Vento, M.; Petkov, N. Trainable cosfire filters for vessel delineation with application to retinal images. Med. Image Anal. 2015, 19, 46–57. [Google Scholar] [CrossRef] [Green Version]
  31. Roychowdhury, S.; Koozekanani, D.D.; Parhi, K.K. Blood vessel segmentation of fundus images by major vessel extraction and sub-image classification. IEEE J. Biomed. Health Inform. 2014, 19, 1118–1128. [Google Scholar] [CrossRef]
  32. KManinis, K.; Tuset, J.P.; Arbelaez, P.; Gool, L.V. Deep retinal image understanding. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin, Germany, 2016; pp. 140–148. [Google Scholar]
  33. Xie, S.; Tu, Z. Holistically-nested edge detection. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1395–1403. [Google Scholar]
  34. Alom, Z.; Yakopcic, C.; Hasan, M.; Taha, T.M.; Asari, V.K. Recurrent residual Unet for medical image segmentation. J. Med. Imaging 2019, 6, 014006. [Google Scholar] [CrossRef] [PubMed]
  35. Guo, X.; Chen, C.; Lu, Y.; Meng, K.; Chen, H.; Zhou, K.; Wang, Z.; Xiao, R. Retinal vessel segmentation combined with generative adversarial networks and dense Unet. IEEE Access 2020, 8, 194551–194560. [Google Scholar] [CrossRef]
  36. XDu, X.-F.; Wang, J.-S.; Sun, W.-Z. Unet retinal blood vessel segmentation algorithm based on improved pyramid pooling method and attention mechanism. Phys. Med. Biol. 2021, 66, 175013. [Google Scholar]
  37. Arias, M.E.G.; Santos, D.M.; Borrero, I.P.; Vazquez, M.J.V. A new deep learning method for blood vessel segmentation in retinal images based on convolutional kernels and modified Unet model. Comput. Methods Programs Biomed. 2021, 205, 106081. [Google Scholar] [CrossRef]
  38. Zou, B.; Dai, Y.; He, Q.; Zhu, C.; Liu, G.; Su, Y.; Tang, R. Multi-label classification scheme based on local regression for retinal vessel segmentation. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 18, 2586–2597. [Google Scholar] [CrossRef] [PubMed]
  39. Shi, Z.; Wang, T.; Huang, Z.; Xie, F.; Liu, Z.; Wang, B.; Xu, J. MD-Net: A multi-scale dense network for retinal vessel segmentation. Biomed. Signal Process. Control 2021, 70, 102977. [Google Scholar] [CrossRef]
  40. Yan, H.; Xie, J.; Yue, X.; Wang, J.; Guo, S. MFE-Net: Multi-type feature enhancement net for retinal blood vessel segmentation. In Proceedings of the 2022 5th International Conference on Artificial Intelligence and Big Data (ICAIBD), Chengdu, China, 27–30 May 2022; pp. 51–56. [Google Scholar]
Figure 1. Medical image segmentation. (a) At-Unet; (b) Brain tumor ground truth; (c) AC-Loss; (d) Retinal ground truth, where the yellow boxes indicate the regions where the methods are mis-segmented compared to ground truth.
Figure 1. Medical image segmentation. (a) At-Unet; (b) Brain tumor ground truth; (c) AC-Loss; (d) Retinal ground truth, where the yellow boxes indicate the regions where the methods are mis-segmented compared to ground truth.
Electronics 11 02616 g001
Figure 2. Visualization of the first-order dual-tree complex wavelet scattering transform of brain tumor MRI images.
Figure 2. Visualization of the first-order dual-tree complex wavelet scattering transform of brain tumor MRI images.
Electronics 11 02616 g002
Figure 3. Comparison example of boundary contour segmentation. (a) CE-Loss; (b) AC-Loss; (c) IAC-Loss.
Figure 3. Comparison example of boundary contour segmentation. (a) CE-Loss; (b) AC-Loss; (c) IAC-Loss.
Electronics 11 02616 g003
Figure 4. LSW-Net model. (a) LSW-Net; (b) DTCWT-Scat module.
Figure 4. LSW-Net model. (a) LSW-Net; (b) DTCWT-Scat module.
Electronics 11 02616 g004
Figure 5. Comparison of segmentation results of four brain tumor samples, in which red, yellow, and green indicate necrotic area, enhanced tumor area, and edema area, respectively. (a) FCN; (b) SegNet; (c) At-Unet; (d) LSW-Net (Ours); (e) Ground truth.
Figure 5. Comparison of segmentation results of four brain tumor samples, in which red, yellow, and green indicate necrotic area, enhanced tumor area, and edema area, respectively. (a) FCN; (b) SegNet; (c) At-Unet; (d) LSW-Net (Ours); (e) Ground truth.
Electronics 11 02616 g005
Figure 6. 3D visualization comparison of the segmentation results of two brain tumors, in which red, yellow, and green indicate the necrotic area, the enhanced tumor area, and the edema area, respectively. (a) FCN; (b) SegNet; (c) At-Unet; (d) LSW-Net (Ours); (e) Ground truth.
Figure 6. 3D visualization comparison of the segmentation results of two brain tumors, in which red, yellow, and green indicate the necrotic area, the enhanced tumor area, and the edema area, respectively. (a) FCN; (b) SegNet; (c) At-Unet; (d) LSW-Net (Ours); (e) Ground truth.
Electronics 11 02616 g006
Figure 7. Comparison of LSW-Net + IAC-Loss with advanced segmentation models results on DRIVE. Lines 2 and 4 are the local details. (a) Test Image; (b) HED; (c) DRIU; (d) LSW-Net + IAC-Loss (Ours); (e) Ground truth.
Figure 7. Comparison of LSW-Net + IAC-Loss with advanced segmentation models results on DRIVE. Lines 2 and 4 are the local details. (a) Test Image; (b) HED; (c) DRIU; (d) LSW-Net + IAC-Loss (Ours); (e) Ground truth.
Electronics 11 02616 g007
Figure 8. Comparison of +IAC-Loss with related loss function results. Lines 2 and 4 are the local details respectively. (a) Test Image; (b) +AC-Loss; (c) +CE-Loss; (d) +IAC-Loss; (e) Ground truth.
Figure 8. Comparison of +IAC-Loss with related loss function results. Lines 2 and 4 are the local details respectively. (a) Test Image; (b) +AC-Loss; (c) +CE-Loss; (d) +IAC-Loss; (e) Ground truth.
Electronics 11 02616 g008
Table 1. Evaluation metrics.
Table 1. Evaluation metrics.
MetricDescription
Pre (Precision) T P F P + T P
Dice (Dice coefficient) 2 T P 2 T P + F P + F N
Sen (Sensitivity) T P T P + F N
Spe (Specificity) T N F P + T N
Acc (Accuracy) T P + T N F P + F N + T P + T N
Table 2. Comparison of LSW-Net model with classical segmentation algorithms on BraTS2020.
Table 2. Comparison of LSW-Net model with classical segmentation algorithms on BraTS2020.
MethodPre Dice Sen
ETTCWTAVGETTCWTAVGETTCWTAVG
FCN [7]0.76500.65540.78310.73450.76560.68020.81250.75280.81970.79040.87220.8274
SegNet [11]0.77480.70760.86690.78310.73160.69840.84480.75830.76150.77540.84640.7944
At-Unet [12]0.77640.72350.87910.79300.76460.73120.86000.78530.80800.82400.86650.8328
LSW-Net (Ours)0.83190.74470.90770.82810.79470.74480.87970.80640.81250.83080.86900.8374
Table 3. Comparison of LSW-Net model with some advanced algorithms on BraTS2020.
Table 3. Comparison of LSW-Net model with some advanced algorithms on BraTS2020.
MethodYearDice
ETTCWTAVG
Zhang et al. [24]20190.70700.73800.88500.7767
Li et al. [25]20190.74500.80800.86500.8060
Feng et al. [26]20200.71000.73000.90000.7800
Latif et al. [27]20210.71800.74600.89600.7860
Hao et al. [28]20210.79260.74650.87640.8051
LSW-Net (Ours) 0.79470.74480.87970.8064
Table 4. Comparison of LSW-Net + IAC-Loss model with some advanced models on DRIVE.
Table 4. Comparison of LSW-Net + IAC-Loss model with some advanced models on DRIVE.
MethodYearDiceSenSpeAcc
Cheng et al. [29]2014-0.72520.97980.9474
Azzopardi et al. [30]2015-0.76550.97040.9442
Roychowdhury et al. [31]2016-0.72500.98300.9520
DRIU [32]20160.67010.96960.91150.9165
HED [33]20170.64000.95630.90070.9054
Unet [34]20190.81420.75370.98200.9553
Recurrent Unet [34]20190.81550.77510.98160.9556
R2Unet [34]20190.81710.77920.98130.9556
Guo et al. [35]20200.82150.82830.97260.9542
Du et al. [36]2021-0.78140.98100.9556
Arias et al. [37]2021-0.85970.96900.9563
Zou et al. [38]20210.81290.77610.97920.9519
MD-Net [39]20210.80990.80650.98260.9676
MFE-Net [40]20220.82040.78530.98120.9563
LSW-Net + IAC-Loss (Ours) 0.82160.78760.98370.9565
Table 5. Influence of α balance coefficient on LSW-Net + IAC-Loss model segmentation result indicators.
Table 5. Influence of α balance coefficient on LSW-Net + IAC-Loss model segmentation result indicators.
αPreDiceSenSpeAcc
0.10.85250.82310.95650.79570.9799
0.20.85420.82220.95640.79250.9802
0.30.85880.82160.95650.78760.9837
0.40.86020.82210.95660.78730.9813
0.50.85710.82270.95660.79090.9807
Table 6. Comparison of IAC-Loss with related loss function metrics.
Table 6. Comparison of IAC-Loss with related loss function metrics.
DiceSenSpeAcc
+AC-Loss0.78750.71470.98530.9509
+CE-Loss0.81820.79200.97890.9551
+IAC-Loss (α = 0.1)0.82310.79570.97990.9565
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, R.; Nan, H.; Zou, Y.; Xie, T.; Ye, Z. LSW-Net: A Learning Scattering Wavelet Network for Brain Tumor and Retinal Image Segmentation. Electronics 2022, 11, 2616. https://doi.org/10.3390/electronics11162616

AMA Style

Liu R, Nan H, Zou Y, Xie T, Ye Z. LSW-Net: A Learning Scattering Wavelet Network for Brain Tumor and Retinal Image Segmentation. Electronics. 2022; 11(16):2616. https://doi.org/10.3390/electronics11162616

Chicago/Turabian Style

Liu, Ruihua, Haoyu Nan, Yangyang Zou, Ting Xie, and Zhiyong Ye. 2022. "LSW-Net: A Learning Scattering Wavelet Network for Brain Tumor and Retinal Image Segmentation" Electronics 11, no. 16: 2616. https://doi.org/10.3390/electronics11162616

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop