Next Article in Journal
A Study on Crowd Evacuation Model Considering Squeezing Equilibrium in Crowded Areas
Next Article in Special Issue
Automated Classification of Urinary Cells: Using Convolutional Neural Network Pre-trained on Lung Cells
Previous Article in Journal
Fock-Space Schrieffer–Wolff Transformation: Classically-Assisted Rank-Reduced Quantum Phase Estimation Algorithm
Previous Article in Special Issue
Quantitative Analysis of Retinal Vascular Leakage in Retinal Vasculitis Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Hepatic Vessels Segmentation Using RORPO Vessel Enhancement Filter and 3D V-Net with Variant Dice Loss Function

IT4Innovations, VSB—Technical University of Ostrava, 70800 Ostrava-Poruba, Czech Republic
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(1), 548; https://doi.org/10.3390/app13010548
Submission received: 18 October 2022 / Revised: 20 December 2022 / Accepted: 21 December 2022 / Published: 30 December 2022

Abstract

:
The segmentation of hepatic vessels is crucial for liver surgical planning. It is also a challenging task because of its small diameter. Hepatic vessels are often captured in images of low contrast and resolution. Our research uses filter enhancement to improve their contrast, which helps with their detection and final segmentation. We have designed a specific fusion of the Ranking Orientation Responses of Path Operators (RORPO) enhancement filter with a raw image, and we have compared it with the fusion of different enhancement filters based on Hessian eigenvectors. Additionally, we have evaluated the 3D U-Net and 3D V-Net neural networks as segmentation architectures, and have selected 3D V-Net as a better segmentation architecture in combination with the vessel enhancement technique. Furthermore, to tackle the pixel imbalance between the liver (background) and vessels (foreground), we have examined several variants of the Dice Loss functions, and have selected the Weighted Dice Loss for its performance. We have used public 3D Image Reconstruction for Comparison of Algorithm Database (3D-IRCADb) dataset, in which we have manually improved upon the annotations of vessels, since the dataset has poor-quality annotations for certain patients. The experiments demonstrate that our method achieves a mean dice score of 76.2%, which outperforms other state-of-the-art techniques.

1. Introduction

Liver vessel segmentation from Computed Tomography (CT) images is important before liver surgery resection. Clinicians must know the liver’s morphology and its venous system location and diameter to plan the path of the surgical cutting. Hence, the segmentation of the small structures such as hepatic veins is a crucial task. These structures have a small diameter, and low image contrast and resolution. Today, accurate hepatic vessel labeling still relies primarily on doctors’ manual segmentation, which is time-consuming and depends on the specialists’ expertise and skills. As a result, an autonomous, robust, and accurate hepatic vascular segmentation algorithm is critical and highly desired.
In this work, we address the issues of low-contrast vessels and labour-intensive manual vessel segmentation. We develop a fully automatic segmentation method that provides high accuracy and precision for hepatic veins segmentation. The solution does not require a doctor’s assistance, such as in semi-automatic methods.
Although many vessel segmentation approaches have been investigated by other authors, such as the threshold method, region growing, and morphology-based methods, the state-of-the-art methods use Convolutional Neural Networks (CNNs) and Deep Learning (DL) strategies. We use the DL approach in our method as well. Typically, enhancement filters such as those based on the Hessian matrix are used to improve the contrast and visibility of the vascular structures. As opposed to that, our method is based on the fusion of the raw image with the RORPO (The Ranking Orientations Responses of Path Operators) [1] enhanced image. This is then utilized by the segmentation architecture of 3D U-Net or 3D V-Net. The final segmentation pipeline consists of RORPO and 3D V-Net, which provide the best segmentation results.
In the paper, several experiments are performed, which test the main parts of the proposed method and make comparisons with alternative state-of-the-art approaches. The first experiment compares five classical vessel enhancement filters separately: Frangi, Hessian, Meijering, RORPO, and Sato. The following two experiments evaluate a fusion of different vessel enhancement filters with two variants of segmentation models (3D U-Net and 3D V-Net). The fusion of the RORPO enhancement filter with the CT image to detect the liver vessel structures supported by the 3D V-Net is proven to be the most efficient. We also suggest the best value of the blending coefficient between the RORPO enhanced image and the raw CT image. In such a setup, we further experiment with different variants of dice loss functions to solve the problem of pixel imbalance between the foreground and background categories. The following experiment then performs the ablation study of the selected architecture. In the last experiment, a comparison between our proposed algorithm and other algorithms is provided.
The proposed algorithm is proven to be effective, robust, and accurate for liver vessel segmentation, even for images with low contrast and high noise. Experiments are performed on the public 3D-IRCADb [2] dataset; the average dice score and precision are 76.2% and 77.7%, respectively.
During the work on this topic, we have used 3D Slicer [3] software, for which we have created a software extension that can provide doctors and other healthcare professionals with tools for automatic vessels segmentation, and tools for validating the segmented datasets. Employing deep learning and 3D Slicer, we can segment vascular structures automatically, improve the annotations, automatically collect the annotated datasets, and then fine-tune the neural network models so they can be quickly exposed for further use in automatic segmentation.

2. Related Work

The trend of the proposed segmentation methods leads to the automation of liver vessel segmentation [4,5]. The methods for liver vessel segmentation from CT abdominal images can be divided into several groups: tracking-based algorithms, active contours, and machine learning [6]. The popular approach to segmenting liver vessels is based on machine learning methods and Deep Learning (DL) specifically.
The application of 3D U-Net [7,8,9,10] is often used for the segmentation of liver vessels from CT images. The 3D U-Net is composed of analytical and synthetical parts [11], similar to the standard U-Net architecture, but it uses 3D operators instead. Yu et al. [7] described the 3D Residual U-Net technique, which is built on 3D U-Net with a 3D morphological closure operation in the postprocessing phase. The algorithm was tested on their private CT abdominal dataset. Huang et al. [8] used 3D U-Net with a combination of the variant dice loss function. The segmented veins were more continuous and complete when training was carried out using the enhanced manual expert annotations rather than the original dataset. The tests were carried out using the 3D-IRCADb, SLiver007 [12], and private datasets. Affane et al. [9] experimented with three distinct 3D U-Net approaches: basic U-Net, MultiRes U-Net, and Dense U-Net. On the 3D-IRCADb dataset, they determined that MultiRes U-Net architecture was superior for segmenting hepatic blood veins. Golla et al. [10] employed an ensemble technique. They combined the predictions of networks into ensemble E and averaged the resulting probability for each class. Probability maps of all networks (2D U-Net, 3D U-Net, 2D V-Net, and 3D V-Net) and ensembles were resampled back to the original data, resolution and the maximum probability was applied to the data to extract the predicted segmentation. The approach was tested on the publicly available datasets 3D-IRCADb and MICCAI [12], with unsuitable data being excluded.
V-Net [13] is the next commonly used neural network architecture that has been studied in recent years for segmenting liver blood vessels. Su et al. [14] proposed the DV-Net algorithm. DV-Net represents a V-Net with a dense block structure. They utilize a combination of DCDS (dense connection downsampling approach) and D-BCE loss function for capturing blood vessel structure. The method was tested on 3D-IRCADb. Yang et al. [15] used an improved V-Net to segment liver blood vessels on the 3D-IRCADb dataset. Improved V-Net is based on inter-scale dense connections in the decoder. Altini et al. [16] used 2.5D V-Net for the segmentation of liver blood vessels. They trained the method on the 3D-IRCADb dataset and tested it on their dataset. They used Tversky index-based loss function in combination with 2.5D V-Net. The Tversky index ensures a short convergence time due to the unbalanced voxels problem.
A different approach than the common utilization of either the U-Net or V-Net architecture is applied in [17,18]. In [17], the authors propose a fusion network called TransfusionNet. This method is based on a transformer for semantic segmentation. The authors use the 3D-IRCADb and LiTS [19] datasets for pre-training, and their data for fine-tuning the segmentation network. Xu et al. [18] applied a deep neural network based on the bootstrapping technique to the 3D-IRCADb dataset. A convex combination of the model predictions and preliminary predictions with the subsequent application of a noise filter is used to segment the liver blood vessels.
There is a full body of work that focuses on image enhancement that can be applied before training the segmentation model. This is especially true if complex structures such as vessels are segmented. Shahid and Taj [20] apply a set of preprocessing operations, including filtering via the Frangi filter, to enhance retinal vessels before their segmentation. Soomro et al. [21] use steps to suppress irregular illumination and to improve low and varying contrast via contrast-limited adaptive histogram equalization for the task of retinal vessel segmentation. Blaiech et al. [22] studied the effect of enhancing the 2D images for coronary artery segmentation. Different enhancement methods (Frangi, CLAHE, and RORPO) were used in normal conditions and in the presence of noise. Lamy et al. [23] compared seven different vessel enhancement filters on real (3D-IRCADb) and synthetical (VascuSynth [24]) datasets. RORPO has been evaluated as being the best-performing method on the whole liver area when applied to the real 3D-IRCADb dataset.
With a focus on the use of vessel-enhancing filters as a part of the pipeline for hepatic vessel segmentation, Survarachakan et al. [25] applied four enhancement filters (Hessian, Frangi, Sato, and Meijering), and proposed to fuse their outcomes in two different segmentation designs. The proposed methods are evaluated on the clinical OSLO-COMET dataset.
In our research, we propose an algorithm that effectively fuses the vessel enhancement technique represented by RORPO [1] with the raw CT image, and handles the problem of pixel imbalance by employing a variant of the dice loss function to optimize the segmentation even further. Evaluation is performed on the publicly available 3D-IRCADb dataset. Compared to the above-mentioned related works [16,23,25], our proposed method can provide better segmentation results, and it adopts a much simpler approach of combining properly the original raw image with the enhanced image to tackle the problem of vessel segmentation. Numerous related works likewise fail to account for the foreground–background pixel mismatch. We use Weighted Dice Loss to provide penalties for the number of incorrectly categorized voxels to tackle the problem. Additionally, some of the related research ignores the dataset’s inferior quality annotations for some patients. To ensure that no vessels are missed and that the segmentation accuracy provides an accurate portrayal of various algorithms, we resolve this by enlisting the aid of medical professionals to annotate any missed vessels.

3. Segmentation Platform

Together with the proposed algorithm, we have developed a plugin extension for 3D Slicer [26] that (i) provides a remote AI-Assisted Annotation service (AIAA) to medical doctors from a High-Performance Computing (HPC) cluster, allowing them to use state-of-the-art methods to perform the automatic segmentation of desired tissue from medical images, and (ii) provides a mechanism to collect the segmented and validated data generated in step (i). We can use the stored data to fine-tune the existing neural network models. These enhanced models can then be employed for automatic tissue segmentation in step (i).
Since the current state-of-the-art medical image processing methods are based on Deep Learning (DL), and typically, DL algorithms are trained with a large amount of data, we allow for the use of multiple GPUs during the training phase to produce models of the required quality in a reasonable amount of time. We use the multi-GPU nodes of an HPC cluster to train models from scratch, as well as single GPU nodes to provide model inference using AIAA. All of the connections and data are encrypted using Secure Shell (SSH).
The entire concept is depicted in Figure 1. Two main sections can be distinguished: one runs at a medical doctor’s site in a local hospital (frontend), and the other operates at an HPC cluster facility (backend). The frontend allows the doctor to load, view, and perform automatic segmentation, and improve annotation on the medical data using HPC resources. The backend part provides the computational power and other required features.
The provided concept adopts the Clara Train SDK [27] from NVIDIA. It contains several APIs, such as those for AIAA, and a training framework for DL-based model training. We use Clara Train SDK, version 4.0, which is solely PyTorch-based. The Clara Training Framework builds on the open-source MONAI framework [28], which is specifically devoted to deep learning in healthcare imaging.
Clara’s AIAA is a client-server-based architecture that delivers a C++ or Python client API. Many medical image viewers, such as 3D Slicer (see Figure 2) can be interfaced as clients [29] to obtain the AIAA services offered by the server running on the network. Based on the pre-loaded models on the server, the AIAA can perform the automatic segmentation of specific tissues and display the results. Doctors can edit and postprocess the resulting tasks. A modified version of this extension, accommodated for HPC cluster usage, has been used by medical doctors to provide improved annotations of the 3D-IRCADb dataset.

4. The Hardware Used

For training the models, we utilized cluster nodes with 8 × NVIDIA A100 (40 GB HBM2) per node. Each node is a powerful x86–64 computer, equipped with 2 × 64-core AMD Zen 3 EPYC™ 7763, 2.45 GHz processor and 1024 GB of memory.

5. Datasets and Improved Annotations

Publicly available datasets that contain labeled hepatic vessels are listed in Table 1. 3D-IRCADb is the most commonly used dataset. We use it in our experiments as well. This particular dataset is described in more detail in the following subsection.
The MSD [30] is another publicly available dataset. However, its ground truth is of poor quality, far inferior to the 3D-IRCADb dataset. Therefore, we have not used this dataset at all in this work.
We are also aware of synthetically generated vascular datasets that can be provided by the VascuSynth [24] software. However, since these datasets cannot replace clinical data where patients have curvilinear vessels, we have not used them either.

5.1. Dataset 3D-IRCADb

The proposed method has been evaluated on the publicly available 3D-IRCADb [2] dataset. We have specifically used its subset 3D-IRCADb-01, which is used in most of the similar research, and thus allows for the comparison. The dataset contains 20 CT volumes (10 male and 10 female). Fifteen cases (75%) of this dataset have hepatic tumors. The dataset contains various classes. The classes used in our study are portalvein and venoussystem. We combined them into one class of veins. The pixel spacing varies from 0.56 to 0.84 mm, and the slice thickness varies from 1 to 4 mm.
There are several challenges within the dataset. In some cases, the inferior vena cava is not part of the liver mask. Besides that, one patient had many liver tumors that covered the majority of the liver vessels. Another patient had inner metallic objects, such as stents, that significantly affected the brightness of the volume. Despite these bottlenecks, we kept all of the volumes from the dataset in the experiment to have more training data. Furthermore, slight abnormalities make the network more robust, even if the output quality suffers. Although low-quality data were used, our technique was still able to achieve a high prediction quality. The training and testing sets should equally include cases of liver vessels of similar appearance (both containing difficult and easy cases). To ensure this, we split the dataset as follows. Training set contains patients—2, 5, 7, 12, 10, 11, 13, 16, 18, 19. Validation includes patients—4, 9, 20, 6, 17. Testing set is represented by patients—1, 3, 8, 14, 15.

Improving Annotations of 3D-IRCADb

Since there are some limitations in the annotations of hepatic vessels in the 3D-IRCADb dataset, we decided to improve them. The limitations are mainly because of inadequately annotated hepatic veins, although they are clearly visible and could be marked correctly. Similar problems have been found and tackled by [8,31]. Some of the CT volumes in 3D-IRCADb are under-segmented, while others are over-segmented. These issues might lead to the misinterpretation of results when segmented vessels appear as false positives or false negatives while testing the neural networks [18]. Another major limitation is the inconsistency of the annotated data for the vena cava. The vena cava is visible in some images but not in others. It also makes vessel enhancement difficult [8].
Because the annotated datasets used in [8,31] are not publicly available, the additional labeling of the missing veins in 3D-IRCADb has been performed using the Slicer built-in tools and the extension explained in Section 3. It increased the quality of annotations; see Figure 3. The comparison of the dice scores with the original and improved dataset is in the experimental part of the paper. If it is not stated differently, the experiments used refined annotations since the vessel prediction is better and more continuous with them.

6. Methods

The core of our solution lies in the use of enhancement filters and deep learning methods for vessel segmentation. We perform experiments on the dataset, as explained in Section 5.1. We apply different filters to improve the contrast between the liver and hepatic vessels, and use the enhanced data as input to a segmentation network.

6.1. Vessel Enhancement Filters

We have used five filters to enhance the tubular structures and other structural information of the veins. Similar to Survarachakan et al. [25], we have applied four Hessian-based filters. Besides that, a specific morphological filter has been used as well. Namely, we have used Hessian, Frangi, Meijering, and Sato as the Hessian-based filters, and RORPO as the morphological filter. Figure 4 shows the enhancing effect of these filters on vessels.

6.1.1. Hessian Matrix Computation

The Hessian matrix serves as a fundamental computational method used in all Hessian-based filters. It calculates the local gradation change by performing the second-order partial derivatives of the input image voxel X = (x, y, z) in nine directions. It is defined as
H ( X ) = f x x f x y f x z f y x f y y f y z f z x f z y f z z = 2 f x 2 2 f x y 2 f x z 2 f y x 2 f y 2 2 f y z 2 f z x 2 f z y 2 f z 2
Eigenanalysis is used to avoid computing derivatives in many directions, and to extract only the principal directions. Let e 1 be the eigenvector representing the axial direction, let e 2 , e 3 be the cross-sectional direction of H(X), and the associated λ 1 , λ 2 , λ 3 are the eigenvalues. The eigenvalues should meet the following condition: | λ 1 | | λ 2 | | λ 3 | . The vessel centerline voxel should satisfy the following: λ 2 λ 3 0 , λ 1 0 .

6.1.2. Hessian Vesselness Filter

By integrating the directional gradient and the Hessian matrix, Ng et al. [32] suggested a modified multi-scale Hessian filter. It is assumed that the brightness of a hepatic vessel is high in the middle and progressively declines toward the end, allowing it to be described as a Gaussian structure transverse to its axis. The convolution between the image gradient field and the Gaussian kernel is the basis of each approximation in the Hessian matrix H at a specified scale σ . The eigenvalues λ 1 , λ 2 are calculated and utilized to compute the curve derivation R and similarity measure S. The curvilinear likeliness E is given by
E ( x , y , σ ) = 0 i f   λ 2 < 0 e R 2 β 1 2 [ 1 e s 2 β 2 2 ] o t h e r w i s e
The response of the filter L is set as a maximum of different scales that approximate the size of the ridges. It is expressed as
L ( x , y ) = max σ m i n σ σ m a x [ E ( x , y , σ ) ]
The parameters β 1 , β 2 control the sensitivity of the filter to the measure R and S, respectively.

6.1.3. Frangi Vesselness Filter

The Frangi [33] approach uses all three eigenvalues to discriminate the local orientation pattern. These three eigenvalues serve in the differentiation of blobs R b , and plate-like and line-like structures R a . To decrease the influence of noise S, a Hessian norm measure was developed.
R b = | λ 1 | / | λ 2 λ 3 |
R a = | λ 2 | / | λ 3 |
S = λ 1 2 + λ 2 2 + λ 3 2
They propose the following combination of the components to define a vesselness function,
F = 0 i f   λ 2 > 0   o r   λ 3 > 0 , ( 1 exp ( R a 2 2 α 2 ) ) exp ( R b 2 2 β 2 ) ( 1 exp ( S 2 2 c 2 ) )
The parameters α , β , γ are the thresholds that control the sensitivity of the filter to the measures R a , R b , and S.

6.1.4. Meijering Vesselness Filter

To recognize very elongated structures, Meijering et al. [34] suggested a parameter-free vesselness function. It is based on the following modified Hessian matrix H ( f ) :
H ( f ) = [ h 11 + α 2 ( h 22 + h 33 ) ( 1 α 2 ) h 12 ( 1 α 2 ) h 13 ( 1 α 2 ) h 21 h 22 + α 2 ( h 11 + h 33 ) ( 1 α 2 ) h 23 ( 1 α 2 ) h 31 ( 1 α 2 ) h 32 h 33 + α 2 ( h 11 + h 22 ) ]
In general, α = 1/3. The eigenvalues of H ( f ) with respect to H ( f ) is expressed as
λ i = λ i + α λ j + α λ k
for i j k .
The vesselness is defined by,
F = λ m a x / λ m i n λ m a x < 0 0 λ m a x 0
where, λ m a x = m a x { λ 1 , λ 2 , λ 3 } , which is computed at each voxel, and λ m i n is the minimum of all λ m a x of the image.

6.1.5. Sato Vesselness Filter

Sato et al. [35] suggested a line enhancement filter function that is responsive to varied diameter ranges. Sato et al. sorted the eigenvalues as λ i as λ 1 λ 2 λ 3 . The eigenvector e 1 corresponds to the direction of the putative vessel. | λ 1 | < 0 and | λ 1 | < 0 represent the sizes of the cross-section. The Sato vesselness introduces a ratio of the eigenvalues to obtain a high response in tubular structures. It is provided by,
F = λ c exp ( λ 1 2 2 ( α 1 λ c ) 2 ) ) λ 1 0 , λ c 0 λ c exp ( λ 1 2 2 ( α 2 λ c ) 2 ) ) λ 1 > 0 , λ c 0 0 λ c = 0
where α 1 < α 2 , λ c = m i n { λ 2 , λ 3 } . The parameters α 1 and α 2 control the asymmetrical strength.

6.1.6. RORPO Vesselness Filter

Differential information is not used by the RORPO [1,36] filter. RORPO, on the other hand, is based on the mathematical morphology of the path operators, and is thus defined using adjacency relations. RORPO is semi-global and non-linear. This filter computes the path openings in the seven primary directions, ranks their responses point-by-point, and extracts structures based on low and high responses at each voxel. This noise-resistant filter maintains the intensity of curvilinear shapes while reducing the intensity of other structures.
Path operators with thin voxels oriented along their length are expressed as:
L n R = L m i n × f n R 1 ,
where L n R is path length to detect, L m i n is the minimal path length, the geometric sequence of scales is f ϵ R , and the number of scales is n R .

6.2. Segmentation Pipelines

The three main steps of the segmentation pipelines we have implemented are image data preprocessing, vessel enhancement, and image segmentation. These are explained in the following subsections. First, the image preprocessing is performed, and then specific enhancement filters are applied and either fused before the segmentation or immediately after the segmentation. Figure 5, Figure 6 and Figure 7 show three different approaches.
We have implemented three segmentation pipelines. In the first two pipelines, the fusion of filters are inspired by the approach of Survarachakan et al. [25], and they serve as a comparison. The third pipeline that we propose implements a simple but very effective fusion of raw and filtered images before performing the segmentation.

6.2.1. Data Preprocessing

In our method, the liver area is extracted and cropped according to the liver mask to focus primarily on the liver vascular system. The liver region is extracted in the first step as opposed to the last step in Survarachakan et al. [25]. This helps to remove unnecessary resulting boundaries, and leads to less interference with the background. All scans are converted into 1 × 1 × 1 mm isotropic resolution. The image intensities are windowed to lie within the 80 , 220 Hounsfield Unit (HU) range, and are then mapped to the range 0 , 1 . In addition, the 3D anisotropic diffusion is applied to lower the image noise and to still preserve the significant parts such as edges and lines.

6.2.2. Vessel Enhancement

The enhancement of vessels is the preprocessing step used before the segmentation. The preprocessing steps are implemented using the scikit-image library [37]. This step is a key factor in obtaining significantly better segmentation results. Enhanced images acquired by the application of different filters (see Section 6.1) are utilized either before or following the segmentation.
A method that combines the results of individual vessel enhancement filters simultaneously before segmentation, and which is inspired by the approach of Survarachakan et al. [25], is depicted in Figure 5. The final enhanced image is obtained by averaging the respective pixel values from all filters. The resulting image is then used as an input to the segmentation network. Two variants of this approach have been considered. They either use Hessian, similar to [25], or RORPO as one of the four filters. As proven by experiments, the Hessian filter has the weakest enhancing effect on the segmentation results in terms of the dice score. The RORPO filter, on the other hand, has the strongest enhancing effect. Therefore, the initial idea of this method was to use the strongest combination of four filters that would outperform the solution presented in [25]. We indicate this method as FilterAdded, and distinguish whether Hessian or RORPO has been used. The fusion of filters is defined as:
F i l t e r A d d e d H e = H e + F r + M e + S a number of filters
and
F i l t e r A d d e d R o = R o + F r + M e + S a number of filters
where H e , R o , F r , M e , and S a , stand for the Hessian, RORPO, Frangi, Meijering, and Sato filter outputs.
Inspired by the second approach of Survarachakan et al. [25], a better overall result can also be achieved by combining the prediction results from four trained models, each of which used a different vessel improvement filter. Only the pixels predicted from at least two independent filters were taken into account when combining the results, in order to reduce false positives. The approach is depicted in Figure 6. Similar to the previous approach, we opt between two variants of applied filters, either with the Hessian or with the RORPO. Again, the variant with the RORPO filter brings better results, as shown in the experiments. We indicate this method as SegAdded, and distinguish whether Hessian or RORPO has been used. A combination of the enhanced outputs is provided in the following manner:
S e g A d d e d H e = ( He Fr ) ( He Me ) ( He Sa ) ( Sa Me ) ( Sa Fr ) ( Me Fr )
and
S e g A d d e d R o = ( Ro Fr ) ( Ro Me ) ( Ro Sa ) ( Sa Me ) ( Sa Fr ) ( Me Fr )
where He , Ro , Fr , Me , and Sa , stand for the Hessian, RORPO, Frangi, Meijering and Sato filtered and segmented outputs, respectively.
The third and most effective approach that we have implemented operates with the RORPO filter and makes a fusion of the enhanced image with the preprocessed raw image before the segmentation. The detailed pipeline is shown in Figure 7. The linear blending fuses the original and enhanced images to create input for the segmentation network.

6.2.3. Linear Blending

The blending method applies a linear combination of two source images, and it is expressed as:
g ( x ) = β · f 0 ( x ) + α · f 1 ( x ) ,
where f 0 ( x ) represents the original image, f 1 ( x ) stands for RORPO enhanced image, g ( x ) provides the blended image, and x represents the pixel coordinate. Both images have to be of the same size and type. Coefficient α lies in the interval 0 , 1 and determines how much the enhanced image will prevail over the original image. The optimal value of Coefficient α can be determined by running experiments in the stated interval, with an appropriate α increment. Coefficient β depends on α and it is expressed as:
β = ( 1.0 α )

6.3. Segmentation Model

We have experimented with two types of segmentation architectures: 3D U-Net and 3D V-Net. They are often used to segment tissues from medical images.
The 3D U-Net architecture on which we conducted the experiments is shown in Figure 8. Each layer of the network has a contracting encoding path and an expanding decoding path with a skip connection between them. The encoding path extracts hidden features from the input data, which are then reconstructed by the decoding path, to provide the prediction. The skip connection is used to retain more semantic information, and thus adds the initial feature map to the final feature map. The encoder downsamples the data using strided convolutions of the residual unit in the spatial dimensions by a factor of 2, followed by a Batch Normalization (BN), and a Parametric Rectified Linear Unit (PReLU) as an activation function. During backpropagation, the PReLu learns a small slope parameter, which multiplies with the negative input to better adapt to the other parameters (such as weights and biases). The decoder upsamples using 2 × 2 × 2 up-convolution (transpose convolutions) with stride 2. These convolutions occur at the beginning of each block.
The 3D V-Net implementation is based on Milletari’s [13] design of the 3D V-Net. The 3D V-Net is intended for 3D medical image segmentation and is shown in Figure 9. The network has encoding and decoding blocks, as well as an additional convolutional block in between, similar to the 3D U-Net. The encoder and decoder are also coupled by skip connections, but instead of 2 × 2 × 2 max-pooling, down-convolutions with stride 2 × 2 × 2 and kernel-size 2 × 2 × 2 are used for encoding. The strided deconvolution is used for decoding. The top-level blocks have one convolutional layer, the second-level blocks have two, and the remaining blocks have three convolutional layers. The number of channels in each block is the same as in the U-Net configuration, but all convolutions extract spatial features using 5 × 5 × 5 kernels.
In the case of hepatic vessel segmentation, the network performs better if the image data are preprocessed (see Section 6.2.1).

6.3.1. Data Augmentation

We employ data augmentation in our approach to improve the amount of input data, and to decrease overfitting while training. In each spatial dimension, we apply symmetric padding to the data. We crop random fixed-size sections with a foreground or a background voxel in the center. We rotate the data by 90 in all three spatial planes. We also use a random offset to adjust the intensity.

6.3.2. Parameter Settings and Training

We have used the same parameter settings for all experiments reported in this paper, unless otherwise stated. We have used a batch size of 2, the Adam optimizer with a learning rate of 2.0 e 4 , and 1000 of epochs. A learning rate scheduler has been applied to adjust the initial learning rate and to further decrease it. The learning rate scheduler is defined as l r = i n i t _ l r × ( 1 e p o c h n E p o c h ) p o w e r , where i n i t _ l r = 0.0002 , n E p o c h = 5000 and p o w e r = 0.1 . We have used a Dropout of 0 for 3D U-Net, and 0.5 for 3D V-Net. We have used ReLU as the activation function.

6.3.3. Loss Function

During the training process, the loss function assists in minimizing the error for each training sample. Because liver vessels make up a minor percentage of the liver, imbalanced foreground (the liver vessels) and background (the liver) classes frequently induce prediction deviation, biasing the classification to the background with more voxels. Therefore, we have experimented with different loss functions. We have used the Weighted Dice Loss function [38] as a final selection. It uses the dice coefficient to place penalties for the number of misclassified voxels, to solve the high-class imbalance problem and to enhance segmentation accuracy; see Table 2.

7. Experiments and Results

We have tested the proposed algorithms on the 3D-IRCADb dataset. If not otherwise stated, all of the experiments were performed using improved labels in the 3D-IRCADb dataset. We have used typical metrics to compare the implemented methods. Specifically, these include the Dice Score, the Accuracy, the Sensitivity, and the Precision. Specific equations for the metrics are expressed in Table 3.

7.1. Experiment 1: Comparison of Different Vessel Enhancement Filters

First, the five vascular enhancement filters were compared: Hessian, Frangi, Meijering, RORPO, and Sato. The five different neural networks were trained to compare the filters in terms of their effect on vessel segmentation. In this experiment, the 3D U-Net architecture was used. The processing workflow from the raw input to the segmented output is shown in Figure 10. Table 4 provides the comparison of different filters concerning segmentation quality measures. The table is divided into two sections. The first one provides results using original annotations, and the second one uses improved annotations.

7.2. Experiment 2: Fusion of Vessel Enhancement Filters

As described in Section 6.2.2, we have used three different segmentation pipelines that leverage different vessel enhancement filters to improve segmentation results. Specifically, the pipelines named as F i l t e r A d d e d H e , F i l t e r A d d e d R o , S e g A d d e d H e , S e g A d d e d R o , and our method (the RORPO filter with a linearly blended input image) were implemented and compared. The dice scores of different approaches performed on the test data are shown in Table 5.

Comparison of Segmentation Pipelines

We have compared all three segmentation pipelines and have used either 3D U-Net or 3D V-Net as the segmentation architecture. The Segmentation results are shown in Table 5. Our approach provides the best performance in this comparison. As for the specific segmentation network, the 3D V-Net leads to better results than the 3D U-Net if applied in our method.

7.3. Experiment 3: Linear Blending with the RORPO Filter

Based on the results from the previous experiment, it is obvious that the linear blending with just the RORPO filter outperforms approaches where multiple filters are combined. Therefore, the focus was aimed on using only this method and elaborating whether the blending between the original and enhanced image was set to its optimum, or whether it could improve the results even further by setting it to a different value. As described by Equations (17) and (18), the blending is controlled by setting the α value from the interval < 0 , 1 > . The effect of different α values is shown in Figure 11.

7.3.1. Comparison of Segmentation Models with Different Blending Values

We have compared our method of linear blending with the RORPO filterm using either 3D U-Net or 3D V-Net as the segmentation architecture. The highest dice score is achieved by blending with 60 % ( α = 0.6 ) of the image enhanced usin the RORPO, and 40 % ( β = 0.4 ) of the original image in both 3D U-Net- and 3D V-Net-based architectures; see Figure 11. It also shows that segmentation with 3D V-Net provides better dice scores and more consistent results over a wider spectrum of α values than 3D U-Net. As an acceptable region for blending value, the region between 40 and 60% is set.

7.3.2. Comparison of Different Loss Functions

We have evaluated different types of loss functions and their effects on the segmentation quality. The evaluation was performed on the best-performing setup with the 3D V-Net as a segmenter. The results for all metrics are shown in Table 6.
The best-performing loss function in our model is the Weighted Dice Loss. It operates with the parameter β l (see Table 2) to tackle the problem of class imbalance. The higher the value of β l , the greater the penalty for misclassified voxels. Unfortunately, this also slows down the gradient flow and increases the risk of sticking in the local minima. When β l is in the range of 5–7, optimal outcomes are obtained. We have found β l = 6 to be the best in our study.

7.4. Experiment 4: Ablation Study on the Final Segmentation Pipeline

We have performed the ablation of a varying number of convolutional filters at all stages of the 3D V-Net architecture after training. We have found that certain stages are more significant than others for the quality of veins segmentation.
The V-Net is divided into several stages (the stages are shown in Figure 9) that operate at different resolutions during the down- as well as upsampling phases. Each stage consists of one to three convolutional layers, with each layer having a specific number of channels (filters). We ablated groups of similar filters at every stage by increasing the percentage ( 25 % , 50 % , and 75 % ) of similar filters. Since the number of filters varies from stage to stage, it means that the same percentage may correspond to a different number of ablated filters. The similarity between filters within a group was calculated based on the absolute Euclidean distance of the normalized filter weights. Ablations were performed by manually setting the weights and biases of all incoming connections of a filter to 0, thus eliminating any activation of that filter.
Figure 12 shows the dice scores for ablations of 25 % , 50 % , and 75 % at every stage of the V-Net. Each data point shows the effect of a specific stage ablation. The dice score is affected noticeably more in stages ’in 16’, ’down 64’, and ’up 64’, than in the other stages. Furthermore, the effect of the largest ablation has a stronger influence on some stages than on others. For instance, stage ’up 32’ shows a noticeably larger drop in dice score for 75 % of the ablated filters, compared to 50 % , while the stages ’down 256’ and ’up 256’ are largely unaffected.

7.5. Segmentation Results

Figure 13 shows the segmentation results of the proposed algorithm on the improved 3D-IRCADb dataset. The provided method can predict the vascular structure well, and the results are close to the ground truth.

7.6. Experiment 5: Comparison of Performance between the Proposed Algorithm and Other Algorithms

We have compared our segmentation approach with the work of others. Specifically, with works [4,8,14,15,16,18,23], which combine different approaches to provide liver vessel segmentation.
In the case of [4,8,14,15,18], it was impossible to make a fair comparison. The source code for these methods was not available, so it was not possible to train and test these methods on the same enhanced 3D-IRCADb dataset that we used in our work. We provide only an indirect comparison with the numerical results as presented in those papers. A comparison is provided in the first part of Table 7.
We were able to objectively evaluate the performance of methods from [16,23] since their code is available. The authors of the paper [23] published the code on GitHub and have also created an online demonstration tool for using the filters. We tested the RORPO filter with default parameter settings. Then, the result was thresholded. The parameter n b T h r e s h o l d s = 200 means that each patient is thresholded with a 200 x different value. This value is increased by a very small step of 0.005 and lies in the range of < 0 , 1 > . As opposed to that, the authors of the paper [16] implemented 2.5D V-Net with Tversky loss function. They combined patients from the 3D-IRCADb dataset, and CT scans from the Polyclinic of Bari. The 2.5D V-Net approach processes five slices as five channels, since the network uses only 2D convolutional layers. The random patches of five consecutive slices are sent to the network for training. To test the network, the sliding window is used. However, we infer sub-volumes of dimensions 512 × 512 × 5 , and only the middle slice is used for prediction on the middle slice. The results of these methods and our proposed method are provided in the second part of Table 7.

8. Discussion

In the first experiment, different vascular enhancement filters were compared. The task was to find out which of the five filters was the best, in terms of the segmentation quality measures. From the visualizations of the filters’ enhancing effects in Figure 4, it can be seen that the RORPO filter performs very well. There is a minimum amount of image noise. In addition, there are no visible boundaries between the liver and the background that could affect the result. Although the image enhanced by the Frangi filter seems promising as well, the vessels are relatively thin. Visible boundaries and some noise are typical for the Meijering filter and the Sato filter. In terms of the segmentation quality, the dice scores for the original and the improved annotations prove that the RORPO is the most effective filter, and has permitted us to improve the dice score from 62.7% to 69.7% and to reduce the mean standard deviation from 13.8% to 4.8%. The improvement in dice score and reduction in standard deviation shows that the model can segment vessels more efficiently in different test patients.
The second experiment focused on leveraging the vessel enhancement effects of different filters in various segmentation pipelines. In FilerAdded, four enhancement filters were combined, and the model was trained on this combined data. In SegAdded, four models were trained with distinct enhancement filters, and the results were later combined when at least two models were able to predict the vessel pixel. In linear blending, the original image was combined with the RORPO enhanced image. By comparing the different pipelines, the linear blending provides the best results for all observed metrics except sensitivity. Here, the SegAdded approach performs better. Results were also obtained from different segmentation models (3D U-Net and 3D V-Net).
In the third experiment, a detailed analysis of the proposed method was provided. We have explored the effect of RORPO on the original image. We have set the mixing ratio between RORPO and the original image, α , from 0 % to 100 % , with a step of 10 % . The RORPO filter has a significant positive effect on the result. The best segmentation is achieved by using 40 % of the original image and 60 % of RORPO. The acceptable α range is 40–60%. This proves that the role of background information is also helpful in segmenting the region of interest. The results are also performed on 3D U-Net and 3D V-Net, and 3D V-Net performs better than 3D U-Net. This shows that for volumetric input, 3D V-Net performs better. Hence, we used V-Net in all further tests. However, 3D V-Net is more expensive in terms of computing resources, and the time for training is twice as long as that of 3D U-Net. Another evaluation of the proposed method focused on the use of different dice loss functions. The Weighted Dice Loss β l = 6 outperformed other loss functions, achieving a dice score of 76.2% and a mean standard deviation of 3.7%. This loss function can deal with unbalanced classes and enhance segmentation accuracy and sensitivity. Parameter β l = 6 was used to introduce a penalty for misclassified pixels, and also to avoid trapping the loss function in local minima. The β l parameter can be in the range (0, 9). If the β l parameter is set to a lower or higher value than 6, the results are worse overall, since the coefficient penalizes too few or too many false negatives. This approach set the final version of the method and led to the highest dice score, as well as increased precision.
The fourth experiment examined the effects of varying percentages of ablations on the segmentation performance of the 3D V-Net architecture trained on the 3D-IRCADb dataset. As anticipated, the performance typically declined as the number of ablated filters increased. The results of the experiment showed that some filters contributed more to the performance than others. The effect of ablation is much stronger for stages (’down 64’, ’up 64’) for 50 % ablation, and stages ’up 32’ for 75 % ablation. The stages (’down 256’, ’up 256’) showed negligible performance loss, as these stages may have redundant feature representation. This experiment determines the influence of different layers on the segmentation results. By identifying the network’s sensitive areas, we want to modify the model’s architecture in future research.
In the last experiment, we compared the performance of the proposed method with some other algorithms described in the related work. We compared indirectly with [4,8,14,15,18] because it was not possible to re-implement or evaluate these methods with our dataset. A direct comparison was made with [16,23]. These methods were re-implemented based on the available source code and evaluated with our dataset. The authors in [23] just used a thresholding approach that did not provide satisfactory vessel segmentation. The authors of the paper [16] used 2.5D V-Net with a Tversky loss function, yielding results with high variability and a high mean standard deviation. We can state that our method prevails in dice score and specificity value over all of the other methods. In comparison with the two reimplemented methods, we dominate in terms of sensitivity, meaning that there are fewer missing and fragmented vessels. In terms of accuracy, our method lags just slightly behind.
Overall, the method outlined in this paper can automatically segment enough liver vessels to correctly extract the vessels, and it can be further manually corrected using Slicer tools. The automatic segmentation results are visually presented in Section 7.5. The method works on CT obtained from different machines with different acquisition parameters. Our method is also able to segment images with different liver shapes, livers with thinner vessels, and livers with vessels atrophied. The specific combination of the original image with the vessel-enhanced image is the originality of our work, and it outperforms other methods proposed in the literature. The method outlined in this paper have some limitations. The results in some cases have misclassified vessel endings, variability in vessel edges, and some regions are not connected to the main vessels. Finally, we need more data to make the model more robust, test the vessel segmentation, and lower the variability in our experiments.

9. Conclusions and Future Works

The proposed algorithm effectively segments liver vessels with a dice score of 76.2 % and a precision of 77.7 % . The network can automatically segment labeled or even other unmarked liver vessels (that should be labeled) from the CT images. The RORPO filter used for vessel enhancement in conjunction with blending its output with the original CT image has successfully been proven to give better results compared to other state-of-the-art techniques. Thus, adding this information to the enhanced vessels has helped to improve the robustness of the segmentation model. The Weighted Dice Loss metric based on the dice coefficient is used for the loss function computation to improve segmentation accuracy, and to deal with unbalanced foreground and background class voxels. The Weighted Dice Loss function with β l = 6 has been proven to perform the best if it is combined with the 3D V-Net.
This best performing model has been made available to the local hospital’s doctors, who can use it for the initial annotation of new data using the tool described in Section 3.
In the future, we plan to cooperate with medical doctors more closely, extend the training set with new validated data, and continue our research in the area of hepatic vessel segmentation. Additionally, in our future studies, we will focus on experiment pruning and computational cost reduction by understanding the significance of ablated filters.

Author Contributions

Conceptualisation: P.S. (Petr Strakos), P.S. (Petra Svobodova), and K.S.; methodology: P.S. (Petra Svobodova), K.S. and P.S. (Petr Strakos); formal analysis: P.S. (Petra Svobodova), K.S., P.S. (Petr Strakos), and A.V.; investigation: P.S. (Petra Svobodova), K.S. and P.S. (Petr Strakos); data curation: P.S. (Petra Svobodova), K.S., P.S. (Petr Strakos), and A.V.; writing—original draft preparation: P.S. (Petra Svobodova), K.S. and P.S. (Petr Strakos); writing—review and editing: P.S. (Petr Strakos), P.S. (Petra Svobodova), and K.S.; visualization: P.S. (Petra Svobodova) and K.S.; supervision: P.S. (Petr Strakos). All of the authors have read and approved the final manuscript and its submission to the journal. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the doctoral grant competition VSB—Technical University of Ostrava, reg. no. CZ.02.2.69/0.0/0.0/19_073/0016945 within the Operational Programme Research, Development and Education, under project DGS/TEAM/2020-008 ‘Development of a tool for scientific data processing and visualization in VR with multi-user support’.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CTComputed Tomography
HUHounsfield Unit
SSHSecure Shell
RORPORanking Orientation Responses of Path Operators
HPCHigh-Performance Computing
GPUGraphics Processing Unit
APIApplication Programming Interface
FilterAddedFiltered images added
SegAddedSegmentation maps added
BNBatch Normalization
PReLUParametric Rectified Linear Unit
SDKSoftware Development Kit
CNNConvolutional Neural Networks
DLDeep Learning
AIAAAI-Assisted Annotation
MONAIMedical Open Network for AI

References

  1. Merveille, O.; Talbot, H.; Najman, L.; Passat, N. Curvilinear structure analysis by ranking the orientation responses of path operators. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 304–317. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. 3D-IRCADb. 2022. Available online: https://www.ircad.fr/research/data-sets/liver-segmentation-3d-ircadb-01/ (accessed on 17 October 2022).
  3. Fedorov, A.; Beichel, R.; Kalpathy-Cramer, J.; Finet, J.; Fillion-Robin, J.; Pujol, S.; Kikinis, R. 3D Slicer as an image computing platform for the Quantitative Imaging Network. Magn. Reson. Imaging 2012, 30, 1323–1341. [Google Scholar] [CrossRef] [Green Version]
  4. Zhang, R.; Zhou, Z.; Wu, W.; Lin, C.C.; Tsui, P.H.; Wu, S. An Improved Fuzzy Connectedness Method for Automatic Three-Dimensional Liver Vessel Segmentation in CT Images. J. Healthc. Eng. 2018, 2018, 2376317. [Google Scholar] [CrossRef] [PubMed]
  5. Guo, X.; Xiao, R.; Zhang, T.; Chen, C.; Wang, J.; Wang, Z. A novel method to model hepatic vascular network using vessel segmentation, thinning, and completion. Med. Biol. Eng. Comput. 2020, 58, 709–724. [Google Scholar] [CrossRef] [PubMed]
  6. Ciecholewski, M.; Kassjański, M. Computational methods for liver vessel segmentation in medical imaging: A Review. Sensors 2021, 21, 2027. [Google Scholar] [CrossRef] [PubMed]
  7. Yu, W.; Fang, B.; Liu, Y.; Gao, M.; Zheng, S.; Wang, Y. Liver vessels segmentation based on 3D residual U-NET. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 250–254. [Google Scholar] [CrossRef]
  8. Huang, Q.; Sun, J.; Ding, H.; Wang, X.; Wang, G. Robust liver vessel extraction using 3D U-net with variant dice loss function. Comput. Biol. Med. 2018, 101, 153–162. [Google Scholar] [CrossRef]
  9. Affane, A.; Kucharski, A.; Chapuis, P.; Freydier, S.; Lebre, M.; Vacavant, A.; Fabijańska, A. Segmentation of liver anatomy by combining 3D U-net approaches. Appl. Sci. 2021, 11, 4895. [Google Scholar] [CrossRef]
  10. Golla, A.; Bauer, D.F.; Schmidt, R.; Russ, T.; Norenberg, D.; Chung, K.; Zollner, F.G. Convolutional Neural Network Ensemble segmentation with ratio-based sampling for the arteries and veins in abdominal CT scans. IEEE Trans. Biomed. Eng. 2021, 68, 1518–1526. [Google Scholar] [CrossRef]
  11. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Medical Image Computing and Computer-Assisted Intervention, Proceedings of the MICCAI 2016, Athens, Greece, 17–21 October 2016; Springer: Cham, Swtizerland, 2016; pp. 424–432. [Google Scholar] [CrossRef]
  12. Segmentation of the Liver Competition 2007 (SLIVER07). 2019. Available online: https://sliver07.grand-challenge.org/ (accessed on 17 October 2022).
  13. Milletari, F.; Navab, N.; Ahmadi, S. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar] [CrossRef] [Green Version]
  14. Su, J.; Liu, Z.; Zhang, J.; Sheng, V.S.; Song, Y.; Zhu, Y.; Liu, Y. DV-net: Accurate liver vessel segmentation via dense connection model with D-BCE loss function. Knowl.-Based Syst. 2021, 232, 107471. [Google Scholar] [CrossRef]
  15. Yang, J.; Fu, M.; Hu, Y. Liver vessel segmentation based on inter-scale V-net. Math. Biosci. Eng. 2021, 18, 4327–4340. [Google Scholar] [CrossRef]
  16. Altini, N.; Prencipe, B.; Brunetti, A.; Brunetti, G.; Triggiani, V.; Carnimeo, L.; Cascarano, G.D. A tversky loss-based convolutional neural network for liver vessels segmentation. In Intelligent Computing Theories and Application; Springer: Cham, Swtizerland, 2020; pp. 342–354. [Google Scholar] [CrossRef]
  17. Meng, X.; Wang, G.; Zhang, Y.; Shi, X.; Dai, H.; Wang, Z.; Wang, X. Exploiting full Resolution Feature Context for Liver Tumor and Vessel Segmentation via Integrate Framework: Application to Liver Tumor and Vessel 3D Reconstruction under embedded microprocessor. arXiv 2022. [Google Scholar] [CrossRef]
  18. Xu, M.; Wang, Y.; Chi, Y.; Hua, X. Training liver vessel segmentation deep neural networks on noisy labels from contrast CT Imaging. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 1552–1555. [Google Scholar] [CrossRef]
  19. Liver Tumor Segmentation Challenge (LiTS) Dataset. Available online: https://competitions.codalab.org/competitions/17094 (accessed on 17 October 2022).
  20. Shahid, M.; Taj, I.A. Robust retinal vessel segmentation using vessel’s location map and frangi enhancement filter. IET Image Process. 2018, 12, 494–501. [Google Scholar] [CrossRef]
  21. Soomro, T.A.; Afifi, A.J.; Ali Shah, A.; Soomro, S.; Baloch, G.A.; Zheng, L.; Yin, M.; Gao, J. Impact of image enhancement technique on CNN model for retinal blood vessels segmentation. IEEE Access 2019, 7, 158183–158197. [Google Scholar] [CrossRef]
  22. Blaiech, A.G.; Mansour, A.; Kerkeni, A.; Bedoui, M.H.; Ben Abdallah, A. Impact of enhancement for coronary artery segmentation based on deep learning neural network. In Iberian Conference on Pattern Recognition and Image Analysis; Springer: Cham, Swtizerland, 2019. [Google Scholar] [CrossRef]
  23. Lamy, J.; Merveille, O.; Kerautret, B.; Passat, N.; Vacavant, A. Vesselness filters: A survey with benchmarks applied to liver imaging. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 3528–3535. [Google Scholar] [CrossRef]
  24. Hamarneh, G.; Jassi, P. VascuSynth: Simulating vascular trees for generating volumetric image data with ground-truth segmentation and tree analysis. Comput. Med. Imaging Graph. 2010, 34, 605–661. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Survarachakan, S.; Pelanis, E.; Khan, Z.A.; Kumar, R.P.; Edwin, B.; Lindseth, F. Effects of enhancement on Deep Learning based hepatic vessel segmentation. Electronics 2021, 10, 1165. [Google Scholar] [CrossRef]
  26. 3D Slicer. 2022. Available online: https://www.slicer.org/ (accessed on 28 December 2021).
  27. NVIDIA Clara Imaging. 2022. Available online: https://developer.nvidia.com/clara-medical-imaging (accessed on 28 December 2021).
  28. Project MONAI. 2021. Available online: https://monai.io/ (accessed on 28 December 2021).
  29. Ai-Assisted-Annotation-Client. 2022. Available online: https://github.com/NVIDIA/ai-assisted-annotation-client/tree/master/slicer-plugin (accessed on 28 December 2021).
  30. Simpson, A.L.; Antonelli, M.; Bakas, S.; Bilello, M.; Farahani, K.; van Ginneken, B.; Kopp-Schneider, A.; Landman, B.A.; Litjens, G.; Menze, B.; et al. A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv 2019, arXiv:1902.09063. [Google Scholar]
  31. Yan, Q.; Wang, B.; Zhang, W.; Luo, C.; Xu, W.; Xu, Z.; You, Z. Attention-guided deep neural network with multi-scale feature fusion for liver vessel segmentation. IEEE J. Biomed. Health Inform. 2021, 25, 2629–2642. [Google Scholar] [CrossRef]
  32. Ng, C.; Yap, M.H.; Costen, N.; Li, B. Automatic wrinkle detection using Hybrid Hessian filter. In Computer Vision— Proceedings of the ACCV 2014, Singapore, 1–5 November 2014; Springer: Cham, Swtizerland, 2015; pp. 609–622. [Google Scholar] [CrossRef] [Green Version]
  33. Frangi, A.F.; Niessen, W.J.; Vincken, K.L.; Viergever, M.A. Multiscale vessel enhancement filtering. In Medical Image Computing and Computer-Assisted Intervention—Proceedings of the MICCAI’98, Cambridge, MA, USA, 11–13 October 1998; Springer: Berlin/Heidelberg, Germany, 1998; pp. 130–137. [Google Scholar]
  34. Meijering, E.; Jacob, M.; Sarria, J.; Steiner, P.; Hirling, H.; Unser, M. Design and validation of a tool for neurite tracing and analysis in fluorescence microscopy images. Cytometry 2004, 58A, 167–176. [Google Scholar] [CrossRef] [Green Version]
  35. Sato, Y.; Nakajima, S.; Shiraga, N.; Atsumi, H.; Yoshida, S.; Koller, T.; Kikinis, R. Three-dimensional multi-scale line filter for segmentation and visualization of curvilinear structures in medical images. Med. Image Anal. 1998, 2, 143–168. [Google Scholar] [CrossRef]
  36. RORPO GIT. Available online: https://github.com/path-openings/RORPO (accessed on 28 December 2021).
  37. Van der Walt, S.; Schönberger, J.L.; Nunez-Iglesias, J.; Boulogne, F.; Warner, J.D.; Yager, N.; Gouillart, E.; Yu, T. Scikit-image: Image processing in python. PeerJ 2014, 2, e453. [Google Scholar] [CrossRef]
  38. Wu, M.; Qian, Y.; Liao, X.; Wang, Q.; Heng, P.A. Hepatic vessel segmentation based on 3Dswin-transformer with inductive biased multi-head self-attention. arXiv 2021, arXiv:2111.03368. [Google Scholar]
  39. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Cham, Swtizerland, 2017; pp. 240–248. [Google Scholar] [CrossRef] [Green Version]
  40. Jadon, S. A survey of loss functions for semantic segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Viña del Mar, Chile, 27–29 October 2020. [Google Scholar] [CrossRef]
Figure 1. The main concept of the tool for medical image processing and analysis.
Figure 1. The main concept of the tool for medical image processing and analysis.
Applsci 13 00548 g001
Figure 2. Slicer’s Segment Editor with the AIAA extension.
Figure 2. Slicer’s Segment Editor with the AIAA extension.
Applsci 13 00548 g002
Figure 3. Annotations on Patient 1 from 3D-IRCADb dataset. (a) CT image; (b) Original Annotation; (c) Improved Annotations; (d) 3D of Original Annotation; (e) 3D of Improved Annotations.
Figure 3. Annotations on Patient 1 from 3D-IRCADb dataset. (a) CT image; (b) Original Annotation; (c) Improved Annotations; (d) 3D of Original Annotation; (e) 3D of Improved Annotations.
Applsci 13 00548 g003
Figure 4. Comparison of different vessel enhancement filters.
Figure 4. Comparison of different vessel enhancement filters.
Applsci 13 00548 g004
Figure 5. Workflow of fusing different enhancement filters prior to segmentation (FilterAdded).
Figure 5. Workflow of fusing different enhancement filters prior to segmentation (FilterAdded).
Applsci 13 00548 g005
Figure 6. Workflow of fusing vessel enhanced outputs after the segmentation (SegAdded), where X, Y = {He or Ro, Fr, Me, Sa}.
Figure 6. Workflow of fusing vessel enhanced outputs after the segmentation (SegAdded), where X, Y = {He or Ro, Fr, Me, Sa}.
Applsci 13 00548 g006
Figure 7. Workflow of linear blending of original and RORPO enhanced image.
Figure 7. Workflow of linear blending of original and RORPO enhanced image.
Applsci 13 00548 g007
Figure 8. Architecture of 3D U-Net segmentation model.
Figure 8. Architecture of 3D U-Net segmentation model.
Applsci 13 00548 g008
Figure 9. Architecture of 3D V-Net segmentation model.
Figure 9. Architecture of 3D V-Net segmentation model.
Applsci 13 00548 g009
Figure 10. Application of individual filters.
Figure 10. Application of individual filters.
Applsci 13 00548 g010
Figure 11. The average dice score for different blending ratios between the RORPO and the original image while using 3D U-Net and 3D V-Net.
Figure 11. The average dice score for different blending ratios between the RORPO and the original image while using 3D U-Net and 3D V-Net.
Applsci 13 00548 g011
Figure 12. Effects of different ablations amounts in all 3D V-Net stages.
Figure 12. Effects of different ablations amounts in all 3D V-Net stages.
Applsci 13 00548 g012
Figure 13. In each case, the first row shows the ground truth, and the second row shows the predictions in the axial, sagittal, coronal, and 3D views.
Figure 13. In each case, the first row shows the ground truth, and the second row shows the predictions in the axial, sagittal, coronal, and 3D views.
Applsci 13 00548 g013
Table 1. Overview of publicly available datasets.
Table 1. Overview of publicly available datasets.
Name of datasetPublishedNumber of patients
3D-IRCADb [2]5 May 201922 patients
Medical Segmentation Decathlon (MSD) [30]20 December 2020430 patients
Vascular Synthesizer (VascuSynth) [24]March 2013120 synthetic samples
Table 2. Loss Functions: where g i is the ground truth; p i is the prediction.
Table 2. Loss Functions: where g i is the ground truth; p i is the prediction.
Loss FunctionEquationDefinition
Dice [39]
2 i N p i g i i N p i 2 + i N g i 2
Calculates the overlap of g i and p i .
Dice-BCE [40]
1 Dice + λ b L B C E
L B C E = ( p i log ( g i ) + ( 1 p i ) log ( 1 g i ) )
Evaluates the class prediction of each pixel vector, and then averages all pixels. λ b is 0.5.
Log Cosh Dice [40]
log ( cosh ( p i g i ) ) + λ l D i c e
For smoothing variations, can be used for skewed dataset. λ l is 0.5.
Focal [40]
α t 1 L B C E γ log L B C E
L B C E = ( p i log ( g i ) + ( 1 p i ) log ( 1 g i ) )
It up-weights the contributions of difficult examples. Here, γ is 0.5, α t is 0.5.
Dice Focal [40]
D i c e + λ f F o c a l
Variant of Dice loss, with focus on difficult examples. λ f is 0.5.
Tversky [40]
i N p i g i i N p i g i + β t i N ( 1 p i ) g i + α t i N p i ( 1 g i )
Adds weight to False positives and False negatives, where α t is 0.5 and β t is (1 − α t ).
Weighted Dice [38]
1 i N p i g i i N p i g i + α l β l i N ( 1 p i 2 ) i N ( 1 g i 2 )
Data with unbalanced classes, where α l is 0.5 and β l is 6 [38].
Table 3. Measures of segmentation quality. TP and TN stand for True Positive and True Negative (correctly segmented pixels), while FP and FN stand for False Positive and False Negative (incorrectly segmented pixels).
Table 3. Measures of segmentation quality. TP and TN stand for True Positive and True Negative (correctly segmented pixels), while FP and FN stand for False Positive and False Negative (incorrectly segmented pixels).
MetricsEquationExplanation
Dice Score 2 T P 2 T P + F P + F N The overlap of the two segmentations, divided by the total size
Accuracy T P + T N T P + T N + F P + F N Correctly predicted data points out of all the data points
Sensitivity T P T P + F N When identifying TP and the cost of an FN is high
Precision T P T P + F P When identifying TP and the cost of an FP is high
Table 4. Comparison of different enhancement filters on the 3D-IRCADb dataset and 3D U-Net as a segmentation network.
Table 4. Comparison of different enhancement filters on the 3D-IRCADb dataset and 3D U-Net as a segmentation network.
FilterOriginal Annotations
DSCACCSENPRC
Unenhanced 56.4 ± 13.8 98.3 ± 2.3 66.0 ± 14.2 56.5 ± 23.9
Hessian 60.0 ± 12.1 99.0 ± 0.3 62.9 ± 6.2 62.4 ± 21.6
Frangi 62.1 ± 6.6 99.0 ± 0.4 74.6 ± 6.6 53.9 ± 8.6
Meijering 62.4 ± 8.9 99.0 ± 0.7 69.1 ± 9.2 57.6 ± 11.9
Sato 58.3 ± 9.1 98.7 ± 0.9 68.4 ± 7.3 51.5 ± 11.4
RORPO62.7 ± 6.699.2 ± 0.260.7 ± 6.769.4 ± 19.1
FilterImproved annotations
DSCACCSENPRC
Unenhanced 62.7 ± 15.4 98.2 ± 2.3 72.5 ± 13.7 63.2 ± 25.5
Hessian 66.1 ± 8.0 99.0 ± 0.3 68.3 ± 5.3 65.4 ± 13.0
Frangi 69.6 ± 3.9 99.1 ± 0.1 75.7 ± 6.1 64.6 ± 3.7
Meijering 66.7 ± 10.1 98.7 ± 1.0 76.8 ± 6.2 61.2 ± 16.2
Sato 67.0 ± 4.8 98.9 ± 0.4 77.7 ± 7.1 59.2 ± 4.9
RORPO69.7 ± 4.899.2 ± 0.162.9 ± 6.179.1 ± 9.0
Table 5. Comparison of segmentation pipelines with enhancement filters applied to improved 3D-IRCADb with standard Dice Loss function and different networks.
Table 5. Comparison of segmentation pipelines with enhancement filters applied to improved 3D-IRCADb with standard Dice Loss function and different networks.
Filter3D U-Net
DSCACCSENPRC
Unprocessed 62.7 ± 15.4 98.2 ± 2.3 72.5 ± 13.7 63.2 ± 25.5
FilterAdded H e 67.8 ± 4.899.0 ± 0.472.8 ± 5.664.4 ± 9.0
FilterAdded R o 71.7 ± 3.2 99.1 ± 0.2 77.3 ± 5.0 67.1 ± 3.7
SegAdded H e 70.2 ± 4.299.0 ± 0.379.5 ± 6.163.1 ± 5.0
SegAdded R o 70.9 ± 3.8 99.0 ± 0.3 79.3 ± 6.1 64.3 ± 4.2
Our method74.4 ± 4.699.3 ± 0.172.7 ± 5.876.5 ± 6.4
Filter3D V-Net
DSCACCSENPRC
Unprocessed 68.4 ± 12.2 98.7 ± 1.1 80.7 ± 3.6 61.3 ± 17.3
FilterAdded H e 68.8 ± 4.4 99.1 ± 0.2 69.5 ± 4.9 68.3 ± 6.0
FilterAdded R o 73.5 ± 3.8 99.2 ± 0.1 75.9 ± 4.8 71.5 ± 5.3
SegAdded H e 72.9 ± 3.5 99.1 ± 0.1 81.5 ± 4.4 66.1 ± 4.4
SegAdded R o 72.7 ± 3.7 99.1 ± 0.1 82.1 ± 3.8 65.5 ± 5.0
Our method75.4 ± 4.399.3 ± 0.1 76.9 ± 4.1 74.4 ± 7.8
Table 6. Comparison of different loss functions tested on the improved 3D-IRCADb dataset and the 3D V-Net architecture.
Table 6. Comparison of different loss functions tested on the improved 3D-IRCADb dataset and the 3D V-Net architecture.
Loss Functions3D V-Net
DSCACCSENPRC
Dice Loss 75.4 ± 4.3 99.3 ± 0.1 76.9 ± 4.1 74.4 ± 7.8
Dice-BCE Loss 75.9 ± 3.5 99.3 ± 0.1 75.5 ± 3.1 76.8 ± 7.1
Log Cosh Dice Loss 74.6 ± 3.3 99.3 ± 0.1 73.5 ± 4.2 76.3 ± 7.4
Focal Loss 74.4 ± 3.7 99.3 ± 0.1 70.2 ± 4.0 79.6 ± 8.0
Dice Focal Loss 75.2 ± 4.3 99.3 ± 0.1 75.2 ± 4.3 75.6 ± 7.0
Tversky Loss 76.0 ± 4.2 99.3 ± 0.1 75.7 ± 3.9 76.8 ± 7.1
Weighted Dice Loss β l = 676.2 ± 3.799.3 ± 0.175.0 ± 3.477.7 ± 6.1
Table 7. Comparison of the proposed method, with related works.
Table 7. Comparison of the proposed method, with related works.
FiltersPublishedDice scoreAccuracySensitivitySpecificity
Indirect comparison
Zhang et al. [4]201867.396.473.797.4
Huang et al. [8]201875.397.676.798.8
Su et al. [14]202175.46-76.9-
Yang et al. [15]202171.698.575.499.5
Xu et al. [18]202068.799.878.699.2
Direct comparison (re-implemented related works)
Altini et al. [16]202065.0 ± 9.799.8 ± 0.452.3 ± 10.888.9 ± 6.2
Lamy et al. [23]202059.6 ± 5.898.9 ± 0.355.9 ± 4.264.8 ± 12.0
Our method202276.2 ± 3.799.3 ± 0.175.0 ± 3.499.7 ± 0.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Svobodova, P.; Sethia, K.; Strakos, P.; Varysova, A. Automatic Hepatic Vessels Segmentation Using RORPO Vessel Enhancement Filter and 3D V-Net with Variant Dice Loss Function. Appl. Sci. 2023, 13, 548. https://doi.org/10.3390/app13010548

AMA Style

Svobodova P, Sethia K, Strakos P, Varysova A. Automatic Hepatic Vessels Segmentation Using RORPO Vessel Enhancement Filter and 3D V-Net with Variant Dice Loss Function. Applied Sciences. 2023; 13(1):548. https://doi.org/10.3390/app13010548

Chicago/Turabian Style

Svobodova, Petra, Khyati Sethia, Petr Strakos, and Alice Varysova. 2023. "Automatic Hepatic Vessels Segmentation Using RORPO Vessel Enhancement Filter and 3D V-Net with Variant Dice Loss Function" Applied Sciences 13, no. 1: 548. https://doi.org/10.3390/app13010548

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop