Next Article in Journal
Harnessing Carrot Discard as a Novel Feedstock for 2,3-Butanediol Bioproduction: A Comparison of Fermentation Strategies and Bacillus Performance
Previous Article in Journal
Deep Learning-Based Layout Analysis Method for Complex Layout Image Elements
Previous Article in Special Issue
Automatic Identification and Description of Jewelry Through Computer Vision and Neural Networks for Translators and Interpreters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable Liver Segmentation and Volume Assessment Using Parallel Cropping

1
ORACLE, Dubai P.O. Box 500099, United Arab Emirates
2
Department of Electronic Systems, Aalborg Universitet, 2450 Copenhagen, Denmark
3
The Maersk Mc-Kinney Moller Institute, University of Southern Denmark, 5230 Odense, Denmark
4
NVIDIA, 8004 Zürich, Switzerland
5
Maimonides Biomedical Research Institute of Cordoba (IMIBIC), Reina Sofia University Hospital, 14008 Córdoba, Spain
6
Department of Electronic and Computer Engineering, Universidad de Córdoba, 14071 Córdoba, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(14), 7807; https://doi.org/10.3390/app15147807
Submission received: 5 June 2025 / Revised: 23 June 2025 / Accepted: 3 July 2025 / Published: 11 July 2025
(This article belongs to the Special Issue Image Processing and Computer Vision Applications)

Abstract

Accurate liver segmentation and volume estimation from CT images are critical for diagnosis, surgical planning, and treatment monitoring. This paper proposes a GPU-accelerated voxel-level cropping method that localizes the liver region in a single pass, significantly reducing unnecessary computation and memory transfers. We integrate this pre-processing step into two segmentation pipelines: a traditional Chan-Vese model and a deep learning U-Net trained on the LiTS dataset. After segmentation, a seeded region growing algorithm is used for 3D liver volume assessment. Our method reduces unnecessary image data by an average of 90%, speeds up segmentation by 1.39× for Chan-Vese, and improves dice scores from 0.938 to 0.960. When integrated into U-Net pipelines, the post-processed dice score rises drastically from 0.521 to 0.956. Additionally, the voxel-based cropping approach achieves a 2.29× acceleration compared to state-of-the-art slice-based methods in 3D volume assessment. Our results demonstrate high segmentation accuracy and precise volume estimates with errors below 2.5%. This proposal offers a scalable, interpretable, efficient liver segmentation and volume assessment solution. It eliminates unwanted artifacts and facilitates real-time deployment in clinical environments where transparency and resource constraints are critical. It is also tested in other anatomical structures such as skin, lungs, and vessels, enabling broader applicability in medical imaging.

1. Introduction

The liver is one of the most vital organs in the human body. Liver diseases are on the rise [1]; statistics show that more than 360 million people worldwide suffer from abnormal liver dysfunction. Each year, approximately 1.1 million people are diagnosed with new hepatocellular carcinoma [2]. The incidence of liver cancer is ranked sixth among all cancers, while it is fourth in terms of mortality [3]. The liver tumor is characterized by easy relapse, rapid deterioration, and easy metastasis. The success rate for liver surgery is relatively low; therefore, there is an urgent demand to improve it.
Accurate liver segmentation and volume assessment are critical in clinical settings, particularly for surgical planning, tumor monitoring, and treatment response evaluation. However, these tasks remain challenging due to the complex and variable anatomy of the liver, the low contrast between adjacent organs in computed tomography (CT) images, and the frequent presence of imaging artifacts. Moreover, manual segmentation is time-consuming and subject to inter-operator variability, underscoring the need for automated, reliable, and interpretable solutions. It remains challenging due to anatomical variations and pathological conditions that affect its morphology. Recent advances in artificial intelligence (AI), particularly in medical image analysis, have enabled substantial improvements in segmentation performance [4]. Nonetheless, many AI-based approaches still suffer from poor interpretability and high computational requirements.
Deep learning (DL) techniques, such as convolutional neural networks and U-Net architectures, have shown high accuracy in liver segmentation tasks. However, they often function as black boxes, lacking interpretability, and demand extensive labeled datasets and powerful hardware. Traditional methods like Chan-Vese offer transparent segmentation logic, but suffer from slow execution, especially with high-resolution medical volumes. These limitations underscore the need for a hybrid approach that offers both speed and explainability while maintaining competitive accuracy. Beyond liver imaging, recent deep learning research has shown significant advances in dermatology and breast cancer screening. For example, CNN-based systems have been successfully tailored to enhance recognition of skin lesions [5], while augmentation strategies have improved deep learning robustness in mammography [6].
This study addresses these limitations by introducing an explainable approach that significantly reduces computation time and enhances accuracy using a ’Parallel Cropping’ technique, overcoming the current constraints identified in recent literature.
Liver segmentation and volume estimation are challenging tasks in medical imaging [7]. The low contrast between the adjacent organs and the liver is the main problem with liver segmentation [8]. Liver images contain a lot of unnecessary data that affects the overall process’s speed and the liver segmentation’s accuracy. Many algorithms have recently been developed for liver segmentation. We classify these approaches into two categories: traditional and learning approaches. Seeded region growth stands out as a traditional approach for segmentation and volume assessment [9,10,11,12]. Xuesong et al. proposed automatic liver segmentation using a multi-dimensional graph cut with shape information [13]. U-Net is a convolutional network approach used for biomedical image segmentation [14,15]. U-Net applied to liver segmentation sometimes segments the part outside the liver boundary [14]. In contrast, the liver segmentation based on Chan-Vese is slow due to the unwanted data in the medical image [16,17].
Learning-based algorithms require extensive datasets for training and validation, utilizing complete liver images obtained by CT [18]. DL solutions require high computational resources and labeled data availability, which is not always available; it is currently scarce. Numerous studies have advanced the field of liver segmentation using deep learning models. Transformer-based architectures such as Swin-Transformer and hybrid attention-based networks like DRDA-Net have recently demonstrated superior performance on liver and tumor segmentation tasks [19]. These models excel in accuracy but are computationally intensive and require extensive labeled datasets, which limit their real-time deployment and interpretability. Our work takes a complementary path by focusing on pre-processing techniques that can be integrated into such pipelines, reducing input complexity and accelerating inference.
So, the most common liver dataset is LiTS [20], which provides only 131 livers. Recent DL-based proposals in prestigious journals have used LiTS. RMAU-Net [21] consists of a segmentation network incorporating residual and multi-scale attention mechanisms to enhance liver and tumor segmentation from CT scans. Ref. [22] explores a hybrid approach combining deep residual networks with a biplane segmentation method to optimize liver boundary detection and volume estimation. Both articles present high-performing segmentation methods. Their reliance on computational power and large datasets limits their applicability in real-time, resource-limited environments.
DL models operate as black boxes. The decision-making process behind the segmentation results is not easily interpretable, making it difficult for clinicians to understand why the model produces certain outputs, especially in edge cases or when anomalies are present. In the medical field, transparency and the ability to explain outcomes are crucial for gaining trust and acceptance from healthcare professionals. The lack of interpretability in these DL models poses a barrier to clinical adoption, as practitioners might be hesitant to rely on a system that cannot provide clear justifications for its results. When a model performs poorly or produces incorrect segmentations, it is difficult to diagnose and address the cause of the error without a clear understanding of how the model processes the data. This lack of insight makes it challenging to fine-tune or adapt the model for different datasets or clinical environments, limiting its flexibility and robustness.
In contrast, traditional methods provide transparency and robustness but suffer from slow execution, particularly when handling large volumetric medical images containing extensive irrelevant regions outside the organ of interest (RoI) [23]. These unwanted elements within liver images pose challenges for real-time segmentation implementation, detrimentally affecting segmentation accuracy [9]. Processing medical images entails managing vast amounts of data, often taxing memory resources [24]. When executed on a graphics processing unit (GPU), insufficient GPU memory hampers the segmentation process [25]. GPUs are pivotal in accelerating processes within image-guided surgery (IGS) [9,10]. Employing cropping techniques aids in eliminating unwanted elements from medical images, enhancing overall process efficiency.
We aim to achieve a rapid parallel cropped image for liver segmentation and volume assessment while minimizing intermediate memory transfers between the central processing unit (CPU) and GPU. This strategy reduces processing overhead on unnecessary image elements, obtaining significant speed improvements. Our proposal is based on the persistent thread block (PT) approach [26]. The PT approach relies on the number of computing blocks available. In cases where there are insufficient computing blocks to process image elements independently, the remaining elements’ workload is distributed among existing blocks. These blocks iterate over all image elements until complete processing, utilizing a grid-stride loop described by Gupta et al. [26]. Persistence and grid-stride loop-based cropping offer the flexibility to harness parallelism while minimizing the processing of unwanted voxel elements. This technique consists of two parts: first, calculating diagonally opposite cropping points in a single GPU kernel call, followed by efficient data transfer from the original to the cropped image.
The main contributions of this work are as follows: (1) A novel voxel-based parallel cropping algorithm is used that significantly reduces unnecessary image content before segmentation. (2) We demonstrate that this approach improves segmentation accuracy and computational efficiency when integrated with traditional (Chan-Vese) and deep learning (U-Net) models, but other models could be integrated, such as U-net++ [27], Swin-UNet [28] or TransUNet [29]. (3) Our voxel-based approach computes cropping boundaries in a single pass across all dimensions, unlike previous slice-based methods in 3D volume assessment, minimizing memory transfers and redundant kernel launches, resulting in a 2.29× speedup without sacrificing accuracy. (4) A clinically interpretable, resource-efficient alternative to DL-based segmentation pipelines is provided, offering practical deployment advantages for real-time medical imaging applications. A new pipeline diagram (Figure 1) illustrates our method’s integration into existing segmentation workflows.
The rest of the paper is structured as follows. Section 2 briefly reviews the relevant background and motivation for crop images and liver segmentation. Section 3 explains the parallel GPU approaches for automatic and semiautomatic cropping. Section 4 and Section 5 discuss the application of parallel cropping to liver segmentation and volume assessment. Section 6 analyzes performance results and the persistent and grid-stride loop-based parallel cropping comparison for liver segmentation and volume assessment. Section 7 summarizes the main conclusions.

2. Background and Motivation

In medical image analysis, particularly for liver segmentation, input data often contains substantial irrelevant information that does not contribute to the target RoI. Processing these unnecessary voxels increases memory usage, causes longer computation times, and reduces overall system responsiveness. Traditional slice-based cropping approaches involve multiple kernel invocations and memory transfers between CPU and GPU. Removing unnecessary data reduces memory usage, avoiding redundant memory transfers and unnecessary computations and, thus, reducing execution time [9,30]. Cropping helps select the desired RoI of the targeted application in the analysis of medical imaging data. Due to poor image quality, the object’s boundary in RoI is sometimes diffused in a CT liver image, making it challenging to find the RoI automatically [31].
Previous works such as [9,30] have illustrated the computational drawbacks of these multi-pass approaches. We propose a GPU-based voxel-level cropping technique integrated into the segmentation pipeline to discard irrelevant data in a single pass, dramatically reducing kernel calls and intermediate memory operations.
From a clinical standpoint, accelerating liver segmentation pipelines is essential for time-sensitive workflows such as preoperative planning or intraoperative navigation [7]. In hospitals with constrained computational resources, the proposed cropping method enables the deployment of segmentation models with smaller memory footprints and faster execution, facilitating broader accessibility and practical integration into daily radiological operations [23,31].
Yu et al. [32] proposed the creation of a robust segmentation model that maintains high accuracy even when incomplete overlap between liver lesions in different imaging phases exists. Ao et al. [33] proposed a novel Transformer–CNN combined network using multi-scale feature learning for three-dimensional (3D) medical image segmentation. Smistad et al. [30,34] have discussed parallel cropping for image segmentation. Tran et al. explored U-Net derived from a traditional U-Net for the liver and liver–tumor segmentation. Their method utilized dilated convolution and dense structure [35]. In [36], Liu et al. used artificial intelligence based on the k-means clustering method for liver segmentation. Their model shows improved performance when compared with the region-growing method. Li et al. [37] proposed Attention U-Net++ (a nested attention-aware segmentation network) with deep supervised encoder–decoder architecture and a redesigned dense skip connection. Fan et al. [38] presented a multi-scale attention net combing self-attention algorithm that adaptively integrates local features with global dependencies. Nawandhar et al. [39] developed a stratified squamous epithelium biopsy image classifier-based sequential computing environment for image segmentation. Finally, Gul et al. [23] provided an overview of the available deep learning approaches for segmenting the liver and detecting liver tumors and their evaluation metrics, including accuracy, volume overlap error, dice coefficient, and mean square distance.
Most recently, some DL implementations have been developed for liver and tumor segmentation. Ref. [40] proposes a deep learning architecture based on the Swin-Transformer structure, which is employed to capture the spatial relationships and global information for liver and tumor segmentation. The DRDA-Net framework [41] analyzes CT liver images, integrating deep residual learning and dual-attention mechanisms to extract local and global contextual features. Finally, ref. [42] presents a hybrid ResUNet model combining the ResNet and U-Net models.
They crop the medical image dataset before processing. The CPU allocates the memory equivalent to the cropped size to copy data to the cropped image on the GPU. Further seeded region growing is performed for image segmentation. This is the simplistic representation of the work by Smistad et al. [30]. They have proposed a non-PT (non-persistent thread) approach for cropping and seeded region growth. Persistence implies the number of active computing threads processing a complete image. If the number of active computing threads is lesser than the image elements, then the task of the remaining image elements is distributed amongst the threads. The total number of active persistent threads implies the grid size.
Smistad et al. [30] implemented a slice-based parallel cropping algorithm. Slice-based parallel cropping needs three kernel invocations to estimate cropping points and one kernel invocation to generate cropped data within cropping points. There are four kernel calls for 3D parallel cropping, which impacts the performance. Once the boundary is detected using cropping points, it is necessary to crop the image properly. Over-cropping of the liver crops part of the liver and destroys the segmentation quality. Under-cropping crops a small amount of data from the image, resulting in a limited speedup. It is necessary to balance the speedup and accuracy using cropping and parallelization to avoid the above-mentioned problems. Hence, we propose persistence and voxel-based parallel cropping to improve the performance of liver segmentation and volume assessment by reducing unwanted computations and avoiding intermediate memory transfers between CPU and GPU. We discuss parallel cropping in the following section in more detail.

3. Methodology

Our hypothesis proposes a voxel-based approach to compute all cropping boundaries in a single pass across all dimensions. This should minimize memory transfers and redundant kernel launches, increasing performance without sacrificing accuracy. We propose a GPU-optimized parallel cropping method that replaces traditional multi-pass, slice-based approaches with a single, volume-level processing step to achieve this. The technique leverages persistent threads and grid-stride loops to efficiently localize and extract the RoI, improving speed and memory utilization.
We describe two novel cropping techniques: semiautomatic and automatic. At first, semiautomatic cropping is proper for low-quality images with unclear boundaries, requiring minimal user interaction to indicate two diagonally opposite points that surround the liver. Automatic cropping, on the other hand, is fully data-driven and relies on voxel-level thresholding to identify the RoI without manual input.
Both cropping variants are implemented as GPU kernels that compute the RoI’s bounding coordinates and then transfer only the relevant voxel data to a new memory buffer. This compact representation is then passed to the segmentation module, significantly reducing input size and enabling faster downstream processing.
We integrate the proposed cropping step into two segmentation pipelines: traditional Chan-Vese and deep learning-based models. The Chan-Vese segmentation method is particularly well-suited for cropped images due to its robustness and mathematical transparency. It iteratively evolves a contour to fit the liver boundary and its performance benefits significantly from the reduced input size achieved via cropping.
In parallel, we apply the cropped images to a U-Net model trained on the LiTS dataset. The reduction in input complexity accelerates the model’s inference time. However, resizing steps introduced by the fixed-size input requirement may reduce raw segmentation accuracy. To address this, a post-processing filter is applied to retain only the largest connected component, effectively eliminating false positives. This combination enables high-precision segmentation with reduced computational demands.
Both segmentation outputs—whether from Chan-Vese or post-processed U-Net—are inputs to the final volume assessment stage. We compute the total liver volume by applying a 3D-seeded region growing on the segmented volume. The accuracy and efficiency of this stage are directly enhanced by the compactness of the cropped data, which reduces voxel traversal and memory overhead.

3.1. Semiautomatic Cropping

CT liver images are sometimes poor in image quality. The boundary of the liver in poor-quality images is diffused sometimes, requiring manual intervention during the processing of the liver images. In semiautomatic cropping, at first, two diagonally opposite points are provided by the clinician expert. Then, we calculate the number of voxels inside the cuboid (in 3D volume, as shown in Figure 2) within the cropping points to allocate the required GPU memory for the cropped image in the GPU.
The CPU calls a kernel on the GPU to copy cropped data from the original image. The advantage is computing units process tiles within the cropped image rather than processing the whole image, as done in the non-PT approach. The process is shown in Algorithm 1.
For each voxel in parallel, the GPU calculates the position of the voxel on the original and cropped image given by “index” and “index_crop”, respectively. It checks if this voxel is within the diagonally opposite cropping points. The data within the cropping points is copied on the GPU to create a cropped image. The semiautomatic cropping is essential where the boundaries are diffused enough that it is difficult to calculate the RoI automatically. This study explores liver segmentation as one of the applications for semiautomatic cropping. In the next section, we discuss automatic cropping for liver volume assessment.
Algorithm 1: Copy Cropped Data
  1:
for each voxel from crop size in parallel
  2:
index=g(voxel.x, voxel.y, voxel.z);
  3:
index_crop=f(voxel_crop.x, voxel_crop.y, voxel_crop.z);
  4:
if  v o x e l . x > = l o w e r 0 && v o x e l . x < u p p e r 0  then
  5:
    if  v o x e l . y > = l o w e r 1 && v o x e l . y < u p p e r 1  then
  6:
        if  v o x e l . z > = l o w e r 2 && v o x e l . z < u p p e r 2  then
  7:
            d a t a _ c r o p [ i n d e x _ c r o p ] = d a t a [ i n d e x ] ;
  8:
        end if
  9:
    end if
10:
end for

3.2. Automatic Cropping

We aim to automate the process of cropping in medical image analysis. Automatic cropping does not involve any user interaction. This approach finds diagonally opposite points automatically to generate the cropped image. For each voxel, the method determines if it belongs to the RoI. The process is shown in Algorithm 2. There are two diagonal opposite points, lower (i.e., lower0, lower1, lower2) and upper (i.e., upper0, upper1, upper2), as shown in Figure 2. We initialize lower points by the size of the image and upper points to 0. For each voxel in parallel, if it takes a value of 1, the lower values (lower0, lower1, lower2) are assigned to the minimum values of the indices, and the upper values (upper0, upper1, upper2) are assigned to the maximum values of the indices from the voxels in x, y and z directions.
Algorithm 2: Voxel-based Cropping
  1:
initialization of global variables;
  2:
lower0=width; lower1=height; lower2=depth;
  3:
upper0=upper1=upper2=0;
  4:
for each voxel in parallel perform
  5:
if threshold0<value(voxel)<threshold1 then
  6:
    lower0=atomicMin(&lower0,voxel.x);
  7:
    lower1=atomicMin(&lower1,voxel.y);
  8:
    lower2=atomicMin(&lower2,voxel.z);
  9:
    upper0=atomicMax(&upper0,voxel.x);
10:
    upper1=atomicMax(&upper1,voxel.y)
11:
    upper2=atomicMax(&upper2,voxel.z)
12:
end if
The voxels are processed by threads in parallel and the threads may overlap inaccurate values of the diagonally opposite points. Hence, these operations of finding maximum and minimum values need to be atomic so that valid data is assigned to each of the diagonally opposite points, “lower” and “upper”. When all the voxels are processed, the final RoI is represented by a cuboid shape, as shown in Figure 2. Once the cropping points are obtained, the task is to create a cropped image applying the same approach described in Algorithm 1 for semiautomatic cropping. We use automatic cropping for 3D volume assessment and discuss the applications of fast parallel cropping to liver segmentation and 3D volume assessment in the following sections.

3.3. Post-Processing

We apply a post-processing step after the initial prediction to improve the quality of U-Net-based liver segmentation. This step filters out small, spurious regions by retaining only the largest connected component in the segmentation mask. The rationale is that the liver typically forms the most voluminous continuous structure in the field of view.
The post-processing procedure is presented in Algorithm 3.
Algorithm 3: Post-processing after U-Net Segmentation
  • Require: Binary segmentation mask SegmentationMask from U-Net
  • Ensure: Filtered mask CleanedMask
  1:
Label all connected components in SegmentationMask
  2:
for each component do
  3:
    Compute voxel count (volume)
  4:
end for
  5:
Identify the largest component L max
  6:
CleanedMask← binary mask containing only L max
  7:
return CleanedMask
This step is particularly effective in removing false positives that result from U-Net’s tendency to segment outside the target region when encountering low-contrast structures or noise. We observed that applying this step leads to a marked improvement in Dice scores and contour accuracy.
Despite U-Net’s high accuracy on the full dataset, it occasionally segments structures outside the liver boundary, particularly in low-contrast regions. This often results in unwanted elements (false positives), as shown in U-Net, Figures 5–12 where the model erroneously includes adjacent organs. To correct this, a post-processing step was introduced to retain only the largest connected component in the output mask (U-Net Figures 13–16, effectively isolating the liver and improving segmentation fidelity.
Finally, it is necessary to resize cropped inputs back to the fixed input shape required by U-Net.

4. Application to Liver Segmentation

This study examines the impact of the cropping on the liver segmentation and the application of semiautomatic cropping to segment the liver. The Chan-Vese implementation by Siri and Latte [16] and the U-Net framework by Ronneberger et al. [14] are discussed in this section for liver segmentation. We crop the original liver slice using the proposed voxel-based parallel cropping approach. The cropped liver slice is provided as an input to both models, i.e., traditional liver segmentation using Chan-Vese and learning-based liver segmentation using U-Net. The study analyzes segmented liver output from both models and the impact of cropping on liver segmentation.
Liver segmentation using the Chan-Vese and U-Net approaches is shown in Figure 3. We apply parallel cropping on the input liver slice to obtain the cropped liver slice. On the left, the Chan-Vese algorithm is applied, obtaining the final segmentation. Similarly, on the right, liver segmentation using U-Net is computed. The U-Net result presents an unwanted segmentation outside the liver, on the right. This unwanted segmentation outside the liver is removed in the post-processing step, obtaining the image’s maximum component and filtering out the other unwanted segmentation. U-Net has been trained using the (LiTS) dataset [20], which is composed of 131 CT liver volumes (58638 CT liver slices). We used 100 volumes for training, 10 for validation, and 21 for testing.
The U-Net model was trained for liver segmentation using 58,638 axial slices derived from 131 annotated CT volumes in the LiTS dataset. We applied the Adam optimizer with a learning rate of 10 3 , β values ( 0 . 9 ,   0 . 999 ) , and a weight decay of 10 5 for regularization. Training was performed for 250–300 epochs with early stopping and learning rate scheduling, which allowed fine segmentation convergence without overfitting. To improve generalization, we employed data augmentation including random horizontal and vertical flips as well as random rotations.
The loss function for binary segmentation combined binary cross-entropy (BCE) and Dice loss in an equal proportion:
L = 0.5 · L BCE + 0.5 · L Dice ,
where L BCE represents binary cross-entropy loss, and L Dice denotes the Dice loss computed between the predicted and ground truth masks. This formulation balances pixel-wise classification accuracy with region-level overlap, which is particularly important for liver boundary precision and accurate volume estimation.
We assess the quality of liver segmentation on original and cropped images and discuss the results and improvement in segmentation quality in the performance evaluation section. In the next section, the study analyzes liver volume assessment as an application for 3D cropping.

5. Application to 3D Volume Assessment

The study aims to assess liver volume as efficiently as possible. We propose fast parallel voxel-based 3D cropping for volume assessment. Parallel automatic cropping is implemented as the initial stage, and 3D seeded region growing is applied. The complete process is described below.
By applying parallel voxel-based cropping before seeded region growing (SRG), we significantly reduce the number of voxels involved in the segmentation process. This reduction lowers memory consumption on the GPU and accelerates the segmentation computation. The cropping ensures that seeded region growing operates within a tightly bounded region of interest (RoI), enhancing efficiency without sacrificing accuracy. This is particularly beneficial in high-resolution CT volumes, where traditional full-volume SRG methods are prohibitively expensive.
The SRG algorithm is implemented using GPU-based persistent threads with kernel relaunching (KTRL) to enable effective exploration of 3D neighborhoods. After cropping, the seed point is automatically determined within the binary mask, and parallel region growing is used to aggregate connected liver tissue voxels. The volume is computed as the product of labeled voxel count and voxel spacing, yielding precise liver volume estimates directly from CT data. The proposed parallel approach for 3D volume assessment is shown in Figure 4.
The CPU stores the 3D images on the GPU. We apply PT and voxel-based parallel cropping to obtain the necessary RoI on the CT liver image. The 3D volume of the desired organ inside this RoI is assessed using PT and shared memory-based SRG. The parallel implementation of SRG is carried out using kernel termination and relaunch (KTRL) [34]. The SRG algorithm is implemented using GPU-based persistent threads with kernel relaunching (KTRL) to enable effective exploration of 3D neighborhoods. After cropping, the seed point is automatically determined within the binary mask, and parallel region growing is used to aggregate connected liver tissue voxels. Volume is computed as the product of labeled voxel count and voxel spacing, yielding precise liver volume estimates directly from CT data. The proposed parallel approach for 3D volume assessment is shown in Figure 4.
The CPU transfers the CT liver image to the GPU. The liver image is cropped to obtain the necessary RoI. We apply KTRL-based SRG to the cropped image. The volume of the desired organ is assessed using KTRL-based SRG on the static RoI of the image tiles, and the processed image is stored back on the CPU.
To demonstrate the generalizability of the proposed cropping method, we evaluated its performance on multiple datasets from the 3D Image Reconstruction for Comparison of Algorithm Database (3DIRACDb [43]), including vessel, skin, and lung CT volumes. In all cases, the voxel-based cropping consistently accelerated the preprocessing and volume estimation stages when compared to conventional slice-based methods. This shows that our approach is not limited to liver imaging but is broadly applicable to other anatomical structures.
In the next section, we evaluate the impact of cropping on liver segmentation and volume assessment, discuss the speedup obtained by cropping, and analyze the segmentation accuracy.

6. Experimentation and Discussion

In this section, we analyze the performance improvements achieved through the proposed parallel cropping techniques, focusing on liver segmentation and volume assessment. Our evaluation emphasizes the gains in both accuracy and processing speed. The comparison between voxel- and slice-based cropping is discussed for the liver volume assessment. We used Intel(R) Core(TM) i7-7700HQ CPU @ 2.80GHz RAM 24 GB, NVIDIA GPU GeForce GTX 1050 (RAM 4GB), OpenCL 1.2 and CUDA Toolkit 10.1 for the implementation.

6.1. Liver Segmentation

This study aims to analyze the impact of cropping on liver segmentation. The segmented results using Chan-Vese [16] and U-Net [14] are shown for four slices. We show the liver segmentation on the original CT liver slices in Figure 5, Figure 6, Figure 7 and Figure 8. The liver segmentation on cropped liver slices is shown in Figure 9, Figure 10, Figure 11 and Figure 12. We apply the U-Net and Chan-Vese methods to the original and cropped liver slices. In previous figures the Chan-Vese model does not show any unwanted elements in the segmentation; however, the U-Net model presents unwanted elements. Figure 13, Figure 14, Figure 15 and Figure 16 present a segmentation comparison between the ground truth, the Chan-Vese approach, and the U-Net algorithm obtained from cropped slices 1–4.
The impact of cropping on liver segmentation was assessed using both the Chan-Vese and U-Net models.
When applied to original CT liver slices, the Chan-Vese model achieved an average dice score of 0.938, with an average processing time of 0.791 s per slice. After applying the cropping technique, the dice score improved to 0.960, and the processing time was reduced to 0.571 s per slice, resulting in an average speedup of 1.387 times. These results affirm that cropping helps to eliminate irrelevant data, focusing computations on the liver, which improves both accuracy and efficiency.
The U-Net model, however, initially showed a decrease in performance when tested on cropped slices, with the dice score dropping from 0.521 to 0.431. This decline was attributed to the resizing of cropped slices to match the input size required by U-Net, which impacted accuracy. Nonetheless, after applying post-processing to remove unwanted segmented areas, the dice score increased to 0.956. Although U-Net’s dice score is comparable to that of Chan-Vese after post-processing, the Chan-Vese model consistently outperformed U-Net in speed and segmentation accuracy on cropped images.
When the slices are cropped, the dice score (defined by Zhao et. al. [44]) (0.96 ± 0.005) and speedup are improved in the Chan-Vese model, as shown in Table 1. But the dice score is reduced in U-Net-based liver segmentation before post-processing (0.431 ± 0.04), as shown in Table 2. Dice score reduction due to cropping in U-Net in Table 2 is due to the fact that the slices are resized again to map the input of the U-Net network. U-Net is trained on original images and tested on original slices. The study examines U-Net trained on cropped images for testing cropped slices. When we perform post-processing after U-Net-based liver segmentation, this takes the maximum component in the image and filter out the other segmented regions. This results in a better dice score (0.956 ± 0.006) after post-processing, as shown in Table 2, and the boundaries are more confined to the liver using U-Net-based liver segmentation after post-processing, as shown in the figures dedicated to the cropped CT slice comparison. We do discuss speedups on U-Net, as it takes the fixed-size image as an input to the network.
The dice score is reduced when the input CT slice is cropped using U-Net-based liver segmentation, as shown in Table 2. But it is improved when we apply the post-processing to remove the unwanted segmentation. The Chan-Vese model for liver segmentation removes the unwanted segmentation completely, giving a better average dice score (0.96 ± 0.005) and speedup (1.387 ± 0.1), as shown in Table 1. The average time taken by Chan-Vese-based liver segmentation is 0.571 ± 0.038 s after cropping, providing an average speedup of 1.387 (±0.1) times in comparison to the original liver slice (0.791 ± 0.058 s on a liver image of size 512 × 512) on an NVIDIA GPU GeForce GTX 1050 with RAM 4 GB. The average dice score obtained by Chan-Vese (0.96 ± 0.005 in Table 1) is better compared to U-Net (0.956 ± 0.006 in Table 2) for liver segmentation in cropped images.
Finally, in all four cases, the U-Net model erroneously segmented adjacent soft tissues with similar intensity values, leading to significant false positives. However, post-processing was able to isolate the true liver region effectively. This highlights both the limitations of slice-wise DL inference and the value of morphology-aware corrections.

6.2. Liver Volume Assessment

In this section, we performed liver volume estimation using an SRG algorithm on the 3D binary masks. The algorithm operates within the cropped region and computes volume as the total number of labeled voxels multiplied by voxel size. Table 3 gives the analysis of semiautomatic 3D cropping showing the estimated liver volumes compared to ground truth. Results show strong agreement (mean absolute error below 2.5%) across both Chan-Vese and U-Net pipelines. The exact cropped size (ECS) is obtained by counting the number of voxels inside the cuboid, which is defined automatically from ground truths validated by clinicians. The average time to crop the liver slice volume is 8.841 ± 1.505 s. The average semiautomatic cropping time per slice is 11.603 ± 1.913 ms. The average percentage of maximum unnecessary data to be reduced is 92.031 ± 2.731% and semiautomatic cropping helps reduce 90.007 ± 3.406% (average) from the volume of liver slices.
We compare the performance on LiTS liver data [20] testing on an NVIDIA GPU GeForce GTX 1050. The voxel-based parallel approach improves cropping by 2.289 times compared to slice-based cropping proposed by Smistad et al. [30]. The cropped volume of liver (with voxels 3097775) is shown in Figure 17. The proposed parallel cropping algorithm crops liver volume of 512 × 512 × 816 to 285 × 229 × 190 in 44.6 ms.
The time required for cropping is reduced, as four kernel calls are involved (three for finding the cropping points in each dimension and one for copying cropped data) in the state of the art. The average time taken by the non-PT and slice-based parallel cropping is 103.529 ± 8.31 ms, whereas the proposed voxel-based parallel cropping takes 45.227 ± 4.9 ms, providing an average speedup of 2.289 times compared to the state of the art. The average time for the volume assessment of the liver is 0.781 ± 0.271 s.

6.3. Volume Assessment Evaluation of Other Datasets

To assess the generalizability of our cropping and volume assessment pipeline, we applied it to several cases from the 3DIRACDb dataset, including CT scans of skin, lung, and vascular structures. These datasets lack ground truth liver labels but serve as a testbed for efficiency and applicability across anatomical domains.
We assess the volume of the liver dataset from the LiTS challenge (Christ [20]). We analyze the proposal in vessel, liver, skin, and lung datasets from 3D Image Reconstruction for Comparison of Algorithm Database (3DIRACDb by Li et al. [43]). A total of 28 persistent blocks are used for SRG on an NVIDIA GPU RTX 2070. The proposed voxel-based cropping approach outperforms slice-based cropping for all the datasets, as shown in Figure 18. The speedups obtained by the voxel-based parallel cropping with respect to (w.r.t.) liver, vessel, skin, and lung datasets are 2.35, 1.89, 1.23, and 2.22 times, respectively, compared to the slice-based cropping.
In all cases, voxel-based cropping successfully localized target regions and enabled rapid volume estimation via SRG. As shown in Table 4, the runtime performance remained consistent, with average speedups of 2.1–2.5× over slice-based methods. Results from non-liver datasets indicate that our pipeline is adaptable to other organs, supporting its deployment in more diverse clinical imaging workflows.
The generalizability of the proposed cropping method was further supported by experiments on 3DIRACDb, which includes non-liver CT data such as vessels, lungs, and skin. Despite anatomical and intensity differences, our voxel-based cropping maintained consistent speedups and localization precision across all evaluated domains. In [46], the compatibility of preprocessing approaches with variable contrast regimes suggests that voxel-based cropping can be extended to other modalities, such as MRI or PET, especially when combined with contrast normalization techniques.

7. Conclusions

This study analyzes automatic and semiautomatic parallel cropping techniques used to segment and assess the volume of the liver in medical imaging. U-Net produces unwanted segmentation outside the liver’s RoI. However, this was effectively filtered, which increased the dice scores. The average percentage of unwanted data using semiautomatic cropping is 90 ± 3.4%. This implies that only 10% of data are useful from CT volumes. Semi-automatic cropping successfully eliminated irrelevant data. Our proposal improves the accuracy and significantly reduces unnecessary computations.
Applying our proposed method to U-Net for liver segmentation, the dice score impressively rises from 0.52 to 0.94, demonstrating a significant leap in accuracy and the potential to drastically improve clinical outcomes. The semi-automatic cropping technique demonstrated higher robustness when dealing with low-quality images with unclear liver boundaries. This approach ensures that cropping remains precise and efficient even when fully automated techniques fail to identify the correct RoI.
The proposed persistence-based approach was 2.29× faster than traditional slice-based cropping methods, because our voxel-based approach computes cropping boundaries in a single pass across all dimensions. Although the focus was on liver segmentation, the results indicate that this method can also be applied to other areas of medical imaging, such as the volumetric assessment of blood vessels, skin, and lungs.
DL models are popular but operate as black boxes. The decision-making process behind the segmentation results is not interpretable, making it difficult to understand why the model produces certain outputs, especially in edge cases or when anomalies are present. Clinical decisions require transparency and confidence. The lack of interpretability in DL models poses a barrier to clinical adoption, as practitioners might be hesitant to rely on a system that cannot provide clear justifications for its results. Furthermore, incorrect segmentations are difficult to diagnose, and it is difficult to address the cause of the error without a clear understanding of how the model processes the data. This lack of insight makes it challenging to fine-tune or adapt the model for different datasets or clinical environments, limiting its flexibility and robustness.
Our parallel cropping proposal is a robust and explainable alternative to deep learning solutions in scenarios where computational resources are limited, labeled data availability is scarce, and rapid integration and greater interpretability are essential in the clinical setting. This approach allows for efficient and accurate results without the high costs and complexity associated with DL models, making it a practical solution for real-time clinical applications.
Future work will explore adaptive architectures such as UNet++ [27], Swin-UNet [28], TransUNet [29], and other new models with dynamic input resolution. We also aim to further quantify uncertainty and integrate explainability metrics into the segmentation pipeline. We are also interested in implementing these algorithms in the emerging and promising Processing-In-Memory (PIM) paradigm [47].

Author Contributions

Conceptualization, N.S., J.G.-L. and J.O.; methodology, N.S., J.G.-L. and J.O.; software, N.S., N.B.G., S.K.K.; validation, N.S., J.G.-L. and J.O.; formal analysis, N.S., J.G.-L. and J.O.; investigation, N.S., J.G.-L. and J.O.; resources, J.O.; data curation, N.B.G. AND S.K.K.; writing—original draft preparation, N.S., N.B.G., S.K.K. and J.O.; writing—review and editing, J.O.; visualization, N.S.; supervision, J.O.; project administration, J.O.; funding acquisition, J.O. All authors have read and agreed to the published version of the manuscript.

Funding

The work is supported by the project High Performance soft tissue Navigation (HiPerNav). This project has been received from the European Union Horizon 2020 research and innovation program under grant agreement No. 722068.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

We thank The Intervention Centre, Oslo University Hospital, Norway for providing the CT images with ground truths for the clinical validation of the liver segmentation. Thanks to PPIT_2022E_026434 (DESIDERIO) DEsarrollo de SIstemas DEep-learning, Robótica e IOt, and, PP-RC-OPI-HIPERNAV IoT Next Generation, Towards 2030 Horizon from the Universidad de Cordoba.

Conflicts of Interest

Author Nitin Satpute was employed by the company ORACLE. Author Juan Gómez-Luna was employed by the company NVIDIA. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Cassinotto, C.; Denys, A.; Gay, F.; Duran, R.; Hocquelet, A.; Piron, L.; Guiu, B. Radiofrequency ablation of liver tumors: No difference in the ablation zone volume between cirrhotic and healthy liver. Cardiovasc. Interv. Radiol. 2018, 41, 905–911. [Google Scholar] [CrossRef] [PubMed]
  2. Sun, Y.; Zhang, N.; Ding, Y.L.; Yu, L.J.; Cai, J.; Ma, D.; Yang, W.; Lu, W.K.; Niu, J.L. Effect of lipid metabolism disorder on liver function in patients with malignant tumors after chemotherapy: A case-control study. Lipids Health Dis. 2019, 18, 1–7. [Google Scholar] [CrossRef] [PubMed]
  3. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed]
  4. Kufel, J.; Bargieł-Łączek, K.; Kocot, S.; Koźlik, M.; Bartnikowska, W.; Janik, M.; Czogalik, Ł.; Dudek, P.; Magiera, M.; Lis, A.; et al. What Is Machine Learning, Artificial Neural Networks and Deep Learning?—Examples of Practical Applications in Medicine. Diagnostics 2023, 13, 2582. [Google Scholar] [CrossRef]
  5. Ali, S.; Wang, X.; Kumar, A. Enhancing Recognition and Categorization of Skin Lesions with Tailored Deep Learning Architectures. Comput. Biol. Med. 2023, 155, 106693. [Google Scholar]
  6. Liu, M.; Zhao, L.; Tan, J. Optimized Deep Learning for Mammography: Augmentation Strategies and Transfer Learning for Improved Classification. Diagnostics 2023, 13, 352. [Google Scholar]
  7. Wei, D.; Jiang, Y.; Zhou, X.; Wu, D.; Feng, X. A Review of Advancements and Challenges in Liver Segmentation. J. Imaging 2024, 10, 202. [Google Scholar] [CrossRef]
  8. Chen, J.J.; Kutluk, A.; Hu, Y.T.; Hamit, M. Liver hydatid CT image segmentation based on localizing region active contours and modified parametric active contours. In Proceedings of the 2014 7th International Conference on Biomedical Engineering and Informatics, Dalian, China, 14–16 October 2014; pp. 217–221. [Google Scholar] [CrossRef]
  9. Satpute, N.; Naseem, R.; Pelanis, E.; Gomez-Luna, J.; Alaya Cheikh, F.; Elle, O.J.; Olivares, J. GPU acceleration of liver enhancement for tumor segmentation. Comput. Methods Programs Biomed. 2020, 184, 105285. [Google Scholar] [CrossRef]
  10. Satpute, N.; Naseem, R.; Palomar, R.; Zachariadis, O.; Gómez-Luna, J.; Cheikh, F.A.; Olivares, J. Fast parallel vessel segmentation. Comput. Methods Programs Biomed. 2020, 192, 105430. [Google Scholar] [CrossRef]
  11. Palomar, R.; Cheikh, F.A.; Edwin, B.; Fretland, Å.; Beghdadi, A.; Elle, O.J. A novel method for planning liver resections using deformable Bézier surfaces and distance maps. Comput. Methods Programs Biomed. 2017, 144, 135–145. [Google Scholar] [CrossRef]
  12. Zhou, Y.; Huang, W.; Dong, P.; Xia, Y.; Wang, S. D-UNet: A Dimension-Fusion U Shape Network for Chronic Stroke Lesion Segmentation. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 940–950. [Google Scholar] [CrossRef] [PubMed]
  13. Lu, X.; Xie, Q.; Zha, Y.; Wang, D. Fully automatic liver segmentation combining multi-dimensional graph cut with shape information in 3D CT images. Sci. Rep. 2018, 8, 10700. [Google Scholar] [CrossRef] [PubMed]
  14. Ronneberger, O.; P.Fischer.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; Lecture Notes in Computer Science. Springer: Berlin/Heidelberg, Germany, 2015; Volume 9351, pp. 234–241. [Google Scholar]
  15. Abdel-Nabi, H.; Ali, M.Z.; Awajan, A. A multi-scale 3-stacked-layer coned U-net framework for tumor segmentation in whole slide images. Biomed. Signal Process. Control 2023, 86, 105273. [Google Scholar] [CrossRef]
  16. Cohen, R. The Chan-Vese Algorithm. arXiv 2011, arXiv:1107.2782. [Google Scholar] [CrossRef]
  17. Satpute, N.; Gomez-Luna, J.; Olivares, J. Accelerating Chan–Vese model with cross-modality guided contrast enhancement for liver segmentation. Comput. Biol. Med. 2020, 124, 103930. [Google Scholar] [CrossRef]
  18. Nayantara, P.V.; Kamath, S.; Manjunath, K.; Rajagopal, K. Computer-aided diagnosis of liver lesions using CT images: A systematic review. Comput. Biol. Med. 2020, 127, 104035. [Google Scholar] [CrossRef]
  19. Chattopadhyay, S.; Dey, A.; Singh, P.K.; Sarkar, R. DRDA-Net: Dense residual dual-shuffle attention network for breast cancer classification using histopathological images. Comput. Biol. Med. 2022, 145, 105437. [Google Scholar] [CrossRef]
  20. Bilic, P.; Christ, P.; Li, H.B.; Vorontsov, E.; Ben-Cohen, A.; Kaissis, G.; Szeskin, A.; Jacobs, C.; Mamani, G.E.H.; Chartrand, G.; et al. The Liver Tumor Segmentation Benchmark (LiTS). Med. Image Anal. 2023, 84, 102680. [Google Scholar] [CrossRef]
  21. Jiang, L.; Ou, J.; Liu, R.; Zou, Y.; Xie, T.; Xiao, H.; Bai, T. RMAU-Net: Residual Multi-Scale Attention U-Net For liver and tumor segmentation in CT images. Comput. Biol. Med. 2023, 158, 106838. [Google Scholar] [CrossRef]
  22. Chen, Y.; Zheng, C.; Zhou, T.; Feng, L.; Liu, L.; Zeng, Q.; Wang, G. A deep residual attention-based U-Net with a biplane joint method for liver segmentation from CT scans. Comput. Biol. Med. 2023, 152, 106421. [Google Scholar] [CrossRef]
  23. Gul, S.; Khan, M.S.; Bibi, A.; Khandakar, A.; Ayari, M.A.; Chowdhury, M.E. Deep learning techniques for liver and liver tumor segmentation: A review. Comput. Biol. Med. 2022, 147, 105620. [Google Scholar] [CrossRef] [PubMed]
  24. Palomar, R.; Gomez-Luna, J.; Cheikh, F.A.; Olivares, J.; Elle, O.J. High-Performance Computation of Bézier Surfaces on Parallel and Heterogeneous Platforms. Int. J. Parallel Program. 2018, 46, 1035–1062. [Google Scholar] [CrossRef]
  25. Liu, P.; Zheng, G. CVCL: Context-aware Voxel-wise Contrastive Learning for label-efficient multi-organ segmentation. Comput. Biol. Med. 2023, 160, 106995. [Google Scholar] [CrossRef] [PubMed]
  26. Gupta, K.; Stuart, J.A.; Owens, J.D. A study of persistent threads style GPU programming for GPGPU workloads. In Proceedings of the Innovative Parallel Computing-Foundations & Applications of GPU, Manycore, and Heterogeneous Systems (INPAR 2012), San Jose, CA, USA, 13–14 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1–14. [Google Scholar]
  27. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support (DLMIA 2018, ML-CDS 2018), Granada, Spain, 20 September 2018; Lecture Notes in Computer Science. Springer: Berlin/Heidelberg, Germany, 2018; Volume 11045, pp. 3–11. [Google Scholar] [CrossRef]
  28. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, Q.; Tian, Q.; Wang, L. Swin-Unet: Unet-like pure transformer for medical image segmentation. arXiv 2021, arXiv:2105.05537. [Google Scholar]
  29. Chen, J.; Lu, Y.; Yu, Q.; Luo, T.; Adeli, E.; Wang, Y.; Lu, L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 66–75. [Google Scholar]
  30. Smistad, E.; Elster, A.C.; Lindseth, F. GPU accelerated segmentation and centerline extraction of tubular structures from medical images. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 561–575. [Google Scholar] [CrossRef]
  31. Lin, Y.; Wang, J.; Liu, Q.; Zhang, K.; Liu, M.; Wang, Y. CFANet: Context fusing attentional network for preoperative CT image segmentation in robotic surgery. Comput. Biol. Med. 2024, 171, 108115. [Google Scholar] [CrossRef]
  32. Yu, W.; Wang, M.; Zhang, Y.; Zhao, L. Reciprocal cross-modal guidance for liver lesion segmentation from multiple phases under incomplete overlap. Biomed. Signal Process. Control 2024, 88, 105561. [Google Scholar] [CrossRef]
  33. Ao, Y.; Shi, W.; Ji, B.; Miao, Y.; He, W.; Jiang, Z. MS-TCNet: An effective Transformer–CNN combined network using multi-scale feature learning for 3D medical image segmentation. Comput. Biol. Med. 2024, 170, 108057. [Google Scholar] [CrossRef]
  34. Smistad, E. Seeded Region Growing. 2015. Available online: https://github.com/smistad/FAST/tree/master/source/FAST/Algorithms/ (accessed on 22 June 2025).
  35. Tran, S.T.; Cheng, C.H.; Liu, D.G. A Multiple Layer U-Net, U Net n , for Liver and Liver Tumor Segmentation in CT. IEEE Access 2021, 9, 3752–3764. [Google Scholar] [CrossRef]
  36. Liu, L.; Wang, L.; Xu, D.; Zhang, H.; Sharma, A.; Tiwari, S.; Kaur, M.; Khurana, M.; Shah, M.A. CT image segmentation method of liver tumor based on artificial intelligence enabled medical imaging. Math. Probl. Eng. 2021, 2021, 9919507. [Google Scholar] [CrossRef]
  37. Li, C.; Tan, Y.; Chen, W.; Luo, X.; Gao, Y.; Jia, X.; Wang, Z. Attention Unet++: A Nested Attention-Aware U-Net for Liver CT Image Segmentation. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 345–349. [Google Scholar] [CrossRef]
  38. Fan, T.; Wang, G.; Li, Y.; Wang, H. MA-Net: A Multi-Scale Attention Network for Liver and Tumor Segmentation. IEEE Access 2020, 8, 179656–179665. [Google Scholar] [CrossRef]
  39. Nawandhar, A.; Kumar, N.; Yamujala, L. Gpu accelerated stratified squamous epithelium biopsy image segmentation for OSCC detector and classifier. Biomed. Signal Process. Control 2021, 64, 102258. [Google Scholar] [CrossRef]
  40. Chen, Z.; Dou, M.; Luo, X.; Yao, Y. Enhanced Liver and Tumor Segmentation Using a Self-Supervised Swin-Transformer-Based Framework with Multitask Learning and Attention Mechanisms. Appl. Sci. 2025, 15, 3985. [Google Scholar] [CrossRef]
  41. Idress, W.M.; Zhao, Y.; Abouda, K.A.; Yang, S. DRDA-Net: Deep Residual Dual-Attention Network with Multi-Scale Approach for Enhancing Liver and Tumor Segmentation from CT Images. Appl. Sci. 2025, 15, 2311. [Google Scholar] [CrossRef]
  42. Muhammad, S.; Zhang, J. Segmentation of Liver Tumors by Monai and PyTorch in CT Images with Deep Learning Techniques. Appl. Sci. 2024, 14, 5144. [Google Scholar] [CrossRef]
  43. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.; Heng, P. H-DenseUNet: Hybrid Densely Connected UNet for Liver and Tumor Segmentation From CT Volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef]
  44. Zhao, Y.; Li, H.; Wan, S.; Sekuboyina, A.; Hu, X.; Tetteh, G.; Piraud, M.; Menze, B. Knowledge-Aided Convolutional Neural Network for Small Organ Segmentation. IEEE J. Biomed. Health Inform. 2019, 23, 1363–1373. [Google Scholar] [CrossRef]
  45. Huang, Q.; Ding, H.; Wang, X.; Wang, G. Robust extraction for low-contrast liver tumors using modified adaptive likelihood estimation. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 1565–1578. [Google Scholar] [CrossRef]
  46. Naseem, R.; Khan, Z.A.; Satpute, N.; Beghdadi, A.; Cheikh, F.A.; Olivares, J. Cross-Modality Guided Contrast Enhancement for Improved Liver Tumor Image Segmentation. IEEE Access 2021, 9, 118154–118167. [Google Scholar] [CrossRef]
  47. Martínez, H.; Gómez-Luna, J.; Palomar, R.; Olivares, J. In-memory operators for medical image processing. Future Gener. Comput. Syst. 2025, 174, 107939. [Google Scholar] [CrossRef]
Figure 1. Proposed pipeline integrating voxel-level parallel cropping with Chan-Vese and U-Net liver segmentation, followed by volume assessment.
Figure 1. Proposed pipeline integrating voxel-level parallel cropping with Chan-Vese and U-Net liver segmentation, followed by volume assessment.
Applsci 15 07807 g001
Figure 2. Liver volume inside cuboid.
Figure 2. Liver volume inside cuboid.
Applsci 15 07807 g002
Figure 3. Liver segmentation: Chan-Vese (left) and U-Net (right).
Figure 3. Liver segmentation: Chan-Vese (left) and U-Net (right).
Applsci 15 07807 g003
Figure 4. Proposed parallel volume assessment.
Figure 4. Proposed parallel volume assessment.
Applsci 15 07807 g004
Figure 5. Original CT slice 1.
Figure 5. Original CT slice 1.
Applsci 15 07807 g005
Figure 6. Original CT slice 2.
Figure 6. Original CT slice 2.
Applsci 15 07807 g006
Figure 7. Original CT slice 3.
Figure 7. Original CT slice 3.
Applsci 15 07807 g007
Figure 8. Original CT slice 4.
Figure 8. Original CT slice 4.
Applsci 15 07807 g008
Figure 9. Cropped CT slice 1 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Figure 9. Cropped CT slice 1 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Applsci 15 07807 g009
Figure 10. Cropped CT slice 2 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Figure 10. Cropped CT slice 2 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Applsci 15 07807 g010
Figure 11. Cropped CT slice 3 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Figure 11. Cropped CT slice 3 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Applsci 15 07807 g011
Figure 12. Cropped CT slice 4 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Figure 12. Cropped CT slice 4 (Note: Cropped slices are resized for proper view, as U-Net provides fixed size output).
Applsci 15 07807 g012
Figure 13. Cropped CT slice 1.
Figure 13. Cropped CT slice 1.
Applsci 15 07807 g013
Figure 14. Cropped CT slice 2.
Figure 14. Cropped CT slice 2.
Applsci 15 07807 g014
Figure 15. Cropped CT slice 3.
Figure 15. Cropped CT slice 3.
Applsci 15 07807 g015
Figure 16. Cropped CT slice 4.
Figure 16. Cropped CT slice 4.
Applsci 15 07807 g016
Figure 17. Cropped liver volume (size 285 × 229 × 190) from CT image of size 512 × 512 × 816.
Figure 17. Cropped liver volume (size 285 × 229 × 190) from CT image of size 512 × 512 × 816.
Applsci 15 07807 g017
Figure 18. Speedup concerning non-PT and slice-based cropping.
Figure 18. Speedup concerning non-PT and slice-based cropping.
Applsci 15 07807 g018
Table 1. Liver segmentation using Chan-Vese.
Table 1. Liver segmentation using Chan-Vese.
Chan-Vese
Volume #
Dice ScoreTime in sec per SliceSpeedup w.r.t. Original
(512 × 512 × n)OriginalCroppingOriginalCropping
1 (n = 128)0.9450.9610.810.571.421
2 (n = 171)0.9380.9670.880.561.571
3 (n = 199)0.9320.9580.820.591.39
4 (n = 89)0.9280.9510.730.511.431
5 (n = 138)0.9390.9620.710.551.291
6 (n = 134)0.9420.9580.810.631.286
7 (n = 150)0.9410.9610.780.591.322
Average0.9380.960.7910.5711.387
Std. Dev.0.0060.0050.0580.0380.1
Table 2. Liver segmentation using U-Net.
Table 2. Liver segmentation using U-Net.
U-NetDice ScoreDice After Post-Processing
Volume #
(512 × 512 × n)
OriginalCroppingOriginalCropping
1 (n = 128)0.5610.4890.9320.951
2 (n = 171)0.5120.4510.9360.953
3 (n = 199)0.5210.4420.9340.956
4 (n = 89)0.5310.4510.9410.957
5 (n = 138)0.5370.4230.9510.967
6 (n = 134)0.4920.3710.9410.954
7 (n = 150)0.4940.3890.9430.951
Average0.5210.4310.940.956
Std. Dev.0.0250.040.0060.006
Table 3. Analysis of semiautomatic 3D cropping.
Table 3. Analysis of semiautomatic 3D cropping.
SizeECSSCSnSCT (s)SCTPS (ms)% of MUD% of UDR
512 × 512 × 816285 × 229 × 190305 × 260 × 2308167.218.8494.20391.473
512 × 512 × 629264 × 242 × 227297 × 250 × 2406298.7813.9691.20589.193
512 × 512 × 671326 × 290 × 233337 × 313 × 2536716.8910.2787.47784.828
512 × 512 × 722305 × 257 × 300317 × 281 × 3217227.9811.0587.57684.892
512 × 512 × 841344 × 203 × 264381 × 221 × 3248419.8111.6691.63887.626
512 × 512 × 771290 × 228 × 249323 × 242 × 2607717.549.7891.85489.945
512 × 512 × 836308 × 271 × 180320 × 279 × 19783611.3413.5693.14491.974
512 × 512 × 751253 × 262 × 190271 × 269 × 20875110.8714.4793.60392.298
512 × 512 × 756284 × 220 × 195297 × 241 × 2037569.2412.2293.85292.668
512 × 512 × 856186 × 252 × 203191 × 271 × 2098568.7510.2295.76095.179
Average8.8411.6092.03190.007
Standard Deviation1.5051.9132.7313.406
Size = original size of liver volume; ECS = exact cropped size; SCS = semiautomatic cropped size; n = number of original slices; SCT = semiautomatic cropping time; SCTPS= semiautomatic cropping time per slice; % of MUD (% of maximum unnecessary data) = (1 − (ECS/size)) × 100; % of UDR (% of unnecessary data reduced) = (1 − (SCS/size)) × 100.
Table 4. Performance comparison between GPU approaches for LiTS liver dataset [20,45].
Table 4. Performance comparison between GPU approaches for LiTS liver dataset [20,45].
Liver Volume AssessmentTime (ms) for CroppingTime (s)
Size (Grown Voxels)Cropped SizeNon PT and SlicePT and VoxelVol. Asses.
512 × 512 × 816
(3,097,775)
285 × 229 × 190113.544.60.55
512 × 512 × 629
(3,681,197)
264 × 242 × 22784.8736.650.63
512 × 512 × 671
(4,588,020)
326 × 290 × 23397.0255.011.08
512 × 512 × 722
(5,710,514)
305 × 257 × 300102.4846.861.28
512 × 512 × 841
(4,068,110)
344 × 203 × 264111.8047.870.98
512 × 512 × 771
(3,140,419)
290 × 228 × 24999.1048.510.96
512 × 512 × 836
(3,373,371)
308 × 271 × 180106.3145.100.68
512 × 512 × 751
(2,781,767)
253 × 262 × 190108.1140.980.55
512 × 512 × 756
(3,120,233)
284 × 220 × 195105.3042.670.61
512 × 512 × 856
(1,092,424)
186 × 252 × 203106.8044.020.49
Average103.5345.230.78
Std. Dev.8.314.90.27
Speedup12.29×-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Satpute, N.; Gaikwad, N.B.; Khare, S.K.; Gómez-Luna, J.; Olivares, J. Explainable Liver Segmentation and Volume Assessment Using Parallel Cropping. Appl. Sci. 2025, 15, 7807. https://doi.org/10.3390/app15147807

AMA Style

Satpute N, Gaikwad NB, Khare SK, Gómez-Luna J, Olivares J. Explainable Liver Segmentation and Volume Assessment Using Parallel Cropping. Applied Sciences. 2025; 15(14):7807. https://doi.org/10.3390/app15147807

Chicago/Turabian Style

Satpute, Nitin, Nikhil B. Gaikwad, Smith K. Khare, Juan Gómez-Luna, and Joaquín Olivares. 2025. "Explainable Liver Segmentation and Volume Assessment Using Parallel Cropping" Applied Sciences 15, no. 14: 7807. https://doi.org/10.3390/app15147807

APA Style

Satpute, N., Gaikwad, N. B., Khare, S. K., Gómez-Luna, J., & Olivares, J. (2025). Explainable Liver Segmentation and Volume Assessment Using Parallel Cropping. Applied Sciences, 15(14), 7807. https://doi.org/10.3390/app15147807

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop