Next Article in Journal
Extracellular Vesicles in Cardiac Amyloidosis: From Pathogenesis to Clinical Applications
Previous Article in Journal
Clinical Profile of a Series of Left-Sided Prosthetic Valve Endocarditis: Revisiting Surgical Indications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Liver Tumor Segmentation from Computed Tomography Scans with a Gradient-Enhanced Network

1
School of Applied Artificial Intelligence and Entrepreneurship, Handong Global University, Pohang 37554, Republic of Korea
2
Graduate School of Artificial Intelligence, Pohang University of Science and Technology, Pohang 37673, Republic of Korea
3
Graduate School of Artificial Intelligence, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
4
Department of Surgery, Jeonbuk National University Medical School and Hospital, Jeonju 54907, Republic of Korea
5
Research Institute of Clinical Medicine, Jeonbuk National University, Jeonju 54907, Republic of Korea
6
Biomedical Research Institute, Jeonbuk National University Hospital, Jeonju 54907, Republic of Korea
7
Department of Radiology, Jeonbuk National University Medical School and Hospital, Jeonju 54907, Republic of Korea
8
Department of Industrial Management and Engineering, Pohang University of Science and Technology, Pohang 37673, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2026, 16(3), 429; https://doi.org/10.3390/diagnostics16030429 (registering DOI)
Submission received: 5 December 2025 / Revised: 20 January 2026 / Accepted: 24 January 2026 / Published: 1 February 2026
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

Background/Objectives: This study aimed to develop a fully automatic method for liver tumor segmentation based on our previously developed gradient-enhanced network G-UNETR++. Methods: The proposed method consists of segmentation of the full liver region from computed tomography (CT) images using G-UNETR++, masking the CT images with the extracted liver region to exclude non-liver regions, and liver tumor segmentation from the masked CT images, also using G-UNETR++. To train and evaluate the model, a total of 131 CT scans (97 for training, 20 for validation, and 20 for testing) from the publicly available LiTS dataset were used. Furthermore, another public dataset, the 3DIRCADb dataset consisting of 20 CT scans was used for cross-validation of the effectiveness and generalizability of our method. Results: Experimental results showed that our method outperformed state-of-the-art models over both the LiTS dataset and the 3DIRCADb dataset, with an average dice score of 0.844 and 0.832 over the two datasets, respectively. Conclusions: The proposed method is effective in clinical application to help physicians with liver tumor diagnosis and treatment.

1. Introduction

Liver cancer ranks as the fourth highest cause of death among all malignancies [1], making its diagnosis and surgical intervention critical tasks in the modern medical field. To enhance the success rates of these procedures, accurate segmentation of both the liver and tumors is essential. Computed tomography (CT) has emerged as a representative imaging modality, widely utilized by physicians for analyzing organs and detecting lesions. In clinical practice, physicians typically detect tumors from a patient’s CT scan through visual observation, relying on their knowledge and experience. This approach is subjective and time-demanding, potentially leading to an increased possibility of misdiagnosis or missed diagnosis [2]. Therefore, automatic and accurate liver tumor segmentation from CT scans is needed to assist physicians in detecting liver tumors efficiently and accurately.
With the advancement of deep learning technologies, numerous methods have been proposed for medical image segmentation tasks. Since the introduction of convolutional neural networks (CNNs) [3], significant progress has been made in the field of computer vision, leading to the development of various CNN-based architectures [4,5,6,7,8,9,10,11,12,13,14,15,16,17]. U-Net [4] has especially become one of the most popular network architectures for medical image analyses tasks. For example, Özcan et al. [17] proposed a hybrid model incorporating U-Net and inception models for automatic segmentation of the liver and liver tumors from CT scans. They reported that their model achieved 75.6% and 65.6% of dice similarity coefficient (DSC) over two public datasets, respectively. CNN-based models are able to capture local features but struggle to capture global contexts from medical images due to their localized receptive fields [18].
To address these issues, we aim to develop a two-step method for liver tumor segmentation based on a gradient-enhanced network, referred to as G-UNETR++ [19]. G-UNETR++ is a hybrid network incorporating U-Net, vision transformer (ViT) [20], and gradient-based encoders, originally designed for liver segmentation. ViT is able to capture global contexts by its self-attention mechanisms. Gradient-based encoders further enable G-UNETR++ to learn 3D boundary features of the organs and tissues from medical images. First, the liver is segmented from CT images with G-UNETR++ and the CT images are masked by the extracted liver regions to remove non-liver regions. Then, the regions of liver tumors are delineated with G-UNETR++.

2. Related Work

2.1. CNN-Based Segmentation Networks

CNN-based architectures have been widely adopted for medical image segmentation, with many approaches derived from U-Net [5,6,10,11,12,13]. Milletari et al. [10] proposed a fully convolutional network, optimized using a dice-based loss for volumetric segmentation. Çiçek et al. [5] extended U-Net to 3D by replacing 2D operations with 3D convolutions, enabling effective volumetric analysis. Dou et al. [11] introduced a deeply supervised 3D network that improved liver segmentation performance on CT volumes. To enhance boundary delineation, Roth et al. [13] proposed a two-stage FCN, achieving higher dice scores in multi-organ CT segmentation. Despite these advances, CNN-based models are limited in capturing long-range dependencies and global contextual information.

2.2. Transformer-Based and Hybrid Segmentation Networks

Dosovitskiy et al. [20] introduced attention mechanisms to computer vision, motivating transformer-based medical image segmentation models. Pure transformer architectures without convolution have been proposed for both 2D and 3D segmentation tasks [21,22]. Cao et al. [22] designed a U-Net-like transformer with shifted window attention for encoder–decoder learning, achieving strong results on 2D medical datasets. Karimi et al. [21] introduced a convolution-free 3D model that applies self-attention across neighboring volumetric patches.
To combine local feature extraction with global context modeling, several studies proposed hybrid CNN–transformer architectures [23,24,25,26,27,28,29,30,31]. Oktay et al. [23] enhanced U-Net with attention gates in skip connections, improving abdominal multi-organ segmentation. Valanarasu et al. [24] proposed axial attention for more efficient positional encoding. TransUNet [30] employed a CNN-based encoder followed by transformer blocks and a CNN decoder for precise localization. Similarly, Xie et al. [31] combined CNN feature extraction with transformer-based contextual modeling for 3D segmentation. Hatamizadeh et al. [25] eliminated convolutional encoders and directly tokenized image patches using a transformer, while retaining a skip-connected decoder for multi-scale prediction. Zhou et al. [26] interleaved convolution and self-attention to learn volumetric representations and achieved superior liver segmentation performance on the Synapse dataset compared to UNETR. Shaker et al. [27] further introduced an efficient paired attention block that jointly models spatial and channel dependencies, improving segmentation accuracy and computational efficiency; however, its liver segmentation performance remained slightly lower than nnFormer, and tumor segmentation results were not reported.

2.3. Liver and Tumor Segmentation Networks

Several segmentation networks have been specifically designed for liver and tumor segmentation [2,32,33,34,35,36,37,38]. Jin et al. [32] replaced standard U-Net convolutional blocks with residual blocks to enhance feature learning and mitigate gradient vanishing, while integrating multi-scale attention to fuse hierarchical features. Chen et al. [2] proposed MS-FANet, incorporating residual attention blocks and multi-scale atrous downsampling to better capture tumor size and shape variations, achieving dice scores of 0.742 and 0.780 on two public datasets. Jiang et al. [34] introduced RMAU-Net, which integrates squeeze-and-excitation mechanisms and multi-scale feature fusion to model both inter-channel and spatial relationships, achieving dice scores of 0.762 and 0.831 on the same datasets. Muhammad and Zhang [35] leveraged the hybrid ResUNet model, a combination of both the ResNet and UNet models developed by the Monai 0.6 and PyTorch 1.10 frameworks and achieved a dice score of 0.87 for liver tumor segmentation over the public MSD Task03 Liver dataset. Yashaswini et al. [36] applied U-Net and a modified ResUNet to liver and tumor segmentation from CT scans using the 3DIRCADb dataset, demonstrating the effectiveness of fully convolutional networks for semantic segmentation. The ResUNet achieved superior performance, with dice scores of 91.44% for liver segmentation and 75.84% for tumor segmentation. Balaguer-Montero et al. [37] developed a fully automated liver tumor detection and segmentation system based on nnU-Net and achieved a dice score of 81.72% at lesion level. Zhang et al. [38] introduced a novel liver tumor segmentation framework that combines deformable attention, global context modeling, and dual cross-scale feature fusion to handle complex and irregular tumor structures. Their model achieved a dice score of 81.33% on their internal test set. Despite these efforts, accurate tumor segmentation remains challenging due to the heterogeneous appearance and complex morphology of liver tumors.

3. Materials and Methods

3.1. Data Preparation

Two public CT datasets, including the MICCAI 2017 LiTS dataset (LiTS) [39] and the 3DIRCADb dataset [40], were used in this study. The LiTS dataset is a public dataset from the liver tumor segmentation challenge held at ISBI 2017 and MICCAI 2017. It is the most widely used dataset for liver and tumor segmentation research. The LiTS dataset contains patients with diverse types of liver tumor diseases, consisting of primary tumor disease, such as hepatocellular carcinoma and cholangiocarcinoma, and secondary liver tumors, such as metastases from colorectal, breast, and lung primary cancers. The LiTS dataset, primarily focusing on portal venous phase CT scans, comprises a training set of 131 CT scans and a separate test set of 70 CT scans. The number of CT slices in each scan varies from 42 to 1026, with an axial resolution of 512 × 512 pixels, in-plane voxel dimensions of 0.55 to 1.0 mm and a slice thickness ranging from 0.45 to 6.0 mm. The training dataset was manually labeled by four radiologists from six clinical sites worldwide, whereas labels of the test set are not publicly available. It is challenging to segment tumors using this dataset because of the significant variations in slice thickness, scan image storage direction, image quality, and spatial resolution.
The 3DIRCADb dataset is another publicly available dataset that provides more complex data on the liver and tumors. The 3DIRCADb-01 dataset consists of enhanced CT scans of 10 females and 10 males, with 75% of cases having hepatic tumors, while the 3DIRCADb-02 dataset comprises two 3D CT scans. The voxel dimensions of the dataset are [0.56–0.87, 0.56–0.87, 1.6–4.0] mm, with an axial resolution of 512 × 512 pixels and the number of slices in each scan varying between 74 and 260. In some cases, the liver and tumors have low contrast and overlapping regions, which makes the tumor segmentation a challenging task. Labels of the dataset are publicly available.
In this study, the 131 CT scans from the LiTS dataset were randomly split into 97 cases for training, 20 for validation, and 20 for testing. Next, the 3DIRCADb-01 dataset was used for cross-validation of the effectiveness and generalizability of our method.

3.2. Data Preprocessing and Augmentation

Preprocessing was performed to normalize the Hounsfield Unit values of all CT scans to a range of −250 to 250 to enhance the visibility of the liver and tumor regions, as shown in Figure 1. All CT scans were resized from 512 × 512 to 256 × 256 to reduce computational cost. The same data augmentation strategies as G-UNERT++ [19], such as random rotation of 90, 180, and 270 degrees, random scaling, random mirroring, and random intensity shifting were applied.

3.3. Deep Learning Model Preparation and Training

Figure 2 shows the overall pipeline of liver tumor segmentation based on our G-UNETR++ model. First, the liver is segmented from CT images using G-UNETR++. Then the extracted liver region is multiplied with the original CT images to exclude non-liver regions. Lastly, liver tumors are delineated from the liver-only CT images using G-UNETR++. Figure 3 shows the architecture of our G-UNETR++ model. The model consists of a hierarchical encoder–decoder structure with skip connections between the encoder and the decoder, efficient paired attention (EPA) blocks, and convolutional blocks to obtain segmentation results. The encoder scheme consists of three 4-stage encoders in parallel one with CT scan f ( x , y , z ) as input, and the other two with second-order partial derivatives 2 f ( x , y , z ) x z and 2 f ( x , y , z ) y z as their inputs to learn 3D geometric features such as the boundaries between different organs and tissues especially along the z-axis. In the first stage, patch embedding is performed, where the input volume f R H × W × D is divided into non-overlapping 3D patches f p R N × ( P 1 ,   P 2 ,   P 3 ) of size (P1, P2, P3), producing a sequence with length N = ( H P 1 × W P 2 × D P 3 ) . These patches are projected into C 1 channels to form a H P 1 × W P 2 × D P 3 × C 1 feature map, using a patch resolution of (4, 4, 2), followed by an EPA block. The remaining stages downsample the feature maps by a factor of two using non-overlapping convolutions, each followed by an EPA block. At every stage, features from the three encoders are fused via element-wise summation. The EPA block employs spatial and channel attention with shared keys and queries and separate value projections to jointly model spatial–channel features [27]. The decoder consists of four stages with deconvolution-based upsampling to progressively increase resolution. EPA blocks are used at all but the final stage. Skip connections fuse encoder and decoder features at each scale to recover spatial information, while channel dimensions are halved between stages. In the final stage, fused features are passed through 3 × 3 × 3 and 1 × 1 × 1 convolutions to generate voxel-wise segmentation outputs.

3.4. Loss Function

We proposed different hybrid loss functions for liver segmentation and tumor segmentation. For liver segmentation, a hybrid loss function L l i v e r was proposed, consisting of dice loss L d i c e , cross-entropy (CE) loss L C E , and Hausdorff distance (HD) loss [41] L H D . For tumor segmentation, a hybrid loss function L t u m o r was proposed, consisting of dice loss, focal loss [42] L f o c a l , and HD loss to address the class imbalance issue in tumor segmentation. The hybrid loss function for liver segmentation is defined as follows:
L l i v e r = α 1 L d i c e + β 1 L C E + γ 1 L H D
where α 1 = β 1 = γ 1 = 1 . The hybrid loss function for tumor segmentation is defined as follows:
L t u m o r = α 2 L d i c e + β 2 L f o c a l + γ 2 L H D
where α 2 = β 2 = γ 2 = 1 . Dice loss is defined as follows:
L D i c e =   1   2   i = 1 N p i y i i = 1 N p i 2 + i = 1 N y i 2
where N denotes the number of predicted voxels; p i indicates the prediction probability at voxel i ; and y i represents the ground truth at voxel i . CE loss is defined as follows:
L C E y ,   p = log p ,     y = 1 log ( 1 p ) ,     y = 0
where y indicates the ground truth and p denotes the prediction probability. HD loss is defined as follows:
L H D =   max x X   min y Y x y 2
where x X denotes the predicted binary segmentation with a threshold of 0.5 and y Y indicates the ground truth. Focal loss is defined as follows:
L f o c a l ( p t ) = a t ( 1 p t ) γ l o g ( p t )
where p t denotes the predicted probability for the correct class; a t = 1 indicates a weighting factor for class imbalance; and γ = 2 denotes the focusing parameter, controlling the rate at which easy examples are downweighted. A deep supervision technique [43] was applied into our decoder for better training efficiency.

3.5. Model Training

Identical settings were applied to train both the liver segmentation model and the tumor segmentation model. All experiments were performed with an NVIDIA RTX 3090 GPU. The initial learning rate was set to 5 × 10−4 with a poly decay strategy [26]. The Adam optimizer [44] was used with a weight decay of 3 × 10−5. The number of epochs was set as 1000. The batch size was set as 8.

3.6. Post-Processing

By visual inspection, we observed that our model tended to segment tumors with fewer voxels compared to the ground truth, even though the model accurately predicted the locations of the tumors. This undersizing of small tumors became more severe as the size of the tumor decreased in the ground truth. To address this issue, we empirically applied the morphological dilation method twice, with a pixel size of one in each operation to the predicted tumor regions smaller than 100 pixels from a CT slice after exhaustive experimentation.

3.7. Evaluation Metrics

To assess the performance of the proposed tumor segmentation method, we utilized a range of evaluation metrics, including the dice similarity coefficient (DSC) to measure overlap accuracy, volumetric overlap error (VOE) to measure volume consistency, relative absolute volume difference (RAVD) to assess volume disparity, and average symmetric surface distance (ASSD) to quantify surface deviation.
DSC quantifies the overlap between the predicted volumetric output (Pred) and the ground truth (GT). Ranging from 0 to 1, DSC = 1 indicates perfect overlap and segmentation, while DSC = 0 signifies no overlap between the predicted and the ground truth. The formula for DSC is as follows:
D S C = 2 | P r e d     G T | | P r e d | +   | G T |
VOE quantifies the alignment between the predicted and the ground truth, assessing the error rate in segmentation. Ranging from 0 to 1, VOE = 0 indicates perfect overlap, while VOE = 1 signifies no overlap. The formula for VOE is as follows:
V O E = 1   |   P r e d     G T   | |   P r e d   G T   |
RAVD is used to measure the volume discrepancy between the predicted and the ground truth, ranging from 0 to 1, with a value of zero indicating no volume disparity, reflecting a perfect segmentation. RAVD is calculated as follows:
R A V D = | |   P r e d   |     |   G T   | |   G T   |   |
ASSD is the key metric for measuring the average deviation between the surfaces of the predicted and the ground truth, with a value of zero indicating a perfect segmentation. The shortest distance of a voxel x to the set of surface voxels S ( G T ) of G T is defined as follows:
d ( x , S ( G T ) ) = min y S ( GT ) x y
where x y denotes the Euclidean distance between x and y . Then ASSD is defined as follows:
A S S D = 1 | S ( P r e d ) | + | S ( G T ) | × ( x S ( P r e d ) d ( x ,   S ( G T ) ) + y S ( G T ) d ( y , S ( P r e d ) ) )

4. Results

We compared our model with state-of-the-art models for liver tumor segmentation, including HDU-Net [45], ResUNet [46], MS-FANet [2], HFRU-Net [47], and RMAU-Net [34]. Table 1 shows the comparison results for liver tumor segmentation over the LiTS and 3DIRCADb datasets. For both datasets, our model outperformed state-of-the-art models. For the LiTS dataset, our model showed a significant improvement in liver tumor segmentation compared with the other models. Our model achieved the best performance in terms of DSC (0.844), VOE (0.263), and ASSD (1.317 mm), with RAVD being the only exception, which indicates the high effectiveness of our model in accurately capturing tumor regions. Next, for the 3DIRCADb dataset, our model also achieved the best performance in terms of DSC (0.832) and ASSD (1.682), except for VOE and RAVD. The consistently high performance of our model across different datasets highlights its robustness and generalizability.
Figure 4 and Figure 5 illustrate some examples of the liver tumor segmentation results with our model from the LiTS dataset and the 3DIRCADb dataset, respectively. Visual inspection of the segmentation results reveals that the predicted tumor regions are properly aligned with the ground truth in terms of both the number of tumors and their morphology.
We conducted an ablation study to demonstrate the effectiveness of the proposed post-processing method. First, we evaluated the post-processing method on the LiTS and 3DIRCADb datasets. As shown in Table 2, for the LiTS dataset, there is a slight performance difference between the results with post-processing and those without post-processing, whereas for the 3DIRCADb dataset, the dice score has been improved by 2.9% and all other metrics have also been improved. Therefore, the proposed post-processing method is effective in improving tumor segmentation results.
To determine the right number of times of the morphological dilation operation in our post-processing method, we applied the morphological dilation operation from one to three times to compare their performance. As shown in Table 3, for the LiTS dataset, the best performance was achieved with the one-time application of the dilation operation, but the differences with the two-time application and the three-time application were small, whereas for the 3DIRCADb dataset, the results showed that applying dilation twice resulted in the best performance in terms of all evaluation metrics. Therefore, the post-processing method that applies the morphological dilation twice was adopted for our liver tumor segmentation method. As shown in Figure 6, after post-processing, the dilated tumor boundary is closer to the ground truth tumor boundary.
We conducted various analyses to further evaluate our model. Figure 7 shows the plotting of the training loss and validation loss, indicating that no overfitting occurred during our model development process. Table 4 summarizes the complexity of our model. Figure 8 presents the precision-recall plotting of our model over the LiTS dataset and the 3DIRACDb dataset, respectively.

5. Discussion

This study proposed a solid deep learning method for liver tumor segmentation based on G-UNETR++, a gradient-enhanced network originally developed in our previous study for liver segmentation from CT images. The proposed method consists of two steps: (1) segmentation of the liver from a CT scan by G-UNETR++ and masking the CT scan with the extracted liver region to exclude non-liver regions, and (2) segmentation of tumors from the masked CT scan by G-UNETR++.
Performance comparisons on the LiTS dataset show that our method outperforms existing state-of-the-art methods, achieving an average DSC of 0.844. To assess generalizability, the model was evaluated on one unseen dataset, 3DIRACDb, where it achieved an average DSC of 0.832. These results demonstrate strong robustness across datasets acquired under diverse conditions. The inclusion of two gradient-enhanced encoders ensures that our model effectively captures 3D geometric features. Furthermore, the proposed hybrid loss function can handle the class imbalance issue with focal loss and deal with difficult cases in tumor segmentation to ensure boundary precision with HD loss. Lastly, the post-processing method reduces tumor segmentation error by morphological dilation operation, especially when the tumor size is small.
Nevertheless, our study has limitations in certain aspects. The introduction of the two gradient-enhanced encoders increases the complexity level of our model and therefore the computational cost. Furthermore, the tumor segmentation performance can be affected by the liver segmentation performance. The proposed method achieved high performance on both the LiTS dataset and the 3DIRCADb dataset, with an average DSC of 97.38% and 97.50% for liver segmentation, respectively. However, in some cases, tumors may be missing in the segmented liver region, as shown in Figure 9c. Then, in the masking step, those tumors will be excluded from the CT images masked by the extracted liver region, which will be used as the input for tumor segmentation. In that case, those tumors will be missing in tumor segmentation, as shown in Figure 9d. In addition, our model may occasionally miss segmenting some extremely small tumors, as shown in Figure 10. This could be caused by tumor size imbalance and biased annotation for small lesions in the public datasets used for our model development. For our future work, the proposed two-step liver tumor segmentation pipeline can be reduced to a single-step pipeline by applying an end-to-end framework. This would enable a direct extraction of tumors from the original CT scans, thereby reducing computational costs and potentially enhancing the robustness of the segmentation process. Furthermore, the sizes of the datasets used for model training and evaluation are relatively small. To strengthen the experimental validity, the performance of the proposed model needs to be further evaluated using k-fold cross-validation or a larger dataset. Lastly, the ablation study is limited in scope and primarily focuses on the effect of post-processing. Though the ablation study in our previous work [19] concluded that the inclusion of the proposed gradient-enhanced encoders and a hybrid loss function that incorporates the HD loss are effective in improving the performance of liver segmentation, their effects on liver tumor segmentation were not studied in the current study. For future work, a comprehensive ablation study that analyzes the impacts of the gradient-enhanced encoders, the proposed hybrid loss function, and the EPA blocks on improving liver tumor segmentation performance will be conducted.

6. Conclusions

This study presents a deep learning framework for liver tumor segmentation from CT images using a two-step pipeline consisting of liver extraction followed by tumor segmentation. The proposed method achieves competitive performance, outperforming existing approaches on the LiTS dataset with a DSC of 0.844, and demonstrating strong generalizability on the unseen 3DIRACDb dataset with a DSC of 0.832.
The incorporation of gradient-enhanced encoders enables the effective learning of 3D geometric features, while the hybrid loss function addresses class imbalance and improves boundary delineation. In addition, post-processing with morphological dilation reduces segmentation errors, particularly for small tumors. Despite these advantages, the model’s complexity and reliance on accurate liver segmentation remain limitations, and very small tumors may still be missed.
Future work will focus on developing an end-to-end, single-stage framework to reduce computational cost and improve robustness, as well as validating the method using k-fold cross-validation or larger datasets with comprehensive ablation studies.
The proposed method shows strong clinical potential by accurately segmenting liver tumors from CT scans. The generated segmentation results allow radiologists to efficiently review tumor regions and assess imaging characteristics for diagnosis, supporting informed treatment planning by physicians.

Author Contributions

Conceptualization, X.Y.; methodology, H.S., K.H., S.L., and X.Y.; software, H.S., K.H., and S.L.; validation, H.S., K.H., S.L., S.K., and X.Y.; formal analysis, H.S., K.H., and S.L.; investigation, H.S., K.H., S.L., S.K., and X.Y.; resources, X.Y. and J.D.Y.; data curation, H.S., K.H., S.L., H.P., S.K., and J.S.; writing—original draft, H.S., K.H., S.L., H.P., S.K., and J.K.; writing—review and editing, X.Y., J.D.Y., H.C.Y., and H.Y.; visualization, H.S., K.H., S.L., S.K., J.K., and X.Y.; supervision, X.Y. and J.D.Y.; project administration, X.Y. and J.D.Y.; funding acquisition, X.Y. and J.D.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by No. of 202500500001 of Handong Global University Research Grants and research funds for newly appointed professors of Jeonbuk National University in 2024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ahn, S.H.; Yeo, A.U.; Kim, K.H.; Kim, C.; Goh, Y.; Cho, S.; Lee, S.B.; Lim, Y.K.; Kim, H.; Shin, D.; et al. Comparative clinical evaluation of atlas and deep-learning-based auto-segmentation of organ structures in liver cancer. Radiat. Oncol. 2019, 14, 213. [Google Scholar]
  2. Chen, Y.; Zheng, C.; Zhang, W.; Lin, H.; Chen, W.; Zhang, G.; Xu, G.; Wu, F. MS-FANet: Multi-scale feature attention network for liver tumor segmentation. Comput. Biol. Med. 2023, 163, 107208. [Google Scholar] [CrossRef] [PubMed]
  3. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  4. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  5. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016, Athens, Greece, 17–21 October 2016. [Google Scholar]
  6. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Granada, Spain, 20 September 2018; pp. 3–11. [Google Scholar]
  7. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 2019, 39, 1856–1867. [Google Scholar] [CrossRef] [PubMed]
  8. Cai, S.; Tian, Y.; Lui, H.; Zeng, H.; Wu, Y.; Chen, G. Dense-UNet: A novel multiphoton in vivo cellular image segmentation model based on a convolutional neural network. Quant. Imaging Med. Surg. 2020, 10, 1275. [Google Scholar] [CrossRef]
  9. Kaur, A.; Kaur, L.; Singh, A. GA-UNet: UNet-based framework for segmentation of 2D and 3D medical images applicable on heterogeneous datasets. Neural Comput. Appl. 2021, 33, 14991–15025. [Google Scholar]
  10. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  11. Dou, Q.; Chen, H.; Jin, Y.; Yu, L.; Qin, J.; Heng, P.A. 3D Deeply Supervised Network for Automatic Liver Segmentation from CT Volumes. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016, Athens, Greece, 17–21 October 2016. [Google Scholar]
  12. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.-W.; Heng, P.-A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar]
  13. Roth, H.R.; Oda, H.; Hayashi, Y.; Oda, M.; Shimizu, N.; Fujiwara, M.; Misawa, K.; Mori, K. Hierarchical 3D fully convolutional networks for multi-organ segmentation. arXiv 2017, arXiv:1704.06382. [Google Scholar] [CrossRef]
  14. Christ, P.F.; Ettlinger, F.; Grün, F.; Elshaera, M.E.A.; Lipkova, J.; Schlecht, S.; Ahmaddy, F.; Tatavarty, S.; Bickel, M.; Bilic, P.; et al. Automatic liver and tumor segmentation of CT and MRI volumes using cascaded fully convolutional neural networks. arXiv 2017, arXiv:1702.05970. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Pan, X.; Li, C.; Wu, T. 3D liver and tumor segmentation with CNNs based on region and distance metrics. Appl. Sci. 2020, 10, 3794. [Google Scholar] [CrossRef]
  16. Zhao, Z.; Ma, Z.; Liu, Y.; Zeng, Z.; Chow, P.K. Multi-Slice Dense-Sparse Learning for Efficient Liver and Tumor Segmentation. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Virtual, 1–5 November 2021. [Google Scholar]
  17. Özcan, F.; Uçan, O.N.; Karaçam, S.; Tunçman, D. Fully automatic liver and tumor segmentation from CT images using an AIM-Unet. Bioengineering 2023, 10, 215. [Google Scholar] [CrossRef] [PubMed]
  18. Hu, H.; Zhang, Z.; Xie, Z.; Lin, S. Local Relation Networks for Image Recognition. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3464–3473. [Google Scholar]
  19. Lee, S.; Han, K.; Shin, H.; Park, H.; Kim, S.; Kim, J.; Yang, X.; Yang, J.D.; Yu, H.C.; You, H. G-UNETR++: A gradient-enhanced network for accurate and robust liver segmentation from computed tomography images. Appl. Sci. 2025, 15, 837. [Google Scholar] [CrossRef]
  20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  21. Karimi, D.; Vasylechko, S.D.; Gholipour, A. Convolution-free medical image segmentation using transformers. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Strasbourg, France, 27 September–1 October 2021; pp. 78–88. [Google Scholar]
  22. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Tel Aviv, Israel, 23–27 October 2022; pp. 205–218. [Google Scholar]
  23. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  24. Valanarasu, J.M.J.; Oza, P.; Hacihaliloglu, I.; Patel, V.M. Medical transformer: Gated axial-attention for medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Strasbourg, France, 27 September–1 October 2021; pp. 36–46. [Google Scholar]
  25. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.R.; Xu, D. UNETR: Transformers for 3D Medical Image Segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 574–584. [Google Scholar]
  26. Zhou, H.-Y.; Guo, J.; Zhang, Y.; Han, X.; Wang, L.; Yu, Y. nnFormer: Volumetric medical image segmentation via a 3D transformer. IEEE Trans. Image Process. 2023, 32, 4036–4045. [Google Scholar] [CrossRef]
  27. Shaker, A.; Maaz, M.; Rasheed, H.; Khan, S.; Yang, M.H.; Khan, F.S. UNETR++: Delving into Efficient and Accurate 3D Medical Image Segmentation. IEEE Trans. Med. Imaging 2024, 43, 3088–3100. [Google Scholar] [CrossRef] [PubMed]
  28. Fang, X.; Yan, P. Multi-organ segmentation over partially labeled datasets with multi-scale feature abstraction. IEEE Trans. Med. Imaging 2020, 39, 3619–3629. [Google Scholar]
  29. Zhang, Y.; Liu, H.; Hu, Q. Transfuse: Fusing transformers and cnns for medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Strasbourg, France, 27 September–1 October 2021; pp. 14–24. [Google Scholar]
  30. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  31. Xie, Y.; Zhang, J.; Shen, C.; Xia, Y. CoTr: Efficiently bridging CNN and transformer for 3D medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Strasbourg, France, 27 September–1 October 2021; pp. 171–180. [Google Scholar]
  32. Jin, Q.; Meng, Z.; Sun, C.; Cui, H.; Su, R. RA-UNet: A hybrid deep attention-aware network to extract liver and tumor in CT scans. Front. Bioeng. Biotechnol. 2020, 8, 1471. [Google Scholar]
  33. Zhang, C.; Lu, J.; Hua, Q.; Li, C.; Wang, P. SAA-Net: U shaped network with Scale-Axis-Attention for liver tumor segmentation. Biomed. Signal Process. Control 2022, 73, 103460. [Google Scholar]
  34. Jiang, L.; Ou, J.; Liu, R.; Zou, Y.; Xie, T.; Xiao, H.; Bai, T. RMAU-Net: Residual multi-scale attention U-Net for liver and tumor segmentation in CT images. Comput. Biol. Med. 2023, 158, 106838. [Google Scholar]
  35. Muhammad, S.; Zhang, J. Segmentation of Liver Tumors by Monai and PyTorch in CT Images with Deep Learning Techniques. Appl. Sci. 2024, 14, 5144. [Google Scholar] [CrossRef]
  36. Yashaswini, G.N.; Manjunath, R.V.; Shubha, B.; Prabha, P.; Aishwarya, N.; Manu, H.M. Deep learning technique for automatic liver and liver tumor segmentation in CT images. J. Liver Transplant. 2025, 17, 100251. [Google Scholar] [CrossRef]
  37. Balaguer-Montero, M.; Marcos Morales, A.; Ligero, M.; Zatse, C.; Leiva, D.; Atlagich, L.M.; Staikoglou, N.; Viaplana, C.; Monreal, C.; Mateo, J.; et al. A CT-based deep learning-driven tool for automatic liver tumor detection and delineation in patients with cancer. Cell Rep. Med. 2025, 6, 102032. [Google Scholar]
  38. Zhang, C.; Wang, L.; Zhang, C.; Zhang, Y.; Li, J.; Wang, P. Liver Tumor Segmentation Based on Multi-Scale Deformable Feature Fusion and Global Context Awareness. Biomimetics 2025, 10, 576. [Google Scholar] [CrossRef]
  39. Bilic, P.; Christ, P.; Li, H.B.; Vorontsov, E.; Ben-Cohen, A.; Kaissis, G.; Szeskin, A.; Jacobs, C.; Mamani, G.E.H.; Chartrand, G.; et al. The liver tumor segmentation benchmark (LiTS). Med. Image Anal. 2023, 84, 102680. [Google Scholar] [CrossRef] [PubMed]
  40. Soler, L.; Hosttettle, A.; Charnoz, A.; Fasquel, J.; Moreau, J. 3D Image Reconstruction for Comparison of Algorithm Database: A Patient Specific Anatomical and Medical Image Database. Available online: https://www.ircad.fr/research/data-sets/liver-segmentation-3d-ircadb-01/ (accessed on 5 December 2025).
  41. Karimi, D.; Salcudean, S.E. Reducing the hausdorff distance in medical image segmentation with convolutional neural networks. IEEE Trans. Med. Imaging 2020, 39, 499–513. [Google Scholar] [CrossRef]
  42. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  43. Lee, C.Y.; Xie, S.; Gallagher, P.; Zhang, Z.; Tu, Z. Deeply-Supervised Nets. In Proceedings of the Artificial Intelligence and Statistics, San Diego, CA, USA, 9–12 May 2015; pp. 562–570. [Google Scholar]
  44. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  45. Chen, Y.; Hu, F.; Wang, Y.; Zheng, C. Hybrid-attention densely connected U-Net with GAP for extracting livers from CT volumes. Med. Phys. 2022, 49, 1015–1033. [Google Scholar] [PubMed]
  46. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [Google Scholar]
  47. Kushnure, D.T.; Talbar, S.N. HFRU-Net: High-level feature fusion and recalibration unet for automatic liver and tumor segmentation in CT images. Comput. Methods Programs Biomed. 2022, 213, 106501. [Google Scholar]
Figure 1. CT image processing performed to enhance the visibility of the liver and tumor regions by normalizing the Hounsfield Unit values to the range of −250 to 250: (a) the original CT slice; (b) the enhanced CT slice.
Figure 1. CT image processing performed to enhance the visibility of the liver and tumor regions by normalizing the Hounsfield Unit values to the range of −250 to 250: (a) the original CT slice; (b) the enhanced CT slice.
Diagnostics 16 00429 g001
Figure 2. The proposed pipeline for liver tumor segmentation.
Figure 2. The proposed pipeline for liver tumor segmentation.
Diagnostics 16 00429 g002
Figure 3. Architecture of G-UNETR++ for liver segmentation and tumor segmentation.
Figure 3. Architecture of G-UNETR++ for liver segmentation and tumor segmentation.
Diagnostics 16 00429 g003
Figure 4. Examples of tumor segmentation results of the proposed model over the LiTS dataset. Upper: the ground truth; bottom: the predicted tumor regions.
Figure 4. Examples of tumor segmentation results of the proposed model over the LiTS dataset. Upper: the ground truth; bottom: the predicted tumor regions.
Diagnostics 16 00429 g004
Figure 5. Examples of tumor segmentation results of the proposed model over the 3DIRCADb dataset. Upper: the ground truth; bottom: the predicted tumor regions.
Figure 5. Examples of tumor segmentation results of the proposed model over the 3DIRCADb dataset. Upper: the ground truth; bottom: the predicted tumor regions.
Diagnostics 16 00429 g005
Figure 6. An illustration of the post-processing result: (a) the ground truth tumor boundary (yellow); (b) the predicted tumor boundary before post-processing (blue); (c) the dilated tumor boundary after post-processing (green); (d) the overlay of the ground truth tumor boundary (yellow) and the dilated tumor boundary (green).
Figure 6. An illustration of the post-processing result: (a) the ground truth tumor boundary (yellow); (b) the predicted tumor boundary before post-processing (blue); (c) the dilated tumor boundary after post-processing (green); (d) the overlay of the ground truth tumor boundary (yellow) and the dilated tumor boundary (green).
Diagnostics 16 00429 g006
Figure 7. Plotting of training loss and validation loss during our model development.
Figure 7. Plotting of training loss and validation loss during our model development.
Diagnostics 16 00429 g007
Figure 8. Precision-recall plotting for liver tumor segmentation over the LiTS dataset (left) and the 3DIRCADb dataset (right).
Figure 8. Precision-recall plotting for liver tumor segmentation over the LiTS dataset (left) and the 3DIRCADb dataset (right).
Diagnostics 16 00429 g008
Figure 9. The following are missing in the final tumor segmentation results due to the missing tumors in liver segmentation results: (a) the original CT slice; (b) the ground truth for the liver (red) and tumors (orange); (c) the liver segmentation result with two tumors missing (highlighted in dotted circles); (d) the tumor segmentation results (orange) with two tumors missing (highlighted in dotted circles).
Figure 9. The following are missing in the final tumor segmentation results due to the missing tumors in liver segmentation results: (a) the original CT slice; (b) the ground truth for the liver (red) and tumors (orange); (c) the liver segmentation result with two tumors missing (highlighted in dotted circles); (d) the tumor segmentation results (orange) with two tumors missing (highlighted in dotted circles).
Diagnostics 16 00429 g009
Figure 10. An extremely small tumor missing in the final tumor segmentation results: (a) the original CT slice; (b) the ground truth for the liver (red) and tumors (orange) with an extremely small tumor (highlighted in a dotted circle); (c) the liver segmentation result; (d) the tumor segmentation result (orange) with one tumor missing (highlighted in a dotted circle).
Figure 10. An extremely small tumor missing in the final tumor segmentation results: (a) the original CT slice; (b) the ground truth for the liver (red) and tumors (orange) with an extremely small tumor (highlighted in a dotted circle); (c) the liver segmentation result; (d) the tumor segmentation result (orange) with one tumor missing (highlighted in a dotted circle).
Diagnostics 16 00429 g010
Table 1. A comparison of the proposed model with state-of-the-art models for liver tumor segmentation over the LiTS and 3DIRCADb datasets.
Table 1. A comparison of the proposed model with state-of-the-art models for liver tumor segmentation over the LiTS and 3DIRCADb datasets.
DatasetMethodDSCVOERAVDASSD
LiTSHDU-Net [45]0.7110.4010.0237.201
ResUNet [46]0.7050.3950.5348.286
MS-FANet [2]0.7420.3670.1075.996
HFRU-Net [47]0.749 ± 0.1070.380 ± 0.1280.218 ± 0.152-
RMAU-Net [34]0.762 ± 0.1180.371 ± 0.1350.012 ± 0.291-
The proposed0.844 ± 0.0780.263 ± 0.1140.133 ± 0.1431.317 ± 0.645
3DIRCADbHDU-Net0.6920.3824.83516.516
ResUNet0.7390.3570.1027.817
MS-FANet0.7800.3130.1555.346
HFRU-Net0.789 ± 0.1110.326 ± 0.1420.033 ± 0.170-
RMAU-Net0.831 ± 0.0950.275 ± 0.1250.126 ± 0.186-
The proposed0.832 ± 0.0600.283 ± 0.0850.138 ± 0.1111.682 ± 1.029
Table 2. The ablation study conducted to evaluate the performance of the proposed post-processing method in liver tumor segmentation with the LiTS and 3DIRCADb datasets.
Table 2. The ablation study conducted to evaluate the performance of the proposed post-processing method in liver tumor segmentation with the LiTS and 3DIRCADb datasets.
DatasetMethodDSCVOERAVDASSD
LiTSWithout post-processing0.8450.2610.1431.267
With post-processing0.8440.2630.1331.317
3DIRCADbWithout post-processing0.8030.3130.1981.784
With post-processing0.8320.2830.1381.682
Table 3. The ablation study conducted to determine the right number of dilation times for the proposed post-processing method in liver tumor segmentation with the LiTS and 3DIRCADb datasets.
Table 3. The ablation study conducted to determine the right number of dilation times for the proposed post-processing method in liver tumor segmentation with the LiTS and 3DIRCADb datasets.
DatasetDilation TimesDSCVOERAVDASSD
LiTS10.8450.2610.1381.311
20.8440.2630.1331.317
30.8410.2670.1271.327
3DIRCADb10.8270.2910.1641.712
20.8320.2830.1381.682
30.8240.2940.1921.693
Table 4. A summary of the complexity level of the proposed model.
Table 4. A summary of the complexity level of the proposed model.
Number of ParametersFloating Point Operations/SecondTraining Time/EpochInference Time/CT Scan
LiTS Dataset3DIRACDb Dataset
97.73 M73.12 G26 min176.9 ± 124.8 s43.8 ± 12.8 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shin, H.; Han, K.; Lee, S.; Park, H.; Kim, S.; Kim, J.; Yang, X.; Yang, J.D.; Song, J.; Yu, H.C.; et al. Deep Learning-Based Liver Tumor Segmentation from Computed Tomography Scans with a Gradient-Enhanced Network. Diagnostics 2026, 16, 429. https://doi.org/10.3390/diagnostics16030429

AMA Style

Shin H, Han K, Lee S, Park H, Kim S, Kim J, Yang X, Yang JD, Song J, Yu HC, et al. Deep Learning-Based Liver Tumor Segmentation from Computed Tomography Scans with a Gradient-Enhanced Network. Diagnostics. 2026; 16(3):429. https://doi.org/10.3390/diagnostics16030429

Chicago/Turabian Style

Shin, Hangyeul, Kyujin Han, Seungyoo Lee, Harin Park, Seunghyon Kim, Jeonghun Kim, Xiaopeng Yang, Jae Do Yang, Jisoo Song, Hee Chul Yu, and et al. 2026. "Deep Learning-Based Liver Tumor Segmentation from Computed Tomography Scans with a Gradient-Enhanced Network" Diagnostics 16, no. 3: 429. https://doi.org/10.3390/diagnostics16030429

APA Style

Shin, H., Han, K., Lee, S., Park, H., Kim, S., Kim, J., Yang, X., Yang, J. D., Song, J., Yu, H. C., & You, H. (2026). Deep Learning-Based Liver Tumor Segmentation from Computed Tomography Scans with a Gradient-Enhanced Network. Diagnostics, 16(3), 429. https://doi.org/10.3390/diagnostics16030429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop