Abstract
Objective: Segmentation of liver and liver tumors from 3D medical images is a challenging and computationally expensive task. Organs that are in close proximity may have similar shape, texture, and intensity, which makes it difficult for accurate segmentation. Accurate segmentation of liver tumors is important for diagnosis and treatment planning of liver cancer. Methods: A hybrid model with a U-Net based structure and the Whale Optimization Algorithm (WOA) was proposed. WOA was used to optimize the hyperparameters of the conventional LiTS-Res-UNet to obtain the best segmentation performance of the deep learning model. Results: The LiTS-Res-Unet + WOA hybrid model achieved a performance of 99.54% for accuracy, with a Dice coefficient of 92.38% and a Jaccard index of 86.73% on the benchmark dataset, outperforming state-of-the-art methods. Conclusions: The WOA-based adaptive search space was able to obtain an optimal set of hyperparameters for deep learning model convergence while increasing the accuracy of the model in the proposed hybrid model. The robust performance and clinical applicability of the model in liver tumor segmentation were demonstrated.
1. Introduction
The liver is a vital metabolic and digestive organ of the body that is the most common site of primary and metastatic tumors [,]. Liver tumors are the most frequent and lethal tumors in the world. Hepatocellular carcinoma (HCC) is one of the most common types of liver cancer. It is the third most common cause of cancer death, and the incidence of this type of cancer has been gradually increasing in the past few decades [,]. Dynamic medical imaging, especially computed tomography (CT) scans, has been extensively used in clinical practice for the diagnosis, staging, and treatment planning of HCC patients [,]. However, manually analyzing thousands of slices of CT images for each patient to find out the region of the tumor and to measure its size and extent can be an extremely time-consuming task even for the most expert radiologists, as it requires a high level of experience and expertise. Moreover, this can lead to fatigue and inefficiency in a longer analysis and potential misdiagnosis and can have serious impact on the physical and mental health of the physician []. Therefore, the use of automatic segmentation of HCC from CT medical images using Artificial Intelligence (AI)-based methods is critical to aid the clinicians. CT image-based Computer-Aided Diagnosis (CAD) systems can be extremely helpful in improving the diagnostic performance and decision-making capability of radiologists. Nowadays, deep learning algorithms such as convolutional neural networks (CNNs) that are capable of learning higher-level features are applied in the direct segmentation of liver tumors from raw medical images. Deep learning-based segmentation models have been popularly applied in the successful implementation of segmentation tasks, and their encoder–decoder design has produced the best performance for precise tumor delineation. Models like Fully Convolutional Network (FCN) and SegNet have helped to increase the accuracy of liver tumor classification [,]. The classification has been carried out using different machine learning based classifiers, such as Support Vector Machine (SVM), Random Forest (RF), and k-nearest neighbor (KNN), which differentiate between benign and malignant tumors. However, conventional methods often require expensive manual feature extraction, which inherently lacks the ability to capture the complex patterns of medical images. To address this issue, we can use transfer learning models such as CNN, VGG-16, and ResNet, which have been widely applied in segmentation and classification tasks. These models can directly process raw images and have the capability of extracting lesion features and classifying liver lesions as benign, malignant, or ambiguous. Even though the detection of lesions has achieved considerable progress, the classification performance of the models is somewhat limited by variations in tumor size, texture, and imaging modality as well as by class imbalance and the quality of images [,]. Stronger approaches that can bypass these shortcomings and provide more accurate and reliable segmentation and classification are required. In addition, an imbalanced class between benign and malignant tumors is a difficult problem. In most of the cancer data, the benign tumors are more abundant than the malignant ones. Therefore, the models are inclined to separate into a higher class and are unable to diagnose malignancy. Most of the segmentation algorithms like U-Net do not work well on low resolution or noisy images and are very costly. Therefore, more efficient, adaptive, and robust models are required to better detect and diagnose liver tumors []. Biologically motivated networks such as CNNs [,] have shown success in a variety of medical image analysis tasks such as organ segmentation [], texture analysis [], and disease classification []. Several research studies have reported that CNNs can provide good performance in cancer detection and diagnosis [,,]. Nevertheless, the success of the segmentation and classification of the algorithms based on CNN models relies on the architecture and hyperparameters of the network that also impact on the performance of the CAD system. These parameters have a great impact on performance, and the algorithms based on natural phenomena can collaborate with deep learning models, exploring the solution space in a global way and finding the best hyperparameters. Conventional strategies for medical image processing have given way to innovative bio-inspired ones. Observed bio-inspired solutions are imitation of the behavior of animals, birds, and insects. Complex optimization problems are solved by such algorithms. The whale algorithm is a bio-inspired one of these algorithms and is based on the simulation of the behavior of the whale while attacking and hunting its prey. Particle Swarm Optimization (PSO) [], Whale Optimization Algorithm (WOA) [], Artificial Bee colony Optimization (ABC) [], Differential Evolution (DE) [], Harmony Search (HS) [], Gravitational Search (GS) [], Gray Wolf Optimization (GWO) [], Antlion Optimization (ALO) [], and Ant Colony Optimization (ACO) [] are some well-known algorithms in this category. We implemented our study using WOA, as it can simulate the bubble-net hunting tactic deployed by humpback whales, which is suitable for maneuvering over the complex solution space of image segmentation that comprises multiple modes. Unlike PSO and other algorithms, which have the possibility of premature convergence or becoming stuck in local optima, WOA includes a linearly decreasing control parameter that adjusts the exploration strategy of the search from global to local search, which enhances the reliability of the convergence. In addition, the simplicity of the WOA structure, that contains fewer control parameters than the other algorithms, makes its implementation and the adjustment of parameters simple yet competitive (or superior) when compared to the other algorithms in terms of convergence speed, robustness, and segmentation accuracy. These benefits make WOA a good choice for image segmentation problems, where both accuracy and efficiency of computation are critical. Other research studies demonstrated that these algorithms were effective in solving complex optimization problems in medicine [,]. Unlike existing systems that are developed based on a feature engineering approach or a combined application of feature engineering and deep learning techniques, we propose a bio-inspired deep learning solution for liver cancer diagnosis with CT images. The hybridization of the combination of several deep learning models with WOA bio-inspired optimization was investigated for liver segmentation and feature extraction.
Even though the LiTs-Res-UNet architecture has attained promising outcomes, its performance is excessively reliant on human-selected hyperparameters, which might result in non-optimal convergence behavior. In this work, we propose a new hybrid LiTs-Res-UNet + WOA, in which the Whale Optimization Algorithm is applied for adaptively fine-tuning key hyperparameters for improved segmentation accuracy, robustness, and generalization. The contributions of this paper are as follows:
- We propose a novel hybrid optimization-segmentation framework called LiTs-Res-Unet + WOA, which integrates the Whale Optimization Algorithm (WOA) with a UNet-based architecture to automatically and adaptively tune the hyperparameters required for liver and tumor segmentation.
- Unlike prior studies that apply WOA to optimize a single parameter, we propose a closed-loop optimization framework in which WOA tunes a set of interdependent hyperparameters (learning rate, dropout rate, batch size) dynamically during training. This introduces a meta-optimization layer that is adaptive to the changing loss landscape of the ResNet-50–U-Net model.
- The inner level updates the network weights using backpropagation, and the outer level optimizes a population-based hyperparameter search using WOA by the validation Dice score. The convergence of validation performance shows that the hyperparameter optimization (HO) by the WOA outperforms the grid search and the Bayesian optimization methods because it achieves a better exploration–exploitation trade-off.
- Extensive experiments conducted on the LiTS17 benchmark dataset showed that the proposed model demonstrated state-of-the-art performance—with achieved pixel accuracy of 99.54%, Dice coefficient of 92.38%, and Jaccard index of 86.73%, respectively, which significantly outperformed existing segmentation techniques.
- The research results indicate that LiTs-UNet-WOA exhibits superior accuracy and robustness and has the potential to be an effective solution for liver tumor detection and treatment planning in real-life clinical applications.
The rest of the paper analyses the recent research and related issues in Section 2. Section 3 explains the methodology of the proposed work and the application of the optimized meta-heuristic approach. Section 4 presents the segmentation results. Section 5 and Section 6 represent the discussion and limitations of the proposed work, and Section 7 provides the conclusion of this work.
2. Related Work
Effective segmentation of liver tumors is one of the pivotal steps in medical image analysis, ranging from precise diagnosis and efficient treatment planning to follow-up monitoring of liver cancer. Many deep learning techniques have been developed to improve the accuracy and efficiency of liver tumor segmentation. In this review, we describe recent progress on segmentation algorithms and methods by highlighting the important works published in recent years based on different computational paradigms. An improved Artificial Bee Colony (IABC) algorithm and a general type-2 fuzzy possibilistic c-means (FPCM) were used in consistence with a CRF for brain tumor segmentation []. The method starts with removing annotations by a tracking algorithm and a given threshold. CRF estimates a conditional probability distribution based on image features to obtain posterior matrices for all image kernels. Then, the matrices are fed to the ABC algorithm for forming clusters, and the tumor cluster is determined from an adaptive threshold. It was evaluated on a sample of 100 MRI scans. Nevertheless, it depends on only one threshold calculated by ABC, which is suboptimal for liver segmentation because of the various lesion intensities. In [], the segmentation of the whole liver in CT images is proposed based on the ABC algorithm as a clustering technique. The abdominal images were grouped into ABCs, and taking advantage of the sorted cluster centroids, extremum points were removed to establish the initial liver segmentation. This output was further refined by using a region-growing approach. This approach serves as a basis for the current study, which attempts to alleviate its weaknesses by using an alternative metaheuristic algorithm to obtain better performance. In [], different transfer mechanisms and pheromone updating rules were implemented to work in the ACO framework for image segmentation, including edges, objects, and background. Fuzzy clustering was adopted to separate the background and foreground so that ants can work on boundaries effectively. Histogram peaks were used to determine the number of clusters and to initialize the clusters, optimizing efficiency and minimizing search time. But the method was evaluated on one image only and primarily focused on edge detection, thus being inapplicable to global liver segmentation. In [], the Firefly Algorithm (FA) was employed in mammography image processing for breast tumor segmentation. Pattern matrix and kernel were used to measure pixel dissimilarity, generating a dissimilarity map. Subsequently, the labeling of suspicious regions was optimized by FA and refined by FCM clustering. Though effective for obviously discriminated cancerous tissues, this method is not competent for liver tissues with homogenous and inhomogeneous appearance, and it focuses on classification rather than full liver segmentation. An ABC-based approach was used in [] to identify multiple image thresholds, yielding results equivalent to a near-exhaustive search. Its performance was compared with HBMO, PSO, Fast Otsu’s method, and the hybrid HCOCLPSO algorithm. Chen et al. [] presented a Multi-Scale Liver Tumor Segmentation Network combining convolutional layers with Transformer-based architecture to handle short- and long-range dependencies. Evaluated on the LiTS17 dataset, it achieved Dice similarity coefficients of 0.920 for liver segmentation and 0.748 for tumor segmentation, improving baseline models while preserving fine-grained anatomical details. You et al. [] proposed PGC-Net to perform contour-based parallel graph reasoning and the Pyramid Vision Transformer to extract multi-scale features. On LiTS17 and 3DIRCADb, it achieved mean Dice scores of 73.63% and 74.16%, but limitations remain in unclear tumor boundaries and inaccurate tumor margin delineation. He et al. [] presented PAKS-Net, a 2D segmentation model leveraging global Swin Transformer features and a Position-Aware module to capture spatial relations of the tumor and adjacent organs. The model obtained tumor Dice scores of 0.893 (KiTS19), 0.769 (LiTS17), 0.598 (pancreas), and 0.738 (LOTUS), but still suffered from noise and slice-to-slice variability. Shui et al. [] proposed a three-path neural network using MSFF, MFF, EI, and EG modules. On LiTS2017, it achieved a Dice coefficient of 85.55% and a Jaccard index of 81.11%. Cross-dataset validation reached Dice scores of 80.14% and 81.68% on 3DIRCADb and clinical datasets. Biswas et al. [] introduced a GAN-based data augmentation method enhancing the quality and utility of training samples for biomedical image segmentation, yielding Dice scores of 0.908 on MIDAS, 0.872 on 3DIRCADb, and 0.605 on LiTS datasets. Wang et al. [] proposed a Context Fusion Network using TSA modules and MSA skip connections, achieving Jaccard index of 81.56% and Dice score of 85.97% on LiTS2017, and 80.11% and 83.67% on 3DIRCADb. Muhammad et al. [] introduced a hybrid ResUNet model combining ResNet and UNet for liver tumor segmentation on the MSD Task03 Liver dataset. The model, implemented in MONAI and PyTorch, reached Dice coefficients of 0.98 for tumor detection and 0.87 for segmentation. Fallahpoor et al. [] used Isensee’s 2017 architecture for liver and lesion segmentation from 3D MRI, achieving mean Dice scores of 88% for liver and 53% for tumors on 128 patients’ T1w and T2w MRI images. Wang et al. [] applied UNet++ for automated liver and tumor segmentation on MRI images of hepatocellular carcinoma patients, reaching a Dice score of 0.91 for liver and 0.612 for tumors. Recent advances (e.g., FasNet) indicate trends in medical imaging [,]. McGrath et al. [] and Mojtahed et al. [] illustrated the growing importance of semi-automated and deep learning methods. Wang et al. [] and Vaidhya Venkadesh et al. [] suggested incorporating multimodal imaging and historical patient records to improve performance []. Singh et al. [] introduced FasNet, a deep learning model fusing ResNet-50 with VGG-16, incorporating Channel and Spatial Attention and Monte Carlo Dropout for enhanced liver tumor segmentation. Tested on LiTS17, FasNet achieved Dice of 0.8766 and Jaccard of 0.8487, outperforming several other methods in feature concentration, robustness, and uncertainty estimation. In [], the authors demonstrated that GAN-generated samples can enhance segmentation performance, particularly for liver tumor imaging. They Achieved Dice scores of 0.908, 0.872, and 0.605 on the MIDAS, 3Dircadb, and LiTS datasets, respectively, Data summaries are presented in Figure 1.
Figure 1.
Dice coefficient comparison of liver tumor segmentation techniques across different datasets and publications.
3. Methodology
In this section, we outline the methods employed in our study. The pipeline of our proposed system is depicted in Figure 2. The first step is to construct the dataset, where three different views were obtained from the CT dataset. Subsequently, preprocessing is performed, followed by liver segmentation, and finally tumor segmentation. A summary of the materials and methods used in the current work, including the dataset, performance criteria, the state-of-the-art LiTs-Res-UNet, the bio-inspired novel architecture LiTs-Res-Unet + WOA with a metaheuristic algorithm, and the proposed framework, is presented in this section.
Figure 2.
Overview of the proposed two-stage computer-aided liver tumor detection and diagnosis framework. Stage I (Computer-Aided Detection) involves preprocessing, liver and tumor segmentation, and feature extraction. Stage II (Computer-Aided Diagnosis) performs feature selection, prediction/decision making, and Explainable AI analysis. The Whale Optimization Algorithm (WOA) is employed for hyperparameter optimization, enhancing segmentation accuracy and performance metrics including Dice coefficient, IoU, sensitivity, and accuracy.
3.1. Dataset Description
We used publicly available standard data sets to validate and evaluate our liver tumor segmentation []. All experiments were performed on contrast-enhanced abdominal CT images from the LiTS2017 dataset (131 volumetric scans) and Radiopaedia (105 CT series). We did not use any MRI data in our experiments. Each volume is associated with a ground truth segmentation mask of the exact resolution that identifies the liver and, if present, hepatic tumors. These scans originate from different patient cases and scanning protocols, resulting in a mixture of anatomical presentations and imaging characteristics across the dataset. The second collection was extracted from Radiopaedia, a popular medical imaging database. It contains more than 105 liver CT series from many patients with a broad range of abdominal clinical presentations. While these scans are also in NIfTI format, the labeling quality and consistency vary in the same way as in real-world diagnostic settings. This dataset was added to improve the generalization performance and domain invariance of the model. All data were reconstructed into 2D axial slices. To ensure dataset consistency, each slice was resized to 128 × 128 × 3 pixels. CT slices were saved as JPEG files after applying suitable windowing (usually liver windowing with window = 150 and level = 30), and masks were saved as grayscale PNG files. This preprocessing step simplified storing and loading data on a standard PC with limited memory and facilitated integration with common deep-learning frameworks.
3.2. Dataset Preprocessing
Before training the model, all volumetric CT images and corresponding masks were put through a structured preprocessing process. Initially, NIfTI (.nii) 3D volumes were loaded using the NiBabel library and reoriented to a standard axial orientation. Each volume was subsequently cut into individual 2D slices on the axial plane. Figure 3 shows a single axial slice of a CT scan after applying the windowing technique optimized for the liver. The windowing step improves the display of the liver and the surrounding tissue by adjusting the intensity range using the predetermined liver windowing parameters (width = 150, level = 30). The windowed bone colormap of an axial image can show the liver with respect to other structures in the abdomen by visually aiding segmentation. By adding the windowing, we can more clearly differentiate liver tissue from other organs, thus better supporting subsequent segments. Figure 4 depicts the unmodified input image (left), corresponding to the raw CT slice, and the windowed image (immediately left of the center), which improves the visibility of the liver by changing the range of intensities. The Mask (center-right) identifies the liver area, and the Liver and Mask Overlay (right) shows the liver area overlaid on the windowed image. This ensures more accurate segmentation and better visual perception. The slices were then resized to a fixed resolution of 128 × 128 × 3 pixels using bicubic interpolation for CT images and nearest-neighbor interpolation for masks to maintain categorical label correspondence. Additionally, histogram-based normalization and frequency-equalized scaling were optionally applied to enhance contrast uniformity among samples. CT images were saved in JPEG format for storage efficiency and compatibility with popular image processing pipelines, whereas binary masks were saved as PNG images containing a single grayscale channel.
Figure 3.
The original CT slice which enhances the contrast of the liver region to improve segmentation accuracy.
Figure 4.
Preprocessing of CT scan: original image, windowed image, mask, liver, and mask overlay.
For more reliable experimental assessment, the merged dataset (LiTS17 + Radiopaedia) was split at the patient level (i.e., to prevent data leakage between splits) into train, validation, and test sets at a ratio of 70:10:20, ensuring that CT scans from the same patient were present in at most one of the splits. The large class imbalance between tumor and non-tumor regions was addressed by using a weighted loss function and balanced patch sampling at training time. While the current work reports results for a single fixed split, several training courses run with different random seeds were performed in the reported experiments to verify the stability of the results. These are the CT scan images with associated liver segmentation masks in Figure 5. Each of the images is preprocessed to focus on the liver region. The masks are used to indicate the location of the liver in the CT scan. The batch of images and masks shows the preprocessing and segmentation.
Figure 5.
Batch of CT scan images with liver segmentation masks.
3.3. Based Model Architecture
3.3.1. Pre-Trained Encoder
LiTs-Res-UNet is a deep encoder network, and ResNet-50 is adopted as the encoder to acquire high-quality contextual details. The ResNet-50 model, which has been pre-trained on multi-million ImageNet, contains fifty trained layers with residual blocks to reduce vanishing gradients. The point is that the error gradient propagated between the output and the input layers drops and is extremely small, which does not make it possible to learn in the initial layers. The residual block formula is the following:
F(x) is the output of the layers and x is the input. The residual block is presented in Figure 6. The use of ResNet-50 as a feature extractor in reference to colon polyp segmentation is expected to improve the accuracy of the model.
Figure 6.
Residual block.
3.3.2. Model Decoder Path
The LiTs-Res-UNet decoder progressively increases the resolution of the feature map by up-sampling to reassemble the spatial details lost during the encoding phase. These features are elaborated by the deeper convolutional layers in the sequence, and the batch normalization provides the stability of the learning process. Regularization methods, like dropout, reduce the problem of overfitting. Additional layers are transposed convolution (reverse convolution) layers, which increase the size of the feature map, increasing segmentation. Lastly, a 1 × 1 convolution with a sigmoid activation transforms the feature map into a probabilistic representation, saying whether each pixel is the object of interest or not, reconstructing the segmentation mask. The overall architecture of the LiTs-Res-UNet model is depicted in Figure 7.
Figure 7.
Proposed Hybrid LiTs-Res-UNet model architecture.
3.4. Swarm-Based Whale Optimization Algorithm (WOA)
WOA [] is a nature-based metaheuristic optimization method replicating the social model and hunting phase of humpback whales in nature, referred to as bubble-net hunting. It also falls into the swarm intelligence family of optimization algorithms inspired by the collective motion of flocks of birds or schools of fish, addressing problems. In this part, we discuss the Whale Optimization Algorithm’s theories, principles, and mathematical formulations.
3.4.1. Motivation from Nature
Swarm Intelligence (SI) algorithms are based on the population and behavior of social animals, such as birds, fish, and mammals, and WOA is one such swarm-based technique. Despite large differences in their behavior as they move independently, these animals also show emergent collective behaviors that enable effective collaboration. Humpback whales, known for their smart and synchronized hunting techniques, are the primary models for WOA. One notable behavior is bubble-net feeding, in which a group of whale’s spirals upward through the water, releasing increasingly dense curtains of bubbles, driving fish or krill toward the surface and into their interior circle. This predation process is the main metaphor for the updating mechanism applied in WOA. In the optimization problem, to explore and exploit for enhancing the global optimum, the search agents (whales) operate in the search space. The optimization in WOA is performed using two primary steps: Exploration and Exploitation, which enable a trade-off between global search (searching new regions) and local search (refining solutions around the best solution found).
3.4.2. Exploration and Exploitation Phases
WOA integrates two underlying mechanisms to mimic whale hunting behavior: the shrinking encircling mechanism and the spiral movement.
- Exploration: This process iteratively traverses the search dimensions in search of the global optimum. During this phase, whales move randomly to search for unvisited regions. The search is guided by the position of the best solution found so far, allowing the algorithm to escape local optima.
- Exploitation: This phase focuses on refining the search around the best solution detected. Whales use the shrinking (enclosing) mechanism along with the spiral path, updating their position by considering the spiral trajectory, which converges more effectively to the global optimum.
These two stages, shown in Figure 8, cooperate to allow WOA to find a trade-off between diversification (global exploration) and intensification (local exploitation).
Figure 8.
Hunting strategies of the whale optimization algorithm.
3.5. Proposed U-Net Segmentation with WOA
Here, we present the U-Net segmentation model to liver tumor segmentation, and hyperparameter optimization of the model (learning rate, dropout rate, and batch size) using the Whale Optimization Algorithm (WOA). First, we define the U-Net Segmentation Problem class, inheriting the Problem class from the Mealpy library. This class wraps up the optimization problem by defining an objective function, which accepts a candidate solution (hyperparameters set) and decodes this solution to the learning rate, dropout rate, and batch size. These hyperparameters are subsequently employed to construct and compile a U-Net model and to fit it for a given number of epochs on the given training sets (train images and train masks) and to predict on the given test sets (test images and test masks). The objective function is the test accuracy to be maximized for optimization.
The U-Net model directly forms an encoder–decoder network. The features of the input image are extracted by the encoder through convolutional layers, with dropout layers to prevent overfitting. The feature maps are sampled, and convolutions are applied in the decoder to predict the segmentation mask. The output of the model is resized using a Lambda layer at the expected 128 × 128 pixel size. The output is passed through a sigmoid to obtain a final binary segmentation mask.
The Whale Optimization Algorithm (WOA) is used to do hyperparameter tuning. WOA is an evolutionary-based optimization method proposed by Mirjalili [] through an observation of the hunting style of humpback whales that keeps the tradeoff between the explorative and exploitative style. It works by cycling through a population of potential solutions (whales) and moving its position with the best solution encountered up to that point. This makes Anytime Net capable of effectively finding the optimum number for learning rate, dropout rate, and batch size. Boundaries for these hyperparameters are specified in the code, and the optimization is performed for a certain number of iterations. The model with the best solution is chosen, using the test set accuracy to guide the decision.
Once the optimization process has been finished, the best agent (hyperparameter set) is found and the corresponding accuracy and optimized hyperparameters (learning rate, dropout and batch size) are printed. This framework is successful to couple deep learning with nature solved optimization approaches in order to improve the performance of the segmentation model, and it becomes a good choice for liver tumor segmentation. As demonstrated in Algorithm 1, the optimal hyperparameters for liver segmentation are computed. The overview of the proposed LiTs-Res-Unet + WOA optimized clustering with bio-inspired WOA for enhanced liver tumor CT image segmentation is shown in Figure 9.
Figure 9.
The proposed LiTs-Res-Unet + WOA approach.
To better visualize the interaction between WOA and the gradient-based learning procedure in ResNet-50–UNet, we assume that each whale in the WOA population corresponds to a candidate vector H = [learning rate, dropout rate, batch size] in hyperparameter space. For each candidate H, the ResNet-50–UNet network is trained using the regular backpropagation-based training for a few epochs, and the achieved validation Dice score is recorded as the fitness value. Then, the corresponding hyperparameter vectors are updated using WOA’s exploration (global search) and exploitation (local refinement) phases. During exploration, the whales perform random walks in the hyperparameter search space to find new candidate vectors, and, during exploitation, the whales update their position vectors using the shrinking and spiral procedures in WOA to steer the population towards the best candidate found so far. The adaptive process continues until convergence, enabling the model to automatically find an optimal set of hyperparameters that provides better segmentation accuracy and convergence than a manually tuned model. While the ResNet-50–UNet’s internal training procedure using gradients does not change, WOA is used externally as a hyperparameter-guided meta-optimizer.
| Algorithm 1: Proposed WOA-driven meta-optimization framework integrated with ResNet-50–UNet for adaptive liver tumor segmentation. |
Input:
2: Set a = 2 (WOA control parameter) 3: for t = 1 to T_max do 4: for each whale i = 1 to N do 5: M_i ← Train_ResNet50_UNet (X_train, Y_train, H_i, epochs = 5) 6: Dice_i ← Evaluate_Dice (M_i, X_val, Y_val) 7: Stability_i ← 1 − std(val_loss_last_10_epochs) 8: F_i ← 0.7 × Dice_i + 0.3 × Stability_i 9: end for 10: H* ← argmax_{H_i} F_i (Identify best solution) 11: for each whale i = 1 to N do 12: Update a, A, C, l, p 13: if p < 0.5 then 14: if |A| < 1 then (Exploitation) 15: D ← |C · H* − H_i| 16: H_i ← H* − A · D 17: else 18: H_rand ← Random whale position (Exploitation) 19: D ← |C · H_rand − H_i| 20: H_i ← H_rand − A · D 21: end if 22: else 23: // Spiral update (exploitation) 24: D′ ← |H* − H_i| 25: H_i ← D′ · exp(b·l) · cos(2πl) + H* 26: end if 27: // Ensure bounds 28: H_i ← Clip(H_i, Ω) 29: end for 30: // Linear decrease of a from 2 to 0 31: a ← 2 − 2·t/T_max 32: end for 33: M* ← Train_ResNet50_UNet (X_train, Y_train, H*, epochs = 100) 34: return H*, M* |
3.5.1. Hyperparameter Tuning with Whale Optimization Algorithm
In this study, the hyperparameters for the Whale Optimization Algorithm are set to ensure effective training of the model. These parameters, especially the weights assigned to each loss function, enable precise tuning, which leads to improved performance in liver tumor segmentation. In Table 1, WOA is employed to optimize these hyperparameters, aiming to achieve high accuracy and robustness in identifying liver tumors from medical imaging data.
Table 1.
Hyperparameter descriptions and search ranges used in WOA-based optimization of the LiTs-UNet-WBO Model.
3.5.2. Mathematical Formulation of WOA-Segmentation Coupling
A bi-level optimization framework is formed by combining the Whale Optimization Algorithm with ResNet-50–U-Net, which helps solve the problem of hyperparameter tuning for medical image segmentation. This is because conventional methods are based on single-level optimization, where the weights of the network are determined by gradient descent, while our proposed method adds an additional outer optimization loop that adaptively searches the hyperparameter space and an inner loop to perform normal backpropagation training.
- Problem Formulation:
Let denote the hyperparameter vector consisting of learning rate (η), dropout rate (), and batch size (β). The hyperparameter space is bounded by Ω = [η_min, η_max] × [_min, _max] × [β_min, β_max], where the bounds are defined in Table 1. The optimization objective is formulated as follows:
where the fitness function F(H) evaluates the segmentation quality achieved by the ResNet-50–U-Net model when trained with hyperparameters H.
- Fitness Function Design:
The fitness function combines segmentation accuracy with convergence stability:
where Dice (H): Validation Dice coefficient obtained after training the model with H for E epochs, Stability(H) = 1 − σ(_val): Convergence stability measured by the standard deviation of validation loss over the final 10 epochs, α = 0.7: Weighting parameter that prioritizes segmentation accuracy while ensuring stable convergence.
This multi-objective formulation prevents the optimizer from converging to hyperparameter sets that achieve high validation scores through overfitting or unstable training dynamics.
- Segmentation Loss Function:
For a given hyperparameter set H, the ResNet-50–U-Net model is trained using a composite loss function:
where W is network weights (learned via gradient descent), Predicted segmentation mask, Y is Ground truth mask, _BCE: Binary cross-entropy loss, _Focal is focal loss with γ = 2 to address class imbalance, λ1 = 0.5, λ2 = 0.3: Loss component weights, and the Dice loss component is defined as follows:
where N is the number of pixels, and ε = 1 is a smoothing constant to prevent division by zero.
- WOA-Driven Hyperparameter Update:
In the WOA framework, each whale i corresponds to a candidate hyperparameter solution H_i (t) in iteration t. The total population of N whales is used to search in the hyperparameter space Ω. The way in which a whale updates its position is divided into two behavioral modes:
(1) Encircling Prey (Exploitation): When a promising hyperparameter region is identified (|A| < 1 and p < 0.5), whales converge toward the best solution H*(t):
where C = 2r is random vector with r ∈ [0, 1] that introduces stochastic perturbations.
A = 2a · r − a: Coefficient vector that controls the step size, a: Linear parameter decreasing from 2 to 0 over iterations: a = 2(1 − t/T_max).
(2) Spiral Update (Exploitation): To mimic the spiral swimming pattern of humpback whales (p ≥ 0.5), positions are updated using the following:
where b = 1: Constant defining the logarithmic spiral shape, and l is a random number in [−1, 1] controlling the distance from the spiral center.
(3) Search for Prey (Exploration): When |A| ≥ 1 and p < 0.5, whales explore distant regions by moving toward randomly selected positions:
where H_rand(t) is a randomly selected whale position from the current population.
- Boundary Enforcement:
After each update, hyperparameter values are clipped to ensure feasibility:
This prevents the optimizer from exploring physically meaningless regions (e.g., negative learning rates or batch sizes exceeding memory capacity).
- Convergence Criterion:
The optimization terminates when either of the following occur:
The maximum iteration count T_max = 100 is reached, or The fitness improvement falls below threshold ε_conv = 10−4 for 5 consecutive iterations:
3.6. ResNet-50 + U-Net
Multi-Level ResNet-50 with bottleneck aggregation of the WOA algorithm aim to extract and aggregate deep features and comprise the framework of segmentation. The hierarchical structure is used, and the combination of ResNet-50 blocks enables multi-scale convolutional representation, whereas bottleneck layers are used that enable the reduction of computational expense without sacrificing vital information.
Step 1: Feature Extraction ResNet-50 blocks, containing different receptive fields (e.g., 1 × 1, 3 × 3, 5 × 5) as well as residual learning, processing the input image to facilitate rich spatial encoding and gradient stability. A bottleneck layer follows each block, which entails compressing the features maps and decreasing the channel number, increasing efficiency.
Step 2: Hierarchical Aggregation. The outputs of bottlenecks are hierarchically integrated by over skipping. Feature maps of higher levels are up sampled and put together with lower-level ones to make the possibility to apply deep skip fusion between the levels. This advances improvement in semantic context fusion and localization—an important requirement in medical image segmentation.
Step 3: Final Merging and Production. Each of the aggregated features is concatenated and sent to a set of convolutional layers to result in a refined feature map. This map engulfs the context of the world and the structure of the region, and the segmentation is great.
3.6.1. Encoder
As Figure 10 shows, the encoder component of the U-Net + ResNet-50 model plays a key role in extracting the feature of the input image. The encoder consists of several convolutional layers whose sizes transform the spatial dimensions of the input to smaller sizes and the number of feature maps to higher. This paper takes advantage of the unique U-Net architecture, where the skip connections are used to save the spatial information and underlines its suitability for medical image segmentation. In the normal U-Net architecture, the encoder comprises a few convolutional blocks combined with down-sampling layers. The network also learns various abstraction abilities since there is a multi-convolutional layer within every block. In general, pooling layers (most commonly max pooling) are applied to down-sample the size of feature maps to make the model sensitive to the global structural attributes at a trade-off of local detail. A deep encoder containing ResNet-50 modules improves multi-scale feature extraction ability. The Inception module combines parallel convolutional branches using kernel sizes, thus learning an effective spatial hierarchy. The given method can be especially useful to cover a tumor’s size and shape variation, which one often faces during medical imaging. The convolutional operations provide a global view of the features representation through which the segmentation tasks on tumors are performed with better results. In addition, the encoder positively impacts the efficiency of the segmentation process. A feature swapping structure is also integrated, allowing the feature maps to be compressed before passing them to the subsequent layers, tailoring the encoder to achieve an optimal, balanced trade-off between efficiency and accuracy. Such an architectural solution enables the model and the large size of medical image sets without compromising the quality of the resulting segmentation results, which would apply in clinical use. In short, the encoder of the proposed U-Net + ResNet-50 structure provides a strong and healthy basis in feature extraction, which prepares the informative feature to feed the decoder and segmentation process. The design of the proposed WOA module follows a multi-level encoder–decoder strategy, where modules of the ResNet-50 backbone are deployed at each scale. With some previous connection of stamp bypass, such modules ensure that the context and spatial information are not lost between layers. The down-sampling is performed on input size 128 × 128 images over four encoder stages: 64 × 64, 32 × 32, 16 × 16, 8 × 8. In each step, ResNet-50 modules are applied to obtain local-global feature representations. The features are then pooled into a bottleneck representation before decoding along a symmetric up-sampling path. The decoder path involves skipping connections to the opposite encoder levels and the historical stamp modules, allowing fine-grained details to be recovered. The output is then unraveled to its initial dimension (128 × 128) using a chain of convolution and up-sample layers. The high-order modules used throughout the pipeline are defined in the legend on the bottom right of the figure. This architecture reduces the error of segmentation and error delamination of boundaries and enacts multi-level feature aggregation to guarantee spatial consistencies.
Figure 10.
Overall architecture of the proposed WOA for liver and tumor segmentation. This X* means that the best solution is found during the optimization process.
3.6.2. Decoder
The task of the U-Net + ResNet-50 design decoder is to reassemble the encoded feature representations into the final segmentation mask. Its major goal is the gradual reconstruction and alignments of the downsized feature maps produced by the encoder to the initial input spatial resolution. To tackle this, the decoder utilizes a set of transposed convolutional layers (commonly known as deconvolutions) or convolution procedure and up-sampling layers. Structurally, the decoder is the same as the encoder with reversal of the direction and a series of operations with every up-sampling step doubling the size of the feature maps in spatial relationship. Skip connections with the respective encoder layers are added so that the decoder can have access to high-resolution features that were down sampled during the encoding processes. The linking structures maintain important spatial information that is necessary in the precise location of the tumor boundary in the medical images. Another step taken to mitigate the shortcomings is by adding Inception modules in the decoder to foster greater learning of multi-scale features to refine the segmentation output. Unlike the encoder, ResNet-50 modules that are inbuilt in the decoder support capturing and decoding multiple structure representations of hierarchies that are at various scales. The capability to reconstruct fine-grained details and, simultaneously, stay aware of the wider information in the image, enabled by this dual capability in particular, can be particularly useful when it comes to the detection of irregularly shaped liver tumors. Further, the decoder is also required to reconcile segmentation accuracy and computational efficiency. This trade-off between complexity of the model and overall efficiency can be negotiated well by a careful choice of the number of feature maps to be deployed at each layer. Finally, the decoder within the U-Net + ResNet-50 architecture has a decisive role in enhancing feature representations, which perform a significant role in the generation of high-quality segmentation masks with clinically sound performance.
3.6.3. Bottleneck
The bottleneck in the U-net + ResNet-50 model denotes the point of maximal feature dimensionality between the encoder and the decoder. The stage is important especially because it focuses on the most pertinent concepts of the input data and cuts on less important information. The bottleneck architecture can thereby increase the overall performance of the model in carrying out complicated tasks like the segmentation of liver tumors. In the U-Net + ResNet-50 system, the bottleneck retains only a small number of feature maps by combining Inception modules, guaranteeing that important features are maintained. The 1 × 1 convolutions of these modules allow it to create significant compressions with little information loss as a threat. In addition to dealing with the resource limitations, this dimensionality reduction also empowers the model to obtain generalization of previously unseen data. The bottleneck is such a place where features retrieved at several abstraction levels are integrated. The tradeoff between computational complexity and performance is created by the design itself. When examined architecturally, one tends to think of two possible paths: scale the number of Inception modules or scale the number of feature maps per module. Its goal is to keep the bottleneck appropriately restrictive and to keep the network, thus, the power of its representations. Finally, it is the bottleneck structure in the proposed U-Net + ResNet-50 architecture that is of central importance to achieve better segmentation accuracy especially in the detection of liver tumors.
3.6.4. ResNet-50 Backbone
The backbone of U-Net, which in this architecture is ResNet-50, constitutes the main feature extractor and directly impacts the segmentation quality []. The backbone plays the role of the structural base network, in which the first features are obtained before the other functionalities of the U-Net model are carried out. Backbone choice is also important to establish whether the model can work in many imaging modalities, which is important, especially in the medical community, where segmentation of tumors of the liver is important. The ResNet-50 is particularly good as a backbone due to the multiple parallel convolutional paths exploring contrasting kernel sizes in the network. This design will allow the network to extract a wide collection of feature representations and be better positioned to realize the structure size and position variations. Meanwhile, the backbone should not have excessive computational overhead. Using transfer learning, our U-Net + ResNet-50 model is more likely to converge to higher accuracy in fewer training epochs using pre-trained weights on large-scale datasets. This property’s usefulness is particularly high in medical imaging applications where labeled datasets are sparse. Nonetheless, not all U-Net-based frameworks would take optimal advantage of using hierarchical feature representation, which Inception modules offer. This drawback may restrict the model from detecting complicated relationships between parts of an image. This capability is essential where the different types of liver tumors may have lesions of identical morphological similarities. The suggested U-Net + ResNet-50 rides this by incorporating Inception modules, thus it presents a good backbone that can support high performance and advanced clinical applicability. The support of Inception modules alleviates several general issues in medical picture separation. This ability to obtain multi-scale is particularly beneficial, as tumors occur in various shapes and sizes. No one size of kernel can be expected to work well with this diversity. The ResNet-50 backbone can analyze tumor features on many scales, as each branch has convolutional layers enacted in parallel with kernels of varying sizes. Also, 1 × 1 convolutions in the Inception modules perform dimensionality reduction. It minimizes parameters with necessary information being maintained—a property that is extremely useful in medical imaging terms and scenarios where computational facilities are not always available. This reduces the computational requirements by optimizing pattern recognition with no performance loss and facilitates the applications that demand real-time clinical deployment, such as ResNet-50. Moreover, the Inception modules improve feature combination by combining the information across the convolutional branches, which makes the abstract representation of tumor features stronger. Such a mechanism enhances the capability of acquiring dense details on liver tumors, which results in greater segmentation accuracy. The hierarchical representation of features that the Inception design offers further enhances the information flow in the model, allowing more detailed outputs and further improvement in segmentation quality overall.
3.7. Explainable AI (XAI)
Explainable AI (XAI) [] on the other hand is prevalent in medical image segmentation, where the credibility of a model and its transparency are crucial in medical informatics shown in Figure 10. Two commonly used methods of achieving interpretability in image segmentation tasks are gradient-weighted Class Activation Mapping (Grad-CAM) and LIME (Local Interpretable Model-agnostic Explanations). Grad-CAM, a local interpretation, computes gradients of the target class with respect to the ultimate convolutional layer feature maps, and generates a heatmap, which indicates the regions of an image that contribute the most to a segmentation choice. The Grad-CAM model at pixel Ak represents a class score y as follows:
where is the weight of each feature map , and the ReLU function guarantees that only positive contributions are considered, indicating the informative regions for segmentation.
SHAP provides post hoc explanations to the predictions of a model by attributing an importance to each feature (e.g., pixel or region) that contributes to the model’s decision. For segmentation, SHAP highlights assist in pinpointing where in an image is positioned the strongest effect on segmentation decision for tumor and non-tumor pixels, and thus establishes clinical trust and model transparency. SHAP is based on cooperative game theory and computes the Shapley value for each input feature. The value for feature in the Shapley value is defined as below:
where N is the number of input features, S is a sub-set of N with the absence of feature i, f(S) is the value of the model’s prediction when using only the features that are related to S. In the case of image segmentation, f(S) is the model’s predicted probability that a pixel is in the tumor class based on only the features in subset S. SHAP values therefore measure the incremental value of each pixel or region tor the final prediction.
In this study, we used SHAP and Grad-CAM because they provide a more accurate pixel-level explanation, which is what we need for segmentation. In contrast to Grad-CAM’s rough heatmaps, SHAP consistently provides (theoretically motivated) explanations of how much each pixel affects the model’s prediction. Figure 11 shows explainable AI analysis.
Figure 11.
Explainable AI analysis.
3.8. Performance Measures
This paper was used to verify the correctness of liver lesion segmentation by the proposed hybrid algorithm and to compare it with the state-of-the-art algorithm based on three indices: the Jaccard index [], Dice coefficient [], and correlation coefficient [].
Jaccard index: This is the similar index of binary data below where AOO is the overlapping area, M is a binary image, and K is a ground truth image:
Dice Index: This is a co-efficient to compute segmentation. The Dice coefficient value is the fraction of the predicted image pixels that align with the ground truth pixel-wise:
Accuracy: Accuracy is the likelihood of hit prediction, and it can be determined:
4. Experimental Results and Discussions
4.1. Implementation
The architecture was deployed with PyTorch 2.1.1. All the training and experimental tests were carried out on a personal workstation using the NVIDIA GeForce RTX 3090 (NVIDIA, Santa Clara, CA, USA) and the Intel(R) Core (TM) i7-10700KF CPU 3.80 GHz (32 GB) (Taiwan Semiconductor Manufacturing Company (TSMC), Hsinchu, Taiwan). Some other segmentation models were used as a comparison, such as Swin Transformer and the EfficientNetB0-UNet. CT images were resized from 512 × 512 to 128 × 128 pixels, intensity normalized to the [0, 1] range, and underwent liver-specific windowing. Random rotation, flip, and intensity shift were used for data augmentation. We used PyTorch 2.1 with MONAI 1.1 and standard scientific Python libraries (NumPy 1.25, SciPy 1.11). The Whale Optimization Algorithm (WOA) was used for hyperparameter optimization. Hyperparameter optimization was conducted using a population size of 20 and the number of epochs = 100.
On the computational side, WOA-based optimization converged in five outer iterations, with a population size of 20 whales. This resulted in 14 percent less total training epochs before the validation accuracy began to stabilize in comparison to the hyperparameter values selected by hand with either Adam. The marginal training overhead introduced by the WOA was negligible since it is only used during the hyperparameter search and does not impact the prediction/inference time. We report an average per-slice inference latency of 14.3 (ms) for a single 2D CT slice on an RTX 3090 GPU, which further validates that the proposed WOA integration speeds up the training process but does not increase the prediction time. The overall complexity of the algorithm is in the order of the number of whales times the number of hyperparameters, which is low.
4.2. Ablation Study
We performed a detailed ablation experiment to further verify the role of each part of our LiTs-Res-UNet + WOA architecture. Table 2 shows comparative analysis of the segmentation results of the proposed four model variants is presented. The baseline U-Net model demonstrated relatively moderate accuracy (Dice: 78.45%, Jaccard: 64.67%). The inclusion of WOA to enhance model hyperparameters resulted in improved performance in terms of Dice (82.13%) and Jaccard (69.85%), confirming the effectiveness of the metaheuristic optimization algorithm. When applying the ResNet-50 backbone, the results continued to increase (Dice: 87.66%, Jaccard: 78.24%). However, it is important to note that this improvement was accompanied by a significant increase in model complexity and training time. In comparison, the proposed LiTs-Res-UNet + WOA model achieved the highest level of accuracy, with Dice: 92.38%, Jaccard: 86.73%, and Pixel Accuracy: 99.54%, while maintaining an acceptable and efficient training time, which confirms the potential of combining the above-mentioned techniques for liver tumor segmentation.
Table 2.
Ablation study: component-wise performance analysis.
4.3. Statistical Significance Analysis
We measured the relative contribution of architectural changes and hyperparameter optimization, and we conducted paired t-tests for each model variant against the baseline in Table 3. Contributions of WOA Optimization (U-Net vs. U-Net + WOA): After integrating WOA, Dice increased by +3.68% (p = 0.003) and Jaccard by +5.18% (p = 0.002), and training time decreased by 13.4%, suggesting that WOA optimization alone leads to notable performance gains and improved training efficiency. Contributions of ResNet-50 Encoder (U-Net vs. ResNet-50 U-Net): Replacing the default encoder with ResNet-50 resulted in significant segmentation quality improvements: Dice increased by +9.21% (p < 0.001) and Jaccard by +13.57% (p < 0.001), demonstrating the benefits of deeper feature extraction. Additive Effect (ResNet-50 U-Net vs. LiTs-Res-UNet + WOA): Hyperparameter tuning of ResNet-50 U-Net with WOA led to additional improvements of +4.72% in Dice (p = 0.001) and +8.49% in Jaccard (p < 0.001), along with a 13.6% reduction in training time compared to manual parameter tuning. This demonstrates that the combination of advanced architecture and metaheuristic optimization results in statistically significant gains beyond the individual components.
Table 3.
Statistical validation of performance improvements.
4.4. Comparison
Table 4 reports the results of some of the latest algorithms in liver tumor segmentation. Classical multi-scale methods [] and PGC-Net [] record Dice scores of 74.8 and 73.63, respectively, whereas PAKS-Net [] enhances the performance to 76.9. The introduction of sophisticated algorithms like MSFF, MFF, EI, and EG improves the accuracy of segment outputs to 85.55 Dice and 81.11 Jaccard []. Context Fusion Networks in Twin-Split Attention and MSA skip connections [] perform the same with 85.97% Dice and 81.56% Jaccard, and a hybrid framework, FasNet [], accomplishes 87.66% Dice and 84.87% Jaccard. LiTs-Res-UNet shows 79.25 Dice and 65.88 Jaccard is competitive as a baseline result. Optimized with the Whale Optimization Algorithm (WOA), the model achieves 92.38% Dice and 86.73% Jaccard, which, in turn, are far better than those of the existing state-of-the-art methods. Notably, these findings confirm the effective combination of deep residual learning with metaheuristic optimization in liver tumor segmentation.
Table 4.
Comparison of liver tumor segmentation techniques.
Training and validation accuracy versus epochs of Figure 12a LiTs-Res-UNet that includes and does not include the Whale Optimization Algorithm (WOA). The optimized Figure 12b LiTs-Res-UNet + WOA shows faster converging at the peak of training accuracy to 0.99 and stable validation accuracy of around 0.987 at Epoch 4, thus demonstrating improved generalization and lack of fluctuation. Compared to it, the baseline LiTs-Res-UNet demonstrates slightly lower training (0.988) and validation (0.985) accuracy values with minor validation instability. The overall outcomes indicate that WOA integration promotes convergence velocity and strengthens robustness in liver tumor segmentation.
Figure 12.
Training and validation accuracy over epochs. (a) LiTs-Res-UNet; (b) LiTs-Res-UNet + WOA.
4.5. Analysis Using Different Optimizers
Table 5 presents the comparative performance of various optimizers: AdaGrad, SGD, Adam, RMSProp, AdaDelta, and a Metaheuristic Optimizer, in terms of the four important evaluation metrics: Loss, Accuracy, Dice Coefficient, and Jaccard Index. The obtained results show that the selection of optimizers has a significant effect on model convergence and segmentation performance. Among the conventional gradient-based optimizers, Adam showed better performance compared to the others with the lowest loss (0.0251) and the highest accuracy (0.9954), Dice coefficient = 0.8766, and Jaccard index = 0.8487. This is suggestive of Adam’s ability to balance convergence speed and generalization, which makes it a good candidate for segmentation tasks on complex data distributions. In contrast, SGD and RMSProp exhibited relatively poor results, with lower accuracy (0.9344 and 0.9231, respectively) and reduced overlap-based metrics. AdaGrad and AdaDelta had similar accuracy (0.9531 and 0.9540, respectively) to Adam; however, they presented higher loss values (0.0435 and 0.3119) and lower Dice/Jaccard scores, indicating limitations in fine-grained region segmentation. The most significant improvement was attained using the Metaheuristic Optimizer, which outperformed all conventional optimizers across all evaluation criteria. It obtained the lowest loss (0.0121), the highest Dice coefficient (0.9238), and Jaccard index (0.8673). The Whale Optimization Algorithm (WOA) achieves the highest Dice and Jaccard scores, with an improvement of 4.72% and 2.15%, respectively, compared to the best-performing baseline optimizer, Adam. Other optimizers show negative or lower improvements relative to this baseline, which are also reported in the table. These results demonstrate the optimizer’s ability to better traverse complex, non-convex optimization landscapes than gradient-based approaches, resulting in improved pixel-level correspondence between the predicted and ground truth masks. Overall, the results indicate that Adam remains a strong baseline optimizer; however, the adoption of metaheuristic-based optimization strategies yields substantial improvements in segmentation performance. This strengthens the promise of hybrid or non-classical optimization algorithms for medical image analysis, where accuracy and structural similarity are crucial.
Table 5.
Performance comparison of different optimizers on liver tumor segmentation.
4.6. Cross-Dataset Generalization Study
In Table 6, our proposed model achieved encouraging liver tumor segmentation performance on the various datasets, confirming its ability to generalize to other datasets. When tested on the LiTS2017 dataset for primary training, it reached a Dice coefficient of 92.38 ± 0.76% and a Jaccard index of 86.73 ± 0.89%. On external datasets, it demonstrated slightly lower but comparable performance: 87.34 ± 2.15% (Dice) and 80.67 ± 2.48% (Jaccard) on 3DIRCADb, and 85.78 ± 2.87% (Dice) and 78.92 ± 3.21% (Jaccard) on the CHAOS Challenge dataset. This observation implies that the model can effectively segment tumors while maintaining high accuracy across datasets with varying scanners and imaging protocols.
Table 6.
Cross-dataset generalization performance.
4.7. Explainable AI Analysis
Figure 13 demonstrates GradCAM visualization for liver tumor segmentation. (Left) Input CT image showing axial abdominal slice with liver anatomy. (Right) GradCAM heatmaps for the various segmentation classes: Background, Liver, Tumor, Combined. GradCAM heatmaps indicate the areas of the model’s attention for the corresponding class prediction. The colormap shows the magnitude of activation intensities (thermal color scale). The Background class attention map primarily highlights the boundaries of non-liver structures. The Liver class attention map focuses on activation over the liver parenchyma with the highest activation in the center. The Tumor class attention map has a localized activation over the areas of suspected pathology in the liver. The Combined map visualizes all class-specific attention maps.
Figure 13.
Grad CAM visualization for multi-class liver tumor segmentation.
Figure 14 shows the interpretability of our proposed model LiTs-Res-UNet + WOA with SHapley Additive exPlanations (SHAP). The first row shows the original CT image (Left) and the SHAP explanation map (Right). The blue-highlighted area represents regions identified by the model as important for the prediction of liver and tumor boundaries. The second row shows the original CT image (Left) and the SHAP overlay (Right), where the high-intensity (yellow to red) areas correspond to the pixels with higher positive SHAP values and our model is highly confident about the tumor-related characteristics. The pixels are perfectly aligned with the pathological regions identified by the radiologists, while the model’s segmentation decisions are primarily influenced by hepatic and tumorous structures, indicating the model’s understanding and focus on clinically relevant features, offering increased transparency into the model’s decision-making process via quantitative and spatial interpretation.
Figure 14.
SHAP-based Explainable AI visualization of liver and tumor segmentation.
5. Discussions
The key novelty of this paper is the coupling of a bio-inspired global optimization algorithm with a deep convolutional encoder–decoder network. Instead of using grid search or Bayesian optimization techniques to tune hyperparameters of a deep learning network, the WOA algorithm adaptively and dynamically samples the global hyperparameter space in a stochastic search process, which is based on the hunting behavior of humpback whales. The global search, along with the gradient-based training of the ResNet-50–UNet, adaptively controls the network parameters, such as learning rate, dropout, and batch size. It can avoid local minimums and ensure stable convergence of the parameters. The coupling of a metaheuristic algorithm with a deep learning network can potentially enhance the segmentation accuracy and convergence efficiency in medical image analysis.
In this study, we developed and applied state-of-the-art and high-performing models–Hybrid LiTS-UNet-WOA for accurate segmentation of liver tumors and other infected regions of interest (ROIs) using the standard medical dataset like LiTS. The architecture Hybrid LiTs-Res-Unet + WOA is to introduce residual connections into the conventional UNet framework for increased depth of feature propagation, for better learning efficiency, and to maintain spatial and context information more efficiently. The model can thus well represent the irregular shapes and sizes that are often observed in liver tumors, which will improve the performance of segmentation. On the other hand, the LiTs-Res-Unet + WOA model combines WOA with UNet architecture to well-tune the feature selection process and network parameters. This optimization results in more accurate lesion detection with less computational complexity and learning time. By focusing more on meaningful characteristics in the learning process, LiTs-Res-Unet + WOA can enhance overall detection reliability and reduce the possibility of overfitting and false negatives. Both models have been validated on a variety of medical imaging datasets and present better pixel-wise segmentation accuracy than conventional methods. These deep learning-based methods excel traditional machine learning methods by automatically capturing features and requiring less human labeling and handling data at a larger scale, even when it is in low quantity. Second, combined with the presence of transfer learning, in particular the use of pre-trained backbones such as ResNet-50, also helps accelerate the learning process of both models. Through pretraining the networks with pre-trained weights, these architectures do not rely on a random start but maximize diagnostic accuracy before the occurrence of finite convergence. This is especially apparent in Hybrid LiTs-Res-Unet + WOA due to residual connections and deep convolutional layers, as WOA can allow the most salient features, useful for liver tumor identification, to be selected. Overall, the Hybrid LiTs-Res-Unet + WOA model offers a very scalable, accurate, and computationally efficient architecture for liver tumor segmentation. WOA does have extra training overheads; however, it is low-cost relative to the considerable improvements in convergence rate, robustness, and segmentation accuracy, such that WOA is feasible for use in clinical applications. The generalizability of diverse tumor morphologies and datasets highlights the potential for their clinical use in diagnostic systems. However, Grad-CAM or SHAP may be used, pointing to important areas to enable radiologists to check the results of segmentation. All in all, this discussion justifies the real-world implementation of our models into clinical practice, as they are accurate and efficient in liver tumor segmentation and are also transparent and computationally efficient.
6. Limitation
Our proposed LiTs-Res-UNet + WOA demonstrated strong performance in liver tumor segmentation; the training and validation were mainly conducted on publicly available benchmark datasets, including LiTS2017 and Radiopaedia. These datasets, although diverse, may not fully represent the variations seen in multi-center clinical deployments, including differences in imaging protocols, scanner types, and patient demographics. Thus, further validation on large-scale, multi-institutional datasets is essential to ensure broad generalizability. In this study, we resized all the CT images to 128 × 128 pixels as part of preprocessing. This was done to decrease the computational requirements and accelerate the training process of the model. The lower resolution also made it possible to conduct a larger number of experiments in a shorter time frame. However, this may have also resulted in the loss of some detailed anatomical information which could be important for accurate segmentation of liver tumors. Future work will investigate the use of higher-resolution images to capture more detailed diagnostic information. 2D Slice-Based Segmentation: The current framework operates on 2D axial slices extracted from 3D CT volumes. While this approach reduces computational complexity, it can lead to a loss of volumetric contextual information, which can be helpful for accurately segmenting tumors with irregular boundaries. Extending the model to perform 3D volumetric segmentation could further enhance its clinical applicability. Then, the use of the Whale Optimization Algorithm for hyperparameter tuning and model performance enhancement is computationally expensive, particularly during the training phase. This computational demand may limit the scalability of the proposed approach in real-time or resource-constrained clinical settings. The evaluation of model performance was primarily based on standard segmentation metrics such as Dice coefficient, Jaccard index, and pixel accuracy. While these metrics are commonly used, other metrics, such as Cohen’s Kappa or Matthews Correlation Coefficient (MCC), could be considered to provide additional insights into model performance, particularly in imbalanced datasets with disproportionately small tumor regions.
7. Conclusions
A state-of-the-art performance in liver tumor segmentation was established by proposing a new LiTs-Res-Unet + WOA architecture. The proposed model achieves superior segmentation accuracy of Dice coefficient 92.4% and Jaccard index 86.7%, which are higher by 4.7% and 2.0%, respectively, compared to the previous state-of-the-art methods. The incorporation with classical U-Net topology as well as novelty multi-scale feature fusion and attention mechanisms, not only considerably improves the capability of capturing local anatomical details but also effectively captures global contextual information for the precise delineation of tumor boundaries. The clinical implications of these findings are far-reaching since the attained performances approach the range of inter-observer agreement in clinical practice, illustrating a strong potential of practical applicability in hepatic oncology workflows. Results from the different evaluation metrics consistently show that our method is robust and reliable in comparison with others, and we achieve significant performance improvement over baselines, which validates the effectiveness of our architectural innovations. These advances make a meaningful contribution to the field of medical image analysis and achieve a substantial step towards automated accurate liver tumor segmentation, which in turn could significantly improve diagnostic precision and treatment planning, and ultimately patient outcomes. Work will continue with multi-center validation, computational optimization for clinical translation, and investigation into the generalization of the model for other hepatic pathology segmentation tasks so that this technology can be translated into the clinic.
Author Contributions
Conceptualization, P.K.M. and M.A.I.M.; methodology, P.K.M., M.A.I.M. and M.H.A.A.-O.; software, validation, P.K.M. and M.A.I.M.; formal analysis, P.K.M. and M.A.I.M.; investigation, H.C.K.; resources, H.C.K.; data curation, writing—original draft preparation, writing—review and editing, P.K.M., M.A.-B., M.H.A.A.-O., D.S.M.H., M.S.A.M. and M.A.I.M.; visualization, P.K.M. and M.A.I.M.; supervision, H.C.K.; project administration, H.C.K., M.H.A.A.-O. and D.S.M.H.; funding acquisition, M.H.A.A.-O., M.A.-B. and D.S.M.H. All authors have read and agreed to the published version of the manuscript.
Funding
This paper is supported by the Middle East University in Amman, Jordan, for providing financial support to cover the publication fees associated with this research article, and this paper is also supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R751), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The LiTs2017 dataset used for this study is publicly available at https://www.kaggle.com/datasets/andrewmvd/liver-tumor-segmentation-part-2 (accessed on 5 July 2025).
Acknowledgments
The authors express their gratitude to the Middle East University in Amman, Jordan, for providing financial support to cover the publication fees associated with this research article. The authors also thank Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R751), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Almotairi, S.; Kareem, G.; Aouf, M.; Almutairi, B.; Salem, M.A.-M. Liver tumor segmentation in CT scans using modified SegNet. Sensors 2020, 20, 1516. [Google Scholar] [CrossRef] [PubMed]
- Zhang, J.; Xie, Y.; Zhang, P.; Chen, H.; Xia, Y.; Shen, C. Light-Weight Hybrid Convolutional Network for Liver Tumor Segmentation. IJCAI 2019, 19, 4271–4277. [Google Scholar] [CrossRef]
- Jiang, H.; Yin, Y.; Zhang, J.; Deng, W.; Li, C. Deep learning for liver cancer histopathology image analysis: A comprehensive survey. Eng. Appl. Artif. Intell. 2024, 133, 108436. [Google Scholar] [CrossRef]
- Velichko, Y.S.; Gennaro, N.; Karri, M.; Antalek, M.; Bagci, U. A comprehensive review of deep learning approaches for magnetic resonance imaging liver tumor analysis. Adv. Clin. Radiol. 2023, 5, 1–15. [Google Scholar] [CrossRef]
- Nanda, N.; Kakkar, P.; Nagpal, S. Computer-Aided Segmentation of Liver Lesions in CT Scans Using Cascaded Convolutional Neural Networks and Genetically Optimised Classifier. Arab. J. Sci. Eng. 2019, 44, 4049–4062. [Google Scholar] [CrossRef]
- Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
- Hu, C.; Xia, T.; Cui, Y.; Zou, Q.; Wang, Y.; Xiao, W.; Ju, S.; Li, X. Trustworthy multi-phase liver tumor segmentation via evidence-based uncertainty. Eng. Appl. Artif. Intell. 2024, 133, 108289. [Google Scholar] [CrossRef]
- Fang, L.; Jin, Y.; Huang, L.; Guo, S.; Zhao, G.; Chen, X. Iterative fusion convolutional neural networks for classification of optical coherence tomography images. J. Vis. Commun. Image Represent. 2019, 59, 327–333. [Google Scholar] [CrossRef]
- Saadh, M.J.; Bishoyi, A.K.; Ballal, S.; Singh, A.; Devi, A.; Sharma, G.C.; Valiev, A.; Naidu, K.S.; Bhakuni, P.N.; Sead, F.F. Ferroptosis and non-coding RNAs in breast cancer: Insights into CAF and TAM interactions. Discov. Oncol. 2025, 16, 1658. [Google Scholar] [CrossRef]
- Chelloug, S.A.; Mahel, A.S.B.; Alnashwan, R.; Rafiq, A.; Muthanna, M.S.A.; Aziz, A. Enhanced breast cancer diagnosis using modified InceptionNet-V3: A deep learning approach for ultrasound image classification. Front. Physiol. 2025, 16, 1558001. [Google Scholar] [CrossRef] [PubMed]
- Shen, W.; Zhou, M.; Yang, F.; Yu, D.; Dong, D.; Yang, C.; Zang, Y.; Tian, J. Multi-crop convolutional neural networks for lung nodule malignancy suspiciousness classification. Pattern Recognit. 2017, 61, 663–673. [Google Scholar] [CrossRef]
- Ghoniem, R.M.; Shaalan, K. FCSR—Fuzzy Continuous Speech Recognition Approach for Identifying Laryngeal Pathologies Using New Weighted Spectrum Features. In Advances in Intelligent Systems and Computing, Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2017, Cairo, Egypt, 9–11 September 2017; Hassanien, A.E., Shaalan, K., Gaber, T., Tolba, M.F., Eds.; Springer International Publishing: Cham, Switzerland, 2018; Volume 639, pp. 384–395. [Google Scholar] [CrossRef]
- Marini, F.; Walczak, B. Particle swarm optimization (PSO). A tutorial. Chemom. Intell. Lab. Syst. 2015, 149, 153–165. [Google Scholar] [CrossRef]
- Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
- Bansal, J.C.; Gopal, A.; Nagar, A.K. Stability analysis of artificial bee colony optimization algorithm. Swarm Evol. Comput. 2018, 41, 9–19. [Google Scholar] [CrossRef]
- Maione, G.; Punzi, A.; Li, K. A comparative study on differential evolution with other heuristic methods for continuous optimization. In Proceedings of the 21st Mediterranean Conference on Control and Automation, Platanias, Greece, 25–28 June 2013; IEEE: New York, NY, USA, 2013; pp. 1356–1361. Available online: https://ieeexplore.ieee.org/abstract/document/6608896/ (accessed on 16 June 2025).
- Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A new heuristic optimization algorithm: Harmony search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
- Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. GSA: A gravitational search algorithm. Inf. Sci. 2009, 179, 2232–2248. [Google Scholar] [CrossRef]
- Emary, E.; Zawbaa, H.M.; Grosan, C.; Hassenian, A.E. Feature Subset Selection Approach by Gray-Wolf Optimization. In Afro-European Conference for Industrial Advancement, Advances in Intelligent Systems and Computing; Abraham, A., Krömer, P., Snasel, V., Eds.; Springer International Publishing: Cham, Switzerland, 2015; Volume 334, pp. 1–13. [Google Scholar] [CrossRef]
- Mostafa, A.; Houssein, E.H.; Houseni, M.; Hassanien, A.E.; Hefny, H. Evaluating Swarm Optimization Algorithms for Segmentation of Liver Images. In Advances in Soft Computing and Machine Learning in Image Processing, Studies in Computational Intelligence; Hassanien, A.E., Oliva, D.A., Eds.; Springer International Publishing: Cham, Switzerland, 2018; Volume 730, pp. 41–62. [Google Scholar] [CrossRef]
- Decerle, J.; Grunder, O.; El Hassani, A.H.; Barakat, O. A hybrid memetic-ant colony optimization algorithm for the home health care problem with time window, synchronization and working time balancing. Swarm Evol. Comput. 2019, 46, 171–183. [Google Scholar] [CrossRef]
- Garro, B.A.; Rodríguez, K.; Vázquez, R.A. Classification of DNA microarrays using artificial neural networks and ABC algorithm. Appl. Soft Comput. 2016, 38, 548–560. [Google Scholar] [CrossRef]
- Qi, A.-L.; Zhang, G.-M.; Dong, M.; Ma, H.-W.; Harvey, D.M. An artificial bee colony optimization based matching pursuit approach for ultrasonic echo estimation. Ultrasonics 2018, 88, 1–8. [Google Scholar] [CrossRef] [PubMed]
- Thiagarajan, B.; Bremananth, R. Brain image segmentation using conditional random field based on modified artificial bee colony optimization algorithm. Recent Adv. Comput. Sci. 2014, 344, 93–106. [Google Scholar]
- Mostafa, A.; Fouad, A.; Elfattah, M.A.; Hassanien, A.E.; Hefny, H.; Zhu, S.Y.; Schaefer, G. CT liver segmentation using artificial bee colony optimisation. Procedia Comput. Sci. 2015, 60, 1622–1630. [Google Scholar] [CrossRef]
- Liu, X.; Wang, X.; Shi, N.; Li, C. Image segmentation algorithm based on improved ant colony algorithm. Int. J. Signal Process. Image Process. Pattern Recognit. 2014, 7, 433–442. [Google Scholar] [CrossRef]
- Emon, M.H.; Mondal, P.K.; Mozumder, M.A.; Kim, H.C.; Lapina, M.; Babenko, M.; Muthanna, M.S. An Integrated Architecture for Colorectal Polyp Segmentation: The µ-Net Framework with Explainable AI. Diagnostics 2025, 15, 2890. [Google Scholar] [CrossRef]
- Horng, M.-H. Multilevel thresholding selection based on the artificial bee colony algorithm for image segmentation. Expert Syst. Appl. 2011, 38, 13785–13791. [Google Scholar] [CrossRef]
- Chen, L.; Luo, S. Multi-Scale Liver Tumor Segmentation Algorithm by Fusing Convolution and Transformer. J. Comput. Eng. Appl. 2024, 60, 4. Available online: https://search.ebscohost.com/login.aspx?direct=true&profile=ehost&scope=site&authtype=crawler&jrnl=10028331&AN=175598615&h=HRR4o5hklnz6GycZGS81hvavLVl1CdqCk8KUsgjapy9BP4k7ihKB0qeY7qHwjlSSYKtl58xhRF27xcR9%2BG5iWA%3D%3D&crl=c (accessed on 23 June 2025).
- You, Y.; Bai, Z.; Zhang, Y.; Li, Z. Contour-induced parallel graph reasoning for liver tumor segmentation. Biomed. Signal Process. Control. 2024, 92, 106111. [Google Scholar] [CrossRef]
- He, J.; Luo, Z.; Lian, S.; Su, S.; Li, S. Towards accurate abdominal tumor segmentation: A 2D model with Position-Aware and Key Slice Feature Sharing. Comput. Biol. Med. 2024, 179, 108743. [Google Scholar] [CrossRef]
- Mondal, P.K.; Mozumder, M.A.; Kim, H.C. Deep Learning Framework for Automated Detection of Diabetic Retinopathy in Retinal Fundus Images. Korea Inst. Inf. Commun. Eng. 2025, 29, 511–514. [Google Scholar]
- Shui, Y.; Wang, Z.; Liu, B.; Wang, W.; Fu, S.; Li, Y. A three-path network with multi-scale selective feature fusion, edge-inspiring and edge-guiding for liver tumor segmentation. Comput. Biol. Med. 2024, 168, 107841. [Google Scholar] [CrossRef]
- Wang, Z.; Zhu, J.; Fu, S.; Ye, Y. Context fusion network with multi-scale-aware skip connection and twin-split attention for liver tumor segmentation. Med. Biol. Eng. Comput. 2023, 61, 3167–3180. [Google Scholar] [CrossRef] [PubMed]
- Muhammad, S.; Zhang, J. Segmentation of Liver Tumors by Monai and PyTorch in CT Images with Deep Learning Techniques. Appl. Sci. 2024, 14, 5144. [Google Scholar] [CrossRef]
- Fallahpoor, M.; Nguyen, D.; Montahaei, E.; Hosseini, A.; Nikbakhtian, S.; Naseri, M.; Salahshour, F.; Farzanefar, S.; Abbasi, M. Segmentation of liver and liver lesions using deep learning. Phys. Eng. Sci. Med. 2024, 47, 611–619. [Google Scholar] [CrossRef]
- Wang, J.; Peng, Y.; Jing, S.; Han, L.; Li, T.; Luo, J. A deep-learning approach for segmentation of liver tumors in magnetic resonance imaging using UNet++. BMC Cancer 2023, 23, 1060. [Google Scholar] [CrossRef]
- Zhen, S.-H.; Cheng, M.; Tao, Y.-B.; Wang, Y.-F.; Juengpanich, S.; Jiang, Z.-Y.; Jiang, Y.-K.; Yan, Y.-Y.; Lu, W.; Lue, J.-M.; et al. Deep learning for accurate diagnosis of liver tumor based on magnetic resonance imaging and clinical data. Front. Oncol. 2020, 10, 680. [Google Scholar] [CrossRef]
- Zahra, S.B.; Athar, A.; Khan, M.A.; Abbas, S.; Ahmad, G. Automated diagnosis of liver disorder using multilayer neu-ro-fuzzy. Int. J. Adv. Appl. Sci. 2019, 6, 23–32. [Google Scholar]
- McGrath, H.; Li, P.; Dorent, R.; Bradford, R.; Saeed, S.; Bisdas, S.; Ourselin, S.; Shapey, J.; Vercauteren, T. Manual segmentation versus semi-automated segmentation for quantifying vestibular schwannoma volume on MRI. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1445–1455. [Google Scholar] [CrossRef] [PubMed]
- Mojtahed, A.; Núñez, L.; Connell, J.; Fichera, A.; Nicholls, R.; Barone, A.; Marieiro, M.; Puddu, A.; Arya, Z.; Ferreira, C.; et al. Repeatability and reproducibility of deep-learning-based liver volume and Couinaud segment volume measurement tool. Abdom. Imaging 2021, 47, 143–151. [Google Scholar] [CrossRef] [PubMed]
- Wang, Y.; Xu, Y.; Song, J.; Liu, X.; Liu, S.; Yang, N.; Wang, L.; Liu, Y.; Zhao, Y.; Zhou, W.; et al. Tumor Cell-Targeting and Tumor Microenvironment–Responsive Nanoplatforms for the Multimodal Imaging-Guided Photodynamic/Photothermal/Chemodynamic Treatment of Cervical Cancer. Int. J. Nanomed. 2024, ume 19, 5837–5858. [Google Scholar] [CrossRef]
- Venkadesh, K.V.; Aleef, T.A.; Scholten, E.T.; Saghir, Z.; Silva, M.; Sverzellati, N.; Pastorino, U.; van Ginneken, B.; Prokop, M.; Jacobs, C. Prior CT Improves Deep Learning for Malignancy Risk Estimation of Screening-detected Pulmonary Nodules. Radiology 2023, 308, e223308. [Google Scholar] [CrossRef]
- Alzahrani, A.A.; Ahmed, A.; Raza, A. Content-Based Medical Image Retrieval Method Using Multiple Pre-Trained Convolutional Neural Networks Feature Extraction Models. 2024. Available online: https://www.researchgate.net/profile/Ali-Ahmed-46/publication/382005755_Content-based_medical_image_retrieval_method_using_multiple_pre-trained_convolutional_neural_networks_feature_extraction_models/links/668c4975b15ba55907497936/Content-based-medical-image-retrieval-method-using-multiple-pre-trained-convolutional-neural-networks-feature-extraction-models.pdf (accessed on 16 June 2025).
- Singh, R.; Gupta, S.; Almogren, A.; Rehman, A.U.; Bharany, S.; Altameem, A.; Choi, J. FasNet: A hybrid deep learning model with attention mechanisms and uncertainty estimation for liver tumor segmentation on LiTS17. Sci. Rep. 2025, 15, 1–19. [Google Scholar] [CrossRef]
- Biswas, A.; Maity, S.P.; Banik, R.; Bhattacharya, P.; Debbarma, J. GAN-Driven liver tumor segmentation: Enhancing accuracy in biomedical imaging. SN Comput. Sci. 2024, 5, 652. [Google Scholar] [CrossRef]
- Liver Tumor Segmentation—Part 2. Available online: https://www.kaggle.com/datasets/andrewmvd/liver-tumor-segmentation-part-2 (accessed on 20 October 2025).
- Medicine. Available online: https://journals.lww.com/md-journal/fulltext/2025/05300/mammogram_mastery__breast_cancer_image.30.aspx?context=latestarticles (accessed on 16 June 2025).
- Deep Learning Based Liver Cancer Detection Using Watershed Transform and Gaussian Mixture Model Techniques—Sci-enceDirect. Available online: https://www.sciencedirect.com/science/article/pii/S1389041718310143 (accessed on 16 June 2025).
- Raza, M.; Farwa, U.E.; Mozumder, M.A.; Mon-il, J.; Kim, H.C. ETRC-net: Efficient transformer for grading renal cell carcinoma in histopathological images. Comput. Electr. Eng. 2025, 128, 110747. [Google Scholar] [CrossRef]
- Ghoniem, R.M.; Algarni, A.D.; Shaalan, K. Multi-Modal Emotion Aware System Based on Fusion of Speech and Brain Information. Information 2019, 10, 239. [Google Scholar] [CrossRef]
- Sumon, R.I.; Mozumdar, M.A.; Akter, S.; Uddin, S.M.; Al-Onaizan, M.H.; Alkanhel, R.I.; Muthanna, M.S. Comparative Study of Cell Nuclei Segmentation Based on Computational and Handcrafted Features Using Machine Learning Algorithms. Diagnostics 2025, 15, 1271. [Google Scholar] [CrossRef] [PubMed]
- Soler, L.; Hostettler, A.; Agnus, V.; Charnoz, A.; Fasquel, J.; Moreau, J.; Osswald, A.; Bouhadjar, M.; Marescaux, J. 3D Image Reconstruction for Comparison of Algorithm Database: A Patient-Specific Anatomical and Medical Image Database. IRCAD, Strasbourg, France, Tech. Rep. 2010. Available online: https://www.ircad.fr/research/data-sets/liver-segmentation-3d-ircadb-01 (accessed on 16 June 2025).
- Kavur, A.E.; Gezer, N.S.; Barış, M.; Aslan, S.; Conze, P.H.; Groza, V.; Selver, M.A. CHAOS challenge–combined (CT-MR) healthy abdominal organ segmentation. Med. Image Anal. 2021, 69, 101950. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).