Next Article in Journal
Tackling Imbalanced Data in Chronic Obstructive Pulmonary Disease Diagnosis: An Ensemble Learning Approach with Synthetic Data Generation
Next Article in Special Issue
Harmonic Ratio Analysis in Magnetic Particle Imaging Enables Differentiation of Malignant and Benign Human Breast Tissues: A Feasibility Study
Previous Article in Journal
Integrating the Contrasting Perspectives Between the Constrained Disorder Principle and Deterministic Optical Nanoscopy: Enhancing Information Extraction from Imaging of Complex Systems
Previous Article in Special Issue
Radiomics in Action: Multimodal Synergies for Imaging Biomarkers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MedSegNet10: A Publicly Accessible Network Repository for Split Federated Medical Image Segmentation

by
Chamani Shiranthika
*,†,
Zahra Hafezi Kafshgari
*,†,
Hadi Hadizadeh
and
Parvaneh Saeedi
School of Engineering Science, Simon Fraser University, Burnaby, BC V5A 1S6, Canada
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2026, 13(1), 104; https://doi.org/10.3390/bioengineering13010104
Submission received: 30 October 2025 / Revised: 26 December 2025 / Accepted: 10 January 2026 / Published: 15 January 2026
(This article belongs to the Special Issue Medical Imaging Analysis: Current and Future Trends)

Abstract

Machine Learning (ML) and Deep Learning (DL) have shown significant promise in healthcare, particularly in medical image segmentation, which is crucial for accurate disease diagnosis and treatment planning. Despite their potential, challenges such as data privacy concerns, limited annotated data, and inadequate training data persist. Decentralized learning approaches such as federated learning (FL), split learning (SL), and split federated learning (SplitFed/SFL) address these issues effectively. This paper introduces “MedSegNet10,” a publicly accessible repository designed for medical image segmentation using split-federated learning. MedSegNet10 provides a collection of pre-trained neural network architectures optimized for various medical image types, including microscopic images of human blastocysts, dermatoscopic images of skin lesions, and endoscopic images of lesions, polyps, and ulcers. MedSegNet10 implements SplitFed versions of ten established segmentation architectures, enabling collaborative training without centralizing raw data and labels, reducing the computational load required at client sites. This repository supports researchers, practitioners, trainees, and data scientists, aiming to advance medical image segmentation while maintaining patient data privacy.

1. Introduction

Medical image segmentation plays a crucial role in the healthcare sector, serving as a foundational task for precise diagnosis, treatment planning, and patient care [1,2]. The emergence of decentralized training approaches, such as federated learning (FL) [3], split learning (SL) [4], and split federated learning (SplitFed learning/SFL) [5], has opened up new possibilities for collaborative, privacy-preserving, and resource-constrained model training across hospitals, research institutes, and healthcare organizations. Despite the considerable potential of decentralized learning approaches, their implementation for medical image segmentation still faces several practical challenges. One notable obstacle is the limited availability of well-designed and rigorously evaluated deep learning networks that can be readily adapted to SplitFed training pipelines. Addressing this gap requires accessible, standardized, and benchmarked resources that allow researchers to support reproducible experimentation and facilitate fair architectural comparison.
Alongside these technical challenges, interpretability has become a parallel requirement for the successful deployment of AI systems in real clinical environments. Clinicians expect models not only to perform accurately but also to provide insights that are consistent with clinical reasoning. Meta-analyses such as [6,7] underscore that transparency, usability, and clinician-aligned explanations are crucial for trustworthy AI-assisted decision support. This growing emphasis on interpretability poses additional challenges for decentralized learning: if medical image segmentation models are to be trained collaboratively across institutions, they must also provide insight into how decisions are made, particularly in high-stakes medical settings.
In this paper, we introduce the first documented repository of SplitFed networks—MedSegNet10 tailored specifically for medical image segmentation tasks. Aligned with the recent evolution of semantic segmentation networks, we curate ten top-performing models based on their demonstrated effectiveness in medical imaging [8,9]. These include UNet [10], SegNet [11], DeepLabV3 [12], DeepLabV3+ [13], RefineNet [14], CGNet [15], SUNet [16], DUCK-Net [17], Attention UNet [18], and Swin-UNet [19]. We train and evaluate all SplitFed variants using three datasets with distinct imaging characteristics: Human Against Machine (HAM10K) [20], KVASIR-SEG [21], and our proprietary Blastocysts dataset [22]. Ultimately, the repository aims to provide a well-structured, reproducible SplitFed benchmark that future work can expand to additional imaging domains and that can be enriched with interpretability-oriented components in subsequent research.
The primary objective of this work is to develop and release a publicly accessible repository of SplitFed-ready medical image segmentation networks. To this end, we standardize, benchmark, and streamline SplitFed development by rigorously implementing and faithfully reproducing SplitFed variants of ten well-established segmentation models. MedSegNet10 is designed as a practical and comprehensive resource, providing both novice and experienced practitioners with a unified platform for implementing and evaluating SplitFed-based segmentation. It integrates multiple architectures within a single framework, enabling systematic and rigorous comparisons of structural properties, performance characteristics, and SplitFed behaviour across models. Rather than introducing new SplitFed algorithms or novel segmentation architectures, this work focuses on establishing the first unified, reusable, and reproducible SplitFed repository for medical image segmentation. All included networks adhere to standardized split-point definitions, consistent client–server partitioning, and harmonized training pipelines. By creating this consistent experimental foundation, MedSegNet10 addresses a critical gap in the literature and provides essential infrastructure for rigorous cross-architecture comparison and future SplitFed research.
The structure of this paper is as follows. Section 2 provides an overview of SplitFed fundamentals and the current state-of-the-art in SplitFed for medical image segmentation, including its integration with emerging transformer-based architectures. It also reviews publicly available federated network repositories, outlines key concepts in semantic segmentation, and describes the architectures of our implemented split models, concluding with a discussion of our split-point selection strategy. Section 3 presents the experimental setup, results, and evaluation. Section 4 discusses the current limitations of MedSegNet10 and outlines potential future directions. Finally, Section 5 concludes the paper.

2. Related Works

This section reviews related work on SplitFed learning, publicly available federated repositories, and semantic segmentation, including notable semantic segmentation networks. Also, it offers insights into our decisions on split point selection for designing our split networks.

2.1. Split Federated Learning (SplitFed)

The technical evolution of SplitFed research reflects a progression from foundational feasibility studies toward more specialized problem domains. The initial SplitFed work [5] established the core client–server partitioning mechanism and demonstrated its advantages in privacy preservation achieved via FL and reduced client workload achieved via SL. However, it did not address heterogeneous data distributions or noisy communication channels. Subsequent studies targeted specific weaknesses of the original formulation: SplitAvg [23] introduced heterogeneity-aware aggregation to improve generalization on non-IID data, Dynamic corrected SplitFed [24] incorporated homomorphic encryption to mitigate security risks, and Quality-Adaptive SplitFed [25] addressed label imperfections—a recurring challenge in medical datasets. Smart SplitFed [26] further tackled robustness under unstable or noisy communication, a practical issue in multi-hospital deployments. More recently, transformer-based SplitFed extensions [27,28,29,30] explored architectures capable of capturing long-range dependencies, demonstrating improved performance in high-dimensional medical imaging tasks but at the cost of increased communication overhead and memory usage on the server. The studies [30,31,32] highlight the effectiveness of SplitFed transformers used in medical imaging tasks.
Despite this progression, prior studies remain isolated in scope: each investigates a single architecture under a narrow set of conditions, with no standardization of split-point selections, model partitioning strategies, or training pipelines across architectures. As a result, direct comparison among SplitFed variants is currently difficult, and the field lacks a unified benchmark for evaluating model behaviour across diverse network architectures and medical imaging domains. This fragmentation motivates the development of MedSegNet10, which consolidates these design elements into a consistent and reproducible SplitFed framework.
Figure 1 illustrates our SplitFed architecture, considering each client’s network backbone as the UNet. In this U-shaped architecture, each client has two split points: the front-end (FE) and back-end (BE) sub-models, with the server (SS) sub-model in between. The FE sub-model exclusively accesses the images, while the BE sub-model solely interacts with the ground truth (GT) data. Both the FE and BE sub-models constitute a small portion of the UNet architecture and are located on the client side, while the server sub-model, containing the majority of the layers, handles the main computational burden of training. This splitting process was consistently applied across all ten selected models.

2.2. Publicly Available Federated Repositories

Various public repositories for federated learning (FL), particularly those with annotated medical images [33], alongside numerous federated data analysis programs [34,35,36], are available for public access. Notable federated network repositories include TensorFlow Federated [37], Google Federated Research [38], LEAF [39], Fedmint [40], Federated Scope [41], FedLab [42], FATE [43], FedIIC [44], QuickSQL [45], GraphQL-Mesh [46], MedAugment [47], OpenMined’s PyGrid [48], FedCT [49], and COALA [50]. Among these, FATE, MedAugment, OpenMined’s PyGrid, and FedCT are specifically designed for medical imaging tasks.
Although these repositories provide valuable tools for federated experimentation, they primarily focus on general-purpose FL pipelines, data-sharing frameworks, or isolated model implementations. Existing platforms do not provide a unified, reusable collection of SplitFed architectures for medical image segmentation, and they do not standardize split-point definitions, client-server model partitioning, or consistent training pipelines across multiple segmentation networks. Prior SplitFed studies typically evaluate a single model or a narrow experimental setup, limiting reproducibility and cross-architecture comparison. These limitations reveal the need for a robust and adaptable SplitFed network repository that supports collaborative and reproducible research. MedSegNet10 directly addresses this gap by streamlining the implementation, training, and evaluation of SplitFed architectures in medical image segmentation.

2.3. Image Segmentation Models

Image segmentation is essential in healthcare for the accurate identification and extraction of regions within medical images, which is critical for disease detection, diagnosis, and treatment planning [8]. Enhancing the capability of algorithms to analyze complex medical data plays a vital role in the interpretation of medical images. Deep learning applications, such as SplitFed, further emphasize the significance of image segmentation by managing large-scale data across institutions while preserving privacy, thereby improving accuracy and efficiency. Semantic segmentation takes this further by classifying each pixel within identified regions, such as tumors or organs, providing detailed representations crucial for precise diagnosis and treatment. This section highlights MedSegNet10’s contributions to advancing semantic segmentation in medical imaging, addressing its complexities, and enhancing image content understanding.
Figure 2 illustrates a timeline of prominent semantic segmentation networks that have emerged in recent decades, with their full names and corresponding references listed in the order shown in the figure. These include:
  • FCN [51]: Fully Connected Network.
  • UNet [10]: UNetwork.
  • SegNet [11]: Segmentation Network.
  • PSPnet [52]: Pyramid Scene Parsing Network.
  • ENet [53]: Efficient Neural Network.
  • RefineNet: Refining Segmentation-based Network.
  • DeepLab [12,13,54,55]: Deep Labelling for Semantic Image Segmentation.
  • Attention UNet: Attention-based UNet.
  • FPN [56]: Feature Pyramid Network.
  • Mask RCNN [57]: Mask Region-based Convolutional Neural Network.
  • PANet [58]: Path Aggregation Network.
  • BiseNet [59]: Bilateral Segmentation Network.
  • HRNet [60]: High-Resolution Network.
  • OCRNet [61]: Object-Contextual Representations for Semantic Segmentation.
  • DANet [62]: Dual Attention Network.
  • CCNet [63]: Criss-Cross Attention Network.
  • SETR [64]: Spatially Enhanced Transformer.
  • UPerNet [65]: Unified Perceptual Parsing Network.
  • FastFCN [66]: Fast Fully Convolutional Network.
  • SUNet [16]: Strong UNet.
  • FANet [67]: Feature Aggregation Network.
  • DMNet [68]: Dense Multi-scale Network.
  • CGNet [15]: Context-Guided Network.
  • DETR [69]: DEtection Transformer.
  • PraNet [70]: Parallel Reverse Attention Network.
  • ViT [71]: Vision Transformer.
  • Swin-UNet: Swin Transformer-based UNet.
  • MSRF-Net [72]: Multi-Scale Residual Fusion Network.
  • T2T-ViT [73]: Token-to-Token Vision Transformer.
  • VAN [74]: Visual Attention Network.
  • CSwin Transformer [75]: Cross-Stage win transformer.
  • DUCK-Net [17]: Deep Understanding Convolutional Kernel Network.
  • ST-UNet [76]: Spatiotemporal UNet.
  • SAM [77]: Segment Anything.
  • VM-UNet [78]: Vision Mamba UNet.
  • HC-Mamba [79]: Hybrid-convolution version of Vision Mamba.
  • EoMT [80]: Encoder-only Mask Transformer.
  • Med-SA [81]: Medical SAM Adapter.
We implemented both the Split and SplitFed versions of the networks marked with a green star in Figure 2. The following sections provide detailed descriptions of these implemented networks.

2.3.1. UNet

UNet can be considered as the most well-known architecture specially designed for image segmentation tasks [10]. It embodies the encoder–decoder architecture with skip connections. UNet employs a process of gradually upsampling the features extracted by its encoder using transpose convolutions. By incorporating skip connections, UNet reaches greater depths than many other SoTA networks, resulting in significantly higher-quality outputs. Figure 3a shows the split version of the UNet used in this work.

2.3.2. SegNet

The SegNet architecture shares significant similarities with UNet as it adheres to an encoder–decoder framework [11]. One of the distinguishing characteristics of SegNet from UNet is the absence of skip connections. Furthermore, SegNet diverges from the conventional upsampling operation found in many encoder–decoder architectures. Instead of traditional upsampling techniques, SegNet employs an approach referred to as “max unpooling”. This technique mitigates the need to learn how to upscale both the final output score and the feature maps from earlier layers. In a conventional encoder–decoder structure, such upscaling operations typically demand substantial learning. By integrating max unpooling, SegNet optimally addresses this challenge. Figure 3b depicts the split version of the SegNet used in our work.

2.3.3. DeepLab

DeepLab was developed by Google [54] and has evolved into several versions, including DeepLab [54], DeepLabV1 [54], DeepLabV2 [55], DeepLabV3 [12] and DeepLabV3+ [13].
DeepLabV1 addresses those challenges presented by the SoTA networks, especially the FCN. It addressed the challenge of reduced feature resolution by employing atrous convolutions for upsampling. It addressed the challenge of reduced localization accuracy, due to DCNN invariance, by performing a post-processing procedure via conditional random fields (CRFs) [54]. DeepLabV2 addresses the challenge of handling objects at multiple scales by introducing the Atrous Spatial Pyramid Pooling (ASPP) method [55]. DeepLabV3 represents a significant advancement over DeepLabV2. Unlike DeepLabV2, which relies on the VGG16 backbone and a simpler ASPP module, DeepLabV3 introduces flexibility with diverse backbones like ResNet [82] and Xception [83]. The ASPP module is enhanced with parallel atrous convolutions, capturing multi-scale information more effectively. DeepLabV3 also incorporates global context through image-level features, distinguishing it from its predecessor. The subsequent DeepLabV3+ introduces additional efficiency with depthwise separable convolution. In summary, DeepLabV3 refines segmentation through advanced techniques, offering improved performance and adaptability over DeepLabV2 [12]. Figure 4a shows the split version of DeepLabV3 with ResNet50 as the backbone used in our work. DeepLabV3+ refines the segmentation output even further, especially along object boundaries. This is achieved using atrous separable convolutions in the encoder–decoder architecture and modifying the backbone version of the Xception network [83]. Figure 4b shows the split version of the DeepLabV3+ used in this work.

2.3.4. RefineNet

RefineNet is a generic multi-path refinement network that explicitly exploits all the information available along the down-sampling process to enable high-resolution prediction using long-range residual connections [14]. The authors highlighted limitations associated with typical CNNs, FCNs, and dilated convolutions. CNNs suffer from the downscaling of the feature maps, while FCNs output low-resolution predictions. Dilated convolutions are computationally expensive to train and could quickly reach memory limits. In RefineNet, fine-grained features from earlier convolutions are used directly to improve the deeper layers to capture high-level semantic features. This is called multi-path refinement. Chained residual pooling is also introduced by this work, where rich background contexts are captured in an efficient manner. Some variants of RefineNet are also proposed as a single RefineNet, 2-cascaded RefineNet, and 4-cascaded 2-scale RefineNet that upgrade its overall flexibility. Figure 5a shows the split version of RefineNet used in this work.

2.3.5. SUNet

SUNet [16] is another U-shaped encoder–decoder network based on the inception module [84] and the dense block [85] to enhance the feature extraction and information reuse capabilities of the network. The idea behind the invention of SUNet is to make the standard UNet stronger in both width and depth. Four versions of SUNet are proposed as SUNet-V1, SUNet-V2, SUNet-V3, and SUNet-V4. As the version number increases, an improvement in classification and segmentation accuracy has been observed. It was originally invented as a federated brain tumor segmentation network. Figure 5b shows the structure of the split SUNet-V4 architecture.

2.3.6. CGNet

CGNet is a recent network architecture designed for efficient and accurate semantic segmentation [15]. It stands out for having a lightweight architecture that benefits from context information to enhance segmentation performance. CGNet employs context blocks to capture contextual information at different scales, which helps the model to better understand the relationships between different objects and parts of an image. To increase the receptive field without adding to the computational overhead, CGNet utilizes dilated convolutions in the context blocks. Dilated convolutions allow the network to incorporate a larger context while keeping the number of parameters and computations low. The CGNet architecture uses feature fusion modules to combine features from different levels of the network. Such a fusion helps in integrating multi-scale information and improving segmentation performance. The network incorporates skip connections to help propagate information across different layers of the network. These connections enable better gradient flow during training and facilitate the overall optimization process. Despite being lightweight, CGNet delivers competitive performance on various semantic segmentation benchmarks, demonstrating its effectiveness in producing accurate segmentation results. Figure 6a shows the split version of CGNet used in our work.

2.3.7. DUCK-Net

The DUCK-Net [17] uses an encoder–decoder architecture and has two main parts: an FCN block called DUCK that utilizes six different types of convolutional blocks at the same time, and a secondary UNet that keeps the low-level information intact. At each step, the DUCK block replaces the traditional pair of 3 × 3 convolutional blocks used in UNet. This allows the model to capture more details at each step while sacrificing finer, low-level information. Residual blocks are utilized in the last downsampling operation. Figure 6b shows the split version of DUCK-Net used in this work.

2.3.8. Attention UNet

The Attention UNet, a modified UNet architecture for semantic image segmentation [18], integrates attention mechanisms [86] to enhance fine-grained detail capture and precise object localization. This attention mechanism enables the network to focus on informative regions while suppressing less relevant areas. Following an encoder–decoder structure, Attention UNet extracts hierarchical representations using the encoder and employs attention gates (channel-spatial gates) in the decoder. These gates assign weights to different regions in the feature maps, enhancing the focus on more important details during up-sampling. This attention-driven approach enhances the accuracy of segmentation masks by selectively emphasizing critical features, aiding in precise object and boundary localization. The split version of the Attention UNet used in our work is shown in Figure 7a.

2.3.9. Swin-UNet

Swin-UNet is the first pure transformer-based U-shaped architecture [19], which combines the architectural advantages of the UNet and the Swin transformer [87]. This network includes the components of patch partition blocks, linear embedding blocks, Swin transformer blocks, patch merging blocks, patch expanding blocks, and linear projection blocks. Three skip connections are used to fuse the multi-scale features from the encoder with the upscaled features. Swin-UNet is famous for comparatively better performance than others due to its efficient attention mechanism, hierarchical feature representation, versatility, fewer parameters, and architectural design. Figure 7b shows the split version of the Swin-UNet used in this work.

2.4. Decision on Split Points Selection

In a SplitFed network, the selection of the most suitable split points is a crucial task. This choice of where to split the model is important to maintain the performance, communication, and overall efficiency of the SplitFed network [88]. While the ten architectures used in this study differ structurally, our split-point design follows a consistent principle: (i) the FE contains low-level feature extractors that interface with sensitive data, (ii) the SS contains the majority of parameters and computational load, and (iii) the BE handles prediction and gradients tied to GT labels to fully preserve the privacy. Although the proportion of layers varies across architectures, these design principles ensure functional consistency and reproducibility.
  • Task-specific concerns: The choice of split points is often guided by the nature of the machine learning task. For instance, in natural language processing, splits should occur at layers that capture semantic features, whereas in computer vision, splits should be at layers that capture high-level visual features.
  • Communication constraints: Split points should be strategically selected to minimize the overall computational load and communication costs associated with information transfer. This involves choosing points where computations are most intensive or sensitive, thus reducing overall latency and communication overhead.
  • Model architecture: Split points are selected at layers representing high-level features to enable clients to effectively learn task-specific details, ensuring that the model architecture supports the desired learning outcomes. Moreover, the edge blocks maintain the same dimensions, which is necessary for backpropagating gradients in the backward pass. Each sub-model generates its own gradients, making consistent dimensionality crucial.
  • Privacy and security concerns: To maintain data privacy, splits must be designed so that sensitive data remains on the client side. This approach involves creating two distinct model parts on the client side, with the front end handling sensitive data and the back end managing sensitive GTs.
  • Computational capabilities of clients: Split points are chosen to ensure that clients perform minimal computations, allowing those with limited resources to participate in the SplitFed training process without facing computational constraints.
In our experiments, we approached the determination of split points by evaluating each network individually. We carefully considered one or more of the criteria outlined above to ensure that the split points were chosen in a way that optimally addressed the specific characteristics and requirements of each network. This methodical approach allowed us to tailor our decisions to the unique aspects of each network, thereby enhancing the overall effectiveness of our experiments.

3. Experiments

In this section, we detail our experimental setup and results, followed by an evaluation that includes comparisons of performance and computational complexity.

3.1. Experimental Setup

We carried out the experiments with the three medical datasets introduced in Section 1:
-
Blastocyst dataset [22]: includes 801 Blastocyst RGB images along with their GTs created for a multi-class embryo segmentation task. Each image is segmented into five classes: zona pellucida (ZP), trophectoderm (TE), blastocoel (BL), inner cell mass (ICM), and background (BG).
-
HAM10K dataset [20]: The Human Against Machine dataset contains 10,015 dermatoscopic RGB images along with the corresponding binary GT masks, representing seven different types of skin lesions, including melanoma and benign conditions. Each image is segmented into two classes: skin lesion and background.
-
KVASIR-SEG dataset [21]: contains 1000 annotated endoscopic RGB images of polyps from colonoscopy procedures, each paired with a binary GT segmentation mask. Each image is segmented into two classes: abnormal condition (such as a lesion, polyp, or ulcer) and background.
We randomly distributed samples from each dataset between clients. For the Blastocyst dataset, samples were assigned to four clients with 110, 90, 200, and 300 samples, respectively, with an additional 101 samples reserved for testing. The HAM10K dataset was divided among ten clients, receiving 1176, 588, 305, 941, 1058, 1294, 648, 942, 883, and 1132 samples each, with 1000 samples set aside for testing. The KVASIR-SEG dataset was partitioned into four clients with 125, 175, 275, and 325 samples, respectively, and 100 images were reserved for testing. In each segmentation task, 85% of the data was used for training and 15% for validation by each client.
All samples were resized to uniform dimensions of 240 × 240 pixels. Based on the network architecture and the dataset, we used several augmentation techniques to improve individual model training performance. Those include horizontal and vertical flipping, rotating, RGB shifting, normalizing, random brightness contrasting, etc. All experiments used the same loss function—Soft Dice Loss [89], Adam optimizer, and stopping criteria. We used network-specific and dataset-specific initial learning rates. Intersection Over Union (IoU) averaged over the sample size (average IoU), or equally the average Jaccard Index [90] was used as the performance metric. For the Blastocyst dataset, we utilized the average IoU of the TE, ZP, BL, and ICM components. Each centralized model was trained for 120 epochs. Each SplitFed model was trained for 10 global communication rounds, while each client trained their local models for 12 local epochs. The only difference between the three settings—centralized, locally centralized, and SplitFed is the training paradigm itself. This ensures a controlled comparison, ensuring fairness and transparency.
The training process begins with each client receiving randomly initialized copies of the FE and BE sub-models, along with a copy of the server sub-model. Clients then train these models locally for a fixed number of epochs, collaborating with their respective server sub-model copies. After local training, clients send the weights of their FE and BE sub-models to the server, where they are aggregated to form an updated global model. The updated sub-models are then distributed back to the clients for local validation, completing one global epoch. During training, clients send features from the FE sub-model to the server, which processes them and returns the output for BE processing. The BE sub-model generates predictions, computes the loss, and initiates back-propagation, with gradient updates passed from the BE sub-model to the server and then back to the FE sub-model.
Experiments were conducted on the Graham, Narval, and Cedar clusters using high-performance computing resources provided by Digital Research Alliance of Canada (https://ccdb.alliancecan.ca). We used a simple Linux utility for resource management scripts (SLURM) to request 1 GPU, 8 CPU cores, and 64 GB RAM per experiment on a single node. We have added the environment.yml file to the repository to support reproducibility. The codebase follows a standard PyTorch (v2.0.1) dataset/dataloader structure, allowing users to integrate their own datasets by implementing a dataset class consistent with the existing format. Actual deployments of the models are outside the scope of this initial release but will be included in future iterations of the repository. The initial release of the repository is available to download at https://vault.sfu.ca/index.php/s/ryhf6t12O0sobuX (password upon request to the authors).

3.2. Experimental Results

The following sections outline our experimental results. We show the results in two sections: Quantitative results and Qualitative results.

3.2.1. Quantitative Results

We consider cases:
  • Centralized learning on full data: We initially trained each network without splitting for image segmentation. We utilized the entire data from the three datasets separately. The average IoUs for all data samples in each set for the centralized models are displayed in the C column of Table 1.
  • Centralized learning locally at each client: Secondly, we trained each client’s local data in a client-specific, centralized manner to ensure a fair comparison. In this step, each client trained the networks without data splitting. We recorded the IoUs for each client and computed the average, which is presented in the L column for each segmentation task in Table 1.
  • SplitFed learning: Thirdly, we trained the SplitFed networks in collaboration with all clients. The IoUs of the SplitFed models are recorded in the S column for each segmentation task in Table 1.

3.2.2. Qualitative Results

We included a visual comparison of three random samples from each test set during the SplitFed model training, as shown in Table 2.

3.3. Evaluation

3.3.1. Testing Performance Comparison

When evaluating the cases described in Section 3.2.1, we expect the following IoU behaviors: Centralized learning on full data (C) should achieve the highest average IoUs due to access to the complete dataset, enabling better generalization. SplitFed learning should follow with slightly lower average IoUs, as collaboration among clients is effective but may face communication and synchronization challenges. Centralized learning locally (L) at each client is expected to yield the lowest average IoUs, as limited data at each client reduces generalization and segmentation performance. Table 1 confirms this: the highest average IoUs (of individual models) and the highest average over models are in the C column, followed by the S column, with the lowest in the L column. When looking at the average over the models (last row), the C column in the Blastocyst dataset showed 10.61% higher average IoU than the L column, and 1.26% higher average IoU than the S column. C column in the HAM10K dataset showed 5% higher average IoU than the L column, and 0.87% higher average IoU than the S column. C column in the KVASIR-SEG dataset showed 12.83% higher average IoU than the L column, and 2.01% higher average IoU than the S column.
Table 3 shows the three top-performing models for each case of the centralized, locally centralized, and SplitFed for the three segmentation tasks. As seen from the results, for smaller datasets like the Blastocyst dataset, models like Attention UNet and SUNet perform better during SplitFed, likely due to their ability to better generalize with limited data. Complex models like DeepLabV3 still perform well in both centralized and SplitFed, due to their pattern recognition capabilities. In the HAM10K dataset, which is much larger, more complex models such as CGNet and DeepLabV3 excel in the centralized case because they can better utilize the complete dataset for complex feature extraction. Even in SplitFed, models like DeepLabV3 and UNet continue to perform well, benefiting from efficient feature extraction suited to larger datasets. In the KVASIR-SEG dataset, models like DeepLabV3 and CGNet consistently perform well across all cases, showing their versatility. DUCK-Net also performs strongly in centralized cases, likely due to its efficiency in handling polyp segmentation tasks. However, in SplitFed cases, the Duck block at the edge layers appears to rely on tight synchronization, which may result in inefficiencies and slower convergence.
For the purpose of better comparison, we graphically represent the results in Table 1 in Figure 8, Figure 9 and Figure 10. Figure 8 summarizes the segmentation performance of various methods for the Blastocyst dataset, while Figure 9 shows the results for the HAM10K dataset. Comparisons for the KVASIR-SEG dataset are displayed in Figure 10. In these figures, the performance of the centralized model is represented in blue, the performance of the locally centralized models is depicted in orange, and the performance of the SplitFed models is illustrated in gray. From these figures, the SplitFed models have demonstrated superior performance compared to the locally centralized models, with performances closer to their corresponding centralized models.

3.3.2. Comparison of Computational Complexity

We measured each model’s complexity with floating-point operations (FLOPs) by using the ptflops package in Python 3.9 [91]. Table 4 outlines the number of trainable parameters of each network, the measured FLOPs, the number of trainable parameters, and the FLOPs value with respect to UNet as the anchor. According to these metrics, CGNet has the least number of trainable parameters and the lowest FLOPs values compared to the other networks. RefineNet has the highest percentage of trainable parameters, while Attention UNet has the highest FLOPs value.
Further, Table 5 presents the layer proportions, trainable parameter distribution, and FLOPs for each architecture, highlighting how model complexity is partitioned across the FE, SS, and BE components in the SplitFed setup.

3.3.3. Performance Comparison with Other Existing Methods

Table 6 presents the average IoUs of the other existing methods for semantic segmentation with the Blastocyst, HAM10K, and KVASIR-SEG datasets. For fairness and consistency, we only included studies that explicitly report average IoUs or average Jaccard Indices on the same datasets used in this work. Several mainstream segmentation models, such as SAM [77] do not provide reproducible public benchmarks or pretrained models for these datasets due to data availability, privacy restrictions, or data incompatibility. As a result, they cannot be reliably included in a quantitative comparison. In particular, the Blastocyst dataset is one of our proprietary datasets, so comparisons with SoTA are missing. Studies that solely mentioned accuracy metrics or any other metrics were intentionally omitted from the evaluation.

4. Limitations & Future Works

Throughout this work, we identified several limitations of the current MedSegNet10 release, as well as promising directions for further research. These are outlined as follows:
  • First, our evaluation is limited to three distinct and commonly studied publicly available image types. Although we considered both multi-class (Blastocyst) and binary (HAM10K and KVASIR-SEG) segmentation datasets with varying sample sizes and styles to broaden the scope of generalization, these datasets may still not fully reflect the diversity of imaging characteristics, modalities, and annotation practices encountered in large-scale clinical deployments. Consequently, the reported results should not be regarded as definitive evidence of cross-domain generalizability. Future work will involve evaluating MedSegNet10 across a wider range of imaging modalities (e.g., CT, MRI, and multimodal acquisitions), annotation practices, and datasets to more comprehensively assess its robustness.
  • Second, the datasets used in this study are modest in size. This limitation reflects the broader reality of medical imaging research- large open-source datasets are rare, expert-annotated data are costly, many image types are difficult to obtain, and privacy regulations restrict access. Consequently, it is naturally infeasible to evaluate decentralized frameworks on large-scale public datasets simply because such resources do not exist for many medical imaging tasks. MedSegNet10 should therefore be viewed as a foundational resource rather than a demonstration of large-scale scalability. Future work will involve expanding MedSegNet10 using larger institutional datasets and multi-centre cohorts, enabling more rigorous evaluation under realistic data volumes, acquisition variability, and deployment conditions.
  • Third, the current experiments intentionally adopt IID data partitions to establish a controlled baseline and support benchmark reproducibility. This design choice does not reflect real clinical environments, where hospitals often exhibit strongly non-IID data distributions due to differing demographics, imaging devices, and annotation protocols. Non-IID robustness is a central challenge in federated learning, and evaluating MedSegNet10 under a range of realistic non-IID scenarios is an important direction for future work.
  • Finally, recent capability-oriented reviews in smart healthcare [105] highlight how AI contributes to integrated monitoring [105,106], remote diagnostics [107], decentralized decision support [108], and data-driven hospital systems [109]. SplitFed architectures align naturally with these developments because they enable collaboration across institutions while preserving data privacy. Incorporating SplitFed networks from MedSegNet10 into smart-healthcare frameworks could facilitate interoperable, privacy-preserving segmentation tools that operate across hospitals or global imaging networks. This integration represents a promising avenue for extending MedSegNet10 beyond standalone model training towards deployment in real clinical infrastructures.

5. Conclusions

In this work, we introduced MedSegNet10, a new publicly accessible repository for medical image segmentation utilizing the SplitFed learning mechanism. Our main goal was to create a resource allowing researchers to easily integrate SplitFed models into their own applications without extensive customization by allowing the reuse of the networks in the MedSegNet10 repository. We designed, implemented, and optimized SplitFed versions of ten selected models to simplify the process and make the strengths of SplitFed accessible to a wider audience. In addition, we outlined identified limitations of the current MedSegNet10 release, with several future research directions in medical image segmentation that could benefit from SplitFed learning. MedSegNet10 provides a structured starting point for future work on SplitFed architectures, including extensions to more diverse datasets and modalities, non-IID clinical scenarios, and interpretability-oriented smart healthcare model designs.

Author Contributions

Conceptualization, C.S. and Z.H.K.; methodology, C.S. and Z.H.K.; software, C.S. and Z.H.K.; validation, C.S. and Z.H.K.; formal analysis, C.S. and Z.H.K.; investigation, H.H. and P.S.; resources, P.S.; data curation, C.S., Z.H.K. and P.S.; writing—original draft preparation, C.S. and Z.H.K.; writing—review and editing, H.H. and P.S.; visualization, C.S. and Z.H.K.; supervision, P.S.; project administration, P.S.; funding acquisition, P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study is funded by the Natural Sciences and Engineering Research Council (NSERC) of Canada.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study only uses publicly available medical imaging datasets released with appropriate ethical approvals by their providers.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Alzahrani, Y.; Boufama, B. Biomedical image segmentation: A survey. SN Comput. Sci. 2021, 2, 310. [Google Scholar] [CrossRef]
  2. Narayan, V.; Faiz, M.; Mall, P.K.; Srivastava, S. A Comprehensive Review of Various Approach for Medical Image Segmentation and Disease Prediction. Wirel. Pers. Commun. 2023, 132, 1819–1848. [Google Scholar] [CrossRef]
  3. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the AISTATS, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  4. Gupta, O.; Raskar, R. Distributed learning of deep neural network over multiple agents. J. Netw. Comput. 2018, 116, 1–8. [Google Scholar] [CrossRef]
  5. Thapa, C.; Arachchige, P.C.M.; Camtepe, S.; Sun, L. SplitFed: When Federated Learning Meets Split Learning. In Proceedings of the AAAI, Virtual, 22 February–1 March 2022; Volume 36, pp. 8485–8493. [Google Scholar]
  6. Abbas, Q.; Jeong, W.; Lee, S.W. Explainable AI in clinical decision support systems: A meta-analysis of methods, applications, and usability challenges. Healthcare 2025, 13, 2154. [Google Scholar] [CrossRef] [PubMed]
  7. Barragan-Montero, A.; Bibal, A.; Dastarac, M.H.; Draguet, C.; Valdes, G.; Nguyen, D.; Willems, S.; Vandewinckele, L.; Holmström, M.; Löfman, F.; et al. Towards a safe and efficient clinical implementation of machine learning in radiation oncology by exploring model interpretability, explainability and data-model dependency. Phys. Med. Biol. 2022, 67, 11TR01. [Google Scholar] [CrossRef]
  8. Guo, Y.; Liu, Y.; Georgiou, T.; Lew, M.S. A review of semantic segmentation using deep neural networks. Int. J. Multimed. Inf. Retr. 2018, 7, 87–93. [Google Scholar] [CrossRef]
  9. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the CVPR, Virtual, 19–25 June 2021; pp. 7262–7272. [Google Scholar]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the MICCAI, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  11. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder–decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  12. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  13. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  14. Lin, G.; Milan, A.; Shen, C.; Reid, I. Refinenet: Multi-path refinement networks for high-resolution semantic segmentation. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 1925–1934. [Google Scholar]
  15. Wu, T.; Tang, S.; Zhang, R.; Cao, J.; Zhang, Y. CGNet: A light-weight context guided network for semantic segmentation. IEEE Trans. Image Process. 2020, 30, 1169–1179. [Google Scholar] [CrossRef]
  16. Yi, L.; Zhang, J.; Zhang, R.; Shi, J.; Wang, G.; Liu, X. SU-Net: An efficient encoder–decoder model of federated learning for brain tumor segmentation. In Proceedings of the ICANN, Bratislava, Slovakia, 15–18 September 2020; pp. 761–773. [Google Scholar]
  17. Dumitru, R.G.; Peteleaza, D.; Craciun, C. Using DUCK-Net for polyp image segmentation. Sci. Rep. 2023, 13, 9803. [Google Scholar] [CrossRef]
  18. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  19. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the ECCV, Tel Aviv, Israel, 23–27 October 2022; pp. 205–218. [Google Scholar]
  20. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 1–9. [Google Scholar] [CrossRef]
  21. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Halvorsen, P.; de Lange, T.; Johansen, D.; Johansen, H.D. Kvasir-seg: A segmented polyp dataset. In Proceedings of the MMM, Daejeon, South Korea, 5–8 January 2020; pp. 451–462. [Google Scholar]
  22. Lockhart, L.; Saeedi, P.; Au, J.; Havelock, J. Multi-Label Classification for Automatic Human Blastocyst Grading with Severely Imbalanced Data. In Proceedings of the MMSP, Kuala Lumpur, Malaysia, 27–29 September 2019; pp. 1–6. [Google Scholar]
  23. Zhang, M.; Qu, L.; Singh, P.; Kalpathy-Cramer, J.; Rubin, D.L. Splitavg: A heterogeneity-aware federated deep learning method for medical imaging. IEEE J. Biomed. Health Inform. 2022, 26, 4635–4644. [Google Scholar] [CrossRef]
  24. Yang, Z.; Chen, Y.; Huangfu, H.; Ran, M.; Wang, H.; Li, X.; Zhang, Y. Dynamic corrected split federated learning with homomorphic encryption for u-shaped medical image networks. IEEE J. Biomed. Health Inform. 2023, 27, 5946–5957. [Google Scholar] [CrossRef]
  25. Kafshgari, Z.H.; Shiranthika, C.; Saeedi, P.; Bajić, I.V. Quality-adaptive split-federated learning for segmenting medical images with inaccurate annotations. In Proceedings of the IEEE ISBI, Cartagena de Indias, Colombia, 18–21 April 2023; pp. 1–5. [Google Scholar]
  26. Kafshgari, Z.H.; Bajić, I.V.; Saeedi, P. Smart split-federated learning over noisy channels for embryo image segmentation. In Proceedings of the IEEE ICASSP, Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  27. Baek, S.; Park, J.; Vepakomma, P.; Raskar, R.; Bennis, M.; Kim, S.L. Visual transformer meets cutmix for improved accuracy, communication efficiency, and data privacy in split learning. arXiv 2022, arXiv:2207.00234. [Google Scholar] [CrossRef]
  28. Qu, L.; Zhou, Y.; Liang, P.P.; Xia, Y.; Wang, F.; Adeli, E.; Fei-Fei, L.; Rubin, D. Rethinking architecture design for tackling data heterogeneity in federated learning. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022; pp. 10061–10071. [Google Scholar]
  29. Oh, S.; Park, J.; Baek, S.; Nam, H.; Vepakomma, P.; Raskar, R.; Bennis, M.; Kim, S.L. Differentially private cutmix for split learning with vision transformer. arXiv 2022, arXiv:2210.15986. [Google Scholar] [CrossRef]
  30. Park, S.; Kim, G.; Kim, J.; Kim, B.; Ye, J.C. Federated split vision transformer for COVID-19 CXR diagnosis using task-agnostic training. arXiv 2021, arXiv:2111.01338. [Google Scholar] [CrossRef]
  31. Park, S.; Ye, J.C. Multi-Task Distributed Learning Using Vision Transformer With Random Patch Permutation. IEEE Trans. Med. Imaging 2023, 42, 2091–2105. [Google Scholar] [CrossRef] [PubMed]
  32. Almalik, F.; Alkhunaizi, N.; Almakky, I.; Nandakumar, K. FeSViBS: Federated Split Learning of Vision Transformer with Block Sampling. arXiv 2023, arXiv:2306.14638. [Google Scholar]
  33. Shiranthika, C.; Saeedi, P.; Bajić, I.V. Decentralized Learning in Healthcare: A Review of Emerging Techniques. IEEE Access 2023, 11, 54188–54209. [Google Scholar] [CrossRef]
  34. Carter, K.W.; Francis, R.W.; Carter, K.; Francis, R.; Bresnahan, M.; Gissler, M.; Grønborg, T.; Gross, R.; Gunnes, N.; Hammond, G.; et al. ViPAR: A software platform for the Virtual Pooling and Analysis of Research Data. Int. J. Epidemiol. 2016, 45, 408–416. [Google Scholar] [CrossRef] [PubMed]
  35. Gazula, H.; Kelly, R.; Romero, J.; Verner, E.; Baker, B.T.; Silva, R.F.; Imtiaz, H.; Saha, D.K.; Raja, R.; Turner, J.A.; et al. COINSTAC: Collaborative informatics and neuroimaging suite toolkit for anonymous computation. JOSS 2020, 5, 2166. [Google Scholar] [CrossRef]
  36. Marcus, D.S.; Olsen, T.R.; Ramaratnam, M.; Buckner, R.L. The Extensible Neuroimaging Archive Toolkit: An informatics platform for managing, exploring, and sharing neuroimaging data. Neuroinformatics 2007, 5, 11–33. [Google Scholar] [CrossRef]
  37. Tensorflow Federated. Available online: https://github.com/tensorflow/federated/ (accessed on 8 May 2025).
  38. Google Federated Research. Available online: https://github.com/google-research/federated/ (accessed on 8 May 2025).
  39. LEAF. Available online: https://github.com/TalwalkarLab/leaf/ (accessed on 8 May 2025).
  40. Fedmint. Available online: https://github.com/fedimint/fedimint/ (accessed on 8 May 2025).
  41. Xie, Y.; Wang, Z.; Gao, D.; Chen, D.; Yao, L.; Kuang, W.; Li, Y.; Ding, B.; Zhou, J. FederatedScope: A Flexible Federated Learning Platform for Heterogeneity. Proc. VLDB Endow. 2023, 16, 1059–1072. [Google Scholar] [CrossRef]
  42. Zeng, D.; Liang, S.; Hu, X.; Wang, H.; Xu, Z. FedLab: A flexible federated learning framework. arXiv 2021, arXiv:2107.11621. [Google Scholar]
  43. FATE. Available online: https://github.com/FederatedAI/FATE/ (accessed on 8 May 2025).
  44. Wu, N.; Yu, L.; Yang, X.; Cheng, K.T.; Yan, Z. FedIIC: Towards Robust Federated Learning for Class-Imbalanced Medical Image Classification. In Proceedings of the MICCAI, Vancouver, Canada, 8–12, October, 2023. [Google Scholar]
  45. Quicksql. Available online: https://github.com/Qihoo360/Quicksql/ (accessed on 8 May 2025).
  46. Graphql-mesh. Available online: https://github.com/ardatan/graphql-mesh/ (accessed on 8 May 2025).
  47. MedAugment. Available online: https://github.com/NUS-Tim/MedAugment/ (accessed on 8 May 2025).
  48. OpenMined’s PyGrid. Available online: https://github.com/OpenMined/PyGrid-deprecated---see-PySyft-/ (accessed on 8 May 2025).
  49. FedCT. Available online: https://github.com/evison/FedCT/ (accessed on 8 May 2025).
  50. Zhuang, W.; Xu, J.; Chen, C.; Li, J.; Lyu, L. COALA: A Practical and Vision-Centric Federated Learning Platform. In Proceedings of the ICML, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  51. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the CVPR, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  52. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  53. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. Enet: A deep neural network architecture for real-time semantic segmentation. arXiv 2016, arXiv:1606.02147. [Google Scholar] [CrossRef]
  54. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  55. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  56. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  57. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 2961–2969. [Google Scholar]
  58. Wang, K.; Liew, J.H.; Zou, Y.; Zhou, D.; Feng, J. PANet: Few-shot image semantic segmentation with prototype alignment. In Proceedings of the ICCV, Seoul, South Korea, 27 October–2 November 2019; pp. 9197–9206. [Google Scholar]
  59. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. BiSeNet: Bilateral segmentation network for real-time semantic segmentation. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  60. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the CVPR, Long Beach, CA, USA, 16–20 June 2019; pp. 5686–5696. [Google Scholar] [CrossRef]
  61. Yuan, Y.; Chen, X.; Wang, J. Object-contextual representations for semantic segmentation. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020; pp. 173–190. [Google Scholar]
  62. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the CVPR, Long Beach, CA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
  63. Huang, Z.; Wang, X.; Huang, L.; Huang, C.; Wei, Y.; Liu, W. CCNet: Criss-cross attention for semantic segmentation. In Proceedings of the CVPR, Long Beach, CA, 16–20 June 2019; pp. 603–612. [Google Scholar]
  64. Heo, B.; Yun, S.; Han, D.; Chun, S.; Choe, J.; Oh, S.J. Rethinking spatial dimensions of vision transformers. In Proceedings of the ICCV, Virtual, 11–17 October 2021; pp. 11936–11945. [Google Scholar]
  65. Xiao, T.; Liu, Y.; Zhou, B.; Jiang, Y.; Sun, J. Unified perceptual parsing for scene understanding. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018; pp. 418–434. [Google Scholar]
  66. Wu, H.; Zhang, J.; Huang, K.; Liang, K.; Yu, Y. FastFCN: Rethinking dilated convolution in the backbone for semantic segmentation. arXiv 2019, arXiv:1903.11816. [Google Scholar] [CrossRef]
  67. Singha, T.; Pham, D.S.; Krishna, A. FANet: Feature aggregation network for semantic segmentation. In Proceedings of the DICTA, Melbourne, Australia, 29 November–2 December 2020; pp. 1–8. [Google Scholar]
  68. Fang, K.; Li, W.J. DMNet: Difference minimization network for semi-supervised segmentation in medical images. In Proceedings of the MICCAI, Lima, Peru, 4–8 October 2020; pp. 532–541. [Google Scholar]
  69. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the ECCV, Virtual, 23–28 August 2020; pp. 213–229. [Google Scholar]
  70. Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. PraNet: Parallel reverse attention network for polyp segmentation. In Proceedings of the MICCAI, Lima, Peru, 4–8 October 2020; pp. 263–273. [Google Scholar]
  71. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  72. Srivastava, A.; Jha, D.; Chanda, S.; Pal, U.; Johansen, H.D.; Johansen, D.; Riegler, M.A.; Ali, S.; Halvorsen, P. MSRF-Net: A multi-scale residual fusion network for biomedical image segmentation. IEEE J. Biomed. Inform. 2021, 26, 2252–2263. [Google Scholar] [CrossRef] [PubMed]
  73. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Jiang, Z.H.; Tay, F.E.; Feng, J.; Yan, S. Tokens-to-token ViT: Training vision transformers from scratch on imagenet. In Proceedings of the ICCV, Virtual, 11–17 October 2021; pp. 558–567. [Google Scholar]
  74. Guo, M.H.; Lu, C.Z.; Liu, Z.N.; Cheng, M.M.; Hu, S.M. Visual attention network. Comput. Vis. Media 2023, 9, 733–752. [Google Scholar] [CrossRef]
  75. Yang, H.; Yang, D. CSwin-PNet: A CNN-Swin Transformer combined pyramid network for breast lesion segmentation in ultrasound images. Expert Syst. Appl. 2023, 213, 119024. [Google Scholar] [CrossRef]
  76. Zhang, J.; Qin, Q.; Ye, Q.; Ruan, T. ST-UNet: Swin transformer boosted U-net with cross-layer feature enhancement for medical image segmentation. Comput. Biol. Med. 2023, 153, 106516. [Google Scholar] [CrossRef]
  77. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the CVPR, Vancouver, British Columbia, Canada, 18–22 June 2023; pp. 4015–4026. [Google Scholar]
  78. Ruan, J.; Li, J.; Xiang, S. VM-Unet: Vision mamba unet for medical image segmentation. arXiv 2024, arXiv:2402.02491. [Google Scholar] [CrossRef]
  79. Xu, J. HC-Mamba: Vision mamba with hybrid convolutional techniques for medical image segmentation. arXiv 2024, arXiv:2405.05007. [Google Scholar] [CrossRef]
  80. Kerssies, T.; Cavagnero, N.; Hermans, A.; Norouzi, N.; Averta, G.; Leibe, B.; Dubbelman, G.; de Geus, D. Your VIT is secretly an image segmentation model. In Proceedings of the CVPR, Nashville, TN, USA, 11–15 June 2025; pp. 25303–25313. [Google Scholar]
  81. Wu, J.; Wang, Z.; Hong, M.; Ji, W.; Fu, H.; Xu, Y.; Xu, M.; Jin, Y. Medical sam adapter: Adapting segment anything model for medical image segmentation. Med. Image Anal. 2025, 102, 103547. [Google Scholar] [CrossRef]
  82. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the CVPR, Las Vegas, NV, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  83. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  84. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the CVPR, Las Vegas, NV, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar]
  85. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the CVPR, Honolulu, HI, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  86. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  87. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the ICCV, Virtual, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  88. Shiranthika, C.; Kafshgari, Z.H.; Saeedi, P.; Bajić, I.V. SplitFed resilience to packet loss: Where to split, that is the question. In Proceedings of the MICCAI DeCaF, Vancouver, BC, Canada, 8–12 October 2023. [Google Scholar]
  89. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Proceedings of the MICCAI DLMIA, Quebec City, QC, Canada, 10–14 September 2017; pp. 240–248. [Google Scholar]
  90. Cox, M.A.A.; Cox, T.F. Multidimensional Scaling. In Handbook of Data Visualization; Springer: Berlin/Heidelberg, Germany, 2008; pp. 315–347. [Google Scholar]
  91. FLOPs Counting Tool for Neural Networks in Pytorch Framework. Available online: https://pypi.org/project/ptflops/ (accessed on 19 April 2025).
  92. Rad, R.M.; Saeedi, P.; Au, J.; Havelock, J. BLAST-NET: Semantic segmentation of human blastocyst components via cascaded atrous pyramid and dense progressive upsampling. In Proceedings of the ICIP, Taipei, Taiwan, 22-25 September 2019; pp. 1865–1869. [Google Scholar]
  93. Bdair, T.; Navab, N.; Albarqouni, S. FedPerl: Semi-supervised peer learning for skin lesion classification. In Proceedings of the MICCAI, Virtual, 27 September–1 October 2021; pp. 336–346. [Google Scholar]
  94. Wicaksana, J.; Yan, Z.; Zhang, D.; Huang, X.; Wu, H.; Yang, X.; Cheng, K.T. FedMix: Mixed supervised federated learning for medical image segmentation. IEEE Trans. Med. Imaging 2022, 42, 1955–1968. [Google Scholar] [CrossRef] [PubMed]
  95. Ruan, J.; Xiang, S.; Xie, M.; Liu, T.; Fu, Y. MALUNet: A multi-attention and light-weight unet for skin lesion segmentation. In Proceedings of the BIBM, Las Vegas, NV, USA and Changsha, Hunan, China, 6–9 December 2022; pp. 1150–1156. [Google Scholar]
  96. Yang, T.; Xu, J.; Zhu, M.; An, S.; Gong, M.; Zhu, H. FedZaCt: Federated learning with Z average and cross-teaching on image segmentation. Electronics 2022, 11, 3262. [Google Scholar] [CrossRef]
  97. Chen, Z.; Yang, C.; Zhu, M.; Peng, Z.; Yuan, Y. Personalized retrogress-resilient federated learning toward imbalanced medical data. IEEE Trans. Med. Imaging 2022, 41, 3663–3674. [Google Scholar] [CrossRef]
  98. Liu, F.; Yang, F. Medical image segmentation based on federated distillation optimization learning on non-IID data. In Proceedings of the IEEE ICIC, Zhengzhou, China, 10–13 August 2023; pp. 347–358. [Google Scholar]
  99. Sanderson, E.; Matuszewski, B.J. FCN-transformer feature fusion for polyp segmentation. In Proceedings of the MIUA, Cambridge, UK, 27–29 July 2022; pp. 892–907. [Google Scholar]
  100. Subedi, R.; Gaire, R.R.; Ali, S.; Nguyen, A.; Stoyanov, D.; Bhattarai, B. A Client-Server deep federated learning for cross-domain surgical image segmentation. arXiv 2023, arXiv:2306.08720. [Google Scholar]
  101. Tomar, N.K.; Jha, D.; Bagci, U. DilatedSegnet: A deep dilated segmentation network for polyp segmentation. In Proceedings of the MMM, Qui Nhon, Vietnam, 5–8 April 2023; pp. 334–344. [Google Scholar]
  102. Duc, N.T.; Oanh, N.T.; Thuy, N.T.; Triet, T.M.; Dinh, V.S. ColonFormer: An efficient transformer based method for colon polyp segmentation. IEEE Access 2022, 10, 80575–80586. [Google Scholar] [CrossRef]
  103. Wang, J.; Huang, Q.; Tang, F.; Meng, J.; Su, J.; Song, S. Stepwise feature fusion: Local guides global. In Proceedings of the MICCAI, Singapore, 18–22 September 2022; pp. 110–120. [Google Scholar]
  104. Zhu, M.; Chen, Z.; Yuan, Y. FedDM: Federated weakly supervised segmentation via annotation calibration and gradient de-conflicting. IEEE Trans. Med. Imaging 2023, 42, 1632–1643. [Google Scholar] [CrossRef] [PubMed]
  105. Nasr, M.; Islam, M.M.; Shehata, S.; Karray, F.; Quintana, Y. Smart healthcare in the age of AI: Recent advances, challenges, and future prospects. IEEE Access 2021, 9, 145248–145270. [Google Scholar] [CrossRef]
  106. Rath, K.C.; Khang, A.; Rath, S.K.; Satapathy, N.; Satapathy, S.K.; Kar, S. Artificial intelligence (AI)-enabled technology in medicine-advancing holistic healthcare monitoring and control systems. In Computer Vision and AI-Integrated IoT Technologies in the Medical Ecosystem; CRC Press: Boca Raton, FL, USA, 2024; pp. 87–108. [Google Scholar]
  107. Mansour, R.F.; El Amraoui, A.; Nouaouri, I.; Díaz, V.G.; Gupta, D.; Kumar, S. Artificial intelligence and internet of things enabled disease diagnosis model for smart healthcare systems. IEEE Access 2021, 9, 45137–45146. [Google Scholar] [CrossRef]
  108. Moreira, M.W.; Rodrigues, J.J.; Korotaev, V.; Al-Muhtadi, J.; Kumar, N. A comprehensive review on smart decision support systems for health care. IEEE Syst. J. 2019, 13, 3536–3545. [Google Scholar] [CrossRef]
  109. Katal, N. Ai-driven healthcare services and infrastructure in smart cities. In Smart Cities; CRC Press: Boca Raton, FL, USA, 2024; pp. 150–170. [Google Scholar]
Figure 1. SplitFed architecture with multiple clients [5], where each client uses the split UNet as its model [33].
Figure 1. SplitFed architecture with multiple clients [5], where each client uses the split UNet as its model [33].
Bioengineering 13 00104 g001
Figure 2. Semantic segmentation networks’ development timeline. Networks marked with green stars are those included in our work.
Figure 2. Semantic segmentation networks’ development timeline. Networks marked with green stars are those included in our work.
Bioengineering 13 00104 g002
Figure 3. Split UNet and SegNet architectures. (a) Split UNet architecture. Ash-colored segments are the layers on the client side, and pink-colored segments are the layers on the server side. Note that this naming convention is applied to all other network figures in this paper. (b) Split SegNet architecture. A similar architecture to Split UNet without the skip connections.
Figure 3. Split UNet and SegNet architectures. (a) Split UNet architecture. Ash-colored segments are the layers on the client side, and pink-colored segments are the layers on the server side. Note that this naming convention is applied to all other network figures in this paper. (b) Split SegNet architecture. A similar architecture to Split UNet without the skip connections.
Bioengineering 13 00104 g003
Figure 4. Split DeepLab architectures. (a) Split DeepLabV3 with ResNet50 as the backbone. (b) Split DeepLabV3+ architecture with the modified version of Xception as the backbone.
Figure 4. Split DeepLab architectures. (a) Split DeepLabV3 with ResNet50 as the backbone. (b) Split DeepLabV3+ architecture with the modified version of Xception as the backbone.
Bioengineering 13 00104 g004
Figure 5. Split RefineNet and SUNet architectures. (a) Split RefineNet architecture. (b) Split SUNet architecture, including inception-dense modules in both encoder and decoder.
Figure 5. Split RefineNet and SUNet architectures. (a) Split RefineNet architecture. (b) Split SUNet architecture, including inception-dense modules in both encoder and decoder.
Bioengineering 13 00104 g005
Figure 6. Split CGNet and DUCK-Net architectures. (a) Split CGNet architecture. (b) Split version of the DUCK-Net architecture.
Figure 6. Split CGNet and DUCK-Net architectures. (a) Split CGNet architecture. (b) Split version of the DUCK-Net architecture.
Bioengineering 13 00104 g006
Figure 7. Split Attention UNet and Swin-UNet architectures. (a) Split Attention UNet architecture. (b) Split Swin-UNet architecture.
Figure 7. Split Attention UNet and Swin-UNet architectures. (a) Split Attention UNet architecture. (b) Split Swin-UNet architecture.
Bioengineering 13 00104 g007
Figure 8. Comparison of results: Blastocyst Dataset. The blue-colored segment is applied to the centralized models; the orange-colored segment is applied to the locally centralized models; and the ash-colored segment is applied to the SplitFed models. Note that the same convention is applied in Figure 9 and Figure 10.
Figure 8. Comparison of results: Blastocyst Dataset. The blue-colored segment is applied to the centralized models; the orange-colored segment is applied to the locally centralized models; and the ash-colored segment is applied to the SplitFed models. Note that the same convention is applied in Figure 9 and Figure 10.
Bioengineering 13 00104 g008
Figure 9. Comparison of results: HAM10K Dataset.
Figure 9. Comparison of results: HAM10K Dataset.
Bioengineering 13 00104 g009
Figure 10. Comparison of results: KVASIR-SEG Dataset.
Figure 10. Comparison of results: KVASIR-SEG Dataset.
Bioengineering 13 00104 g010
Table 1. Performance Comparison: Centralized (C), Locally Centralized (L), and SplitFed (S) Models (Average IoU metrics).
Table 1. Performance Comparison: Centralized (C), Locally Centralized (L), and SplitFed (S) Models (Average IoU metrics).
ModelBlastocyst DatasetHAM10K DatasetKVASIR-SEG Dataset
C L S C L S C L S
UNet0.86430.77260.85930.86720.83200.86400.82710.69460.8042
SegNet0.84750.74160.84750.84260.77730.86200.73370.57130.7669
SUNet0.84870.75660.85040.86790.82410.85390.72800.60060.7233
DeepLabV30.87680.80160.83690.86990.77150.86960.84380.77150.8262
DeepLabV3+0.87740.68340.85910.87150.83110.82620.82640.69650.8278
RefineNet0.78810.69480.81810.85840.81610.84030.70830.66690.7652
Attention UNet0.86730.69900.86050.86540.82410.86990.82360.69910.7961
Swin-UNet0.80740.62830.81420.84920.77680.84780.78710.54830.6642
CGNet0.84330.72870.78910.87280.83820.84900.83540.68680.8110
DUCK-Net0.87250.79940.83210.86520.83890.86000.88240.77780.7800
Average over models0.84930.73060.83670.86300.81300.85430.79960.67130.7765
Table 2. Qualitative comparison of three samples from SplitFed training.
Table 2. Qualitative comparison of three samples from SplitFed training.
SampleGround TruthUNetSegNetSUNetDeepLab V3DeepLab V3+RefineNetAttention UNetSwin-UNetCGNetDUCK-Net
Blastocyst Dataset
 Bioengineering 13 00104 i001
Blast_PCRM_R14-0411a.BMP
Bioengineering 13 00104 i002Bioengineering 13 00104 i003Bioengineering 13 00104 i004Bioengineering 13 00104 i005Bioengineering 13 00104 i006Bioengineering 13 00104 i007Bioengineering 13 00104 i008Bioengineering 13 00104 i009Bioengineering 13 00104 i010Bioengineering 13 00104 i011Bioengineering 13 00104 i012
HAM10K Dataset
Bioengineering 13 00104 i013
ISIC_0024308.jpg
Bioengineering 13 00104 i014Bioengineering 13 00104 i015Bioengineering 13 00104 i016Bioengineering 13 00104 i017Bioengineering 13 00104 i018Bioengineering 13 00104 i019Bioengineering 13 00104 i020Bioengineering 13 00104 i021Bioengineering 13 00104 i022Bioengineering 13 00104 i023Bioengineering 13 00104 i024
KVASIR-SEG Dataset
  Bioengineering 13 00104 i025
cju7bgnvb1sf808717qa799ir.jpg
Bioengineering 13 00104 i026Bioengineering 13 00104 i027Bioengineering 13 00104 i028Bioengineering 13 00104 i029Bioengineering 13 00104 i030Bioengineering 13 00104 i031Bioengineering 13 00104 i032Bioengineering 13 00104 i033Bioengineering 13 00104 i034Bioengineering 13 00104 i035Bioengineering 13 00104 i036
Table 3. Performance-wise comparison for SplitFed networks (Three top performing networks in each category are listed).
Table 3. Performance-wise comparison for SplitFed networks (Three top performing networks in each category are listed).
ModelCentralized ModelsLocally Centralized ModelsSplitFed Models
Blastocyst datasetDeepLabV3+, DeepLabV3, DUCK-NetDeepLabV3, DUCK-Net, CGNetAttention UNet, DeepLabV3+, SUNet
HAM10K datasetCGNet, DeepLabV3+, DeepLabV3DUCK-Net, CGNet, DeepLabV3+Attention UNet, DeepLabV3, UNet
KVASIR-SEG datasetDUCK-Net, DeepLabV3, CGNetDUCK-Net, DeepLabV3, DeepLabV3+DeepLabV3+, DeepLabV3, CGNet
Table 4. Computational complexity-wise comparison (FLOPs are measured in “MAC”, which stands for “Multiply Accumulate”).
Table 4. Computational complexity-wise comparison (FLOPs are measured in “MAC”, which stands for “Multiply Accumulate”).
ModelTrainable Parameters (TPs)FLOPsTP as a %. of UNetFLOPs as a %. of UNet
UNet7.76M10.52 GMAC1%1%
SegNet9.44M7.04 GMAC1.22%0.67%
SUNet14.1M24 GMAC1.82%2.28%
DeepLabV328.32M12.83 GMAC3.64%1.21%
DeepLabV3+54.70M15.85 GMAC7.05%1.50%
RefineNet118M50.24 GMAC15.20%4.77%
Attention UNet34.87M51.03 GMAC4.50%4.85%
Swin-UNet41.38M8.67 GMAC5.33%0.82%
CGNet0.30M541.69 MMAC0.039%0.05%
DUCK-Net22.67M12.55 GMAC2.92%1.19%
Table 5. Comparison of layer proportions, trainable parameters (TPs), and FLOPs across FE, SS, and BE for the ten segmentation models.
Table 5. Comparison of layer proportions, trainable parameters (TPs), and FLOPs across FE, SS, and BE for the ten segmentation models.
ModelLayer ProportionsTrainable Parameters (TP)FLOPs
FE SS BE FE SS BE FE SS BE
UNet2
(8.0%)
21
(84.0%)
2
(8.0%)
0.002M
(0.03%)
7.75M
(99.85%)
0.009M
(0.12%)
123.73
MMAC
13.09
GMAC
614.5
MMAC
SegNet2
(6.25%)
28
(87.50%)
2
(6.25%)
0.001M
(0.01%)
9.43M
(99.90%)
0.009M
(0.10%)
0.057
GMAC
8.46
GMAC
0.61
GMAC
SUNet4
(6.78%)
52
(88.14%)
3
(5.08%)
0.006M
(0.47%)
13.91M
(98.72%)
0.11M
(0.81%)
0.415
GMAC
38.817
GMAC
1.030
GMAC
DeepLabV31
(1.61%)
59
(95.16%)
2
(3.23%)
0.009M
(0.03%)
28.24M
(99.69%)
0.074M
(0.26%)
0.31
GMAC
27.98
GMAC
0.77
GMAC
DeepLabV3+1
(1.30%)
73
(94.81%)
3
(3.89%)
0.009M
(0.02%)
54.403M
(99.41%)
0.32M
(0.57%)
0.19
GMAC
31.85
GMAC
0.49
GMAC
RefineNet1
(0.71%)
138
(98.57%)
1
(0.71%)
0.009M
(0.01%)
117.85M
(99.95%)
0.012M
(0.01%)
0.31
GMAC
59.84
GMAC
0.08
GMAC
Attention UNet1
(2.63%)
35
(92.11%)
2
(5.26%)
0.35M
(0.01%)
34.82M
(99.88%)
0.009M
(0.12%)
0.042
GMAC
13.10
GMAC
0.61
GMAC
Swin-UNet2
(6.9%)
24
(82.8%)
3
(10.3%)
0.29M
(0.7%)
39.23M
(94.8%)
1.86M
(4.5%)
0.08
GMAC
8.18
GMAC
0.42
GMAC
CGNet3
(11.11%)
13
(48.15%)
11
(40.74%)
0.02M
(6.69%)
0.06M
(20.13%)
0.22M
(73.18%)
0.32
GMAC
0.58
GMAC
0.66
GMAC
DuckNet2
(6.67%)
26
(86.67%)
2
(6.67%)
0.11M
(0.50%)
21.08M
(93.00%)
1.47M
(6.50%)
0.11
GMAC
11.55
GMAC
0.89
GMAC
Table 6. Performance comparison with other existing methods (NA: Not Applicable/Relevant experimental results are not reported or experiments have not been conducted in those studies).
Table 6. Performance comparison with other existing methods (NA: Not Applicable/Relevant experimental results are not reported or experiments have not been conducted in those studies).
SoTA ResearchCentralized Models IoUFederated Models IoUSplitFed Models IoU
Blastocyst Dataset
Our Previous research0.798 (BLAST-NET [92]), 0.817 [88]0.810 [88]0.825 [88]
HAM10K Dataset
FedPerl (Efficient-Net) [93]0.7690.747NA
FedMix (UNet) [94]NA0.819 ± 1.7NA
MALUNet [95]0.802NANA
FedZaCt (UNet) [96]0.8550.856NA
FedZaCt (DeepLabV3+) [96]0.8610.863NA
Chen et al. [97]NA0.892NA
FedDTM (UNet) [98]NA0.7994NA
KVASIR-SEG Dataset
DUCK-Net [17]0.9502NANA
FCN-Transformer [99]0.9220NANA
MSRF-Net [72]0.8508NANA
PraNet [70]0.7286NANA
HRNetV2 [60]0.8530NANA
Subedi et al. [100]0.810.823NA
DilatedSegNet [101]0.8957NANA
DeepLabV3+ [101]0.8837NANA
Colonformer [102]0.8770.876NA
SSFormer-S [103]0.8743NANA
SSFormer-L [103]0.8905NANA
FedDM [104]0.5275 ± 0.00020.6877 ± 0.0308NA
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shiranthika, C.; Kafshgari, Z.H.; Hadizadeh, H.; Saeedi, P. MedSegNet10: A Publicly Accessible Network Repository for Split Federated Medical Image Segmentation. Bioengineering 2026, 13, 104. https://doi.org/10.3390/bioengineering13010104

AMA Style

Shiranthika C, Kafshgari ZH, Hadizadeh H, Saeedi P. MedSegNet10: A Publicly Accessible Network Repository for Split Federated Medical Image Segmentation. Bioengineering. 2026; 13(1):104. https://doi.org/10.3390/bioengineering13010104

Chicago/Turabian Style

Shiranthika, Chamani, Zahra Hafezi Kafshgari, Hadi Hadizadeh, and Parvaneh Saeedi. 2026. "MedSegNet10: A Publicly Accessible Network Repository for Split Federated Medical Image Segmentation" Bioengineering 13, no. 1: 104. https://doi.org/10.3390/bioengineering13010104

APA Style

Shiranthika, C., Kafshgari, Z. H., Hadizadeh, H., & Saeedi, P. (2026). MedSegNet10: A Publicly Accessible Network Repository for Split Federated Medical Image Segmentation. Bioengineering, 13(1), 104. https://doi.org/10.3390/bioengineering13010104

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop