Next Article in Journal
Wearable Haptic Music Player with Multi-Feature Extraction Using Spectral Flux and Yin Algorithms
Previous Article in Journal
High-Precision Visual SLAM for Dynamic Scenes Using Semantic–Geometric Feature Filtering and NeRF Maps
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Source-Free Domain Adaptation for Medical Image Segmentation via Mutual Information Maximization and Prediction Bank

School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(18), 3656; https://doi.org/10.3390/electronics14183656
Submission received: 1 August 2025 / Revised: 2 September 2025 / Accepted: 12 September 2025 / Published: 15 September 2025

Abstract

Medical image segmentation faces significant challenges due to domain shift between different clinical centers and data privacy restrictions. Current source-free domain adaptation methods for medical images suffer from three critical limitations, including unstable training caused by noisy pseudo-labels and poor handling of foreground-background imbalance where critical structures like optic cup occupy extremely small regions. Additionally, strict privacy regulations often prevent access to source domain data during adaptation. To address these limitations, this paper proposes a source-free domain adaptation approach based on mutual information optimization for fundus image segmentation. The method incorporates a teacher–student network to ensure training stability and a mutual information maximization algorithm to reduce pseudo-label noise naturally. Furthermore, a prediction bank is constructed to handle class imbalance by leveraging complete statistics. Experimental results on fundus segmentation datasets demonstrate better performance, achieving 91.74% average Dice coefficient on Drishti-GS and 87.80% on RIM-ONE-r datasets, outperforming current methods. This work provides a practical solution for cross-institutional medical image analysis while preserving data privacy, with significant potential for eye disease diagnosis and other medical applications requiring robust domain adaptation.

1. Introduction

Medical image segmentation represents a fundamental cornerstone in computer-aided diagnosis systems, where accurate depiction of organic structures directly assists clinical decision-making and patient outcomes [1]. Traditional segmentation models demonstrate exceptional performance when sufficient well-annotated training data is available, yet the high cost of annotation creates significant bottlenecks for model development across diverse clinical scenarios [2]. To address this challenge, unsupervised domain adaptation (UDA) methods have emerged as a solution by training adaptive models that integrate unlabeled target domain images with labeled source domain images. They effectively alleviate the cost of data annotation to some extent. Despite the remarkable performance achieved by UDA approaches [3,4], fundamental limitations persist in medical image segmentation applications. The high sensitivity of medical images concerning privacy and copyright protection poses substantial obstacles, as labeled source domain images are often subject to strict institutional restrictions and may be completely unavailable during the adaptation process.
This ubiquitous practical constraint, commonly encountered in real-world clinical deployments, has driven extensive research into source-free domain adaptation (SFDA) methods [5,6]. These approaches aim to adapt a source-trained model to the target domain using exclusively unlabeled target data, providing a more adaptable and feasible solution for cross-institutional medical image segmentation development while respecting privacy regulations and data sharing limitations. Since training on both the source and target domains simultaneously is not possible, effective denoising plays a key role in this task. Experiments [7] rely on model-generated pseudo-labels, creating a delicate balance between leveraging confident predictions and avoiding error propagation. Advances have introduced various denoising mechanisms to enhance pseudo-label quality, including pixel-level filtering, class-aware thresholding, and uncertainty-based selection criteria.
Despite significant progress in source-free adaptation, three critical limitations persist that particularly affect medical image segmentation performance. The primary challenge involves employing a posterior-like approach for denoising the results, which may lead to error accumulation and performance degradation. Additionally, the absence of ground-truth supervision creates training instability, as models may converge to suboptimal solutions when relying exclusively on self-generated feedback. Most critically for medical applications, existing methods inadequately address the class imbalance characteristic of anatomical structures, where clinically significant regions such as optic cups occupy minimal image areas compared to background tissue. This may lead to biased learning dynamics that favor dominant classes at the expense of diagnostic precision.
To address the aforementioned issues, we propose Mutual Information Maximization and Prediction Bank (MIMPB), a source-free domain adaptation approach for medical image segmentation. The algorithm formulates an optimization problem consisting of two key modules. First, the mutual information maximization module minimizes the relative entropy between pseudo-labels and model output probabilities, effectively reducing the discrepancy between distributions and naturally mitigating pseudo-label noise. This approach aims to fully utilize the information in high-quality labels, thereby reducing the uncertainty in the generated data. Second, the prediction bank effectively addresses the foreground–background imbalance in medical images by analyzing global statistics across the entire dataset. By storing prediction results of all pixels and calculating separate average losses for foreground and background classes, the prediction bank calibrates the loss function to ensure balanced learning between classes, significantly improving segmentation accuracy for minority classes like the optic cup. These two components work synergistically within a teacher–student framework to enhance model stability and performance for source-free domain adaptation in medical image segmentation tasks.
Our contributions can be summarized as follows:
  • First, we introduce a novel mutual information maximization approach that fundamentally addresses pseudo-label noise by formulating an optimization problem from information theory principles rather than relying on post-processing techniques.
  • Second, we develop a comprehensive prediction bank mechanism that effectively tackles the class imbalance problem in medical images by analyzing dataset-wide statistics and dynamically adjusting the loss function weights.
  • Third, we present an enhanced teacher–student framework that seamlessly integrates these aforementioned two components, enabling stable and effective source-free domain adaptation for medical image segmentation without requiring access to source domain data during training.
This paper is organized as follows. Section 2 reviews related work on domain adaptation methods and source-free domain adaptation techniques in medical image segmentation. Section 3 presents our proposed MIMPB framework, including the teacher–student network, mutual information maximization algorithm, and prediction bank mechanism. Experimental results on fundus segmentation datasets with performance comparisons, ablation studies, and parameter analysis are discussed in Section 4. Finally, Section 5 concludes this paper and discusses future research directions.

2. Related Work

Recent developments in segmenting optic disc and cup have increasingly focused on addressing domain shift challenges across different medical institutions and imaging devices. While supervised and unsupervised domain adaptation methods have shown promising results, they often require access to source domain data during the adaptation process. This requirement has led to the emergence of source-free domain adaptation approaches that can adapt pre-trained models using only unlabeled target domain data.

2.1. Domain Adaptation

Domain adaptation methods represent a class of transfer learning techniques designed to address the challenge of domain shift between source and target domains. In the medical imaging field, these methods can be broadly categorized into supervised domain adaptation (SDA) and unsupervised domain adaptation (UDA) approaches, each offering distinct advantages and addressing different scenarios of data availability.
Supervised domain adaptation methods utilize labeled data from both source and target domains to facilitate knowledge transfer. These approaches can be categorized into fine-tuning strategies and multi-step adaptation methods. Fine-tuning strategies involve adapting pre-trained models through parameter adjustment on target domain data. Recent studies [8] have demonstrated that fine-tuning convolutional neural networks pre-trained on brain MRI scans can significantly improve brain lesion segmentation performance using limited target samples. Similarly, successful adaptations have been reported for VGG networks on MRI images [9] and AlexNet architectures for lesion detection [10]. Multi-step adaptation approaches address scenarios with insufficient target samples by employing intermediate adaptation stages. Enhanced approaches [11] first fine-tune a residual network on a relatively large medical dataset then train on smaller target domain data, showing superior performance over direct transfer methods. Semi-supervised domain adaptive methods [12,13] have emerged to address the challenging scenario where both limited annotation and domain shift coexist in medical image segmentation. These approaches utilize small amounts of labeled target data alongside abundant unlabeled data to facilitate domain adaptation, offering the advantage of reduced annotation burden while maintaining robust cross-domain performance through effective knowledge transfer mechanisms.
Unsupervised domain adaptation methods address the more challenging scenario where only labeled source domain data and unlabeled target domain data are available during training. These methods primarily rely on feature alignment or image alignment strategies to mitigate domain shift. Feature alignment approaches [14,15] primarily leverage adversarial learning frameworks to learn domain-invariant representations by constraining feature distributions across domains. These methods help reduce the model’s sensitivity to image quality variations across different domains by encouraging the model to focus on task-relevant features while suppressing domain-specific variations. Researchers [16] have also focused on low-level feature training by fixing the weights of high-level networks, thereby better adapting to image assessment requirements. Image alignment strategies typically employ generative adversarial networks to achieve domain-level alignment. A noise adaptation generative adversarial network [17] has been proposed to address fundus vessel segmentation by treating domain adaptation as a noise style transfer problem, incorporating dual discriminators to ensure content similarity and noise pattern consistency. Other studies [18,19] focus on advanced generative modeling and masked autoencoding techniques to bridge domain gaps by learning shared representations across different medical imaging modalities, so as to improve existing methods and provide practical solutions for multi-institutional medical studies with data sharing restrictions.

2.2. Source-Free Domain Adaptation (SFDA)

Source-free domain adaptation methods can be categorized into entropy-based methods and output alignment-based methods. Entropy-based methods [20,21] utilize a trained model to predict a small amount of unlabeled target domain data. They update the model’s parameters for adaptation to the target domain based on the entropy of the predictions. These methods can enhance discriminability and generalization ability. However, they suffer from several issues, such as updating only the lower-layer batch normalization parameters while neglecting high-level semantic features and encountering uncertainty in supervision information due to differences between the source and target domains, which may lead to erroneous model updates. Output alignment-based methods [22] aim to mitigate uncertainty in predictions. Most approaches use denoising autoencoders or other denoising techniques to refine results. The typical process involves first training a segmentation model on the source domain with label supervision to optimize performance, then allowing the model to adapt to a small amount of unlabeled target domain images. The model is trained and updated using denoised segmentation results, mainly adjusting shallow-layer batch normalization parameters. However, these methods still focus primarily on adapting low-level features of target domain images, providing limited improvements in segmentation accuracy.
Source-free domain adaptation methods can also be broadly divided into generation-based methods [23,24] and pseudo-labeling methods [25,26]. Generation-based methods face challenges in learning to generate meaningful features and have limited applicability. In contrast, pseudo-labeling methods are relatively simpler and more general, with a strong research foundation, particularly in medical image segmentation. For instance, some researchers [27] proposed a method that minimizes an unsupervised entropy loss function defined on target domain data while incorporating domain-invariant prior information from the segmentation region to guide model optimization. This method has been applied to spinal, prostate, and cardiac segmentation tasks. Another study [28] employed pseudo-label generation for source-free domain adaptive image segmentation, integrating self-training in a two-stage adaptation process. During the specific target adaptation stage, the study introduced an ensemble entropy minimization loss and a selective voting strategy to enhance pseudo-label generation by reducing high-entropy regions. In the specific task adaptation stage, the approach leveraged a teacher–student network to further refine pseudo-labels, significantly improving segmentation performance in the target domain. Additionally, the ProSFDA method [29] was proposed as a source-free domain adaptation approach featuring both a generation phase and an adaptation phase. In the generation phase, class-specific source-like images are generated using the statistical information from a pre-trained source model and Fourier transform. In the adaptation phase, a specially designed distillation module enables feature-level adaptation. This module includes a domain distillation loss to constrain relational knowledge transfer and a domain contrastive loss to minimize domain discrepancies.

3. Methdology

Our proposed MIMPB framework includes two key components. They formulate the adaptation problem as an information-theoretic optimization that naturally reduces pseudo-label noise while maintaining training stability. The framework seamlessly integrates mutual information maximization with a novel prediction bank mechanism to achieve balanced learning across imbalanced anatomical structures.

3.1. Framework of MIMPB

Source-free domain adaptation of medical images segmentation aims to transfer a source model f s , which is trained on labeled source domain images X s , to the target domain using only unlabeled target domain images X t . The ground truth labels in the source domain are denoted as Y i { 0 , 1 } H × W × C , where H, W, and C represent the image height, width, and number of categories, respectively, and i indicates the i-th image. If the j-th pixel of the i-th image belongs to the c-th category, then y i j c = 1 ; otherwise, y i j c = 0 .
Based on this, we aim to construct and train a deep neural network f θ : x y . Through training and optimization of this network, the ultimate goal is to achieve accurate segmentation of the optic disc and optic cup. To effectively address key challenges such as data denoising, model training stability, and the imbalance between foreground and background information, we propose a solution based on mutual information optimization. Specifically, this method leverages the properties of mutual information to fundamentally eliminate noise interference in the results. Additionally, a teacher–student network is introduced to enhance the stability of the training process. Moreover, a prediction bank is constructed to balance foreground and background information, thereby improving overall model performance.
The primary objective of the method is to enhance the model’s performance in a specific target domain by leveraging knowledge obtained from the source domain. The model framework is shown in Figure 1. The detailed training procedure is shown in Algorithm 1.
Algorithm 1 Training procedure of MIMPB
Require: 
Source model f s , Target data D T = { X i t } i = 1 N (unlabeled)
Ensure: 
Adapted model parameters θ s , θ t
 1:
begin
 2:
   Initialize θ s , θ t f s
 3:
   Build prediction bank PB with f θ t ( D T )
 4:
   for epoch = 1 to E do
 5:
      for each batch B D T do
 6:
         Generate X w WeakAug ( B ) , X s StrongAug ( B )
 7:
         // Teacher prediction
 8:
          p w σ ( f θ t ( X w ) ) based on Equation (1)
 9:
         // MIM optimization
  10:
         Optimize W via MIM based on Equation (3)
  11:
         Compute optimal pseudo-labels y ^ r e based on Equation (18)
  12:
         // Prediction bank weighting
  13:
         Calculate η f g , η b g from PB based on Equations (20) and (21)
  14:
         // Student prediction and training
  15:
          p s σ ( f θ s ( X s ) )
  16:
          L L s e g (with weight η f g / η b g ) + α L M I based on Equations (11), (22) and (23)
  17:
         Update θ s via gradient descent on L
  18:
         // Teacher update
  19:
          θ t λ θ t + ( 1 λ ) θ s based on Equation (2)
  20:
         Update PB with new predictions
  21:
      end for
  22:
   end for
  23:
end

3.2. Teacher–Student Network

To avoid error accumulation and ensure a more stable training process, we adopt a weak–strong augmentation teacher–student network. The teacher and student models share the same structure but differ in their weight update methods. Specifically, for each image X i , a weakly augmented version X i w is generated through image flipping and resizing and is fed into the teacher network. Meanwhile, a strongly augmented version X i s is generated and input into the student network. The strong augmentation methods include random erasing, contrast adjustment, and impulse noise.
Finally, the teacher network’s output is used to guide the student network, with the mathematical formulation as follows:
L = E x s , y ^ w [ L b c e ] ,
where x s X i s represents the pixels in the image, y ^ w denotes the pseudo-labels obtained from the teacher network, and L b c e refers to the binary cross-entropy loss (BCE). Using this network offers the following advantages: First, given that fundus image datasets are generally small in scale, insufficient training data makes the model highly prone to overfitting. To effectively mitigate this issue, the algorithm incorporates image augmentation techniques to enhance the diversity of the training set. Second, applying different random augmentation methods during training introduces consistency regularization, which helps group images with similar semantics into the same category. This, in turn, facilitates the formation of more discriminative feature representations.
The teacher model does not update its parameters through loss backpropagation but instead updates them using the exponential moving average of the student model’s parameters. The update formula is as follows:
θ t = λ θ t 1 + ( 1 λ ) θ t ,
where θ t represents the teacher model’s parameters at time step t, θ t denotes the student model’s parameters at time step t, and λ is the smoothing coefficient hyperparameter. This network enhances generalization ability, stabilizes training, and effectively utilizes unlabeled data.

3.3. Mutual Information Optimization Algorithm

The algorithm formulates an optimization problem consisting of two parts. The first part minimizes the relative entropy between the pseudo-labels and the model’s output probabilities. The goal is to reduce the gap between the pseudo-label distribution and the probability distribution, thereby learning pseudo-labels that better represent the probability distribution. The second part maximizes the mutual information between the pseudo-labels and the data. This is motivated by the idea that high-quality labels can reduce uncertainty within the data.
Given a segmentation model f θ , its output is a K-dimensional probability vector, where K is the predefined number of segmentation classes. For an input image X = { x i } i = 1 B , the model’s output probabilities are defined as P = { p i } i = 1 B R K , where p i = s o f t m a x ( f θ ( x i ) ) . This method aims to learn the optimized probability results W = { w i } i = 1 B for training the model, where B represents the image size.
Consider the optimization problem:
W = a r g m i n W Δ K 1 B i = 1 B D K L ( w i , p i ) β I ( Y W ; B ) ,
where Y w { 1 , , K } represents the pseudo-labels obtained from the distribution W = { w i } i = 1 B , B { 1 , , B } denotes the pixel position index, and Δ K : = { w R + K | w T 1 K = 1 } . The hyperparameter β [ 0 , 1 ) in Equation (3) is a crucial coefficient that balances the two objectives of the optimization problem. It controls the trade-off between minimizing the KL divergence and maximizing mutual information. A higher value of β places greater emphasis on the mutual information term, encouraging the generation of diverse and well-structured pseudo-labels that can prevent the model from collapsing into trivial solutions. Conversely, a lower value of β ensures that the resulting pseudo-labels W remain closer to the model’s original predictions P. The selection of β thus allows for tuning the balance between model confidence and label diversity.
By expanding the formula in Equation (3) using the definitions of KL divergence and mutual information, we can get:
W i = a r g m i n W Δ K 1 B i = 1 B j = 1 K w i j l o g p i j + 1 β B i = 1 B j = 1 K w i j l o g w i j + β j = 1 K w ^ j l o g w ^ j .
It is clear that this optimization problem is a strictly convex optimization problem. Therefore, the optimization problem is constructed using the Karush–Kuhn–Tucker (KKT) conditions:
L ( W ) = f ( W ) + λ g ( W ) ,
where,
f ( W ) = 1 B i = 1 B j = 1 K w i j l o g p i j + 1 β B i = 1 B j = 1 K w i j l o g w i j + β j = 1 K w ^ j l o g w ^ j ,
g ( W ) = w T 1 K 1 .
Finally, we can get
w i j = p i j 1 1 β w j ^ β 1 β 1 j p i j 1 β w j ^ β 1 β ,
and
w j ^ = [ 1 B i p i j 1 1 β k p i k 1 1 β w k ^ β 1 β ] 1 β .
Based on the above formula, { u j ( n ) } j = 1 I t e r a can be iteratively constructed to obtain the optimal solution, and the iterative formula is:
u j ( n + 1 ) = [ 1 B i p i j 1 1 β k p i k 1 1 β ( u j ( n ) ) β 1 β ] 1 β .
As proved in [30], Equation (10) is a fixed-point iteration for the strictly convex objective and converges to the unique optimum. Through this iterative method, the optimal pseudo-label distribution can be obtained, which is then used for model training. That is, the loss function for a single image can be defined as:
L M I = 1 B × K i = 1 B j = 1 K u i j l o g p i j .
It is important to note that since this algorithm uses a teacher–student network, u i j is the optimal pseudo-label obtained by optimizing the probability values p i j s output by the teacher network. The specific process is not elaborated on here.
The above optimization method minimizes the relative entropy between pseudo labels and the model’s output probabilities. During model training, there is often a discrepancy between the distributions of pseudo labels and output probabilities. Mutual information maximization can effectively reduce this gap, making the two distributions as close as possible. In this way, the model can better exploit the latent features and patterns in the data, thereby gaining a deeper understanding of its essence and underlying structure. Finally, the method leverages the output probabilities of the teacher network to guide those of the student network, further enhancing the stability of model training.

3.4. Prediction Bank

The teacher–student network and mutual information optimization methods effectively address the stability and denoising issues during model training. However, in the fundus segmentation task, there is still the challenge of foreground–background imbalance. Specifically, foreground objects (e.g., the optic cup) usually occupy very small areas in the image, causing the majority of pixels to belong to the background. If binary cross-entropy (BCE) loss is directly used to update the student model, the background class will dominate the loss calculation, weakening the supervisory signal for the foreground class. The prediction bank proposed in this method is designed to alleviate this issue to some extent. In practice, the model performs separate segmentation for the optic cup and the optic disc. When segmenting the cup, the disc and fundus are regarded as background, while when segmenting the disc, the remaining regions serve as background. Thus, the prediction bank is applied in a binary segmentation manner for each structure.
When processing images, the simplest way to address the foreground–background imbalance issue is to separately count the number of pixels belonging to the foreground and background classes in each image and then design a loss weighting function based on these statistics. In standard supervised learning tasks, this strategy may be effective since the labels are reliable. However, for pseudo-labels, performing statistical analysis based on a single image is highly risky. To mitigate this problem, this algorithm constructs a prediction bank to analyze the class imbalance across the entire dataset, using this global information to calibrate the loss for each image. Specifically, the prediction bank stores the prediction results of all pixels in all images and calculates the average loss for the foreground and background:
η k f g = i L i , k × 1 [ y ^ i , k = 1 ] i 1 [ y ^ i , k = 1 ] ,
η k b g = i L i , k × 1 [ y ^ i , k = 0 ] i 1 [ y ^ i , k = 0 ] .
L is the segmentation loss, and its specific form will be detailed in the next section. Here, f g and b g refer to the foreground and background, respectively. This method uses the mean of the loss, rather than averaging based on the number of pixels, because the loss of each pixel reflects the “difficulty level” of that pixel. This property allows pixels with more informative features to be assigned greater weight. The average loss computed based on Formulas (12) and (13) can be further optimized and calibrated as follows:
L ^ b c e = E x , k [ y ^ k l o g ( p k ) + η k f g / η k b g ( 1 y ^ k ) l o g ( 1 p k ) ] .
Additionally, considering that in actual model prediction processes, the vast majority of predictions generally exhibit high confidence, meaning their values are extremely close to 0 or 1, these highly deterministic predictions contain relatively little information in terms of information entropy. Therefore, to more accurately compute the average loss and enhance the model’s learning capability for complex cases, only those pixels with relatively large loss magnitudes are considered during the calculation. This screening mechanism is specifically implemented by adopting a constraint threshold approach:
| p i , k γ | | y ^ i , k γ | > 0.2 ,
where γ is the probability threshold for converting soft probabilities to hard labels, and p i is the output probability value.
Overall, the calibrated loss ensures fair learning across different categories, thereby alleviating the performance degradation problem caused by class imbalance. The model constructs a prediction bank for images and analyzes class imbalance in the dataset from a global perspective. It effectively avoids risks, making the model more robust and reliable. In addition, the model’s decision-making becomes more scientific and reasonable by adopting the mean loss and assigning weights to key pixels according to the pixel-wise loss.

3.5. Objective Function

This section will elaborate on the objective function of the algorithm, focusing on two main components: the segmentation loss and the loss based on mutual information optimization. It is worth noting that the loss related to mutual information optimization has been explained in detail previously, so it will not be repeated here.
The method for constructing the original pseudo-labels, which directly uses the model’s prediction results to generate pseudo-labels, is as follows:
y ^ i , k = 1 [ p i , k > γ ] ,
where, 1 is the indicator function, and γ is the probability threshold for converting soft probabilities to hard labels. p i and y ^ i are the i-th dimensions of p and y ^ , representing the prediction and pseudo-label for the i-th class, respectively.
Most existing works tend to employ two main strategies to optimize the original methods. First, they propose calibration techniques to obtain better pseudo-labels. Second, they measure uncertainty and apply weights during the loss calculation. These improvements do enhance model performance to some extent, but they still fail to overcome stability issues. This is because the presence of noise directly interferes with model performance, and since the prediction results are used to generate pseudo-labels, errors accumulate, which further impacts the reliability of the model. Additionally, these methods generally overlook the imbalance in foreground and background pixel distributions in images, with the foreground region typically occupying a small proportion. Under these conditions, directly using binary cross-entropy loss for probability updates is highly susceptible to interference from inaccurate contextual relationships, leading to the incorrect propagation of background probabilities to the foreground, thus reducing the accuracy of foreground probabilities. To address this, the proposed algorithm not only introduces a prediction bank but also incorporates a correction mechanism for the probabilities, aiming to further enhance the model’s performance and stability. Specifically:
p i , k r e = p i , k m a x j p j , k .
To achieve more accurate probability representation, it is necessary to perform a correction operation on the reduced probabilities based on the maximum value in the image. After calibration using Formula (17), the output of the maximum probability (e.g., when at the center of a certain region) should approach 1. This approach effectively improves the accuracy and reliability of probability calculations, thereby optimizing the performance of the related algorithm when processing data.
Based on this, pseudo-labels y ^ i , k are generated:
y ^ i , k r e = 1 [ p i , k r e > γ ] ,
select p i , k r e that satisfies the following formula:
| p i r e γ | | y ^ i γ | > 0.2 ,
and calculate η k f g and η k b g :
η k f g = i L i , k × 1 [ y ^ i , k r e = 1 ] i 1 [ y ^ i , k r e = 1 ] ,
η k b g = i L i , k × 1 [ y ^ i , k = 0 ] i 1 [ y ^ i , k = 0 ] .
Finally, the segmentation loss is updated using BCE loss:
L ^ b c e r e = E x , k [ y ^ k r e l o g ( p k r e ) + η k f g / η k b g ( 1 y ^ k r e ) l o g ( 1 p k r e ) ] ,
It is important to note that since this algorithm utilizes a teacher–student network, y ^ k r e is a pseudo-label derived from the probability values output by the teacher network. The specific process is not reiterated here.
Building upon the aforementioned segmentation loss, the mutual information optimization loss is incorporated, resulting in the final objective function:
L = L ^ b c e r e + α L M I ,
where α is the weighting coefficient for the mutual information loss. In the next section, a sensitivity analysis of this coefficient will be conducted to explore its impact on model performance.

4. Experiments

In this section, the performance of the proposed algorithm in the retinal image segmentation task will be evaluated. The method will be compared with state-of-the-art methods, and the obtained results will be analyzed. Additionally, ablation experiments will be conducted and examined. Finally, parameter sensitivity evaluation and visualization analysis will be performed.

4.1. Experiment Settings

Datasets: Three datasets were used, described as follows:
REFUGE Dataset [31]: This dataset is designed for glaucoma assessment and contains a total of 1200 fundus images, all derived from clinical data at the Zhongshan Ophthalmic Center, Sun Yat-sen University. It includes precise annotations for optic disc and cup segmentation, as well as clinical labels for glaucoma, making it suitable for tasks such as optic disc and cup segmentation and glaucoma detection.
RIM-ONE Dataset [32]: This dataset consists of three subsets, with RIM-ONE-r3 being one of them, primarily used for research on optic disc and cup segmentation in ophthalmic images. The subset contains 159 fundus images, with segmentation annotations for the optic disc and cup, making it useful for training and evaluating segmentation models for these structures and supporting the diagnosis and study of ophthalmic diseases.
Drishti-GS Dataset [33]: This dataset is intended for retinal optic nerve head segmentation and glaucoma classification, providing a valuable resource for automated glaucoma assessment research. It includes a total of 101 retinal images. The dataset was curated by clinical researchers based on clinical findings during examinations, with glaucoma patients selected accordingly. Each image was annotated by four glaucoma experts with 3, 5, 9, and 20 years of experience, respectively, to capture inter-observer variability. This dataset can be used for tasks such as fundus optic disc and cup segmentation, as well as glaucoma and normal classification.
In the experiment, datasets collected from different clinical centers with distribution differences were used. Specifically, the REFUGE dataset’s training set is set as the source domain, while the publicly available RIM-ONE-r3 dataset and Drishti-GS dataset are used as two different target domains. The source domain contains 400 labeled training images, and the data from the two target domains are divided into training and testing sets at ratios of 99/60 and 50/51, respectively.
Evaluation Metrics: The Dice coefficient and Average Surface Distance (ASD) are used as evaluation metrics.
The Dice coefficient is a statistical measure used to quantify the similarity between two sets. In segmentation tasks, it evaluates the degree of overlap between the predicted segmentation and the ground truth. Given two sets, | A | and | B | (where | A | represents the set of segmented pixels and | B | represents the ground truth pixel set), the Dice coefficient is calculated as follows:
D i c e = 2 | A B | | A | + | B | ,
where | A B | represents the number of pixels that are labeled as the target in both the segmentation result and the ground truth.
The Average Surface Distance (ASD) is a metric used to measure the surface distance between the segmentation result and the ground truth. It primarily focuses on the difference between the segmentation boundary and the actual boundary, reflecting the precision of the segmentation.
Let S 1 be the set of points on the segmented surface and S 2 be the set of points on the ground truth surface. For each point p 1 S 1 , the shortest distance to S 2 is computed as d 1 . Similarly, for each point p 2 S 2 , the shortest distance to S 1 is computed as d 2 .
A S D = p 1 S 1 d 1 + p 2 S 2 d 2 | S 1 | + | S 2 | .
Implementation Details: The backbone, threshold, software, and hardware used in the experiment:
Experiments adopt DeepLabv3+ as the backbone network. For this task, the probability threshold γ for converting soft probabilities into hard labels is set to 0.75. During training, weak augmentation strategies are applied to the input of the teacher network, specifically through image flipping and resizing. Meanwhile, strong augmentation is applied to the input of the student network, incorporating techniques such as random erasing, contrast adjustment, and impulse noise addition. Introducing perturbations to the input helps deviate predictions from pseudo-labels, thereby increasing training difficulty and encouraging the model to enhance its robustness.
The model is trained using the Adam optimizer, with momentum parameters set to 0.9 and 0.99, a learning rate of 2 × 10 3 , and a batch size of 4. The entire framework is implemented in PyTorch 0.4.1 and runs on an NVIDIA GeForce GTX 1080Ti GPU.
In the preprocessing stage, each image is cropped to a region of 512 × 512 pixels to serve as the input to the network.

4.2. Comparison with Other Methods

The proposed model is compared with several state-of-the-art domain adaptation methods. These include unsupervised domain adaptation (UDA) methods such as the BEAL (Boundary and Entropy-driven Adversarial Learning) method [34] and AdvEnt method [22]. Additionally, source-free domain adaptation (SFDA) methods are discussed, including the SRDA (Source-Relaxed Domain Adaptation) method [35], DAE method [36], DPL (Denoised Pseudo-Labeling) method [37], CBMT (Class-Balanced Mean Teacher) method [38], Crots (Cross-Domain Teacher–Student Learning) method [39], CPR (Context-Aware Pseudo-Label Refinement) method [40] and RDPL method [41]. Table 1 and Table 2 present comparison results on two datasets against various methods. In each column, the best scores are highlighted. For the evaluation metrics, the up-arrow symbol “↑” indicates that a higher score represents better performance, while the down-arrow symbol “↓” signifies that a lower score is better. The symbol “-” indicates that the method did not report corresponding results, while “±” represents the standard deviation of the sample in the dataset. The “S-F” label denotes whether the method is source-free.
During the evaluation, the results corresponding to “W/o DA” (without domain adaptation) in the table can be seen as the lower bound of model performance, while the “Oracle” results, representing supervised learning, serve as the upper bound. From the data in the table, it is evident that the evaluation results of both metrics significantly outperform the most advanced source-free domain adaptation methods and, in some aspects, even show performance improvements over traditional unsupervised domain adaptation methods. Notably, on the Drishti-GS dataset, this method demonstrates a significant performance improvement compared to previous research results.
Specifically, according to the data in Table 1, the proposed method achieves the best performance in the disc and cup segmentation tasks on the RIM-ONE-r dataset. Additionally, an analysis of the variance data reveals that this method demonstrates high stability, which reflects its generalization ability, further validating the effectiveness of the teacher–student network. At the same time, the proposed method has surpassed traditional unsupervised domain adaptation methods such as BEAL, indicating that the algorithm can effectively adapt to the target domain even without access to source domain data.
Based on the data in Table 2, the method shows optimal performance in the cup segmentation task on the Drishti-GS dataset. However, in the disc segmentation task, its performance is slightly inferior to the CBMT method. Nevertheless, when considering the overall results, the average DICE coefficient of this method is still 1.27% higher than that of the CBMT method. The improvement in cup segmentation performance strongly validates the effectiveness of the prediction library and probability calibration strategy introduced to address the foreground–background imbalance problem.
In summary, these results confirm that the denoising method optimized using mutual information is effective. This method not only enhances the model’s performance but also strengthens its stability. It helps address potential noise issues in pre-trained model results by naturally removing noise through an entropy-based approach, thereby increasing the information content.

4.3. Ablation Study Results

This section will validate the effectiveness of each module and setting in the algorithm through a series of ablation experiments. Specifically, Table 3 shows the results of the ablation experiments on the mutual information module and the prediction library, which were conducted on the RIM-ONE-r dataset and the Drishti-GS dataset.
By carefully comparing the first and third rows of each dataset, the effectiveness of the prediction library can be clearly observed. Notably, on the Drishti-GS dataset, the prediction library led to a significant 10% improvement in the DICE coefficient. This result confirms that the module not only effectively enhances the model’s performance but also alleviates the issue caused by the imbalance between the foreground and background. By encouraging the model to focus more on foreground information, the model’s performance is further improved.
Additionally, by analyzing the second and third rows of data for each model, the effectiveness of the mutual information module can be clearly observed. From an information theory perspective, the core principle of the mutual information module is to use the theory of entropy to accurately denoise the model’s output. During domain adaptation, various noise interferences inevitably occur, which can affect the accuracy and reliability of model predictions. The mutual information module can, to some extent, identify and remove these interfering signals, making the model’s output more accurate. This denoising strategy based on information entropy theory has been fully validated in the experimental data comparison, demonstrating its effectiveness in improving model performance.
A deeper analysis reveals the modules’ complementary roles. The MIM module enhances the supervisory signal by filtering noise from pseudo-labels, which is crucial for preventing error accumulation in the self-training process. Concurrently, the Prediction Bank module directly tackles severe class imbalance by re-weighting the loss based on global statistics, forcing the model to focus on vital foreground regions. This synergy ensures that the cleaner labels provided by MIM are effectively utilized. However, we also identified a trade-off in Table 3 for the RIM-ONE-r dataset: while our full model achieves the best Dice score, the MIM module alone yields a slightly better ASD for the optic cup. This suggests the combined approach offers the best overall balance between regional overlap and boundary precision.
It is particularly worth emphasizing that when the prediction library and the mutual information module are combined, they create a synergistic effect that enhances the model’s performance. This combination is not just a simple addition but an interaction and optimization within the model. The prediction library provides the model with rich and targeted prior information, while the mutual information module further optimizes the data quality, complementing each other. This collaboration enables the model to more accurately extract key features when processing various complex data, thereby improving the model’s performance.
In addition, this section also explores different data augmentation methods. Table 4 visually presents the experimental results under various data augmentation strategies. The first row, “Weak Augmentation,” indicates that both the teacher and student networks used weak data augmentation; the second row, “Strong Augmentation,” shows that both networks applied strong data augmentation; and the method proposed in this paper uses weak data augmentation for the teacher network and strong data augmentation for the student network.
Upon analyzing the experimental results, it is evident that when both models use strong data augmentation, the performance is less than satisfactory. This is likely because strong data augmentation alters the images significantly, causing them to lose some of their original key information, which negatively affects the model’s performance. In contrast, the approach used in this paper applies weak data augmentation to the teacher network, effectively ensuring the reliability and stability of the teacher network’s output. On this basis, the student network is further guided to tackle more challenging tasks, thereby improving the student network’s ability to extract information and ultimately enhancing both the performance and stability of the model.

4.4. Parameter Analysis

This study also evaluates the sensitivity of hyperparameter weights. First, Table 5 presents the sensitivity analysis results for the teacher network update coefficient. Analyzing the average results, it is observed that the experimental outcomes remain largely consistent with different update coefficients. However, upon closer inspection, the model performs optimally when the update coefficient is set to 0.99. This result demonstrates that the model exhibits good robustness when faced with different update coefficients. It indicates that the model can flexibly adapt to changes in hyperparameters within a certain range while maintaining relatively stable performance, providing a degree of assurance for its widespread use in practical applications.
Additionally, this section performs a sensitivity analysis of the loss coefficient α , as shown in Figure 2. The DICE coefficient here represents the average value of the cup and disc results. From the analysis, it can be observed that the DICE coefficients corresponding to different loss coefficients generally fall within the range of 90–92%. This data strongly demonstrates the algorithm’s stability with respect to the loss coefficient. Further observation reveals that when α is set to 1.0, the model shows the best performance. This phenomenon suggests that when the ratio of segmentation loss to mutual information loss is balanced, the model achieves the optimal segmentation results.

4.5. Visualization

To intuitively assess the performance of the segmentation algorithm, this section presents a visual analysis of the segmentation results. As shown in Figure 3, the visualization is presented in a four-row layout. The top two rows display the segmentation results for the Drishti-GS dataset, while the bottom two rows focus on the RIM-ONE-r dataset. Each row follows a consistent four-column structure.
The first column shows the original images, which retain their initial state in the dataset and provide a basic reference for subsequent comparative analysis. The second column displays the ground truth labels, which are annotated by professionals and serve as the standard for evaluating the accuracy of the segmentation algorithm. The third column presents the segmentation results generated by the source domain pre-trained model. The fourth column shows the segmentation results after adaptation with the method proposed in this paper.
The method proposed in this study fully considers the characteristics of medical images and provides targeted optimization for pre-trained models. From the visual results, it is clear that the segmentation accuracy has been significantly improved after processing with the proposed method. The previously blurred boundaries of the target region are now clear and well defined, closely aligning with the boundaries defined by the ground truth labels. Additionally, the segmentation edges are smoother. Furthermore, jagged lines are noticeably reduced, which not only enhances the visual quality of the segmentation results but, more importantly, provides more accurate and reliable image analysis for doctors in practical applications, such as medical diagnostic assistance systems.
In conclusion, the observation and analysis of the visualization results validate the effectiveness of the proposed method in improving segmentation accuracy and its ability to denoise images.

5. Conclusions and Discussion

In this paper, we addressed the critical challenges of domain shift and data privacy in medical image segmentation by proposing a novel source-free unsupervised domain adaptation algorithm, MIMPB. Our method achieved state-of-the-art performance on public fundus segmentation benchmarks, demonstrating that models can be effectively adapted to new clinical data without accessing sensitive source data. The core contributions establish a strong framework for privacy-preserving AI deployment across different healthcare institutions.
The clinical relevance of this work lies in its potential to be seamlessly integrated into computer-aided diagnostic systems for ophthalmology. For instance, an automated glaucoma screening tool equipped with our method could quickly adapt to new imaging devices or patient populations within a hospital, thus providing consistent and accurate optic disc and cup segmentation. This would reduce the need for costly data re-annotation and decrease the workload of specialists. It would also facilitate large-scale, reliable disease screening. Beyond ophthalmology, the principles of our source-free approach are broadly applicable to other medical imaging modalities, such as MRI and CT scans, where domain shifts are a common obstacle. It also holds promise for non-medical fields where data is isolated due to privacy or commercial sensitivities.
Despite its promising results, this study has limitations. Its efficacy has only been validated on fundus images, and its generalization to modalities with fundamentally different characteristics remains to be explored. Moreover, the performance of the adaptation process is inherently linked to the quality of the initial source-trained model. Future research will therefore focus on validating the MIMPB framework on a wider range of medical imaging tasks, including brain tumor and organ segmentation. By addressing these challenges, we believe the framework can evolve into a truly practical and scalable solution, thereby accelerating the adoption of adaptive AI in diverse medical scenarios.

Author Contributions

Conceptualization, H.W. and Y.Z.; methodology, Y.Z.; software, H.W.; validation, H.W., Y.Z. and X.L.; formal analysis, H.W.; investigation, Y.Z.; resources, X.L.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, H.W.; visualization, Y.Z.; supervision, X.L.; project administration, H.W.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Innovation Plan of Shanghai Science and Technology Commission under Grant 22511106005.

Data Availability Statement

All datasets analyzed in this study are publicly available and can be found at the locations specified in the references [31,32,33]. The code supporting the findings of this study is not publicly available at this time due to project and funding requirements. However, it may be made available from the corresponding author upon reasonable request, pending necessary approvals.

Acknowledgments

The authors would like to thank the High Performance Computing Center of Shanghai University and Shanghai Engineering Research Center of Intelligent Computing System for providing computing resources and technical support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yao, W.; Bai, J.; Liao, W.; Chen, Y.; Liu, M.; Xie, Y. From cnn to transformer: A review of medical image segmentation models. J. Imaging Inform. Med. 2024, 37, 1529–1547. [Google Scholar] [CrossRef] [PubMed]
  2. Guan, H.; Liu, M. Domain Adaptation for Medical Image Analysis: A Survey. IEEE Trans. Biomed. Eng. 2022, 69, 1173–1185. [Google Scholar] [CrossRef] [PubMed]
  3. Yang, M.; Wu, Z.; Zheng, H.; Huang, L.; Ding, W.; Pan, L.; Yin, L. Cross-modality medical image segmentation via enhanced feature alignment and cross pseudo supervision learning. Diagnostics 2024, 14, 1751. [Google Scholar] [CrossRef] [PubMed]
  4. Shen, Y.; Sheng, B.; Fang, R.; Li, H.; Dai, L.; Stolte, S.; Qin, J.; Jia, W.; Shen, D. Domain-invariant interpretable fundus image quality assessment. Med. Image Anal. 2020, 61, 101654. [Google Scholar] [CrossRef] [PubMed]
  5. Yu, Q.; Xi, N.; Yuan, J.; Zhou, Z.; Dang, K.; Ding, X. Source-free domain adaptation for medical image segmentation via prototype-anchored feature alignment and contrastive learning. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 3–12. [Google Scholar]
  6. Shen, W.; Wang, Q.; Jiang, H.; Li, S.; Yin, J. Unsupervised Domain Adaptation for Semantic Segmentation via Self-Supervision. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 12–16 July 2021; pp. 2747–2750. [Google Scholar]
  7. Xu, Z.; Lu, D.; Wang, Y.; Luo, J.; Wei, D.; Zheng, Y.; Tong, R.K.y. Denoising for relaxing: Unsupervised domain adaptive fundus image segmentation without source data. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 214–224. [Google Scholar]
  8. Ghafoorian, M.; Mehrtash, A.; Kapur, T.; Karssemeijer, N.; Marchiori, E.; Pesteie, M.; Guttmann, C.R.G.; de Leeuw, F.E.; Tempany, C.M.; van Ginneken, B.; et al. Transfer Learning for Domain Adaptation in MRI: Application in Brain Lesion Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention—MICCAI 2017, Quebec, QC, Canada, 10–14 September 2017; pp. 516–524. [Google Scholar]
  9. Swati, Z.N.K.; Zhao, Q.; Kabir, M.; Ali, F.; Ali, Z.; Ahmed, S.; Lu, J. Brain tumor classification for MR images using transfer learning and fine-tuning. Comput. Med. Imaging Graphs 2019, 75, 34–46. [Google Scholar] [CrossRef] [PubMed]
  10. Samala, R.K.; Chan, H.P.; Hadjiiski, L.; Helvie, M.A.; Richter, C.; Cha, K. Cross-domain and multi-task transfer learning of deep convolutional neural network for breast cancer diagnosis in digital breast tomosynthesis. In Proceedings of the Medical Imaging 2018: Computer-Aided Diagnosis, Houston, TX, USA, 12–15 February 2018; pp. 172–178. [Google Scholar]
  11. Gu, Y.; Ge, Z.; Bonnington, C.P.; Zhou, J. Progressive Transfer Learning and Adversarial Domain Adaptation for Cross-Domain Skin Disease Classification. IEEE J. Biomed. Health Inform. 2020, 24, 1379–1393. [Google Scholar] [CrossRef] [PubMed]
  12. Liu, X.; Xing, F.; Shusharina, N.; Lim, R.; Jay Kuo, C.C.; El Fakhri, G.; Woo, J. Act: Semi-supervised domain-adaptive medical image segmentation with asymmetric co-training. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 66–76. [Google Scholar]
  13. Ma, Q.; Zhang, J.; Qi, L.; Yu, Q.; Shi, Y.; Gao, Y. Constructing and exploring intermediate domains in mixed domain semi-supervised medical image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 11642–11651. [Google Scholar]
  14. Kamnitsas, K.; Baumgartner, C.; Ledig, C.; Newcombe, V.; Simpson, J.; Kane, A.; Menon, D.; Nori, A.; Criminisi, A.; Rueckert, D.; et al. Unsupervised domain adaptation in brain lesion segmentation with adversarial networks. In Proceedings of the International Conference on Information Processing in Medical Imaging, Boone, NC, USA, 25–30 June 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 597–609. [Google Scholar]
  15. Wang, S.; Yu, L.; Yang, X.; Fu, C.W.; Heng, P.A. Patch-Based Output Space Adversarial Learning for Joint Optic Disc and Cup Segmentation. IEEE Trans. Med. Imaging 2019, 38, 2485–2495. [Google Scholar] [CrossRef] [PubMed]
  16. Dou, Q.; Ouyang, C.; Chen, C.; Chen, H.; Glocker, B.; Zhuang, X.; Heng, P.A. Pnp-adanet: Plug-and-play adversarial domain adaptation network at unpaired cross-modality cardiac segmentation. IEEE Access 2019, 7, 99065–99076. [Google Scholar] [CrossRef]
  17. Zhang, T.; Cheng, J.; Fu, H.; Gu, Z.; Xiao, Y.; Zhou, K.; Gao, S.; Zheng, R.; Liu, J. Noise Adaptation Generative Adversarial Network for Medical Image Analysis. IEEE Trans. Med. Imaging 2020, 39, 1149–1159. [Google Scholar] [CrossRef] [PubMed]
  18. Ji, W.; Chung, A.C. Diffusion-based domain adaptation for medical image segmentation using stochastic step alignment. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024; pp. 188–198. [Google Scholar]
  19. Zhang, X.; Wu, Y.; Angelini, E.; Li, A.; Guo, J.; Rasmussen, J.M.; O’Connor, T.G.; Wadhwa, P.D.; Jackowski, A.P.; Li, H.; et al. MAPSeg: Unified Unsupervised Domain Adaptation for Heterogeneous Medical Image Segmentation Based on 3D Masked Autoencoding and Pseudo-Labeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024. [Google Scholar]
  20. Kim, Y.; Cho, D.; Han, K.; Panda, P.; Hong, S. Domain Adaptation Without Source Data. IEEE Trans. Artif. Intell. 2021, 2, 508–518. [Google Scholar] [CrossRef]
  21. Li, X.; Chen, W.; Xie, D.; Yang, S.; Yuan, P.; Pu, S.; Zhuang, Y. A free lunch for unsupervised domain adaptive object detection without source data. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, pp. 8474–8481. [Google Scholar]
  22. Vu, T.H.; Jain, H.; Bucher, M.; Cord, M.; Pérez, P. ADVENT: Adversarial Entropy Minimization for Domain Adaptation in Semantic Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2512–2521. [Google Scholar]
  23. Hou, Y.; Zheng, L. Source free domain adaptation with image translation. arXiv 2020, arXiv:2008.07514. [Google Scholar]
  24. Li, R.; Jiao, Q.; Cao, W.; Wong, H.S.; Wu, S. Model Adaptation: Unsupervised Domain Adaptation Without Source Data. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9638–9647. [Google Scholar]
  25. Iqbal, J.; Ali, M. Mlsl: Multi-level self-supervised learning for domain adaptation with spatially independent and semantically consistent labeling. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1864–1873. [Google Scholar]
  26. Li, G.; Kang, G.; Liu, W.; Wei, Y.; Yang, Y. Content-consistent matching for domain adaptive semantic segmentation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 440–456. [Google Scholar]
  27. Bateson, M.; Kervadec, H.; Dolz, J.; Lombaert, H.; Ayed, I.B. Source-free domain adaptation for image segmentation. Med. Image Anal. 2022, 82, 102617. [Google Scholar] [PubMed]
  28. VS, V.; Valanarasu, J.M.J.; Patel, V.M. Target and task specific source-free domain adaptive image segmentation. arXiv 2022, arXiv:2203.15792. [Google Scholar]
  29. Yang, C.; Guo, X.; Chen, Z.; Yuan, Y. Source free domain adaptation for medical image segmentation with fourier style mining. Med. Image Anal. 2022, 79, 102457. [Google Scholar] [CrossRef] [PubMed]
  30. Lee, D.H.; Choi, S.; Kim, H.; Chung, S.Y. Unsupervised Visual Representation Learning via Mutual Information Regularized Assignment. arXiv 2022, arXiv:2211.02284. [Google Scholar] [CrossRef]
  31. Orlando, J.I.; Fu, H.; Breda, J.B.; Van Keer, K.; Bathula, D.R.; Diaz-Pinto, A.; Fang, R.; Heng, P.A.; Kim, J.; Lee, J.; et al. Refuge challenge: A unified framework for evaluating automated methods for glaucoma assessment from fundus photographs. Med. Image Anal. 2020, 59, 101570. [Google Scholar] [CrossRef] [PubMed]
  32. Fumero, F.; Alayon, S.; Sanchez, J.L.; Sigut, J.; Gonzalez-Hernandez, M. RIM-ONE: An open retinal image database for optic nerve evaluation. In Proceedings of the 2011 24th International Symposium on Computer-Based Medical Systems (CBMS), Bristol, UK, 27–30 June 2011; pp. 1–6. [Google Scholar]
  33. Sivaswamy, J.; Krishnadas, S.; Chakravarty, A.; Joshi, G.; Ujjwal. A comprehensive retinal image dataset for the assessment of glaucoma from the optic nerve head analysis. JSM Biomed. Imaging Data Pap. 2015, 2, 1004. [Google Scholar]
  34. Wang, S.; Yu, L.; Li, K.; Yang, X.; Fu, C.W.; Heng, P.A. Boundary and Entropy-driven Adversarial Learning for Fundus Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  35. Bateson, M.; Kervadec, H.; Dolz, J.; Lombaert, H.; Ayed, I.B. Source-Relaxed Domain Adaptation for Image Segmentation. arXiv 2020, arXiv:2005.03697v1. [Google Scholar]
  36. Karani, N.; Erdil, E.; Chaitanya, K.; Konukoglu, E. Test-time adaptable neural networks for robust medical image segmentation. Med. Image Anal. 2021, 68, 101907. [Google Scholar] [CrossRef] [PubMed]
  37. Chen, C.; Liu, Q.; Jin, Y.; Dou, Q.; Heng, P.A. Source-Free Domain Adaptive Fundus Image Segmentation with Denoised Pseudo-Labeling. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021. [Google Scholar]
  38. Tang, L.; Li, K.; He, C.; Zhang, Y.; Li, X. Source-Free Domain Adaptive Fundus Image Segmentation with Class-Balanced Mean Teacher. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; Springer: Cham, Switzerland, 2023; pp. 684–694. [Google Scholar]
  39. Luo, X.; Chen, W.; Liang, Z.; Yang, L.; Wang, S.; Li, C. Crots: Cross-Domain Teacher–Student Learning for Source-Free Domain Adaptive Semantic Segmentation. Int. J. Comput. Vis. 2024, 132, 20–39. [Google Scholar] [CrossRef]
  40. Huai, Z.; Ding, X.; Li, Y.; Li, X. Context-aware pseudo-label refinement for source-free domain adaptive fundus image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; Springer: Cham, Switzerland, 2023; pp. 618–628. [Google Scholar]
  41. Li, L.; Zhou, Y.; Yang, G. Robust source-free domain adaptation for fundus image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 1–10 January 2024; pp. 7840–7849. [Google Scholar]
Figure 1. The framework of MIMPB. First, the model undergoes pretraining using labeled source domain data. During this phase, the labeled source domain data is fully utilized to enable the model to learn essential features and patterns. After pretraining, the pretrained model is used to initialize the target domain model. During the training phase in the target domain, the model is divided into a teacher model and a student model. Both models are initialized using the pretrained source model. However, their parameter update mechanisms differ during training. The teacher model does not update its parameters directly; instead, its parameters are updated using the exponential moving average of the student model’s training results. Furthermore, the algorithm optimizes the teacher model’s results through mutual information optimization, which guides the student network’s training in a more precise direction. Additionally, a prediction bank is utilized to address the imbalance between foreground and background data.
Figure 1. The framework of MIMPB. First, the model undergoes pretraining using labeled source domain data. During this phase, the labeled source domain data is fully utilized to enable the model to learn essential features and patterns. After pretraining, the pretrained model is used to initialize the target domain model. During the training phase in the target domain, the model is divided into a teacher model and a student model. Both models are initialized using the pretrained source model. However, their parameter update mechanisms differ during training. The teacher model does not update its parameters directly; instead, its parameters are updated using the exponential moving average of the student model’s training results. Furthermore, the algorithm optimizes the teacher model’s results through mutual information optimization, which guides the student network’s training in a more precise direction. Additionally, a prediction bank is utilized to address the imbalance between foreground and background data.
Electronics 14 03656 g001
Figure 2. Sensitivity analysis of loss coefficient α . The blue stars indicate the Dice coefficient values corresponding to each α .
Figure 2. Sensitivity analysis of loss coefficient α . The blue stars indicate the Dice coefficient values corresponding to each α .
Electronics 14 03656 g002
Figure 3. Visual comparison of optic disc and cup segmentation before and after domain adaptation.
Figure 3. Visual comparison of optic disc and cup segmentation before and after domain adaptation.
Electronics 14 03656 g003
Table 1. Performance comparison on RIM-ONE-r dataset.
Table 1. Performance comparison on RIM-ONE-r dataset.
MethodS-FOptic DiscOptic CupAverage
Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓
RIM-ONE-r Dataset
W/o DA 83.18 ± 6.4624.15 ± 15.5874.51 ± 16.4014.44 ± 11.2778.85 ± 1.4319.47 ± 13.43
Oracle 96.80-85.60-91.20-
BEAL×89.80-81.00-85.40-
AdvEnt×89.73 ± 3.669.84 ± 3.8677.99 ± 21.087.57 ± 4.2483.86 ± 12.378.71 ± 4.05
SRDA89.37 ± 2.709.91 ± 2.4577.61 ± 13.5810.15 ± 5.7583.49 ± 6.9310.03 ± 4.10
DAE89.08 ± 3.3211.63 ± 6.8479.01 ± 12.8210.31 ± 8.4584.05 ± 8.0710.97 ± 7.65
DPL90.13 ± 3.069.43 ± 3.4679.78 ± 11.059.01 ± 5.5984.85 ± 7.069.22 ± 4.53
CBMT93.36 ± 4.076.20 ± 4.7981.16 ± 14.718.37 ± 6.9987.26 ± 9.397.29 ± 5.89
Crots92.93 ± 3.616.36 ± 3.3880.19 ± 2.136.84 ± 4.5786.56 ± 2.876.60 ± 3.98
CPR91.72 ± 7.376.80 ± 5.1978.56 ± 2.037.65 ± 5.3385.14 ± 4.707.22 ± 5.26
RDPL91.70 ± 3.887.62 ± 3.8478.55 ± 2.207.48 ± 4.4285.13 ± 3.047.55 ± 4.13
OURS93.73 ± 2.875.53 ± 2.4181.86 ± 10.107.91 ± 3.7887.80 ± 6.496.72 ± 3.10
Table 2. Performance comparison on Drishti-GS dataset.
Table 2. Performance comparison on Drishti-GS dataset.
MethodS-FOptic DiscOptic CupAverage
Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓
Drishti-GS Dataset
W/o DA 93.84 ± 2.919.05 ± 7.5083.36 ± 11.9511.39 ± 6.3086.60 ± 7.4310.22 ± 6.90
Oracle 97.40-90.10-93.75-
BEAL [34]×96.10-86.20-91.15-
AdvEnt [22]×96.16 ± 1.654.36 ± 1.8382.75 ± 11.0811.36 ± 7.2289.46 ± 6.377.86 ± 4.53
SRDA [35]96.22 ± 1.304.88 ± 3.4780.67 ± 11.7813.12 ± 6.4888.45 ± 6.549.00 ± 4.98
DAE [36]94.04 ± 2.858.79 ± 7.4583.11 ± 11.8911.56 ± 6.3288.58 ± 7.3710.18 ± 6.89
DPL [37]96.39 ± 1.334.08 ± 1.4983.53 ± 17.8011.39 ± 10.1889.96 ± 9.577.74 ± 5.84
CBMT [38]96.61 ± 1.453.85 ± 1.6384.33 ± 11.7010.30 ± 5.8890.47 ± 6.587.08 ± 3.76
Crots [39]96.58 ± 1.783.90 ± 2.1883.18 ± 11.3611.22 ± 6.7989.88 ± 6.577.56 ± 4.49
CPR [40]90.12 ± 2.4210.63 ± 2.1075.04 ± 12.9816.69 ± 9.0982.58 ± 7.7013.66 ± 5.59
RDPL [41]96.64 ± 1.493.78 ± 1.6184.33 ± 12.1710.30 ± 5.9890.49 ± 6.837.04 ± 3.80
OURS96.51 ± 1.183.90 ± 1.2986.97 ± 11.888.62 ± 5.3291.74 ± 6.536.26 ± 3.31
Table 3. Ablation study results on RIM-ONE-r and Drishti-GS.
Table 3. Ablation study results on RIM-ONE-r and Drishti-GS.
MIM ModulePB ModuleOptic DiscOptic CupAverage
Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓
RIM-ONE-r Dataset
93.04 ± 3.136.02 ± 2.4180.45 ± 14.647.70 ± 3.9686.75 ± 8.896.87 ± 3.19
92.98 ± 3.196.20 ± 2.6975.20 ± 21.638.92 ± 5.8884.09 ± 12.417.56 ± 4.29
93.73 ± 2.875.53 ± 2.4181.86 ± 10.107.91 ± 3.7887.80 ± 6.496.72 ± 3.10
Drishti-GS Dataset
86.89 ± 3.7313.96 ± 2.3477.73 ± 10.7512.62 ± 6.1182.31 ± 7.2414.29 ± 4.23
89.16 ± 2.4311.63 ± 2.1767.62 ± 9.5121.14 ± 7.7178.39 ± 5.9716.39 ± 4.94
96.51 ± 1.183.90 ± 1.2986.97 ± 11.888.62 ± 5.3291.74 ± 6.536.26 ± 3.31
Table 4. Segmentation performance with different data augmentation strategies.
Table 4. Segmentation performance with different data augmentation strategies.
AugmentationOptic DiscOptic CupAverage
Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓
Drishti-GS Dataset
Weak Augmentation93.36 ± 1.467.65 ± 1.7183.01 ± 11.8211.20 ± 5.3288.19 ± 6.649.42 ± 7.03
Strong Augmentation92.95 ± 1.548.15 ± 1.7984.88 ± 12.029.93 ± 5.2288.92 ± 6.789.04 ± 3.51
Ours96.51 ± 1.183.90 ± 1.2986.97 ± 11.888.62 ± 5.3291.74 ± 6.536.26 ± 3.31
Table 5. Teacher–student update coefficient sensitivity analysis.
Table 5. Teacher–student update coefficient sensitivity analysis.
Update CoefficientOptic DiscOptic CupAverage
Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓Dice [%] ↑ASD [Pixel] ↓
Drishti-GS Dataset
0.995.34 ± 1.555.23 ± 1.6285.29 ± 11.959.79 ± 5.1990.31 ± 6.757.51 ± 3.41
0.9593.36 ± 1.707.75 ± 1.7185.53 ± 11.709.50 ± 4.8189.44 ± 6.708.50 ± 4.21
0.9695.07 ± 1.435.58 ± 1.5485.77 ± 13.519.42 ± 6.1490.42 ± 7.477.48 ± 4.51
0.9796.19 ± 1.514.28 ± 1.6685.54 ± 13.779.59 ± 6.3490.86 ± 7.646.94 ± 4.65
0.9894.55 ± 1.646.26 ± 1.9885.77 ± 13.959.39 ± 6.4990.16 ± 7.807.82 ± 5.89
0.9996.51 ± 1.183.90 ± 1.2986.97 ± 11.888.62 ± 5.3291.74 ± 6.536.26 ± 3.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, H.; Zhou, Y.; Li, X. Source-Free Domain Adaptation for Medical Image Segmentation via Mutual Information Maximization and Prediction Bank. Electronics 2025, 14, 3656. https://doi.org/10.3390/electronics14183656

AMA Style

Wu H, Zhou Y, Li X. Source-Free Domain Adaptation for Medical Image Segmentation via Mutual Information Maximization and Prediction Bank. Electronics. 2025; 14(18):3656. https://doi.org/10.3390/electronics14183656

Chicago/Turabian Style

Wu, Hongzhen, Yue Zhou, and Xiaoqiang Li. 2025. "Source-Free Domain Adaptation for Medical Image Segmentation via Mutual Information Maximization and Prediction Bank" Electronics 14, no. 18: 3656. https://doi.org/10.3390/electronics14183656

APA Style

Wu, H., Zhou, Y., & Li, X. (2025). Source-Free Domain Adaptation for Medical Image Segmentation via Mutual Information Maximization and Prediction Bank. Electronics, 14(18), 3656. https://doi.org/10.3390/electronics14183656

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop