Next Article in Journal
Efficient Multi-Target Localization Using Dynamic UAV Clusters
Previous Article in Journal
A Quasi-Distributed Crack Sensor Based on Weakly Coupled Vertical U-Shaped Ring Array
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bridging Domain Gaps in Computational Pathology: A Comparative Study of Adaptation Strategies

by
João D. Nunes
1,2,*,
Diana Montezuma
3,4,
Domingos Oliveira
3,
Tania Pereira
1,5,
Inti Zlobec
6,
Isabel Macedo Pinto
3 and
Jaime S. Cardoso
1,2
1
Institute for Systems and Computer Engineering, Technology and Science, 4200-465 Porto, Portugal
2
Faculty of Engineering, University of Porto, 4200-465 Porto, Portugal
3
IMP Diagnostics, 4150-146 Porto, Portugal
4
Cancer Biology and Epigenetics Group, Research Center of Portuguese Oncology Institute of Porto/RISE@Research Center of Portuguese Oncology Institute of Porto (Health Research Network), Portuguese Oncology Institute of Porto/Porto Comprehensive Cancer Centre Raquel Seruca, R. Dr. António Bernardino de Almeida, 4200-072 Porto, Portugal
5
FCTUC—Faculty of Sciences and Technology, University of Coimbra, 3004-516 Coimbra, Portugal
6
Institute of Tissue Medicine and Pathology, University of Bern, 3008 Bern, Switzerland
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(9), 2856; https://doi.org/10.3390/s25092856
Submission received: 25 February 2025 / Revised: 19 April 2025 / Accepted: 29 April 2025 / Published: 30 April 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Due to the high variability in Hematoxylin and Eosin (H&E)-stained Whole Slide Images (WSIs), hidden stratification, and batch effects, generalizing beyond the training distribution is one of the main challenges in Deep Learning (DL) for Computational Pathology (CPath). But although DL depends on large volumes of diverse and annotated data, it is common to have a significant number of annotated samples from one or multiple source distributions, and another partially annotated or unlabeled dataset representing a target distribution for which we want to generalize, the so-called Domain Adaptation (DA). In this work, we focus on the task of generalizing from a single source distribution to a target domain. As it is still not clear which domain adaptation strategy is best suited for CPath, we evaluate three different DA strategies, namely FixMatch, CycleGAN, and a self-supervised feature extractor, and show that DA is still a challenge in CPath.

1. Introduction

Computational Pathology (CPath) is one of the critical fields where Machine Learning (ML) is revolutionizing clinical practice by increasing productivity, reducing costs, improving the precision of diagnosis, and enabling more effective treatments. But a challenge with microscopy imaging, from which Hematoxylin and Eosin (H&E)-stained Whole Slide Images (WSIs) is one of the most widely available modalities, is the high variability in image chromaticity and morphological appearance, aside from batch effects and hidden stratification that prevent generalization beyond the training data distribution. Deep Learning (DL) is the go-to solution for WSI analysis, with the top-down approach one of the most common methods [1]. The strategy consists of splitting the gigapixel WSIs into multiple lower-dimensional tiles, subsequently encoded into tile-level features (or scores), followed by an aggregation function for slide-level predictions. WSI analysis is thus naturally a Multiple Instance Learning (MIL) paradigm, and its success strongly depends on the quality of tile-level representations. Indeed, although ImageNet pre-training and self-supervised learning on unlabeled histopathology datasets were shown to be successful for tile-level feature extraction [2,3,4,5,6], task-specific fine-tuning could translate to better results [1,7,8,9].
DL relies on large volumes of diverse and annotated datasets that accurately reflect the statistical distribution of the entire population. However, these datasets are often difficult to collect and annotate in CPath, leading to biased solutions that often fail even for small distribution shifts. Therefore, algorithms with outstanding results on independent and identically distributed (IID) samples frequently depend on dataset-specific attributes, i.e., spurious correlations that do not hold in related but distinct distributions (i.e., domains). Nevertheless, a common setting is to have access to a significant number of annotated samples from one or multiple source distributions and another (frequently smaller) dataset representing a target distribution with little to no annotations for which we want to generalize, the so-called Domain Adaptation (DA) paradigm. Indeed, factors such as stain variability, blurriness, and resolution inconsistencies severely affect the generalization of DL models for CPath. Nonetheless, several strategies have been proposed to address these issues, like Generative Adversarial Networks (GANs) to deal with stain variability or adversarial learning, which learn invariant features. In this work, we focus on the challenging scenario of single-source DA, which aims to transfer knowledge from a single-source distribution to a target domain. However, as it is still not clear which DA strategy is best suited for CPath, we evaluate and compare three different approaches, namely a weakly supervised learning (WSL) strategy based on consistency regularization (FixMatch [10]), a domain transfer approach with CycleGAN [11], and a method based on self-supervised feature extraction with a foundation model (Phikon [12]). We explore the complementarity of these methods and compare these strategies in semi-supervised DA (SSDA) and unsupervised DA (UDA) to clarify how performance degrades when no annotations from the target distribution are available. We thus expect to shed light on the effectiveness and limitations of each method in addressing DA in CPath. Moreover, to the best of our knowledge, we are the first to address the curse of domain-invariant representations in CPath. Besides this introduction, this work is structured as follows: In Section 2, we start by a review of works similar to ours; then, in Section 3, we discuss the methods we consider in this work, while in Section 4 we describe the experimental details of this work. After, in Section 5, we present the results and provide a discussion in Section 6, and, finally, in Section 7, we unveil the main conclusions of this work.

2. Related Work

The task of DA usually assumes that the conditional distribution of the labels given the data remains stable across source and target domains ( P ( Y | X t a r g e t ) = P ( Y | X s o u r c e ) ) but that the marginal distributions can change ( P ( X t a r g e t ) P ( X s o u r c e ) ). This is the so-called covariate shift assumption and it motivates DA techniques from different angles. For example, reweighting samples from the source domain to match the target distribution [13], learning representations that are transportable across domains [14], using optimal transport to map the source distribution to the target domain [15], or the widely adopted strategy of learning domain-invariant representations [16]. However, a limitation is that the covariate shift assumption could be violated, meaning domain invariant representations hinder the generalization of the model to the target domain [17]. This is referred to as the curse of domain-invariant representations, with recent approaches empirically demonstrating the effectiveness of consistency regularization [18,19] to tackle this issue.
Specifically, in CPath, Ren et al. [20] suggest domain-adversarial networks for a prostate classification task. Koohbanani et al. [21] propose a multi-task learning framework for oral squamous cell carcinoma and breast cancer metastasis classification that combines SL with self-supervised pretext tasks to learn invariant representations without requiring annotations from the target domain. In turn, Abbet et al. [22] design a self-supervised method with curriculum learning to learn invariant features for colorectal cancer tissue detection, whereas Marini et al. [23] suggest a robust stain color augmentation technique, while also demonstrating its generalization effectiveness when combined with domain adversarial neural networks. Other approaches focus on Generative Adversarial Networks (GANs) for stain transfer between domains [24,25]. Yet, the usefulness of the various DA techniques available in computer vision still requires clarification in CPath. Moreover, to the best of our knowledge, no previous method in CPath addresses the curse of domain-invariant representations.

3. Materials and Methods

In this section, we present the methods that we address in this work. We start by detailing the theoretical foundation of the method based on FixMatch, followed by the formal description of CycleGAN, and then the self-supervised feature extraction strategy. We opt for these methods, as they can work without multiple source domains. We also describe the aggregation function we consider to evaluate the learned features. Figure 1 illustrates the methods we consider in this work.

3.1. Domain Adaptation with FixMatch

To learn effective tile-level features, we resort to two different approaches: (a) SL using only source domain data annotated at the tile level, and (b) a consistency regularization approach, namely FixMatch [10], using both source and target domains. Consistency regularization is based on the smoothness assumption of semi-supervised learning (SmSL), stating that close data points in the input space are likely to share the same label. Therefore, considering a label-preserving transformation, t, over some transformation space, T , the estimated labels for a given sample, x, and its transformed view, t ( x ) , should be the same. In the context of SSDA, strong data augmentations could be interpreted as creating novel domains [18]. Therefore, consistency regularization promotes features that generalize well across domains [18]. Moreover, some augmented samples may fall in higher-density regions over the distribution of the source domain(s), meaning empirical risk minimization on the source distribution with consistency regularization on samples from the target domain should also minimize the error on the target distribution [18]. Fixmatch [10] enforces consistency by minimizing the cross-entropy between the most confident predictions for weakly augmented views of an unlabeled input and a strongly augmented version of the same sample. In essence, the strategy suggests a two-term compound loss, one for annotated data and one for unlabeled inputs.
Given two sets of labeled, L = { ( x l 1 , y l 1 ) , , ( x l N , y l N ) } (source domain), and unlabeled, U = { ( x u 1 , y u 1 ) , , ( x u M , y u M ) } (out of distribution target domain), images, sampled, respectively, in batches of size B and μ B , with μ the proportion of unlabeled samples, FixMatch starts by computing a pseudo-label for unlabeled images. This is achieved by computing hard pseudo-labels, q ^ u m = arg max ( q u m ) from the conditional distribution of the targets given a weakly augmented view of the unlabeled image, t ( x u m ) , (or unlabeled pixels in the case of partially annotated images), q u m = p ( y u m | t ( x u m ) ) . Then, consistency is imposed by minimizing the cross-entropy between the conditional distribution of the class given a strongly augmented view of the input, q u m = p ( y u m | t ( x u m ) ) , and confident pseudo-labels, max ( q u m ) > τ :
L u = 1 μ B m = 1 μ B = 1 ( max ( q u m ) > τ ) H ( q ^ u m | q u m ) ,
with q ^ u m = arg   max ( p ( y u m | t ( x u m ) ) ) , τ a confidence threshold, H the cross-entropy loss, and 1 the indicator function.
For labeled images, the standard cross-entropy loss is adopted:
L s = 1 B S b = 1 B H ( p b | p ( y | t ( x l b ) ) )
with p b the ground truth (one-hot encoded) label.
The FixMatch objective thus becomes:
L = L s + λ L u
where λ is a hyperparameter weighting the relative contribution of the unsupervised loss.
An advantage of consistency regularizatoin with FixMatch is that it is effective in dealing with the curse of domain-invariant representations [18,19] where the assumption that the conditional distribution of the labels given the data remains stable across source and target domains ( P ( Y | X t a r g e t ) = P ( Y | X s o u r c e ) ) is violated.

3.2. Domain Transformation Using CycleGAN

CycleGANs enable domain translation with unpaired images by use of cycle-consistency. The task of transferring images from source, X, to target domain, Z, involves a generator, G, and a discriminator, D 1 , to learn a mapping, G : X Z , where G tries to generate images G ( x ) that resemble images from domain Z [11]. In turn, the role of the discriminator is to distinguish between real, z, and generated samples, G ( x ) . The networks are optimized adversarially with the following objective (G tries to minimize the objective while D 1 tries to maximize it):
L G A N ( G , D 1 , x , z ) = log D 1 + 1 log D 1 ( G ( x ) )
In theory, adversarial training can produce outputs with the same distribution as the target domain. However, provided the discriminator objective is fulfilled, the generator can transform the input image into any image from the target domain [24]. When dealing with unpaired image-to-image translation, cycle-consistency ensures that, when an image is translated from one domain to another, it can later be translated back to its original domain, maintaining its essential features and structure [11]. This is achieved through a second generator, F, and discriminator pair and enforcing F ( G ( x ) ) x . The cycle-consistency loss is defined as
L c y c l e = E x [ | | G ( F ( x ) ) x | | 1 ] + E z [ | | F ( G ( z ) ) z | | 1 ]

3.3. Self-Supervised Feature Extraction

Another promising and growing direction to extract tile-level features is domain-specific foundation models pre-trained with self-supervised learning on large-scale histopathology datasets [6,12]. These models were shown to be significantly more successful than ImageNet pre-trained models [5] and can be directly used for feature extraction or fine-tuned. In this work, we explore Phikon [12], a Vision Transformer (ViT) feature extractor, pre-trained on 40 million H&E-stained tiles from TCGA [26,27] using Masked Image Modeling (MIM) and the iBOT framework [28]. We then compare the generalization of this strategy with FixMatch [10], CycleGAN, and the supervised baseline.

3.4. Evaluation Protocol: Grading of Colorectal Biopsies and Polypectomies

We focus on the task of grading colorectal biopsies and polypectomies (excluding surgical specimens) into three classes: non-neoplastic (NNeo), low-grade (LG) lesions, and high-grade (HG) lesions (high-grade dysplasia and adenocarcinoma) [9]. To compare the effectiveness of the feature learning approaches, we consider training a slide-level classifier using a state-of-the-art aggregation function, namely TransMIL [2]. We evaluate the representations in SSDA and UDA. In the semi-supervised setting, we assume access to a small amount of slide-level annotated data from the target domain used to optimize a slide-level classifier, whereas, in the unsupervised setting, we have only labels from the source domain for training (at both tile and slide-level). Moreover, we consider an additional paradigm wherein we train TransMIL using a combination of samples from both domains.

4. Experimental Details

Code to reproduce the experiments is available at the following URL: https://github.com/joao-nunes/domain-adaptation-cpath (accessed on 28 April 2025).

4.1. Datasets

We resort to two H&E-stained WSI datasets, namely CRS10K (IMP Diagnostics, Porto, Portugal) [9,29,30] (available online: https://doi.org/10.25747/fb1q-j507 (accessed on 28 April 2025) and BernCRC (IGMP, Bern, Switzerland) containing samples of colorectal biopsies and polypectomies (excluding surgical specimens) with annotations for the dysplasia grade into Non-neoplastic (NNeo); Low-grade (LG); and High-grade (HG) lesions. The slides from CRS10K were digitized with Leica GT450 WSI scanners, at 40× magnification. All samples from CRS10K are annotated at the slide level, while roughly 9% of the WSIs (967) also contain instance-level annotations. From the annotated tiles of CRS10K, we reserve 200 WSIs for testing and 200 WSIs for validation, while the remaining data are used to train a tile-level classifier for subsequent feature extraction. In turn, BernCRC was digitized with 3DHistech P250 WSI scanners and is only labeled at the slide level. Figure 2 presents sample H&E-stained tiles from each dataset. The staincolor distribution shift between source and target data is evident.

4.2. Data Pre-Processing

As in the work by Neto et al. [9], to isolate the colorful tissue from the meaningless white background, the pre-processing of the slides consists of an automatic tissue segmentation with Otsu’s thresholding on the saturation (S) channel of HSV color space, leading to the separation between tissue regions and the background. Subsequently, a 32 × 32 down-sampled version of the generated mask is used to extract WSI tiles with dimension 512 × 512 pixels at 40 × magnification and containing only tissue (i.e., a 100 % tissue threshold is used).

4.3. Data Augmentation

For both the supervised baseline and FixMatch strategy (strong augmentations), we adopt a label-preserving data augmentation policy that was previously shown to be successful in CPath [8,31]. Table 1 presents the details of the transformations we use and the corresponding hyperparameters. For data augmentation, we use the python libraries Albumentations version 1.3.0, and scikit-image version 0.19.3. Figure 3 shows some sample H&E-stained tiles from CRS10K before and after data augmentation.

4.4. Training

4.4.1. Tile-Level Feature Learning

We make use of ResNet-34 as the feature extractor, pretrained on ImageNet and fine-tuned using both SL on the source domain and FixMatch.
As first suggested in the work by Oliveira et al. [29], we treat the grading of colorectal biopsies and polypectomies as an ordinal regression problem and adopt a soft version of QWK [32] as the loss function, which weighs misclassifications differently depending on the distance between the predicted and ground truth label.
As in the work by Neto et al. [9], we adopt as an optimizer stochastic gradient descent (SGD) with momentum = 0.9 , a learning rate of 1 × 10 4 , and a weight decay of 3 × 10 4 . To train the FixMatch method, we consider λ = 0.3 and τ = 0.9 . We use a batch size, B, of 32 for the annotated samples and μ = 0.5 , i.e., μ B = 16 , for the unlabeled inputs. In addition, we include a warm-up period of 5000 iterations before introducing the FixMatch regularization term. To generate weakly augmented views, we only normalize the input with mean = 0.5 , and std = 0.1 . We adopt stratified k-fold cross-validation (k = 5) for BernCRC data and divide the data into train, validation, and test splits using, respectively, 67.5%, 12.5%, and 20% of the data. The tile-level feature learning experiments are implemented in python 3.10.8 with Pytorch version 2.0.0 and a single Tesla V100 32GB GPU. Due to computational limits and long run-times, hyperparameter optimization was limited. Our hyperparameters thus derive from similar works and from practicalities, and are likely to be sub-optimal.

4.4.2. CycleGAN

We use the architectures suggested in the work by Zhu et al. [11]. The generator consists of a stride-1 convolution with kernel size equal to 7, followed by two convolutions with stride-2 and kernel size 3, 15 residual blocks, followed by two convolutions with up-sampling, and a final convolution with stride 1 and kernel size 7. All convolutions are followed by instance normalization and ReLU activation. For the discriminator network, we are inspired by PatchGAN [24]. As in de Bel et al. [24], the discriminator consists of four convolutional layers with 64, 128, 256, and 512 filters, respectively, and with kernel size of 4 and stride 2. A final convolution ensures the output is a single filter map. All convolutions except for the last were followed by an instance normalization layer and a leaky ReLU with the leak parameter equal to 0.2. As data augmentation, we consider only normalization with m e a n = 0.5 and s t d = 0.5.

4.4.3. TransMIL

We use the official code base when implementing TransMIL [2]. We train the model for 50 epochs using LookAhead [33] optimizer with RAdam [34], a batch size B equal to 1, learning rate 1 × 10 4 , and weight decay 1 × 10 5 . For testing, we select the model with the best validation accuracy. For the slide-level aggregation experiments, we select a subset of 1000 WSIs (CRS1K) from the CRS10K training data (source domain) annotated at the slide level. We perform validation and testing using the publicly available validation and test sets from CRS10K. In turn, regarding the target domain (BernCRC data), we adopt a stratified k-fold cross-validation (k = 5) and divide the data into train, validation, and test splits using, respectively, 67.5%, 12.5%, and 20% of the data. To optimize TransMIL using samples from both domains, we train the model using two batches of data at each iteration (one from each domain). For the loss term corresponding to each mini-batch, we perform hyperparameter tuning with a random search using the CRS1K training set and one of the five folds from BernCRC and select the values with the highest validation accuracy. We observed θ 1 = 0.3 (CRS1K) and θ 2 = 0.7 (BernCRC) translated to the best validation results. The experiments for slide-level aggregation are implemented in python 3.10.8 with Pytorch version 2.0.0 and a single Nvidia GeForce 1080ti 11 GB GPU.

4.5. Evaluation Metrics

4.5.1. Quadratic Weighted Kappa

The Quadratic Weighted Kappa (QWK) measures the agreement between two raters. This metric typically varies from 0 (random agreement between raters) to 1 (complete agreement between raters). It is an appropriate metric for ordinal data as it weighs misclassifications differently depending on the distance between the predicted and actual classes, according to the following equation:
Q W K = 1 i , j = 1 n w i , j x i , j i , j = 1 n w i , j m i , j
where w i , j belongs to the weight matrix, x i , j belongs to the observed matrix, and m i , j are elements in the expected matrices. The n × n matrix of weights w is computed based on the difference between the actual and predicted class, as follows:
w i , j = ( i j ) 2 ( n 1 ) 2

4.5.2. Accuracy

Accuracy is a metric that measures how often an ML model is correct. It is defined as the proportion of correct predictions over the total number of predictions:
A c c = T P + T N T P + F P + T N + F N
With T P the number of true positive predictions, T N the true negative predictions, F P the number of false positives, and F N the number of false negative predictions.

4.5.3. Sensitivity

Sensitivity or True Positive Rate (TPR) defines the number of true positives that are correctly identified by the model:
S e n s = T P T P + F N

4.5.4. Specificity

Specificity or True Negative Rate (TNR) measures the proportion of true negatives that are correctly identified by the model:
S p e c = T N T N + F P

4.5.5. Precision

Precision quantifies the proportion of true positive predictions among all positive predictions made by the model:
P r e c = T P T P + F P

4.5.6. F1 Score

The F1 score is the harmonic mean between precision and recall:
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

4.5.7. Area Under the Receiver Operating Characteristic Curve (AUC)

The Area Under the Receiver Operating Characteristic Curve (AUC) summarizes the trade-off between the TPR and false positive rate (FPR) across all possible decision thresholds. An ML model with perfect discriminative ability will have an AUC of 1, while a model without discriminative power will have an AUC of 0.5. AUC presents two important advantages: it is scale-invariant as well as classification threshold-invariant.

5. Results

5.1. Source Domain Results

Table 2 and Table 3 report the results of model evaluation on IID samples. We observe high performance regardless of the feature extractor learning strategy and that incorporating unlabeled samples from an additional data distribution leads to slight improvements over all metrics, thus hinting that the adopted consistency regularization technique enables learning more generalizable representations.

5.2. Unsupervised Domain Adaptation

Table 4 presents the Fréchet Inception Distance (FID) between BernCRC (target) transformed to CRS10K (source) and BernCRC (target) to CRS10K (source). As we observe, the FID metric is smaller between the source and transformed target than between the source and target, thus suggesting the visual quality and realism of the domain transformation approach.
Table 5 and Table 6 report the results in the UDA paradigm for, respectively, colorectal tissue classification into NNeo, LG, HG, and binary classification. Figure 4 summarizes the main metrics (QWK, Accuracy, and Binary Accuracy) for the UDA task. Single Source DA is still a challenging scenario as, although we observe outstanding results when the training and testing data originate from the same distribution, the performance greatly decreases in the target domain.

5.3. Semi-Supervised Domain Adaptation

Table 7 and Table 8 report the results in the SSDA paradigm for, respectively, grading of colorectal biopsies and polypectomies (NNeo, LG, and HG) and binary classification. Figure 5 summarizes the main metrics (QWK, Accuracy, and Binary Accuracy) for the SSDA task.

6. Discussion

6.1. Unsupervised Domain Adaptation

The results suggest that FixMatch is a better strategy to learn more generalizable tile-level representations, as in the UDA paradigm, the MIL aggregator (TransMIL) achieves higher performance (e.g., an improvement of +14.09% in the QWK metric, +12.78% in accuracy, and +10.13% in binary accuracy for FixMatch) when compared with only SL on the source domain. We argue that this is due to the ability of FixMatch to learn representations that are transferable across the two domains. Nonetheless, the best results are observed with CycleGAN (e.g., an improvement of 16.46% in QWK, 20.63% in accuracy and 12.18% in binary accuracy when compared with supervised learning using only source domain data) showing that transforming images from one domain to another is a competitive strategy. Nevertheless, there is still potential for improvement. We suggest that the limitations of Cycle-GAN derive from synthetic image generation artifacts and from tissue-specific artifacts that are ineffectively modeled, like blurriness, tissue folds, or air bubbles. Regarding the FixMatch strategy, although it enables learning more general features shared across the two domains, it still cannot completely model the stain variability, as it focuses on data augmentation instead of modeling the domain-specific stain variability. This might suggest that these methods are complementary, as while the FixMatch strategy enables learning features shared across the two domains, Cycle-GAN mitigates stain variability. Another finding of this work is that data augmentation significantly improves performance compared to no data augmentation, with a 32.78% increase in QWK and a 21.1% increase in accuracy. However, a limitation of this study is that we only tested one combination of hyperparameters for data augmentation due to the long run times, which may lead to suboptimal results. Furthermore, it would be beneficial to compare our results with the existing literature. Unfortunately, there is limited research on domain adaptation in computational pathology, which makes direct comparisons difficult.

6.2. Semi-Supervised Domain Adaptation

The best results are achieved when slide-level-labeled samples from the target domain are available to optimize the slide-level classifier. In this setting, we achieve competitive performance despite the strategy used to learn tile-level representations. Nonetheless, the MIL aggregator with the FixMatch feature extractor achieves superior generalization (e.g., QWK increases +1.95%, and accuracy improves +0.95% (BernCRC)), while a top accuracy is also achieved when training TransMIL with source (CRS1K) and target (BernCRC) data. Compared with the UDA scenario, we observe significant improvements in performance in all metrics, which reflects that, despite the same tile feature extractor, the MIL paradigm is still subject to the simplicity bias, meaning the model is prone to depend on domain-specific spurious correlations. In turn, CycleGAN achieves a QWK of 0.6088 and an accuracy of 83.24 %.
Another finding of this work is that task-specific fine-tuning should be the preferred strategy to learn effective tile-level representations for grading of colorectal biopsies and polypectomies as the fine-tuned models compare favourably with self-supervised learning on large-scale histopathology datasets. However, the performance of the optimized WSI classifiers on top of Phikon tile embeddings is quite competitive with the fine-tuned models, hinting that further performance improvements can be observed after fine-tuning Phikon.
Overall, DA in CPath is a challenging objective, and the problem should be tackled from two fronts: On the one hand, there is the need to learn effective tile-level representations; on the other hand, the aggregation of instance features into a robust global slide representation is still prone to the simplicity bias. Therefore, DA and generalization techniques should also be adopted at this level.

7. Conclusions

In this work, we show that a WSL method based on consistency regularization, namely FixMatch, is a promising strategy for UDA in CPath, as it achieves superior generalization when compared to SL when no labeled data from the target domain are available. However, we observe the best results in UDA with CycleGAN, hinting that transforming images from source to target domain enables the reduction in stain variability that hinders DL model performance. In turn, we find that slide-level labels from the target domain are essential to learning an optimal slide-level classifier in the target domain, thus demonstrating that, despite the tile-level feature encoder, slide-level classifiers are still subject to the simplicity bias. Overall, generalizing to out-of-distribution samples is still challenging in CPath, and despite the promises of DA, there are still a few limitations, like hidden stratification and batch effects, stain, blur, and digitization artifacts, or over-representation of certain subgroups in the training set. Future work should focus on comparing other DA strategies and better understanding these issues to develop more robust models for WSI analysis. Another possible direction is the combination of the FixMatch strategy with CycleGAN, taking advantage of the complementarity of the techniques for more robust classifiers. Other possible future directions are multitask and multimodal learning. Whereas multitask learning allows for learning more general features through a shared backbone in multitask models [35,36], multimodal learning allows the model to take advantage of the complementarity of multiple data modalities like molecular profiling and H&E-stained WSI data [37]. However, this comes with the challenge of collecting paired multimodal data.

Author Contributions

Conceptualization, J.D.N., T.P. and J.S.C.; methodology, J.D.N.; software, J.D.N.; validation, D.M., D.O. and I.M.P.; formal analysis, J.D.N., D.M., T.P. and J.S.C.; investigation, J.D.N. and D.M.; resources, I.M.P. and Inti Zlobec; data curation, D.M., D.O. and I.Z.; writing—original draft preparation, J.D.N.; writing—review and editing, T.P., D.M. and J.S.C.; visualization, D.M.; supervision, T.P. and J.S.C.; project administration, T.P. and J.S.C.; funding acquisition, D.M., J.S.C. and J.D.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work is financed by National Funds through the Portuguese funding agency, FCT—Fundação para a Ciência e a Tecnologia, within the PhD grant 2022.12385.BD. In addition, IMP Diagnostics also provided funding support for this study.

Institutional Review Board Statement

All procedures were in accordance with the ethical standards of the 1964 Helsinki declaration and its later amendments and comparable ethical standards. All data was anonymized and data collection and usage were performed in accordance with the General Data Protection Regulation (GDPR) and national laws and regulations. Ethical approval was waived by the local Ethics Committee of INESC TEC in view of the retrospective nature of the study and all the procedures being performed were part of the routine care.

Informed Consent Statement

Patient consent was waived because all data were anonymized and data collection and usage were performed in accordance with the General Data Protection Regulation (GDPR) and national laws and regulations.

Data Availability Statement

The original CRS10K data presented in the study are openly available in INESC TEC research data repository at https://doi.org/10.25747/fb1q-j507 (accessed on 28 April 2025). The BernCRC dataset presented in this article is not readily available because of the informed user consent and the data transfer agreement that allow us to use the data for research, but we do not have permission to make the data publicly available.

Conflicts of Interest

Diana Montezuma, Domingos Oliveira and Isabel Macedo Pinto are employed by IMP Diagnostics.

Abbreviations

The following abbreviations are used in this manuscript:
AccAccuracy
AUCArea Under the Receiver Operating Characteristic Curve
CPathComputational Pathology
DADomain Adaptation
DLDeep Learning
H&EHematoxylin and Eosin
HGHigh-Grade
IIDIndependent and Identically Distributed
LGLow-Grade
MILMultiple Instance Learning
MLMachine Learning
NNeoNon-neoplastic
PrecPrecision
QWKQuadratic Weighted Kappa
SensSensitivity
SLSupervised Learning
SmSLSemi-supervised Learning
SpecSpecificity
SSDASemi-supervised Domain Adaptation
UDAUnsupervised Domain Adaptation
WSIWhole Slide Image
WSLWeakly Supervised Learning

References

  1. Bilal, M.; Jewsbury, R.; Wang, R.; AlGhamdi, H.M.; Asif, A.; Eastwood, M.; Rajpoot, N. An aggregation of aggregation methods in computational pathology. Med. Image Anal. 2023, 88, 102885. [Google Scholar] [CrossRef] [PubMed]
  2. Shao, Z.; Bian, H.; Chen, Y.; Wang, Y.; Zhang, J.; Ji, X.; Zhang, Y. TransMIL: Transformer based Correlated Multiple Instance Learning for Whole Slide Image Classification. In Proceedings of the NeurIPS, online, 6–14 December 2021; pp. 2136–2147. [Google Scholar]
  3. Li, B.; Li, Y.; Eliceiri, K.W. Dual-Stream Multiple Instance Learning Network for Whole Slide Image Classification With Self-Supervised Contrastive Learning. Computer Vision Foundation/IEEE. In Proceedings of the CVPR, Nashville, TN, USA, 20–25 June 2021; pp. 14318–14328. [Google Scholar]
  4. Chen, R.J.; Chen, C.; Li, Y.; Chen, T.Y.; Trister, A.D.; Krishnan, R.G.; Mahmood, F. Scaling Vision Transformers to Gigapixel Images via Hierarchical Self-Supervised Learning. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 16123–16134. [Google Scholar]
  5. Kang, M.; Song, H.; Park, S.; Yoo, D.; Pereira, S. Benchmarking Self-Supervised Learning on Diverse Pathology Datasets. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 3344–3354. [Google Scholar]
  6. Wang, X.; Yang, S.; Zhang, J.; Wang, M.; Zhang, J.; Yang, W.; Huang, J.; Han, X. Transformer-based unsupervised contrastive learning for histopathological image classification. Med. Image Anal. 2022, 81, 102559. [Google Scholar] [CrossRef] [PubMed]
  7. Ciga, O.; Xu, T.; Martel, A.L. Self supervised contrastive learning for digital histopathology. Mach. Learn. Appl. 2022, 7, 100198. [Google Scholar] [CrossRef]
  8. Srinidhi, C.L.; Kim, S.W.; Chen, F.; Martel, A.L. Self-supervised driven consistency training for annotation efficient histopathology image analysis. Med. Image Anal. 2022, 75, 102256. [Google Scholar] [CrossRef]
  9. Neto, P.C.; Montezuma, D.; Oliveira, S.P.; Oliveira, D.; Fraga, J.; Monteiro, A.; Monteiro, J.; Ribeiro, L.; Gonçalves, S.; Reinhard, S.; et al. An interpretable machine learning system for colorectal cancer diagnosis from pathology slides. Npj Precis. Oncol. 2024, 8, 56. [Google Scholar] [CrossRef]
  10. Sohn, K.; Berthelot, D.; Carlini, N.; Zhang, Z.; Zhang, H.; Raffel, C.; Cubuk, E.D.; Kurakin, A.; Li, C. FixMatch: Simplifying Semi-Supervised Learning with Consistency and Confidence. In Proceedings of the NeurIPS, online, 6–12 December 2020. [Google Scholar]
  11. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar] [CrossRef]
  12. Filiot, A.; Ghermi, R.; Olivier, A.; Jacob, P.; Fidon, L.; Mac Kain, A.; Saillard, C.; Schiratti, J.B. Scaling Self-Supervised Learning for Histopathology with Masked Image Modeling. medRxiv 2023, 2023-07. [Google Scholar]
  13. Cortes, C.; Mansour, Y.; Mohri, M. Learning Bounds for Importance Weighting. In NIPS; Curran Associates, Inc.: Newry, UK, 2010; pp. 442–450. [Google Scholar]
  14. Zhao, P.; Wang, W.; Lu, Y.; Liu, H.; Yao, S. Transfer robust sparse coding based on graph and joint distribution adaption for image representation. Knowl. Based Syst. 2018, 147, 1–11. [Google Scholar] [CrossRef]
  15. Courty, N.; Flamary, R.; Tuia, D.; Rakotomamonjy, A. Optimal Transport for Domain Adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1853–1865. [Google Scholar] [CrossRef]
  16. Ferreira, P.M.; Pernes, D.; Rebelo, A.; Cardoso, J.S. DeSIRe: Deep Signer-Invariant Representations for Sign Language Recognition. IEEE Trans. Syst. Man. Cybern. Syst. 2021, 51, 5830–5845. [Google Scholar] [CrossRef]
  17. Zhao, H.; des Combes, R.T.; Zhang, K.; Gordon, G.J. On Learning Invariant Representations for Domain Adaptation. In Proceedings of the Proceedings of Machine Learning Research, ICML, PMLR, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 7523–7532. [Google Scholar]
  18. Pernes, D.; Cardoso, J.S. Tackling unsupervised multi-source domain adaptation with optimism and consistency. Expert Syst. Appl. 2022, 194, 116486. [Google Scholar] [CrossRef]
  19. Berthelot, D.; Roelofs, R.; Sohn, K.; Carlini, N.; Kurakin, A. Adamatch: A unified approach to semi-supervised learning and domain adaptation. arXiv 2021, arXiv:2106.04732. [Google Scholar]
  20. Ren, J.; Hacihaliloglu, I.; Singer, E.A.; Foran, D.J.; Qi, X. Adversarial Domain Adaptation for Classification of Prostate Histopathology Whole-Slide Images. In Lecture Notes in Computer Science, Proceedings of the MICCAI (2), Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11071, pp. 201–209. [Google Scholar]
  21. Koohbanani, N.A.; Unnikrishnan, B.; Khurram, S.A.; Krishnaswamy, P.; Rajpoot, N.M. Self-Path: Self-Supervision for Classification of Pathology Images With Limited Annotations. IEEE Trans. Med. Imaging 2021, 40, 2845–2856. [Google Scholar] [CrossRef] [PubMed]
  22. Abbet, C.; Studer, L.; Fischer, A.; Dawson, H.; Zlobec, I.; Bozorgtabar, B.; Thiran, J. Self-rule to multi-adapt: Generalized multi-source feature learning using unsupervised domain adaptation for colorectal cancer tissue detection. Med. Image Anal. 2022, 79, 102473. [Google Scholar] [CrossRef] [PubMed]
  23. Marini, N.; Otalora, S.; Wodzinski, M.; Tomassini, S.; Dragoni, A.F.; Marchand-Maillet, S.; Morales, J.P.D.; Duran-Lopez, L.; Vatrano, S.; Müller, H.; et al. Data-driven color augmentation for H&E stained images in computational pathology. J. Pathol. Inform. 2023, 14, 100183. [Google Scholar] [CrossRef]
  24. de Bel, T.; Bokhorst, J.M.; van der Laak, J.; Litjens, G. Residual cyclegan for robust domain transformation of histopathological tissue slides. Med. Image Anal. 2021, 70, 102004. [Google Scholar] [CrossRef]
  25. Runz, M.; Rusche, D.; Schmidt, S.; Weihrauch, M.R.; Hesser, J.; Weis, C.A. Normalization of HE-stained histological images using cycle consistent generative adversarial networks. Diagn. Pathol. 2021, 16, 71. [Google Scholar] [CrossRef]
  26. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef]
  27. Kirk, S.; Lee, Y.; Sadow, C.; Levine, S.; Roche, C.; Bonaccio, E.; Filiippini, J. Radiology Data from The Cancer Genome Atlas Colon Adenocarcinoma [TCGA-COAD] Collection. The Cancer Imaging Archive. 2016. Available online: https://www.cancerimagingarchive.net/collection/tcga-coad/ (accessed on 29 May 2020).
  28. Zhou, J.; Wei, C.; Wang, H.; Shen, W.; Xie, C.; Yuille, A.L.; Kong, T. Image BERT Pre-training with Online Tokenizer. In Proceedings of the ICLR, online, 25 April 2022; Available online: https://arxiv.org/abs/2111.07832 (accessed on 25 April 2022).
  29. Oliveira, S.P.; Neto, P.C.; Fraga, J.; Montezuma, D.; Monteiro, A.; Monteiro, J.; Ribeiro, L.; Gonçalves, S.; Pinto, I.M.; Cardoso, J.S. CAD systems for colorectal cancer from WSI are still not ready for clinical acceptance. Sci. Rep. 2021, 11, 14358. [Google Scholar] [CrossRef]
  30. Neto, P.C.; Oliveira, S.P.; Montezuma, D.; Fraga, J.; Monteiro, A.; Ribeiro, L.; Gonçalves, S.; Pinto, I.M.; Cardoso, J.S. iMIL4PATH: A semi-supervised interpretable approach for colorectal whole-slide images. Cancers 2022, 14, 2489. [Google Scholar] [CrossRef]
  31. Tellez, D.; Litjens, G.; Bándi, P.; Bulten, W.; Bokhorst, J.; Ciompi, F.; van der Laak, J. Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology. Med. Image Anal. 2019, 58, 101544. [Google Scholar] [CrossRef]
  32. de La Torre, J.; Puig, D.; Valls, A. Weighted kappa loss function for multi-class classification of ordinal data in deep learning. Pattern Recognit. Lett. 2018, 105, 144–154. [Google Scholar] [CrossRef]
  33. Zhang, M.R.; Lucas, J.; Ba, J.; Hinton, G.E. Lookahead Optimizer: K steps forward, 1 step back. In Proceedings of the NeurIPS, Vancouver, Canada, 8–14 December 2019; pp. 9593–9604. [Google Scholar]
  34. Liu, L.; Jiang, H.; He, P.; Chen, W.; Liu, X.; Gao, J.; Han, J. On the Variance of the Adaptive Learning Rate and Beyond. In Proceedings of the ICLR, Addis Ababa, Ethiopia, 30 April 2020; Available online: https://iclr.cc/virtual_2020/poster_rkgz2aEKDr.html (accessed on 30 April 2020).
  35. Graham, S.; Vu, Q.D.; Jahanifar, M.; Raza, S.E.A.; Minhas, F.; Snead, D.; Rajpoot, N. One model is all you need: Multi-task learning enables simultaneous histology image segmentation and classification. Med. Image Anal. 2023, 83, 102685. [Google Scholar] [CrossRef] [PubMed]
  36. Yu, X.; Liang, X.; Zhou, Z.; Zhang, B. Multi-task learning for hand heat trace time estimation and identity recognition. Exp. Syst. Appl. 2014, 255, 124551. [Google Scholar] [CrossRef]
  37. Chen, R.J.; Lu, M.Y.; Williamson, D.F.K.; Chen, T.Y.; Lipkova, J.; Noor, Z.; Shaban, M.; Shady, M.; Williams, M.; Joo, B.; et al. Pan-cancer integrative histology-genomic analysis via multimodal deep learning. Cancer Cell 2022, 40, 865–878.e6. [Google Scholar] [CrossRef]
Figure 1. Illustration of the proposed framework for weakly supervised domain adaptive colorectal tissue classification. Top: Fixmatch based strategy; Middle: CycleGAN; Bottom: Phikon feature extractor.
Figure 1. Illustration of the proposed framework for weakly supervised domain adaptive colorectal tissue classification. Top: Fixmatch based strategy; Middle: CycleGAN; Bottom: Phikon feature extractor.
Sensors 25 02856 g001
Figure 2. H&E-stained images from CRS10K—IMP Diagnostics (source) and BernCRC (target). The stain color distribution shift between source and target data is quite evident.
Figure 2. H&E-stained images from CRS10K—IMP Diagnostics (source) and BernCRC (target). The stain color distribution shift between source and target data is quite evident.
Sensors 25 02856 g002
Figure 3. Illustration of data augmentation on H&E-stained images from CRS10K. Stain color variation after data augmentation is the most salient feature.
Figure 3. Illustration of data augmentation on H&E-stained images from CRS10K. Stain color variation after data augmentation is the most salient feature.
Sensors 25 02856 g003
Figure 4. Summary of results for Unsupervised Domain Adaptation.
Figure 4. Summary of results for Unsupervised Domain Adaptation.
Sensors 25 02856 g004
Figure 5. Summary of results for Semi-supervised Domain Adaptation.
Figure 5. Summary of results for Semi-supervised Domain Adaptation.
Sensors 25 02856 g005
Table 1. Data augmentation tailored to CPath as proposed in the work by Tellez et al. [31].
Table 1. Data augmentation tailored to CPath as proposed in the work by Tellez et al. [31].
TransformConfounding Factor(s)Hyperparameters
Hue ShiftStainingshift limits = [ 0.125 , 0.125 ]
Contrast ShiftStaining/Whole-slide scannershift limits = [ 0.2 , 0.2 ]
Saturation ShiftStaining/Whole-slide scannershift limits = [ 0.125 , 0.125 ]
HED JitterStaining α = 0.1 β = 0.0075
Flips (Horizontal/Vertical)Viewpoint-
RotationViewpointlimits = [ 90 , 90 ]
ScalingWhole-slide scannerlimits = [ 0.8 , 1.2 ]
Gaussian BlurWhole-slide scannermax kernel size = 15, σ limits = [ 0.1 , 2 ]
Gaussian NoiseWhole-slide scanner μ = 0 , σ = [ 0 , 0.1 ]
Table 2. Results on the source domain (CRS10K) test set for grading of colorectal biopsies and polypectomies into Non-neoplastic (NNeo); Low-grade (LG), and High-grade (HG) lesions. We report the mean and standard deviation of the k-fold (k = 5) cross-validation on the CRS10K test set. Values in bold indicate the best result.
Table 2. Results on the source domain (CRS10K) test set for grading of colorectal biopsies and polypectomies into Non-neoplastic (NNeo); Low-grade (LG), and High-grade (HG) lesions. We report the mean and standard deviation of the k-fold (k = 5) cross-validation on the CRS10K test set. Values in bold indicate the best result.
MethodMIL Train/Val SetQWKAccAUC (NNeo vs. Rest)AUC (LG vs. Rest)AUC (HG vs. Rest)
SL (no data aug)CRS1K0.87590.89190.99110.96270.9661
SL (CRS1K)CRS1K0.89770.90970.99550.97070.9780
FixMatchCRS1K0.9120 ± 0.00960.9222 ± 0.00520.9956 ± 0.00110.9768 ± 0.00210.9815 ± 0.0031
FixMatchBernCRC+CRS1K0.9189 ± 0.00860.9289 ± 0.00600.9953 ± 0.00060.9781 ± 0.00160.9834 ± 0.0011
Table 3. Results on the source domain (CRS10K) test set for binary classification of colorectal dysplasia. To perform binary classification, we combine the Low-grade (LG) and High-grade (HG) labels into a single class. We perform K-fold cross-validation and report the mean and standard deviation on the CRS10K test set. Values in bold indicate the best result.
Table 3. Results on the source domain (CRS10K) test set for binary classification of colorectal dysplasia. To perform binary classification, we combine the Low-grade (LG) and High-grade (HG) labels into a single class. We perform K-fold cross-validation and report the mean and standard deviation on the CRS10K test set. Values in bold indicate the best result.
MethodTrain/Val setAcc (Bin)SensSpecPrecF1
SL (no data aug)CRS1K0.96880.98190.91570.97920.9464
SL (CRS1K)CRS1K0.97440.98470.93260.98330.9573
FixMatchCRS1K0.9792 ± 0.00440.9864 ± 0.00270.9506 ± 0.01570.9877 ± 0.00380.9688 ± 0.0100
FixMatchBernCRC+CRS1K0.9810 ± 0.00280.9878 ± 0.00100.9539 ± 0.01200.9886 ± 0.00300.9709 ± 0.0076
Table 4. Fréchet Inception Distance (FID) between BernCRC (target) transformed to CRS10K (source) and BernCRC (target) to CRS10K (source).
Table 4. Fréchet Inception Distance (FID) between BernCRC (target) transformed to CRS10K (source) and BernCRC (target) to CRS10K (source).
FID BernCRC (Target) Transformed—CRS10K (Source)FID BernCRC (Target)—crs10k (Source)
9.994021.2732
Table 5. Results of Unsupervised Domain Adaptation (UDA) for colorectal tissue classification into Non-neoplastic (NNeo); Low-grade (LG); and High-grade (HG) lesions. We report the mean and standard deviation of the k-fold (k = 5) cross-validation on BernCRC. Values in bold indicate the best result.
Table 5. Results of Unsupervised Domain Adaptation (UDA) for colorectal tissue classification into Non-neoplastic (NNeo); Low-grade (LG); and High-grade (HG) lesions. We report the mean and standard deviation of the k-fold (k = 5) cross-validation on BernCRC. Values in bold indicate the best result.
MethodMIL Train/Val SetQWKAccAUC (NNeo vs. Rest)AUC (LG vs. Rest)AUC (HG vs. Rest)
SL (no data aug)CRS1K0.0746 ± 0.05290.4045 ± 0.04090.7034 ± 0.08000.5593 ± 0.07120.3055 ± 0.0433
SL (CRS1K)CRS1K0.4024 ± 0.07920.6155 ± 0.04430.9302 ± 0.03430.8025 ± 0.07930.7159 ± 0.1220
FixMatchCRS1K0.5433 ± 0.15610.7433 ± 0.02880.9607 ± 0.02880.8330 ± 0.08870.8187 ± 0.1128
CycleGANCRS1K0.5670 ± 0.06850.8218 ± 0.04240.9748 ± 0.00910.8654 ± 0.06680.7435 ± 0.0814
Table 6. Results of the Unsupervised Domain Adaptation (UDA) for binary classification of colorectal dysplasia. To perform binary classification, we combine the Low-grade (LG) and High-grade (HG) labels into a single class. We report the mean and standard deviation of the k-fold (k = 5) cross-validation in BernCRC. Values in bold indicate the best result.
Table 6. Results of the Unsupervised Domain Adaptation (UDA) for binary classification of colorectal dysplasia. To perform binary classification, we combine the Low-grade (LG) and High-grade (HG) labels into a single class. We report the mean and standard deviation of the k-fold (k = 5) cross-validation in BernCRC. Values in bold indicate the best result.
MethodMIL Train/Val SetAcc (bin)SensSpecPrecF1
SL (no data aug)CRS1K0.4618 ± 0.06490.4000 ± 0.08950.8333 ± 0.14910.9389 ± 0.05600.8797 ± 0.1084
SL (CRS1K)CRS1K0.8049 ± 0.05020.7835 ± 0.06300.9333 ± 0.08170.9856 ± 0.01770.9576 ± 0.0519
FixMatchCRS1K0.9062 ± 0.04720.9265 ± 0.05580.8000 ± 0.26670.9650 ± 0.046080.8529 ± 0.2030
CycleGANCRS1K0.9267 ± 0.04730.91362 ± 0.05641 ± 01 ± 01 ± 0
Table 7. Results of Semi-supervised Domain Adaptation (SSDA) for grading of colorectal biopsies and polypectomies into Non-neoplastic (NNeo); Low-grade (LG); and High-grade (HG) lesions. We report the mean and standard deviation of the k-fold (k = 5) cross-validation in BernCRC. Values in bold indicate the best result.
Table 7. Results of Semi-supervised Domain Adaptation (SSDA) for grading of colorectal biopsies and polypectomies into Non-neoplastic (NNeo); Low-grade (LG); and High-grade (HG) lesions. We report the mean and standard deviation of the k-fold (k = 5) cross-validation in BernCRC. Values in bold indicate the best result.
MethodMIL Train/Val SetQWKAccAUC (NNeo vs. Rest)AUC (LG vs. Rest)AUC (HG vs. Rest)
SL (no data aug)BernCRC0.6678 ± 0.24690.8394 ± 0.12280.9277 ± 0.04340.8236 ± 0.12480.8450 ± 0.1641
SL (CRS1K)BernCRC0.6816 ± 0.15410.8435 ± 0.07470.98662 ± 0.01070.8238 ± 0.10050.7767 ± 0.1707
PhikonBernCRC0.6786 ± 0.06570.8103 ± 0.04760.9949 ± 0.00780.8793 ± 0.02870.8283 ± 0.0395
FixMatchBernCRC0.7011 ± 0.07480.8532 ± 0.04310.9906 ± 0.00860.8503 ± 0.06300.8181 ± 0.0645
FixMatchBernCRC+CRS1K0.7093 ± 0.10370.8531 ± 0.04640.9865 ± 0.01450.8466 ± 0.04840.8241 ± 0.0573
CycleGANBernCRC0.6088 ± 0.08280.8324 ± 0.04880.9914 ± 0.00950.8642 ± 0.07540.8071 ± 0.1150
Table 8. Results of Semi-supervised Domain Adaptation (SSDA) for binary classification of colorectal dysplasia. To perform binary classification, we combine the Low-grade (LG) and High-grade (HG) labels into a single class. We report the mean and standard deviation of the k-fold (k = 5) cross-validation in BernCRC. Values in bold indicate the best result.
Table 8. Results of Semi-supervised Domain Adaptation (SSDA) for binary classification of colorectal dysplasia. To perform binary classification, we combine the Low-grade (LG) and High-grade (HG) labels into a single class. We report the mean and standard deviation of the k-fold (k = 5) cross-validation in BernCRC. Values in bold indicate the best result.
MethodMIL Train/Val SetAcc (Bin)SensSpecPrecF1
SL (no data aug)BernCRC0.9410 ± 0.03120.9812 ± 0.03750.7000 ± 0.12650.9523 ± 0.02180.8016 ± 0.0954
SL (CRS1K)BernCRC0.9476 ± 0.01580.9691 ± 0.01980.8201 ± 0.10660.9699 ± 0.01860.8860 ± 0.0699
PhikonBernCRC0.9846 ± 0.02050.9939 ± 0.01210.9333 ± 0.08170.9881 ± 0.01460.9586 ± 0.0507
FixMatchBernCRC0.9636 ± 0.02620.9814 ± 0.02490.8600 ± 0.12720.9763 ± 0.02210.9108 ± 0.0828
FixMatchBernCRC+CRS1K0.9583 ± 0.04190.9691 ± 0.02750.9000 ± 0.13330.9814 ± 0.02450.9351 ± 0.0875
CycleGANBernCRC0.9476 ± 0.01580.9629 ± 0.02300.8600 ± 0.12720.9761 ± 0.02170.9107 ± 0.0826
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nunes, J.D.; Montezuma, D.; Oliveira, D.; Pereira, T.; Zlobec, I.; Pinto, I.M.; Cardoso, J.S. Bridging Domain Gaps in Computational Pathology: A Comparative Study of Adaptation Strategies. Sensors 2025, 25, 2856. https://doi.org/10.3390/s25092856

AMA Style

Nunes JD, Montezuma D, Oliveira D, Pereira T, Zlobec I, Pinto IM, Cardoso JS. Bridging Domain Gaps in Computational Pathology: A Comparative Study of Adaptation Strategies. Sensors. 2025; 25(9):2856. https://doi.org/10.3390/s25092856

Chicago/Turabian Style

Nunes, João D., Diana Montezuma, Domingos Oliveira, Tania Pereira, Inti Zlobec, Isabel Macedo Pinto, and Jaime S. Cardoso. 2025. "Bridging Domain Gaps in Computational Pathology: A Comparative Study of Adaptation Strategies" Sensors 25, no. 9: 2856. https://doi.org/10.3390/s25092856

APA Style

Nunes, J. D., Montezuma, D., Oliveira, D., Pereira, T., Zlobec, I., Pinto, I. M., & Cardoso, J. S. (2025). Bridging Domain Gaps in Computational Pathology: A Comparative Study of Adaptation Strategies. Sensors, 25(9), 2856. https://doi.org/10.3390/s25092856

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop