Next Article in Journal
Contemporary Mechanical Support Devices for Temporary and Long-Term Applications
Next Article in Special Issue
Operations Research for Pediatric Elective Surgery Planning: Example of a Mathematical Model
Previous Article in Journal
The Effect of Genipin Matrix Augmentation on the Retention of Glycosaminoglycans in the Intervertebral Disc—A Pilot Study
Previous Article in Special Issue
Digital Twin and Artificial Intelligence Technologies to Assess the Type IA Endoleak
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

A Systematic Review of Contrastive Learning in Medical AI: Foundations, Biomedical Modalities, and Future Directions

by
George Obaido
1,*,
Ibomoiye Domor Mienye
1,
Kehinde Aruleba
1,
Chidozie Williams Chukwu
2,
Ebenezer Esenogho
1,* and
Cameron Modisane
1
1
Center for Artificial Intelligence and Multidisciplinary Innovations, Department of Auditing, College of Accounting Sciences, University of South Africa, Pretoria 0002, South Africa
2
Department of Mathematical Sciences, Georgia Southern University, Statesboro, GA 30460, USA
*
Authors to whom correspondence should be addressed.
Bioengineering 2026, 13(2), 176; https://doi.org/10.3390/bioengineering13020176
Submission received: 25 December 2025 / Revised: 24 January 2026 / Accepted: 26 January 2026 / Published: 2 February 2026

Abstract

Medical artificial intelligence (AI) systems depend heavily on high-quality data representations to support accurate prediction, diagnosis, and clinical decision-making. However, the availability of large, well-annotated medical datasets is often constrained by cost, privacy concerns, and the need for expert labeling, motivating growing interest in self-supervised representation learning. Among these approaches, contrastive learning has emerged as one of the most influential paradigms, driving major advances in representation learning across computer vision and natural language processing. This paper presents a comprehensive review of contrastive learning in medical AI, highlighting its theoretical foundations, methodological developments, and practical applications in medical imaging, electronic health records, physiological signal analysis, and genomics. Furthermore, we identify recurring challenges, including pair construction, sensitivity to data augmentations, and inconsistencies in evaluation protocols, while discussing emerging trends such as multimodal alignment, federated learning, and privacy-preserving frameworks. Through a synthesis of current developments and open research directions, this review provides insights to advance data-efficient, reliable, and generalizable medical AI systems.

1. Introduction

The increasing digitalization of healthcare has led to an unprecedented accumulation of multimodal medical data, including imaging, clinical records, physiological signals, and genomics [1,2,3,4]. These data provide vast opportunities for artificial intelligence (AI) to improve diagnosis, prognosis, and treatment planning. However, the effectiveness of AI systems in healthcare remains constrained by the scarcity of annotated datasets and the high cost of expert labeling. Medical annotation often requires domain specialists and is subject to inter-observer variability, making the implementation of large-scale supervised learning difficult [5,6,7]. This has stimulated the growing adoption of self-supervised learning (SSL) approaches, which exploit large volumes of unlabeled data to learn meaningful representations that can generalize across downstream tasks with minimal supervision.
Contrastive learning (CL) has become one of the most prominent paradigms within SSL. It operates by comparing data pairs to bring similar instances closer in representation space while pushing dissimilar ones apart [8,9,10,11,12]. Unlike traditional supervised methods, CL relies on instance discrimination and data augmentations to build robust representations without requiring human annotations. Recent studies have demonstrated the effectiveness of CL for medical representation learning by adapting foundational CL frameworks, such as SimCLR, MoCo, BYOL, and SwAV, to domain-specific data characteristics [13,14,15]. For example, Azizi et al. [16] introduced Multi-Instance Contrastive Learning (MICLe) to exploit multiple images per patient case during self-supervised pretraining, improving label efficiency and downstream performance in medical image classification. Sowrirajan et al. [17] proposed MoCo-CXR, a Momentum Contrast adaptation for chest X-ray interpretation, showing improved representation quality and transferability, particularly in low-label regimes and external datasets. Beyond image-only CL, Zhang et al. [18] presented ConVIRT, which aligns medical images with paired radiology reports via a bidirectional contrastive objective, yielding substantial gains in data-efficient learning and retrieval-based evaluation. In physiological signal modeling, Diamant et al. [19] developed Patient Contrastive Learning of Representations (PCLR) for ECGs, using patient identity over time to define positives and demonstrating improved clinical prediction performance across multiple downstream tasks. Consequently, these methods have been extended to medical domains, where data scarcity and heterogeneity remain pressing challenges. The ability of CL to learn invariant and transferable features makes it particularly suitable for applications in medical imaging, electronic health records (EHRs), and multi-omics analysis.
Several recent reviews have examined self-supervised and CL, although their coverage of medical applications is often limited or modality-specific. Jaiswal et al. [20] provided an early survey outlining theoretical principles and algorithmic variants, but did not address domain-specific adaptations for medical data. Gui et al. [21] broadened the scope to include generative and clustering-based self-supervised methods, yet discussion of healthcare contexts remained minimal. Hu et al. [22] presented a comprehensive and systematic survey of CL, summarizing core principles and a universal CL framework, and synthesizing advances across key components such as augmentations, sampling strategies, architectures, and loss functions. Liu [23] reviewed CL for visual representation learning, highlighting key components, limitations, and practical strategies for improving CL pipelines in computer vision. More medically focused reviews include studies, such as Shurrab et al. [24] and several newer works that explicitly target clinical settings. Wang et al. [25] reviewed predictive and contrastive self-supervised learning for medical images, with emphasis on how natural image SSL methods are adapted for medical data. Huang et al. [26] systematically reviewed self-supervised learning for medical image classification across studies published between 2012 and 2022. VanBerlo et al. [27] surveyed evidence on the impact of self-supervised pretraining across imaging modalities, including X-ray, CT, MRI, and ultrasound, with attention to comparisons against supervised baselines and transfer learning protocols. Table 1 summarizes representative review papers on CL, highlighting their scope and coverage of medical applications.
In contrast to prior reviews that primarily emphasize general CL foundations or focus narrowly on medical imaging, this review provides a cross-modality synthesis spanning medical imaging, EHRs, physiological signals, genomics and proteomics, and multimodal vision language systems. We further emphasize clinically grounded design choices for medical CL, including pairing strategies, augmentation validity, evaluation regimes, and reporting requirements for reproducibility and external validation.
While informative, these studies share common limitations. They often restrict attention to a single modality, typically medical imaging, provide limited cross-modality synthesis, and lack an operational taxonomy that clearly links CL design choices to clinical data characteristics. In addition, evaluation practices are frequently underanalyzed, with insufficient emphasis on external validation, robustness, reproducibility, and cross-domain generalization. To address these gaps, this paper provides a comprehensive review of CL in medical AI, covering its theoretical foundations, methodological advances, and practical applications across diverse biomedical data modalities. The main contributions of this study are as follows:
  • We propose an operational, domain-aware taxonomy of medical CL methods grounded in core design components, including loss functions, positive and negative pairing strategies, augmentation policies, and evaluation regimes.
  • We synthesize CL applications across key medical modalities, including medical imaging, electronic health records, physiological time series, genomics and proteomics, and multimodal vision language systems, highlighting modality-specific challenges and transferable design patterns.
  • We critically examine methodological limitations in the literature, with emphasis on evaluation heterogeneity, reproducibility constraints, data access limitations, and the need for standardized benchmarking and external validation.
  • We provide practical guidance and prioritized future directions for clinically trustworthy CL, focusing on robustness to distribution shift, interpretability, fairness, privacy-preserving training, and deployment considerations.
The remainder of the paper is organized as follows. Section 2 presents the methodology of this review. Section 3 describes the foundations of CL. Section 4 reviews applications across medical imaging, electronic health records, genomics and proteomics, multimodal learning, and physiological signal analysis. Section 5 discusses key challenges and limitations. Section 6 outlines future research directions, and Section 7 concludes the paper.

2. Methodology

We conducted a systematic review to identify and synthesize CL methods applied to medical and biomedical data. The review question was “How are CL objectives designed and evaluated across medical modalities, and what methodological practices enable robust, reproducible, and clinically meaningful performance?”

2.1. Databases and Search Strategy

We searched PubMed, IEEE Xplore, ACM Digital Library, Web of Science, Scopus, and arXiv for studies published between January 2019 and October 2025. The time window captures the maturation of modern CL frameworks and their adoption in medical AI. Searches were executed on 31 October 2025.
Search queries combined contrastive/self-supervised learning terms with medical domain keywords. Representative query patterns included:
  • (“contrastive learning” OR “self-supervised”) AND (medical OR healthcare OR clinical)
  • (“contrastive learning” OR SimCLR OR MoCo OR BYOL) AND (radiology OR “chest x-ray” OR MRI OR CT)
  • (“contrastive learning” OR “self-supervised”) AND (“electronic health record” OR EHR OR “clinical notes”)
  • (“contrastive learning” OR “representation learning”) AND (ECG OR EEG OR “physiological signals”)
  • (“contrastive learning” OR “self-supervised”) AND (genomics OR proteomics OR “single-cell”)
Search strings were adapted to database-specific syntax and indexing conventions. Backward and forward citation chasing was performed for key studies to identify additional relevant articles.

2.2. Eligibility Criteria

We included:
  • Peer-reviewed journal articles and full conference papers; influential preprints were included selectively when they introduced widely adopted methods, benchmarks, or were heavily cited in subsequent peer-reviewed work.
  • Studies that explicitly employed CL or closely related objectives (e.g., InfoNCE, supervised contrastive loss, MoCo/SimCLR/BYOL-style frameworks, CLIP-style alignment).
  • Studies using medical or biomedical data (medical imaging, EHRs, physiological time series, genomics/proteomics, pathology, or multimodal combinations).
We excluded:
  • Non-medical studies without biomedical datasets or clinically motivated tasks.
  • Abstract-only records, posters lacking sufficient methodological detail, editorials, commentaries, and theses.
  • Non-validated technical reports and preprints without experimental results, or preprints superseded by peer-reviewed versions.
When multiple papers described incremental versions of the same method, we prioritized the most comprehensive peer-reviewed version and retained earlier versions for historical context when needed.

2.3. Study Selection

Records were deduplicated, then screened by title/abstract, followed by full-text assessment using the eligibility criteria. Screening was performed independently by G.O. and I.D.M. Disagreements were resolved by consensus, and unresolved conflicts were adjudicated by K.A., E.E., C.M., and C.W.C. Reasons for full-text exclusion were recorded (e.g., non-medical domain, insufficient methodological detail, no contrastive objective, no empirical evaluation). Of 612 identified records, 300 were screened after duplicate and automated exclusions. Following a full-text assessment of 94 reports, 38 studies were included. Figure 1 summarizes the selection process.

2.4. Data Extraction

We extracted study characteristics using a standardized template and cross-checked entries for consistency. For each study, we recorded the following:
  • Modality and dataset(s);
  • Contrastive formulation (loss, supervision regime, pairing strategy, negative sampling);
  • Encoder architecture and pretraining scale;
  • Downstream task(s), evaluation regime (linear probe, fine-tuning, few-shot/zero-shot), and metric(s);
  • Reported limitations, external validation, and reproducibility artifacts (e.g., code availability).

2.5. Taxonomy of Medical Contrastive Learning

To consolidate the heterogeneous methodological landscape of medical CL, we introduce a taxonomy that groups approaches according to a small set of design dimensions that materially affect (i) the clinical invariances encoded in learned representations, (ii) the source and strength of supervision used during pretraining, and (iii) the evidentiary standard used to claim downstream benefit. The goal of this taxonomy is twofold. First, it provides a consistent analytic framework for comparing studies across modalities and clinical endpoints. Second, it serves as a reporting scaffold by making explicit the minimum set of methodological decisions that must be specified to support reproducibility and meaningful clinical interpretation.
Table 2 summarizes the taxonomy dimensions applied throughout this manuscript. The loss family or objective defines the learning signal and distinguishes objectives that rely on explicit negatives, such as InfoNCE or NT-Xent, incorporate labels during pretraining through supervised CL, adopt negative-free teacher–student distillation, or implement clustering-based consistency. The pairing strategy specifies what constitutes a positive relation in clinical data, including augmentations of the same instance, same-class positives, temporally adjacent windows, and patient-level correspondences across time. This choice is particularly consequential in medical datasets because cohort structure, repeated measures, and phenotype similarity can lead to false negatives and shortcut learning unless pairing rules are explicitly controlled. Augmentation and view design operationalize invariances and should be clinically plausible, since modality-inappropriate transformations may suppress pathology in imaging, distort physiological rhythms in biosignals, or alter semantic content in clinical text.
The label regime captures the extent of supervision used during representation learning and distinguishes unsupervised settings from weakly supervised, semi-supervised, and fully supervised pretraining. This distinction is essential for fair comparison because label access during pretraining changes both the information content of the learning signal and the interpretation of label-efficiency gains. The evaluation protocol defines how representation quality is assessed, including the downstream training regime, linear probing versus fine-tuning, evaluation under label scarcity through label-fraction sweeps, and testing under clinically relevant distribution shifts such as temporal, site, scanner, or device, and demographic shift. Calibration and uncertainty reporting are also important when risk prediction is a primary endpoint. Finally, studies are grouped by task family, including classification, segmentation, retrieval, and prognosis, since both contrastive design choices and evaluation expectations are strongly task-dependent and should be compared under aligned clinical objectives.

2.6. Quality and Reporting Appraisal

Because medical CL studies vary widely in design and reporting, we used a lightweight methodological checklist focused on reproducibility and clinical validity (Table A1; Appendix B). The checklist was used to characterize the evidence base and identify common gaps; it was not used to exclude studies.

2.7. Protocol Registration

No protocol was preregistered. This review was conducted to map and synthesize rapidly evolving methodological literature. To support transparency and reproducibility, we provide database-specific search strategies (Appendix A) and an explicit appraisal checklist (Table A1; Appendix B).

3. Overview of Contrastive Learning

CL is a self-supervised learning technique designed to learn effective representations from unlabeled data by distinguishing between positive and negative pairs [11,22,28,29,30]. Unlike traditional supervised learning, which heavily depends on labeled data to guide the learning process, CL utilizes the intrinsic structure of the data to derive semantically meaningful representations [20,21,31,32]. This makes it especially effective in situations where labeled data is limited or difficult to obtain. The concept of CL originated from the broader field of self-supervised learning, a paradigm that has seen increasing interest due to its ability to leverage vast amounts of unlabeled data. The early developments of CL date back to the 1990s with the introduction of metric learning and Siamese networks, which were designed to learn similarity metrics between pairs of data points [33]. These early forms laid the groundwork for the evolution of CL into a robust tool for modern machine learning applications. Figure 2 illustrates the general workflow of CL, in which an encoder parameterized by θ maps an anchor, a positive, and a negative example into a shared embedding space. The goal is to maximize similarity between the anchor and positive pair while minimizing similarity to the negative.
Originally introduced for verification tasks using Siamese networks, contrastive objectives gained traction through contrastive predictive coding (CPC) and frameworks, such as SimCLR and MoCo, which demonstrated that high-quality representations could be learned from unlabeled data at scale [20,34,35,36]. These advances catalyzed rapid adoption across fields including medical imaging, text analysis, and multimodal learning [37,38].
Figure 2. Steps in the CL process, adapted from [39].
Figure 2. Steps in the CL process, adapted from [39].
Bioengineering 13 00176 g002
Furthermore, the theoretical foundations of CL are rooted in the idea of representation learning through similarity and dissimilarity. The main goal is to learn a function f θ that maps input data points to a representation space where semantically similar inputs are closer together, while semantically dissimilar inputs are further apart [21]. This is operationalized through a contrastive loss function, commonly referred to as Noise Contrastive Estimation (InfoNCE) loss. The InfoNCE loss is defined as follows:
L = log exp ( sim ( z i , z j ) / τ ) k = 1 K exp ( sim ( z i , z k ) / τ ) ,
where z i and z j are the embeddings of the anchor and the positive sample, sim ( z i , z j ) denotes the similarity between two embeddings, typically calculated using cosine similarity: sim ( z i , z j ) = z i · z j z i z j , τ is a temperature parameter that regulates the sharpness of the similarity distribution, and the denominator sums over all K samples in the dataset, encompassing both positive and negative pairs.
This loss function seeks to maximize the similarity between the embeddings of positive pairs while minimizing the similarity between the anchor and negative samples [40,41]. The efficacy of CL is determined by several key factors. The selection of positive and negative pairs is critical, as it directly influences the quality of the learned representations. For instance, utilizing multiple views or augmentations of the same data point as positive pairs can help the model learn invariances to specific transformations [42]. Additionally, the use of larger batch sizes or memory banks can provide the model with a richer set of negative samples, enhancing the learning process. The choice of data augmentation strategies is also vital, as it determines the types of invariances the model will learn.

3.1. Variants and Extensions of Contrastive Learning

To address practical limitations such as negative sampling bias, augmentation sensitivity, modality heterogeneity, and temporal dependence, CL has evolved into several complementary variants. In medical AI, these variants differ primarily in whether they rely on explicit negatives, how positives are constructed (instance-level vs patient-level), and whether the objective aligns representations within a single modality or across modalities.

3.1.1. InfoNCE-Based Contrastive Learning (SimCLR, MoCo, SupCon)

The most widely used family of methods is based on the InfoNCE objective, which maximizes similarity between positive pairs while contrasting them against a set of negatives. SimCLR forms positive pairs using two augmented views of the same sample and relies on large batch sizes to provide many in-batch negatives [43]. Its loss can be written as follows:
L SimCLR = i I log exp ( sim ( z i , z j ) / τ ) k = 1 2 N [ k i ] exp ( sim ( z i , z k ) / τ )
SimCLR relies on carefully designed stochastic augmentations to define positive pairs, meaning that the choice of augmentation policy strongly determines what invariances are learned [44,45,46]. In medical imaging, this is particularly important because aggressive transformations, such as heavy cropping, blurring, or color jitter, can remove or distort subtle pathology signals and may unintentionally encourage shortcut learning. Because SimCLR treats all other samples in the minibatch as negatives, it benefits substantially from large batch sizes, which increase the number and diversity of in-batch negatives and improve representation quality. However, this assumption can be problematic in clinical datasets where semantically similar cases, such as patients sharing the same diagnosis or repeated examinations from related cohorts, may appear in the same batch, creating false negatives that reduce downstream performance and calibration.
MoCo improves scalability by maintaining a queue (memory bank) of negative keys and using a momentum encoder to stabilize feature representations across iterations [47,48,49]. The MoCo loss function is defined similarly to the InfoNCE loss but incorporates a momentum encoder:
L MoCo = log exp ( sim ( q , k + ) / τ ) i = 0 K exp ( sim ( q , k i ) / τ )
where q is the query embedding from the current batch, k + is the key embedding of the positive sample, k i are the embeddings of negative samples stored in the memory bank, and τ is the temperature parameter [47,50,51]. Through maintaining a queue of negative samples and using a slowly updated encoder, MoCo effectively improves the model’s ability to learn discriminative features.
Supervised Contrastive Learning (SupCon) extends the same principle to labeled settings by treating all samples from the same class as positives [29]. This formulation is relevant to medical AI for fine-tuning and hybrid training regimes where limited labels are available:
L sup = i I 1 | P ( i ) | p P ( i ) log exp ( sim ( z i , z p ) / τ ) a A ( i ) exp ( sim ( z i , z a ) / τ ) .
Although effective, InfoNCE-based methods can suffer from false negatives and batch-size dependence, which are amplified in medical datasets where patients may share similar phenotypes or repeated examinations.

3.1.2. Negative-Free Self-Distillation (BYOL, DINO, SimSiam)

A second major family of approaches removes explicit negative samples and instead relies on self-distillation or cross-view prediction between augmented views. This design is particularly appealing for medical AI, where (i) batch sizes are often constrained by high-resolution imaging or long physiological sequences, (ii) datasets are highly imbalanced, and (iii) many samples can be semantically similar due to shared diagnoses, repeated examinations, or cohort effects. In such settings, InfoNCE-style negative sampling can introduce harmful false negatives, weakening representation quality and downstream calibration.
Bootstrap Your Own Latent (BYOL) learns representations using an online network that predicts the embedding produced by a slowly evolving target network, with the target parameters updated via exponential moving average [52,53,54]. By eliminating dependence on large numbers of negatives, BYOL reduces the need for very large batches and can be more stable under clinical class imbalance and limited-label regimes.
DINO (Self-Distillation with No Labels) similarly adopts a teacher–student paradigm, training the student to match the teacher’s output distribution under different augmentations [55,56,57,58]. The objective is typically defined as a cross-entropy loss between teacher and student predictions:
L DINO = x X p teacher ( x ) log p student ( x ) .
DINO has shown strong performance in representation learning and is often used in medical imaging pipelines where interpretability-relevant attention maps and robust global features are desired.
SimSiam learns representations by predicting one augmented view from another while employing stop-gradient operations to prevent collapse [59,60,61,62]. In biomedical settings, negative-free objectives are especially useful for learning invariances in modalities such as histopathology, radiology, ECG, and EEG, where clinically meaningful similarity can occur across patients and where treating similar cases as negatives may degrade transfer performance. Overall, negative-free self-distillation provides a practical alternative to InfoNCE-based CL when negative sampling is unreliable or batch scaling is infeasible.

3.1.3. Clustering-Based Self-Supervised Learning (SwAV)

Clustering-based CL replaces explicit pairwise instance discrimination with a clustering objective. SwAV (Swapping Assignments between Views) performs online clustering and encourages consistent cluster assignments across augmentations [63]. This can reduce dependence on large numbers of negatives and may better preserve subtle pathology features by avoiding overly aggressive augmentation policies. Such methods are relevant in medical imaging, where representation stability and texture-level features are critical.

3.1.4. Multimodal Contrastive Alignment (CLIP, ConVIRT, BioViL)

CL has been extended beyond single-modality learning to align heterogeneous biomedical modalities into a shared embedding space. The most influential paradigm is CLIP-style vision–language pretraining, which trains an image encoder and a text encoder jointly using paired image–text data [64]. Given a minibatch of N paired samples { ( x i , t i ) } i = 1 N , encoders produce normalized embeddings z i I and z i T . Training maximizes similarity for matched pairs and minimizes similarity for mismatched pairs using a bidirectional retrieval objective (image-to-text and text-to-image), typically implemented as symmetric cross-entropy over in-batch negatives:
L CLIP = 1 2 1 N i = 1 N log exp ( z i I , z i T / τ ) j = 1 N exp ( z i I , z j T / τ ) 1 N i = 1 N log exp ( z i T , z i I / τ ) j = 1 N exp ( z i T , z j I / τ ) ,
where τ is a temperature parameter and · , · denotes cosine similarity. This formulation produces aligned multimodal representations that support retrieval and prompt-based zero-shot transfer: downstream classification can be performed by comparing an image embedding with text embeddings of label prompts (e.g., “no pleural effusion” vs “pleural effusion”), enabling label-free inference.
In medical AI, CLIP-style objectives have enabled major advances in radiology and pathology by leveraging radiology reports or biomedical captions as weak supervision. ConVIRT aligned chest X-rays with paired radiology reports via bidirectional CL, improving label efficiency and retrieval-based evaluation [18]. Subsequent work strengthened alignment through finer-grained supervision: GLoRIA introduced global–local alignment between image regions and report phrases to improve grounding and reduce spurious correlations [65]. BioViL further improved semantics by incorporating biomedical language pretraining, yielding stronger transferable representations and improved zero-shot performance [66]. Extensions, such as BioViL-T, incorporate temporal alignment across prior and current studies, improving progression-sensitive recognition [67]. Prompt-based adaptations, such as CXR-CLIP, integrate radiologist-defined class prompts to enhance clinical interpretability and reduce label ambiguity [68]. Zero-shot clinical deployment has also been explored through CLIP-based radiology models such as CheXzero [69].
A central limitation in clinical CLIP-style learning is that reports are noisy supervision: they contain negation (e.g., “no pneumothorax”), uncertainty (e.g., “cannot exclude”), templated phrases, and study-level context that may not map cleanly to image-level findings. Misalignment can create false negatives (a finding present in the image but unmentioned in the text) and shortcut learning driven by site, protocol, or demographics. Consequently, medical variants often incorporate filtering, entity extraction, uncertainty handling, or knowledge-aware objectives. For example, MedCLIP reduces dependency on strictly paired data by decoupling image and text corpora and using knowledge-enhanced supervision, improving robustness under limited or noisy pairings [70]. Beyond radiology, pathology vision–language models such as PLIP and CONCH scale CLIP-style pretraining to histopathology captions, enabling strong zero-shot transfer across datasets and tasks [71,72]. Large biomedical foundation models such as BiomedCLIP and PMC-CLIP further extend multimodal contrastive alignment by leveraging literature-scale figure–caption corpora to support broad biomedical retrieval, transfer, and few-shot adaptation [73,74].

3.1.5. Temporal and Patient-Aware Objectives (CPC, CLOCS/PCLR)

Medical data often exhibit sequential structure and repeated measurements, motivating contrastive objectives that exploit temporal and patient-identity information. Contrastive Predictive Coding (CPC) learns representations by predicting future latent embeddings in a sequence, using a contrastive objective to distinguish true future representations from negatives [75]. CPC is well-suited to physiological signals such as ECG and EEG, where long-range temporal dependence is clinically meaningful.
Patient-aware extensions define positives using identity or longitudinal structure. For example, Patient Contrastive Learning of Representations (PCLR) treats ECG recordings from the same patient as positives and recordings from different patients as negatives, improving generalization across downstream clinical prediction tasks [19]. Similarly, CLOCS introduces spatiotemporal contrastive structure by aligning signals across time and across leads, improving robustness under lead variations and temporal drift [76]. These objectives are particularly relevant in medicine, where patient-level consistency and longitudinal trajectories are central to clinical decision-making.

3.1.6. Optimization Refinements (Hard Negatives, Debiased Objectives)

Medical datasets frequently contain semantically similar cases, repeated examinations, and cohort effects, which makes naive negative sampling prone to false negatives. This issue is especially acute in clinical cohorts where different patients may share the same diagnosis or phenotype, and where repeated studies from the same hospital or scanner can introduce hidden correlations. Hard negative mining and debiased contrastive losses mitigate these issues by reweighting negatives, correcting sampling bias, or explicitly filtering likely false negatives [77,78].
In practice, hard negative mining prioritizes negatives that are close to the anchor in representation space (i.e., most confusing samples), which can sharpen decision boundaries but may also amplify errors if hard negatives are actually clinically similar positives. Debiased objectives address the mismatch between the InfoNCE assumption that all negatives are truly dissimilar and real-world medical data, where batch negatives may contain unobserved positives due to label scarcity, weak supervision, and reporting noise. These refinements improve robustness under clinical heterogeneity and reduce representation shortcuts driven by site, protocol, demographic confounders, or disease prevalence patterns.
A common refinement is to apply weights to negatives in the denominator of InfoNCE, emphasizing harder (more similar) negatives:
L HN = log exp ( sim ( z , z + ) / τ ) exp ( sim ( z , z + ) / τ ) + k = 1 K w k exp ( sim ( z , z k ) / τ )
where w k increases with similarity (e.g., w k exp ( sim ( z , z k ) / β ) ) so that negatives closer to the anchor contribute more strongly. In medical data, this strategy must be used cautiously because clinically similar cases may be incorrectly treated as negatives.
To reduce the impact of false negatives, debiased CL modifies the negative term by accounting for the probability that some negatives are actually positives. Following the debiased contrastive objective [77,79,80], the loss can be written as follows:
L Debiased = log exp ( sim ( z , z + ) / τ ) exp ( sim ( z , z + ) / τ ) + K · p ˜ neg
where p ˜ neg is a corrected negative expectation term that subtracts the estimated contribution of false negatives. This formulation is particularly relevant in medical AI because minibatches may contain semantically similar patients (shared condition, demographic similarity, repeated study types), even when labels are missing or incomplete [81,82,83].

3.2. Datasets and Benchmarks for Medical Contrastive Learning

A key barrier to fair comparison in medical CL is the diversity of datasets, label taxonomies, and evaluation protocols used across modalities. To improve reproducibility and enable more meaningful benchmarking, we summarize widely used public datasets and community benchmarks for contrastive and self-supervised representation learning in medical AI, spanning imaging, EHRs, physiological signals, and omics. We also highlight common evaluation protocols (linear probing, fine-tuning, few-shot transfer, external validation) and task-appropriate metrics that recur across the literature. Table 3 presents representative public datasets and community benchmarks commonly used in medical CL, grouped by modality, together with typical downstream tasks used for evaluation.
Chest radiography is the most common benchmark for medical CL because it provides large-scale paired image–report corpora and standardized multi-label tasks. Prominent resources include MIMIC-CXR (images with reports) [84] and CheXpert [85], alongside earlier large-scale X-ray datasets such as NIH ChestX-ray14 [86]. Beyond X-rays, neuroimaging benchmarks such as ADNI support Alzheimer’s disease research and multimodal clinical studies [87]. For segmentation and structured imaging tasks, community challenges such as BraTS (brain tumor MRI segmentation) provide standardized datasets, splits, and metrics [88,89]. Dermatology imaging is frequently benchmarked via the ISIC Archive and its associated tasks [90]. In computational pathology, whole-slide image benchmarks and paired resources have grown rapidly, including CAMELYON-style lymph node metastasis benchmarks and large-scale cancer slide repositories derived from TCGA and related infrastructures [91,92].
EHR-based CL often builds patient representations from longitudinal visits and mixed structured/unstructured data. The most widely used public critical-care benchmarks are MIMIC-III [93] and MIMIC-IV [94], with complementary ICU cohorts such as the eICU Collaborative Research Database [95]. These datasets support mortality prediction, length-of-stay prediction, readmission risk, phenotyping, and treatment trajectory modeling, typically under strong temporal and site-specific confounding.
Self-supervised and CL on physiological signals are commonly evaluated on ECG and EEG corpora with large unlabeled volumes and clinically meaningful downstream tasks. PTB-XL is a standard ECG benchmark supporting multi-label ECG classification and transfer learning [96]. EEG benchmarks often use large clinical corpora, such as the TUH EEG dataset for seizure detection and broader EEG event modeling [97].
Omics benchmarks include bulk genomics and transcriptomics resources (e.g., TCGA [98]) and tissue-specific expression atlases (e.g., GTEx [99]). Large biobanks, such as UK Biobank, further enable genotype–phenotype analysis for representation learning and risk modeling [100]. In parallel, single-cell transcriptomics has become a prominent benchmark for CL because of its sparsity and strong batch effects, supported by atlas-scale efforts and curated reference datasets [101].
Table 3. Representative public datasets and benchmarks commonly used in medical contrastive learning, grouped by modality.
Table 3. Representative public datasets and benchmarks commonly used in medical contrastive learning, grouped by modality.
ModalityRepresentative Datasets/BenchmarksTypical Downstream Tasks
Chest imagingMIMIC-CXR [84]; CheXpert [85]; NIH ChestX-ray14 [86]Multi-label classification (AUROC); label-scarce transfer; external validation; retrieval and zero-shot classification
General medical imagingBraTS [88,89]; ISIC Archive [90]; ADNI [87]Segmentation (Dice/HD95); lesion classification; progression prediction; robustness across scanners and time
Computational pathologyCAMELYON [91,92]; TCGA-derived slide repositories [98]WSI classification; region retrieval; weakly supervised detection; cross-site generalization
EHR (clinical records)MIMIC-III [93]; MIMIC-IV [94]; eICU [95]Mortality and LOS prediction; readmission; phenotyping; temporal outcome modeling; multimodal fusion with notes
Physiological signalsPTB-XL [96]; TUH EEG [97]Diagnosis classification; seizure detection; patient-level transfer; low-label training
Genomics and transcriptomicsTCGA [98]; GTEx [99]; UK Biobank [100]Subtype prediction; survival and risk modeling; representation transfer across cohorts; cross-tissue generalization
Single-cell omicsHuman Cell Atlas [101]Cell-type clustering; batch correction; rare cell detection; perturbation response prediction
Across modalities, medical CL is typically evaluated under one or more of the following regimes: (i) linear probing, where the encoder is frozen, and a lightweight classifier is trained to test representation quality under minimal supervision; (ii) full fine-tuning, where the pretrained encoder is adapted end-to-end to measure downstream task performance under realistic clinical training conditions; (iii) few-shot or label-scarce transfer, which evaluates label efficiency by measuring performance as a function of annotated data fraction, such as 1%, 10%, and 100%; and (iv) external validation, where models are evaluated across institutions, scanners, devices, acquisition protocols, or time periods to quantify robustness to domain shift and distribution drift.
Evaluation metrics are task dependent [102,103,104]. For multi-label classification in imaging, particularly chest radiography, performance is commonly reported using the area under the receiver operating characteristic curve (AUROC), and the area under the precision–recall curve (AUPRC), often complemented by macro-F1 or clinically meaningful sensitivity and specificity at fixed thresholds. In segmentation tasks, overlap and boundary-based measures, such as the Dice coefficient and the 95th percentile Hausdorff distance (HD95), are widely used. For prediction tasks on electronic health records and physiological signals, patient-level discrimination metrics such as AUROC and AUPRC are frequently reported, and several studies additionally include calibration metrics such as the Brier score or expected calibration error to assess whether probabilistic outputs are clinically reliable. For vision–language models, benchmarking extends beyond classification to evaluate multimodal alignment and grounding. Common benchmarks include cross-modal retrieval using Recall@K and median rank, phrase grounding scores, and zero-shot classification using prompt embeddings that map clinical labels into the text representation space. Importantly, medical CL studies increasingly distinguish between representation evaluation protocols such as linear probing and deployment-relevant protocols such as full fine-tuning and external validation, since gains under frozen-feature testing do not necessarily translate to robustness, calibration, or clinical reliability in real-world workflows. Table 4 presents common benchmarking protocols and evaluation metrics used to assess medical CL models across modalities, including linear probing, full fine-tuning, few-shot transfer, and external validation under distribution shift.

4. Applications of Contrastive Learning in Medical AI

CL has shown significant promise in the field of medical AI due to its ability to learn effective representations from limited or unlabeled data. This capability is particularly valuable in medical contexts, where labeled data can be scarce or expensive to obtain. Figure 3 presents several applications of CL in healthcare. The following subsections explore the applications of CL across various domains within medical AI. A full summary of the included studies is presented in Appendix C.

4.1. Medical Imaging

Medical imaging is one of the most prominent areas where CL has been applied successfully. Clinical imaging pipelines commonly target disease classification, lesion detection, organ or tumor segmentation, retrieval, and anomaly detection. However, robust model development is often limited by the scarcity of high-quality expert annotations, inter-observer variability, and substantial heterogeneity across scanners, acquisition protocols, and institutions. CL addresses these challenges by leveraging large-scale unlabeled imaging repositories to learn transferable representations, improving label efficiency, and supporting more reliable generalization across clinical settings.
A common strategy is to pretrain encoders on unlabeled images using contrastive objectives that enforce invariance across augmentations while preserving clinically meaningful structure. For example, Azizi et al. [16] applied CL to medical imaging tasks such as classification and segmentation by learning representations from unlabeled medical images. They utilized SimCLR to pretrain a model on a large dataset of unlabeled chest X-rays and then fine-tuned the encoder on smaller labeled datasets for downstream tasks. Their results demonstrated consistent performance gains compared to training from scratch, highlighting the value of contrastive pretraining in low-label medical imaging scenarios.
Beyond fully self-supervised pretraining, CL has also been adopted for semi-supervised imaging pipelines. Chaitanya et al. [105] developed a CL approach for semi-supervised learning in medical imaging, with the objective of improving robustness under limited labels and anatomical variability. Their method learns consistent representations across augmented views of the same image as positive pairs while contrasting against other images as negatives, leading to improved performance in MRI classification and segmentation tasks.
Histopathology has emerged as another high-impact imaging domain for CL due to the scale and complexity of whole-slide images and the cost of expert annotations. Ciga et al. [106] explored self-supervised CL for histopathological image analysis and showed that the learned representations could distinguish tissue patterns and malignancy-related morphology effectively. Their findings suggest that contrastive objectives can capture subtle texture and micro-structural cues important for cancer detection and grading, even when labeled samples are limited.
CL has also been tailored to dense prediction tasks such as segmentation, where structural consistency and multi-scale context are essential. Guo et al. [107] proposed a CL framework for cardiac MRI segmentation that incorporates a multi-scale contrastive loss to learn representations at different spatial resolutions. This design supports learning both global anatomical structures and local pathological variations. Their results demonstrated improved segmentation accuracy for cardiac structures such as the myocardium and ventricles, illustrating the usefulness of contrastive objectives for robust segmentation in heterogeneous imaging data.
In addition to supervised and semi-supervised pipelines, CL has been applied to unsupervised anomaly detection, where abnormal findings may be rare, diverse, and costly to label. Luo et al. [108] developed a self-supervised contrastive framework for anomaly detection in brain MRI. Their approach contrasts normal and abnormal patches and uses augmented views of the same patch as positives, allowing the model to learn representations sensitive to subtle pathological changes. The study reported improved anomaly detection performance over prior baselines, suggesting that CL can support early detection of neurological abnormalities in settings where annotated anomalies are scarce.

4.2. Electronic Health Records

EHRs represent a major application area for CL in medical AI because they capture longitudinal patient trajectories at scale. EHR data are inherently heterogeneous, combining structured variables (e.g., laboratory values, vital signs, diagnosis codes, medications) with unstructured clinical narratives. Compared to imaging, EHR modeling is additionally challenged by irregular sampling, missingness, temporal drift, high dimensionality, and institutional variation in coding and documentation practices. These characteristics often limit the portability of supervised models and increase reliance on large labeled cohorts that may not generalize across healthcare systems. CL offers a promising alternative by enabling representation learning from unlabeled EHR sequences, producing patient embeddings that can transfer effectively across predictive tasks and label-scarce clinical settings.
A common EHR contrastive paradigm constructs multiple views of the same patient record through stochastic perturbations, temporal cropping, modality masking, or aggregation windows, treating these as positive pairs while contrasting against other patients as negatives. Krishnan et al. [109] applied CL to EHR data through a self-supervised framework that generates augmented views of patient histories. By treating augmented versions of the same patient’s record as positives and records from different patients as negatives, the model learns patient representations that preserve temporal and clinical structure. Their experiments showed improved performance compared to standard supervised baselines across multiple tasks, including mortality prediction and heart failure diagnosis, highlighting the role of contrastive pretraining in improving clinical risk stratification under limited labeled data.
Beyond general-purpose representation learning, contrastive learning has also been applied to construct task-ready patient embeddings that support common operational outcomes in healthcare systems. Pick et al. [110] developed a contrastive framework for learning patient-level representations for hospital mortality and length-of-stay prediction. Kerdabadi et al. [111] proposed an ontology-aware temporal contrastive survival framework that learns patient embeddings using temporally distinctive patterns and hardness-aware negatives, demonstrating improved acute kidney injury survival risk prediction. Liu et al. [112] addressed irregular sampling and missingness through a contrastive imputation–prediction network, in which contrastive objectives guide representation learning during data reconstruction, leading to improved in-hospital mortality prediction. Zang and Wang [113] adopted a supervised contrastive framework for longitudinal EHR classification, using label-informed positives to tighten outcome-specific clusters. By explicitly modeling similarity and dissimilarity between patient trajectories, their approach improved downstream risk prediction performance, suggesting that contrastive objectives can better capture latent clinical states compared to purely supervised training on sparse labels.
EHR data are increasingly multimodal, motivating contrastive objectives that align complementary patient information sources rather than learning from each modality independently. Sun et al. [114] developed a CL framework for multimodal EHR integration, aligning structured signals such as laboratory results and vital signs with representations derived from unstructured clinical notes. Positive pairs were constructed by matching structured and unstructured views from the same patient, while negatives were formed using cross-patient mismatches. This alignment improved the quality of patient representations and increased performance in tasks including disease progression prediction and complication risk identification. These results support the growing view that CL is particularly well-suited for EHR settings where clinically meaningful information is distributed across heterogeneous modalities.
Finally, a major barrier to real-world EHR modeling is scale. Health systems produce massive longitudinal datasets that require scalable training strategies and efficient distributed computation. Cai et al. [115] addressed this challenge by proposing a distributed CL framework designed for large-scale EHR repositories. Their work highlights that CL can be extended to big-data healthcare settings and suggests practical pathways for training robust patient representation models across diverse populations.

4.3. Genomics and Proteomics

In genomics and proteomics, CL offers substantial advantages for analyzing high-dimensional biomedical data, where labeled outcomes are often limited and experimental noise can be substantial. These domains generate massive quantities of molecular measurements, including DNA variation, gene expression profiles, epigenetic regulation signals, and protein sequences and structures. However, learning clinically meaningful representations remains challenging due to extreme feature dimensionality, sparsity (especially in single-cell assays), batch effects, confounding biological and technical variation, and the need to integrate signals across heterogeneous omics layers. CL is particularly well-suited to this setting because it can learn robust molecular representations by exploiting intrinsic structure, such as similarity across biological replicates, related pathways, shared molecular functions, or paired measurements across modalities, without requiring extensive annotation.
Within genomics, Zhong et al. [116] applied CL to identify disease-associated genetic signals using a multi-scale contrastive learning (MSCL) framework. MSCL was designed to capture genetic interactions across multiple levels of granularity, from local gene patterns to broader pathway-level relationships. By defining positive pairs using sequences from the same genomic regions and negatives from distinct regions, the model learns representations that differentiate healthy from diseased samples and enhances the detection of genetic markers. This multi-scale formulation reflects a key advantage of CL for genomics: contrastive objectives can encode biologically meaningful similarity under complex, non-linear genetic interactions that are difficult to capture with standard supervised pipelines.
A second major direction is multi-omics integration, where the objective is to learn unified molecular patient representations across complementary assays. Liu et al. [117] advanced CL for genomics by developing a framework (MoHeG/GenCL) that aligns heterogeneous omics layers such as genomics, transcriptomics, and epigenomics. Contrastive alignment across modalities enables the model to learn consistent patient-level embeddings that capture regulatory interactions between genes and downstream functional effects. Such unified representations improve interpretability and predictive power for disease susceptibility and precision medicine tasks, particularly in multifactorial diseases where interactions across molecular layers are essential.
Single-cell genomics further amplifies the relevance of CL, since scRNA-seq data are highly sparse, noisy, and sensitive to batch effects, yet rich in unlabeled biological structure. Li et al. [118] proposed a CL framework for scRNA-seq representation learning that constructs positive pairs between cells with similar expression patterns while contrasting against dissimilar cells. This approach improves clustering of cell types and supports the discovery of rare populations, which is critical for understanding tumor microenvironments, immune heterogeneity, and neurodegenerative processes. More broadly, CL offers a natural mechanism for learning invariances to technical noise while retaining biologically meaningful discriminative structure.
In proteomics, CL has been applied to learn protein representations that reflect functional and structural similarity. Bepler and Berger [119] used CL to derive protein sequence representations by forming positive pairs using different conformations or states of the same protein and negatives from unrelated proteins. This formulation improves representation quality for tasks including protein function prediction and interaction modeling, supporting downstream applications in drug discovery and protein engineering.
At the protein interaction level, Zhang et al. [120] developed Pepharmony, a CL approach for predicting protein–protein interactions (PPIs) by integrating both sequence and structural information. Positive pairs were constructed from interacting protein conformations, while negatives corresponded to non-interacting proteins. The resulting embeddings improved PPI prediction accuracy, highlighting that contrastive objectives can capture complex molecular compatibility signals that are essential for understanding disease mechanisms and discovering therapeutic targets.

4.4. Multimodal and Cross-Domain Learning

Integrating heterogeneous medical modalities, such as imaging, clinical text, and structured patient data, is a major direction in medical AI, because clinically meaningful decision-making often requires joint reasoning across multiple information sources. CL provides a principled mechanism for multimodal fusion by aligning representations from different modalities into a shared embedding space, enabling label-efficient learning, retrieval, and zero-shot transfer.
Early medical vision–language contrastive models focused on radiology report alignment. Zhang et al. [18] introduced ConVIRT, which learns chest X-ray representations by aligning images with paired radiology reports using a bidirectional contrastive objective. ConVIRT demonstrated substantial label efficiency, requiring only 10% of labeled data relative to an ImageNet-initialized baseline to achieve similar or better performance across four downstream tasks. Building on this paradigm, Huang et al. [65] proposed GLoRIA, which strengthens supervision through both global and local alignment between image regions and report phrases. On MIMIC-CXR, GLoRIA achieved a precision@5 of 69.24% for image-to-text retrieval compared to 66.98% for ConVIRT, and attained CheXpert AUROC scores of 0.926, 0.943, and 0.950 when fine-tuned with 1%, 10%, and 100% labeled data, respectively. Boecking et al. [66] further improved semantic alignment with BioViL by incorporating domain-specific biomedical language pretraining. BioViL achieved zero-shot accuracy, F1, and AUROC of 0.732, 0.665, and 0.831, respectively, and reached a linear-probe AUROC of up to 0.891 on RSNA pneumonia classification, establishing a strong benchmark for biomedical vision–language representation learning.
More recent work has emphasized the importance of temporality and structured clinical priors. Bannur et al. [67] proposed BioViL-T, extending vision–language pretraining with temporal alignment across prior and current chest X-rays. BioViL-T improved progression classification, phrase grounding, and report generation, highlighting the clinical relevance of longitudinal contrastive structure. Similarly, You et al. [68] developed CXR-CLIP, integrating radiologist-defined prompts with both image–label and image–text supervision to improve clinical interpretability and downstream robustness. Collectively, these works suggest that incorporating temporal cues, richer language semantics, and supervised priors improves the clinical validity of contrastive multimodal representations.
Contrastive pretraining has also enabled clinically meaningful zero-shot transfer through large-scale vision–language alignment. Tiu et al. [69] introduced CheXzero, a CLIP-based model trained on unannotated chest X-rays and reports. In a reader study, CheXzero achieved multi-label classification performance statistically indistinguishable from board-certified radiologists on CheXpert, with no significant differences in Matthews correlation coefficient across five evaluated pathologies. To improve robustness under limited or noisy pairings, Wang et al. [70] proposed MedCLIP, which decouples image and text corpora and uses a knowledge-aware matching loss to mitigate false negatives. Using only 20,000 pretraining pairs, MedCLIP achieved a zero-shot accuracy of 44.8%, surpassing GLoRIA (43.3% with 191,000 pairs) and ConVIRT (42.2% with 369,000 pairs) under identical evaluation settings.
Beyond radiology, multimodal CL has advanced computational pathology and strengthened cross-domain generalization. Huang et al. [71] proposed PLIP, a pathology vision–language foundation model trained on OpenPath image–caption pairs. PLIP achieved zero-shot F1 scores between 0.565 and 0.832 across four external datasets, outperforming prior vision–language models that achieved F1 between 0.030 and 0.481. Lu et al. [72] introduced CONCH, trained on over 1.17 million histopathology image–caption pairs, demonstrating state-of-the-art performance across classification, retrieval, captioning, and segmentation tasks. These findings indicate that scaling multimodal contrastive pretraining improves transferability in histopathology, where domain shift across scanners, staining, and cohorts is a pervasive challenge.
Scaling has also expanded to literature-based biomedical corpora, enabling more generalizable biomedical foundation models. Zhang et al. [73] developed BiomedCLIP, pretrained on 15 million image–text pairs from PubMed Central. BiomedCLIP achieved 56% and 77% top-1 and top-5 retrieval accuracy on a 725,000-pair held-out set and demonstrated strong zero- and few-shot performance across radiology and pathology benchmarks, often surpassing domain-specific models such as ConVIRT and GLoRIA. Similarly, Lin et al. [74] proposed PMC-CLIP, pretrained on 1.6 million biomedical figure–caption pairs, improving medical visual question answering and retrieval performance, and highlighting the value of literature-derived multimodal alignment when clinical pairings are scarce.
Recent innovations have moved beyond representation learning to enable fine-grained localization and segmentation. Huang et al. [121] introduced MaCo, which applies masked CL with correlation weighting to chest X-rays and improves both zero-shot and supervised recognition of localized findings. Koleilat et al. [122] combined contrastive vision–language models with the Segment Anything Model to enable text-driven segmentation across ultrasound, MRI, and CT datasets, achieving strong performance without explicit segmentation annotations. Overall, these developments demonstrate the growing versatility of multimodal and cross-domain CL, enabling efficient, interpretable, and transferable medical AI systems that bridge visual and textual clinical evidence.

4.5. Time-Series and Physiological Signal Analysis

Medical time-series data, such as ECG, EEG, respiratory signals, and vital signs, present major opportunities for CL because large volumes of unlabeled recordings are routinely collected in clinical practice. However, these signals are also characterized by strong temporal dependencies, noise, missingness, and substantial inter-patient variability. A systematic review by Liu et al. [123] covering 43 studies on self-supervised CL for medical time series reported that most approaches rely on standard augmentations (e.g., scaling, jittering, cropping) and encoder architectures such as 1D CNNs or Transformers. The review further emphasized the need for hierarchical and patient-aware contrastive objectives to better capture long-range dependencies and clinically meaningful temporal consistency.
Diamant et al. [19] introduced Patient Contrastive Learning of Representations (PCLR), which defines positive pairs as ECG recordings from the same patient and negatives as recordings from different patients. Using a dataset of more than 3.2 million 12-lead ECGs, their results showed that linear models trained on PCLR representations achieved an average 51% improvement across downstream tasks, including sex classification, age regression, left ventricular hypertrophy, and atrial fibrillation detection, compared to models trained from scratch. Relative to alternative pretraining strategies, PCLR achieved a 47% average gain on three of four tasks and yielded a 9% improvement over the strongest baseline per task.
Yuan et al. [124] proposed poly-window CL, which samples multiple overlapping temporal windows from each ECG as positive pairs rather than relying on only two augmented views. On the PTB-XL dataset, this approach achieved AUROC 0.891 compared to 0.888 for conventional two-view CL, and an F1 score of 0.680 versus 0.679, while reducing pretraining time by 14.8%. These findings suggest that explicitly modeling intra-record temporal relationships can improve both efficiency and representation quality.
Wang et al. [125] developed COMET, a hierarchical CL framework that organizes data at multiple levels, including observation, sample, trial, and patient, and applies contrastive objectives across these granularities. COMET demonstrated improvements over six baselines across ECG and EEG datasets targeting myocardial infarction, Alzheimer’s disease, and Parkinson’s disease tasks, particularly in low-label settings (10% and 1%). Chen et al. [76] introduced CLOCS (Contrastive Learning of Cardiac Signals across Space, Time, and Patients), which aligns temporal segments and ECG leads to improve robustness under lead variation and temporal drift. These hierarchical and spatiotemporal approaches extend CL beyond simple view augmentation toward multi-level temporal consistency.
Raghu et al. [126] explored multimodal extensions by pretraining contrastive models on physiological time series combined with structured clinical variables such as laboratory values and vital signs. Their results indicated consistent downstream gains compared to baseline pretraining methods, highlighting the value of multimodal temporal alignment for capturing richer clinical context. Guo et al. [127] proposed a Multi-Scale and multimodal contrastive learning network (MBSL) for biomedical signals, leveraging cross-modal contrastive objectives between modalities such as respiration, heart rate, and motion sensors. MBSL reduced mean absolute error by 33.9% for respiration rate prediction, by 13.8% for exercise heart rate estimation, and improved activity recognition accuracy and F1 scores by 1.41% and 1.14%, respectively, compared to state-of-the-art baselines.
To address false negatives arising from batch sampling in large clinical cohorts, Sun et al. [128] proposed a Patient Memory Queue (PMQ) mechanism that maintains a memory bank of intra-patient and inter-patient samples during contrastive pretraining. Across three public ECG datasets and varying label ratios, PMQ outperformed existing contrastive methods in both classification accuracy and robustness to label scarcity. These patient-aware memory designs reflect a broader trend in physiological CL toward objectives that explicitly encode patient identity and long-term temporal consistency.
Overall, CL for physiological time series has progressed from basic augmentation-driven pipelines toward more structured paradigms, including multi-window sampling, hierarchical CL, cross-modal alignment, and patient-aware memory mechanisms. These innovations enable models to better capture temporal dynamics and inter-patient invariances inherent in clinical signals, improving robustness and label efficiency in data-limited healthcare settings. Table 5 summarizes key applications of contrastive learning across different medical domains.

5. Challenges, Limitations, and Practical Considerations

Despite strong empirical results, translating contrastive learning into clinically reliable and reproducible medical AI remains challenging. These challenges arise at multiple levels, including pair construction, augmentation design, evaluation methodology, multimodal supervision quality, fairness and privacy risks, optimization stability, and deployment constraints.

5.1. Pair Construction and Clinical Semantics

Most contrastive objectives rely on constructing positive and negative pairs, yet defining such pairs in clinical settings is nontrivial. In standard instance discrimination, samples from different patients are treated as negatives by default. However, medical datasets frequently contain semantically similar cases across patients (e.g., shared phenotypes, repeated staging patterns, common radiographic presentations), which can produce false negatives. False negatives attenuate disease-relevant signal by pushing clinically similar cases apart in embedding space and may bias learned representations toward spurious correlates. In addition, patient identity, acquisition protocol, scanner vendor, and site effects may leak into representations if sampling is not controlled. When paired views are constructed within a narrow subset of acquisition conditions, the contrastive objective may prioritize hospital-specific or device-specific features over pathology features. This shortcut learning can yield high in-domain performance while degrading cross-site generalization.
Clinical datasets also exhibit longitudinal structure, with repeated measurements over time. If not accounted for, longitudinal leakage can inflate downstream evaluation: pretraining and fine-tuning may implicitly learn patient-specific signatures rather than clinically generalizable features. Even when patient-level splits are used, subtle overlaps can persist when repeated exams, segments, or derived patches are not tracked carefully. These issues motivate representation-level interpretability and alignment audits to ensure that embedding similarity reflects clinical semantics.

5.2. Augmentation and View Design in Medical Data

A key assumption in contrastive learning is that two augmented views of the same instance preserve semantic identity. However, augmentation policies calibrated for natural images do not always transfer to medical data. Geometric transformations, cropping, and intensity perturbations can erase subtle findings, distort anatomical context, or remove small lesions. For example, aggressive cropping can eliminate peripheral abnormalities, while contrast jittering can suppress radiological cues tied to tissue density.
In medical imaging, defining clinically valid invariances requires domain expertise. Certain transformations may be valid for some tasks (e.g., rotation for dermoscopy) but invalid for others (e.g., left-right flips in chest X-rays). Similarly, time-series augmentations such as jittering or permutation can disrupt clinically meaningful rhythms in ECG or EEG. Without medically validated augmentations and consistent reporting of augmentation policy, contrastive pretraining may learn invariances that suppress pathologies or amplify confounders.

5.3. Heterogeneity Across Modalities and Tasks

A central challenge for medical CL is heterogeneity across modalities. Imaging, EHR, omics, clinical notes, and physiological signals have fundamentally different noise processes, feature semantics, and temporal structure. A contrastive formulation that works well for radiology may fail in EHR due to sparsity, irregular sampling, and missingness patterns that encode care pathways rather than disease. In genomics and proteomics, the definition of positives and negatives often depends on biological priors, and naive sampling can embed batch effects rather than functional similarity.
Furthermore, clinical tasks span diagnosis, prognosis, retrieval, segmentation, progression monitoring, and phenotyping. Contrastive objectives optimized for global representations may underperform in localization-sensitive tasks (e.g., lesion detection or segmentation), where patch selection and spatial correspondence matter. This mismatch between representation objective and clinical task introduces uncertainty about which CL choices yield clinically meaningful embeddings.

5.4. Reproducibility and Reporting Gaps

Reproducibility is constrained by restricted data access, limited release of pretraining corpora, and incomplete reporting. Many papers under-specify key components such as augmentation policy, pairing strategy, preprocessing, tokenizer, and text normalization steps, or hyperparameters (batch size, temperature, queue length). Even when code is released, private clinical datasets and institutional pipelines limit replicability.
CL outcomes are highly sensitive to training details, and small differences in preprocessing can lead to nontrivial changes in downstream performance. This sensitivity makes it difficult to assess whether improvements arise from methodological novelty or from differences in training scale and implementation.

5.5. Multimodal Alignment and Weak Supervision Risks

Multimodal contrastive alignment introduces additional concerns. Free-text reports include negations, hedging, and section-specific context that can misalign with image-level findings. A pathology may be present but unmentioned in the report, producing implicit false negatives. Conversely, templated phrases or clinical history can produce matches unrelated to the actual imaging evidence.
Weak supervision mined from reports may propagate label noise if entity linking, uncertainty handling, and negation detection are not explicit. These issues can amplify bias and reduce trustworthiness in downstream clinical interpretation. Additionally, multimodal pretraining can inadvertently learn shortcuts via hospital-specific language patterns, scanner metadata, or demographic correlates embedded in reporting style.

5.6. Interpretability and Explainability

Interpretability is central to clinical validity, yet it remains underdeveloped in medical CL. Unlike supervised models, where explanations can be tied directly to task labels, CL pretraining optimizes representation geometry via similarity objectives (e.g., InfoNCE), which can inadvertently encode a mixture of clinically meaningful factors (e.g., anatomy and pathology) and nuisance variables (e.g., site, device, acquisition protocol, reporting style). This is particularly problematic in multi-centre settings, where site effects may dominate embedding structure even when downstream metrics appear strong. Consequently, explainability for CL should emphasize representation-level interpretability, understanding what factors structure the embedding space, rather than only post-hoc explanation of downstream predictions. A practical CL-specific approach is to perform alignment audits to verify whether embedding similarity corresponds to clinical semantics. For example, nearest-neighbor retrieval in embedding space can reveal whether clinically similar cases are clustered together or whether representations separate primarily by hospital/site or scanner vendor. Likewise, embedding visualizations (e.g., UMAP/t-SNE) overlaid with metadata such as site, device, or demographic variables can expose confounding and shortcut learning. Since CL aims to learn invariances defined by augmentations, interpretability can also be operationalized as invariance auditing: representations should remain stable under clinically irrelevant perturbations (e.g., mild intensity changes) while remaining sensitive to clinically meaningful changes.
Caution is warranted when using saliency methods (e.g., Grad-CAM) on downstream classifiers trained atop CL encoders. Such explanations may not reflect what the contrastively learned representation encodes, and may amplify spurious shortcuts (e.g., text markers, laterality cues, portable vs stationary scanner artifacts). We therefore recommend pairing saliency with representation-level audits (retrieval inspection, metadata overlays, invariance tests) and documenting interpretability failure cases, as a minimum standard for clinical transparency.
At minimum, medical CL studies should report: (i) embedding-space retrieval examples with clinician review; (ii) representation visualizations with site/device overlays; (iii) invariance sensitivity tests aligned with clinical plausibility; and (iv) interpretability failure cases highlighting confounding or shortcut reliance.

5.7. Fairness, Bias, and Clinical Validity

Fairness concerns remain under-addressed in medical CL. Self-supervised objectives do not eliminate bias. Instead, they may learn and amplify latent dataset biases. If the pretraining distribution is skewed, for example, through over-representation of certain populations or hospital systems, embeddings may transfer poorly to underrepresented groups. Bias may manifest through subgroup performance gaps, miscalibration, or uneven representation quality.
Clinical validity also requires interpretability and subgroup evaluation. Yet many studies rely primarily on global metrics and do not assess whether learned representations preserve clinically relevant features consistently across demographic groups, comorbidity profiles, or rare disease subpopulations.

6. Discussion and Future Research Directions

This review shows that CL has evolved from a general-purpose self-supervised paradigm into a versatile representation learning toolkit for medical AI. Across imaging, electronic health records, physiological signals, omics, and multimodal settings, contrastive objectives have consistently demonstrated improved label efficiency and transferable feature learning, especially when pretrained on large-scale unlabeled corpora. Nevertheless, the current evidence base remains uneven across modalities and clinical tasks. Imaging and vision–language learning have benefited from large datasets and standardized benchmarks, whereas time-series, EHR, and omics studies are still characterized by heterogeneous experimental designs, inconsistent evaluation protocols, and limited external validation. As a result, many reported gains remain difficult to compare across studies, and the extent to which improvements persist under realistic clinical deployment conditions is often unclear.
A unifying insight across modalities is that the construction of positive and negative pairs fundamentally determines clinical meaning, downstream robustness, and the risk of shortcut learning. Instance-level positives generated through stochastic augmentations can work well in imaging, but medical settings frequently benefit from pairing strategies grounded in clinical structure, such as patient-level positives derived from repeated exams, longitudinal studies, adjacent slices, multi-view imaging, or multiple recordings from the same patient [76,129]. Conversely, standard negative sampling assumptions often break down in clinical cohorts, where semantically similar patients can appear in the same batch and inadvertently become false negatives. This issue can attenuate disease-relevant signal, degrade calibration, and bias representations toward acquisition artifacts and cohort effects. These risks are amplified in multi-center settings where site-specific protocols, scanner differences, demographic variation, and institutional workflows introduce confounders that may be spuriously predictive, thereby undermining generalization [129,130].
Another consistent finding is that scale improves performance but does not guarantee reliability. Large pretraining corpora may increase representation quality, yet they can simultaneously amplify dataset biases and confounding patterns unless sampling and evaluation explicitly account for distribution shift. In particular, multimodal contrastive alignment has enabled major advances in radiology and pathology by leveraging paired image–report data and CLIP-style retrieval objectives, enabling strong zero-shot and label-free inference [18,66,69]. However, multimodal supervision introduces new failure modes. Radiology reports frequently contain negation, uncertainty, templated phrasing, and contextual descriptions that do not map cleanly to image-level labels, making report–image supervision inherently noisy. Without explicit handling of uncertainty and structured clinical semantics, vision–language alignment can produce misleading associations, propagate label noise, or encourage superficial lexical matching rather than clinically grounded representations.
From a practical clinical AI systems perspective, these findings imply that CL should be treated not merely as a performance optimization tool but as an upstream design decision that shapes downstream validity, safety, and reproducibility. Method choices must be aligned with deployment realities: pairing strategies should reflect clinical semantics rather than convenience; evaluation protocols should match intended clinical use; and robustness under domain shift should be treated as a minimum requirement rather than an optional extension. In addition, clinical adoption depends on more than accuracy. Reporting should increasingly incorporate calibration, uncertainty, and subgroup performance, since miscalibration or uneven error rates across groups can directly translate into inequitable or unsafe decision support [131,132]. Interpretability should also be treated as a core evaluation dimension rather than a future add-on. Even when contrastive pretraining improves end-task scores, the clinical trustworthiness of representations remains limited if models are opaque, brittle under shift, or reliant on shortcuts. Therefore, explanation methods and grounding analyses should be included in model validation pipelines where applicable [133,134,135].
Future research should prioritize directions that maximize clinical impact while closing the methodological gaps identified across the literature. A near-term priority is to establish standardized reporting and benchmarking practices that enable fair comparison across studies, including clear documentation of patient-level split rules, pretraining corpus composition, pairing heuristics, augmentation policies, and key hyperparameters. Similarly, external validation should become routine, with experiments that explicitly examine multi-site generalization, temporal drift, and device/protocol variation [129,136]. Augmentation strategies should be clinically validated, since transformations suitable for natural images may suppress subtle pathology signals or distort medically relevant textures and boundaries. In parallel, the field would benefit from more operational guidance that maps typical medical conditions and data regimes to appropriate contrastive objective families, supporting decision-making such as when negative-free objectives may be preferable, when patient-aware pairing should be emphasized, and when supervised contrastive objectives offer more stable learning in heavily imbalanced multi-label settings.
Beyond near-term standardization, methodological progress should increasingly target known clinical failure modes. Promising directions include debiased objectives, calibrated hard-negative strategies, and supervision schemes that integrate patient groups or clinically meaningful priors. Reducing shortcut learning will likely require approaches that explicitly address confounding, including strategies inspired by invariant learning or causal representation learning [137,138]. For multimodal systems, robust alignment will depend on clinical language understanding that can model negation, uncertainty, and structured report content; entity-level alignment and section-aware objectives may reduce label noise and improve faithfulness. Another important long-term direction concerns privacy-preserving learning, particularly because medical pretraining often requires sensitive data at scale. Federated CL, secure aggregation, and differential privacy mechanisms may provide pathways toward multi-institution representation learning while limiting the risks of memorization and re-identification [139,140]. Finally, the clinical value of contrastive pretraining should be validated not only through retrospective benchmarks but also through prospective and workflow-based studies, including reader-in-the-loop evaluation that measures real endpoints such as reduction in annotation burden, improved diagnostic consistency, or improved time-to-decision [141].
Despite its promise, CL also faces limitations that may not be fully resolvable in the short term. Pair construction and negative sampling remain structurally challenging in medicine because clinical similarity is often latent, labels can be incomplete, and cohort overlap can be high. Domain shift is pervasive and can emerge through unmeasured confounders, evolving documentation practices, changes in acquisition protocols, and institutional workflow differences. While CL can reduce sensitivity to shift, it does not eliminate the need for explicit external validation and careful deployment monitoring. In addition, compute and infrastructure requirements continue to constrain accessibility: large-scale pretraining, comprehensive ablations, and multi-center validation are often feasible only for well-resourced institutions, which can bias the research landscape and limit independent reproducibility. Overall, CL is positioned as a foundational paradigm for medical representation learning and multimodal alignment, but its clinical adoption will depend on moving beyond in-domain accuracy toward robust, reproducible, externally validated, and clinically interpretable evidence under realistic deployment conditions.

7. Conclusions

CL has emerged as one of the most influential self-supervised paradigms in medical AI, enabling representation learning from large-scale unlabeled data in settings where annotation is costly, scarce, and heterogeneous. This review synthesized recent progress across medical imaging, electronic health records, physiological time series, omics, and multimodal vision–language systems, and provided an operational taxonomy linking contrastive design choices to clinical data properties. Overall, evidence suggests that contrastive pretraining can improve label efficiency and downstream performance, particularly when positives and augmentations are defined in a clinically meaningful way and evaluation includes external validation.
However, the review also reveals persistent limitations that constrain translation. Reported gains are difficult to compare due to heterogeneous datasets, label taxonomies, and evaluation protocols, and many studies still emphasize in-domain results with limited multi-center validation. Reproducibility remains weak because pretraining corpora, pairing heuristics, augmentation policies, and optimization settings are often incompletely reported, and because many advances rely on proprietary datasets that cannot be audited or replicated. In addition, several limitations may not be resolvable in the short term: false negatives induced by patient similarity, shortcut learning driven by site and acquisition confounders, and noisy supervision in vision–language models arising from negation, uncertainty, and incomplete reporting in free-text clinical notes.
Future progress requires a shift from purely architectural novelty toward clinically grounded evidence. Robust benchmarking with standardized reporting, systematic evaluation under distribution shift, calibration, and subgroup analyses, and privacy-preserving training should become the default expectations. Ultimately, CL is a promising foundation for scalable medical representation learning, but its clinical impact will depend on reproducible methodology and rigorous validation under real-world deployment constraints.

Author Contributions

Conceptualization, G.O., I.D.M., K.A., E.E., C.W.C. and C.M.; methodology, I.D.M.; validation, I.D.M., K.A. and G.O.; investigation, I.D.M. and G.O.; resources, I.D.M., K.A., G.O. and C.W.C.; writing—original draft preparation, G.O., I.D.M., K.A., C.W.C., E.E. and C.M.; writing—review and editing, G.O., I.D.M., K.A., C.W.C., E.E. and C.M.; visualization, I.D.M.; supervision, G.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding authors.

Acknowledgments

The authors sincerely thank the reviewers for their constructive feedback, which helped improve the quality of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AUROCArea Under the Receiver Operating Characteristic Curve
AUPRCArea Under the Precision–Recall Curve
BioViLBiomedical Vision–Language
BioViL-TBiomedical Vision–Language with Temporal Alignment
BiomedCLIPBiomedical CLIP
BYOLBootstrap Your Own Latent
CheXpertCheXpert dataset
CLContrastive Learning
CLOCSContrastive Learning of Cardiac Signals
COMETHierarchical contrastive framework
CONCHHistopathology foundation model
ConVIRTContrastive Learning of Visual Representations from Text
CPCContrastive Predictive Coding
CTComputed Tomography
CXR-CLIPChest X-Ray CLIP
DiceDice coefficient
DINOSelf-Distillation with No Labels
ECGElectrocardiogram
EEGElectroencephalogram
EHR(s)Electronic Health Record(s)
F1F1 score
GLoRIAGlobal–Local image–text alignment in radiology
HD9595th percentile Hausdorff Distance
ICUIntensive Care Unit
IIDIndependent and Identically Distributed
InfoNCEInfo Noise-Contrastive Estimation (loss)
MaCoMasked Contrastive Learning
MBSLMulti-Scale and Multimodal Contrastive Learning
MedCLIPMedical CLIP
mIoUmean Intersection over Union
MIMIC-CXRMedical Information Mart for Intensive Care–Chest X-Ray
MoCoMomentum Contrast
MRIMagnetic Resonance Imaging
MSCLMulti-Scale Contrastive Learning
OODOut-of-Distribution
PCLRPatient Contrastive Learning of Representations
PLIPPathology Language–Image Pretraining
PMQPatient Memory Queue
PPI(s)Protein–Protein Interaction(s)
PMC-CLIPPubMed Central CLIP
RSNARadiological Society of North America
SAMSegment Anything Model
scRNA-seqsingle-cell RNA sequencing
SimCLRSimple Framework for Contrastive Learning of Visual Representations
SSLSelf-Supervised Learning
SupConSupervised Contrastive Learning
SwAVSwapping Assignments between Multiple Views
VQAVisual Question Answering
1D CNNOne-dimensional Convolutional Neural Network

Appendix A. Database-Specific Search Strings

We provide the database-specific search strategies used to support reproducibility. Search queries combined contrastive and self-supervised learning terms (e.g., “contrastive learning”, “self-supervised”, SimCLR, MoCo, BYOL) with medical domain keywords (e.g., medical imaging, EHR, genomics, ECG/EEG). Searches were restricted to January 2019–October 2025.

Appendix A.1. PubMed

Query:
(
(``contrastive learning’’[Title/Abstract] OR ``self-supervised’’[Title/Abstract] OR
``self supervised’’[Title/Abstract] OR ``representation learning’’[Title/Abstract] OR
SimCLR[Title/Abstract] OR MoCo[Title/Abstract] OR BYOL[Title/Abstract] OR
``supervised contrastive’’[Title/Abstract] OR InfoNCE[Title/Abstract] OR CLIP[Title/Abstract])
AND
(medical[Title/Abstract] OR healthcare[Title/Abstract] OR clinical[Title/Abstract] OR
radiology[Title/Abstract] OR ``chest x-ray’’[Title/Abstract] OR CXR[Title/Abstract] OR
MRI[Title/Abstract] OR CT[Title/Abstract] OR pathology[Title/Abstract] OR
``electronic health record’’[Title/Abstract] OR EHR[Title/Abstract] OR ``clinical notes’’[Title/Abstract] OR
ECG[Title/Abstract] OR EEG[Title/Abstract] OR ``physiological signals’’[Title/Abstract] OR
genomics[Title/Abstract] OR proteomics[Title/Abstract] OR ``single-cell’’[Title/Abstract] OR ``multi-omics’’[Title/Abstract])
)
  • Filters: Publication dates: 1 January 2019–31 October 2025; Article types: journal articles and conference papers where applicable; Language: English.

Appendix A.2. IEEE Xplore

Query (All Metadata):
(``contrastive learning’’ OR ``self-supervised’’ OR ``self supervised’’ OR ``representation learning’’
OR SimCLR OR MoCo OR BYOL OR ``supervised contrastive’’ OR InfoNCE OR CLIP)
AND
(medical OR healthcare OR clinical OR radiology OR ``chest x-ray’’ OR CXR OR MRI OR CT OR pathology
OR ``electronic health record’’ OR EHR OR ``clinical notes’’
OR ECG OR EEG OR ``physiological signals’’
OR genomics OR proteomics OR ``single-cell’’ OR ``multi-omics’’)
  • Filters: Years: 2019–2025; Content type: Journals and Conferences; Fields searched: All Metadata.

Appendix A.3. ACM Digital Library

Query:
(``contrastive learning’’ OR ``self-supervised’’ OR ``representation learning’’ OR SimCLR OR MoCo OR BYOL
OR ``supervised contrastive’’ OR InfoNCE OR CLIP)
AND
(medical OR healthcare OR clinical OR radiology OR ``chest x-ray’’ OR MRI OR CT OR pathology
OR ``electronic health record’’ OR EHR OR ``clinical notes’’
OR ECG OR EEG OR ``physiological signals’’
OR genomics OR proteomics OR ``single-cell’’ OR ``multi-omics’’)
  • Filters: Publication years: 2019–2025.

Appendix A.4. Web of Science

Query (Topic search, TS):
TS=(
(``contrastive learning’’ OR ``self-supervised’’ OR ``self supervised’’ OR ``representation learning’’
OR SimCLR OR MoCo OR BYOL OR ``supervised contrastive’’ OR InfoNCE OR CLIP)
AND
(medical OR healthcare OR clinical OR radiology OR ``chest x-ray’’ OR CXR OR MRI OR CT OR pathology
OR ``electronic health record’’ OR EHR OR ``clinical notes’’
OR ECG OR EEG OR ``physiological signals’’
OR genomics OR proteomics OR ``single-cell’’ OR ``multi-omics’’)
)
  • Filters: Timespan: 2019–2025; Document types: Article, Proceedings Paper; Language: English.

Appendix A.5. Scopus

Query (TITLE-ABS-KEY):
TITLE-ABS-KEY(
(``contrastive learning’’ OR ``self-supervised’’ OR ``self supervised’’ OR ``representation learning’’
OR SimCLR OR MoCo OR BYOL OR ``supervised contrastive’’ OR InfoNCE OR CLIP)
AND
(medical OR healthcare OR clinical OR radiology OR ``chest x-ray’’ OR CXR OR MRI OR CT OR pathology
OR ``electronic health record’’ OR EHR OR ``clinical notes’’
OR ECG OR EEG OR ``physiological signals’’
OR genomics OR proteomics OR ``single-cell’’ OR ``multi-omics’’)
)
  • Filters: Year: 2019–2025; Document type: Article, Conference Paper; Language: English.

Appendix A.6. arXiv

Query:
(``contrastive learning’’ OR ``self-supervised’’ OR ``representation learning’’ OR SimCLR OR MoCo OR BYOL
OR ``supervised contrastive’’ OR InfoNCE OR CLIP)
AND
(medical OR healthcare OR clinical OR radiology OR ``chest x-ray’’ OR MRI OR CT OR pathology
OR ``electronic health record’’ OR EHR OR ``clinical notes’’
OR ECG OR EEG OR ``physiological signals’’
OR genomics OR proteomics OR ``single-cell’’ OR ``multi-omics’’)
  • Filters: Years: 2019–2025; Categories considered: cs.LG, cs.CV, cs.AI, eess.IV, q-bio.QM.

Appendix B. Reporting Checklist

Table A1. Lightweight methodological and reporting checklist used to characterize included medical contrastive learning studies. Items are recorded as Yes/No/Unclear. The checklist is descriptive and was not used to exclude studies.
Table A1. Lightweight methodological and reporting checklist used to characterize included medical contrastive learning studies. Items are recorded as Yes/No/Unclear. The checklist is descriptive and was not used to exclude studies.
Checklist ItemRationale for Medical CL
Patient-level split reported (when applicable)Reduces leakage from repeated exams/segments across train and test.
External validation (cross-site/device/cohort)Supports generalization beyond in-domain evaluation.
Pretraining corpus described (size, source, pairing)Enables reproducibility and interpretation of representation quality.
Pairing strategy specified (what defines positives/negatives)Central determinant of clinical semantics and false-negative risk.
Augmentation policy specified and clinically justifiedPrevents invariances that erase subtle pathology or amplify artifacts.
Evaluation regime clearly stated (linear probe, fine-tune, few-shot, zero-shot)Needed for fair comparison across studies.
Primary metric(s) justified for task (e.g., AUROC vs F1)Metrics have different clinical meanings under imbalance and thresholding.
Baselines appropriate (supervised, SSL, ImageNet, CL variants)Prevents inflated claims due to weak comparisons.
Code/model released or sufficient implementation detail providedEnables replication and reuse.
Reproducibility constraints discussed (private data, restricted access)Helps interpret evidence strength and limitations.

Appendix C. Included Studies Summary Table

This section presents a structured taxonomy of the included contrastive learning studies (n = 38). For each study, we summarize (i) biomedical modality and dataset(s), (ii) clinical prediction or analysis task, (iii) contrastive learning formulation and pairing strategy, (iv) label regime and evaluation protocol, and (v) the primary metric used for reporting performance. This taxonomy is intended to improve reproducibility, support cross-modality synthesis, and clarify how contrastive design choices map to clinical data characteristics.
Table A2. Summary of included studies (n = 38): modality, dataset, task, contrastive formulation, label regime, evaluation protocol, and key metric.
Table A2. Summary of included studies (n = 38): modality, dataset, task, contrastive formulation, label regime, evaluation protocol, and key metric.
ModalityStudyYearDataset(s)TaskContrastive FormulationLabel RegimeEval ProtocolKey Metric
CXRAzizi et al. [16]2021Large-scale chest X-rays (multi-source CXR)Classification/transferImage–image CL (SimCLR-style): augmented views positives; others negativesSSL pretrain; supervised transferIID and external transfer evaluationAUROC/Acc
MRIChaitanya et al. [105]2020Cardiac/brain MRI segmentation benchmarksSegmentation (limited labels)Semi-supervised CL: consistency + image-view positives/negativesSS (few labels + unlabeled)CV or held-out splitDice/mIoU
HistopathologyCiga et al. [106]2022WSI patch datasets (public pathology cohorts)Cancer/tissue classificationPatch–patch SSL CL (augmentation-invariant representations)SSL pretrain; supervised fine-tunePatient-level IID split (where applicable)AUROC/F1
Cardiac MRIGuo et al. [107]2023Cardiac MRI segmentation dataset(s)Segmentation (myocardium/ventricles)Multi-scale CL loss (global/local scale alignment)Supervised + auxiliary CLTrain/val/test or CVDice
Brain MRILuo et al. [108]2023Brain MRI anomaly datasetsAnomaly detection/localizationPatch-level CL (normality representation learning)SSL/weak supervisionIID split; lesion localization evalAUROC/
AUPRC
EHRKrishnan et al. [109]2022ICU/hospital EHR cohort(s)Mortality/HF predictionPatient view–view CL: time-window crops, code dropout/shuffleSSL + supervised downstreamIID and/or temporal splitAUROC
EHRPick et al. [110]2024Hospital/ICU EHR cohort(s)Mortality + length-of-stayPatient embedding CL (same-patient/clinical-views positives)SSL/weak sup + supervised evalIID or temporal splitAUROC/
MAE
EHR (codes + notes)Sun et al. [114]2024Paired structured EHR + clinical notesProgression/
complications
Cross-modal CL (align code-based and note-based embeddings)Weak supervision (paired modalities)IID; optional cross-site testAUROC
EHR (large-scale)Cai et al. [115]2024Large EHR warehouse(s)Generalizable patient modelingDistributed CL pretraining (large memory bank/shards)SSL pretrain; supervised fine-tuneCross-population/cross-site when availableAUROC
EHR survivalKerdabadi et al. [111]2023Large longitudinal EHR cohort for AKI risk forecastingSurvival risk prediction (AKI)Temporal distinctiveness CL: hardness-aware negative sampling; time-aware pairsSupervised survival + CL auxiliary lossPatient-level split; temporal evaluationC-index/
AUROC
EHR (missing/irregular)Liu et al. [112]2023Two real-world EHR datasets (ICU/hospital cohorts)In-hospital mortality prediction (with imputation)CL-imputation–prediction network: patient stratification + contrastive repr. learningSupervised prediction + CL auxiliaryIID split; patient-level where applicableAUROC
EHRZang and Wang [113]2021Longitudinal EHR cohorts (risk prediction benchmarks)Mortality + phenotyping (multi-label)Supervised CL loss: contrastive cross-entropy + supervised CL regularizerSupervisedIID splitAUROC/
micro-F1
GenomicsZhong et al. [116]2024Bulk genomics datasets (expression/pathways)Biomarker discovery/predictionMulti-scale CL (gene-level + pathway-level alignment)Supervised + CL regularization or SSL + FTIID; CV commonAUROC/F1
ProteinsBepler and Berger [119]2021Protein sequence corpora (+ structures)Function/structure predictionContrastive protein sequence representationsSSL pretrain; supervised downstreamIID; family holdout possibleAcc/AUROC
Multi-omicsLiu et al. [117]2022Multi-omics cohorts (paired omics)Disease susceptibility/outcome predictionCross-omics CL: align patient embeddings across modalitiesWeak sup (paired omics) + supervised taskIID; CVAUROC
scRNA-seqLi et al. [118]2024Multiple scRNA-seq datasetsClustering/rare cell discoveryCell–cell CL with augmented count views; batch robustnessSSLCross-dataset/batch-aware evaluationARI/NMI
Seq + structureZhang et al. [120]2024Paired peptide sequence–structure datasetsPPI predictionSequence–structure CL alignmentWeak sup (paired modalities) + supervisedIID + cold-start splitAUROC
CXR + reportZhang et al. [18]2022Paired CXR–report corpora (e.g., MIMIC-CXR)CXR repr. learning; classification/retrievalImage–text CL (CLIP-style)Weak sup (paired)IID; external transferAUROC/
Recall@K
CXR + reportHuang et al. [65]2021MIMIC-CXRClassification/retrieval/
grounding
Global–local image–text CL (region–phrase + global)Weak sup (paired)IID; external optionalAUROC/
Recall@K
CXR + reportBoecking et al. [66]2022Large CXR-report corporaZS/FS radiology benchmarksBioViL vision–language CLWeak sup (paired) + ZS evalIID; externalAUROC
CXR (temporal)Bannur et al. [67]2023MIMIC-CXR (longitudinal pairs)Disease progression trackingTemporal CL across patient studies/timeWeak sup (paired + temporal IDs)Patient split; temporal testAUROC
CXR + promptYou et al. [68]2023CheXpert/MIMIC-CXRPrompt-based recognitionPrompt/label-text alignment CLWeak sup + supervised finetuneIID; externalAUROC
CXR (ZS)Tiu et al. [69]2022Pretrain paired CXR–report; eval CheXpertZero-shot multi-label classificationCLIP-style image–text CL; prompt inferenceWeak sup + ZS evalExternal benchmarkAUROC
Radiology VLPWang et al. [70]2022Paired + unpaired image/text corporaRadiology classification robustnessDecoupled/knowledge-aware CL matchingWeak supervisionIID; external validationAUROC
Pathology VLPHuang et al. [71]2023Pathology image–text corporaZero-shot pathology transferCLIP-style PLIPWeak sup (paired)Cross-dataset transferAUROC
Histo VLPLu et al. [72]20241.17 M histopathology image–caption pairsRetrieval/segmentation transferLarge-scale image–text CLWeak supMulti-task transfer evaluationAUROC/
Dice
Biomedical VLPZhang et al. [73]2023PubMed-scale biomedical image–textZero/few-shot across tasksCLIP-style biomedical CLWeak supBenchmark suiteAUROC/
Recall@K
PMC VLPLin et al. [74]20231.6 M PMC figure–caption pairsVQA/retrievalFigure–caption image–text CLWeak supBenchmark evaluationVQA Acc/
Recall@K
CXR masked CLHuang et al. [121]2024MIMIC-CXR/CheXpertZS + localizationMasked CL + image–text alignmentWeak supIID + external evaluationAUROC
Segmentation (multi)Koleilat et al. [122]2024Ultrasound/MRI/CT segmentation setsText-driven segmentationMedCLIP + SAM; text prompts guide masksWeak sup (prompts)Multi-dataset transferDice/mIoU
EHR (structured)Liu et al. [112]2023MIMIC-III; eICUClinical time-series imputation + in-hospital mortality predictionImputation: unsupervised CL (patient vs augmented view)Hybrid: SSL (imputation) + supervised (mortality labels)Random split 70/15/15 train/val/test; repeated 10 runsImputation: MAE/MRE; Prediction: AUROC/AUPRC
ECGDiamant et al. [19]2022Repeated ECG per patient cohortsCardiac disease predictionPatient-level CL (same-patient positives)SSL + supervised downstreamPatient splitAUROC
ECGYuan et al. [124]2025ECG recordings (windowed)Classification repr. learningPoly-window CL (overlap-aware positives)SSLIID/patient splitAUROC
ECG/EEGWang et al. [125]2023ECG and/or EEG datasetsFew-label classificationHierarchical CL (instance/segment/patient)SSL/SS + supervised FTFew-label protocolAUROC/F1
ECGChen et al. [76]2025Multi-lead ECG datasetsRobust classification (lead/time shift)Spatiotemporal CL (lead-wise + time-wise invariance)SSLPatient split; robustness testsAUROC
Physio + labsRaghu et al. [126]2022Multimodal ICU time-seriesOutcome predictionMultimodal temporal CL (align modalities/time contexts)Weak sup (paired)Temporal or IID splitAUROC
WearablesGuo et al. [127]2025Respiration/HR/
motion datasets
Multi-task inferenceMulti-scale multimodal CLSSL + supervised tasksIID/subject splitAUROC/F1
ECGSun et al. [128]2025Repeated ECG per patient datasetsReduce false negatives in CL pretrainingPatient memory queue (intra-patient positives)SSL + supervised downstreamPatient splitAUROC

References

  1. Parvin, N.; Joo, S.W.; Jung, J.H.; Mandal, T.K. Multimodal AI in Biomedicine: Pioneering the Future of Biomaterials, Diagnostics, and Personalized Healthcare. Nanomaterials 2025, 15, 895. [Google Scholar] [CrossRef] [PubMed]
  2. Nazir, A.; Hussain, A.; Singh, M.; Assad, A. Deep learning in medicine: Advancing healthcare with intelligent solutions and the future of holography imaging in early diagnosis. Multimed. Tools Appl. 2025, 84, 17677–17740. [Google Scholar] [CrossRef]
  3. Mienye, I.D.; Swart, T.G.; Obaido, G.; Jordan, M.; Ilono, P. Deep convolutional neural networks in medical image analysis: A review. Information 2025, 16, 195. [Google Scholar] [CrossRef]
  4. Mienye, I.D.; Jere, N.; Obaido, G.; Ogunruku, O.O.; Esenogho, E.; Modisane, C. Large language models: An overview of foundational architectures, recent trends, and a new taxonomy. Discov. Appl. Sci. 2025, 7, 1027. [Google Scholar] [CrossRef]
  5. Nichyporuk, B.; Cardinell, J.; Szeto, J.; Mehta, R.; Falet, J.P.R.; Arnold, D.L.; Tsaftaris, S.A.; Arbel, T. Rethinking generalization: The impact of annotation style on medical image segmentation. arXiv 2022, arXiv:2210.17398. [Google Scholar] [CrossRef]
  6. Daneshjou, R.; Yuksekgonul, M.; Cai, Z.R.; Novoa, R.; Zou, J.Y. Skincon: A skin disease dataset densely annotated by domain experts for fine-grained debugging and analysis. Adv. Neural Inf. Process. Syst. 2022, 35, 18157–18167. [Google Scholar]
  7. Krenzer, A.; Makowski, K.; Hekalo, A.; Fitting, D.; Troya, J.; Zoller, W.G.; Hann, A.; Puppe, F. Fast machine learning annotation in the medical domain: A semi-automated video annotation tool for gastroenterologists. Biomed. Eng. Online 2022, 21, 33. [Google Scholar] [CrossRef]
  8. Chen, H.; Gouin-Vallerand, C.; Bouchard, K.; Gaboury, S.; Couture, M.; Bier, N.; Giroux, S. Contrastive Self-Supervised Learning for Sensor-Based Human Activity Recognition: A Review. IEEE Access 2024, 12, 152511–152531. [Google Scholar] [CrossRef]
  9. Liu, S.; Zhao, L.; Chen, D.; Song, Z. Contrastive learning for image complexity representation. arXiv 2024, arXiv:2408.03230. [Google Scholar] [CrossRef]
  10. Ren, X.; Wei, W.; Xia, L.; Huang, C. A comprehensive survey on self-supervised learning for recommendation. ACM Comput. Surv. 2025, 58, 1–38. [Google Scholar] [CrossRef]
  11. Prince, J.S.; Alvarez, G.A.; Konkle, T. Contrastive learning explains the emergence and function of visual category-selective regions. Sci. Adv. 2024, 10, eadl1776. [Google Scholar] [CrossRef]
  12. Albelwi, S. Survey on self-supervised learning: Auxiliary pretext tasks and contrastive learning methods in imaging. Entropy 2022, 24, 551. [Google Scholar] [CrossRef] [PubMed]
  13. Khan, A.; Asmatullah, L.; Malik, A.; Khan, S.; Asif, H. A Survey on Self-supervised Contrastive Learning for Multimodal Text-Image Analysis. arXiv 2025, arXiv:2503.11101. [Google Scholar]
  14. Zeng, D.; Wu, Y.; Hu, X.; Xu, X.; Shi, Y. Contrastive learning with synthetic positives. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 430–447. [Google Scholar]
  15. Xu, Z.; Dai, Y.; Liu, F.; Wu, B.; Chen, W.; Shi, L. Swin MoCo: Improving parotid gland MRI segmentation using contrastive learning. Med. Phys. 2024, 51, 5295–5307. [Google Scholar] [CrossRef] [PubMed]
  16. Azizi, S.; Mustafa, B.; Ryan, F.; Beaver, Z.; Freyberg, J.; Deaton, J.; Loh, A.; Karthikesalingam, A.; Kornblith, S.; Chen, T.; et al. Big self-supervised models advance medical image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 3478–3488. [Google Scholar]
  17. Sowrirajan, H.; Yang, J.; Ng, A.Y.; Rajpurkar, P. Moco pretraining improves representation and transferability of chest x-ray models. In Proceedings of the Medical Imaging with Deep Learning, Lübeck, Germany, 7–9 July 2021; PMLR: New York, NY, USA, 2021; pp. 728–744. [Google Scholar]
  18. Zhang, Y.; Jiang, H.; Miura, Y.; Manning, C.D.; Langlotz, C.P. Contrastive learning of medical visual representations from paired images and text. In Proceedings of the Machine Learning for Healthcare Conference, Durham, NC, USA, 5–6 August 2022; PMLR: New York, NY, USA, 2022; pp. 2–25. [Google Scholar]
  19. Diamant, N.; Reinertsen, E.; Song, S.; Aguirre, A.D.; Stultz, C.M.; Batra, P. Patient contrastive learning: A performant, expressive, and practical approach to electrocardiogram modeling. PLoS Comput. Biol. 2022, 18, e1009862. [Google Scholar] [CrossRef]
  20. Jaiswal, A.; Babu, A.R.; Zadeh, M.Z.; Banerjee, D.; Makedon, F. A survey on contrastive self-supervised learning. Technologies 2020, 9, 2. [Google Scholar] [CrossRef]
  21. Gui, J.; Chen, T.; Zhang, J.; Cao, Q.; Sun, Z.; Luo, H.; Tao, D. A Survey on Self-supervised Learning: Algorithms, Applications, and Future Trends. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 9052–9071. [Google Scholar] [CrossRef]
  22. Hu, H.; Wang, X.; Zhang, Y.; Chen, Q.; Guan, Q. A comprehensive survey on contrastive learning. Neurocomputing 2024, 610, 128645. [Google Scholar] [CrossRef]
  23. Liu, R. Understand and improve contrastive learning methods for visual representation: A review. arXiv 2021, arXiv:2106.03259. [Google Scholar] [CrossRef]
  24. Shurrab, S.; Duwairi, R. Self-supervised learning methods and applications in medical imaging analysis: A survey. PeerJ Comput. Sci. 2022, 8, e1045. [Google Scholar] [CrossRef]
  25. Wang, W.C.; Ahn, E.; Feng, D.; Kim, J. A review of predictive and contrastive self-supervised learning for medical images. Mach. Intell. Res. 2023, 20, 483–513. [Google Scholar] [CrossRef]
  26. Huang, S.C.; Pareek, A.; Jensen, M.; Lungren, M.P.; Yeung, S.; Chaudhari, A.S. Self-supervised learning for medical image classification: A systematic review and implementation guidelines. NPJ Digit. Med. 2023, 6, 74. [Google Scholar] [CrossRef] [PubMed]
  27. VanBerlo, B.; Hoey, J.; Wong, A. A survey of the impact of self-supervised pretraining for diagnostic tasks in medical X-ray, CT, MRI, and ultrasound. BMC Med. Imaging 2024, 24, 79. [Google Scholar] [CrossRef] [PubMed]
  28. Yeh, C.H.; Hong, C.Y.; Hsu, Y.C.; Liu, T.L.; Chen, Y.; LeCun, Y. Decoupled contrastive learning. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 668–684. [Google Scholar]
  29. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised contrastive learning. Adv. Neural Inf. Process. Syst. 2020, 33, 18661–18673. [Google Scholar]
  30. Tian, Y.; Sun, C.; Poole, B.; Krishnan, D.; Schmid, C.; Isola, P. What makes for good views for contrastive learning? Adv. Neural Inf. Process. Syst. 2020, 33, 6827–6839. [Google Scholar]
  31. Kalantidis, Y.; Sariyildiz, M.B.; Pion, N.; Weinzaepfel, P.; Larlus, D. Hard negative mixing for contrastive learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21798–21809. [Google Scholar]
  32. Wu, J.; Chen, J.; Wu, J.; Shi, W.; Wang, X.; He, X. Understanding contrastive learning via distributionally robust optimization. Adv. Neural Inf. Process. Syst. 2024, 36, 23297–23320. [Google Scholar]
  33. Le-Khac, P.H.; Healy, G.; Smeaton, A.F. Contrastive representation learning: A framework and review. IEEE Access 2020, 8, 193907–193934. [Google Scholar] [CrossRef]
  34. Falcon, W.; Cho, K. A framework for contrastive self-supervised learning and designing a new approach. arXiv 2020, arXiv:2009.00104. [Google Scholar] [CrossRef]
  35. Peng, X.; Wang, K.; Zhu, Z.; Wang, M.; You, Y. Crafting better contrastive views for siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16031–16040. [Google Scholar]
  36. Kim, B.; Ye, J.C. Energy-based contrastive learning of visual representations. Adv. Neural Inf. Process. Syst. 2022, 35, 4358–4369. [Google Scholar]
  37. Wu, L.; Zhuang, J.; Chen, H. Voco: A simple-yet-effective volume contrastive learning framework for 3d medical image analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 22873–22882. [Google Scholar]
  38. Tang, C.; Zeng, X.; Zhou, L.; Zhou, Q.; Wang, P.; Wu, X.; Ren, H.; Zhou, J.; Wang, Y. Semi-supervised medical image segmentation via hard positives oriented contrastive learning. Pattern Recognit. 2024, 146, 110020. [Google Scholar] [CrossRef]
  39. Kundu, R. The Beginner’s Guide to Contrastive Learning. 2022. Available online: https://www.v7labs.com/blog/contrastive-learning-guide (accessed on 10 October 2025).
  40. Zhang, C.; Zhang, K.; Pham, T.X.; Niu, A.; Qiao, Z.; Yoo, C.D.; Kweon, I.S. Dual temperature helps contrastive learning without many negative samples: Towards understanding and simplifying moco. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14441–14450. [Google Scholar]
  41. Hoffmann, D.T.; Behrmann, N.; Gall, J.; Brox, T.; Noroozi, M. Ranking info noise contrastive estimation: Boosting contrastive learning via ranked positives. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 897–905. [Google Scholar]
  42. Xu, L.; Xie, H.; Li, Z.; Wang, F.L.; Wang, W.; Li, Q. Contrastive learning models for sentence representations. ACM Trans. Intell. Syst. Technol. 2023, 14, 1–34. [Google Scholar] [CrossRef]
  43. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020; PMLR: New York, NY, USA, 2020; pp. 1597–1607. [Google Scholar]
  44. Zhang, H.; Cao, Y. Understanding the benefits of simclr pre-training in two-layer convolutional neural networks. arXiv 2024, arXiv:2409.18685. [Google Scholar]
  45. Bunyang, S.; Thedwichienchai, N.; Pintong, K.; Lael, N.; Kunaborimas, W.; Boonrat, P.; Siriborvornratanakul, T. Self-supervised learning advanced plant disease image classification with SimCLR. Adv. Comput. Intell. 2023, 3, 18. [Google Scholar] [CrossRef]
  46. Fırıldak, K.; Çelik, G.; Talu, M.F. SimCLR-based Self-Supervised Learning Approach for Limited Brain MRI and Unlabeled Images. Bitlis Eren Üniv. Fen Bilim. Derg. 2024, 13, 1304–1313. [Google Scholar] [CrossRef]
  47. Chen, X.; Fan, H.; Girshick, R.; He, K. Improved baselines with momentum contrastive learning. arXiv 2020, arXiv:2003.04297. [Google Scholar] [CrossRef]
  48. Li, Y.; Liu, Q.; Zhou, L.; Zhao, W.; Tian, Y.; Zhang, W. Improved contrastive learning with MoCo framework. In Proceedings of the 2023 3rd International Conference on Consumer Electronics and Computer Engineering (ICCECE), Guangzhou, China, 6–8 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 729–732. [Google Scholar]
  49. He, Y.; Wang, X.; Shi, T. Ddpm-moco: Advancing industrial surface defect generation and detection with generative and contrastive learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Jeju, Republic of Korea, 3–9 August 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 34–49. [Google Scholar]
  50. Xie, E.; Ding, J.; Wang, W.; Zhan, X.; Xu, H.; Sun, P.; Li, Z.; Luo, P. Detco: Unsupervised contrastive learning for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 8392–8401. [Google Scholar]
  51. Xiao, T.; Wang, X.; Efros, A.A.; Darrell, T. What should not be contrastive in contrastive learning. arXiv 2020, arXiv:2008.05659. [Google Scholar] [CrossRef]
  52. Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent-a new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
  53. Deng, Z.; Man, J.; Song, Z.; Yang, G. A Few-Shot Anomaly Detection Method Based on BYOL Contrastive Learning Framework. In Proceedings of the 2024 4th International Conference on Robotics, Automation and Intelligent Control (ICRAIC), Changsha, China, 6–9 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 439–443. [Google Scholar]
  54. Richemond, P.H.; Grill, J.B.; Altché, F.; Tallec, C.; Strub, F.; Brock, A.; Smith, S.; De, S.; Pascanu, R.; Piot, B.; et al. Byol works even without batch statistics. arXiv 2020, arXiv:2010.10241. [Google Scholar] [CrossRef]
  55. Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; Joulin, A. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 9650–9660. [Google Scholar]
  56. Takanami, K.; Takahashi, T.; Sakata, A. The effect of optimal self-distillation in noisy gaussian mixture model. arXiv 2025, arXiv:2501.16226. [Google Scholar] [CrossRef]
  57. Wang, Q.; Mao, Z.; Gao, J.; Zhang, Y. Document-level relation extraction with progressive self-distillation. ACM Trans. Inf. Syst. 2024, 42, 1–34. [Google Scholar] [CrossRef]
  58. Tong, S.; Xia, Z.; Alahi, A.; He, X.; Shi, Y. GeoDistill: Geometry-Guided Self-Distillation for Weakly Supervised Cross-View Localization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Honolulu, HI, USA, 19–23 October 2025; pp. 25357–25366. [Google Scholar]
  59. Chen, X.; He, K. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15750–15758. [Google Scholar]
  60. Zhang, C.; Zhang, K.; Zhang, C.; Pham, T.X.; Yoo, C.D.; Kweon, I.S. How does simsiam avoid collapse without negative samples? A unified understanding with self-supervised contrastive learning. arXiv 2022, arXiv:2203.16262. [Google Scholar]
  61. Lu, Y.; Jha, A.; Deng, R.; Huo, Y. Contrastive learning meets transfer learning: A case study in medical image analysis. In Proceedings of the Medical Imaging 2022: Computer-Aided Diagnosis, San Diego, CA, USA, 20–24 February 2022; SPIE: Bellingham, WA, USA, 2022; Volume 12033, pp. 729–736. [Google Scholar]
  62. Khorram, S.; Kim, J.; Tripathi, A.; Lu, H.; Zhang, Q.; Sak, H. Contrastive siamese network for semi-supervised speech recognition. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 7207–7211. [Google Scholar]
  63. Caron, M.; Misra, I.; Mairal, J.; Goyal, P.; Bojanowski, P.; Joulin, A. Unsupervised learning of visual features by contrasting cluster assignments. Adv. Neural Inf. Process. Syst. 2020, 33, 9912–9924. [Google Scholar]
  64. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; PMLR: New York, NY, USA, 2021; pp. 8748–8763. [Google Scholar]
  65. Huang, S.C.; Shen, L.; Lungren, M.P.; Yeung, S. Gloria: A multimodal global-local representation learning framework for label-efficient medical image recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 3942–3951. [Google Scholar]
  66. Boecking, B.; Usuyama, N.; Bannur, S.; Castro, D.C.; Schwaighofer, A.; Hyland, S.; Wetscherek, M.; Naumann, T.; Nori, A.; Alvarez-Valle, J.; et al. Making the most of text semantics to improve biomedical vision–language processing. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–21. [Google Scholar]
  67. Bannur, S.; Hyland, S.; Liu, Q.; Perez-Garcia, F.; Ilse, M.; Castro, D.C.; Boecking, B.; Sharma, H.; Bouzid, K.; Thieme, A.; et al. Learning to exploit temporal structure for biomedical vision-language processing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 15016–15027. [Google Scholar]
  68. You, K.; Gu, J.; Ham, J.; Park, B.; Kim, J.; Hong, E.K.; Baek, W.; Roh, B. Cxr-clip: Toward large scale chest x-ray language-image pre-training. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 101–111. [Google Scholar]
  69. Tiu, E.; Talius, E.; Patel, P.; Langlotz, C.P.; Ng, A.Y.; Rajpurkar, P. Expert-level detection of pathologies from unannotated chest X-ray images via self-supervised learning. Nat. Biomed. Eng. 2022, 6, 1399–1406. [Google Scholar] [CrossRef] [PubMed]
  70. Wang, Z.; Wu, Z.; Agarwal, D.; Sun, J. Medclip: Contrastive learning from unpaired medical images and text. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; Volume 2022, p. 3876. [Google Scholar]
  71. Huang, Z.; Bianchi, F.; Yuksekgonul, M.; Montine, T.J.; Zou, J. A visual–language foundation model for pathology image analysis using medical twitter. Nat. Med. 2023, 29, 2307–2316. [Google Scholar] [CrossRef]
  72. Lu, M.Y.; Chen, B.; Williamson, D.F.; Chen, R.J.; Liang, I.; Ding, T.; Jaume, G.; Odintsov, I.; Le, L.P.; Gerber, G.; et al. A visual-language foundation model for computational pathology. Nat. Med. 2024, 30, 863–874. [Google Scholar] [CrossRef]
  73. Zhang, S.; Xu, Y.; Usuyama, N.; Xu, H.; Bagga, J.; Tinn, R.; Preston, S.; Rao, R.; Wei, M.; Valluri, N.; et al. Biomedclip: A multimodal biomedical foundation model pretrained from fifteen million scientific image-text pairs. arXiv 2023, arXiv:2303.00915. [Google Scholar]
  74. Lin, W.; Zhao, Z.; Zhang, X.; Wu, C.; Zhang, Y.; Wang, Y.; Xie, W. Pmc-clip: Contrastive language-image pre-training using biomedical documents. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 525–536. [Google Scholar]
  75. van den Oord, A.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  76. Chen, W.; Wang, H.; Zhang, L.; Zhang, M. Temporal and spatial self supervised learning methods for electrocardiograms. Sci. Rep. 2025, 15, 6029. [Google Scholar] [CrossRef]
  77. Chuang, C.Y.; Robinson, J.; Lin, Y.C.; Torralba, A.; Jegelka, S. Debiased contrastive learning. Adv. Neural Inf. Process. Syst. 2020, 33, 8765–8775. [Google Scholar]
  78. Robinson, J.; Sun, L.; Yu, K.; Batmanghelich, K.; Jegelka, S.; Sra, S. Can contrastive learning avoid shortcut solutions? Adv. Neural Inf. Process. Syst. 2021, 34, 4974–4986. [Google Scholar] [PubMed]
  79. Jang, T.; Wang, X. Difficulty-based sampling for debiased contrastive representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 24039–24048. [Google Scholar]
  80. Biswas, D.; Tešić, J. Unsupervised domain adaptation with debiased contrastive learning and support-set guided pseudolabeling for remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 3197–3210. [Google Scholar] [CrossRef]
  81. Agarwal, A.; Banerjee, T.; Romine, W.L.; Cajita, M. Debias-clr: A contrastive learning based debiasing method for algorithmic fairness in healthcare applications. In Proceedings of the 2024 IEEE International Conference on Big Data (BigData), Washington DC, USA, 15–18 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 6411–6419. [Google Scholar]
  82. Yun, B.; Zhao, S.; Li, Q.; Kot, A.; Wang, Y. Debiasing Medical Knowledge for Prompting Universal Model in CT Image Segmentation. IEEE Trans. Med. Imaging 2025, 44, 5142–5154. [Google Scholar] [CrossRef] [PubMed]
  83. Tang, P.; Ouyang, C.; Liu, Y. Debiasing medication recommendation with counterfactual analysis. In Proceedings of the International Conference on Neural Information Processing, Changsha, China, 20–23 November 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 426–438. [Google Scholar]
  84. Johnson, A.E.W.; Pollard, T.J.; Berkowitz, S.J.; Greenbaum, N.R.; Lungren, M.P.; Deng, C.Y.; Mark, R.G.; Horng, S. MIMIC-CXR: A large publicly available database of labeled chest radiographs. arXiv 2019, arXiv:1901.07042. [Google Scholar]
  85. Irvin, J.; Rajpurkar, P.; Ko, M.; Yu, Y.; Ciurea-Ilinca, M.; Chute, C.; Marklund, H.; Haghgoo, B.; Ball, R.L.; Shpanskaya, K.; et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison. Proc. AAAI Conf. Artif. Intell. 2019, 33, 590–597. [Google Scholar] [CrossRef]
  86. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-ray8: Hospital-scale Chest X-ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar] [CrossRef]
  87. Weiner, M.W.; Veitch, D.P.; Aisen, P.S.; Beckett, L.A.; Cairns, N.J.; Green, R.C.; Harvey, D.; Jack, C.R.; Jagust, W.; Liu, E.; et al. The Alzheimer’s Disease Neuroimaging Initiative: A review of papers published since its inception. Alzheimer’s Dement. 2015, 11, e1–e120. [Google Scholar] [CrossRef]
  88. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
  89. Baid, U.; Ghodasara, S.; Mohan, S.; Bilello, M.; Calabrese, E.; Colak, E.; Farahani, K.; Kalpathy-Cramer, J.; Kitamura, F.C.; Pati, S.; et al. The rsna-asnr-miccai brats 2021 benchmark on brain tumor segmentation and radiogenomic classification. arXiv 2021, arXiv:2107.02314. [Google Scholar]
  90. Cassidy, B.; Kendrick, C.; Brodzicki, A.; Jaworek-Korjakowska, J.; Yap, M.H. Analysis of the ISIC image datasets: Usage, benchmarks and recommendations. Med. Image Anal. 2022, 75, 102305. [Google Scholar] [CrossRef]
  91. Bejnordi, B.E.; Veta, M.; Van Diest, P.J.; Van Ginneken, B.; Karssemeijer, N.; Litjens, G.; Van Der Laak, J.A.; Hermsen, M.; Manson, Q.F.; Balkenhol, M.; et al. Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. JAMA 2017, 318, 2199–2210. [Google Scholar] [CrossRef]
  92. Underwood, T. Pan-cancer analysis of whole genomes. Nature 2020, 578, 82–93. [Google Scholar] [CrossRef]
  93. Johnson, A.E.; Pollard, T.J.; Shen, L.; Lehman, L.w.H.; Feng, M.; Ghassemi, M.; Moody, B.; Szolovits, P.; Anthony Celi, L.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 160035. [Google Scholar] [CrossRef]
  94. Johnson, A.E.; Bulgarelli, L.; Shen, L.; Gayles, A.; Shammout, A.; Horng, S.; Pollard, T.J.; Hao, S.; Moody, B.; Gow, B.; et al. MIMIC-IV, a freely accessible electronic health record dataset. Sci. Data 2023, 10, 1. [Google Scholar] [CrossRef]
  95. Pollard, T.J.; Johnson, A.E.; Raffa, J.D.; Celi, L.A.; Mark, R.G.; Badawi, O. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Sci. Data 2018, 5, 180178. [Google Scholar] [CrossRef] [PubMed]
  96. Wagner, P.; Strodthoff, N.; Bousseljot, R.D.; Kreiseler, D.; Lunze, F.I.; Samek, W.; Schaeffter, T. PTB-XL, a large publicly available electrocardiography dataset. Sci. Data 2020, 7, 1–15. [Google Scholar] [CrossRef] [PubMed]
  97. Obeid, I.; Picone, J. The temple university hospital EEG data corpus. Front. Neurosci. 2016, 10, 196. [Google Scholar] [CrossRef] [PubMed]
  98. Weinstein, J.N.; Collisson, E.A.; Mills, G.B.; Shaw, K.R.; Ozenberger, B.A.; Ellrott, K.; Shmulevich, I.; Sander, C.; Stuart, J.M. The cancer genome atlas pan-cancer analysis project. Nat. Genet. 2013, 45, 1113–1120. [Google Scholar] [CrossRef]
  99. Lonsdale, J.; Thomas, J.; Salvatore, M.; Phillips, R.; Lo, E.; Shad, S.; Hasz, R.; Walters, G.; Garcia, F.; Young, N.; et al. The genotype-tissue expression (GTEx) project. Nat. Genet. 2013, 45, 580–585. [Google Scholar] [CrossRef]
  100. Bycroft, C.; Freeman, C.; Petkova, D.; Band, G.; Elliott, L.T.; Sharp, K.; Motyer, A.; Vukcevic, D.; Delaneau, O.; O’Connell, J.; et al. The UK Biobank resource with deep phenotyping and genomic data. Nature 2018, 562, 203–209. [Google Scholar] [CrossRef]
  101. Regev, A.; Teichmann, S.A.; Lander, E.S.; Amit, I.; Benoist, C.; Birney, E.; Bodenmiller, B.; Campbell, P.; Carninci, P.; Clatworthy, M.; et al. The human cell atlas. elife 2017, 6, e27041. [Google Scholar] [CrossRef]
  102. Hasanah, U.; Leu, J.S.; Avian, C.; Azmi, I.; Prakosa, S.W. A systematic review of multilabel chest X-ray classification using deep learning. Multimed. Tools Appl. 2025, 84, 26719–26753. [Google Scholar] [CrossRef]
  103. Bhusal, D.; Panday, S.P. Multi-label classification of thoracic diseases using dense convolutional network on chest radiographs. arXiv 2022, arXiv:2202.03583. [Google Scholar]
  104. Sammani, F.; Joukovsky, B.; Deligiannis, N. Visualizing and understanding contrastive learning. IEEE Trans. Image Process. 2023, 33, 541–555. [Google Scholar] [CrossRef] [PubMed]
  105. Chaitanya, K.; Erdil, E.; Karani, N.; Konukoglu, E. Contrastive learning of global and local features for medical image segmentation with limited annotations. Adv. Neural Inf. Process. Syst. 2020, 33, 12546–12558. [Google Scholar]
  106. Ciga, O.; Xu, T.; Martel, A.L. Self supervised contrastive learning for digital histopathology. Mach. Learn. Appl. 2022, 7, 100198. [Google Scholar] [CrossRef]
  107. Guo, Z.; Zhang, Y.; Qiu, Z.; Dong, S.; He, S.; Gao, H.; Zhang, J.; Chen, Y.; He, B.; Kong, Z.; et al. An improved contrastive learning network for semi-supervised multi-structure segmentation in echocardiography. Front. Cardiovasc. Med. 2023, 10, 1266260. [Google Scholar] [CrossRef]
  108. Luo, G.; Xie, W.; Gao, R.; Zheng, T.; Chen, L.; Sun, H. Unsupervised anomaly detection in brain MRI: Learning abstract distribution from massive healthy brains. Comput. Biol. Med. 2023, 154, 106610. [Google Scholar] [CrossRef]
  109. Krishnan, R.; Rajpurkar, P.; Topol, E.J. Self-supervised learning in medicine and healthcare. Nat. Biomed. Eng. 2022, 6, 1346–1352. [Google Scholar] [CrossRef]
  110. Pick, F.; Xie, X.; Wu, L.Y. Contrastive Multitask Transformer for Hospital Mortality and Length-of-Stay Prediction. In Proceedings of the International Conference on AI in Healthcare, Laguna Hills, CA, USA, 5–7 February 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 134–145. [Google Scholar]
  111. Nayebi Kerdabadi, M.; Hadizadeh Moghaddam, A.; Liu, B.; Liu, M.; Yao, Z. Contrastive learning of temporal distinctiveness for survival analysis in electronic health records. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; pp. 1897–1906. [Google Scholar]
  112. Liu, Y.; Zhang, Z.; Qin, S.; Salim, F.D.; Yepes, A.J. Contrastive learning-based imputation-prediction networks for in-hospital mortality risk modeling using ehrs. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Turin, Italy, 18–22 September 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 428–443. [Google Scholar]
  113. Zang, C.; Wang, F. Scehr: Supervised contrastive learning for clinical risk prediction using electronic health records. In Proceedings of the IEEE International Conference on Data Mining, Auckland, New Zealand, 7–10 December 2021; Volume 2021, p. 857. [Google Scholar]
  114. Sun, M.; Yang, X.; Niu, J.; Gu, Y.; Wang, C.; Zhang, W. A cross-modal clinical prediction system for intensive care unit patient outcome. Knowl.-Based Syst. 2024, 283, 111160. [Google Scholar] [CrossRef]
  115. Cai, T.; Huang, F.; Nakada, R.; Zhang, L.; Zhou, D. Contrastive Learning on Multimodal Analysis of Electronic Health Records. arXiv 2024, arXiv:2403.14926. [Google Scholar] [CrossRef]
  116. Zhong, X.; Batmanghelich, K.; Sun, L. Enhancing Biomedical Multi-modal Representation Learning with Multi-scale Pre-training and Perturbed Report Discrimination. In Proceedings of the 2024 IEEE Conference on Artificial Intelligence (CAI), Singapore, 25–27 June 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 480–485. [Google Scholar]
  117. Liu, X.; Xu, X.; Xu, X.; Li, X.; Xie, G. Representation Learning for Multi-omics Data with Heterogeneous Gene Regulatory Network. In Proceedings of the 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA, 9–12 December 2021; pp. 702–705. [Google Scholar] [CrossRef]
  118. Li, S.; Ma, J.; Zhao, T.; Jia, Y.; Liu, B.; Luo, R.; Huang, Y. CellContrast: Reconstructing spatial relationships in single-cell RNA sequencing data via deep contrastive learning. Patterns 2024, 5, 101022. [Google Scholar] [CrossRef] [PubMed]
  119. Bepler, T.; Berger, B. Learning the protein language: Evolution, structure, and function. Cell Syst. 2021, 12, 654–669. [Google Scholar] [CrossRef] [PubMed]
  120. Zhang, R.; Wu, H.; Liu, C.; Li, H.; Wu, Y.; Li, K.; Wang, Y.; Deng, Y.; Chen, J.; Zhou, F.; et al. Pepharmony: A multi-view contrastive learning framework for integrated sequence and structure-based peptide encoding. arXiv 2024, arXiv:2401.11360. [Google Scholar] [CrossRef] [PubMed]
  121. Huang, W.; Li, C.; Zhou, H.Y.; Yang, H.; Liu, J.; Liang, Y.; Zheng, H.; Zhang, S.; Wang, S. Enhancing representation in radiography-reports foundation model: A granular alignment algorithm using masked contrastive learning. Nat. Commun. 2024, 15, 7620. [Google Scholar] [CrossRef]
  122. Koleilat, T.; Asgariandehkordi, H.; Rivaz, H.; Xiao, Y. Medclip-sam: Bridging text and image towards universal medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 643–653. [Google Scholar]
  123. Liu, Z.; Alavi, A.; Li, M.; Zhang, X. Self-supervised contrastive learning for medical time series: A systematic review. Sensors 2023, 23, 4221. [Google Scholar] [CrossRef]
  124. Yuan, Y.; Van Duyn, J.; Yan, R.; Huang, Z.; Vesal, S.; Plis, S.; Hu, X.; Kwak, G.H.; Xiao, R.; Fedorov, A. Learning ECG Representations via Poly-Window Contrastive Learning. arXiv 2025, arXiv:2508.15225. [Google Scholar] [CrossRef]
  125. Wang, Y.; Han, Y.; Wang, H.; Zhang, X. Contrast everything: A hierarchical contrastive framework for medical time-series. Adv. Neural Inf. Process. Syst. 2023, 36, 55694–55717. [Google Scholar]
  126. Raghu, A.; Chandak, P.; Alam, R.; Guttag, J.; Stultz, C. Contrastive pre-training for multimodal medical time series. In Proceedings of the NeurIPS 2022 Workshop on Learning from Time Series for Health, New Orleans, LA, USA, 2 December 2022. [Google Scholar]
  127. Guo, H.; Xu, X.; Wu, H.; Liu, B.; Xia, J.; Cheng, Y.; Guo, Q.; Chen, Y.; Xu, T.; Wang, J.; et al. Multi-scale and multi-modal contrastive learning network for biomedical time series. Biomed. Signal Process. Control 2025, 106, 107697. [Google Scholar] [CrossRef]
  128. Sun, X.; Yang, Y.; Dong, X. Enhancing Contrastive Learning-based Electrocardiogram Pretrained Model with Patient Memory Queue. arXiv 2025, arXiv:2506.06310. [Google Scholar]
  129. Zech, J.R.; Badgeley, M.A.; Liu, M.; Costa, A.B.; Titano, J.J.; Oermann, E.K. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLoS Med. 2018, 15, e1002683. [Google Scholar] [CrossRef]
  130. Oakden-Rayner, L.; Dunnmon, J.; Carneiro, G.; Ré, C. Hidden stratification causes clinically meaningful failures in machine learning for medical imaging. In Proceedings of the ACM Conference on Health, Inference, and Learning, Toronto, ON, Canada, 2–4 April 2020; pp. 151–159. [Google Scholar]
  131. Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On calibration of modern neural networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; PMLR: New York, NY, USA, 2017; pp. 1321–1330. [Google Scholar]
  132. Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A. A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
  133. Samek, W.; Wiegand, T.; Müller, K.R. Explainable Artificial Intelligence: Understanding, Visualizing and Interpreting Deep Learning Models. arXiv 2017, arXiv:1708.08296. [Google Scholar] [CrossRef]
  134. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  135. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  136. Finlayson, S.G.; Bowers, J.D.; Ito, J.; Zittrain, J.L.; Beam, A.L.; Kohane, I.S. Adversarial attacks on medical machine learning. Science 2019, 363, 1287–1289. [Google Scholar] [CrossRef]
  137. Arjovsky, M.; Bottou, L.; Gulrajani, I.; Lopez-Paz, D. Invariant Risk Minimization. arXiv 2019, arXiv:1907.02893. [Google Scholar]
  138. Schölkopf, B.; Locatello, F.; Bauer, S.; Ke, N.R.; Kalchbrenner, N.; Goyal, A.; Bengio, Y. Toward Causal Representation Learning. Proc. IEEE 2021, 109, 612–634. [Google Scholar] [CrossRef]
  139. McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), Ft. Lauderdale, FL, USA, 20–22 April 2017; PMLR: New York, NY, USA, 2017; pp. 1273–1282. [Google Scholar]
  140. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep Learning with Differential Privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (CCS), Vienna, Austria, 24–28 October 2016; ACM: New York, NY, USA, 2016; pp. 308–318. [Google Scholar]
  141. Rajpurkar, P.; Chen, E.; Banerjee, O.; Topol, E.J. AI in health and medicine. Nat. Med. 2022, 28, 31–38. [Google Scholar] [CrossRef]
Figure 1. PRISMA diagram of the literature selection process.
Figure 1. PRISMA diagram of the literature selection process.
Bioengineering 13 00176 g001
Figure 3. Application Areas of Contrastive Learning in Medical AI.
Figure 3. Application Areas of Contrastive Learning in Medical AI.
Bioengineering 13 00176 g003
Table 1. Summary of representative review papers on contrastive and self-supervised learning, highlighting the scope and coverage of medical applications.
Table 1. Summary of representative review papers on contrastive and self-supervised learning, highlighting the scope and coverage of medical applications.
ReviewYearPrimary ScopeMedical CoverageNotes for Positioning in This Paper
Jaiswal et al. [20]2020General CL and SSL overviewLimitedEarly overview of CL principles and variants.
Gui et al. [21]2024Broad SSL surveyLimitedIncludes contrastive, generative, and clustering-based SSL methods.
Hu et al. [22]2024General CL surveyIndirectPresents a universal CL framework and component-level advances.
Liu [23]2021CL for visual representationIndirectVision-focused synthesis of CL components, limitations, and improvements.
Shurrab et al. [24]2022SSL in medical imagingImaging onlyFocuses on imaging applications with limited clinical data coverage.
Wang et al. [25]2023Medical imaging SSL and CLImaging onlyEmphasizes adaptation of natural image SSL methods to medical imaging.
Huang et al. [26]2023Systematic review of SSL for medical imagingImaging onlySystematic evidence synthesis and reporting trends for image classification studies.
VanBerlo et al. [27]2024Evidence review of SSL pretraining in imagingImaging onlyHighlights comparisons against supervised baselines and transfer protocols.
This review2025CL in medical AIImaging, EHR, signals, omics, multimodalCross modality synthesis with operational taxonomy and evaluation guidance.
Table 2. Taxonomy dimensions for medical contrastive learning studies.
Table 2. Taxonomy dimensions for medical contrastive learning studies.
DimensionDefinitionCommon Categories/Options
Loss family/
objective
The contrastive or self-supervised learning signal is used to shape representations.InfoNCE/NT-Xent; supervised contrastive; distillation/negative-free; clustering consistency; cross-modal retrieval loss; temporal predictive losses (e.g., CPC).
Pairing strategy (positives)How positives are constructed, i.e., what the method enforces to be similar.Augmented views of the same instance; same-class positives; patient-aware positives; longitudinal (same patient across time); cross-modal paired (image–text, ECG–EHR); region–text phrase.
Augmentations/
views
Transformations are used to generate alternative views and define invariances.Imaging: mild geometry/intensity, modality-specific; signals: masking/jitter/windowing; EHR: time masking/cropping; text: entity-aware processing; multi-view sampling (slices/windows).
Label regimeHow labels are used during pretraining (if at all) and during evaluation.Unsupervised (no labels); weak/self-labels; supervised contrastive; semi-supervised (mix of labeled/unlabeled); label-scarce learning curves.
Evaluation protocolHow the downstream benefit is tested, including whether shift and calibration are considered.Linear probe; full fine-tuning; few-shot/label fractions; external validation (site/device/time); temporal split; retrieval evaluation (Recall@K); calibration (ECE/Brier) for risk models.
Task familyClinical endpoint and decision context used for evaluation.Classification (diagnosis/risk); survival/prognosis; segmentation; detection; retrieval/triage; phenotype discovery/clustering; zero-shot/VLM prompting.
Table 4. Common benchmarking protocols and metrics for evaluating medical contrastive learning across modalities.
Table 4. Common benchmarking protocols and metrics for evaluating medical contrastive learning across modalities.
ProtocolWhat It TestsTypical Metrics (Examples)
Linear probeRepresentation quality under minimal supervisionAUROC/AUPRC (multi-label); accuracy; macro-F1 (class imbalance); calibration metrics when reported
Full fine-tuningEnd-task performance with task-specific adaptationAUROC/AUPRC; Dice/HD95 for segmentation; patient-level metrics for clinical outcomes
Few-shot and label-scarceLabel efficiency and robustness when annotation is limitedAUROC/F1 vs label fraction; confidence intervals across seeds; sensitivity to class imbalance
External validation and site shiftGeneralization across institutions, devices, protocols, and timePerformance drop under shift; subgroup robustness; calibration under shift; domain-wise reporting
Retrieval and alignment (vision–language)Cross-modal grounding and matching fidelityRecall@K; median rank; phrase grounding scores; zero-shot AUROC using text prompts
Table 5. Summary of applications of contrastive learning in medical AI.
Table 5. Summary of applications of contrastive learning in medical AI.
Application DomainAuthor(s)YearMethodApplication
Medical ImagingAzizi et al. [16]2021SimCLR-based pretraining on unlabeled chest X-raysLearned transferable visual representations for medical image classification and segmentation using unlabeled data.
Chaitanya et al. [105]2020Semi-supervised contrastive framework for MRIImproved segmentation and classification in MRI with limited labels using augmented positive and negative pairs.
Ciga et al. [106]2022Self-supervised contrastive learning for histopathologyEnhanced cancer detection and tissue differentiation in biopsy samples using augmentation-invariant representations.
Guo et al. [107]2023Multi-scale contrastive loss for cardiac MRI segmentationCaptured both global and local structures, improving accuracy of myocardium and ventricle segmentation.
Luo et al. [108]2023Self-supervised anomaly detection using contrastive lossDetected abnormal regions in brain MRI scans by distinguishing normal and pathological patches.
Electronic Health RecordsKrishnan et al. [109]2022Self-supervised contrastive learning on augmented EHR viewsModeled temporal and clinical correlations for mortality and heart failure prediction.
Pick et al. [110]2024Contrastive patient representation learningImproved prediction of hospital mortality and length-of-stay through patient-level embeddings.
Sun et al. [114]2024Cross-modal contrastive framework for EHR integrationAligned structured and unstructured EHR data to predict disease progression and complications.
Cai et al. [115]2024Distributed large-scale contrastive learningScalable training on large EHR datasets for improved generalization across patient populations.
Kerdabadi et al. [111]2023ontology-aware temporal contrastive survivalLearns temporally distinctive EHR embeddings with hardness-aware negatives for AKI survival risk prediction.
Liu et al. [112]2023Contrastive imputation–prediction network (CL-IPN)Contrastive-enhanced imputation with patient stratification improves in-hospital mortality prediction under missing/irregular EHRs.
Zang and Wang [113]2021Supervised Contrastive framework using longitudinal EHRUnified supervised contrastive loss improves EHR classification outcomes.
Genomics and ProteomicsZhong et al. [116]2024Multi-scale contrastive learning (MSCL) for genomicsIdentified disease-associated genetic markers by modeling gene and pathway-level interactions.
Bepler and Berger [119]2021Contrastive protein sequence representation learningLearned structural and functional protein embeddings for improved function prediction and drug discovery.
Liu et al. [117]2022Multi-omics contrastive learning (MoHeG/GenCL)Integrated genomics, transcriptomics, and epigenomics to predict disease susceptibility and treatment outcomes.
Li et al. [118]2024CellContrast for single-cell RNA sequencingEnhanced clustering and identification of rare cell types in scRNA-seq data.
Zhang et al. [120]2024Pepharmony: sequence–structure contrastive learningPredicted protein–protein interactions with improved accuracy using multimodal peptide representations.
Multimodal and Cross-Domain LearningZhang et al. [18]2022ConVIRT (image–text alignment)Learned chest X-ray representations by aligning images and radiology reports with bidirectional contrastive loss.
Huang et al. [65]2021GLoRIA (global–local image–text alignment)Improved retrieval and classification on MIMIC-CXR through local region–phrase alignment.
Boecking et al. [66]2022BioViL (biomedical vision–language model)Enhanced zero-shot radiology performance using domain-specific text pretraining.
Bannur et al. [67]2023BioViL-T (temporal alignment)Improved disease progression tracking in chest X-rays via temporal contrastive learning.
You et al. [68]2023CXR-CLIP (prompt-based multimodal CL)Combined image–label and image–text supervision for robust chest X-ray recognition.
Tiu et al. [69]2022CheXzero (CLIP-style vision–language model)Achieved radiologist-level zero-shot classification on the CheXpert benchmark.
Wang et al. [70]2022MedCLIP (knowledge-aware matching loss)Reduced false negatives in radiology by decoupling image–text corpora for efficient pretraining.
Huang et al. [71]2023PLIP (pathology vision–language foundation model)Achieved state-of-the-art performance in pathology classification and zero-shot transfer.
Lu et al. [72]2024CONCH (large-scale histopathology pretraining)Trained on 1.17 M image–caption pairs for generalizable pathology retrieval and segmentation.
Zhang et al. [73]2023BiomedCLIP (PubMed multimodal foundation model)Pretrained on 15 M image–text pairs for broad biomedical zero/few-shot applications.
Lin et al. [74]2023PMC-CLIP (literature-derived pretraining)Improved biomedical VQA and retrieval from 1.6M figure–caption pairs.
Huang et al. [121]2024MaCo (masked contrastive learning)Applied to chest X-rays for enhanced zero-shot and localized recognition.
Koleilat et al. [122]2024MedCLIP + SAM (text-driven segmentation)Enabled multimodal segmentation across ultrasound, MRI, and CT without explicit labels.
Time-Series and Physiological SignalsLiu et al. [123]2023Systematic review of contrastive time-series methodsIdentified key design trends in self-supervised ECG/EEG contrastive learning.
Diamant et al. [19]2022PCLR (patient-level contrastive learning)Leveraged same-patient ECGs to improve cardiac disease prediction tasks.
Yuan et al. [124]2025Poly-window contrastive learningModeled temporal overlap in ECGs to enhance representation efficiency.
Wang et al. [125]2023COMET (hierarchical contrastive framework)Applied multi-level contrastive learning for ECG and EEG classification with few labels.
Chen et al. [76]2025CLOCS (spatiotemporal contrastive model)Improved robustness in cardiac signals under lead and time variation.
Raghu et al. [126]2022Multimodal temporal contrastive pretrainingIntegrated physiological signals with lab and vitals data for outcome prediction.
Guo et al. [127]2025MBSL (multi-scale multimodal contrastive learning)Combined respiration, heart rate, and motion signals for multi-task biomedical inference.
Sun et al. [128]2025PMQ (patient memory queue)Mitigated false negatives in ECG pretraining by leveraging intra-patient memory banks.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Obaido, G.; Mienye, I.D.; Aruleba, K.; Chukwu, C.W.; Esenogho, E.; Modisane, C. A Systematic Review of Contrastive Learning in Medical AI: Foundations, Biomedical Modalities, and Future Directions. Bioengineering 2026, 13, 176. https://doi.org/10.3390/bioengineering13020176

AMA Style

Obaido G, Mienye ID, Aruleba K, Chukwu CW, Esenogho E, Modisane C. A Systematic Review of Contrastive Learning in Medical AI: Foundations, Biomedical Modalities, and Future Directions. Bioengineering. 2026; 13(2):176. https://doi.org/10.3390/bioengineering13020176

Chicago/Turabian Style

Obaido, George, Ibomoiye Domor Mienye, Kehinde Aruleba, Chidozie Williams Chukwu, Ebenezer Esenogho, and Cameron Modisane. 2026. "A Systematic Review of Contrastive Learning in Medical AI: Foundations, Biomedical Modalities, and Future Directions" Bioengineering 13, no. 2: 176. https://doi.org/10.3390/bioengineering13020176

APA Style

Obaido, G., Mienye, I. D., Aruleba, K., Chukwu, C. W., Esenogho, E., & Modisane, C. (2026). A Systematic Review of Contrastive Learning in Medical AI: Foundations, Biomedical Modalities, and Future Directions. Bioengineering, 13(2), 176. https://doi.org/10.3390/bioengineering13020176

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop