Next Article in Journal
Whole-Genome Sequence and Fermentation Characteristics of Enterobacter hormaechei UW0SKVC1: A Promising Candidate for Detoxification of Lignocellulosic Biomass Hydrolysates and Production of Value-Added Chemicals
Next Article in Special Issue
Multi-Scale Learning with Sparse Residual Network for Explainable Multi-Disease Diagnosis in OCT Images
Previous Article in Journal
Guar-Based Injectable Hydrogel for Drug Delivery and In Vitro Bone Cell Growth
Previous Article in Special Issue
Discriminative-Region Multi-Label Classification of Ultra-Widefield Fundus Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-FI: Self-Supervised Learning for Disease Diagnosis in Fundus Images

1
Department of AI Systems Engineering, Sungkyunkwan University, Suwon 16419, Republic of Korea
2
College of Computing and Informatics, Sungkyunkwan University, Suwon 16419, Republic of Korea
3
Sungkyun AI Research Institute, Sungkyunkwan University, Suwon 16419, Republic of Korea
4
Department of Ophthalmology, Kangbuk Samsung Hospital, School of Medicine, Sungkyunkwan University, Suwon 16419, Republic of Korea
5
Biomedical Institute for Convergence, Sungkyunkwan University, Suwon 16419, Republic of Korea
6
Department of Electrical and Computer Engineering, Sungkyunkwan University, Suwon 16419, Republic of Korea
*
Authors to whom correspondence should be addressed.
Bioengineering 2023, 10(9), 1089; https://doi.org/10.3390/bioengineering10091089
Submission received: 14 August 2023 / Revised: 13 September 2023 / Accepted: 14 September 2023 / Published: 16 September 2023
(This article belongs to the Special Issue Biomedical Imaging and Analysis of the Eye)

Abstract

:
Self-supervised learning has been successful in computer vision, and its application to medical imaging has shown great promise. This study proposes a novel self-supervised learning method for medical image classification, specifically targeting ultra-wide-field fundus images (UFI). The proposed method utilizes contrastive learning to pre-train a deep learning model and then fine-tune it with a small set of labeled images. This approach reduces the reliance on labeled data, which is often limited and costly to obtain, and has the potential to improve disease detection in UFI. This method employs two contrastive learning techniques, namely bi-lateral contrastive learning and multi-modality pre-training, to form positive pairs using the data correlation. Bi-lateral learning fuses multiple views of the same patient’s images, and multi-modality pre-training leverages the complementary information between UFI and conventional fundus images (CFI) to form positive pairs. The results show that the proposed contrastive learning method achieves state-of-the-art performance with an area under the receiver operating characteristic curve (AUC) score of 86.96, outperforming other approaches. The findings suggest that self-supervised learning is a promising direction for medical image analysis, with potential applications in various clinical settings.

Graphical Abstract

1. Introduction

Deep learning has established itself as a cornerstone technique in computer vision, driving significant advancements in various applications, including but not limited to object detection, segmentation, and classification [1]. Despite its powerful capabilities, these deep learning models are not without their challenges. They necessitate a substantial quantity of labeled data to deliver a high degree of accuracy [2]. The acquisition of such data, particularly in the domain of medical imaging, can pose a substantial challenge, both in terms of feasibility and associated costs [3]. One potential solution to mitigate this dependency on labeled data is self-supervised learning (SSL). This approach harnesses the inherent structure and context intrinsic to the data to learn meaningful representations [4]. Conceptually, SSL falls under the broader category of unsupervised learning, wherein the model is trained to predict certain aspects of the data based on its own learned representations, thereby effectively reducing the requirement for external annotations [5,6]. SSL methods have already showcased compelling results in the field of natural image processing [7]. Consequently, the prospect of their successful application within medical imaging is a compelling avenue that merits further exploration [8].
Ultra-wide-field fundus images (UFI) and conventional fundus images (CFI) are pivotal diagnostic instruments in the field of ophthalmology, providing indispensable insights into the retina, thereby facilitating the diagnosis and subsequent management of an array of ocular conditions [9,10,11]. UFI images afford an expansive glimpse of the retina, capturing up to 200 degrees, and thus offering a more comprehensive field of view compared to CFI images, which are limited to a range of 30 to 50 degrees of the retina [12,13]. Despite their advantages, diagnosing ocular diseases from UFI images presents several complexities. These challenges stem from the intrinsic intricacies of the images themselves, the requirement of expert personnel for accurate interpretation, and the scarcity of suitably labeled data for machine learning purposes [14]. The evolution of accurate and efficient methodologies for disease diagnosis from UFI images holds immense potential. With such advances, patient outcomes could be markedly improved, aiding in earlier diagnosis, more precise management plans, and, consequently, ultimately enhancing overall prognosis.
Image classification in medical imaging involves the automatic assignment of predefined labels to medical images based on their visual content. In the context of this study, the term “image classification” is used to refer to the automated categorization of medical images, which serves a role analogous to preliminary disease diagnosis. Existing literature highlights the potential of self-supervised learning (SSL) in medical imaging, specifically its role in enhancing the accuracy and minimizing dependence on labeled data. Despite this, the majority of self-supervised learning methods have predominantly focused on natural images, which encompass everyday scenes and objects often used in computer vision research. This has resulted in a research gap pertaining to the application of SSL techniques to the domain of ultra-wide-field fundus images (UFI) in the context of medical imaging [15]. This paper, therefore, embarks on an exploration of SSL performance in the domain of medical image classification using UFI images [16]. Among various SSL approaches, contrastive learning has garnered considerable attention as an efficacious strategy to derive useful representations from unlabeled data [7]. Its core principle rests on maximizing the similarity between various augmented perspectives of an identical image while concurrently reducing the similarity amongst augmented views of disparate images. Accomplishing this necessitates the design and implementation of several loss functions such as InfoNCE loss and NT-Xent loss [17,18]. Recent studies have expanded upon this by investigating the effectiveness of employing various augmentation strategies—including but not limited to cutout, mixup, and random erasing—to enhance the performance of contrastive learning [19]. Furthermore, certain investigations have adopted a task-specific approach, tailoring contrastive learning to particular tasks such as object detection and image segmentation, and yielding promising results in the process [20,21,22]. Overall, these pioneering studies not only validated the effectiveness of contrastive learning as a powerful tool for unsupervised learning but also lay a robust foundation for prospective research in this captivating field.
In this paper, we introduce an SSL strategy that harnesses contrastive learning principles to train a deep learning model for disease classification in ultra-wide-field fundus images (UFI). The proposed methodology, self-FI, employs bi-lateral contrastive learning and multi-modality pre-training to extract the correlation between UFI and conventional fundus images (CFI), illustrated in Figure 1. We pre-train the model on an extensive array of unlabeled UFI and CFI images before fine-tuning it on a smaller, labeled UFI dataset. To evaluate the effectiveness of the proposed method, we utilize a publicly accessible dataset and compare our approach against other prevalent techniques. Our pivotal contributions in this work are multifaceted and are outlined as follows:
  • We demonstrate that our self-supervised learning method outperforms conventional supervised learning in UFI classification, achieving unrivaled results with an AUC score of 86.96. This underlines the potential of self-supervised learning as an effective instrument for medical image analysis, especially in contexts where labeled data are either scarce or prohibitively expensive to procure.
  • We propose a novel contrastive learning technique that employs a new loss function, thereby enhancing the quality of the learned representations. Our approach draws upon both intra- and inter-image correlations to derive a feature space that is optimally suited for UFI classification. We substantiate the effectiveness of our method through a series of experiments and validations, including an examination of learned feature maps and an assessment of different hyperparameters.
  • We offer comprehensive heatmaps of our models for each of the six diseases encompassed in our dataset using GradCAM [23]. This provides an in-depth insight into the model’s focus on each disease. We validate our model by juxtaposing it against a range of other methodologies, including supervised learning and transfer learning from preexisting models. Our proposed model outperforms these comparative methods, emphasizing the crucial role of self-supervised learning in medical image analysis.
  • Our work signifies a significant leap forward in the application of self-supervised learning for disease diagnosis in retinal fundus images. By harnessing the capabilities of contrastive learning and our innovative loss function, we demonstrate the feasibility of achieving state-of-the-art results with a limited quantity of labeled data. Our approach presents a novel pathway for medical practitioners to diagnose patients efficiently and accurately, thereby enhancing patient outcomes and the overall quality of care.
The remainder of this paper is organized as follows: Section 2 provides detail about our dataset and novel methodology, leveraging bi-lateral contrastive learning and multi-modality pre-training. Section 3 presents the results, showcasing our model’s performance compared to other techniques, and includes an analysis of our model’s disease focus via heatmaps and evaluation. Finally, Section 4 encapsulates our discussion and conclusions, reflecting on our findings, the implications of our proposed method, and potential directions for future research in this promising field.

2. Materials and Methods

2.1. Dataset

The dataset used in this study, which was collected from KangBuk Samsung Hospital, comprised ultra-wide-field fundus images (UFI) and conventional fundus images (CFI) from 2 January 2006 to 31 December 2019. To respect patient privacy and confidentiality, all personally identifiable information was removed from the images before use. This study adhered to the tenets of the Declaration of Helsinki, and the protocol was reviewed and approved by the Institutional Review Boards (IRB) of Kangbuk Samsung Hospital (No. KBSMC 2020-01-031-001). This is a retrospective study of medical records, and our data were fully anonymized. Therefore, the IRB waived the requirement for informed consent. Based on the correlations, we formed sub-datasets from the dataset, shown in Figure 2, as follows:
  • UFI unlabeled: The UFI unlabeled subset is a diverse collection of 15,706 ultra-wide-field fundus images, including UFI images from UFI Left–Right and UFI–CFI Pair. Without any specific annotations, these images span across a variety of eye conditions and also include images of healthy eyes. This collection is an invaluable asset for self-supervised and unsupervised learning approaches.
  • UFI Left–Right: Comprising 8382 fundus images of various eye conditions and healthy eyes, the UFI Left–Right pair dataset is specifically assembled to assist tasks that require data in pairs, such as stereo image matching. Each image in this set has a corresponding left and right pair.
  • UFI–CFI: The UFI–CFI pair subset consists of 7166 fundus images, incorporating both ultra-wide-field fundus images (UFI) and conventional fundus images (CFI). Created to explore the benefits of multi-modality pre-training through self-supervised learning, this set spans across all diseases and healthy eye conditions, thereby serving as a rich resource for disease diagnostic research.
  • UFI Labeled: The UFI labeled subset is a curated collection of ultra-wide-field fundus images, annotated by a team of experienced ophthalmologists, and consists a total of 3285 images. Primarily intended for supervised learning approaches that require labeled data for training, these annotations offer vital details about the presence and severity of various eye diseases, including diabetic retinopathy, age-related macular degeneration, and glaucoma.

2.2. Methodology

In this study, we use a two-step methodology to improve the classification performance in retinal disease diagnosis, as shown in Figure 3. The first step is to perform self-supervised pre-training, where the objective is to identify and learn features inherent in medical fundus images. For this, we employ a contrastive learning approach using three unique types of image pairs: pair-instance, bi-lateral, and multi-modality. This step allows the model to capture the general features and structures within fundus images, but without focusing on disease-specific characteristics. Once the model is proficient in recognizing these features, we transition to the second phase, which is supervised fine-tuning. During this phase, the pre-trained model is further refined on a labeled dataset specifically designed for retinal disease diagnosis. This fine-tuning ensures that the model, initialized with features learned from fundus images, becomes adept at identifying and classifying retinal diseases. In essence, the first phase prepares the groundwork through broad feature learning, while the second phase tailors the model for the specific task of disease diagnosis.
Contrastive learning is a self-supervised approach that leverages the inherent structure of data to learn meaningful representations without relying on explicit labels. Central to this paradigm are the concepts of positive and negative pairs. A positive pair typically consists of two augmented versions of the same data instance, implying that they share a common underlying source of information and, therefore, should be close in the learned feature space. Conversely, a negative pair comprises two distinct data instances, suggesting that they come from different sources and should thus be far apart in the feature space. The objective of self-contrastive learning is to pull positive pairs close together while pushing negative pairs apart in a representation space, as shown in Figure 2. To this end, the InfoNCE loss, or noise-contrastive estimation loss, is often employed. InfoNCE loss quantifies the similarity between representations of positive pairs in relation to a set of negative samples. By minimizing this loss, models are trained to recognize and emphasize the differences between positive and negative pairs, ultimately leading to a rich and discriminative feature space that captures the essential characteristics of the data.
In general, the generation of a positive pair is initiated by selecting an individual image u from the dataset. Subsequently, two distinct data augmentations are applied to this image to produce two correlated but unique views u i and u j . These augmented views serve as a positive pair and undergo further processing through a neural network architecture. Initially, both u i and u j are passed through an encoder f, which in our case is implemented using a ResNet architecture. The encoder transforms the input images into high-dimensional feature vectors f ( u i ) and f ( u j ) . To facilitate the subsequent contrastive learning, a projection head g is employed to map these feature vectors into a representation space where contrastive loss is more stable and easier to optimize. Mathematically, the image representation z is obtained as z i = g ( f ( u i ) ) . The projection head g itself is a neural network layer characterized by its weight matrix W. It receives the encoder’s output f ( u i ) and maps it through a ReLU non-linearity ω to produce the final output representation h, given by h = ω ( W · f ( z i ) ) . Figure 4 illustrates the framework of pair-instance, bi-lateral, and multi-modality, where bi-lateral and multi-modality not only use one individual u, but rather in two pairs ( u l u r and u c ) to form positive samples.

2.2.1. Pair-Instance Pre-Training

For pair-instance pre-training, we employ data augmentation methods, similar to SimCLR [18], to create a positive pair. This method focuses on distinguishing between pairs of differently augmented instances derived from the same image. The underlying goal of SimCLR is to amplify the similarity between representations stemming from two unique augmentations of a single image while concurrently minimizing any similarity present among the representations of distinct images. This approach presents a stark contrast to traditional supervised learning methodologies that rely heavily on explicitly provided labels, thereby allowing SimCLR to learn directly from the inherent structure of the data. SimCLR follows a systematic two-stage training regimen, with the first stage dedicated to feature extraction from the input image using a neural network encoder, specifically the ResNet34 [24]. The second stage is oriented towards the transformation of this high-dimensional feature vector into a space of lower dimensionality using a projection head. This component consists of one or more fully connected layers, each bolstered with non-linear activations.
Given an image u, two distinct augmentations u i and u j are applied to generate two correlated, yet distinct, views u i and u j of the same image. The aim is to maximize the similarity between the encoded projections of the positive pair, relative to all other negative pairs in the same batch. First, in the encoding stage, each image u from the positive pair is passed through a neural network base encoder f ( · ) and projection head g ( · ) . The representation results are represented as z i and z j . The similarity metric s i m ( z i , z j ) is defined using the cosine similarity as:
s i m ( z i , z j ) = z i · z j z i 2 × z j 2
where z i · z j represents the dot product of z i and z j , and z i 2 × z j 2 represents the L2 norms (Euclidean norms). The output of the cosine similarity ranges between −1 and 1, where a value close to 1 implies that the vectors are similar, and a value close to −1 implies that the vectors are dissimilar.
The operational pipeline of SimCLR commences with encoding two distinct input images—both being post data augmentation—using an encoder base network denoted as f ( · ) . This step results in the generation of feature representations that are further transformed by a projection head g ( · ) . Central to SimCLR is the presumption that a pair of images created by applying various augmentations to a single source image share a common ground of information and, hence, are considered a positive pair. However, this approach does not come without limitations. The acquisition of knowledge remains confined to a single image, thus restricting the model’s ability to harness wider contextual understanding. To address this constraint, we propose two novel strategies for contrastive pre-training: bi-lateral pre-training and multi-modality pre-training. The former entails training with left–right eye images from ultra-wide-field fundus (UFI) images, thereby acknowledging the inherent symmetry in anatomical structures and the potential commonalities in pathological occurrences. In contrast, the latter exploits the combination of UFI and conventional fundus Images (CFI) for training, leveraging the differing perspectives and resolutions of these image types.

2.2.2. Bi-Lateral Pre-Training

For fundus images, it is common to have a pair of two images from the same patient, capturing the left and right body parts of the patient. These images are presumed to encapsulate identical patient-specific information, albeit presenting differing views and image structures. This characteristic proves advantageous for the model, as it facilitates learning of representations from multiple angles and viewpoints, thus enhancing its understanding of image patterns. This is the basis of our bi-lateral pre-training, where we leverage the left and right eye images from ultra-wide-field fundus imaging, treating them as a positive pair. The nuanced differences between these paired images equip the model with a broader perspective and greater feature comprehension, which ultimately refines its performance in subsequent disease classification tasks. The inherent differences between these paired images, which span distinct angles and capture varied structural intricacies, imbue the model with a broader perspective. It exposes the model to diverse and complex feature variations, which contribute towards its ability to understand and interpret patterns more effectively. By incorporating this information from paired images into its learning process, the model is equipped with a more nuanced understanding of the data. This understanding extends beyond what could be derived from an individual image, thereby enabling the model to better distinguish between normal and pathological conditions.
Suppose u l and u r represent the left and right eye images, respectively, from ultra-wide-field fundus (UFI) imaging of a given patient. The aim of the bi-lateral pre-training is to maximize the similarity between these two representations. This can be formulated as a contrastive loss function:
L B i l a t e r a l = log e x p ( s i m ( z l , z r / τ ) k = 1 K e x p ( s i m ( z l , z k ) / τ )
where f ( · ) is the encoder network that maps an image to a feature representation, s i m ( · ) is a similarity function such as cosine similarity, K is the total number of negative samples, and τ is a temperature parameter that adjusts the concentration of the distribution. In this setup, the encoder network f ( · ) is trained to extract features such that the representations of the left and right eye images ( z l and z r ) are similar. This is achieved by minimizing the contrastive loss L B i l a t e r a l . For a given positive pair ( u l , u r ) in the dataset, the loss aims to increase the similarity between z l and z r while reducing the similarity between z l and z k , where z k is a negative sample that is not related to z l . The objective is to learn a robust representation that captures patient-specific information shared between left and right eye images while distinguishing between different patients.
Bi-lateral pre-training is designed to augment the model’s grasp of the unique attributes present within ultra-wide-field fundus images (UFI). By presenting the model with paired left–right eye images, we aim to empower it with a broader perspective that captures intricate variations within the images. This technique enriches the model’s ability to discern nuanced textures, patterns, and structural elements, which in turn contribute to a more comprehensive understanding of the dataset. Although the immediate focus is on image comprehension, the gained insights indirectly enhance the model’s capacity for accurate disease classification during subsequent fine-tuning.

2.2.3. Multi-Modality Pre-Training

For the dual-modality pre-training, we introduce another contrastive learning strategy that capitalizes on the combination of UFI and CFI. Each type of imaging provides unique and complementary information: whereas UFIs offer a wider field of view encompassing peripheral regions, CFIs offer a focused, detailed visualization of the vasculature. Suppose U and C are the UFI and CFI images, respectively, of the same patient. The objective of the dual-modality pre-training is to maximize the similarity between these two representations. This can be expressed mathematically using the contrastive loss function:
L m u l t i m o d a l i t y = log e x p ( s i m ( z U , z C / τ ) k = 1 K e x p ( s i m ( z U , z k / τ )
In this formulation, f ( · ) is again the encoder network mapping an image to a feature representation, s i m ( · ) is the similarity function such as cosine similarity, K is the total number of negative samples, and τ is the temperature parameter. In this scenario, we train the encoder network f ( · ) to extract features such that the representations of the UFI and CFI images ( z U and z C ) are similar. The contrastive loss L m u l t i m o d a l i t y is minimized during training. The idea is to learn a robust representation that effectively fuses the complementary information from UFI and CFI images, thereby enhancing the model’s overall understanding and classification capabilities.
The rationale behind training UFI and CFI together lies in the recognition that each image modality holds distinct attributes that complement one another. UFI images encapsulate a panoramic view of the retina, revealing features that may extend beyond the scope of conventional images. Conversely, CFI images, although limited in their field of view, offer high-resolution details that might not be as prevalent in UFI images. By training the model on this diverse combination, we aim to exploit the unique advantages of each modality to provide a comprehensive understanding of retinal structures. The multi-modality approach enriches the model’s ability to capture the intricate variations and patterns that characterize retinal images, which subsequently contributes to more effective disease classification during the fine-tuning phase.
Dual-modality pre-training fundamentally relies on the premise that the UFI and CFI images, despite their differing scopes, contain overlapping and mutually informative data about a patient’s condition. By maximizing the similarity between the representations of UFI and CFI images, the model can take advantage of these shared insights, ultimately enhancing the versatility and robustness of the learned representations. Moreover, this approach benefits from the distinct attributes of UFI and CFI images. The UFI’s wide-field view is particularly useful in diseases where peripheral retinal information is crucial. On the other hand, CFI’s high-resolution imaging of the central retina is beneficial for diseases predominantly affecting the central field. By coupling these two modalities, the model has access to a more holistic understanding of the patient’s condition, broadening its diagnostic capabilities. The effectiveness of dual-modality pre-training is reliant on the careful selection of the similarity function, s i m ( · ) , and the optimization of the temperature parameter, τ . These choices are pivotal in maintaining the balance between the model’s ability to distinguish the informative patterns in the data and its capacity to avoid overfitting. Furthermore, the selection of negative samples in the denominator of the contrastive loss function is a key factor influencing the model’s performance. In the context of dual-modality pre-training, negative samples should ideally be chosen to encourage the model to distinguish between different patients, rather than merely different imaging modalities.

2.2.4. Self-FI: Self-Supervised Learning Model on Retinal Fundus Images for Disease Diagnosis

Our proposed methodology harnesses the power of three distinctive pre-training approaches: pair-instance, bi-lateral, and multi-modality pre-training, as shown in Figure 5. This unique amalgamation of techniques serves to learn high-quality representations from a diverse, unlabeled dataset of UFI and CFI images. In the SimCLR pre-training scheme, a pair of augmented views derived from the same image forms the “positive” pair. Bi-lateral pre-training further enriches the model’s understanding by treating each pair of left–right UFI images as a “positive” instance, thus allowing the model to appreciate the inherent similarities in bi-lateral anatomical structures. To exploit the complementary information embedded in different imaging modalities, we design multi-modality pre-training, where a “positive” pair comprises a UFI and a CFI image from the same patient. Contrastive learning thrives not just on the identification of “positive” pairs, but also on the distinction from “negative” pairs. To this end, for every “positive” pair, we introduce a “negative” pair composed of an image randomly selected from the remainder of the dataset. This forces the model to discern between images exhibiting similar and dissimilar features. The success of our method hinges on a novel loss function that amalgamates the losses from each of the pre-training strategies. The final loss, L t o t a l , is the sum of the losses from SimCLR, bi-lateral, and multi-modality pre-training, each weighted by a coefficient optimized through rigorous hyperparameter tuning. This composite loss function is represented as:
L t o t a l = L p a i r i n s t a n c e + L b i l a t e r a l + L m u l t i m o d a l i t y
The method is summarized in Algorithm 1.
Algorithm 1: Proposed contrastive learning methods.
Bioengineering 10 01089 i001

2.3. Implementation Settings

To assess the effectiveness of our proposed contrastive learning approaches, we deploy a ResNet34 neural network as our foundational model, denoted by f ( · ) , paired with a projection head g ( · ) . This projection head comprises a fully connected layer sequence: Linear → ReLu → Linear. This sequence maps the high-dimensional image feature vectors, generated by f ( · ) , into a lower-dimensional feature space. For model training, we employ the AdamW optimizer with a learning rate of 0.00005, weight decay of 0.0001, and temperature of 0.07. The model is trained over 3000 epochs with a batch size of 128 and a 128-dimensional embedding. In the pre-training phase, images are resized to 224 × 224 pixels, and a range of augmentation techniques are used to yield positive pairs. These techniques encompass horizontal and vertical flipping, random cropping, rotation (up to 180 degrees), color adjustment (brightness, contrast, saturation, hue), grayscale application, and Gaussian blurring. Each pre-training sub-dataset is divided into 80:20 for training and validating, respectively. We utilize the SimCLR framework for pair-instance pre-training, bi-lateral pre-training, and multi-modality pre-training. The implementation of our proposed method is constructed using Pytorch [25] and is executed on four GeForce RTX 2080 Ti GPUs. These GPUs afford the necessary computational resources to manage the expansive dataset and the complexity of the model. The GPUs’ parallel processing capability enables us to train our model effectively and expedite pre-training using the proposed methods. Once the foundational model is trained, we leverage it to fine-tune the downstream multi-label classification task. We retain the pre-trained weights of the ResNet34 while substituting the projection head with a fully connected layer. This layer yields an output vector with a size corresponding to the number of classes represented in the dataset. For this phase, we train the model for 100 epochs, with a batch size of 64. The model is optimized using the AdamW optimizer with a learning rate of 0.001 and a weight decay of 0.001. To mitigate overfitting, we incorporate a dropout rate of 0.5 into the fully connected layer. During the fine-tuning process, we resort to a larger image size of 512 × 512 pixels, granting the model access to a richer set of details. Of those, 2629 images are used for training, 492 images are used for validating, and 164 images are used for testing. There is no overlap or mixing of data between the training and validation sets; all data in the validation set came from individuals who were entirely distinct from those in the training set. The training set was not exposed to the algorithm prior to the actual training process and stayed unprocessed and separated to ensure the model’s validity. In a bid to boost the model’s robustness, we engaged a suite of random augmentation techniques, such as horizontal/vertical flip, random cropping, rotation (up to 180 degrees), color adjustment (brightness, contrast, saturation, hue), grayscale application, and Gaussian blurring. Using the PyTorch framework, we applied on-the-fly data augmentation during each batch processing. This involves dynamically generating augmented images through transformations such as random rotations, flips, and color adjustments. These augmented images increase the diversity of the samples fed to the model during training. Training of the model is performed on an NVIDIA GeForce RTX 3080Ti GPU, which ensures swift training times because of its impressive memory and computational capacity.

2.4. Measurements Metrics

AUC Score: The area under the receiver operating characteristic curve (AUC) is a prevalent metric employed to evaluate binary classification models. It quantifies the model’s aptitude in differentiating between positive and negative examples across an array of thresholds. An ideal classifier commands an AUC of 1, whereas a random classifier holds an AUC of 0.5. The mathematical representation for AUC is as follows:
A U C = 1 2 i = 1 n 1 ( x i + 1 x i ) ( y i + y i + 1 )
where n is the total number of thresholds, and x i and y i correspond to the false positive rate and true positive rate, respectively, at the ith threshold.
Accuracy: Accuracy is a commonly deployed metric to assess classification models. It computes the ratio of accurately classified samples to the overall count of samples. The formula for accuracy is as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
where T P represents the count of true positives, T N stands for the count of true negatives, F P is the count of false positives, and F N refers to the count of false negatives.
F1 Score: The F1 score is an indicator of the equilibrium between precision and recall for binary classification models. It is the harmonic mean of precision and recall, and ranges from 0 (worst case) to 1 (best case). The formula for the F1 score is as follows:
F 1 = 2 · p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l
where p r e c i s i o n is the ratio of true positives to the sum of true positives and false positives, and r e c a l l is the ratio of true positives to the sum of true positives and false negatives.
Precision and Recall: Precision and recall are two popular metrics employed for evaluating binary classification models. Precision computes the proportion of true positives out of the total count of predicted positives, and recall calculates the proportion of true positives out of the total count of actual positives. The formulas for precision and recall are given below:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where T P denotes the count of true positives, F P is the count of false positives, and F N refers to the count of false negatives.

3. Results

3.1. Comparison with Other Methods

Table 1 illustrates the performance of the supervised baseline along with three self-supervised learning methods in detecting six prevalent diseases in UFI and CFI images. The self-supervised approaches exhibit superior mean AUC scores compared to the supervised baseline, underscoring the efficacy of self-supervised learning in medical image classification. Notably, our proposed technique yielded the highest AUC score of 86.96, outstripping the supervised baseline by 1.06%. This suggests that our technique is more proficient in identifying UFI diseases from fundus images. In addition, our method secured the best accuracy of 89.50%, F1 score of 62.51%, precision of 71.80%, and recall of 64.54% among all examined methods. These results extol the superior performance of our proposed approach compared to alternative self-supervised learning methods and the supervised baseline. The increased accuracy, F1 score, precision, and recall denote that our approach can discern diseases more precisely and dependably than other methods, a critical aspect for early disease diagnosis and effective treatment planning.
In Table 1, it is notable that the AUC score provides a measure of a model’s ability to distinguish between classes across various decision thresholds and is generally less affected by the class imbalance. In contrast, F1 score, precision, and recall are sensitive to the specific decision threshold applied and can vary based on the class distributions. Each instance in our dataset could belong to multiple classes, potentially leading to the variance in these metrics. The F1 score, precision, and recall metrics are influenced by the varying number of positive and negative samples for different classes. Even if a model has a high AUC or accuracy, these metrics can differ based on how the model performs on individual classes. Different models like no pre-train, supervised, SimCLR, Barlow Twins, and our proposed method may exhibit varying sensitivities to these factors. Understanding these nuances offers a more comprehensive view of the model’s overall performance and accounts for the observed disparities among the metrics.
Table 2 presents the mean AUC scores for different diseases achieved by various methods, including no pre-train, bi-lateral, multi-modality, and our proposed method. Our method consistently outperformed the other approaches across most diseases. Specifically, our method achieved the highest AUC scores for DM retinopathy (93.36), epiretinal membrane (84.73), and glaucoma suspect (97.29), demonstrating its effectiveness in accurately detecting these conditions. In addition, our method obtained competitive AUC scores for macular degeneration (87.39), retinal break (93.08), and retinal vein occlusion (69.49) compared to the other methods. These results highlight the accuracy of our proposed method in image classification, particularly for DM retinopathy and glaucoma suspects. The high AUC scores obtained by our method underscore its potential for improving the diagnosis and treatment of retinal diseases, contributing to enhanced patient care and outcomes.
The heatmaps generated by deep learning models can provide valuable insights into the areas of fundus images that are most indicative of a particular disease. In our study, we generated heatmaps for each of the six diseases in UFI images using our proposed contrastive learning method and compared them to those generated by a supervised learning method. Figure 6 demonstrates a compelling visual representation of the capabilities of our proposed model in comparison to conventional methods. Notably, the yellow boxes in the figure demarcate the areas correctly identified by our deep learning model as being afflicted by disease. These correctly identified regions reflect our model’s superior ability to interpret and understand the underlying features of fundus images, enabling it to pinpoint disease-afflicted areas with remarkable accuracy. On the contrary, the red boxes in the figure indicate the areas that were incorrectly identified by the traditional supervised learning model. This clear distinction between correctly and incorrectly identified regions reiterates the shortcomings of the supervised learning models, particularly in comparison to the advanced capabilities of our proposed model. In our study, we utilized distinct methods for visualization and interpretability of the learned representations, specifically employing Grad-CAM for the models utilizing InfoNCE loss and Score-CAM for the Barlow Twins model, since Barlow Twins uses a different loss function than the other methods (Barlow Twins loss). Our deep learning model exhibits a clear capacity to identify all disease-afflicted areas of the fundus image, a feat which conventional supervised learning models fell short of achieving. A detailed evaluation of our results reveals that self-supervised learning methods, and, more specifically, our unique contrastive learning method, generate heatmaps that exhibit an enhanced degree of accuracy, zeroing in on the disease-related regions with unparalleled focus when compared to the heatmaps generated by supervised learning methods.
The performance of different models for classifying diseases was evaluated using receiver operating characteristic (ROC) curves, as shown in Figure 7. The ROC curves provide an effective way to visualize the trade-offs between the true positive rate (sensitivity) and the false positive rate (1-specificity) across different threshold settings. Each curve represents a different model’s ability to discriminate between positive and negative instances for each disease class. Models that closely follow the upper left border of the ROC space indicate a better performance, whereas models that lie closer to the diagonal line reflect less discriminative power. The area under the curve (AUC) values further quantify the overall ability of each model to distinguish between the classes, with higher AUC values suggesting better model performance. In addition to the ROC curves, the performance of our proposed model was also evaluated using confusion matrices for every disease class, shown in Figure 8. A confusion matrix offers a more granular view of the model’s performance, breaking it down into true positives, true negatives, false positives, and false negatives. This allows for the calculation of various metrics, such as sensitivity, specificity, and F1-score. All of the models exhibit high true positive rates across multiple classes, indicating good sensitivity in detecting diseases. At the same time, the true negative rates are substantial, reflecting that the models are adept at correctly identifying non-disease instances. It is worth noting that the total number of instances in each confusion matrix remains constant across different classes, as each instance may belong to multiple classes in our multi-label, multi-class setting.
Figure 9 presents the training and validation loss curves for five different models: random initialization, supervised learning, SimCLR, Barlow Twins, and our proposed method. Each curve represents the model’s loss over a series of training epochs, offering insight into the model’s learning efficiency and its ability to generalize. Most notably, the curve for our proposed method consistently exhibits the lowest loss values across the epochs and demonstrates a continuous downward trend, indicating optimal learning and generalization capabilities. This performance underscores the efficiency of our method in minimizing the loss function compared to other established techniques. In particular, the declining nature of the loss curve for our proposed method suggests that the model is effectively learning the underlying patterns in the data without signs of overfitting. This outcome is further supported by our use of weight decay as a regularization technique and periodic validation every iteration.

3.2. Ablation Study

Table 3 shows a comparative analysis of the performance of different self-supervised learning methods in disease detection. We utilized several metrics for this comparison, namely the area under the curve (AUC) score, accuracy, F1 score, precision, and recall. Starting with the SimCLR method, it yielded an AUC score of 85.09, an accuracy of 88.25%, an F1 score of 57.42, precision of 67.70, and a recall of 64.18. Bi-lateral learning showed a slightly better AUC score of 85.47, although its accuracy was slightly lower at 88.13%. The F1 score for bi-lateral was 57.16, with a precision of 63.18 and recall of 63.93. The multi-modal learning method presented a more significant improvement, with an AUC score of 85.90, accuracy of 89.13%, an F1 score of 61.54, precision of 67.99, and a recall of 64.19. However, our proposed method clearly outperformed the others. It achieved the highest AUC score of 86.96, an accuracy of 89.50%, F1 score of 62.51, precision of 71.80, and a recall of 64.54. These results provide empirical support to our claim that our proposed self-supervised learning model offers a substantial improvement over existing methods across multiple performance metrics in classification.
To examine the impact of various self-supervised method combinations on image classification efficiency, we conducted an ablation study, integrating distinct self-supervised methods. Each combination is trained with the same settings and environment as our proposed method, but with different datasets and loss. Specifically, method 1 + 2 uses L p a i r i n s t a n c e + L B i l a t e r a l , method 2 + 3 uses L B i l a t e r a l + L m u l t i m o d a l i t y , and method 1 + 3 uses L p a i r i n s t a n c e + L m u l t i m o d a l i t y . Table 4 encapsulates the performance of these combinations in terms of mean AUC, accuracy, F1 score, precision, and recall. Our study demonstrated that the triumvirate of pair-instance, bi-lateral, and multi-modality procured the superior mean AUC score of 86.96. This performance slightly edged out the amalgamation of SimCLR and bi-lateral (86.75) as well as the bi-lateral and multi-modality ensemble (86.51). However, the fusion of bi-lateral and multi-modality secured the leading F1 score of 63.06, demonstrating an optimal balance between precision and recall. Furthermore, the pair-instance and bi-lateral combination achieved the highest recall score of 67.88, signifying an enhanced capacity to detect positive instances. The result of this exploration suggests that blending multiple self-supervised methods can bolster the performance of the model. Furthermore, these combinations could establish diverse trade-offs among different evaluation metrics. This informs us that choosing the right blend of methods may be application-specific, whereby the optimal combination may differ based on the diagnostic criteria of the disease or the medical scenario at hand. For example, in cases where a high recall is imperative to avoid missing any potential positive cases (in conditions with severe consequences or when early diagnosis significantly improves prognosis), the pair-instance with bi-lateral combination may be most suitable. Conversely, when it is crucial to have a balanced model in terms of both recall and precision (to avoid an excessive number of false alarms or when resources for further testing are limited), the bi-lateral and multi-modality pair could be the preferred choice. Such insights, derived from our ablation study, are instrumental in tailoring self-supervised learning models to particular disease diagnosis tasks, thereby advancing precision medicine and contributing to better patient outcomes.

3.3. Contribution of Each Pairing Method on the Improvement of the Performance

To quantify the individual improvement brought by each method, we subtract the baseline performance metrics (termed as “no pre-train”) from the metrics achieved when the respective method was applied. Formally, for a given performance metric M and a pre-training method M i , the individual improvement I i is calculated as:
I i = M i M No Pre - train
This calculates the net increase in performance solely due to the implementation of a particular pre-training strategy. Subsequently, we determine the incremental benefit provided by a method when it is amalgamated with another. This additional improvement A i j for a pair of methods M i and M j is computed as:
A i j = M i + j M j
This measures the extra performance enhancement due to the synergistic interaction of the two methods. For instance, when calculating the improvement introduced by the pair-instance method in conjunction with the bi-lateral method, the additional improvement is obtained by subtracting the performance of the bi-lateral method alone from the performance delivered by the combination of the pair-instance and bi-lateral methods. This calculation manifests the additional contribution the pair-instance method proffers when paired with the bi-lateral method. We repeat this process for all pairs and average the additional improvements. The total contribution C i of a method is then computed as the sum of its individual improvement and the mean of its additional improvements when combined with the other methods:
C i = I i + 1 n j A i j
This formula ensures that we capture the total contribution of each method, considering both its standalone impact and the augmentations it imparts when coupled with others. Figure 10 shows a detailed comparative analysis of the performance improvements delivered by our three individual self-supervised learning methods—pair-instance, bi-lateral, and multi-modality—against a baseline model with no pre-training. The performance improvements were calculated based on five evaluation metrics: AUC score, accuracy, F1 score, precision, and recall. To ascertain the performance improvement for each method, we subtracted the metric score of the baseline model from the score achieved by each method. The result is a measure of the performance gain attributable to the incorporation of each self-supervised learning method. The bar graph in the figure visually showcases the performance enhancements for each method. As per the numerical values indicated, the multi-modality method registers the highest improvement across all performance metrics, implying that this method is the most beneficial of the three. It is closely followed by the bi-lateral method, which demonstrates substantial improvements as well, albeit slightly less than the multi-modality approach. The pair-instance method also outperforms the baseline, but its gains are comparatively modest. These results reinforce the merit of our proposed self-supervised learning strategies in enhancing the accuracy of deep learning models, surpassing the non-pretrained model in all respects. The bar graph in the figure visually showcases the performance enhancements for each method. Significantly, the multi-modality method registers the highest improvement across all performance metrics. This dominant contribution underscores the importance of correlating multiple modalities within medical images, emphasizing that the interplay between different image forms is vital in capturing intricate and critical features that enhance diagnostic accuracy. The bi-lateral method, though marginally surpassed by the multi-modality approach, also demonstrates substantial improvements. It leverages the left and right perspectives of anatomical structures within medical imaging, allowing the model to grasp nuanced differences and gain a holistic understanding. Such bi-lateral insights into the data structure enable the model to recognize intricate patterns that would otherwise be elusive, thereby significantly boosting its performance.

4. Discussion and Conclusions

In conclusion, we introduced a contrastive learning approach for pre-training deep neural networks in the context of retinal disease diagnosis. Our proposed approach leveraged the power of three different types of image pairs, including pair-instance, bi-lateral, and multi-modality, to create a diverse and comprehensive set of positive and negative pairs for contrastive learning. Our experimental results demonstrated that our method surpassed the performance of the supervised learning baseline, achieving state-of-the-art results across four major retinal disease diagnosis tasks. These findings underscore the effectiveness of self-supervised learning in the field of retinal imaging, highlighting its potential for advancing early diagnosis and treatment strategies.
The strength of our findings lies in their potential to revolutionize early diagnosis and treatment planning for retinal diseases. The application of self-supervised learning, as demonstrated in our study, brings a new dimension to the field by addressing challenges such as limited labeled data and manual annotation effort. This has profound implications for not only improving diagnostic accuracy but also for reducing the cost and time involved in retinal disease management. In medical fields where early diagnosis can dramatically impact patient outcomes, any advancement in accuracy is particularly crucial. Our method’s enhanced performance could potentially lead to more effective treatments and interventions, ultimately improving patient prognosis and quality of life.
Furthermore, our work underscores the potential of self-supervised learning to revolutionize medical imaging, offering a flexible and data-efficient alternative to traditional supervised approaches. We acknowledge that ongoing research is essential to explore the method’s generalization potential across diverse clinical scenarios and diverse datasets. Our study also opens avenues for incorporating additional modalities and clinical data, deepening the synergy between machine learning and medical expertise. Although our method shows promise, it is not without limitations.
The study was focused on a specific subset of retinal diseases, and the algorithms have yet to be validated on a diverse demographic or against rare retinal conditions. Another limitation of this study is the relatively small size of the dataset used for training and validating the models. The paucity of data may not capture the full complexity and variability associated with medical fundus images and the diseases they are meant to help diagnose. This limitation may affect the generalizability of our model to a broader, more diverse population. Future research could benefit from utilizing larger and more diverse datasets to validate the effectiveness of the proposed algorithm and to potentially improve its diagnostic capabilities. Furthermore, a comprehensive hyperparameter optimization was not conducted. Future work will focus on a more rigorous optimization of the model’s hyperparameters, employing techniques such as grid search or Bayesian optimization.
Moving forward, our future work will explore the impact of incorporating different modalities and data sources in our proposed method. For instance, alongside fundus images, we aim to incorporate additional imaging modalities such as optical coherence tomography (OCT) and angiography to create multi-modal image pairs [27,28]. Furthermore, we are interested in investigating the integration of other modalities, such as clinical data and genetic information, in conjunction with retinal images to enhance the model’s accuracy [29]. Additionally, we plan to assess the generalizability of our method to other medical imaging domains beyond retinal disease diagnosis [30]. As deep learning models become more complex, there is also a need to develop efficient and scalable self-supervised learning methods capable of handling large-scale medical image datasets [31]. We anticipate that our work will inspire future research endeavors in this direction, ultimately contributing to improved disease diagnosis and treatment outcomes.

Author Contributions

T.D.N.: conceptualization, methodology, writing; D.-T.L.: validation, formal analysis, writing—review and editing, supervision; J.B.: writing—review; S.K.: data curation; S.J.S.: conceptualization, writing—review, data curation; H.C.: writing—review and editing, supervision, funding acquisition, project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Institute of Information & Communications Technology Planning & Evaluation IITP grant funded by the Korea government (MSIT) under the ICT Creative Consilience program (IITP-2022-2020-0-0182), Artificial Intelligence Graduate School Program (Sungkyunkwan University) (No.2019-0-00421), Artificial Intelligence Innovation Hub (No.2021-0-02068), and KBSMC-SKKU Future Clinical Convergence Academic Research Program, Kangbuk Samsung Hospital & Sungkyunkwan University, 2023.

Institutional Review Board Statement

We note that this study adhered to the tenets of the Declaration of Helsinki, and the protocol was reviewed and approved by the Institutional Review Boards (IRB) of Kangbuk Samsung Hospital (No. KBSMC 2020-01-031-001). This is a retrospective study of medical records, and our data were fully anonymized. Therefore, the IRB waived the requirement for informed consent.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request (data availability is decided by the IRB of Kangbuk Samsung Hospital on each request).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  2. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 843–852. [Google Scholar]
  3. Chartrand, G.; Cheng, P.M.; Vorontsov, E.; Drozdzal, M.; Turcotte, S.; Pal, C.J.; Kadoury, S.; Tang, A. Deep learning: A primer for radiologists. Radiographics 2017, 37, 2113–2131. [Google Scholar] [CrossRef] [PubMed]
  4. Jing, L.; Tian, Y. Self-supervised visual feature learning with deep neural networks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4037–4058. [Google Scholar] [CrossRef] [PubMed]
  5. Zhu, J.; Li, Y.; Ding, L.; Zhou, S.K. Aggregative Self-supervised Feature Learning from Limited Medical Images. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Proceedings of the 25th International Conference, Singapore, 18–22 September 2022; Proceedings, Part VIII; Springer Nature: Cham, Switzerland, 2006; pp. 57–66. [Google Scholar]
  6. Koohbanani, N.A.; Unnikrishnan, B.; Khurram, S.A.; Krishnaswamy, P.; Rajpoot, N. Self-path: Self-supervision for classification of pathology images with limited annotations. IEEE Trans. Med. Imaging 2021, 40, 2845–2856. [Google Scholar] [CrossRef] [PubMed]
  7. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  8. Xu, J. A review of self-supervised learning methods in the field of medical image analysis. Int. J. Image, Graph. Signal Process. (IJIGSP) 2021, 13, 33–46. [Google Scholar] [CrossRef]
  9. Ahn, S.; Pham, Q.T.; Shin, J.; Song, S.J. Future image synthesis for diabetic retinopathy based on the lesion occurrence probability. Electronics 2021, 10, 726. [Google Scholar] [CrossRef]
  10. Lee, J.; Lee, J.; Cho, S.; Song, J.; Lee, M.; Kim, S.H.; Lee, J.Y.; Shin, D.H.; Kim, J.M.; Bae, J.H.; et al. Development of decision support software for deep learning-based automated retinal disease screening using relatively limited fundus photograph data. Electronics 2021, 10, 163. [Google Scholar] [CrossRef]
  11. Tan, C.S.; Chew, M.C.; van Hemert, J.; Singer, M.A.; Bell, D.; Sadda, S.R. Measuring the precise area of peripheral retinal non-perfusion using ultra-widefield imaging and its correlation with the ischaemic index. Br. J. Ophthalmol. 2016, 100, 235–239. [Google Scholar] [CrossRef] [PubMed]
  12. Witmer, M.T.; Parlitsis, G.; Patel, S.; Kiss, S. Comparison of ultra-widefield fluorescein angiography with the Heidelberg Spectralis® noncontact ultra-widefield module versus the Optos® Optomap®. Clin. Ophthalmol. 2013, 7, 389–394. [Google Scholar] [CrossRef] [PubMed]
  13. Silva, P.S.; Cavallerano, J.D.; Haddad, N.M.; Kwak, H.; Dyer, K.H.; Omar, A.F.; Shikari, H.; Aiello, L.M.; Sun, J.K.; Aiello, L.P. Peripheral lesions identified on ultrawide field imaging predict increased risk of diabetic retinopathy progression over 4 years. Ophthalmology 2015, 122, 949–956. [Google Scholar] [CrossRef] [PubMed]
  14. Ting, D.S.W.; Pasquale, L.R.; Peng, L.; Campbell, J.P.; Lee, A.Y.; Raman, R.; Tan, G.S.W.; Schmetterer, L.; Keane, P.A.; Wong, T.Y. Artificial intelligence and deep learning in ophthalmology. Br. J. Ophthalmol. 2019, 103, 167–175. [Google Scholar] [CrossRef] [PubMed]
  15. Zhou, Q.; Zou, H.; Wang, Z. Long-Tailed Multi-label Retinal Diseases Recognition via Relational Learning and Knowledge Distillation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Proceedings of the 25th International Conference, Singapore, 18–22 September 2022; Proceedings, Part II; Springer Nature: Cham, Switzerland, 2022; pp. 709–718. [Google Scholar]
  16. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; He, Q. A comprehensive survey on transfer learning. Proc. IEEE 2019, 109, 43–76. [Google Scholar] [CrossRef]
  17. Oord, A.V.D.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  18. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020. [Google Scholar]
  19. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation strategies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 113–123. [Google Scholar]
  20. Jiang, Z.; Zhang, H.; Wang, Y.; Ko, S.B. Retinal blood vessel segmentation using fully convolutional network with transfer learning. Comput. Med. Imaging Graph. 2018, 68, 1–15. [Google Scholar] [CrossRef] [PubMed]
  21. Mahapatra, D.; Ge, Z. Training data independent image registration with gans using transfer learning and segmentation information. In Proceedings of the IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 709–713. [Google Scholar]
  22. Oh, K.; Kang, H.M.; Leem, D.; Lee, H.; Seo, K.Y.; Yoon, S. Early detection of diabetic retinopathy based on deep learning and ultra-wide-field fundus images. Sci. Rep. 2021, 11, 1897. [Google Scholar] [CrossRef] [PubMed]
  23. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  25. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Chintala, S. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Nice, France, 2019; Volume 32, pp. 8024–8035. Available online: http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf (accessed on 13 August 2023).
  26. Zbontar, J.; Jing, L.; Misra, I.; LeCun, Y.; Deny, S. Barlow twins: Self-supervised learning via redundancy reduction. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 12310–12320. [Google Scholar]
  27. Sadda, S.R.; Wu, Z.; Walsh, A.C.; Richine, L.; Dougall, J.; Cortez, R.; LaBree, L.D. Errors in retinal thickness measurements obtained by optical coherence tomography. Ophthalmology 2006, 113, 285–293. [Google Scholar] [CrossRef] [PubMed]
  28. DeBuc, D.C. A review on the structural and functional characterization of the retina with optical coherence tomography. Eye Contact Lens Sci. Clin. Pract. 2012, 38, 272. [Google Scholar]
  29. Ritchie, M.D.; Holzinger, E.R.; Li, R.; Pendergrass, S.A.; Kim, D. Methods of integrating data to uncover genotype—Phenotype interactions. Nat. Rev. Genet. 2015, 16, 85–97. [Google Scholar] [CrossRef] [PubMed]
  30. Shen, D.; Wu, G.; Suk, H.I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed]
  31. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Implicit data correlation between fundus images. Bi-lateral and multi-modality pairs are positive samples, and images from different patients are considered as negative samples.
Figure 1. Implicit data correlation between fundus images. Bi-lateral and multi-modality pairs are positive samples, and images from different patients are considered as negative samples.
Bioengineering 10 01089 g001
Figure 2. Summarization the of pre-training and fine-tuning dataset. DM: DM retinopathy, EM: epiretinal membrane, GS: glaucoma suspect, MD: macular degeneration, RB: retinal break, RVO: retinal vein occlusion.
Figure 2. Summarization the of pre-training and fine-tuning dataset. DM: DM retinopathy, EM: epiretinal membrane, GS: glaucoma suspect, MD: macular degeneration, RB: retinal break, RVO: retinal vein occlusion.
Bioengineering 10 01089 g002
Figure 3. The general process of classification process on fundus image using contrastive pre-training. Positive and negative pairs are used to pre-train the model backbone, then the backbone is used for fine-tuning with labeled data.
Figure 3. The general process of classification process on fundus image using contrastive pre-training. Positive and negative pairs are used to pre-train the model backbone, then the backbone is used for fine-tuning with labeled data.
Bioengineering 10 01089 g003
Figure 4. The framework of contrastive learning for pair-instance, bi-lateral, and multi-modality pre-training. The images are augmented to form positive views (for pair-instance) and new augmented view (for bi-lateral and multi-modality). The encoder f and projection head g try to maximize the agreement using contrastive loss); x x ˙ / x ^ denotes the data augmentation step.
Figure 4. The framework of contrastive learning for pair-instance, bi-lateral, and multi-modality pre-training. The images are augmented to form positive views (for pair-instance) and new augmented view (for bi-lateral and multi-modality). The encoder f and projection head g try to maximize the agreement using contrastive loss); x x ˙ / x ^ denotes the data augmentation step.
Bioengineering 10 01089 g004
Figure 5. Overview of our proposed multi-modality contrastive learning method. The model is pre-trained on a combination of UFI and CFI using pair-instance, bi-lateral, and multi-modality techniques.
Figure 5. Overview of our proposed multi-modality contrastive learning method. The model is pre-trained on a combination of UFI and CFI using pair-instance, bi-lateral, and multi-modality techniques.
Bioengineering 10 01089 g005
Figure 6. Side-by-side comparison of the models’ heatmaps, including traditional supervised learning and our proposed contrastive learning model. Yellow boxes represent accurately identified disease-afflicted areas, demonstrating the precision of our model. Conversely, red boxes indicate the areas incorrectly identified by the traditional model.
Figure 6. Side-by-side comparison of the models’ heatmaps, including traditional supervised learning and our proposed contrastive learning model. Yellow boxes represent accurately identified disease-afflicted areas, demonstrating the precision of our model. Conversely, red boxes indicate the areas incorrectly identified by the traditional model.
Bioengineering 10 01089 g006
Figure 7. ROC curves of different models for each disease. The dashed line in the middle indicates a random classifier. The curve that is further to the left corner indicates a better model.
Figure 7. ROC curves of different models for each disease. The dashed line in the middle indicates a random classifier. The curve that is further to the left corner indicates a better model.
Bioengineering 10 01089 g007
Figure 8. The confusion matrix for each disease of our proposed model.
Figure 8. The confusion matrix for each disease of our proposed model.
Bioengineering 10 01089 g008
Figure 9. Training and validation loss of different models. The figure shows the convergence of training and validation loss over epochs, indicating a low likelihood of overfitting.
Figure 9. Training and validation loss of different models. The figure shows the convergence of training and validation loss over epochs, indicating a low likelihood of overfitting.
Bioengineering 10 01089 g009
Figure 10. Comparative performance improvement of the three individual self-supervised learning methods—pair-instance, bi-lateral, and multi-modality—over the baseline model (no pre-train). Each bar represents the increase in the respective performance metric (AUC score, accuracy, F1 score, precision, recall) attributable to the implementation of the corresponding method.
Figure 10. Comparative performance improvement of the three individual self-supervised learning methods—pair-instance, bi-lateral, and multi-modality—over the baseline model (no pre-train). Each bar represents the increase in the respective performance metric (AUC score, accuracy, F1 score, precision, recall) attributable to the implementation of the corresponding method.
Bioengineering 10 01089 g010
Table 1. Performance of different approaches on image classification, including random initiation weights, supervised learning, simCLR, and our proposed method.
Table 1. Performance of different approaches on image classification, including random initiation weights, supervised learning, simCLR, and our proposed method.
MethodsAUC ScoreAccuracyF1 ScorePrecisionRecall
No pre-train72.9584.2128.2939.4748.29
Supervised (ImageNet)84.5787.6757.1167.0361.22
SimCLR [18]85.0988.2557.4267.7064.18
Barlow Twins [26]85.4789.4640.1455.9353.69
Ours86.9689.5062.5171.8064.54
Texts in bold denote the best performance.
Table 2. Mean AUC score of each disease.
Table 2. Mean AUC score of each disease.
MethodsDMEMGSMDRBRVO
No pre-train74.3954.2484.4872.4768.4264.75
Supervised (ImageNet)91.1657.5683.3184.6292.6690.05
SimCLR [18]90.3179.4886.1984.6392.7766.68
Barlow Twins [26]95.6564.5283.1786.7985.4583.93
Ours93.3684.7397.2987.3993.0869.49
DM: DM retinopathy, EM: epiretinal membrane, GS: glaucoma suspect, MD: macular degeneration, RB: retinal break, RVO: retinal vein occlusion. Texts in bold denote the best performance.
Table 3. Performance of different self-supervised learning methods on image classification.
Table 3. Performance of different self-supervised learning methods on image classification.
MethodAUC ScoreAccuracyF1 ScorePrecisionRecall
Pair-instance85.0988.2557.4267.7064.18
Bi-lateral85.4788.1357.1663.1863.93
Multi-modality85.9089.1361.5467.9964.19
Ours86.9689.5062.5171.8064.54
Texts in bold denote the best performance.
Table 4. Performance of supervised baseline and three self-supervised learning methods on image classification.
Table 4. Performance of supervised baseline and three self-supervised learning methods on image classification.
Pair-InstanceBi-LateralMulti-ModalityAUC ScoreAccuracyF1 ScorePrecisionRecall
86.7588.9259.5568.7762.08
86.5588.7161.1966.4661.99
86.5188.8363.0670.5364.94
86.9689.5062.5171.8064.54
Texts in bold denote the best performance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nguyen, T.D.; Le, D.-T.; Bum, J.; Kim, S.; Song, S.J.; Choo, H. Self-FI: Self-Supervised Learning for Disease Diagnosis in Fundus Images. Bioengineering 2023, 10, 1089. https://doi.org/10.3390/bioengineering10091089

AMA Style

Nguyen TD, Le D-T, Bum J, Kim S, Song SJ, Choo H. Self-FI: Self-Supervised Learning for Disease Diagnosis in Fundus Images. Bioengineering. 2023; 10(9):1089. https://doi.org/10.3390/bioengineering10091089

Chicago/Turabian Style

Nguyen, Toan Duc, Duc-Tai Le, Junghyun Bum, Seongho Kim, Su Jeong Song, and Hyunseung Choo. 2023. "Self-FI: Self-Supervised Learning for Disease Diagnosis in Fundus Images" Bioengineering 10, no. 9: 1089. https://doi.org/10.3390/bioengineering10091089

APA Style

Nguyen, T. D., Le, D. -T., Bum, J., Kim, S., Song, S. J., & Choo, H. (2023). Self-FI: Self-Supervised Learning for Disease Diagnosis in Fundus Images. Bioengineering, 10(9), 1089. https://doi.org/10.3390/bioengineering10091089

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop