Next Article in Journal
Occam’s Razor for Big Data? On Detecting Quality in Large Unstructured Datasets
Next Article in Special Issue
Special Issue on Using Machine Learning Algorithms in the Prediction of Kyphosis Disease: A Comparative Study
Previous Article in Journal
Identification of Alzheimer’s Disease on the Basis of a Voxel-Wise Approach
Previous Article in Special Issue
Deep Convolutional Neural Network Model for Automated Diagnosis of Schizophrenia Using EEG Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Medinoid: Computer-Aided Diagnosis and Localization of Glaucoma Using Deep Learning

1
Center for Biotech Data Science, Ghent University Global Campus, Incheon 21985, Korea
2
IDLab, ELIS, Ghent University, 9000 Gent, Belgium
3
Department of Ophthalmology, Samsung Medical Center, Seoul 06351, Korea
4
Sungkyunkwan University School of Medicine, Seoul 06351, Korea
5
Department of Nuclear Medicine, Medical AI Research Lab, Samsung Medical Center, Seoul 06351, Korea
*
Authors to whom correspondence should be addressed.
This work is an extended version of a paper published in the proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), held in Madrid, Spain, 3–6 December 2018.
These authors contributed equally to this work.
Appl. Sci. 2019, 9(15), 3064; https://doi.org/10.3390/app9153064
Submission received: 4 July 2019 / Revised: 25 July 2019 / Accepted: 25 July 2019 / Published: 29 July 2019
(This article belongs to the Special Issue Machine Learning for Biomedical Data Analysis)

Abstract

:
Glaucoma is a leading eye disease, causing vision loss by gradually affecting peripheral vision if left untreated. Current diagnosis of glaucoma is performed by ophthalmologists, human experts who typically need to analyze different types of medical images generated by different types of medical equipment: fundus, Retinal Nerve Fiber Layer (RNFL), Optical Coherence Tomography (OCT) disc, OCT macula, perimetry, and/or perimetry deviation. Capturing and analyzing these medical images is labor intensive and time consuming. In this paper, we present a novel approach for glaucoma diagnosis and localization, only relying on fundus images that are analyzed by making use of state-of-the-art deep learning techniques. Specifically, our approach towards glaucoma diagnosis and localization leverages Convolutional Neural Networks (CNNs) and Gradient-weighted Class Activation Mapping (Grad-CAM), respectively. We built and evaluated different predictive models using a large set of fundus images, collected and labeled by ophthalmologists at Samsung Medical Center (SMC). Our experimental results demonstrate that our most effective predictive model is able to achieve a high diagnosis accuracy of 96%, as well as a high sensitivity of 96% and a high specificity of 100% for Dataset-Optic Disc (OD), a set of center-cropped fundus images highlighting the optic disc. Furthermore, we present Medinoid, a publicly-available prototype web application for computer-aided diagnosis and localization of glaucoma, integrating our most effective predictive model in its back-end.

1. Introduction

Glaucoma is a progressive optic neuropathy [1], mostly manifesting itself in the area between the optic disc and the macula. The progress of glaucoma usually remains undetected until the optic nerve gets irreversibly damaged. This damage may result in varying degrees of permanent vision loss [2], as illustrated in Figure 1. The earlier glaucoma is diagnosed and treated, the less patients suffer from irreversible disease progress leading to blindness.
The global prevalence of glaucoma is approximately 3–5% for people aged 40–80 years. Specifically, the number of people aged 40–80 years and affected by glaucoma worldwide was estimated to be 64 million in 2013, and this number is expected to increase to 76 million in 2020 and to 112 million in 2040 [3].
As glaucoma progresses, glaucomatous optic disc changes (e.g., Optic Nerve Head (ONH) rim thinning or notching) and parapapillary retinal nerve fiber defects become representative morphological patterns [4,5,6]. As those patterns can be captured by fundus images, an ophthalmologist is able to diagnose glaucoma by manual screening of these images. However, due to individual diversity in optic disc and retina morphology [7,8,9], a human diagnosis usually does not only require fundus images, but also other types of medical images (Retinal Nerve Fiber Layer (RNFL), Optical Coherence Tomography (OCT) disc/macula, perimetry) in order to achieve high accuracy, sensitivity, and specificity values. Furthermore, the whole process of image capturing and manual screening is typically labor intensive and time consuming. Therefore, various Computer-Aided Diagnosis (CAD) techniques have been developed for identifying glaucoma, with the aim of improving the visual interpretation of fundus images by ophthalmologists. The latter type of images is commonly available in hospitals world-wide, given that their acquisition is relatively inexpensive compared to the acquisition of other types of medical images.
In the area of computer-aided glaucoma diagnosis, we can identify two major research directions: (1) glaucoma detection and (2) Optic Disc (OD) segmentation. When scanning the literature on glaucoma detection, we can observe that the accuracy of classification may vary significantly. In particular, as shown in Table 1, the classification accuracy may range from 80%–96%, depending on the feature extraction method and the type of classifier used.
Features are often generated by higher order spectral transforms [10,11,12], wavelet transforms [13,14], and/or thresholding [15,16]. It is also common to apply one or more feature extraction methods to OD images in order to compute the Cup-to-Disc Ratio (CDR) and/or the ISNT (Inferior, Superior, Nasal, Temporal) rule. The extracted features can then be fed into classifiers like k-Nearest Neighbors (k-NN), naive Bayes, Artificial Neural Networks (ANNs), Support Vector Machines (SVMs), Sequential Minimal Optimizations (SMOs), and Random Forests (RFs). The use of Convolutional Neural Networks (CNNs) has recently also become popular [17].
Based on Table 1, we can point out two additional concerns. First, many studies only make use of a limited number of images, with the number of images often varying between 50 and 200. This makes the classifiers used prone to fitting only to the given image distributions. As a result, it is highly likely that the proposed approaches do not generalize well to other sets of fundus images. Second, many studies do not provide a comprehensive quantitative analysis in terms of accuracy, sensitivity, and specificity.
When scanning the literature on OD segmentation, we can observe efforts on ONH localization using local contrast enhancement, brightest image pixel selection, and the circular Hough transform [18]. However, the use of brightness does not always allow finding the ONH, thus sometimes leading to the wrong diagnosis result. Another popular method makes use of Principal Component Analysis (PCA) [19] as a feature extractor. Moreover, algorithms such as the normalized correlation coefficient [20] and pyramidal decomposition [21] have been introduced for ONH localization. These approaches are relatively more accurate than their predecessors, while gradually reducing the time complexity. Nevertheless, when the ONH is not clearly visible in an input image, then the accuracy drops significantly [22].
A recent effort on ONH localization leveraged CNNs [27], using U-Net and polar transformations. However, this approach only segments the optic disc and the optic cup for measuring the CDR, given that the latter is considered to be an important factor for glaucoma diagnosis. We can therefore identify two drawbacks. First, from a clinical point of view, there is the possibility of having the wrong diagnosis outcome, given that a deviating CDR does not necessarily point to the presence of glaucoma. Second, the aforementioned approach only works in a supervised learning context. In other words, if segmentation annotations are not available for a particular dataset, then the approach at hand cannot be applied.
In this paper, we propose a novel approach towards computer-aided identification of glaucoma, making it possible to both diagnose and localize this eye disease in fundus images with a high effectiveness. More precisely, we can summarize our contributions as follows:
  • High diagnosis effectiveness: Our predictive model comes with a high diagnosis accuracy of 96%, as well as a high sensitivity of 96% and a high specificity of 100% for Dataset-OD, a set of center-cropped fundus images highlighting the optic disc. Current state-of-the-art techniques for computer-aided diagnosis using deep learning [17] have an accuracy of 94%, a sensitivity of 92%, and a specificity of 96%. Furthermore, our predictive model is able to achieve an AUC score of 99%.
  • Glaucomatous area localization: Our predictive model does not only come with a high diagnosis accuracy; it is also able to localize the glaucomatous area, thus helping ophthalmologists obtain a diagnosis that is more trustworthy.
  • Probability of glaucoma diagnosis: Upon diagnosis of glaucoma, we provide a probability that represents the diagnostic confidence.
  • Use of 1903 and 220 fundus images for training and testing, respectively: The dataset used was curated by ophthalmologists working at Samsung Medical Center in Seoul, Korea. More details about this dataset, which can be made available upon request, can be found in Section 3. Furthermore, we also validated the effectiveness of our predictive model using an external dataset [28].
  • Medinoid: We have developed Medinoid (http://www.medinoid.org), a publicly-available prototype web application for computer-aided diagnosis and localization of glaucoma in fundus images, integrating our predictive model. This web application is intuitive to use for medical doctors, as well as for people who have difficulties in accessing human experts and/or specialized medical imaging equipment.
Note that this paper is an extension of work previously presented at a conference [29], leveraging deeper models so as to be able to increase the effectiveness of diagnosis and additional datasets so as to be able to better evaluate generalization power. Furthermore, we present more in-depth experimental results and introduce Medinoid, a web application for both glaucoma diagnosis and localization.
The remainder of this paper is organized as follows. In Section 2, we provide an in-depth overview of our approach towards glaucoma diagnosis and localization. In Section 3, we outline our experimental setup, and in Section 4, we subsequently discuss the results obtained. Finally, we provide conclusions and directions for future research in Section 5.

2. Methodology

Our proposed approach encompasses two major computer vision tasks, namely image classification and localization. The first task outputs the diagnosis result (normal or glaucomatous) and the diagnostic confidence. Upon diagnosis of glaucoma, the second task outputs the most suspicious area in the given fundus image.
At the core of the classification task is a CNN architecture that acts as a feature extractor, returning a predicted output y i for the class i. The softmax function σ = e y i / k e y k is then applied to calculate the probability of each class k K . We used the resulting probability of the predicted output as a measure of the diagnosis confidence, given that the softmax function normalizes the output into a distribution of K probabilities.
In our research, through a thorough comparison of various CNN architectures and datasets, we focused on obtaining a high diagnosis accuracy and a localization technique that is clinically meaningful. More details about the different CNN architectures used and the localization technique employed can be found below.

2.1. Convolutional Neural Networks

Given their high effectiveness, CNNs are currently the most widely-used technique in image classification [30,31,32]. Their strength stems from the use of convolutional filters that are composed of neurons, also known as receptive fields. Inspired by biological processes, the neurons convolve over local regions in the input layer and respond to certain patterns, as the visual cortex does to stimuli in a local space. This is an important characteristic because, unlike the full connectivity of ordinary neural networks, this local connectivity enables CNNs to handle high-dimensional inputs such as natural images by substantially reducing the number of parameters to compute. In addition, CNNs are able to learn and classify hierarchical features directly from raw input images (end-to-end learning), thus not needing extraction of hand-crafted features. As CNNs have been demonstrated to be highly effective across various general image classification tasks, including the ImageNet Large-Scale Visual Recognition Challenge (ILSVRC), which comes with more than 1.2 millions images distributed over 1000 classes [33], medical image analysis using CNNs followed quickly [34], targeting use cases such as diabetic retinopathy diagnosis [35] and chest pathology analysis [36,37].
To achieve our goal of effective glaucoma diagnosis, we explored three representative CNNs: (1) VGG Networks (VGGNets) [38], (2) Inception Networks (InceptionNets) [31], and (3) Residual Networks (ResNets) [39]. Compared with their predecessors such as LeNetand AlexNet, these networks are all based on deeper architectures, all achieving higher accuracies while maintaining lower error rates. A comparison of the accuracies achieved by the different types of networks can be found in Table 2.
The aforementioned CNNs were able to obtain high levels of accuracy thanks to the availability of ImageNet. This implies that, to achieve similar levels of effectiveness for our use case of glaucoma diagnosis, we need a dataset of similar size to ImageNet or we need to modify the CNNs in question so that they are able to mitigate overfitting issues caused by the usage of a limited amount of data [34]. As such, starting from the CNNs in question, we incorporated several regularization techniques, including early stopping, data augmentation, dropout layers, and transfer learning [41], with the aim of achieving better model generalization.
In what follows, we explain the main characteristics of each model, with Table 3 summarizing the modifications we made.
VGGNets: Developed at and named after Oxford’s Visual Geometry Group (VGG), VGGNets [38] were the first models to combine small receptive fields with a size of 3 × 3 with multiple CNN layers. VGGNets are constructed using several convolutional blocks that contain two to three convolutional layers, followed by a max pooling layer. VGG-16 is currently the most representative network, having a relatively low number of parameters, but a higher top 1 classification accuracy compared to VGG-19. Like ResNets, VGGNets require input images to be normalized, resulting in zero mean and unit variance. Furthermore, VGGNets and ResNets make use of the same input size.
VGG-16 in its original form uses Stochastic Gradient Descent (SGD) as an optimizer, also employing learning rate decay. In our research, however, we made use of the Adaptive Moment estimation (ADAM) optimizer [42], so as to be able to overcome a number of drawbacks of SGD, including a slow convergence and a high error fluctuation.
ResNets: Microsoft Research introduced ResNets in 2016 [39]. These models obtained the first place in the ImageNet and the COCO2015 competitions, covering the tasks of image classification, semantic segmentation, and object detection. The core idea behind ResNets is the use of a residual network, which leverages skip connections to jump over some layers, reducing the impact of vanishing gradients, as there are fewer layers through which to propagate. As such, the use of residual networks makes it possible to train networks that are deeper, obtaining lower training errors and higher levels of accuracy. Depending on the number of residual networks, ResNets come with certain variations. In our research, we used a ResNet with 152 layers (i.e., ResNet-152), given that this network, among the different ResNets available, is able to obtain some of the highest accuracy levels.
InceptionNets: Unlike VGGNets and ResNets, InceptionNets [31] make it possible to better focus on the location of features that are important for the purpose of classification. Since salient parts can have different sizes per image, choosing the right receptive field size is difficult. In addition, in deep networks, there is the issue of overfitting and the issue of vanishing gradients. Therefore, Google Research introduced a new module, called “inception”, having several receptive fields with a different size [43]. Specifically, filter sizes of 1 × 1 , 3 × 3 , and 5 × 5 were used, making the network wider. After max pooling, each output is concatenated and then sent to the next inception module. Over the course of time, inception modules have been modified and improved, leading to more powerful InceptionNets. Inception-ResNet-v2 and Inception-v4 are currently the most effective networks. By default, InceptionNets take as input images with a size of 299 × 299 × 3 , with pixel values belonging to [−1, 1]. The optimizer used is also different from the previous networks: InceptionNets typically make use of the Root Mean Squared Propagation (RMSProp) optimizer [44].

2.2. Grad-CAM

In practice, most medical images do not have any localization information, making it impossible to apply a state-of-the-art image segmentation approach like U-Net [30]. Therefore, we adopted Gradient-weighted Class Activation Mapping (Grad-CAM) [45] to localize glaucoma, a weakly-supervised learning approach, forgoing the need to give explicit segmentation information to a network. Grad-CAM generalizes the Class Activation Mapping (CAM) method originally proposed in [46], integrating gradient weights. CAM visualizes what CNNs look at when they classify the input, highlighting activations in the feature maps of the last convolutional layer, using the predicted output and backpropagation.
To produce the final localization map L G r a d C A M g of the glaucomatous area, we first calculate the gradients of the predicted output y g for the class glaucoma with respect to the feature maps A k of the last convolutional layer, having width u and height v, where k { 1 , 2 , , K } , and with K denoting the total number of feature maps. The result y g A k then goes through a global-average pooling process, calculating the weights α k g of the glaucoma class as follows:
α k g = 1 u × v i j y g A k .
Equation (1) can be interpreted as the importance of each feature map for the glaucomatous class. The weights and the corresponding feature maps are linearly combined and subsequently given as an input to a Rectified Linear Unit (ReLU) [47], with the latter only selecting the positive activations. The overall process can be formally summarized as follows:
L G r a d C A M g = ReLU ( k α k g A k ) .

3. Experimental Setup

3.1. Dataset: Fundus Images

Digital color fundus images (this study followed all guidelines for experimental investigation on human subjects, was approved by the Samsung Medical Center Institutional Review Board, and adhered to the tenets of the Declaration of Helsinki) were used from a database created by the glaucoma clinic of Samsung Medical Center in Seoul, Korea. All patients gave their informed consent to have their data used in our experiments. When the fundus images were taken, the camera had a magnification setting of 30 or 35 . Moreover, the macula was center-positioned near the intersection of the crosshairs in the ocular. A fixation target was used to direct the gaze of the subject in the primary (straight ahead) position, so that the optic disc did not appear directly behind the lens. The color images were evaluated for quality by an experienced investigator. If the quality was not deemed adequate for assessment of the key features of the studied eye (such as optic disc morphology) and if no irremediable cause of inadequate quality was present (such as lens opacities or a pupil that did not dilate adequately), then the images were retaken.
The fundus dataset used for training contained 540 normal eye images and 1363 glaucomatous eye images, thus resulting in the use of a total of 1903 eye images. This dataset was randomly split into two sets, namely a training and a validation set, using a split ratio of approximately 4:1 (that is, the training dataset contained 1503 images and the validation dataset 400 images). The training set was used to fine-tune our models, whereas the validation set was employed to assess the generalization power of the models created during training. Apart from these two datasets, for testing purposes, 220 unseen fundus images were used: 55 normal eye images and 165 glaucomatous eye images. The images used came with different spatial resolutions, ranging from 1172–2500 pixels horizontally and from 1500–3200 pixels vertically.
Starting from the set of original fundus images, we created two derived datasets, as shown in Figure 2. For the first derived dataset, we center-cropped all images using a 1:1 ratio by cutting off unnecessary image areas (e.g., a black background sometimes containing image numbers and text). For the second derived dataset, we cropped the optical disc part from each image. The former set of images was named Dataset-C, and the latter set of images was named Dataset-OD. We created the two different datasets in order to be able to investigate the importance of the optic disc, given that most related work only made use of this part of the fundus images.

3.2. Implementation

Our approach towards glaucoma diagnosis and localization was implemented in Python 2.7, using Tensorflow 1.7, with all pre-trained models taken from Tensorflow Slim (https://github.com/tensorflow/models/tree/master/research/slim). Furthermore, our approach was run using two Intel(R) Xeon(R) E5-2620 2.4-GHz CPUs and an NVIDIA GeForce GTX TITAN X GPU.
We focused on regularizing and generalizing the different predictive models created by making use of different techniques, for instance mitigating overfitting and bias by adding a dropout layer as a regularizer at the end of each convolutional block, residual block, or inception block, depending on the type of model used. Our overall experimental setup can be found in Table 3.
Data pre-processing: In this phase, all the input images were re-sized and normalized before training, taking into account the pre-trained model settings [31,38,39]. For VGG-16-* and ResNet-152-B, we re-sized the input to 224 × 224 × 3 and normalized the input using the following RGB means: R:105.51, G:54.52, and B:16.19 (as derived from the images in the training set). For ResNet-152-Mand Inception-v4-*, after re-sizing the input to 299 × 299 × 3 , we scaled the input values to the range [−1, 1] instead of [0, 255].
Data augmentation: Overfitting occurs frequently when combining a small-sized dataset with a deep neural network, as was the case for the given set of fundus images and the architectures used, with the latter coming with a large number of parameters that needed to be computed. In order to alleviate overfitting, we incorporated data augmentation, artificially amplifying the volume of data during training through image processing techniques (e.g., horizontal flipping, contrast adjustment, and brightness adjustment), and where these techniques are often used in the context of computer vision tasks. Based on a comparison of the accuracy of different data augmentation techniques in [48], we applied several random data augmentation methods, also taking into account the different models used (that is, for each model, we used different types of data augmentation).
Fine-tuning: When training from scratch, a model often initializes all parameters using random Gaussian distributions. This approach towards initialization is usually less accurate, with the model at hand needing more time to converge [34,49]. To avoid initialization issues, we employed models pre-trained on ImageNet, transferring their weights to our models, hereby also allowing for better generalization. Starting from these pre-trained weights, we fine-tuned all layers using the hyperparameter values shown in Table 4, over Dataset-C and Dataset-OD. Note that the batch size used was 32.

3.3. Evaluation

Other than validation during the training phase, we implemented two more approaches to evaluate our model generalization. First, as explained in Section 3.1, we created a new dataset for testing purposes only. Second, we made use of RIM-ONEr3 [28], a publicly-available dataset of fundus images, to evaluate our model generalizability. Evaluation results obtained for both datasets are reported in the following section using several metrics: accuracy, recall (sensitivity), precision, F1 score, specificity, and Receiver Operating Characteristic-Area Under the Curve (ROC-AUC). These metrics were calculated as follows:
  • Accuracy = tp + tn tp + fp + tn + fn
  • Recall (sensitivity) = tp tp + fn
  • Precision = tp tp + fp
  • F1 score = 2 · recall · precision recall + precision
  • Specificity = tn tn + fp
In the above, true positive (tp) denotes the number of glaucomatous images that have been classified as glaucomatous and true negative (tn) denotes the number of normal images that have been classified as normal. Similarly, false positive (fp) denotes the number of normal images that have been classified as glaucomatous, and false negative (fn) denotes the number of glaucomatous images that have been classified as normal. The first four metrics were used to evaluate the robustness of the deep learning models. The recall, specificity, and ROC-AUC were used to evaluate the models from a clinical point of view.

4. Experimental Results

Our experiments focused on evaluating both the model generalizability and the importance of the OD area in the fundus images. In this context, we first performed a quantitative analysis of the diagnosis accuracy, followed by a qualitative analysis of the effectiveness of localization.

4.1. Diagnosis Accuracy

To investigate the importance of the OD image area, the original dataset was used to create two new datasets, namely Dataset-C and Dataset-OD. The results, as obtained for the testing set and presented in Table 5 and Table 6, demonstrated that the models trained on Dataset-OD were more effective than the models trained on Dataset-C, in terms of all evaluation metrics used, thus confirming that the OD image area is a clinically-important region.
As shown by Table 5, and unlike the accuracy results obtained for the ImageNet classification task (see Table 2), the ResNet-152-M model was able to achieve the highest accuracy of 95% for Dataset-C and the highest accuracy of 96% for Dataset-OD. Compared to the images in ImageNet, the images in Dataset-C and Dataset-OD came with salient areas that had a relatively similar size and location. This homogeneity can for instance not be exploited well by InceptionNets, given the use of different convolutional filter sizes in each inception block, so as to be able to deal with salient areas that may differ from image to image in terms of size and location. Furthermore, when only considering the results presented for Dataset-OD in both Table 5 and Table 6, the ResNet-152-M model was able to outperform all other models, in terms of all metrics presented.
Table 6 additionally presents the sensitivity and the specificity of each model. For both metrics, the ResNet-152-M model, as trained on Dataset-OD, was able to obtain the highest values. This implies that the ResNet-152-M model was able to correctly screen both normal and glaucomatous eyes. This can also be qualitatively interpreted as follows: the ResNet-152-M model was able to provide more benefits at lower costs to patients. Furthermore, for both Dataset-C and Dataset-OD, the ResNet-152-M model obtained an AUC score of 99%, outperforming the other models (see Table 5 for the AUC score of each model).
To evaluate our model generalizability, we investigated two types of learning curves during training, making use of two representative models. Figure 3 illustrates the difference between the worst- and the best-performing models. In particular, Figure 3a,c illustrates the effectiveness of training and validation in terms of accuracy, showing that the accuracy increased for both VGG-16-B and ResNet-152-M as the number of steps increased. Contrarily, in terms of loss, Figure 3b,d shows that the VGG-16-B model did not optimize well at later steps, resulting in overfitting, while the ResNet-152-M model was able to converge well, resulting in good fitting.
Finally, we made use of RIM-ONE r3 [28], an external dataset, to provide a more in-depth evaluation of the generalization power of ResNet-152-M, our most effective model. The total number of fundus images in the RIM-ONE r3 dataset is 159, of which 85 eye images are normal and 74 eye images are glaucomatous. Although the characteristics (angle, eye range, camera model) of the fundus images are different from the fundus images collected at Samsung Medical Center, in general affecting the diagnosis accuracy of deep learning models, we applied our ResNet-152-M model, as trained on Dataset-C, to the RIM-ONE r3 dataset, given the visual resemblance between the two different datasets.
The authors of [17] performed experiments using five-fold cross-validation, with each validation set containing 17 normal and 14 glaucomatous eye images. For our external testing purposes, we split the RIM-ONE r3 dataset into five subsets, randomly selecting 31 images for each subset. The authors of [17] then evaluated the overall accuracy, sensitivity, and specificity. We implemented the same experiment, and the results obtained are summarized in Table 7. Note that the models proposed in [17] may be biased towards the RIM-ONE r3 dataset (the authors of [17] did not provide testing results for a dataset different from the RIM-ONE r3 dataset). Moreover, as also mentioned in [17], the small size of the RIM-ONE r3 dataset made it challenging to apply CNNs, or even most other machine learning techniques.

4.2. Localization of Glaucomatous Areas

Unlike the evaluation of diagnosis accuracy, it is difficult to perform a quantitative evaluation of the effectiveness of localization, given that boundaries of optic disc notching or retinal nerve fiber layer defects cannot be clearly segmented in eye fundus images. In other words, there is no absolute gold standard available for a quantitative analysis of glaucomatous defect localization. Indeed, clinicians typically only confirm the presence of glaucoma by observing structural features [2,52,53]. Therefore, we made use of a qualitative approach towards evaluating the effectiveness of localization, asking clinicians for their informed opinion.
Figure 4 shows the localization results obtained by the *-M models for a fundus image taken from Dataset-C that was diagnosed as glaucomatous. For each model, the layer extracted for the purpose of localization was the last convolutional layer before the global max-pooling layer. Even though all classifiers correctly identified the given fundus image as glaucomatous, the location each network looked at was different, with the VGG-16-M model not succeeding in localizing the correct glaucomatous area in the given input image (as shown in Figure 4b). On the contrary, ResNet-152-M and Inception-v4-M were able to localize the glaucomatous area correctly. Given both Figure 4c and Figure 4d, the ResNet-152-M model produced a more precise localization than the Inception-v4-M model, with the produced localization also being more correct from a clinical point of view.
Figure 5 and Figure 6 visualize how models find the glaucomatous area per ResNet block in ResNet-152-M. Given both figures, we can again confirm that our model looked at the OD area to diagnose glaucoma. Having a look at the internals of ResNet-152-M, as illustrated in Figure 5, the leftmost image was the input image, and from the left to the right, we present the last convolutional layer of each residual block. As the input went from the lower layers to the higher layers, Figure 5 shows where the model looked at in order to make a classification decision, localizing where the glaucomatous region can be found in the heat map.
Both Figure 7 and Figure 8 show localization results obtained for our most effective model, namely ResNet-152-M, for fundus images taken from Dataset-C. In our test results, 209 images were correctly diagnosed and 11 images were wrongly diagnosed, given the human labels. Among the correctly-diagnosed images, the optic nerves in the true positives were all correctly localized, as shown in Figure 7a. For the true negatives, most of them were clinically well localized, as shown in Figure 7b. However, there were some atypical localizations, as indicated by the white arrows in Figure 7c.
Among the wrongly-classified images, there were three false positives and three false negatives. Regarding the false positives, each one of them was diagnosed wrongly for different reasons. In Figure 8a, we can observe the distribution of the normal retinal nerve fibers around each optic disc. In addition, we can observe a normal cup-to-disc ratio and a healthy RNFL in each fundus image. However, the model still showed glaucomatous features around the optic disc. Furthermore, the fundus image in Figure 8b shows a myopic tilted optic disc with large Parapapillary Atrophy (PPA) (white arrows). Though no RNFL defects were shown in the fundus image, the localized result clinically showed a positive diagnosis around the ONH and, in particular, in the inferior optic disc area. When investigating the last image in Figure 8c, we can observe that the image was not clear due to media opacity (as for instance caused by cataract). Therefore, even though the image looked normal, which meant that glaucoma was absent, the model diagnosed the image as glaucomatous.
The remaining erroneous results of the model stemmed from three false negatives. Given Figure 8, we can further analyze why the model was producing the aforementioned false negatives. The first example in Figure 8d was an interesting case because the human labeling was wrong and the model correctly diagnosed the disease. In other words, the image did not have any glaucomatous symptoms, which implied that it was a true negative. The second example in Figure 8e shows a myopic fundus image with a tigroid pattern in the overall retina. Clinically, it was difficult to detect whether RNFL defects were present in this type of retina. Only the cup-to-disc ratio can give information about the presence or absence of glaucoma. In this experiment, the model might have missed the glaucomatous cup-to-ratio morphology. The third image in Figure 8f seemed to point to early-stage glaucoma producing RNFL defects. Given that the RNFL defect width was relatively narrow, as indicated by the white arrows in the figure, it might not be detectable.

4.3. Medinoid

We integrated ResNet-152-M, our most effective model, into the back-end of Medinoid, a publicly-available web application for glaucoma diagnosis and localization. Medinoid takes as input a fundus image (JPG, PNG, or BMP), showing the diagnosis outcome to the user as soon as the given fundus image has been analyzed by the model in the back-end. If the input image has been diagnosed as normal, then the result window will show the decision made, as well as the diagnostic confidence. If the input image has been diagnosed as glaucomatous, then the result window will additionally show the localization of the glaucomatous area. An illustrative screenshot can be found in Figure 9.

5. Conclusions and Future Research

In this paper, we introduced a new computer-aided approach towards glaucoma diagnosis and localization. The predictive CNN-based model we developed, namely ResNet-152-M, obtained the best diagnosis scores among all metrics used. Moreover, using Grad-CAM, a weakly-supervised localization method, our predictive model was able to highlight where a glaucomatous area can be found in a given input image. This demonstrates that the application of deep learning tools to medical images, even though the number of images available for training is typically small, can help doctors in diagnosing glaucoma in a more effective and efficient way. Lastly, we presented Medinoid, a web application that integrates our predictive model into its backend, making it for instance possible to diagnose glaucoma in a constrained medical environment.
If a set of training images is too small or if a set of training images is not representing well the general nature of the classes, then it is easy for a deep learning model to be biased towards the training set. In this context, we were for instance able to make the following observation: the experiments that made use of an external dataset produced lower accuracy scores than the experiments that made use of our own dataset. As a result, to mitigate model bias, future research will focus on incorporating various types of fundus images, for instance taken by other medical equipment using different angles, so as to be able to include more diversity. Furthermore, in future research, we plan to leverage our experience with deep learning-based diagnosis and localization of glaucoma in the context of other diseases, particularly paying attention to disease diagnosis and localization using 3D medical imagery.

Author Contributions

Conceptualization, M.K. and J.C.H.; methodology, M.K.; software, M.K.; validation, M.K., J.C.H., and S.H.H.; formal analysis, J.C.H. and C.K.; resources, J.C.H., S.H.H., and C.K.; data curation, M.K.; writing, original draft preparation, M.K.; writing, review and editing, J.C.H., Olivier Janssens, and W.D.N.; visualization, M.K.; supervision, S.V.H., C.K., and W.D.N.; project administration, W.D.N.; funding acquisition, C.K. and W.D.N.

Funding

The research described in this paper was funded by Ghent University, Ghent University Global Campus, imec, Flanders Innovation & Entrepreneurship (VLAIO), the Fund for Scientific Research-Flanders (FWO-Flanders), and the European Union. This research was also supported by the Basic Science Research Program of the National Research Foundation of Korea (NRF), as funded by the Ministry of Education (2017034834), and an NRF grant of the Korean government (MSIT) (No. 2017R1D1A1B03028735 and No. 2017R1D1A1B03034834).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Burgoyne, C.F.; Downs, J.C.; Bellezza, A.J.; Suh, J.K.F.; Hart, R.T. The optic nerve head as a biomechanical structure: A new paradigm for understanding the role of IOP-related stress and strain in the pathophysiology of glaucomatous optic nerve head damage. Prog. Retin. Eye Res. 2005, 24, 39–73. [Google Scholar] [CrossRef] [PubMed]
  2. Quigley, H.A.; Addicks, E.M.; Green, W.R.; Maumenee, A. Optic nerve damage in human glaucoma: II. The site of injury and susceptibility to damage. Arch. Ophthalmol. 1981, 99, 635–649. [Google Scholar] [CrossRef] [PubMed]
  3. Tham, Y.C.; Li, X.; Wong, T.Y.; Quigley, H.A.; Aung, T.; Cheng, C.Y. Global prevalence of glaucoma and projections of glaucoma burden through 2040: A systematic review and meta-analysis. Ophthalmology 2014, 121, 2081–2090. [Google Scholar] [CrossRef] [PubMed]
  4. Quigley, H.A.; Katz, J.; Derick, R.J.; Gilbert, D.; Sommer, A. An evaluation of optic disc and nerve fiber layer examinations in monitoring progression of early glaucoma damage. Ophthalmology 1992, 99, 19–28. [Google Scholar] [CrossRef]
  5. Schuman, J.S.; Hee, M.R.; Puliafito, C.A.; Wong, C.; Pedut-Kloizman, T.; Lin, C.P.; Hertzmark, E.; Izatt, J.A.; Swanson, E.A.; Fujimoto, J.G. Quantification of nerve fiber layer thickness in normal and glaucomatous eyes using optical coherence tomography: A pilot study. Arch. Ophthalmol. 1995, 113, 586–596. [Google Scholar] [CrossRef] [PubMed]
  6. Zangwill, L.M.; Bowd, C.; Berry, C.C.; Williams, J.; Blumenthal, E.Z.; Sánchez-Galeana, C.A.; Vasile, C.; Weinreb, R.N. Discriminating between normal and glaucomatous eyes using the Heidelberg retina tomograph, GDx nerve fiber analyzer, and optical coherence tomograph. Arch. Ophthalmol. 2001, 119, 985–993. [Google Scholar] [CrossRef] [PubMed]
  7. Savini, G.; Zanini, M.; Carelli, V.; Sadun, A.; Ross-Cisneros, F.; Barboni, P. Correlation between retinal nerve fibre layer thickness and optic nerve head size: An optical coherence tomography study. Br. J. Ophthalmol. 2005, 89, 489–492. [Google Scholar] [CrossRef]
  8. Huang, D.; Chopra, V.; Lu, A.T.H.; Tan, O.; Francis, B.; Varma, R. Does optic nerve head size variation affect circumpapillary retinal nerve fiber layer thickness measurement by optical coherence tomography? Investig. Ophthalmol. Vis. Sci. 2012, 53, 4990–4997. [Google Scholar] [CrossRef]
  9. Girkin, C.A. Differences in optic nerve structure between individuals of predominantly African and European ancestry: Implications for disease detection and pathogenesis. Clin. Ophthalmol. 2008, 2, 65. [Google Scholar] [CrossRef]
  10. Krishnan, M.M.R.; Faust, O. Automated glaucoma detection using hybrid feature extraction in retinal fundus images. J. Mech. Med. Biol. 2013, 13, 1350011. [Google Scholar] [CrossRef]
  11. Bock, R.; Meier, J.; Michelson, G.; Nyúl, L.G.; Hornegger, J. Classifying glaucoma with image-based features from fundus photographs. In Joint Pattern Recognition Symposium; Springer: Berlin, Germany, 2007; pp. 355–364. [Google Scholar]
  12. Acharya, U.R.; Dua, S.; Du, X.; Chua, C.K. Automated diagnosis of glaucoma using texture and higher order spectra features. IEEE Trans. Inf. Technol. Biomed. 2011, 15, 449–455. [Google Scholar] [CrossRef]
  13. Singh, A.; Dutta, M.K.; ParthaSarathi, M.; Uher, V.; Burget, R. Image processing based automatic diagnosis of glaucoma using wavelet features of segmented optic disc from fundus image. Comput. Methods Progr. Biomed. 2016, 124, 108–120. [Google Scholar] [CrossRef]
  14. Dua, S.; Acharya, U.R.; Chowriappa, P.; Sree, S.V. Wavelet-based energy features for glaucomatous image classification. Ieee Trans. Inf. Technol. Biomed. 2012, 16, 80–87. [Google Scholar] [CrossRef]
  15. Issac, A.; Sarathi, M.P.; Dutta, M.K. An adaptive threshold based image processing technique for improved glaucoma detection and classification. Comput. Methods Progr. Biomed. 2015, 122, 229–244. [Google Scholar] [CrossRef]
  16. Nayak, J.; Acharya, R.; Bhat, P.S.; Shetty, N.; Lim, T.C. Automated diagnosis of glaucoma using digital fundus images. J. Med Syst. 2009, 33, 337. [Google Scholar] [CrossRef]
  17. Zilly, J.; Buhmann, J.M.; Mahapatra, D. Glaucoma detection using entropy sampling and ensemble learning for automatic optic cup and disc segmentation. Comput. Med Imaging Graph. 2017, 55, 28–41. [Google Scholar] [CrossRef]
  18. Yin, F.; Liu, J.; Wong, D.W.K.; Tan, N.M.; Cheung, C.; Baskaran, M.; Aung, T.; Wong, T.Y. Automated segmentation of optic disc and optic cup in fundus images for glaucoma diagnosis. In Proceedings of the 2012 25th International Symposium on Computer-Based Medical Systems (CBMS), Rome, Italy, 20–22 June 2012; pp. 1–6. [Google Scholar]
  19. Li, H.; Chutatape, O. Automatic location of optic disk in retinal images. In Proceedings of the 2001 International Conference on Image Processing, Thessaloniki, Greece, 7–10 October 2001; Volume 2, pp. 837–840. [Google Scholar]
  20. Youssif, A.A.H.A.R.; Ghalwash, A.Z.; Ghoneim, A.A.S.A.R. Optic disc detection from normalized digital fundus images by means of a vessels’ direction matched filter. IEEE Trans. Med Imaging 2008, 27, 11–18. [Google Scholar] [CrossRef]
  21. Lalonde, M.; Beaulieu, M.; Gagnon, L. Fast and robust optic disc detection using pyramidal decomposition and Hausdorff-based template matching. IEEE Trans. Med Imaging 2001, 20, 1193–1200. [Google Scholar] [CrossRef]
  22. Haleem, M.S.; Han, L.; van Hemert, J.; Li, B. Automatic extraction of retinal features from colour retinal images for glaucoma diagnosis: A review. Comput. Med Imaging Graph. 2013, 37, 581–596. [Google Scholar] [CrossRef] [Green Version]
  23. Stapor, K.; Świtonski, A.; Chrastek, R.; Michelson, G. Segmentation of fundus eye images using methods of mathematical morphology for glaucoma diagnosis. In Proceedings of the International Conference on Computational Science, Kraków, Poland, 6–9 June 2004; pp. 41–48. [Google Scholar]
  24. Joshi, G.D.; Sivaswamy, J.; Krishnadas, S. Optic disk and cup segmentation from monocular color retinal images for glaucoma assessment. IEEE Trans. Med Imaging 2011, 30, 1192–1205. [Google Scholar] [CrossRef]
  25. Kim, P.Y.; Iftekharuddin, K.M.; Davey, P.G.; Tóth, M.; Garas, A.; Holló, G.; Essock, E.A. Novel fractal feature-based multiclass glaucoma detection and progression prediction. IEEE J. Biomed. Health Informatics 2013, 17, 269–276. [Google Scholar] [CrossRef]
  26. Nyúl, L.G. Retinal image analysis for automated glaucoma risk evaluation. In Proceedings of the MIPPR 2009: Medical Imaging, Parallel Processing of Images, and Optimization Techniques, International Society for Optics and Photonics, Yichang, China, 30 October–1 November 2009; Volume 7497, p. 74971C. [Google Scholar]
  27. Fu, H.; Cheng, J.; Xu, Y.; Wong, D.W.K.; Liu, J.; Cao, X. Joint Optic Disc and Cup Segmentation Based on Multi-label Deep Network and Polar Transformation. IEEE Trans. Med. Imaging 2018, 37, 1597–1605. [Google Scholar] [CrossRef]
  28. Pena-Betancor, C.; Gonzalez-Hernandez, M.; Fumero-Batista, F.; Sigut, J.; Medina-Mesa, E.; Alayon, S.; de la Rosa, M.G. Estimation of the relative amount of hemoglobin in the cup and neuroretinal rim using stereoscopic color fundus images. Investig. Ophthalmol. Vis. Sci. 2015, 56, 1562–1568. [Google Scholar] [CrossRef]
  29. Kim, M.J.; Park, H.M.; Zuallaert, J.; Janssens, O.; Van Hoecke, S.; De Neve, W. Computer-aided diagnosis and localization of glaucoma using deep learning. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 2357–2362. [Google Scholar]
  30. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
  31. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. AAAI 2017, 4, 12. [Google Scholar]
  32. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  33. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  34. Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef]
  35. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. Jama 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  36. Bar, Y.; Diamant, I.; Wolf, L.; Lieberman, S.; Konen, E.; Greenspan, H. Chest pathology detection using deep learning with non-medical training. In Proceedings of the 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), New York, NY, USA, 16–19 April 2015; pp. 294–297. [Google Scholar]
  37. Tang, Y.X.; Tang, Y.B.; Han, M.; Xiao, J.; Summers, R.M. Deep adversarial one-class learning for normal and abnormal chest radiograph classification. In Proceedings of the Medical Imaging 2019: Computer-Aided Diagnosis, International Society for Optics and Photonics, San Diego, CA, USA, 16–21 February 2019; Volume 10950, p. 1095018. [Google Scholar]
  38. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27 June–1 July 2016; pp. 770–778. [Google Scholar]
  40. TensorFlow-Slim Image Classification Model Library. Available online: https://github.com/tensorflow/models/tree/master/research/slim (accessed on 6 April 2018).
  41. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 3320–3328. [Google Scholar]
  42. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–10 June 2015; pp. 1–9. [Google Scholar]
  44. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA: Neural Networks Mach. Learn. 2012, 4, 26–31. [Google Scholar]
  45. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. arXiv 2017, arXiv:1610.02391. [Google Scholar]
  46. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the 2016 IEEE Conference on IEEE, Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  47. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Las Vegas, NV, USA, 27–30 June 2010; pp. 807–814. [Google Scholar]
  48. Hussain, Z.; Gimenez, F.; Yi, D.; Rubin, D. Differential data augmentation techniques for medical imaging classification tasks. AMIA Annu. Symp. Proc. 2017, 2017, 979–984. [Google Scholar]
  49. Torrey, L.; Shavlik, J. Transfer learning. In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques; IGI Global: Hershey, PA, USA, 2010; pp. 242–264. [Google Scholar]
  50. Cheng, J.; Liu, J.; Xu, Y.; Yin, F.; Wong, D.W.K.; Tan, N.M.; Tao, D.; Cheng, C.Y.; Aung, T.; Wong, T.Y. Superpixel classification based optic disc and optic cup segmentation for glaucoma screening. IEEE Trans. Med Imaging 2013, 32, 1019–1032. [Google Scholar] [CrossRef]
  51. Wong, D.; Liu, J.; Lim, J.; Jia, X.; Yin, F.; Li, H.; Wong, T. Level-set based automatic cup-to-disc ratio determination using retinal fundus images in ARGALI. Conf. Proc. IEEE Eng. Med. Biol. Soc. 2008, 2008, 2266–2269. [Google Scholar]
  52. Hoyt, W.; Newman, N. The earliest observable defect in glaucoma? Lancet 1972, 299, 692–693. [Google Scholar] [CrossRef]
  53. Sommer, A.; Miller, N.R.; Pollack, I.; Maumenee, A.E.; George, T. The nerve fiber layer in the diagnosis of glaucoma. Arch. Ophthalmol. 1977, 95, 2149–2156. [Google Scholar] [CrossRef]
Figure 1. Progressive visual loss caused by glaucoma. (a) Normal vision. (b) As glaucoma advances, the field of vision of a patient slowly narrows. (c) Advanced glaucoma without proper treatment leads to substantial vision loss, and to blindness if left untreated.
Figure 1. Progressive visual loss caused by glaucoma. (a) Normal vision. (b) As glaucoma advances, the field of vision of a patient slowly narrows. (c) Advanced glaucoma without proper treatment leads to substantial vision loss, and to blindness if left untreated.
Applsci 09 03064 g001
Figure 2. Taking into account the clinical importance of the Optic Disc (OD) and the macula, the original dataset was used to created two additional datasets, either by center-cropping the original images (Dataset-C) or by extracting the optic disc from the original images (Dataset-OD).
Figure 2. Taking into account the clinical importance of the Optic Disc (OD) and the macula, the original dataset was used to created two additional datasets, either by center-cropping the original images (Dataset-C) or by extracting the optic disc from the original images (Dataset-OD).
Applsci 09 03064 g002
Figure 3. Comparison between the models with the lowest and the highest effectiveness for the training set and the validation set during fine-tuning. (a) VGG-16-B accuracy on Dataset-C. (b) VGG-16-B training and validation (generalization) error on Dataset-C. (c) ResNet-152-M accuracy on Dataset-OD. (d) ResNet-152-M training and validation error on Dataset-OD.
Figure 3. Comparison between the models with the lowest and the highest effectiveness for the training set and the validation set during fine-tuning. (a) VGG-16-B accuracy on Dataset-C. (b) VGG-16-B training and validation (generalization) error on Dataset-C. (c) ResNet-152-M accuracy on Dataset-OD. (d) ResNet-152-M training and validation error on Dataset-OD.
Applsci 09 03064 g003
Figure 4. Localization results produced by the different models for the input image shown in (a). (b) shows the (wrong) localization result produced by VGG-16-M. (c) shows the (correct) localization result produced by Inception-v4-M, and (d) shows the (correct) localization result produced by ResNet-152-M.
Figure 4. Localization results produced by the different models for the input image shown in (a). (b) shows the (wrong) localization result produced by VGG-16-M. (c) shows the (correct) localization result produced by Inception-v4-M, and (d) shows the (correct) localization result produced by ResNet-152-M.
Applsci 09 03064 g004
Figure 5. Dataset-C: Visualization of intermediate layers, at the end of each ResNet block, of ResNet-152-M. The leftmost image is the input image, and the rightmost image is the last convolutional layer of the network. (a) An original funuds image inpu; (b) The 1st ResNet block output; (c) The 2nd ResNet block output; (d) The 3rd ResNet block output; (e) The final output.
Figure 5. Dataset-C: Visualization of intermediate layers, at the end of each ResNet block, of ResNet-152-M. The leftmost image is the input image, and the rightmost image is the last convolutional layer of the network. (a) An original funuds image inpu; (b) The 1st ResNet block output; (c) The 2nd ResNet block output; (d) The 3rd ResNet block output; (e) The final output.
Applsci 09 03064 g005
Figure 6. Dataset-OD: Visualization of intermediate layers, at the end of each ResNet block, of ResNet-152-M. The leftmost image is the input image, and the rightmost image is the last convolutional layer of the network. (a) An original funuds image inpu; (b) The 1st ResNet block output; (c) The 2nd ResNet block output; (d) The 3rd ResNet block output; (e) The final output.
Figure 6. Dataset-OD: Visualization of intermediate layers, at the end of each ResNet block, of ResNet-152-M. The leftmost image is the input image, and the rightmost image is the last convolutional layer of the network. (a) An original funuds image inpu; (b) The 1st ResNet block output; (c) The 2nd ResNet block output; (d) The 3rd ResNet block output; (e) The final output.
Applsci 09 03064 g006
Figure 7. Segmented highlights obtained for glaucomatous areas. (a) True positive localization examples. (b) True negative localization examples. (c) Abnormal localization examples.
Figure 7. Segmented highlights obtained for glaucomatous areas. (a) True positive localization examples. (b) True negative localization examples. (c) Abnormal localization examples.
Applsci 09 03064 g007
Figure 8. False localization examples. The first three figures, from (ac), represent false positive examples, and the next three figures, from (df), represent false negative examples.
Figure 8. False localization examples. The first three figures, from (ac), represent false positive examples, and the next three figures, from (df), represent false negative examples.
Applsci 09 03064 g008
Figure 9. (a) Screenshots of Medinoid, a web application developed in support of computer-aided glaucoma diagnosis. Once a user has uploaded a fundus image, Medinoid will automatically screen the fundus image for the presence of glaucoma using a back-end engine that has been built on top of our most effective predictive model. If the image is diagnosed to be glaucomatous, then Medinoid will print out the diagnostic confidence and visualize the glaucomatous area. (b) Diagnosis result for a glaucomatous eye image, showing the diagnostic confidence and highlighting the glaucomatous area. Grad-CAM, Gradient-weighted Class Activation Mapping.
Figure 9. (a) Screenshots of Medinoid, a web application developed in support of computer-aided glaucoma diagnosis. Once a user has uploaded a fundus image, Medinoid will automatically screen the fundus image for the presence of glaucoma using a back-end engine that has been built on top of our most effective predictive model. If the image is diagnosed to be glaucomatous, then Medinoid will print out the diagnostic confidence and visualize the glaucomatous area. (b) Diagnosis result for a glaucomatous eye image, showing the diagnostic confidence and highlighting the glaucomatous area. Grad-CAM, Gradient-weighted Class Activation Mapping.
Applsci 09 03064 g009
Table 1. Overview of computer-aided diagnosis for glaucoma. The dataset described covers both training and validation sets. Effectiveness results have been rounded to one decimal place, if applicable. CDR, Cup-to-Disc Ratio; SMO, Sequential Minimal Optimization; ISNT, Inferior, Superior, Nasal, Temporal.
Table 1. Overview of computer-aided diagnosis for glaucoma. The dataset described covers both training and validation sets. Effectiveness results have been rounded to one decimal place, if applicable. CDR, Cup-to-Disc Ratio; SMO, Sequential Minimal Optimization; ISNT, Inferior, Superior, Nasal, Temporal.
ReferenceApproachDatasetClassifierEffectiveness
AccuracySensitivitySpecificityAUROC
[17]Entropy sampling, ensemble learning, and CNNsDRISHTI-GS,50 imagesSoftmax logistic classifier94.1%92.3%95.6%N/A
[10]Higher order spectra, trace transform, and discrete wavelet transformPrivate, 60 imagesSVM92%90%93%N/A
[11]Pixel intensity values, texture, spectral features, and histogram parametersPrivate, 200 imagesNaive Bayes, k-NN, and SVM80%N/AN/AN/A
[13]Wavelet feature extraction and classificationPrivate, 63 imagesSVM and k-NN94.7%97.0%93.3%N/A
[18]Circular Hough transform, edge detection, and optimum channel selectionORIGAlight, 650 imagesNACDR error: 0.10N/AN/AN/A
[23]Morphological operations, geodesic reconstruction by dilationPrivate, 50 imagesNA96%N/AN/AN/A
[24]Region-based active contour model, Hough transform, vessel bend detection, and r-bends informationPrivate, 138 imagesNAVertical CDR error: 0.09/0.08 (mean/std.); CDR area ratio: 0.12/0.10N/AN/AN/A
[14]Extraction of energy features using a wavelet transform and classificationPrivate, 60 imagesSVM, SMO, RF, and naive Bayes93%N/AN/AN/A
[15]Adaptive thresholding and ISNT quadrants of neuro-retinal rim area and blood vesselsPrivate, 67 imagesSVM and ANN94.4%100%90%N/A
[16]Morphological operations, thresholding, and classificationPrivate, 61 imagesANN90%100%80%N/A
[12]Combination of texture and higher order spectra and classificationPrivate, 60 imagesRF91%N/AN/AN/A
[25]Fractal analysis and classificationN/ASVM88%93%67%82%
[26]Pixel intensity values, Fast Fourier Transform (FFT) coefficients, and B-spline coefficientsPrivate, not specifiedSVM81%N/AN/AN/A
Table 2. Accuracies of different deep neural network models trained on the ImageNet Large-Scale Visual Recognition Challenge (ILSVRC)-2012-CLS. Top 1 accuracy: the class having the highest probability is the same as the true label. Top 5 accuracy: one class among the top 5 ranked classes (ranked according to decreasing probability) is the same as the true label [40].
Table 2. Accuracies of different deep neural network models trained on the ImageNet Large-Scale Visual Recognition Challenge (ILSVRC)-2012-CLS. Top 1 accuracy: the class having the highest probability is the same as the true label. Top 5 accuracy: one class among the top 5 ranked classes (ranked according to decreasing probability) is the same as the true label [40].
ModelTop 1 AccuracyTop 5 Accuracy
VGG-1671.589.8
VGG-1971.189.8
Inception-v480.295.2
Inception-ResNet-v280.495.3
ResNet-15277.894.1
Table 3. Overview of the different model configurations for VGG-16 [38], ResNet-152 [39], and Inception-v4 [31]. When the name of a model contains the label “B”, this means that the model adopted the settings from the paper that introduced the model in question. When the label “M” has been added to the name of a model, this means that our research context has been taken into account, resulting in the modification of settings, the addition of regularizers such as data augmentation and dropout, and/or a change of the optimizers used. RMSProp, Root Mean Squared Propagation.
Table 3. Overview of the different model configurations for VGG-16 [38], ResNet-152 [39], and Inception-v4 [31]. When the name of a model contains the label “B”, this means that the model adopted the settings from the paper that introduced the model in question. When the label “M” has been added to the name of a model, this means that our research context has been taken into account, resulting in the modification of settings, the addition of regularizers such as data augmentation and dropout, and/or a change of the optimizers used. RMSProp, Root Mean Squared Propagation.
ModelInput SizeOptimizerData Augmentation# of Layers or Blocks
VGG-16-B 224 × 224 × 3 SGDcrop, horizontal flip16 layers
VGG-16-M 224 × 224 × 3 ADAMcrop, horizontal flip, contrast adjustment16 layers
ResNet-152-B 224 × 224 × 3 SGDcrop, horizontal flip152 layers
ResNet-152-M 299 × 299 × 3 ADAMcrop, horizontal flip, contrast adjustment152 layers
Inception-v4-B 299 × 299 × 3 RMSPropcrop, horizontal flip, brightness adjustment, contrast adjustment, hue adjustment, saturation adjustment14 inception blocks
Inception-v4-M 299 × 299 × 3 RMSPropcrop, horizontal flip, contrast adjustment14 inception blocks
Table 4. Hyperparameter values. We used early stopping to alleviate overfitting.
Table 4. Hyperparameter values. We used early stopping to alleviate overfitting.
ModelLearning RateNumber of Steps
VGG-16-B0.0000515,000
VGG-16-M0.0000515,000
ResNet-152-B0.000018500
ResNet-152-M0.000018400
Inception-v4-B0.000515,000
Inception-v4-M0.000515,000
Table 5. Evaluation of model performance for glaucoma diagnosis over the testing dataset.
Table 5. Evaluation of model performance for glaucoma diagnosis over the testing dataset.
DataClassifierAccuracyRecallPrecisionF1 ScoreAUC Score
CVGG-16-B88%88%95%91%93%
VGG-16-M90%88%97%92%94%
ResNet-152-B88%85%99%91%94%
ResNet-152-M95%95%98%96%99%
Inception-v4-B90%90%96%93%97%
Inception-v4-M93%92%99%95%99%
ODVGG-16-B86%95%88%91%93%
VGG-16-M91%90%98%94%96%
ResNet-152-B87%83%100%91%95%
ResNet-152-M96%95%100%97%99%
Inception-v4-B91%88%99%93%97%
Inception-v4-M94%92%99%95%99%
Table 6. Clinical evaluation of model performance for glaucoma diagnosis over the testing dataset.
Table 6. Clinical evaluation of model performance for glaucoma diagnosis over the testing dataset.
DatasetClassifiertnfptpfnSensitivitySpecificity
CVGG-16-B4781461988%85%
VGG-16-M5141461988%93%
ResNet-152-B5321412485%96%
ResNet-152-M523157895%95%
Inception-v4-B4961481790%89%
Inception-v4-M5411511492%98%
ODVGG-16-B3322157895%60%
VGG-16-M5231481790%95%
ResNet-152-B5501372883%100%
ResNet-152-M550157895%100%
Inception-v4-B5321461988%96%
Inception-v4-M5411521392%98%
Table 7. Effectiveness of glaucoma diagnosis for RIM-ONEr3 [28]. Ours refers to the ResNet-152-M model. The ES model of [17] model used Entropy Sampling for feature extraction, whereas the CNN of [17] used a CNN. Both ES [17] and CNN [17] used a softmax classifier. The results were adopted from [17].
Table 7. Effectiveness of glaucoma diagnosis for RIM-ONEr3 [28]. Ours refers to the ResNet-152-M model. The ES model of [17] model used Entropy Sampling for feature extraction, whereas the CNN of [17] used a CNN. Both ES [17] and CNN [17] used a softmax classifier. The results were adopted from [17].
OursES [17]CNN [17][50][51][24]
Accuracy93.594.192.490.289.492.1
Sensitivity92.992.390.187.486.489.8
Specificity92.995.694.392.592.094.0

Share and Cite

MDPI and ACS Style

Kim, M.; Han, J.C.; Hyun, S.H.; Janssens, O.; Van Hoecke, S.; Kee, C.; De Neve, W. Medinoid: Computer-Aided Diagnosis and Localization of Glaucoma Using Deep Learning . Appl. Sci. 2019, 9, 3064. https://doi.org/10.3390/app9153064

AMA Style

Kim M, Han JC, Hyun SH, Janssens O, Van Hoecke S, Kee C, De Neve W. Medinoid: Computer-Aided Diagnosis and Localization of Glaucoma Using Deep Learning . Applied Sciences. 2019; 9(15):3064. https://doi.org/10.3390/app9153064

Chicago/Turabian Style

Kim, Mijung, Jong Chul Han, Seung Hyup Hyun, Olivier Janssens, Sofie Van Hoecke, Changwon Kee, and Wesley De Neve. 2019. "Medinoid: Computer-Aided Diagnosis and Localization of Glaucoma Using Deep Learning " Applied Sciences 9, no. 15: 3064. https://doi.org/10.3390/app9153064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop