Next Article in Journal
A Novel Fuzzy Logic-Based Scheme for Malicious Node Eviction in a Vehicular Ad Hoc Network
Next Article in Special Issue
RGB-Based Triple-Dual-Path Recurrent Network for Underwater Image Dehazing
Previous Article in Journal
Improving the Energy Efficiency of Software-Defined Networks through the Prediction of Network Configurations
Previous Article in Special Issue
Segmentation of Spectral Plant Images Using Generative Adversary Network Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Diabetic Retinopathy in Retinal Fundus Images Using CNN Classification Models

1
School of Computer Science and Engineering, Central South University, Changsha 410083, China
2
College of Literature and Journalism, Central South University, Changsha 410083, China
3
Department of Computer Science, College of Computer and Information Sciences, Majmaah University, Al-Majmaah 11952, Saudi Arabia
4
School of Computer Science, China University of Geosciences, Wuhan 430074, China
5
Department of Ophthalmology, The Second Xiangya Hospital of Central South University, Changsha 410083, China
6
State Key Laboratory for Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
7
Faculty of Engineering, Sana’a University, Sana’a 12544, Yemen
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(17), 2740; https://doi.org/10.3390/electronics11172740
Submission received: 5 August 2022 / Revised: 26 August 2022 / Accepted: 27 August 2022 / Published: 31 August 2022
(This article belongs to the Special Issue Artificial Intelligence (AI) for Image Processing)

Abstract

:
Diabetes is a widespread disease in the world and can lead to diabetic retinopathy, macular edema, and other obvious microvascular complications in the retina of the human eye. This study attempts to detect diabetic retinopathy (DR), which has been the main reason behind the blindness of people in the last decade. Timely or early treatment is necessary to prevent some DR complications and control blood glucose. DR is very difficult to detect in time-consuming manual diagnosis because of its diversity and complexity. This work utilizes a deep learning application, a convolutional neural network (CNN), in fundus photography to distinguish the stages of DR. The images dataset in this study is obtained from Xiangya No. 2 Hospital Ophthalmology (XHO), Changsha, China, which is very large, little and the labels are unbalanced. Thus, this study first solves the problem of the existing dataset by proposing a method that uses preprocessing, regularization, and augmentation steps to increase and prepare the image dataset of XHO for training and improve performance. Then, it takes the advantages of the power of CNN with different residual neural network (ResNet) structures, namely, ResNet-101, ResNet-50, and VggNet-16, to detect DR on XHO datasets. ResNet-101 achieved the maximum level of accuracy, 0.9888, with a training loss of 0.3499 and a testing loss of 0.9882. ResNet-101 is then assessed on 1787 photos from the HRF, STARE, DIARETDB0, and XHO databases, achieving an average accuracy of 0.97, which is greater than prior efforts. Results prove that the CNN model (ResNet-101) has better accuracy than ResNet-50 and VggNet-16 in DR image classification.

1. Introduction

Diabetic retinopathy (DR) is a microvascular entanglement of type 1 and 2 diabetes. DR causes retinal irregularities and is one of the main causes of visual impairment around the world. Approximately 33% of individuals with diabetes have DR, and practically all diabetics will foster it after some time. By 2030, DR is predicted to afflict 191 million individuals [1,2]. Though the visual weakness and deficiency brought about by DR can be prevented [3], early discovery is critical [4]. To guarantee early discovery and brief treatment, current rules propose that individuals with inadequately controlled diabetes ought to be evaluated for DR once a year. Patients previously determined to have DR ought to be checked frequently. Evaluating for DR mainly includes catching a fundus picture of the retina, which is then assessed by expert ophthalmologists. If not treated, diabetic retinopathy progresses through four clinical stages of mild nonproliferative, moderate nonproliferative, and severe nonproliferative, as shown in Figure 1.
The number of retina photographs generated via the screening program will rise with the number of diabetic patients, challenging the provision of specialized eye care for everyone, thereby imposing a huge labor-intensive burden on medical experts and costs for medical services; these issues and the growing waiting list for ophthalmic consultations are the main problems facing public health systems [5,6,7]. These problems can be alleviated by an automated system, which can be used as a support tool for medical experts or a complete diagnostic tool. Many studies have reported on the application of deep learning (DL) algorithms in the automatic detection of DR. These techniques demonstrate the great sensitivity and specificity of automatic detection systems based on deep learning artificial neural networks to the reference DR. In addition, new eye problems such as age-related macular degeneration, glaucoma, and diabetic macular edema have recently been discovered [8,9,10,11].
The automated system must be able to arrange retinal photographs in accordance with the severity of clinical practice, such as the suggested international clinical levels of diabetic macular edema and DR, which are also used in some countries, for it to be practically practicable. According to previous studies, the latest experimental results of the former DR scale can be obtained, but the latter is not used for the experimental classification of maculopathy. The enormous amount of annotated images needed for model learning is a significant barrier to the widespread and successful usage of deep learning systems [12,13,14].
This study aims to use fundus image classification technology to automatically diagnose DR, classify fundus images according to the severity of DR, and realize end-to-end real-time classification from fundus photographs to the patient’s condition. The automatic diagnosis system has a high degree of automation and precision, thereby reducing the pressure on DR diagnosis and treatment. This task uses different image preprocessing methods to extract numerous significant features and then classifies them. Therefore, this work uses three types of CNN architecture (residual neural network (ResNet)-101, ResNet-50, and VggNet-16) to detect the DR of two datasets and evaluates the training accuracy, training loss, and test accuracy of the model.
This paper introduces a CNN method based on deep learning to solve the problem of DR classification. The motivations for employing different CNN-based deep neural networks evolved and obtained a substantial result on ImageNet Challenger, the most important image classification and segmentation challenge in the image-analyzing area. In medical categorization, a CNN-based deep neural system is commonly utilized. Because CNN is such a good feature extractor, using it to identify medical pictures may save time and money on feature engineering. The key benefits of CNN are a low reliance on preprocessing and a reduction in the amount of human labor required to create its features. It is simple to comprehend and apply. Among all picture prediction algorithms, it has the greatest accuracy. It finds the relevant traits without the need for human intervention. In addition, CNN is computationally efficient. Finally, CNN has shown to be quite effective in resolving picture categorization issues. Many image databases’ top performance has been greatly enhanced thanks to research based on CNN.
To the best of our knowledge, this is the first article to discuss the two categories of DR (symptomatic and asymptomatic) using different kinds of CNN architecture (ResNet-101, ResNet-50, and VggNet-16) on our laboratory dataset (obtained from the Xiangya Second Ophthalmic Hospital, Changsha, China, from March 2016 to October 2016). Recently, several new methods were introduced to adapt the CNN to large datasets. Then, we analyzed the performance and function of the different kinds of CNN architecture and used deep CNNs (ResNet-101, ResNet-50, and VggNet-16) to detect DR, ultimately achieving automatic and precise detection to minimize optical impairment. Compared with prior techniques, this work applied enhancements in terms of the convergence time of large-scale experimental datasets and classification performance. The main contributions of this paper are as follows:
  • This work’s laboratory database has a very large size, is little, and has some noise, and deep networks are slightly difficult to train. Thus, this study uses some preprocessing methodologies on the laboratory datasets to increase the training dataset and improve the performance of using a deep training network, such as resizing of input images, augmentation, and data normalization.
  • This study demonstrates that CNN-based approaches could achieve state-of-the-art performance in DR detection. In addition, this work uses CNN methods to find the fundamental picture features relevant to disease grade prediction. Later on, it justifies those aspects in light of expert expertise. Furthermore, this study evaluates different CNNs rigorously, taking into consideration what factors were used to determine illness state, and justifies their clinical utility.
  • Exploiting the advantages of CNN and ResNet with images, this work utilizes ResNet-101, ResNet-50, and VggNet-16 to recognize the slight differences and classify DR images.
  • This study also compares these three networks to determine the best for this kind of image.
The rest of this paper is organized as follows. Section 2 introduces the related work of this study, while Section 3 highlights the suggested methods. Section 4 establishes the experiments and highlights the performance evaluation of the CNN architectures, and Section 5 finally concludes this work.

2. Related Work

The primary research on DR detection essentially divides DR problems into different components, such as vascular, microaneurysm, fovea, exudate, hemorrhage, and optic disc, to analyze the severity of DR. In [15], a multidirectional two-dimensional Gaussian-matched filter is proposed to detect blood vessels. In [16], sinthanayothin attempts are made to localize the disc by discovering the area with the biggest local average gray variance. Baudoin finds microaneurysms in fluoroscopic photographs of the fundus according to morphological methods in [17]. Although these methods can locate the components of the retina well, they usually operate on high-quality images and cannot directly detect the severity of the DR.
For DR harshness classification, many traditional machine learning methods are used to detect DR automatically. Most traditional machine learning methods first extract features by hand, then classify images through traditional classifiers. In [18], a model of different types of retinopathy is developed based on scale-invariant feature transform in [19,20] to generate an image local descriptor. In [21,22], the BoVW scheme is based on the detection of low-level sparse features with an accelerated enhanced features (i.e., SURF) local descriptor and midlevel features in semisoft encoding with the maximum combination. The best representation of BoVW for retinal image classification is area under the receiver operating characteristics curves of 97.8% (exudates) and 93.5% (red lesions), applying a set of data cross-validation protocols. In [23,24], Seoud first generates the injury probability map and then combines its position, size, and probability information to represent the characteristics. The proposed approach achieves an overall FROC curve score of 0.393, while a previous work with images of similar resolution reports a score of 0.233. In [25], by extracting the features of the vascular and exudate areas and the texture features, a support vector machine (SVM) is provided for DR detection using fundus images. In [26], a random forest is used as a basic classifier. In [27], SVM is used; although the SVM and random forest methods can detect DR automatically, they are sensitive to the quality of fundus images and are inappropriate for large-scale datasets. In [28,29], the optic disc is detected using morphological filtering and watershed transformation techniques, The algorithm is tested on a small image database, obtaining a mean sensitivity of 92.8% and a mean predictive value of 92.4%, and compared with a human evaluator in terms of performance.
In recent years, deep learning technology has brought about an extraordinary revolution in the field of computer vision. The use of CNN for image classification has especially attracted the attention of many investigators in this area, including the segmentation of these features and blood vessels, as mentioned in [30,31]. The deep CNN structure was initially proposed to solve the problem of natural image classification. In [32], a CNN model (LeNet-5) is used to extract image features to solve the vascular segmentation problem. These technologies have some limitations. First, given the manual extraction of dataset features by experts, their accuracy cannot be ensured. Second, the dataset is limited and of poor quality, usually consisting of just a few hundred or even hundreds of fundus images, and the acquisition setting is rather simple, making it impossible to compare the algorithms’ performance in the experiment. Since Alex et al. [33,34] won in the 2012 ILSVRC competition, a remarkable improvement in the AlexNet architecture has been proposed, and the CNN’s deep-field computer vision has been extensively applied. Several excellent CNN architectures have also been recommended to further improve CNN’s performance, such as VggNet [35,36], GoogleNet [37], and ResNet [38]; the latest is one of the most important network classification models proposed in 2015. CNN models (AlexNet, VggNet, GoogleNet, and ResNet) are employed in [39] to distinguish the minor changes across image classes for DR detection, with the best classification accuracy of 95.68%. In [40], two models of deep learning are adopted. The first model (CNN512) uses the full image as the input to the CNN model for classification in one of the five DR classes, obtaining 84.1% and 88.6% accuracies on the APTOS and DDR Kaggle 2019 public datasets, respectively, and the second model uses an adopted YOLOv3 model to detect and locate DR lesions, reaching 0.216 mAP in locating lesions in all of the DR detection data.
According to [41], the SVM is suggested to predict DR in three separate classifications, including diabetic normal, nonproliferative retinopathy (non-PDR), and diabetic PDR, based on the features contained in a picture of the entry retinal fundus. In addition, the CNN technique can be applied to DR images. In [42], a four-layer DCNN is used to divide DR into normal, mild, and severe DR. In [43], a universal representational model, which enables the effective CNN classification of target objects inside images at any scale, is presented. The efficacy of this method is proven in a real-world application of detecting lesions in retinal pictures for classes that differ significantly in scale. Recently, Ref. [44] used a CNN based on Vgg16 on the Kaggle DR database [45] and achieved a sensitivity of 0.95 and an accuracy of 0.75 on 5000 verification images. These techniques essentially focus on classification without the capability to locate distinctive areas. Ref. [46] proposes class activation mapping (CAM), which is used to describe the weighted activation mapping after the global average pool [47], to locate different image areas. Inspired by CAM, [30] extended the method [35] from classification to a regression locale and shed light on the DR detection difficulty, which is based on VGG16 without fully connected layers.
According to Ref. [48], an embedded system based on image processing might be used to identify conditions that lead to blindness, such as glaucoma and diabetic retinopathy. For low-cost automated glaucoma and retinopathy diagnosis, a hybrid feature extraction technique is described. The datasets are classified using an artificial neural network classifier. In Ref. [49], initially, the RGB retina image is transformed to grayscale. On this grayscale image, contrast-limited adaptive histogram equalization (CLAHE) is used to adapt the varied intensity variances to a uniform intensity. Then, to reduce background noise and improve blood arteries, morphological opening surgery is performed. Later, using canny edge detection, the perimeter is recovered from the morphing image. After that, gray thresholding is used to remove the region from the morphing image. The retinopathy is then visible in the resulting image.
A methodical investigation of the significance of image processing for diabetic eye disease (DED) categorization is presented in Ref. [50]. Picture quality enhancement, image segmentation (area of interest), image augmentation (geometric transformation), and classification are all steps in the proposed automated classification framework for DED. Traditional image processing methods are combined with a novel built convolution neural network (CNN) architecture to produce the best results. For DED classification tasks, the novel designed CNN paired with the old image processing approach provides the best result in terms of accuracy. In [51], the input image is collected from the Indian Diabetic Retinopathy Image Dataset database, and 13 filters are used to enhance the photos, including smoothing and sharpening filters. Then, using performance measures, the quality of the enhancement algorithms is compared, and superior results are obtained for the Median, Gaussian, Bilateral, Wiener, and partial differential equation filters, which are integrated to improve picture enhancement. The convolutional neural network input is supplied to all of the enhanced filters’ output images, and the results are compared to discover the best enhancement approach.
Researchers working on the retinopathy detection problem ran into a few issues, many of which are unavoidable and for which there is no immediate solution. This is because the field of deep learning is still relatively new to many people, and data collection is always problematic. Data are sparse, and much of what is accessible is unusable for one reason or another. Using CNN to separate the lesions, Figure 2 shows the research gaps in retinopathy. The following are the key drawbacks of traditional approaches and DL architectures, as determined by the preceding study of the existing literature: 1. Dataset is limited, 2. Images that are twisted and blurred, 3. Models that are overfitting, and 4. Limited computing power.

3. Methods

First, some augmenting and preprocessing methods were used to increase the amount of training data and improve the classification effect and performance of the deep learning network used in the laboratory dataset. For example, the input images were resized and augmented, and the data was normalized. Then, the three most advanced CNN models were used to detect DR, and the accuracy of these models in image classification was tested.

3.1. Proposed Solution for Solving the Dataset Issues

3.1.1. Data Preprocessing

Given that low-quality images will produce inaccurate results, preprocessing is an essential operation to improve image quality. The results of preprocessing are regarded as the original input for data training, which divides the images into two categories. The dataset also contains images from patients of different races, age groups, and lighting levels in fundus photography, which will affect the pixel intensity values in the image and produce unnecessary changes independent of the classification level. To solve this problem, this work used the OpenCV library to preprocess the images. The OpenCV library provides two transform functions, cv2.warpAffine and cv2.warpPerspective, and noise. This library is necessary for image processing, as it comprises rich built-in features for quick processing.
The preprocessing consisted of image cropping followed by resizing. Each image was cropped to a square shape, which included the most tightly contained circular area of the fundus. The procedure removed most of the black borders and all of the patient-related annotations from the image data. Each of the cropped images was then resized to 300 × 300 pixels. Here, the creation of multiple resolutions was for the purposes of analyzing the effect of the input image resolution on the classification performance
  • Original fundus image dataset and retinal image grading systems
The research of the present study was conducted in collaboration with Xiangya No. 2 Hospital Ophthalmology (XHO), Changsha, China, a certified provider of diabetic retinopathy screening and monitoring services in China (obtained from the Xiangya Second Ophthalmic Hospital, Changsha, China, from March 2016 to October 2016). The output images were of variable resolutions of 1956 × 1934 pixels.
Each of the retinal images was graded with respect to three different criteria: (i) diabetic retinopathy, (ii) macular edema, and (iii) gradability, and a number of images were not graded to any criteria. Images are graded with the proposed international clinical diabetic retinopathy and macular edema disease severity scales, denoted later as symptomatic and asymptomatic, respectively. Image gradability is a two-stage system, which considers an image to be either gradable or not. All personnel participating in retinopathy assessment had over 10 years’ experience in diabetic retinopathy grading. All images with symptomatic and asymptomatic were graded by two ophthalmologists, both with more than 10 years of experience in grading fundus images. If there was a disagreement in grading, such an image was not included in this study.
Finally, all the images classified into (symptoms 607 and nonsymptoms 1000) images from the training set were used to develop and validate the model. The model was tested on (symptoms 200 and nonsymptoms 122), images as explained in Table 1.
Images from the training dataset (samples in Figure 3) are read using the Tensorflow-GPU library. This library is important for image processing, as it contains rich built-in features for rapid processing. The result of this can be seen in Figure 4b; therefore, preprocessing is the reshaping of input data. The images acquired are of size 1956 × 1934 pixels and occupy a large area of memory space. This demands too much RAM usage, resulting in slower computation. Hence, they are downscaled to 300 × 300 dimensions. The result of an image augmentation can be seen in Figure 4c.

3.1.2. Data Augmentation

The deep learning (DL) model has always been a data requirement due to data noise and dataset limitations. Thus, the data expander is used to generate more data without actually collecting more data to expand the dataset. This operation can be performed using the Keras library’s built-in capabilities. Image rotation (to induce rotation invariance), horizontal flipping, scaling, clipping, and translation for preprocessing are some of the augmentation techniques. To generate new data, a recent work [52] used the most advanced generative adversarial networks (i.e., GAN). Additionally, to eliminate extraneous parts, the fundus photos are cropped to a lesser size.

3.1.3. Utilizing CNN Models

This paper is focused on the use of CNNs in the diagnosis of DR. Automatic feature extraction and efficient computation are hallmarks of CNN and any deep learning paradigm.
The convolution, pool, and activation layers make up the CNN model. Different filters combine the data fed into the network. These filters are similar to traditional image processing filters. However, instead of being explicitly defined, they are self-learning. The convolution layer extracts local features of various positions from the original input or intermediate feature maps with various kernel sizes. The weight-sharing mechanism, local connectivity, and target position invariance are the advantages of the convolution layer. The pooling layer shrinks the feature maps and network parameters while keeping critical information. The most commonly used methods are average pool and maximum pool. The convolution layer is usually followed by the pooling layer. The fully connected layer resembles a traditional neural network and is added to the final pool layer for classification [52].
  • VggNet-16
VggNet-16 is a CNN structure, which won the 2014 ILSVRC (Imagenet) competition. VggNet-16 is regarded as one of the most advanced visualization model architectures to date. The 16 in VggNet-16 means that it has 16 layers of weight. This network is a fairly large network with approximately 138 million parameters. Figure 5c shows the VggNet-16 network structure, the number of parameters for each level, and the detailed performance test for all layers of the network [35,37,39].
VGG16 is a proposed model that compresses the previously successful VGG16 network and improves on the following aspects: (1) smaller model size, (2) quicker speed, (3) leverages residual learning for faster convergence, better generalization, and degradation resolution, and (4) equals the recognition accuracy of the noncompressed model on the very large-scale grand challenge MIT Places 365-Standard image dataset. The suggested model is 88.4% less in size and 23.86% faster in training time than VGG16.
This backs up our contention that the proposed model takes the best features of VGG16 and improves on them.
This work trains deep models of VGGNet-16 to screen DR as deep models achieve great success in many tasks. VGGNet exhibits excellent performance but is not better than ResNet-101 and ResNet-50.
  • ResNet (50 and 101)
Ref. [34] proposes ResNet, which was utilized to successfully train 152 deep neural networks to win the ILSVRC 2015 championship and obtain a 3.57% error rate classification for the top 5 classes using the residual unit, which is extremely good considering the number of parameters is smaller than that in VggNet. HighWay Nets, the ResNet core, uses the skip connection to let some input into (skip) the layer to integrate the information flow indiscriminately, preventing information loss and gradient vanishing (which also suppresses the generation of some noise). Furthermore, noise suppression entails averaging the model, ensuring that the model maintains a balance in training accuracy and generalization. Ultra-deep neural network training and model accuracy can be greatly improved using the ResNet structure. ResNet solves the degradation problem by increasing the CNN depth. The accuracy initially increases, reaches the limit, and then decreases as the depth increases. This phenomenon is not an overfitting issue because the error increases not only in the test samples but also in the training samples. When a shallow medium network satisfies the saturation accuracy and contains several congruence mappings layers, the errors will not accumulate when the network is at its smallest, and the deeper the network is, the fewer the training example errors will be. ResNet was inspired by the idea of employing a congruent mapping to send the previous output directly to the next layer. If a CNN’s input is x and the expected output is H(x), our learning goal is F(x) = H(x) − x when input x is directly transferred to the output.
A deep convolutional neural network with 50 layers is called ResNet-50. The ImageNet database contains a pretrained version of the network that can be loaded. This network has been trained on more than a million images. As a result, the network has acquired knowledge about several rich feature representations for numerous photos. ResNet-101 is a deep convolutional neural network with 101 layers. The ImageNet database, which has been trained on more than a million images, allows you to import a pretrained version of the network. In order to classify photos into 1000 different object categories, the network has already been trained. The network has learnt a range of rich feature representations for a variety of pictures as a consequence.
This study trained deep models VggNet-16, ResNet-50, and ResNet-101 to screen DR as deep models achieve great success in many tasks. For example, these methods have been cited for their good performance in image classification [39]. Figure 5a–c shows the ResNet-101, ResNet-50, and VggNet-16, respectively, network structures used to train our lab dataset for classifying DR images. ResNet-101 provides better results than ResNet-50 and VggNet-16.

3.1.4. Training Process

This work chose the latest CNN models (ResNet-101, ResNet-V1-50, and VGG-16) to train and classify our lab dataset because they are very popular for their effectiveness and robustness. They are designed in such a way that the input retinal image of size 300 × 300 can be fed into the network, which has alternate convolution and pooling layers activated using the ReLU activation function. The latest CNN models have been implemented in Python using Tensorflow. The parameters of the hyperparameter-tuning method are not initialized by the network itself; it is necessary to tune and optimize these parameters according to the results of training the DR image in enhancing performance. In addition, some of these parameters used in the proposed methods are mentioned in Table 2.
In this work, 1607 images from the Xiangya No. 2 Hospital Ophthalmology Changsha China dataset and 1000 images from this dataset are healthy and 607 images are defective. Out of 1000 healthy images, 800 images are used for training, and the remaining 200 images are used for testing; out of the 607 defective images, 485 images are used for training and 122 for testing. The images are fed into the network as tensors, and the network is trained iteratively. The hyperparameters of the proposed models are fine-tuned to achieve better performance. The number of iterations for every retinal image is set as 4000, and the learning rate is chosen as 0.0005. Parameter tuning for the proposed models is given in Table 2.

4. Experiment

This work started by using some preprocessing methods. The input image was resized and augmented and the data normalized to increase the amount of training data and improve the performance of using a deep training network for this work’s laboratory dataset. Then, the three state-of-the-art CNN models were used to detect DR, and the accuracy of these models was tested to classify the images. In addition to measuring the number of correctly and incorrectly classified test images, the images must also be evaluated according to the categories of error classification and discipline the accuracy scores accordingly.

4.1. Dataset and Preprocessing

The images were obtained from Xiangya No. 2 Hospital Ophthalmology (XHO), Changsha, China, for the study of two causes or classes of diabetic patients with and without symptoms. Initially, the datasets were indivisible, unmodified, and dirty due to many datasets covering many diseases, which are not diabetes. We then modified, arranged, and divided them into two appropriate categories of diabetes with and without symptoms. Second, all the images used were preprocessed to avoid noise and poor illumination. Before training the collected dataset, this study used many preprocessing methods to prepare the input images. The preprocessing included cropping and scaling the images. Each image was cropped to a square form, which includes the circular section of the fun-dust that is most securely enclosed. The images acquired are of size 1956 × 1934 pixels and occupy a large space. This requires higher RAM usage, which leads to low computation. Hence, they are downscaled to 300 × 300 dimensions before feeding them into the framework. This helps speed up the performance.
The number of images used with symptoms is 607, while 1000 did not have symptoms. Given the insufficient number of images for training, this study increased the size of the dataset by using the augmentation steps as shown in Table 3. Furthermore, given the presence of noise and the limited number of datasets, increasing the data is always preferred. Therefore, a data expander was used to generate more data without actually collecting more data. This work used two different types of OpenCV methods to increase the training set by rotating images and providing noisy images to improve model accuracy, as shown in Figure 6. The XHO dataset was divided into 800 for the no symptoms class and 485 for the symptoms class in the training dataset before the augmentation operation. The images in the testing dataset are 322. After the augmentation, the datasets increased to 1982 for the no symptoms class and 1204 for the symptoms class in the training dataset. The initial learning rate was 0.001, the momentum was 0.9, the batch size was 800, and the iterations were 3000.

4.2. Results and Discussion

The first experiment was on 16 layers of the VGG network for classifying the input images. The different scales of the 16 CNN layers and the max-pooling layers after each different scale were used to avoid overfitting and speed the processing operation. VGG-16 successfully obtained the training and testing classes, as shown in Table 4. Meanwhile, the accuracy is not very high because of the problem of vanishing gradients, which results in an insufficient number of layers for extracting the many features of the input images.
The second experiment used ResNet-50. This network contains different scales of 50 CNN layers with residual blocks to avoid the vanishing gradients problem. As shown in Figure 5, each block has a particular number of CNN layers and scale, and the input of each block is added to the output of the same block, helping enrich and increase the feature extraction. The accuracy of this experiment is higher than that of VGG-16 because of the ResNet properties.
To improve the model accuracy, we increased the number of layers in Resnet to 101 in the last experiment. The advantage of ResNet-101 is that it has more layers, is easier to overfit, and has a deeper network structure than ResNet-50. ResNet-101 has 101 layers, helping obtain more features of the input images. ResNet-101 achieved the highest accuracy in terms of classifying the input images for both the training and testing datasets, as shown in Table 4 and Table 5. The cross-entropy loss function was used to evaluate the training operation in all experiments to adjust the model weights during training. In addition, we used the arithmetic average to assess the model accuracy when the model was tested with the testing dataset.
The learning performance of the CNN models (ResNet-101, ResNet-50, and VGG-16) are shown graphically in Figure 7a,b, Figure 8a,b and Figure 9a,b, respectively. These figures show the plot of the number of iterations on the x-axis against the accuracy and loss on the y-axis. The ResNet-101 model exhibited promising results and robust stability in classifying the images as either healthy or afflicted by scanning the existence of DR. The designed model proves efficient in classifying the images present in the XHO datasets because the ResNet-101 model has a higher accuracy than ResNet-50 and VggNet-16.
The classification accuracies are shown in Table 4 and Table 5. The tables indicate that the results obtained using the CNN models with data augmentation are better than those obtained without data augmentation because data augmentation can assist CNN models in dealing with minor rotations or translations during data recording. Additionally, because of the imbalanced distribution of the XHO dataset in which the number of normal images was more than that of the DR images, we first classified the images into two DR stages and split them into the training and testing data sets. Then, image preprocessing was performed to improve the quality of the retinal images, which is important because low-quality images can degrade network performance. Thus, ensuring the consistency of all images and the enhancement of the features of the images are important. Thus, as networks and datasets improve and real-time classifications become available, the potential usefulness of CNN models (ResNet-101, ResNet-50, and VGG-16) to DR clinicians will continue to increase.

4.3. Performance Evaluation

To assess the performance of the proposed system ResNet-101 in DR detection on our dataset and other datasets, we used a high-resolution fundus (HRF) image, which contains 30 images of dimension 3304 × 2336 pixels [53]; STructured Analysis of the Retina (STARE), which contains 20 images from the STARE dataset of size 700 × 605 pixels [54]; DIARETDB0, which contains 130 images from the DIARETDB0 dataset of size 1500 × 1152 pixels [55]; 1200 images from the MESSIDOR dataset of resolution 1440 × 960 [56]; and our database XHO, which contains 1607 images of size 300 × 300.
The testing dataset is divided into two categories: images with no DR and images with DR recognized by ophthalmologists who divided the 2987 images with DR into 1089 images. No DR was observed when the suggested approach was tested on normal (no DR) pictures. Table 6 shows a detailed summary of all datasets used.

4.4. Performance Evaluation Criteria

The measures used to evaluate the effectiveness of CNNs include accuracy (ACC), specificity (SP), sensitivity (SEN), area under the ROC curve (AUC), positive predictive value (PPV), negative predictive value (NPV), and F1 score (F1). The amount of correctly identified photos is known as the ACC. The SP is the proportion of correctly categorized photos as normal images, while the SEN is the proportion of correctly classified images as DR images. The ROC curve and the value calculated by ROC, and the AUC provide a graphic representation of the ratio between SEN and SP. PPV is the proportion of correctly categorized DR pictures, whereas NPV is the proportion of correctly labeled normal images. These variables are used to evaluate how ResNet-101 performs in comparison to other algorithms. The following defines these metrics:
SP = TN ( TN + FP )
SEN = TP ( TP + FN )
ACC = ( TN + TP ) ( TN + TP + FN + FP )
PPV = TP ( TP + FP )
NPV = TN ( TN + FN )
F 1 score = 2 Precision Recall Precision + Recall
where false negative (FN) values refer to DR pictures that are categorized as non-DR, whereas false positive (FP) values refer to non-DR images that are labeled as DR. The terms TP and TN stand for true positive and true negative values, respectively, and relate to the DR pictures that are categorized as DR and the non-DR images that are not.
ResNet-101 is implemented utilizing many assessment criteria, including ACC, SEN, SP, and F1. The usage of 2987 retinal fundus pictures from four distinct datasets. For the detection of DR, an assessment of the proposed system ResNet-101 is conducted. The results of DR detection using various datasets were shown in Table 7 for all the accuracy (ACC), specificity (SP), sensitivity (SEN), area under the ROC curve (AUC), and F1 score metrics (F1). The AUC values are evaluated for the HRF, DRIVE, STARE, MESSIDOR, DIARETDB0, and DIARETDB1 datasets. The ROC curve plot is shown in Figure 10. In Table 8, the findings are compared with several current state-of-the-art methods for each dataset to determine their superiority and efficacy.

5. Conclusions

The automatic classification of DR fundus images can aid doctors in the diagnosis of DR and accelerate the diagnosis process. In this study, we first established the preprocessing and regularization steps that our lab dataset images require for maximizing the deep learning system’s functionality. Then, we looked into how much a small number of images used in training affects the performance. Second, our work has shown that the two-class problem for national screening of DR can be approached using a CNN method. The latest CNN models (ResNet-101, ResNet-50, and VggNet-16) show promising signs of being able to learn the features required to classify fundus images. The ResNet-101 model was tested on a set of images from our lab datasets. The network achieved a testing accuracy of 98.8198758%, whereas the network provides a training accuracy of 98.88% with a training loss of 34.99%. Finally, the ResNet-101 method was applied to 1787 color fundus images from the four (HRF, STARE, DIARETDB0, MESSIDOR, and XHO) datasets, and achieved an accuracy of 98%, 100%, 95%, 95.45%, and 97% for the all datasets, respectively. Therefore, the latest CNN models (ResNet-101, ResNet-50, and VggNet-16) can be trained to identify the features of DR in fundus images.
Our data sets’ quality and balance are essential for developing a system for DR detection. To create a balance for these datasets, we plan to merge several datasets in the future. Although the CNN models work well in detecting DR in two stages of our datasets, they have limitations in terms of the classification of system performance, so our goal is to extend and improve these models by adding more layers or by using CNN’s new model design, which can categorize and produce results in real time.

Author Contributions

Conceptualization, A.-O.A. and C.-Z.Z.; methodology, M.A.A.A.-Q.; software, D.A.-A.; validation, A.-O.A., D.A.-A. and S.A.A.; formal analysis, P.-B.O.; investigation, A.-O.A.; resources, S.A.A.; data curation, Y.-L.X.; writing—original draft preparation, A.-O.A.; writing—review and editing, M.A.A.A.-Q.; visualization, C.-Z.Z.; supervision, C.-Z.Z.; project administration, A.-O.A.; funding acquisition, S.A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key R&D Program of China (2018AAA0102100), the Scientific and Technological Innovation Leading Plan of High-tech Industry of Hunan Province (2020GK2021), the Natural Science Foundation of Hunan Province (No. 2022JJ30762), the International Science and Technology Innovation Joint Base of Machine Vision and Medical Image Processing in Hunan Province (2021CB1013), the Key Research and Development Program of Hunan Province (2022SK2054), the 111 project under grant no. B18059.

Data Availability Statement

All of the data in this report come from the Xiangya No. 2 Hospital Ophthalmology (XHO), Changsha, China. In addition, it is available as requested.

Acknowledgments

We want to express our thanks and gratitude to all those who helped us throughout this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Reguant, R.; Brunak, S.; Saha, S. Understanding inherent image features in CNN-based assessment of diabetic retinopathy. Sci. Rep. 2021, 11, 9704. [Google Scholar] [CrossRef]
  2. Aguirre, F.; Brown, A.; Cho, N.H.; Dahlquist, G.; Dodd, S.; Dunning, T.; Hirst, M.; Hwang, C.; Magliano, D.; Patterson, C.; et al. IDF Diabetes Atlas, 6th ed.; International Diabetes Federation: Basel, Switzerland, 2013. [Google Scholar]
  3. Ting, D.S.; Tan, K.A.; Phua, V.; Tan, G.S.; Wong, C.W.; Wong, T.Y. Biomarkers of Diabetic Retinopathy. Curr. Diabetes Rep. 2016, 16, 125. [Google Scholar] [CrossRef]
  4. Bourne, R.R.; Stevens, G.A.; White, R.A.; Smith, J.L.; Flaxman, S.R.; Price, H.; Jonas, J.B.; Keeffe, J.; Leasher, J.; Naidoo, K.; et al. Causes of vision loss worldwide, 1990–2010: A systematic analysis. Lancet Glob. Health 2013, 1, e339–e349. [Google Scholar] [CrossRef]
  5. Cheung, N.; Mitchell, P.; Wong, T.Y. Diabetic retinopathy. Lancet 2010, 376, 124–136. [Google Scholar] [CrossRef]
  6. Kroenke, K. Telemedicine screening for eye disease. JAMA 2015, 313, 1666–1667. [Google Scholar] [CrossRef] [PubMed]
  7. DCCT/EDIC Research Group; Nathan, D.M.; Bebu, I.; Hainsworth, D.; Klein, R.; Tamborlane, W.; Lorenzi, G.; Gubitosi-Klug, R.; Lachin, J.M. Frequency of Evidence-Based Screening for Retinopathy in Type 1 Diabetes. N. Engl. J. Med. 2017, 376, 1507–1516. [Google Scholar]
  8. Chan, J.C.; Malik, V.; Jia, W.; Kadowaki, T.; Yajnik, C.S.; Yoon, K.H.; Hu, F.B. Diabetes in Asia: Epidemiology, risk factors, and pathophysiology. JAMA 2009, 301, 2129–2140. [Google Scholar] [CrossRef] [PubMed]
  9. Congdon, N.G.; Friedman, D.S.; Lietman, T. Important causes of visual impairment in the world today. JAMA 2003, 290, 2057–2060. [Google Scholar] [CrossRef] [PubMed]
  10. Marshall, S.M.; Flyvbjerg, A. Prevention and early detection of vascular complications of diabetes. BMJ 2006, 333, 475–480. [Google Scholar] [CrossRef] [PubMed]
  11. Hutchinson, A.; McIntosh, A.; Peters, J.; O’keeffe, C.; Khunti, K.; Baker, R.; Booth, A. Effectiveness of screening and monitoring tests for diabetic retinopathy—A systematic review. Diabet. Med. 2000, 17, 495–506. [Google Scholar] [CrossRef]
  12. Taylor, R.; Batey, D. Handbook of Retinal Screening in Diabetes: Diagnosis and Management; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  13. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  14. Ting, D.S.W.; Cheung, C.Y.L.; Lim, G.; Tan, G.S.W.; Quang, N.D.; Gan, A.; Hamzah, H.; Garcia-Franco, R.; San Yeo, I.Y.; Lee, S.Y.; et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA 2017, 318, 2211–2223. [Google Scholar] [CrossRef]
  15. Chaudhuri, S.; Chatterjee, S.; Katz, N.; Nelson, M.; Goldbaum, M. Detection of blood vessels in retinal images using two-dimensional matched filters. IEEE Trans. Med. Imaging 1989, 8, 263–269. [Google Scholar] [CrossRef]
  16. Sinthanayothin, C.; Boyce, J.F.; Cook, H.L.; Williamson, T.H. Automated localisation of the optic disc, fovea, and retinal blood vessels from digital colour fundus images. Br. J. Ophthalmol. 1999, 83, 902–910. [Google Scholar] [CrossRef]
  17. Baudoin, C.E.; Lay, B.J.; Klein, J.C. Automatic detection of micro aneurysms in diabetic fluorescein angiography. Rev. D’épidémiol. St. Publique 1984, 32, 254–261. [Google Scholar]
  18. Silberman, N.; Ahrlich, K.; Fergus, R.; Subramanian, L. Case for automated detection of diabetic retinopathy. In Proceedings of the 2010 AAAI Spring Symposium Series, Stanford, CA, USA, 22–24 March 2010; Available online: https://nyuscholars.nyu.edu/en/publications/case-for-automated-detection-of-diabetic-retinopathy (accessed on 24 May 2022).
  19. Lowe, D.G. Distinctive image features from scale-invariant key points. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  20. Lindeberg, T. Scale-invariant feature transform. Scholarpedia 2012, 7, 10491. [Google Scholar] [CrossRef]
  21. Pires, R.; Jelinek, H.F.; Wainer, J.; Valle, E.; Rocha, A. Advancing bag-of-visual-words representations for lesion classification in retinal images. PLoS ONE 2014, 9, e96814. [Google Scholar] [CrossRef]
  22. Bay, H.; Tuytelaars, T.; Van Gool, L. Surf: Speeded up robust features. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2006; pp. 404–417. [Google Scholar]
  23. Séoud, L.; Faucon, T.; Hurtut, T.; Chelbi, J.; Cheriet, F.; Langlois, J.P. Automatic detection of microaneurysms and haemorrhages in fundus images using dynamic shape features. In Proceedings of the 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI), Beijing, China, 29 April 2014–2 May 2014; pp. 101–104. [Google Scholar]
  24. Seoud, L.; Chelbi, J.; Cheriet, F. Automatic grading of diabetic retinopathy on a public database. In Ophthalmic Medical Image Analysis International Workshop; University of Iowa: Iowa City, IA, USA, 2015. [Google Scholar]
  25. Paranjpe, M.J.; Kakatkar, M.N. Automated diabetic retinopathy severity classification using support vector machine. Int. J. Res. Sci. Adv. Technol. 2013, 3, 86–91. [Google Scholar]
  26. Liaw, A.; Wiener, M. Classification and regression by random Forest. R News 2002, 2, 18–22. [Google Scholar]
  27. Joachims, T. Making large-scale SVM learning practical. In Technical Report; University of Dortmund: Dortmund, Germany, 1998; Available online: https://www.econstor.eu/handle/10419/77178 (accessed on 4 August 2022).
  28. Walter, T.; Klein, J.C.; Massin, P.; Erginay, A. A contribution of image processing to the diagnosis of diabetic retinopathy-detection of exudates in color fundus images of the human retina. IEEE Trans. Med. Imaging 2002, 21, 1236–1243. [Google Scholar] [CrossRef] [PubMed]
  29. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In European Conference on Computer Vision, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 818–833. [Google Scholar]
  30. Wang, Z.; Yang, J. Diabetic retinopathy detection via deep convolutional networks for discriminative localization and visual explanation. In Proceedings of the Workshops at the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 8 February 2017. [Google Scholar]
  31. Yang, Y.; Li, T.; Li, W.; Wu, H.; Fan, W.; Zhang, W. Lesion detection and grading of diabetic retinopathy via two-stages deep convolutional neural networks. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2017; pp. 533–540. [Google Scholar]
  32. Wang, S.; Yin, Y.; Cao, G.; Wei, B.; Zheng, Y.; Yang, G. Hierarchical retinal blood vessel segmentation based on feature and ensemble learning. Neurocomputing 2015, 149, 708–717. [Google Scholar] [CrossRef]
  33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  34. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  35. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep Face Recognition. The University of Oxford Is a Collegiate Research University in Oxford, England. 2015. Available online: https://www.robots.ox.ac.uk/~vgg/publications/2015/Parkhi15/parkhi15.pdf (accessed on 15 August 2022).
  36. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:14091556. [Google Scholar]
  37. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  39. Wan, S.; Liang, Y.; Zhang, Y. Deep convolutional neural networks for diabetic retinopathy detection by image classification. Comput. Electr. Eng. 2018, 72, 274–282. [Google Scholar] [CrossRef]
  40. Alyoubi, W.L.; Abulkhair, M.F.; Shalash, W.M. Diabetic Retinopathy Fundus Image Classification and Lesions Localization System Using Deep Learning. Sensors 2021, 21, 3704. [Google Scholar] [CrossRef]
  41. Behera, M.K.; Mishra, R.; Ransingh, A.; Chakravarty, S. Prediction of different stages in Diabetic retinopathy from retinal fundus images using radial basis function based SVM. Indian J. Sci. Technol. 2020, 13, 2030–2040. [Google Scholar] [CrossRef]
  42. Sankar, M.; Batri, K.; Parvathi, R. Earliest diabetic retinopathy classification using deep convolution neural networks. Int. J. Adv. Eng. Technol. 2016, 10, M9. [Google Scholar]
  43. Lim, G.; Lee, M.L.; Hsu, W.; Wong, T.Y. Transformed representations for convolutional neural networks in diabetic retinopathy screening. In Proceedings of the Workshops at the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–28 July 2014. [Google Scholar]
  44. Pratt, H.; Coenen, F.; Broadbent, D.M.; Harding, S.P.; Zheng, Y. Convolutional neural networks for diabetic retinopathy. Procedia Comput. Sci. 2016, 90, 200–205. [Google Scholar] [CrossRef]
  45. Kaggle. Diabetic Rentinopathy Detection Challenge. 2016. Available online: http://www.kaggle.com/c/diabetic-retinopathy-detection/data (accessed on 1 June 2022).
  46. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  47. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv Prepr. 2013, arXiv:1312.4400. [Google Scholar]
  48. Lekshmi, S.A.; Rajathi, G.M. Detection of Glaucoma and Diabetic Retinopathy using Image Processing Technique by Raspberry pi. Indian J. Sci. Technol. 2019, 12, 29. [Google Scholar] [CrossRef]
  49. Bindu, C.H.; Sravanthi, G.S. Retinopathy Detection of EYE Images in Diabetic Patients. Indian J. Sci. Technol. 2019, 12, 3. [Google Scholar] [CrossRef]
  50. Sarki, R.; Ahmed, K.; Wang, H.; Zhang, Y.; Ma, J.; Wang, K. Image Preprocessing in Classification and Identification of Diabetic Eye Diseases. Data Sci. Eng. 2021, 6, 455–471. [Google Scholar] [CrossRef]
  51. Henry, A.G.P.; Jude, A. Convolutional neural-network-based classification of retinal images with different combinations of filtering techniques. Open Comput. Sci. 2021, 11, 480–490. [Google Scholar] [CrossRef]
  52. Frid-Adar, M.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H. Synthetic data augmentation using GAN for improved liver lesion classification. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging, Washington, DC, USA, 4–7 April 2018; pp. 289–293. [Google Scholar]
  53. University of Erlangen-Nuremberg. High Resolution Fundus (HRF) Image Database; University of Erlangen-Nuremberg: Erlangen, Germany, 2014; Available online: http://www5.cs.fau.de/research/data/fundus-images/ (accessed on 2 May 2021).
  54. Goldbaum, M. Structured Analysis of the Retina. 2013. Available online: http://www.ces.clemson.edu/ahoover/stare/ (accessed on 24 May 2021).
  55. DIARETDB0 Database. Medical Image Understanding and Analysis. March 2014. Available online: http://www.it.lut.fi/project/imageret/diaretdb (accessed on 24 March 2022).
  56. Decenciere, E.; Zhang, X.; Cazuguel, G.; Lay, B.; Cochener, B.; Trone, C. Feedback on a publicly distributed image database: The MESSIDOR database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef]
  57. Li, T.; Gao, Y.; Wang, K.; Guo, S.; Liu, H.; Kang, H. Diagnostic assessment of deep learning algorithms for diabetic retinopathy screening. Inf. Sci. 2019, 501, 511–522. [Google Scholar] [CrossRef]
  58. Liu, Y.P.; Li, Z.; Xu, C.; Li, J.; Liang, R. Referable diabetic retinopathy identification from eye fundus images with weighted path for convolutional neural network. Artif. Intell. Med. 2019, 99, 101694. [Google Scholar] [CrossRef]
  59. Jiang, H.; Yang, K.; Gao, M.; Zhang, D.; Ma, H.; Qian, W. An interpretable ensemble deep learning model for diabetic retinopathy disease classification. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 2045–2048. [Google Scholar]
  60. Das, S.; Kharbanda, K.; Suchetha, M.; Raman, R.; Dhas, E. Deep learning architecture based on segmented fundus image features for classification of diabetic retinopathy. Biomed. Signal Processing Control. 2021, 68, 102600. [Google Scholar] [CrossRef]
  61. Wang, X.; Lu, Y.; Wang, Y.; Chen, W.B. Diabetic retinopathy stage classification using convolutional neural networks. In Proceedings of the 2018 IEEE International Conference on Information Reuse and Integration (IRI), Salt Lake City, UT, USA, 6–9 July 2018; pp. 465–471. [Google Scholar]
  62. Khan, S.H.; Abbas, Z.; Rizvi, S.D. Classification of diabetic retinopathy images based on customised CNN architecture. In Proceedings of the 2019 Amity International Conference on Artificial Intelligence (AICAI), Dubai, United Arab Emirates, 4–6 February 2019; pp. 244–248. [Google Scholar]
  63. Shanthi, T.; Sabeenian, R.S. Modified Alexnet architecture for classification of diabetic retinopathy images. Comput. Electr. Eng. 2019, 76, 56–64. [Google Scholar] [CrossRef]
Figure 1. Stages of diabetic retinopathy.
Figure 1. Stages of diabetic retinopathy.
Electronics 11 02740 g001
Figure 2. Study of the research gap of Retinopathy using CNN to segment the lesions.
Figure 2. Study of the research gap of Retinopathy using CNN to segment the lesions.
Electronics 11 02740 g002
Figure 3. Sample frames of the retina images. The first two frames in the top row come from normal subjects, while the two frames in the bottom row come from the patients who have diabetic retinopathy.
Figure 3. Sample frames of the retina images. The first two frames in the top row come from normal subjects, while the two frames in the bottom row come from the patients who have diabetic retinopathy.
Electronics 11 02740 g003
Figure 4. Illustration of the preprocessing and augmentation processes. (a) Original images. (b) Preprocessed images. (c) Augmented images.
Figure 4. Illustration of the preprocessing and augmentation processes. (a) Original images. (b) Preprocessed images. (c) Augmented images.
Electronics 11 02740 g004
Figure 5. Latest CNN (a) ResNet-101, (b) ResNet-V1-50, and (c) VGG-16 Architecture [39].
Figure 5. Latest CNN (a) ResNet-101, (b) ResNet-V1-50, and (c) VGG-16 Architecture [39].
Electronics 11 02740 g005
Figure 6. Some examples of images in the training set and the output of preprocessing to increase the size of the input dataset. (a) Images have symptoms. (b) Images without any symptoms. (c) Images with left and right rotate. (d) Images with noise.
Figure 6. Some examples of images in the training set and the output of preprocessing to increase the size of the input dataset. (a) Images have symptoms. (b) Images without any symptoms. (c) Images with left and right rotate. (d) Images with noise.
Electronics 11 02740 g006
Figure 7. Training process of ResNet-101 for both (a) accuracy and (b) loss.
Figure 7. Training process of ResNet-101 for both (a) accuracy and (b) loss.
Electronics 11 02740 g007aElectronics 11 02740 g007b
Figure 8. Training process of ResNet-50 for both (a) accuracy and (b) loss.
Figure 8. Training process of ResNet-50 for both (a) accuracy and (b) loss.
Electronics 11 02740 g008
Figure 9. Training process of VggNet-16 for both (a) accuracy and (b) loss.
Figure 9. Training process of VggNet-16 for both (a) accuracy and (b) loss.
Electronics 11 02740 g009
Figure 10. ROC plot for retinal HRF, DRIVE, STARE, MESSIDOR, DIARETDB0, and XHO datasets.
Figure 10. ROC plot for retinal HRF, DRIVE, STARE, MESSIDOR, DIARETDB0, and XHO datasets.
Electronics 11 02740 g010
Table 1. Classification dataset.
Table 1. Classification dataset.
Class NameDegree of DRNumber of LabelsFor TrainingFor Testing
Class 0Nonsymptoms1000800200
Class 1Symptoms607485122
Table 2. Parameter tuning in the latest CNN models.
Table 2. Parameter tuning in the latest CNN models.
DescriptionOutput Shape
weight_decay0.00001
num_classes2
batch_size8
val_batch_size32
image_height300
image_width300
image_channels3
num_iters4000
lr [0.0001, 0.001]
momentum0.9
batch_norm_scaleTrue
batch_norm_epsilon1 × 10−5
batch_norm_decay0.997
Table 3. The number of images before and after the augmentation process.
Table 3. The number of images before and after the augmentation process.
ClassAllTraining SetTasting Set
Before
Augmentation
After
Augmentation
No symptoms10008001982200
Symptoms6074851204122
Table 4. Performance comparison with different approaches before augmentation.
Table 4. Performance comparison with different approaches before augmentation.
ModelTraining Accuracy (%)Training Loss (%)Testing Accuracy (%)
ResNet-10180.8840.179.76
ResNet-5079.737.671.8
VggNet-1664.467.3262.11
Table 5. Performance comparison with different approaches after augmentation.
Table 5. Performance comparison with different approaches after augmentation.
ModelTraining Accuracy (%)Training Loss (%)Testing Accuracy (%)
ResNet-10198.8834.9998.82
ResNet-50933491.5
VggNet-1671.3961.4864.11
Table 6. Complete description of database.
Table 6. Complete description of database.
DatasetNumber of ImagesNo DRDR
XHO datasets16071000607
HRF datasets301515
STARE datasets20128
DIARETDB0 datasets13020110
MESSIDOR 1200851349
Total298718981089
Table 7. Summary of DR detection for four different datasets using ResNet-101 CNN method.
Table 7. Summary of DR detection for four different datasets using ResNet-101 CNN method.
DatasetTest ImagesCorrectly DetectedAccuracy (%)Sensitivity
(%)
Specificity
(%)
F1 Score (%)AUC (%)
XHO datasets2001969897.1497.6597.3698.55
HRF datasets303010099.9899.9899.9899.99
STARE datasets20199594.9695.1195.0395.04
DIARETDB0 datasets11010595.4595.3999.3895.4595.46
MESSIDOR34934799.4299.4599.3899.4199.42
Total3603499796.8798.0396.9597.26
Table 8. Comparison between the related works that used CNN with our work to classify DR Images.
Table 8. Comparison between the related works that used CNN with our work to classify DR Images.
Evaluated Parameter [57] [58] [59]Proposed Method [60] [39] [61] [62] [63]
Number of Classes2545
Detect LesionNoNoNoYes
Datasetprivate datasetKaggleMessidorDDR
Performance MeasureACC94.23%88.21%98.7%98.88%63.23%95.6%98.15%96.35%82.84%
AUC0.98230.946-98.55%-0.978---
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Asia, A.-O.; Zhu, C.-Z.; Althubiti, S.A.; Al-Alimi, D.; Xiao, Y.-L.; Ouyang, P.-B.; Al-Qaness, M.A.A. Detection of Diabetic Retinopathy in Retinal Fundus Images Using CNN Classification Models. Electronics 2022, 11, 2740. https://doi.org/10.3390/electronics11172740

AMA Style

Asia A-O, Zhu C-Z, Althubiti SA, Al-Alimi D, Xiao Y-L, Ouyang P-B, Al-Qaness MAA. Detection of Diabetic Retinopathy in Retinal Fundus Images Using CNN Classification Models. Electronics. 2022; 11(17):2740. https://doi.org/10.3390/electronics11172740

Chicago/Turabian Style

Asia, Al-Omaisi, Cheng-Zhang Zhu, Sara A. Althubiti, Dalal Al-Alimi, Ya-Long Xiao, Ping-Bo Ouyang, and Mohammed A. A. Al-Qaness. 2022. "Detection of Diabetic Retinopathy in Retinal Fundus Images Using CNN Classification Models" Electronics 11, no. 17: 2740. https://doi.org/10.3390/electronics11172740

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop