Next Article in Journal
Optical Coherence Tomography Angiography Is a Useful Tool for Distinguishing Primary Raynaud’s Phenomenon from Systemic Sclerosis and/or Very Early Disease of Systemic Sclerosis
Next Article in Special Issue
Evaluation of Augmentation Methods in Classifying Autism Spectrum Disorders from fMRI Data with 3D Convolutional Neural Networks
Previous Article in Journal
Machine Learning Approach for the Prediction of In-Hospital Mortality in Traumatic Brain Injury Using Bio-Clinical Markers at Presentation to the Emergency Department
Previous Article in Special Issue
A Coarse-to-Fine Fusion Network for Small Liver Tumor Detection and Segmentation: A Real-World Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Detection and Classification of Diabetic Retinopathy Using the Improved Pooling Function in the Convolution Neural Network

by
Usharani Bhimavarapu
1,
Nalini Chintalapudi
2 and
Gopi Battineni
2,3,*
1
Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram 522302, India
2
Clinical Research Centre, School of Medicinal and Health Products Sciences, University of Camerino, 62032 Camerino, Italy
3
The Research Centre of the ECE Department, V. R. Siddhartha Engineering College, Vijayawada 520007, India
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(15), 2606; https://doi.org/10.3390/diagnostics13152606
Submission received: 7 July 2023 / Revised: 30 July 2023 / Accepted: 2 August 2023 / Published: 5 August 2023
(This article belongs to the Special Issue Artificial Intelligence in Clinical Medical Imaging)

Abstract

:
Diabetic retinopathy (DR) is an eye disease associated with diabetes that can lead to blindness. Early diagnosis is critical to ensure that patients with diabetes are not affected by blindness. Deep learning plays an important role in diagnosing diabetes, reducing the human effort to diagnose and classify diabetic and non-diabetic patients. The main objective of this study was to provide an improved convolution neural network (CNN) model for automatic DR diagnosis from fundus images. The pooling function increases the receptive field of convolution kernels over layers. It reduces computational complexity and memory requirements because it reduces the resolution of feature maps while preserving the essential characteristics required for subsequent layer processing. In this study, an improved pooling function combined with an activation function in the ResNet-50 model was applied to the retina images in autonomous lesion detection with reduced loss and processing time. The improved ResNet-50 model was trained and tested over the two datasets (i.e., APTOS and Kaggle). The proposed model achieved an accuracy of 98.32% for APTOS and 98.71% for Kaggle datasets. It is proven that the proposed model has produced greater accuracy when compared to their state-of-the-art work in diagnosing DR with retinal fundus images.

1. Introduction

Glucose in the body is converted into energy, which helps with everyday tasks. Diabetes is caused by obesity, poor nutrition, and limited physical activity. However, elevated blood glucose can build up in the blood vessels of several human organs, including the eye. People who have had diabetes for over a decade have the chance of getting diabetic retinopathy (DR) [1]. Globally, the population suffering from diabetes is expected to reach 552 million by 2030 [2]. Preventing visual loss is possible with early detection and sufficient treatment [3]. DR consists of five classes—no DR, mild, moderate, severe, and proliferative.
DR can affect blood vessels, in severe cases damaging, enlarging, or blocking them, or causing leaks; the abnormal growth of blood vessels can cause total blindness. Micro-aneurysms, haemorrhages, and exudates are the major signs of retinal DR. The level of the disease can be identified based on the shape, size, and overall appearance of the lesions. The main benefits of DR screening are its high effectiveness, low cost and minimal reliance on clinicians (i.e., ophthalmologists). The global eye screening tool for DR is the fundus photograph [4]. To prevent diabetes-related blindness, automated screening allows for clinically convenient and cost-effective detection [5].
From the field of computer science, deep learning can be a practical approach to automatic DR detection [6]. A deep learning system automatically identifies the DR with an accuracy that is equal to or better than that of ophthalmologists. The core deep learning model for medical image diagnosis prediction, and classification is the convolution neural network (CNN). However, there is the possibility to improve the performance of the model by tuning the hyperparameters in these deep learning-based models.
CNN models AlexNet and VGGNet-16 have been implemented for this purpose and the results suggest that VGG-19 performs best; however, the DR stages have not been explicitly ranked [7]. A hybrid technique incorporating image processing and deep learning was proposed for the detection and classification of DR in the publicly available dataset MESSIDOR, and Histogram Equalization (HE) and Contrast Limited Adaptive Histogram Equalization (CLAHE) were implemented to improve the contrast of the image [8]. Other CNN models, like Inception V3, Dense 121, Xception, Dense 169, and ResNet 50, have been explored for the enhanced classification of different DR phases [9].
In another study, the authors proposed a framework with a new loss function by implementing mid-level representations to improve DR detection performance [10]. Another report proved that VGGNet produced higher accuracy compared with other CNN models such as AlexNet, GoogleNet, and ResNet for DR classification [11]. A CNN model implementation with data augmentation for DR image classification was presented in [12].
Other frameworks for the early diagnosis and classification of DR were presented for Grampian [13], MESSIDOR [14], and EYEPACS datasets [15]. In [16], the authors mentioned that 90% of accuracy was achieved in diagnosing microaneurysms and extracting and classifying the candidate lesions. All of these existing studies have implemented built-in hyperparameters. However, model performance can be improved by adjusting hyperparameters within deep learning models. To counter the self-strengthening trend and ensure that as many candidate component models as possible have been properly trained, we have added balance loss to our model. The proposed approach could extract key features from the fundus images that can help make an accurate DR diagnosis.

2. Materials and Methods

The objective of the current study was to accurately categorize DR fundus images into different severities. We discussed an automated system for assessing the seriousness of diabetic retinopathy. The classification accuracy for diabetic retinopathy was improved in the current research using a modified CNN architecture. Figure 1 illustrates the proposed framework.

2.1. Dataset Collection

We collected the dataset from two publicly available fundus image datasets, i.e., APTOS [17] and Kaggle [18]. Table 1 tabulates the count for five categories in APTOS and Kaggle datasets. Figure 2 shows the sample fundus images from the two datasets. The first-row fundus images are from APTOS and the second-row fundus images are from the Kaggle dataset.
We employed data augmentation to increase the number of images throughout the training sample. Once provided with more DR to learn from, DL approaches generally improve their performance. Overfitting is avoided and the imbalance in the dataset is corrected by the application of data augmentation. Horizontal shift augmentation was one of the transformations considered for this study; it involves horizontally shifting an image’s pixels while maintaining the original image’s perspective. The dimension of this transition is specified by a number ranging from 0 to 1 and the viewing angle of the original image is preserved. The image can also be rotated with an additional type of transformation by a random amount between 0 and 180 degrees. By employing data augmentation methods, we were able to fix the problem of varying sample sizes and convoluted categorizations. After augmentation, the APTOS dataset classes were evenly distributed for the training set—1805 for NODR, 1850 for Mid, 1988 for Moderate, 1737 for Severe, and 1770 for PDR. After augmentation, the Kaggle dataset classes were evenly distributed for the training set—25,810 for NODR, 24,430 for Mid, 26,460 for Moderate, 25,317 for Severe, and 25,488 for PDR. Figure 3 shows some of the augmentation operations followed in this study. Table 2 tabulates the statistics of the data augmentation operations and the final augmented fundus images of each dataset.

2.2. Pre-Processing

In this study, we implemented the enhanced artificial bee colony (ABC) algorithm to improve the lesions’ visual contents. Consider ξ i , j ϵ D with dimensions PXQ, where the values of P, Q are taken as 512 for every image in the database D.
The mathematical representation of the transformation function,
Ξ f = 1 0 1 x c 1 1 x d 1 d x X 0 v x c 1 1 x d 1 d x ,
where x is an integration variable and c and d are adjustable parameters of a given function where the maximum value of c is compared with d.
We evaluated the fitness function to adjust the values of c and d and also to measure the complete lesion image.
F ( ξ H ( i , j ) ) =   log ( log j = 1 Ψ ) M Ψ E ξ H Y ξ H ,
where j = 1 Ψ represents the total edge intensities of an image evaluated through a canny edge detector.   Y ξ H represents the contrast of the image ξ H (i, j), M Ψ represents the total edge pixels of the processed image, and   E ξ H represents the image entropy ξ H (i, j), represented as:
  E ξ H   = j = 0 m q i log 2 q i ,
where q i represents the ith pixel intensity probability; the max value is 255.
The contrast of the image is represented as:
Y ξ H   =   j = 0 m I Y ξ H I i ,
where I i represents the image blocks and m I represents the mth image block.
The contrasted local band of each block is represented as:
ξ H y I i   = p , q ϵ I Y ξ H p , q = p , q I ξ H p , q ϕ b ξ H p , q ϕ c ,
where p , q represents the pixels of the rows and columns of each block, ϕ b represents the bandpass filter, and ϕ c represents the low pass filter.

2.3. Enhanced ResNet-50

The proposed model consists of convolution blocks and includes the improved pooling function, a drop-out layer, dense layers, and a SoftMax classification layer; Figure 4 presents the improved ResNet-50 model.
Convolution Layer: The convolutional block is the fundamental building component, and each convolution block contains a convolution 2D, an improved activation function, and improved pooling with the average value. The vanishing gradient issue is solved using the improved activation function, simplifying the process so the network can understand and carry out its tasks promptly.
Kernel: The model’s initial layer is the convolution layer. This layer initiates the process by applying the filters, also known as the kernel. The kernel size depends on two values—the width and height of the filter. In this study, we set the size of the filter as 3. This filter enables and identifies the features that help understand low-level visual aspects like edges and curves.
Flattened layer: The flattened layer is located among the convolution and the dense layers. Tensor datatypes are used as inputs for the convolution layers, whereas dense layers demand a one-dimensional layout. The flattened layer was applied to translate the two-dimensional image representation into a one-dimensional input.
Dropout Layer: A dropout value of 0.2 was used in this study, which helps to avoid overfitting. This layer’s function was to turn various components on and off to reduce the model’s complexity and training time. The model thus acquires all the features that are required.
Dense Layer: A single matrix is accepted as input by the dense layer, which produces output based on the characteristics of the matrix. The identification and class labelling of fundus images occurs in these layers. The model’s output is produced by a dense layer with five neurons and an improved activation function, and it assigns the image to one of five categories of diabetes: NoDR, Mild, Moderate, Severe, or Proliferative. After a few layers, the proposed activation is applied; this probability-based activation function measures the number of neurons by the entire number of classes.
Pooling function: The pooling function in the CNN is primarily used to downsample the feature maps and learn deeper image features that are resilient to subtle local alterations. The features from each spatial region are aggregated in this process. Pooling not only expands the receptive field of convolutional kernels across layers but also reduces memory needs and computational complexity by lowering the resolution of the feature maps while keeping critical features required for processing by the following layers. Pooling can be used in medical image analysis to manage variations in lesion sizes and positions [19,20]. Fundus images frequently have many lesions or parts, which causes their distributions of convolutional activations to be exceedingly complex since unimodal distributions cannot adequately capture statistics of convolutional activations, which limits the CNN performance.
We first pass Y throughout a group of prediction layers with parameters θ p , i.e., c θ p ; Y . The weights are outputted throughout by using a fully connected layer with additional noise.
The improved pooling function is presented as:
F k c θ p ; Y   =   T k h C θ p ; Y   + δ . l o g 1 + e x p T k m C θ p ; Y ,
where T k h and T k m are the fully connected layers, the kth parameter and additional noise, δ is the random variable, C θ p ; Y are the learned weights, and the weight function can be represented as:
w k Y   = e x p T O P Q F k c θ p ; Y k = 1 m e x p T O P Q F k c θ p ; Y ,
where TOP-Q are the Q largest weights.
To make learned weights sparse, we maintained the TOP-Q weights and set the remaining ones as negative infinity and we used the improved activation function to normalize all the weights.
We added extra loss using the learned weights:
L s   = 3 β S s = 1 N w k Y s M s = 1 N w k Y s ,
where Y s is the mini-batch training sample, S and M are the standard deviation and the mean, and β is the parameter.
The improved activation function, which was recommended as a replacement for the activation function ReLU, is represented as:
f x =   x / 2 ; i f 2 x < 2 1 ;   i f   x < 2 1 ;   i f   x > 2 .

2.4. Classification

We applied the improved SVM in this study to improve classification accuracy. Initially, the SVM calculates the score for all the extracted features by using linear mapping on feature vectors and uses this to evaluate the loss. The improved SVM uses the linear mapping on extracted features to calculate the feature score for the parts of the region of interest used to differentiate the lesion types, which helps in the evaluation of loss function, which helps to obtain the classification results. Algorithm 1 for the improved SVM is presented below.
Algorithm 1 Improved SVM
    •
Initialize the values in the training set.
    •
Repeat for j = 1 to M.

Calculate the loss using the enhanced optimization for all values of j.
Compare the extracted regions in the liver images.
end
    •
Repeat for every score vector j − 1 to M.

Compute the SVM
argmax((w × p j) + b)
end
    •
Compute for all weights and finally evaluate the output.

3. Results

All experiments were implemented on Keras. The data split was performed based on an 80:20 ratio, where 80% of the data were used for training and 20% for testing. We implemented the proposed pooling function and activation function in the base models VGG-16, DenseNet, ResNet-50, Xception, and AlexNet for the fundus images. Table 3 tabulates the splitting of training and testing sets of fundus images for two augmented datasets.

3.1. Image Enhancement Evaluation

Image enhancement is a vital concept that changes the intensities of the original image to improve the image’s perceptual quality. Figure 5 shows the contrast enhancement results for the APTOS dataset fundus image. Figure 5 compares the proposed model with some other existing enhancement models. Contrast-limited adaptive histogram equalization (CLAHE) models show insufficient image enhancement. The histogram modification framework (HMF) model enhances the image well; however, the hazy look is not adequately removed. The heuristic adaptive histogram equalization (HAHE) model produces an enhanced image with unwanted artefacts visible in the fundus image. The artificial bee colony algorithm (ABC) yields better results than the other existing models; still, it has some viewable artifacts in the fundus image. The proposed model generates an outstanding result compared to all other existing models and successfully improves every minor detail present in the fundus image.
Evaluation and assessment are important for analysing the proposed model performance quantitatively. The proposed image enhancement model is accessed with performance measures such as entropy, peak signal-to-noise ratio (PSNR), the structural similarity index measure (SSIM), gradient magnitude similarity deviation (GMSD), and the patch-based contrast quality index (PCQI) [21,22,23].
Entropy defines the amount of information contained in the processed image.
Entropy = y = 0 255 P n log 2 P n ;
where P(n) represents the probability of the nth level of the image.
PSNR computes the amount of noise content in the processed image.
PSNR =   20 l o g 10 2 1 A B , x = 0 A 1 y = 0 B 1 I 0 x , y I i x y 2 ,
where A, B denotes the image size.
SSIM =   ( 2 μ I i μ I o + A 1 ) ( 2 σ I i σ I o + A 2 ) ( μ I i 2 + μ I o 2 + A 1 ) ( σ I i 2 + σ I o 2 + A 2 ) ,
where μ I i ,   μ I o represents the input and the output intensity values, σ I i , σ I o represent the input and the output standard deviation values, and A1, A2 represent the constant to limit the instability problem.
Table 4 tabulates the average scores for the augmented APTOS dataset. The performance of the proposed model was demonstrated by comparing six state-of-the-art existing models such as Clahe [24], exposure-based sub-image histogram equalization (ESIHE) [25], HAHE [26], BIMEF [27], HMF [28], and ABC. From Table 4, it is clear that the proposed model achieves a higher SSIM value, and its similarity level is up to the mark when compared with the original fundus image. The proposed enhanced model attains a lesser GMSD value for the images and holds more excellent visual quality compared to the other methods. The proposed model gains a higher PSNR value and the noise suppression level is very good compared with that of the other models. The proposed model holds a higher entropy value to the original image and the amount of information preserved is high compared with the state-of-the-art models. The proposed model obtains a more significant PCQI value compared with the other models, and generates a good quality image with minimum structural distortions. The proposed enhanced model offers less running time when compared to the state-of-the-art contrast enhancement models. The running time of the CLAHE and ESIHE models is approximately equal to that of the proposed model. But these models suffer from noise and distortion. From Table 4, we can recognise that the proposed enhanced model is superior in enriching content, maintaining similarity, and suppressing the noise and distortion. The proposed enhanced image enhancement model generated a crisp and clear output.

3.2. Segmentation Comparison

The proposed model obtains more accurate and robust segmentation results. From Figure 6 it can be noticed that the proposed model obtains more accurate results.
Table 5 tabulates the performance of the proposed enhanced ResNet-50 compared to the state-of-the-art models. The proposed system performed very accurately compared with the other lesion segmentation methods in the state-of-the-art models. It saves the obtained accuracy of abnormal fundus images. It achieves accurate, detailed segmentation results with small lesions, so it is the perfect choice for automatic computer-aided diagnosis (CAD) systems that depend on lesion segmentation results as it exceeds the estimations of the alternative models in terms of overall accuracy.

3.3. Evaluation of the APTOS Dataset

Figure 7 illustrates the confusion matrix for the APTOS dataset. We implemented five baseline models—VGG-16, DenseNet, ResNet-50, Xception, and AlexNet—and compared their performances on the APTOS dataset. From these five models, ResNet-50 showed the highest performance.
According to the 5-class confusion matrix mentioned above, the performance of each model was evaluated based on accuracy, recall, precision, and F1-score. Table 6 tabulates the APTOS fundus classification test set results. The improved SVM model achieved the highest accuracy of the remaining classification models. The results show that the augmented APTOS fundus classification for the ResNet-50 model achieves the highest accuracy for the improved SVM model.

3.4. Evaluation of the Kaggle Dataset

Figure 8 illustrates the confusion matrix for the Kaggle dataset. We implemented five baseline models-VGG-16, DenseNet, ResNet-50, Xception, and AlexNet-and compared their performances on the Kaggle dataset. From these five models, ResNet-50 showed the highest performance. In 203 NODR fundus images, the proposed ISVM classifier accurately classified 202 fundus images for the ResNet-50 model. In 54 Mild images, the ISVM classifier accurately classified 54. Out of 69 moderate fundus images, ISVM accurately identified 68. Out of 15 images, ISVM accurately identified 14 for severe, and out of 7 images, ISVM accurately identified 6 for PDR for the ResNet-152 model. For the ResNet-50 model, the SVM classifier accurately identified 201 NODR images, 53 mild and 67 moderate, 14 severe, and 5 for PDR. For the ResNet-152 model, the RF classifier accurately identified 201 NODR images, 53 mild and 66 moderate, 13 severe, and 5 for PDR. For the ResNet-50 model, the NB classifier accurately identified 201 NODR images, 52 mild and 65 for moderate, 12 for severe, and 5 for PDR. Table 6 tabulated the Kaggle classification test set results.
From Table 7, we can see that the improved SVM model achieved the highest accuracy compared to the remaining classification models. The achieved results revealed that the overall testing accuracy and the performance metrics for the improved ResNet-50 with the improved SVM are the most appropriate for diabetic retinopathy detection, with a testing accuracy of 99.9% for fundus images.
Figure 9 presents the evaluation of the performance metrics for the different models. According to the achieved results, overall testing accuracy, and performance metrics, the proposed model is appropriate for detecting and classifying DR with a testing accuracy of 98.32% on the APTOS dataset.
Table 8 tabulates the varying sizes of the training and testing sets and the corresponding mean and standard deviation.

4. Discussion

This study aimed to identify and classify DR based on fundus images from two different datasets. Initially, all the images in the dataset were of different sizes; the images were resized to 225 × 225 using the RGB colour. The hyperparameters were tuned to optimize the proposed model. Model training can be accelerated, and the possibility of performance improved using the pooling function. There is no ideal batch size, and we implemented the experiments with various batch sizes. If we find the suitable batch size in addition to the suitable kernel and hidden layers, the model will yield a high performance. Batch size 64 produces better results than batch sizes 16 or 32. The batch size was 64 for the fundus images because this study’s dataset was large. From previous studies, we observed that the batch sizes, in conjunction with a suitable kernel and hidden layer, will yield a high performance. The parameters (i.e., a batch size of 64, epochs of 1000, and a learning rate of 0.001) were adjusted to achieve a high performance.
After extracting the features, the improved SVM classifies the lesions. In [15], the authors implemented AdaBoost to extract the features and the Gaussian mixture model, KNN, and SVM to classify the lesions and analyse the retina fundus images with different illuminations and views. A new unsupervised approach based on PCA for detecting microaneurysms was presented in [16]. The manual identification and differentiation of diabetic retinopathy from fundus images is time-consuming. Table 9 presents the processing time analysis of the existing techniques for the Kaggle and APTOS datasets to calculate the computation overhead. The achieved results revealed that the overall processing time for the improved SVM classifier is the most appropriate for diabetic retinopathy classification, with a minimum of 14 ms for Kaggle and 15 ms for APTOS datasets.
A study based on feature extraction using the RF model produced 74% accuracy in DR image classification [38]. Another two proposed hybrid models are based on combining the Gaussian mixture model and SVM to diagnose microaneurysms [18] and using KNN for the detection and classification of DR [39]. All the above-discussed studies used the existing classifiers to classify the DR lesions.
Some studies implemented CNN models to perform the binary classification of DR datasets [40,41]. Dropout regularization, augmentation, and pre-processing were performed manually by using the image editing tools in [42]. A deep CNN was proposed by [43] to classify normal and NPDR with two neural networks (i.e., the global and the local) and model performance was evaluated by the kappa score. The main disadvantage of this work is that it classifies only normal and NPPR, but it only works to detect the PDR.
To overcome those issues, the diagnostic results of the proposed model proved that it can achieve a satisfactory diagnostic performance, which can significantly assist the medical professional in the decision-making process in the early stages of detecting the infection, and timely treatment can decrease risk. Automatic screening and differentiation of diabetic retinopathy from fundus images will significantly reduce the effort of the medical professional and accelerate the diagnosis process.
Five class classifications are realized in the model, providing feasibility for the diagnosis of DR and its severity levels. The proposed model for the feature extraction and classification of DR performs better than the state-of-the-art models with high accuracy and less complexity. We will further optimize the model to model the accuracy of DR diagnosis and try to develop a more powerful DR detection model to assist doctors in clinical examinations.
The limitation of this model is that it is trained with only fundus image-level supervision, making it very challenging to accurately locate some minute lesion regions. Next, we need to specify the coarse location of the lesion along with the DR grading, which will help from the perspective of clinical application.

5. Conclusions

High blood pressure leads to DR, which causes retinal damage. Retinal vascularization is damaged by DR and can lead to blindness and potentially death. Fundoscopy examinations, which are time-consuming and expensive, allow ophthalmologists to see retinal vascular swelling. There is a need to automatically identify diabetic retinopathy by examining retinal fundus images. This study proposed an enhanced pooling function technique to minimize the loss to detect retina lesions, and an improved SVM classifier to classify the lesions using linear mapping. Five pre-trained deep learning models were recognized during the selection of the implementation, namely VGG-16, DenseNet, ResNet-50, Inception, and AlexNet. The proposed pooling and activation function results outperformed all the existing models. This study’s proposed model provided efficient accuracy results compared to the existing models.

Author Contributions

Methodology, U.B., N.C. and G.B.; Software, U.B.; Validation, G.B.; Investigation, U.B. and N.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

This study does not include any human subject or clinical trial information.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wild, S.H.; Roglic, G.; Green, A.; Sicree, R.; King, H. Global Prevalence of Diabetes: Estimates for the Year 2000 and Projections for 2030. Diabetes Care 2004, 27, 2569. [Google Scholar] [CrossRef] [Green Version]
  2. Scully, T. Diabetes in numbers. Nature 2012, 485, S2–S3. [Google Scholar] [CrossRef]
  3. Wu, L.; Fernandez-Loaiza, P.; Sauma, J.; Hernandez-Bogantes, E.; Masis, M. Classification of diabetic retinopathy and diabetic macula+r edema. World J. Diabetes 2013, 4, 290. [Google Scholar] [CrossRef]
  4. Khansari, M.M.; O’neill, W.D.; Penn, R.D.; Blair, N.P.; Shahidi, M. Detection of Subclinical Diabetic Retinopathy by Fine Structure Analysis of Retinal Images. J. Ophthalmol. 2019, 2019, 5171965. [Google Scholar] [CrossRef]
  5. Tufail, A.; Rudisill, C.; Egan, C.; Kapetanakis, V.V.; Salas-Vega, S.; Owen, C.G.; Rudnicka, A.R. Automated diabetic retinopathy image assessment software: Diagnostic accuracy and cost-effectiveness compared with human graders. Ophthalmology 2017, 124, 343–351. [Google Scholar] [CrossRef] [Green Version]
  6. Gulshan, V.; Rajan, R.; Widner, K.; Wu, D.; Wubbels, P.; Rhodes, T.; Whitehouse, K.; Coram, M.; Corrado, G.; Ramasamy, K.; et al. Performance of a Deep-Learning Algorithm vs Manual Grading for Detecting Diabetic Retinopathy in India. JAMA Ophthalmol. 2019, 137, 987–993. [Google Scholar] [CrossRef] [Green Version]
  7. García, G.; Gallardo, J.; Mauricio, A.; López, J.; Del Carpio, C. Detection of diabetic retinopathy based on a convolutional neural network using retinal fundus images. In Artificial Neural Networks and Machine Learning–ICANN 2017, Proceedings of the 26th International Conference on Artificial Neural Networks, Alghero, Italy, 11–14 September 2017; Springer International Publishing: Cham, Switzerland, 2017; pp. 635–642, Proceedings, Part II 26. [Google Scholar]
  8. Hemanth, D.J.; Deperlioglu, O.; Kose, U. An enhanced diabetic retinopathy detection and classification approach using deep convolutional neural network. Neural Comput. Appl. 2019, 32, 707–721. [Google Scholar] [CrossRef]
  9. Qummar, S.; Khan, F.G.; Shah, S.; Khan, A.; Shamshirband, S.; Rehman, Z.U.; Khan, I.A.; Jadoon, W. A Deep Learning Ensemble Approach for Diabetic Retinopathy Detection. IEEE Access 2019, 7, 150530–150539. [Google Scholar] [CrossRef]
  10. Costa, P.; Galdran, A.; Smailagic, A.; Campilho, A. A Weakly-Supervised Framework for Interpretable Diabetic Retinopathy Detection on Retinal Images. IEEE Access 2018, 6, 18747–18758. [Google Scholar] [CrossRef]
  11. Wan, S.; Liang, Y.; Zhang, Y. Deep convolutional neural networks for diabetic retinopathy detection by image classification. Comput. Electr. Eng. 2018, 72, 274–282. [Google Scholar] [CrossRef]
  12. Bhatkar, A.P.; Kharat, G.U. Detection of diabetic retinopathy in retinal images using MLP classifier. In Proceedings of the 2015 IEEE International Symposium on Nanoelectronic and Information Systems, Indore, India, 21–23 December 2015; IEEE: New York, NY, USA, 2015; pp. 331–335. [Google Scholar]
  13. Xu, J.; Zhang, X.; Chen, H.; Li, J.; Zhang, J.; Shao, L.; Wang, G. Automatic Analysis of Microaneurysms Turnover to Diagnose the Progression of Diabetic Retinopathy. IEEE Access 2018, 6, 9632–9642. [Google Scholar] [CrossRef]
  14. Antal, B.; Hajdu, A. An Ensemble-Based System for Microaneurysm Detection and Diabetic Retinopathy Grading. IEEE Trans. Biomed. Eng. 2012, 59, 1720–1726. [Google Scholar] [CrossRef] [Green Version]
  15. Dutta, S.; Manideep, B.C.; Basha, S.M.; Caytiles, R.D.; Iyengar, N.C.S.N. Classification of Diabetic Retinopathy Images by Using Deep Learning Models. Int. J. Grid Distrib. Comput. 2018, 11, 99–106. [Google Scholar] [CrossRef]
  16. Lunscher, N.; Chen, M.L.; Jiang, N.; Zelek, J. Automated Screening for Diabetic Retinopathy Using Compact Deep Networks. J. Comput. Vis. Imaging Syst. 2017, 3, 1–3. [Google Scholar] [CrossRef]
  17. Available online: https://www.kaggle.com/competitions/aptos2019-blindness-detection/data (accessed on 2 October 2022).
  18. Available online: https://www.kaggle.com/competitions/diabetic-retinopathy-detection/discussion/234309 (accessed on 2 October 2022).
  19. Nirthika, R.; Manivannan, S.; Ramanan, A.; Wang, R. Pooling in convolutional neural networks for medical image analysis: A survey and an empirical study. Neural Comput. Appl. 2022, 34, 5321–5347. [Google Scholar] [CrossRef]
  20. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  21. Kumar, M.; Bhandari, A.K. Contrast Enhancement Using Novel White Balancing Parameter Optimization for Perceptually Invisible Images. IEEE Trans. Image Process. 2020, 29, 7525–7536. [Google Scholar] [CrossRef]
  22. Niu, Y.; Wu, X.; Shi, G. Image Enhancement by Entropy Maximization and Quantization Resolution Upconversion. IEEE Trans. Image Process. 2016, 25, 4815–4828. [Google Scholar] [CrossRef]
  23. Veluchamy, M.; Bhandari, A.K.; Subramani, B. Optimized Bezier Curve Based Intensity Mapping Scheme for Low Light Image Enhancement. IEEE Trans. Emerg. Top. Comput. Intell. 2021, 6, 602–612. [Google Scholar] [CrossRef]
  24. Pizer, S.M. Contrast-limited adaptive histogram equalization: Speed and effectiveness stephen m. pizer, r. eugene johnston, james p. ericksen, bonnie c. yankaskas, keith e. muller medical image display research group. In Proceedings of the First Conference on Visualization in Biomedical Computing, Atlanta, GA, USA, 22–25 May 1990; Volume 337, p. 2. [Google Scholar]
  25. Singh, K.; Kapoor, R. Image enhancement using Exposure based Sub Image Histogram Equalization. Pattern Recognit. Lett. 2014, 36, 10–14. [Google Scholar] [CrossRef]
  26. Kansal, S.; Tripathi, R.K. New adaptive histogram equalisation heuristic approach for contrast enhancement. IET Image Process. 2020, 14, 1110–1119. [Google Scholar] [CrossRef]
  27. Yang, K.-F.; Zhang, X.-S.; Li, Y.-J. A Biological Vision Inspired Framework for Image Enhancement in Poor Visibility Conditions. IEEE Trans. Image Process. 2019, 29, 1493–1506. [Google Scholar] [CrossRef]
  28. Arici, T.; Dikbas, S.; Altunbasak, Y. A Histogram Modification Framework and Its Application for Image Contrast Enhance-ment. IEEE Trans. Image Process. 2009, 18, 1921–1935. [Google Scholar] [CrossRef]
  29. Mishra, M.; Menon, H.; Mukherjee, A. Characterization of S1 and S2 Heart Sounds Using Stacked Autoencoder and Convo-lutional Neural Network. IEEE Trans. Instrum. Meas. 2018, 68, 3211–3220. [Google Scholar] [CrossRef]
  30. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  31. Li, H.-Y.; Dong, L.; Zhou, W.-D.; Wu, H.-T.; Zhang, R.-H.; Li, Y.-T.; Yu, C.-Y.; Wei, W.-B. Development and validation of medical record-based logistic regression and machine learning models to diagnose diabetic retinopathy. Graefe’s Arch. Clin. Exp. Ophthalmol. 2022, 261, 681–689. [Google Scholar] [CrossRef]
  32. Tsao, H.Y.; Chan, P.Y.; Su, E.C.Y. Predicting diabetic retinopathy and identifying interpretable biomedical features using machine learning algorithms. BMC Bioinform. 2018, 19, 111–121. [Google Scholar] [CrossRef] [Green Version]
  33. Bhatia, K.; Arora, S.; Tomar, R. Diagnosis of diabetic retinopathy using machine learning classification algorithm. In Proceedings of the 2016 2nd International Conference on Next Generation Computing Technologies (NGCT), Dehradun, India, 14–16 October 2016; IEEE: New York, NY, USA, 2016; pp. 347–351. [Google Scholar]
  34. Chen, Y.; Hu, X.; Fan, W.; Shen, L.; Zhang, Z.; Liu, X.; Du, J.; Li, H.; Chen, Y.; Li, H. Fast density peak clustering for large scale data based on kNN. Knowl.-Based Syst. 2019, 187, 104824. [Google Scholar] [CrossRef]
  35. Cao, K.; Xu, J.; Zhao, W.Q. Artificial intelligence on diabetic retinopathy diagnosis: An automatic classification method based on grey level co-occurrence matrix and naive Bayesian model. Int. J. Ophthalmol. 2019, 12, 1158. [Google Scholar] [CrossRef]
  36. Alzami, F.; Megantara, R.A.; Fanani, A.Z. Diabetic retinopathy grade classification based on fractal analysis and random forest. In Proceedings of the 2019 International Seminar on Application for Technology of Information and Communication (iSemantic), Se-marang, Indonesia, 21–22 September 2019; IEEE: New York, NY, USA, 2019; pp. 272–276. [Google Scholar]
  37. Yu, S.; Tan, K.K.; Sng, B.L.; Li, S.; Sia, A.T.H. Lumbar Ultrasound Image Feature Extraction and Classification with Support Vector Machine. Ultrasound Med. Biol. 2015, 41, 2677–2689. [Google Scholar] [CrossRef]
  38. Seoud, L.; Chelbi, J.; Cheriet, F. Automatic grading of diabetic retinopathy on a public database. In Proceedings of the Ophthalmic Medical Image Analysis International Workshop, Munich, Germany, 8 October 2015; University of Iowa: Iowa City, IA, USA, 2015; Volume 2. No. 2015. [Google Scholar]
  39. Savarkar, S.P.; Kalkar, N.; Tade, S.L. Diabetic retinopathy using image processing detection, classification and analysis. Int. J. Adv. Comput. Res. 2013, 3, 285. [Google Scholar]
  40. Gondal, W.M.; Kohler, J.M.; Grzeszick, R.; Fink, G.A.; Hirsch, M. Weakly-supervised localization of diabetic retinopathy lesions in retinal fundus images. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 2069–2073. [Google Scholar] [CrossRef] [Green Version]
  41. Wang, Z.; Yin, Y.; Shi, J.; Fang, W.; Li, H.; Wang, X. Zoom-in-net: Deep mining lesions for diabetic retinopathy detection. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2017, Proceedings of the 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; Springer International Publishing: Cham, Switzerland, 2017; pp. 267–275, Proceedings, Part III 20. [Google Scholar]
  42. Chandrakumar, T.; Kathirvel, R.J.I.J.E.R.T. Classifying diabetic retinopathy using deep learning architecture. Int. J. Eng. Res. Technol. 2016, 5, 19–24. [Google Scholar]
  43. Yang, Y.; Li, T.; Li, W.; Wu, H.; Fan, W.; Zhang, W. Lesion detection and grading of diabetic retinopathy via two-stages deep convolutional neural networks. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2017, Proceedings of the 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; Springer International Publishing: Cham, Swit-zerland, 2017; pp. 533–540, Proceedings, Part III 20. [Google Scholar]
Figure 1. Experimental framework.
Figure 1. Experimental framework.
Diagnostics 13 02606 g001
Figure 2. Multiclass of DR (a) NODR, (b) Mild DR, (c) Moderate DR, (d) Severe DR, and (e) PDR. (First row—APTOS dataset, second row—Kaggle dataset).
Figure 2. Multiclass of DR (a) NODR, (b) Mild DR, (c) Moderate DR, (d) Severe DR, and (e) PDR. (First row—APTOS dataset, second row—Kaggle dataset).
Diagnostics 13 02606 g002
Figure 3. Augmentation (a) Original image, (b) rotation, (c) horizontal flip, (d) brightness, (e) contrast.
Figure 3. Augmentation (a) Original image, (b) rotation, (c) horizontal flip, (d) brightness, (e) contrast.
Diagnostics 13 02606 g003
Figure 4. Improved ResNet-50 model.
Figure 4. Improved ResNet-50 model.
Diagnostics 13 02606 g004
Figure 5. Comparison of the image enhancement of the proposed model with the existing models.
Figure 5. Comparison of the image enhancement of the proposed model with the existing models.
Diagnostics 13 02606 g005
Figure 6. Segmentation results. (a) original image, (b) ground truth, (c) proposed model, (d) DenseNet, (e) Inception, (f) VGG-19, (g) AlexNet.
Figure 6. Segmentation results. (a) original image, (b) ground truth, (c) proposed model, (d) DenseNet, (e) Inception, (f) VGG-19, (g) AlexNet.
Diagnostics 13 02606 g006
Figure 7. Confusion matrix for APTOS augmented dataset on different CNN models.
Figure 7. Confusion matrix for APTOS augmented dataset on different CNN models.
Diagnostics 13 02606 g007
Figure 8. Confusion matrix for Kaggle augmented dataset on different CNN models.
Figure 8. Confusion matrix for Kaggle augmented dataset on different CNN models.
Diagnostics 13 02606 g008aDiagnostics 13 02606 g008b
Figure 9. DR classification comparison of various classifiers of different datasets. It displays the performance results of the two datasets. Pink color represents the messidor and the green represents the APTOS.
Figure 9. DR classification comparison of various classifiers of different datasets. It displays the performance results of the two datasets. Pink color represents the messidor and the green represents the APTOS.
Diagnostics 13 02606 g009
Table 1. Dataset distribution.
Table 1. Dataset distribution.
DatasetNODRMild DRModerate DRSevere DRPDRCount
APTOS18053709991932953662
Kaggle25,8102443529287370835,126
Table 2. Dataset augmentation operations.
Table 2. Dataset augmentation operations.
ClassAPTOSKaggle
OriginalOperationsAugmentedOriginalOperationsAugmented
NoDR18050180525,810025,810
MildDR3705185024431024,430
Moderate DR999219985292526,460
Severe DR193917378732925,317
PDR295617707083625,488
Total3662 916035,126 127,505
Table 3. Augmented dataset image distribution.
Table 3. Augmented dataset image distribution.
ClassAPTOSKaggle
TrainingTestingTrainingTesting
NoDR144436120,6485162
MildDR148037019,5444886
Moderate DR159840021,1665292
Severe DR139034720,2545063
PDR141635420,3905098
Total73281832102,00425,501
Table 4. Average scores for the augmented APTOS dataset.
Table 4. Average scores for the augmented APTOS dataset.
ModelPSNRGMSDEntropySSIMPCQIProcessing Time (s)
Clahe [24]30.830.1637.2630.6341.1390.155
ESIHE [25]31.930.0747.3160.6351.2820.153
HAHE [26]32.820.1257.2260.6931.0010.373
BIMEF [27]31.680.1997.2690.7361.0070.364
HMF [28]32.630.0857.2830.6361.1030.218
ABC34.830.0487.8340.8771.3780.173
Proposed35.560.0377.9350.9831.4840.151
Table 5. Comparison of segmentation results for the APTOS dataset with the state-of-the-art models.
Table 5. Comparison of segmentation results for the APTOS dataset with the state-of-the-art models.
ModelPool + ActAccuracyPrecisionRecall
DenseNet [29]Max + Relu0.94840.83640.9584
Inception [12]Max + Relu0.98470.85780.9848
VGG-19 [30]Max + Relu0.97950.84790.9483
AlexNet [31]Max + Relu0.98580.93780.9847
ResNet-50Proposed0.99861.00001.0000
AlexNetProposed0.99861.00000.9864
DenseNetProposed0.99591.00000.9916
InceptionProposed0.99720.98640.9864
VGG-19Proposed0.99860.98661.0000
Table 6. Performance metrics for APTOS augmented dataset.
Table 6. Performance metrics for APTOS augmented dataset.
CNN ModelClassifierAccuracyPrecisionRecallF1-ScoreClass
DenseNetISVM0.997816590.994459830.994459830.99445983Normal
0.996724890.989247310.994594590.99191375Mild
0.996179040.990024940.992500000.99126092Moderate
0.997270740.991379310.994236310.99280576Severe
0.997816591.00000000.988700560.99431818PDR
SVM0.996179040.988950280.991689750.99031812Normal
0.996179040.989218330.991891890.99055331Mild
0.995087340.987531170.990000000.98876404Moderate
0.996179040.988505750.991354470.98992806Severe
0.995633190.994285710.983050850.98863636PDR
RF0.995633190.988919670.988919670.98891967Normal
0.992903930.981132080.983783780.98245614Mild
0.993449780.982587060.987500000.98503741Moderate
0.995087340.985632180.988472620.98705036Severe
0.993449780.988571430.977401130.98295455PDR
NB0.994541480.983471070.988919670.98618785Normal
0.990720520.973190350.981081080.97711978Mild
0.993995630.985037410.987500000.98626717Moderate
0.992903930.982658960.979827090.98124098Severe
0.992903930.988538680.974576270.98150782PDR
ResNet-50ISVM0.997816590.991735540.997229920.99447514Normal
0.998362450.994609160.99729730.99595142Mild
0.998362450.997493730.99500000.99624531Moderate
0.999454151.000000000.997118160.99855700Severe
0.999454151.000000000.997175140.99858557PDR
SVM0.997270740.991712710.994459830.99308437Normal
0.997270740.991913750.994594590.99325236Mild
0.995633190.987562190.992500000.99002494Moderate
0.997270740.994219650.991354470.99278499Severe
0.998362451.000000000.991525420.99574468PDR
RF0.996179040.988950280.991689750.99031812Normal
0.995633190.986559140.991891890.98921833Mild
0.995633190.990000000.990000000.99000000Moderate
0.996179040.991329480.988472620.98989899Severe
0.996724890.994318180.988700560.99150142PDR
NB0.995087340.983516480.991689750.98758621Normal
0.993995630.981233240.989189190.98519515Mild
0.995087340.989974940.987500000.98873592Moderate
0.993449780.985507250.979827090.98265896Severe
0.995087340.991452990.983050850.98723404PDR
AlexNetISVM0.996179040.988950280.991689750.99031812Normal
0.994541480.986486490.986486490.98648649Mild
0.992358080.982500000.982500000.98250000Moderate
0.996724890.991354470.991354470.99135447Severe
0.997270740.994334280.991525420.99292786PDR
SVM0.994541480.983471070.988919670.98618785Normal
0.994541480.989130430.983783780.98644986Mild
0.992903930.987405540.980000000.98368883Moderate
0.994541480.982808020.988472620.98563218Severe
0.995087340.985915490.988700560.98730606PDR
RF0.993449780.980716250.986149580.98342541Normal
0.991266380.975806450.981081080.97843666Mild
0.991266380.984848480.975000000.97989950Moderate
0.992358080.979827090.979827090.97982709Severe
0.994541480.985875710.985875710.98587571PDR
NB0.992903930.978021980.986149580.98206897Normal
0.989628820.970509380.978378380.97442799Mild
0.990174670.982323230.972500000.97738693Moderate
0.992358080.98260870.976945240.97976879Severe
0.992358080.980225990.980225990.98022599PDR
InceptionISVM0.993995630.978142080.991689750.98486933Normal
0.991266380.973262030.983783780.97849462Mild
0.992903930.992366410.975000000.98360656Moderate
0.993995630.982758620.985590780.98417266Severe
0.995087340.991452990.983050850.98723404PDR
SVM0.993449780.978082190.988919670.98347107Normal
0.991812230.981029810.978378380.97970230Mild
0.989628820.979848870.972500000.97616060Moderate
0.991812230.977011490.979827090.97841727Severe
0.992903930.983002830.980225990.98161245PDR
RF0.991812230.975274730.983379500.97931034Normal
0.989082970.972972970.972972970.97297297Mild
0.987445410.972431080.97000000.97121402Moderate
0.991266380.979710140.97406340.97687861Severe
0.991266380.977401130.977401130.97740113PDR
NB0.990720520.969945360.983379500.97661623Normal
0.989628820.978201630.970270270.97421981Mild
0.986353710.972292190.965000000.96863237Moderate
0.987445410.962857140.971181560.96700143Severe
0.991266380.980113640.974576270.97733711PDR
VGG-19ISVM0.992903930.975409840.988919670.98211829Normal
0.989082970.967914440.978378380.97311828Mild
0.990174670.984771570.970000000.97732997Moderate
0.994541480.988405800.982708930.98554913Severe
0.992903930.983002830.980225990.98161245PDR
SVM0.991812230.972677600.986149580.97936726Normal
0.990720520.980926430.972972970.97693351Mild
0.987445410.977215190.965000000.97106918Moderate
0.991266380.974212030.979827090.97701149Severe
0.989628820.971830990.974576270.97320169PDR
RF0.991266380.972602740.983379500.97796143Normal
0.988537120.972899730.970270270.97158322Mild
0.986899560.974747470.965000000.96984925Moderate
0.989628820.971264370.974063400.97266187Severe
0.987991270.968926550.968926550.96892655PDR
NB0.988537120.961956520.980609420.97119342Normal
0.986353710.964959570.967567570.96626181Mild
0.985807860.972222220.962500000.96733668Moderate
0.987991270.968299710.968299710.96829971Severe
0.986899560.971428570.960451980.96590909PDR
Table 7. Performance metrics for Kaggle augmented dataset.
Table 7. Performance metrics for Kaggle augmented dataset.
CNN ModelClassifierAccuracyPrecisionRecallF1-ScoreClass
DenseNetISVM0.999764720.999225410.999612550.99941894Normal
0.999803930.999590580.999386000.99948828Mild
0.999686290.999055530.999433110.99924428Moderate
0.999607860.999012440.999012440.99901244Severe
0.999568640.999214920.998626910.99892083PDR
SVM0.999647070.999031570.999225110.99912833Normal
0.999607860.998976670.998976670.99897667Mild
0.999529430.998866210.998866210.99886621Moderate
0.999568640.998815170.999012440.99891379Severe
0.999529430.999018840.998626910.99882284PDR
RF0.999568640.999031200.998837660.99893442Normal
0.999490220.998363670.998976670.99867008Mild
0.999490220.998866000.998677250.99877161Moderate
0.999411790.998814230.99822240.99851823Severe
0.999294150.998039220.998430760.99823495PDR
NB0.999294150.998063520.998450210.99825683Normal
0.999137290.997748670.997748670.99774867Mild
0.999215720.997922180.998299320.99811071Moderate
0.999215720.998024890.998024890.99802489Severe
0.999137290.998233220.997449980.99784144PDR
ResNet-50ISVM0.999921570.999806280.999806280.99980628Normal
0.999843140.999590670.999590670.99959067Mild
0.999921570.999811040.999811040.99981104Moderate
0.999843140.999604980.999604980.99960498Severe
0.999843140.999607690.999607690.99960769PDR
SVM0.999725500.999031950.999612550.99932217Normal
0.999725500.999385880.999181330.99928359Mild
0.999764720.999244430.999622070.99943321Moderate
0.999725500.999407350.999209950.99930864Severe
0.999725500.999607460.999019220.99931325PDR
RF0.999607860.998644990.999418830.99903176Normal
0.999568640.998772250.998976670.99887445Mild
0.999686290.999244140.999244140.99924414Moderate
0.999647070.999209800.999012440.99911111Severe
0.999607860.999411070.998626910.99901884PDR
NB0.999529430.998644460.999031380.99883788Normal
0.999254930.997750050.998362670.99805627Mild
0.999411790.998488570.998677250.99858290Moderate
0.999451000.998814460.998419910.99861715Severe
0.999372570.998822140.998038450.99843014PDR
AlexNetISVM0.999882360.999806240.999612550.99970939Normal
0.999843140.999795250.999386000.99959058Mild
0.999764720.999244430.999622070.99943321Moderate
0.999803930.999604900.999407470.99950617Severe
0.999686290.999019610.999411530.99921553PDR
SVM0.999490220.998451110.999031380.99874116Normal
0.999529430.998772000.998772000.99877200Mild
0.999333360.998299640.998488280.99839395Moderate
0.999490220.999011860.998419910.99871580Severe
0.999411790.998626640.998430760.99852869PDR
RF0.999647070.999031570.999225110.99912833Normal
0.999490220.998567630.998772000.99866980Mild
0.999529430.998677750.999055180.99886643Moderate
0.999490220.998814700.998617420.99871605Severe
0.999490220.999018650.998430760.99872461PDR
NB0.999098070.997483060.998062770.99777283Normal
0.999176500.998360990.997339340.99784990Mild
0.999058860.997544390.997921390.99773285Moderate
0.999098070.997826950.997629860.99772840Severe
0.998941220.997254360.997449980.99735216PDR
InceptionISVM0.999803930.999612480.999418830.99951564Normal
0.999725500.999385880.999181330.99928359Mild
0.999843140.999622070.999622070.99962207Moderate
0.999607860.998618510.999407470.99901283Severe
0.999764720.999607540.999215380.99941142PDR
SVM0.999411790.998257500.998837660.99854750Normal
0.999372570.997955010.998772000.99836334Mild
0.999372570.998676750.998299320.99848800Moderate
0.999333360.998616600.998024890.99832066Severe
0.999372570.998626370.998234600.99843045PDR
RF0.999451000.998643940.998643940.99864394Normal
0.999372570.997955010.998772000.99836334Mild
0.999254930.998487430.997921390.99820433Moderate
0.999294150.998222400.99822240.99822240Severe
0.999333360.998430450.99823460.99833252PDR
NB0.998980430.997288920.997675320.99748208Normal
0.998902000.997338250.996930000.99713408Mild
0.998902000.997166600.997543460.99735500Moderate
0.998980430.997432350.997432350.99743235Severe
0.999058860.997841440.997449980.99764567PDR
VGG-19ISVM0.999803930.999806160.999225110.99951555Normal
0.999647070.998976880.999181330.99907910Mild
0.999843140.999810960.999433110.99962200Moderate
0.999451000.998027220.999209950.99861824Severe
0.999686290.999411300.999019220.99921522PDR
SVM0.999372570.998257160.998643940.99845051Normal
0.999333360.997954590.998567340.99826087Mild
0.999333360.998676500.998110360.99839335Moderate
0.999215720.998221700.997827380.99802450Severe
0.999215720.998038450.998038450.99803845PDR
RF0.999254930.997870280.998450210.99816016Normal
0.999294150.997954170.998362670.99815838Mild
0.999176500.998298360.997732430.99801531Moderate
0.999098070.997826950.997629860.99772840Severe
0.999254930.998234260.998038450.99813634PDR
NB0.998902000.997095270.997481600.99728840Normal
0.998784360.996929380.996725340.99682735Mild
0.998862790.997166070.997354500.99726027Moderate
0.998862790.997037910.997234840.99713637Severe
0.999098070.998037290.997449980.99774355PDR
Table 8. Varying training and test size.
Table 8. Varying training and test size.
DatasetTrainingTestingAccuracyMeanStandard
Deviation
APTOS70300.9812250.9825430.0011409
75250.983202
80200.983202
Kaggle70300.9713440.9802370.0080882
75250.982213
80200.987154
Table 9. DR classification comparison of the processing time for the proposed model with different optimizations.
Table 9. DR classification comparison of the processing time for the proposed model with different optimizations.
ClassifierKaggle (s)APTOS (s)
Logistic regression [32]2129
DT [33]1521
KNN [34]2330
NB [35]2025
RF [36]2023
SVM [37]2231
Improved SVM1415
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bhimavarapu, U.; Chintalapudi, N.; Battineni, G. Automatic Detection and Classification of Diabetic Retinopathy Using the Improved Pooling Function in the Convolution Neural Network. Diagnostics 2023, 13, 2606. https://doi.org/10.3390/diagnostics13152606

AMA Style

Bhimavarapu U, Chintalapudi N, Battineni G. Automatic Detection and Classification of Diabetic Retinopathy Using the Improved Pooling Function in the Convolution Neural Network. Diagnostics. 2023; 13(15):2606. https://doi.org/10.3390/diagnostics13152606

Chicago/Turabian Style

Bhimavarapu, Usharani, Nalini Chintalapudi, and Gopi Battineni. 2023. "Automatic Detection and Classification of Diabetic Retinopathy Using the Improved Pooling Function in the Convolution Neural Network" Diagnostics 13, no. 15: 2606. https://doi.org/10.3390/diagnostics13152606

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop