Next Article in Journal
Mathematical Fuzzy Logic in the Emerging Fields of Engineering, Finance, and Computer Sciences
Previous Article in Journal
Joint Approximation of Analytic Functions by Shifts of the Riemann Zeta-Function Twisted by the Gram Function II
 
 
Retraction published on 25 March 2024, see Axioms 2024, 13(4), 215.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RETRACTED: Diabetic Retinopathy Progression Prediction Using a Deep Learning Model

by
Hanan A. Hosni Mahmoud
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
Axioms 2022, 11(11), 614; https://doi.org/10.3390/axioms11110614
Submission received: 9 October 2022 / Revised: 27 October 2022 / Accepted: 1 November 2022 / Published: 4 November 2022 / Retracted: 25 March 2024
(This article belongs to the Special Issue Bio-Informatics and Data Set Analysis)

Abstract

:
Diabetes is an illness that happens with a high level of glucose in the body, and can harm the retina, causing permanent loss vision or diabetic retinopathy. The fundus oculi method comprises detecting the eyes to perform a pathology test. In this research, we implement a method to predict the progress of diabetic retinopathy. There is a research gap that exists for the detection of diabetic retinopathy progression employing deep learning models. Therefore, in this research, we introduce a recurrent CNN (R-CNN) model to detect upcoming visual field inspections to predict diabetic retinopathy progression. A benchmark dataset of 7000 eyes from healthy and diabetic retinopathy progress cases over the years are utilized in this research. Approximately 80% of ocular cases from the dataset is utilized for the training stage, 10% of cases are used for validation, and 10% are used for testing. Six successive visual field tests are used as input and the seventh test is compared with the output of the R-CNN. The precision of the R-CNN is compared with the regression model and the Hidden Markov (HMM) method. The average prediction precision of the R-CNN is considerably greater than both regression and HMM. In the pointwise classification, R-CNN depicts the least classification mean square error among the compared models in most of the tests. Also, R-CNN is found to be the minimum model affected by the deterioration of reliability and diabetic retinopathy severity. Correctly predicting a progressive visual field test with the R-CNN model can aid physicians in making decisions concerning diabetic retinopathy.

1. Introduction

One of the main causes of blindness in the world, diabetic retinopathy is defined by the permanent loss of retinal ganglion cells (RGCs) [1,2]. The vision field gradually deteriorates due to structural abnormalities in the RGCs and the optic nerve head [2]. The evaluation and forecasting of the progressive visual field are thought to be crucial processes for maintaining visual function. However, because to the various random mistakes and fluctuations they contain, visual field tests are prone to unpredictability. Clinical knowledge of the evolution of the visual field is hindered by this variability, which is more noticed in ocular cases with diabetic retinopathy than in normal ocular cases [3].
The past several years have seen a lot of interest in and success with studies on machine learning algorithms used to gauge the course of diabetic retinopathy. Visual field defects are divided into 16 archetypes and their evolution is described by the authors in [4]. They used 12,217 eyes from 7360 cases with several reliable 24-2 visual field recording over 5 years. They achieved 90% accuracy in the ground truth validation of 412 eyes with 29.3% confirmed progression. Variation Bayes Linear Regression (VBLR), a sort of machine learning technique, is applied by the authors in [5], and they showed better prediction than pointwise Linear Regression. Deep learning algorithms have recently been used with great success on a variety of tasks as artificial intelligence has advanced. However, only a few studies have attempted to use deep learning algorithms to detect visual field progression. A single visual field test is used as the input by the authors in [6] to train a convolutional neural network (CNN) to predict progressive visual fields. Measurements from successive 28-2 visual fields from the years 1998 to 2018 were selected from a public medical database. K-fold cross validation testing with a held out test set was utilized. The model was trained on temporal visual field tests to produce the following predictions for using only a single visual field with accuracy of 90.5%. The authors of [7] presented an auto-encoder learned from low-dimensional standard visual fields utilizing 29,161 fields from 3900 cases. The auto-encoder was trained on a 90% of the dataset of random cases. The remaining 10% was used for the progression of testing. The accuracy is about 90% but with higher CPU time.
Applications involving sequential time series and temporal dependencies have been carried out using recurrent neural networks (RNNs), which are artificial convolutional networks with recurrent connections [8,9,10]. These have been used successfully for many years in various tasks involving sequence modeling. An RNN can analyze recent data using historical data. Using the dependencies between sequence elements, RNNs are effective predictors [11,12,13]. The Hidden Markov Model (HMM) and gated recurrent CNN (R-CNN), the two primary RNN variations, attempt to model long-term reliance in lengthy sequences. In contrast to conventional least squares linear regression, HMM has shown a greater ability to predict progressive visual fields according to our prior research. According to the authors in [14], HMM networks could over time detect both local and global patterns in visual fields.
However, the R-CNN performs as well as the HMM and deploys gating units more effectively than a standard HMM [14,15,16,17,18]. Studies in a variety of fields have shown that the R-CNN performs exceptionally well in sequential data analysis when compared to other RNN types [19,20,21,22]. RNN has recently been upgraded to a bidirectional technique by simultaneously training in positive and negative time directions, which improves context understanding [23]. Bidirectional, gated recurrent CNN (R-CNN) can more accurately forecast visual field advancement, since visual field tests are sequential data with many connections between them. Table 1. depicts recent research in diabetic retinopathy prediction intelligent models.
In this paper, we propose a deep learning R-CNN to classify progressive visual field impairment. In this research, we introduce an RNN model and perform performance evaluation and compare the results with regression and HMM models.
The key contribution of this research is to realize higher accuracy by extracting temporal features related to the progress of diabetic retinopathy over 4.3 years on average. The proposed research introduces novel visual field features. Clinical temporal markers of ocular cases are captured from seven successive visual field tests. Temporal features are represented by the input channel to the deep CNN. The dataset is fed to three models, namely: regression, HMM, and R-CNN. A comparison of the three models is performed. The contributions of the proposed research are as follows:
  • The temporal representations of the ocular cases are taken in seven successive visual field tests over 4.3 years to test the progression of the disease and predict the progression using deep learning.
  • The overall accuracy is improved compared to the related work.
The paper is organized as follows: Section 2 describes the materials and methods. Section 3 depicts experimental settings and results. The paper is concluded in Section 4.

2. Dataset

This retroactive research is performed on public datasets of diabetic retinopathy images. The ocular image dataset utilized in this research is gathered from a public diabetic retinopathy database. The progress of diabetic retinopathy is depicted in Figure 1.
Ocular cases are taken from seven successive visual field tests that are utilized in the training and validation datasets. There is no overlap between the training and validation datasets. Eyes with intermission of four years between the first and seventh visual field tests are included. For instance, if there are thirteen successive visual field tests, the first to sixth tests are used as the initial data and the seventh to the test number 12 are used as the subsequent data. The test number 13 is omitted from the database. The tests numbered 6 and 12 are employed in the prediction phase, and the left behind tests are employed in the training, as depicted in Figure 2.

2.1. Random Errors and Fluctuations to Visual Field Tests

Visual field tests of the public dataset are done employing the 28–2 modeling with the Swedish Threshold model on a Field Analyzer III (Karl Xeiss Meditec, Inc., San Francisco, CA, USA). Visual fields are tested for fluctuations and errors comprising eyelid artifacts, lack of attention or tiredness effects. We also omitted artifacts such as faulty fixation or evidence of glaucoma, which can affect the results; any tests with such artifacts were excluded from this study. Visual fields can be unreliable and the loss function can exceeds 35% of errors in false negative rate.
We obtained 8323 visual field data consisting of 6-cells from 6685 eyes of 4593 ocular cases. Datasets from 7051 (85%) and 1272 (15%) are used as the training and test datasets, respectively. A total of 7051 records from the training dataset are randomly split into training and validation datasets at a ratio of 9:1. The validation data are used to prevent overfitting through checking the current fitness of the neural network during training. All 8323 datasets had exactly six visual field tests, and the average follow-up period for each of the six visual field tests is 4.39   ±   1.69 years. Table 2 shows the information for each dataset.

2.2. Training and Testing Dataset

Diabetic retinopathy cases are classified into five classes: normal, mild diabetic retinopathy, moderate diabetic retinopathy, severe diabetic retinopathy, and proliferative diabetic retinopathy as depicted in Table 3. Mild diabetic retinopathy begins with minute alterations in blood vessels, and recovery can still be achieved at this stage. If the patient has not received treatment, the case will progress to moderate diabetic retinopathy accompanied with blood vessel leakage. The diabetic retinopathy will then progress to severe and proliferative cases and may lead to vision impairment [34,35,36].
To classify diabetic retinopathy with better precision using a deep learning model, a large size dataset is required for training. Table 4 depicts more information on the count of images in each diabetic retinopathy class in both the training and testing subsets.

2.3. Visual Field Test

Automated primary tests are done using a Humphrey Visual device 750i (Carl Zeiss Meditec, Boston, MA, USA) with the threshold algorithm (ITA) 28-2 or 34-2. Among the 62 points of the 28-2 test arrangement, two points of biological scotoma are left out, and the remaining points are utilized. The 34-2 test arrangement is transformed to 28-2 utilizing test points. Robust visual field tests are depicted as having a false-positive ratio of less than 35%, a false-negative ratio of less than 35%, and a fixation loss value of less than 35%.

2.4. Convolutional Neural Network

We used the convolutional neural networks HMM and R-CNN. Python language 3.8 (Google Inc., New York, NY, USA) was used to classify the visual field test.
A.
HMM and R-CNN
We constructed a one-layer CNN to train on the structure of the utilized dataset utilizing preprocessed training data. The HMM neural networks are defined in Equations (1)–(5) as follows:
F g = ¥ ( S f I t + S L f L t 1 + P f )
I = ¥ ( S i I t + S L i L t 1 + P i )
O = ¥ ( S o I t + S L o L t 1 + P o )
( A ) t = ( A ) t 1 ( F g ) t + ( I   ) t ( tanh ( S C I t + S L C L t 1 ) + P C )
L t = O   tanh ( ( A ) t 1 )
where, I is the input, O is the output, ¥ is the activation function   S f , S i ,   S o ,   and   S C and P f ,   P i ,   P o ,     and   P C   represent the score and preference bias using various steps in the CNN, respectively, of three gates and a RAM cell and ⨂ is the dot product of two vectors.
The definitions of the input gate I and output gate O are utilized to regulate the memory flow of inputs and outputs into the remaining part of the CNN, while F g   is an auxiliary to the cells which authorizes the output of more weight values from the prior neuron to the subsequent neuron. The data located in the memory cells is influenced by the greater activation values; if the input has greater activation function value, the data will be stored in the memory. In addition, if the output has greater activation function value, it permits the data to go to the following neuron. If not, the input data with the greater weight values will be stored in the memory. The sigmoid and tanh are activation functions. L (t − 1) denotes the previous hidden convolutional layer that computes the weights. After computing Equation (4), (A) t turns into the up-to-date memory cell. Equation (5) displays the dot multiplication of the previous hidden layer value and prior cell. The non-linear property of the gates using tanh and sigmoid are produced, which are depicted in Equations (1)–(5) The quantity 𝑡 − 1 and t are prior and present time periods.
R-CNN is an alternative of HMM which contains two gates—the update and the reset gates. The R-CNN has no extra memory cell to store the data; hence, it can control the data included in the unit, as depicted in Equations (6)–(9).
G u p d a t e = ¥ ( S u I t + S L u L t 1 + P u )
G r e s e t = ¥ ( S r I t + S L r L t 1 + P r )
L t ˜ = t a n h ( S I t + S ( G r e s e t L t 1 ) )
L t = ( 1 ( G u p d a t e ) t ) L t 1 + ( G u p d a t e ) t L t
The update gate G u p d a t e in Equation (6) defines the amount of information that has been changed. In Equation (7), the reset gate G r e s e t   is comparable to G u p d a t e ; if G r e s e t   is reset to zero, it captures the input and overlooks the prior computed state. Furthermore, L t ˜ denotes the equivalent functionality as in R-CNN, and L t computes the linear function between the present L t ˜   and the prior L   t 1 activation function value in Equations (8) and (9).
A reverse and forward R-CNN is joined to formulate an R-CNN convolutional layer. These layers utilize the same input but will learn differently and join the results to produce the output. Deep neural networks can be effective at mapping activation functions and representing variable dependencies [20,21,22]. We can establish that R-CNN has higher performance on the datasets.
B.
Proposed Model and Experimental Results
In the proposed model, the deep learning includes three phases: feeding the data, the sequential neural network layer that is utilized for predictions, and dense layers. The convolutional neural structures for both HMM and R-CNN are depicted in Figure 3.
Figure 3a shows the structure of the HMM method. The HMM model has been introduced in [13].
Figure 3b depicts the structure of the proposed R-CNN model. The input layers comprise five successive data values from I   t 1   to I   t + 3 and the final classification output I   t + 4 , with an output of 60 classes on the final exam.
The sequence diagram of the proposed model is depicted in Figure 4.
The sequential convolutional neural model comprised eight parallel HMM or R-CNN units. The architecture of the HMM and R-CNN models are presented in Figure 3a,b, respectively.
The initial seven units receives 120 features, including 72 deviation values (DV) and 48 pattern values (PTV). Metrics including false negative, false positive and fixation loss ratios are presented as well as the time metric. To enhance the precision of the deep learning prediction, all of the inputs are divided by the average value and normalized. The DV, PTV, and time metric values are partitioned into 60, 60, and 1100 samples, respectively. The final cell utilizes the normalized time metric value. Later, the convolutional neural layer is linked to the following dense layer with 62 neurons which produce the final output and 62 visual predictions where each neuron produces one visual field value.

3. Experiments

3.1. Experimental Results Analyses

The mean square error (MSE) and mean absolute error (MAE) of the DV are utilized as precision measures. The MSE is computed for left and right eyes as depicted in Equations (10) and (11).
M S E = n = 1 52 ( a c t u a l   D V n c l a s s i f i e d   D V n ) 2 62
where, n is the number of the test points of the visual field.
MAE is computed for each test point of the visual field for all eyes using the following Equation:
M A E i = i = 1 n | T D V i P T D V i | n
where, n is the number of eyes, T D V i is the actual case and P T D V i is the predicted case.
A variance analysis is done to evaluate the three models. At the time of rejection of the null hypothesis, the average hypothesis is significant and is adopted, and ad hoc computation is done utilizing a t-test. In all statistical studies, p < 0.05 indicates a significant result.

3.2. Experimental Results

Table 2 depicts the demographic statistics of the testing portion of the dataset. The utmost diagnosis is principal open angle diabetic retinopathy (57.28%). The mean classification time (the time slot between classification and the final visual field test) is 1.02 ± 0.74 years, as depicted previously. The mean MSE and points mean absolute error (PMAE) are depicted in Table 5 and Table 6.
Confusion matrices are presented in Table 7, Table 8 and Table 9 for the regression model, HMM, and the proposed deep learning model for the five classes of diabetic retinopathy.
The classification precision of the R-CNN model is higher than the regression and HMM precisions. The mean square error of the R-CNN is 2.91 ± 1.32 dB and the MSEs of the regression and HMM models are 3.71 ± 3.59 dB and 3.06 ± 3.61 dB, respectively. The differences in the classification errors are significant (F = 45.14, p < 0.0015). The mean square error of the R-CNN is considerably less than the other models (both p < 0.0015).
The count of cases discarded by the mean square error classification error is depicted in Figure 5. The varieties where the classification error of R-CNN comprised 50% or more of the total count of cases are ≤3 dB (630 cases, 42.11%) and 3–4 dB (275 cases, 14.16%). The equivalent ranges of the classification mean error by the regression method are ≤3 dB (429 cases, 31.86%) and 3–4 dB (291 cases, 21.97%), and the results of the HMM are ≤3 dB (605 cases, 39.70%) and 3–4 dB (165 cases, 13.97%).
The visual field mean absolute error is depicted in Figure 6. Of the 62 DV points, R-CNN displayed the least classification error in the different methods. R-CNN displayed considerably higher precision at point 30 (red circles) and point 50 (blue circles) as opposed to the regression and HMM, respectively.
Table 10 depicts the mean square classification error (MSE) for the various field tests, as displayed in Figure 7. The 30–2 field is partitioned into six partitions as presented in [21,22,23]. The eye optic nerve anatomy (visual and temporal) are utilized. In all partitions, the classification errors of the R-CNN are considerably less than regression and HMM (p ≤ 0.0015).
The mean MSE error discarded by the different parameters are depicted in Table 11 and Figure 8. The classification error of R-CNN is considerably less in cases of false positive, false negative, and loss rate as opposed to the other models (p ≤ 0.028); as MD increases, the classification errors of all methods decrease.
The correlation of the classification error and other parameters are depicted in Table 12 and Figure 8. The mean square error increases as the false negative rate increases. In addition, MSE increases as the loss function value increases. MSE decreases as visual field D increases (p < 0.025) in all models, as depicted in Figure 8.

4. Conclusions

In this research, we proposed an intelligent model to predict diabetic retinopathy progression over the years as an auxiliary diagnostic tool. This tool can aid in predicting retinopathy progression at an early stage and help clinical professionals take fast and efficient treatment steps.
In this research, R-CNN displayed a higher classification accuracy than regression and HMM models in all areas of the visual field. Also, the R-CNN network has higher accuracy in the central regions than the other methods. These results are clinically significant due to the protection of the central visual area. This protection is important for the quality of life value of the ocular patients with diabetic retinopathy [27,28,29]. In the present study, R-CNN is the least affected by the worsening of reliability indices. The false negative and fixation losses affected visual field classification among the reliability indices in all models. However, the correlation coefficient of fixation losses is weak, and R2 is also small, indicating that the effect of fixation losses is small. Previous studies reported that false negative rates are associated with visual field assessment, but fixation losses are not. Other studies reported that fixation losses are the most common cause of unreliable visual field classification [32,33,34,35,36].
The limitations of our study include the following: First, there is a lack of generality according to the degree of diabetic retinopathy severity; ocular cases with early diabetic retinopathy with D > −6 dB are included relatively more in the training and test datasets than ocular cases with advanced diabetic retinopathy. This may have affected R-CNN model learning, but it can be more helpful as it reflects the ratio of actual ocular cases in real clinical practice. Second, we did not include clinical data in the training.
For future work, we aim develop a deep learning architecture by adding clinical characteristics to the input data. Comparative time complexity analysis will also be included. In the future we will also create a benchmark dataset for use in experiments. We will employ a preprocessing phase with data augmentation to enhance performance. In addition, a detection technique with bounding boxes can enhance time complexity and precision.
In summary, this study shows that a deep learning architecture utilizing the R-CNN algorithm, a variant of RNN, can predict progressive visual field tests significantly better than the pointwise regression method and HMM algorithms. The R-CNN model is less affected by the reliability indices of the visual field input data. This could aid in decision-making by accurately predicting progressive visual field tests in clinical practice. Our R-CNN algorithm can also help clinicians make treatment decisions for ocular cases that have difficulty undergoing repeated visual field tests.

Funding

This research is funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R113), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Acknowledgments

I would like to thank Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R113), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Resnikoff, S.; Pascolini, D.; Etya’Ale, D.; Kocur, I.; Pararajasegaram, R.; Pokharel, G.P.; Mariotti, S.P. Global data on visual impairment in the year 2002. Bull. World Health Organ. 2004, 9, 844–851. [Google Scholar]
  2. Weinreb, R.N.; Aung, T.; Medeiros, F.A. The Pathophysiology and Treatment of Diabetic retinopathy: A Review. JAMA 2014, 311, 1901. [Google Scholar] [CrossRef]
  3. Henson, D.B.; Chaudry, S.; Artes, P.H.; Faragher, E.B.; Ansons, A. Response Variability in the Visual Field: Comparison of Optic Neuritis, Diabetic retinopathy, Ocular Hypertension, and Normal Eyes. Investig. Ophthalmol. Vis. Sci. 2000, 41, 5. [Google Scholar]
  4. Wang, M.; Shen, L.Q.; Pasquale, L.R.; Petrakos, P.; Formica, S.; Boland, M.V.; Wellik, S.R.; De Moraes, C.G.; Myers, J.S.; Saeedi, O.; et al. An Artificial Intelligence Approach to Detect Visual Field Progression in Diabetic retinopathy Based on Spatial Pattern Analysis. Investig. Ophthalmol. Vis. Sci. 2019, 60, 365. [Google Scholar] [CrossRef] [PubMed]
  5. Murata, H.; Araie, M.; Asaoka, R. A New Approach to Measure Visual Field Progression in Diabetic retinopathy Ocular cases Using Variational Bayes Linear Regression. Investig. Ophthalmol. Vis. Sci. 2014, 55, 8386–8392. [Google Scholar] [CrossRef]
  6. Wen, J.C.; Lee, C.S.; Keane, P.A.; Xiao, S.; Rokem, A.S.; Chen, P.P.; Wu, Y.; Lee, A.Y. Forecasting progressive Humphrey Visual Fields using deep learning. PLoS ONE 2019, 14, e0214875. [Google Scholar] [CrossRef] [PubMed]
  7. Berchuck, S.I.; Mukherjee, S.; Medeiros, F.A. Estimating Rates of Progression and Predicting Progressive Visual Fields in Diabetic retinopathy Using a Deep Variational Autoencoder. Sci. Rep. 2019, 9, 18113. [Google Scholar] [CrossRef]
  8. Jun, T.J.; Eom, Y.; Kim, D.; Kim, C.; Park, J.H.; Nguyen, H.M.; Kim, D. TRk-CNN: Transferable ranking-CNN for image classification of glaucoma, glaucoma suspect, and normal eyes. Expert Syst. Appl. 2021, 182, 115211. [Google Scholar] [CrossRef]
  9. Manias, G.; Kiourtis, A.; Mavrogiorgou, A.; Kyriazis, D. Multilingual Sentiment Analysis on Twitter Data Towards Enhanced Policy Making. In Proceedings of the IFIP International Conference on Artificial Intelligence Applications and Innovations, León, Spain, 14–17 June 2022; Springer: Cham, Switzerland, 2022; pp. 325–337. [Google Scholar]
  10. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  11. Salehinejad, H.; Sankar, S.; Barfett, J.; Colak, E.; Valaee, S. Recent Advances in Recurrent Neural Networks. arXiv 2017, arXiv:1801.01078. [Google Scholar]
  12. Liu, S.; Yang, N.; Li, M.; Zhou, M. A Recursive Recurrent Neural Network for Statistical Machine Translation. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, MD, USA, 22–27 June 2014; Volume 1, pp. 1491–1500. [Google Scholar]
  13. Young, T.; Hazarika, D.; Poria, S.; Cambria, E. Recent Trends in Deep Learning Based Natural Language Processing. IEEE Comput. Intell. Mag. 2018, 13, 55–75. [Google Scholar] [CrossRef]
  14. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  15. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  16. Park, K.; Kim, J.; Lee, J. Visual Field Classification using Recurrent Neural Network. Sci. Rep. 2019, 9, 8385. [Google Scholar] [CrossRef]
  17. Dixit, A.; Yohannan, J.; Boland, M.V. Assessing Diabetic retinopathy Progression Using Machine Learning Trained on Longitudinal Visual Field and Clinical Data. Ophthalmology 2021, 128, 1016–1026. [Google Scholar] [CrossRef] [PubMed]
  18. Lynn, H.M.; Pan, S.B.; Kim, P. A Deep Bidirectional R-CNN Network Model for Biometric Electrocardiogram Classification Based on Recurrent Neural Networks. IEEE Access. 2019, 7, 145395–145405. [Google Scholar] [CrossRef]
  19. Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  20. Khandelwal, S.; Lecouteux, B.; Besacier, L. Comparing R-CNN and Hmm For Automatic Speech Recognition; LIG: Indore, India, 2016. [Google Scholar]
  21. Li, X.; Li, X.; Ma, X.; Xiao, F.; Xiao, C.; Wang, F.; Zhang, S. Time-series production forecasting method based on the integration of Bidirectional Gated Recurrent CNN (R-CNN) network and Sparrow Search Algorithm (SSA). J. Pet. Sci. Eng. 2022, 208, 109309. [Google Scholar] [CrossRef]
  22. Darmawahyuni, A.; Nurmaini, S.; Rachmatullah, M.N.; Firdaus, F.; Tutuko, B. Unidirectional-bidirectional recurrent networks for cardiac disorders classification. Telkomnika 2021, 19, 902. [Google Scholar] [CrossRef]
  23. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
  24. Ferreira, J.; Júnior, A.A.; Galvão, Y.M.; Barros, P.; Fernandes, S.M.M.; Fernandes, B.J. Performance Improvement of Path Planning algorithms with Deep Learning Encoder Model. In Proceedings of the 2020 Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob), Valparaiso, Chile, 26–30 October 2020; pp. 1–6. [Google Scholar]
  25. Pascanu, R.; Gulcehre, C.; Cho, K.; Bengio, Y. How to Construct Deep Recurrent Neural Networks. arXiv 2013, arXiv:1312.6026. [Google Scholar]
  26. Gomes, J.C.; Masood, A.I.; Silva, L.H.d.S.; da Cruz Ferreira, J.R.B.; Júnior, A.A.F.; dos Santos Rocha, A.L.; de Oliveira, L.C.P.; da Silva, N.R.C.; Fernandes, B.J.T.; Dos Santos, W.P. COVID-19 diagnosis by combining RT-PCR and pseudo-convolutional machines to characterize virus sequences. Sci. Rep. 2021, 11, 1–28. [Google Scholar]
  27. Majumder, S.; Elloumi, Y.; Akil, M.; Kachouri, R.; Kehtarnavaz, N. A deep learning-based smartphone app for real-time detection of five stages of diabetic retinopathy. In Real-Time Image Processing and Deep Learning 2020; International Society for Optics and Photonics; SPIE: Bellingham, WA, USA, 2020; Volume 11401, p. 1140106. [Google Scholar]
  28. Asaoka, R.; Murata, H.; Iise, A.; Araie, M. Detecting Preperimetric Diabetic retinopathy with Standard Automated Perimetry Using a Deep Learning Classifier. Ophthalmology 2016, 123, 1974–1980. [Google Scholar] [CrossRef] [PubMed]
  29. Garway-Heath, D.F.; Poinoosawmy, D.; Fitzke, F.W.; Hitchings, R.A. Mapping the Visual Field to the Optic Disc in Normal Tension Diabetic retinopathy Eyes. Ophthalmology 2000, 107, 7. [Google Scholar]
  30. Silva, L.H.d.S.; George, O.d.A.; Fernandes, B.J.; Bezerra, B.L.; Lima, E.B.; Oliveira, S.C. Automatic Optical Inspection for Defective PCB Detection Using Transfer Learning. In Proceedings of the 2019 IEEE Latin American Conference on Computational Intelligence (LA-CCI), Guayaquil, Ecuador, 11–15 November 2019; pp. 1–6. [Google Scholar]
  31. Borrelli, E.; Sacconi, R.; Querques, L.; Battista, M.; Bandello, F.; Querques, G. Quantification of diabetic macular ischemia using novel three-dimensional optical coherence tomography angiography metrics. J. Biophotonics 2020, 13, e202000152. [Google Scholar] [CrossRef]
  32. Zang, P.; Gao, L.; Hormel, T.T.; Wang, J.; You, Q.; Hwang, T.S.; Jia, Y. DcardNet: Diabetic retinopathy classification at multiple levels based on structural and angiographic optical coherence tomography. IEEE Trans. Biomed. Eng. 2020, 68, 1859–1870. [Google Scholar] [CrossRef]
  33. Ayala, A.; Ortiz Figueroa, T.; Fernandes, B.; Cruz, F. Diabetic Retinopathy Improved Detection Using Deep Learning. Appl. Sci. 2021, 11, 11970. [Google Scholar] [CrossRef]
  34. Dai, L.; Wu, L.; Li, H.; Cai, C.; Wu, Q.; Kong, H.; Liu, R.; Wang, X.; Hou, X.; Liu, Y.; et al. A deep learning system for detecting diabetic retinopathy across the disease spectrum. Nat. Commun. 2021, 12, 3242. [Google Scholar] [CrossRef]
  35. Lyu, X.; Jajal, P.; Tahir, M.Z.; Zhang, S. Fractal dimension of retinal vasculature as an image quality metric for automated fundus image analysis systems. Sci. Rep. 2022, 12, 11868. [Google Scholar] [CrossRef]
  36. Pour, E.K.; Rezaee, K.; Azimi, H.; Mirshahvalad, S.M.; Jafari, B.; Fadakar, K.; Faghihi, H.; Mirshahi, A.; Ghassemi, F.; Ebrahimiadib, N.; et al. Automated machine learning–based classification of proliferative and non-proliferative diabetic retinopathy using optical coherence tomography angiography vascular density maps. Graefes Arch. Clin. Exp. Ophthalmol. 2022. [Google Scholar] [CrossRef]
Figure 1. The progress of diabetic retinopathy with image 1 from a mild to a severe case in image 6.
Figure 1. The progress of diabetic retinopathy with image 1 from a mild to a severe case in image 6.
Axioms 11 00614 g001
Figure 2. The timeline of the images of a patient who went through thirteen visual field tests. The visual field test times are specified in the timeline and are utilized in training, and those in the black timeline are utilized in the prediction.
Figure 2. The timeline of the images of a patient who went through thirteen visual field tests. The visual field test times are specified in the timeline and are utilized in training, and those in the black timeline are utilized in the prediction.
Axioms 11 00614 g002
Figure 3. The architecture of the HMM and R-CNN models. HMM = Hidden Markov; R-CNN = recurrent CNN; DV = Deviation value. (a) The structure of the HMM method (b) The structure of the proposed R-CNN model.
Figure 3. The architecture of the HMM and R-CNN models. HMM = Hidden Markov; R-CNN = recurrent CNN; DV = Deviation value. (a) The structure of the HMM method (b) The structure of the proposed R-CNN model.
Axioms 11 00614 g003
Figure 4. The sequence diagram of the model testing.
Figure 4. The sequence diagram of the model testing.
Axioms 11 00614 g004
Figure 5. The number of cases according to the mean prediction error (DB).
Figure 5. The number of cases according to the mean prediction error (DB).
Axioms 11 00614 g005
Figure 6. (A) Mean absolute error (MAE) of classified deviation value (DV). R-CNN had the least classification error (MAE) in most of the points. The more opaque hue indicates greater error. Red dots depict significant variances between the regression and R-CNN; blue dots depict significant variances between HMM and R-CNN (t-test). (B) Normalized squared error (C) Normalized absolute error.
Figure 6. (A) Mean absolute error (MAE) of classified deviation value (DV). R-CNN had the least classification error (MAE) in most of the points. The more opaque hue indicates greater error. Red dots depict significant variances between the regression and R-CNN; blue dots depict significant variances between HMM and R-CNN (t-test). (B) Normalized squared error (C) Normalized absolute error.
Axioms 11 00614 g006
Figure 7. Representation of the optic nerve (a) and spatial field (b,c). (b) The spatial field is partitioned into six parts (c) The spatial field is partitioned into central and marginal parts.
Figure 7. Representation of the optic nerve (a) and spatial field (b,c). (b) The spatial field is partitioned into six parts (c) The spatial field is partitioned into central and marginal parts.
Axioms 11 00614 g007
Figure 8. Regression investigation of classification error (MSE) and other parameters. (a) MSE correlation with false positive rate. (b) MSE correlation with false negative rate. (c) MSE correlation with loss function value. (d) MSE correlation with visual field mean deviation (D).
Figure 8. Regression investigation of classification error (MSE) and other parameters. (a) MSE correlation with false positive rate. (b) MSE correlation with false negative rate. (c) MSE correlation with loss function value. (d) MSE correlation with visual field mean deviation (D).
Axioms 11 00614 g008
Table 1. Recent research in diabetic retinopathy prediction intelligent models.
Table 1. Recent research in diabetic retinopathy prediction intelligent models.
ReferenceResearch DescriptionProposed SolutionDatabaseAverage
Accuracy
[24]Deep learning CNN for early detection of stages of diabetic retinopathyThe model uses markers for classification to predict abnormalities by computing features correlation.980 Fundus oculi
images
91.5%
[25]Deep learning diagnosis of pre-parametric retinopathy due to diabetes with automated perimetry methodologyDeep learning using Fourier polynomialsSmall-sized dataset, cannot be
generalized
91.7%
[26]Cornea classification by mapping visual field of diabetic retinopathy eyesPixel-differentiation of the Fundus oculi
images
2000 Fundus oculi images91.57%
with high recall
[27]Fundus oculi imaging irregularities detection of optical identification of PCB using transfer learningIntelligent classification modelunknown91.7%
[28]Multi-label retinopathy ocular classification of diabetic macular ischemia utilizing 3-D coherence methodDense neural network130092.5%
[29]Quantifying diabetic retinopathy niches using OCT imaging defining DcardNet: multi—
classification at multiple levels based on
structural and angiographic of optical
retinopathy.
Discrete domain-optical analysisFundus oculi image 2100 Fundus oculi images87.93%
[30]Deep image CNN for diabetic retinopathy
diagnosis.
Feature data mining detection in retinal
fundus
950 Fundus oculi
images of five
labelled diabetic
retinopathy cases
89.16%
[31]Automated corneal image analysis with the exclusion of areas that does not indicate
dangerous disease
Regional CNN4130 Fundus oculi images90.97%
[32]Progression diabetic retinopathy in corneal fundus oculi videos using the fractal dimensionImage-Net convolutional neural network1700 videos with 25 frames each88.4%
[33]Deep learning prediction of proliferative
diabetic retinopathy employing optical
angiography vascular density
Geometric parameters1320 3-D Fundus
oculi images
90.7%
Our proposed modelA multitasking fusion deep CNN for
detecting the progression of diabetic retinopathy phases from no-diabetic retinopathy to severe
diabetic retinopathy progression over 4.3 years on average.
14,000 oculi imagesMSE and
p-values are used
Table 2. Characteristics of the datasets.
Table 2. Characteristics of the datasets.
CharacteristicsThe Whole DatasetTraining DataTesting Data
Number of ocular cases (each eye)14,000 (7000)11,200 (5600)2800 (1400)
Age; average ± standard deviation49.96 ± 16.0444.11 ± 14.8849.19 ± 16.84
Initial field: IF (dB); average ± standard deviation−4.89 ± 6.21−4.77 ± 6.16−6.19 ± 6.44
Follow up (years); average ±   standard deviation4.69 ± 2.744.87 ± 2.874.61 ± 1.84
Average number of visual field tests8.48 ± 2.088.82 ± 2.226.00 ± 0.00
IF   6 dB44162688828
5 dB > IF     13 dB1218881226
13 dB > IF1062846208
Dataset extension
Cases of the dataset with eight eyes series822280611282
Follow up (years); average ± standard deviation4.28 ± 1.684.26 ± 1.664.61 ± 1.84
Detection time (years); average ± standard deviation0.84 ± 0.820.82 ± 0.811.00 ± 0.84
IF 5 dB66884861828
5 dB > IF   13 dB14881241226
13 dB < IF12681068208
Table 3. Classes of diabetic retinopathy.
Table 3. Classes of diabetic retinopathy.
Progress in Years02–4 Years4–8 Years8–12 Years>12 Years
Class of diabetic
retinopathy
NormalMild diabetic
retinopathy
Moderate diabetic retinopathySevere diabetic retinopathyProliferate diabetic
retinopathy
Damage to retinaNo
retinopathy
Minute alteration in blood vessels.blood vessels
leakage.
Larger blood leakages and. vessel blockage.Vision loss.
Table 4. The count of images in each diabetic retinopathy classes.
Table 4. The count of images in each diabetic retinopathy classes.
Diabetic Retinopathy Class/Count of ImagesTraining SetTesting Set
Left EyeRight EyeLeft EyeRight Eye
Normal (No diabetic retinopathy)12241226197203
Mild diabetic retinopathy12001231190189
Moderate diabetic retinopathy21022240395390
Severe diabetic retinopathy421448313318
Proliferate diabetic retinopathy353355305300
Table 5. Statistics of the test dataset.
Table 5. Statistics of the test dataset.
Number of Cases
Total2100
Gender, Male (%)1092 (52%)
Diagnosis
    Diabetic retinopathy suspect560
    Primary open angle diabetic retinopathy840
    Pseudo exfoliation diabetic retinopathy100
    Primary angle closure diabetic retinopathy299
    Secondary diabetic retinopathy190
    Others111
Table 6. Comparison of average MSE and P-value between regression method, HMM, and R-CNN.
Table 6. Comparison of average MSE and P-value between regression method, HMM, and R-CNN.
Regression MethodHMMR-CNNANOVA
p-Value
p-Value
Regression Method vs.
R-CNN
HMM vs.
R-CNN
Regression Method vs. HMM
Prediction error,
average ± standard deviation
MSE (dB)5.81 ± 5.895.06 ± 3.615.71 ± 3.53<0.001<0.001<0.001<0.001
PMAE (dB)5.53 ± 0.565.10 ± 0.593.80 ± 0.56<0.001<0.001<0.001<0.001
Table 7. Confusion matrix for the regression model (300 test cases for each class).
Table 7. Confusion matrix for the regression model (300 test cases for each class).
Predicted CasesProliferate
Normal Mild Moderate Severe
Actual Cases Normal (No diabetic
retinopathy)
28012800
Mild diabetic
retinopathy
102701532
Moderate
diabetic retinopathy
116270103
Severe diabetic
retinopathy
0032907
Proliferate diabetic
retinopathy
00225273
Table 8. Confusion matrix for the HMM.
Table 8. Confusion matrix for the HMM.
Predicted CasesProliferate
NormalMildModerateSevere
Actual Cases Normal (No diabetic retinopathy) 28510500
Mild diabetic
retinopathy
10277832
Moderate diabetic
retinopathy
11428050
Severe diabetic
retinopathy
0002937
Proliferate diabetic
retinopathy
00119280
Table 9. Confusion matrix for the R-CNN.
Table 9. Confusion matrix for the R-CNN.
Predicted CasesProliferate
NormalMildModerateSevere
Actual Cases Normal (No diabetic
retinopathy)
2928000
Mild diabetic
retinopathy
3293400
Moderate diabetic
retinopathy
0429051
Severe diabetic
retinopathy
0002955
Proliferate diabetic retinopathy 0005295
Table 10. The prediction error (MSE) by visual field partitions.
Table 10. The prediction error (MSE) by visual field partitions.
Prediction Error (MSE, dB),
Average ± Standard Deviation
p-Value
Regression MethodHMMR-CNNR-CNN vs. HMMR-CNN vs. Regression MethodRegression Method vs. HMM
Spatial4.85 ± 5.084.39 ± 3.864.03 ±   3.55<0.001<0.001<0.001
Temporal4.94 ± 5.534.79 ± 4.534.38 ± 3.91<0.001<0.0010.310
Intertemporal5.58 ± 5.194.78 ± 4.054.54 ±   3.85<0.001<0.001<0.001
Nose angle5.34 ± 5.755.30 ± 4.344.97 ±   4.36<0.001<0.001<0.001
Marginal5.90 ± 4.955.05 ± 3.604.74 ±   3.58<0.001<0.001<0.001
Dominant5.08 ± 5.184.76 ± 4.154.44 ±   3.68<0.0010.001<0.001
Table 11. Mean classification error (MSE) according to reliability metrics.
Table 11. Mean classification error (MSE) according to reliability metrics.
Prediction error (MSE, dB), Average ± Standard
Deviation
Number of Eyesp-Value
Regression MethodHMMR-CNNR-CNN vs. HMMR-CNN vs. Regression MethodRegression Method vs. HMMANOVA
prediction error vs. false positive rate (FPR, %)
FP rate ≤ 25.90 ± 5.435.06 ± 3.654.71 ± 3.55797<0.001<0.001<0.001<0.001
2 < FP 55.75 ± 4.355.18 ± 3.694.80 ± 3.54358<0.001<0.001<0.001<0.001
5 < FP 85.43 ± 3.534.83 ± 3.484.53 ± 3.1873<0.001<0.0010.007<0.001
8 < FP 10.04.90 ± 3.384.74 ± 3.144.45 ± 1.9557<0.0010.0010.431<0.001
FP rate > 105.15 ± 4.195.19 ± 3.544.85 ± 3.4488<0.001<0.001<0.001<0.001
prediction error and false negative (FN rate %)
FN rate ≤ 2.55.34 ± 4.884.58 ± 3.594.33 ± 3.31766<0.001<0.001<0.001<0.001
2 < FN 55.16 ± 3.934.43 ± 1.794.10 ± 1.59155<0.001<0.001<0.001<0.001
5 < FN 85.63 ± 4.035.05 ± 3.415.57 ± 3.06109<0.001<0.0010.007<0.001
8 < FN ≤ 10.05.65 ± 3.915.53 ± 3.055.30 ± 1.8991<0.001<0.001<0.001<0.001
FN rate > 107.43 ± 5.676.36 ± 4.045.95 ± 4.08151<0.001<0.001<0.001<0.001
prediction error vs. loss function (L, %)
L ≤ 35.91 ± 5.885.04 ± 3.754.66 ± 3.53518<0.001<0.001<0.001<0.001
3 < L ≤ 56.55 ± 3.995.99 ± 3.305.17 ± 3.06140.0030.0350.533<0.001
5 < L ≤ 85.59 ± 3.875.08 ± 3.614.71 ± 3.48175<0.001<0.0010.001<0.001
8 < L ≤ 114.95 ± 4.554.05 ± 3.193.86 ± 3.10141<0.001<0.001<0.001<0.001
L > 115.98 ± 3.945.45 ± 3.504.98 ± 3.45545<0.001<0.001<0.001<0.001
Classification error and mean deviation (D, dB)
D < −117.40 ± 5.566.98 ± 3.596.30 ± 3.69340<0.001<0.0010.174<0.001
−11 ≤ D < −86.88 ± 3.866.57 ± 3.055.85 ± 3.1080<0.001<0.0010.339<0.001
−8 ≤ D < −55.99 ± 3.555.54 ± 1.905.03 ± 1.80153<0.001<0.0010.003<0.001
−5 ≤ D < −25.68 ± 4.974.70 ± 1.954.55 ± 1.73378<0.001<0.001<0.001<0.001
−3 ≤ D4.30 ± 4.133.38 ± 1.383.15 ± 1.17553<0.001<0.001<0.001<0.001
Table 12. Correlation coefficients and linear regression analyses between classification error and reliability and between classification error and visual field average deviation.
Table 12. Correlation coefficients and linear regression analyses between classification error and reliability and between classification error and visual field average deviation.
Correlation CoefficientsLinear Regression Analysis
Spearman’s rhop-ValueSlopeIntercept R 2 p-Value
classification error vs. false positive rate
Regression method−0.0240.344−0.0424.9110.0010.329
HMM−0.0430.040−0.0414.1840.0020.048
R-CNN−0.0420.134−0.0383.8040.0020.141
classification error vs. false negative rate
regression method0.444<0.0010.4443.1420.143<0.001
HMM0.443<0.0010.3492.4020.234<0.001
R-CNN0.448<0.0010.3422.4140.249<0.001
classification error vs. fixation loss percentage
regression method0.0830.0030.0114.424<0.0010.424
HMM0.0410.0290.0243.8810.0020.101
R-CNN0.0440.0040.0293.4940.0040.032
classification error vs. average visual field average deviation
regression method−0.441<0.001−0.2243.4030.128<0.001
HMM−0.443<0.001−0.2432.4340.382<0.001
R-CNN−0.444<0.001−0.2182.3430.304<0.001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hosni Mahmoud, H.A. RETRACTED: Diabetic Retinopathy Progression Prediction Using a Deep Learning Model. Axioms 2022, 11, 614. https://doi.org/10.3390/axioms11110614

AMA Style

Hosni Mahmoud HA. RETRACTED: Diabetic Retinopathy Progression Prediction Using a Deep Learning Model. Axioms. 2022; 11(11):614. https://doi.org/10.3390/axioms11110614

Chicago/Turabian Style

Hosni Mahmoud, Hanan A. 2022. "RETRACTED: Diabetic Retinopathy Progression Prediction Using a Deep Learning Model" Axioms 11, no. 11: 614. https://doi.org/10.3390/axioms11110614

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop