Next Article in Journal
GD2 Identifies Cancer Stemness in Glioblastoma and Phytoalexin Library Screen Identifies Potential Novel Natural Inhibitors
Previous Article in Journal
Pridopidine Protects ALS Patient-Derived Neural Progenitor Cells via Sigma-1 Receptor Activation
Previous Article in Special Issue
Diagnostic Performance of Universal Transport Medium for Viral Polymerase Chain Reaction in Aqueous Humor Samples of Suspected Viral Uveitis: A Pilot Methods Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Use of Artificial Intelligence in the Interpretation of Electroretinography (ERG) Studies

1
Department of Ophthalmology, Townsville University Hospital, 100 Angus Smith Drive, Douglas, Townsville, QLD 4814, Australia
2
Faculty of Medicine, University of Queensland, Herston, Brisbane, QLD 4006, Australia
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2026, 27(8), 3491; https://doi.org/10.3390/ijms27083491
Submission received: 5 March 2026 / Revised: 4 April 2026 / Accepted: 6 April 2026 / Published: 14 April 2026
(This article belongs to the Special Issue Advances in the Pathophysiology and Treatment of Eye Diseases)

Abstract

Electroretinograms are an important diagnostic tool to measure retinal electrical activity. However, their interpretation, done by sub-specialised ophthalmologists, can be not only time consuming but also challenging to obtain due to availability. In recent years, studies have investigated the use of artificial intelligence in the analysis of electroretinograms. This systematic review summarises the accuracy of artificial intelligence in interpreting electroretinograms and appraises the studies included. The review comprises primary, peer-reviewed published studies that determined accuracy of artificial intelligence by comparison to an expert ophthalmologist. In the 14 studies retrieved from databases and published between 2006 and 2025, machine learning was the most widely used artificial intelligence, with an accuracy rate between 39.3% and 100%. Overall, the “artificial neural network” machine learning tool was the most accurate. Quality assessment of the studies demonstrated high bias in patient selection but robustness in the methodology for the reference standard, flow and timing. The results revealed potential benefits in the real-world use of artificial intelligence in ophthalmic diagnostic testing; however, the variability in results suggests a requirement for further investigation prior to clinical implementation.

1. Introduction

The electroretinogram (ERG) measures electrical activity of the retina in response to light stimulation [1,2]. Summated cellular signals from different layers of the retina are recorded as a trace. Specifically, phototransduction leads to hyperpolarisation of the retina’s outermost layer, the photoreceptor layer, which contains rod and cone cells [3]. Photoreceptor hyperpolarisation is represented as a negative “a-wave” form in the ERG [4]. In turn, there is a decrease in glutamate release from the photoreceptors to the post-synaptic bipolar cells, located in the inner nuclear layer of the retina [4,5]. Resulting depolarisation of bipolar cells signifies a positive “b-wave” in the ERG [4,5]. Finally, action potentials are produced in post-synaptic cells of the retinal ganglion cell layer. Oscillatory potentials are an ERG wave form representative of the ganglion cell activity [4]. Understanding the cellular components of this electrical circuit can assist in the diagnosis of dysfunction along this pathway.
The clinical utility of ERGs in early diagnosis of ophthalmic conditions is well documented, especially its ability to detect disease where there is no structural change shown on other imaging modalities [6]. It is essential for the diagnosis of a range of retinal disorders, and may also be used to monitor disease progress in dystrophies, forms of uveitis and drug toxicity [1,7,8]. Its role in targeting ganglion cell function, such as in optic neuropathy and early glaucoma, has also been established [1]. Three major types of ERGs exist: full-field (ffERG), multifocal (mfERG) and pattern (PERG), for which clinical applications differ [9].
ERG interpretation is performed by trained ophthalmologists in clinical practice [10]. The process can be time-consuming, and ophthalmologists with the necessary training are not always widely available [10,11]. Accurate diagnosis and staging are key to treatment and controlling disease progression, but can be delayed significantly by resourcing constraints [10].
Artificial intelligence (AI) methods such as machine learning and deep learning have increasingly been explored to aid clinical diagnosis in several fields of medicine, in many instances achieving expert-level performance [10,11,12]. In ophthalmology, which is heavily reliant on specialised investigations in various modalities, the scope for augmentation with computer-aided diagnosis is substantial [13]. Research has demonstrated the capability of machine and deep learning technologies in detecting and staging diseases such as diabetic retinopathy and glaucoma [14,15,16]. Optical coherence tomography, fundus photography and visual field testing are examples of investigations for which interpretation using AI has shown promise [13,15].
Relatively few studies, however, have assessed the applicability of AI to visual electrophysiology. To date, no systematic review has evaluated evidence on the potential role of AI technologies as a diagnostic aid in electroretinography.
This review aims to identify, appraise and synthesise primary literature assessing the accuracy of AI methods in the interpretation of ERG studies.

2. Materials and Methods

2.1. Protocol

This systematic review was conducted in accordance with the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) framework (see Figure 1) [17]. The PRISMA checklist can be found in the Supplementary Materials. This study was prospectively registered in PROSPERO, and the protocol can be accessed through study ID: CRD420261327277.

2.2. Literature Search Strategy

A comprehensive literature search was conducted across PubMed, Medline, Embase, Scopus and Web of Science international databases, from inception to present. Keywords in the search strategy included, but were not limited to, “artificial intelligence”, “machine learning”, “deep learning”, “computational neural network”, “predictive learning models”, “electroretinography” and “electroretinogram”. Keywords were combined using the Boolean operators “OR” and “AND”, as detailed in Supplementary Digital Content Appendix A. The systematic search was performed independently by two researchers (M.H. and A.T.).

2.3. Eligibility Criteria

The following inclusion criteria were defined:
  • Primary research, including comparative, retrospective and clinical research.
  • Publication in a peer-reviewed journal.
  • Use of artificial intelligence (AI) in the evaluation of electroretinography studies.
  • Comparison to ERG evaluation by expert reviewers (Ophthalmologists).
  • Studies conducted across all dates, populations and languages.
The following exclusion criteria were defined:
  • Abstracts without full-text availability.
  • Use of electrophysiology testing other than electroretinography, such as visual evoked potentials.
  • Reviews, conference papers, editorials, author-responses, theses and books.

2.4. Data Extraction and Risk of Bias Assessment

Covidence was used for data extraction and reporting bias. This was completed by two reviewers (M.H. and A.T.) independently, with discussion to resolve conflicts. The following data were extracted from each study: first author, year of publication, study location (country), study design, population (human or non-human, adult, paediatric, gender), mean age (years), disease for diagnosis, type of AI and key findings (accuracy or mean error of the tool). Findings were standardised across studies. Finally, risk of bias was assessed using the QUADAS-2 tool [18]. Developed for diagnostic accuracy studies, this tool assesses the risk of bias in patient selection, the index test, reference standard and flow and timing [18].

3. Results

3.1. Study Selection

The initial search generated 295 studies across the electronic databases. The search consisted of studies up until 24 November 2025 (date of search).
Data management for this study was conducted using EndNote version 21 and Covidence. Initially, duplicates were removed using the EndNote “Find Duplicates” function, with manual confirmation. Then, once imported into Covidence, two duplicates were removed.
Two independent researchers (M.H. and A.T.) conducted the search according to the inclusion and exclusion criteria. Initial screening of studies was based on title and abstract. Eligible papers were then assessed by full-text reviews. Any conflicts were resolved by discussion and consensus between investigators. One study initially appeared to meet inclusion criteria, but on closer analysis was found to evaluate visual evoked potentials predominantly, rather than ERGs [19].
Following implementation of the inclusion and exclusion criteria, this systematic review contained 14 studies. This selection process is illustrated in Figure 1.

3.2. Study Characteristics

A total of 14 studies were included in this review, and this is summarised in Table 1. Of note, two studies were conference presentations only and the full text was inaccessible for two studies [20,21,22,23]. Publication dates were between 2006 and 2025. Noting two studies had inadequate information, the total population in this review was 2997 people. Most studies were conducted in Turkey (5/14), followed by Russia (4/14). Where adequate data was available, weighted mean and standard deviations were performed to encapsulate the population of the study. Only eight of the studies reported age of the participants and they all contained adults with an average age range between 35 and 47 years. While most of the papers were retrospective cohort studies (8/14), there were three prospective cohort studies and one randomised controlled trial. The most commonly used ERG modality was the full-field ERG (7/14). Only two studies used deep learning AI tools, compared to 12 with machine learning AI. Finally, a wide range of conditions were used to test for diagnostic accuracy of the AI tool; however, the majority (5/14) assessed for healthy and unhealthy ERGs.

3.3. Artificial Intelligence Accuracy

The results of the 14 included studies, presented in Table 1, had varied accuracy in the interpretation of ERGs. Only one study found a diagnostic accuracy of 100% by machine learning technology. Specifically, this study used artificial neural networks to diagnose achromatopsia and congenital stationary night blindness [24]. Similarly, artificial neural networks performed well in other studies demonstrating accuracy rates of 92% and 94.2% when diagnosing optic neuritis [28,29]. Support vector machine learning comprised the next most utilised technology. Within one study, there were wide-ranging accuracy rates, where interpretation of macular dysfunction alone and cone and rod dysfunction was highly accurate at 96.7% and 93.8% respectively, while combined macular and cone dysfunction had detection rates as low as 39.3% [11]. An additional study using support vector machine learning demonstrated reasonable precision at 85.3% [27]. Furthermore, the same machine learning approach outperformed other tools within one study, achieving a positive prediction rate of 87.1% [30]. Subsequently, these authors conducted a further study with an equivalent cohort of retinitis pigmentosa participants, and found naïve Bayes machine learning to outperform the support vector machine with a diagnostic accuracy of 82.32% [31]. In a similar cohort of retinitis pigmentosa patients but by different authors, resnet50 was superior to other machine learning tools within the study at 94.9% accuracy [10]. Alternatively, time series forest machine learning had the highest accuracy rate (74%) when diagnosing optic neuropathy in a study evaluating several machine learning tools [25]. When assessing broadly healthy and unhealthy adults and children, machine learning combined with multi-layered signal processing (wavelet 1–4) achieved the highest accuracy in the adult group (83%); in contrast to the paediatric group that achieved, at best, 70% accuracy from machine learning combined with low-level signal processing (wavelet 1–2) [34]. The same author team used decision tree machine learning and had the lowest diagnostic rate, 52% in adults and 40% in children, across all studies in this systematic review [35]. A study found two machine learning groups to have impressive accuracy rates of 95% when diagnosing normal physiology on high noise artefact ERGs [26]. This was superior to the human electrophysiologist who, in comparison to minimal artefact ERGs, had an accuracy rate of 62% [26]. There were only two studies by the same author group that used deep learning tools. They found the greatest diagnostic precision by vision transformer deep learning tool at 84% and 88%, 84.9% and 85%, and 87.5% and 91% for interpreting maximum ERG responses, scotopic ERG responses, and photopic ERG responses respectively [32,33].

3.4. Quality Assessment

The risk of bias in the methodology for each study is displayed in Figure 2. In terms of patient selection, 57% (8/14) of studies had a high risk of bias and the most common explanations for this classification were lack of random or consecutive sampling and absence of inclusion and exclusion criteria. Only 21% (3/14) of studies achieved low bias for patient selection. Most studies (64%) had unclear index testing bias due to deficient information on blind evaluation. Conversely, high index test bias was seen in 21% (3/14) and low bias was seen in 14% (2/14). Bias of reference standard was much reduced, with 57% (8/14) exhibiting low bias and only 7% (1/14) with high bias. About 71% (10/14) of studies had low bias in terms of flow and timing. Consequently, only 7% (1/14) had high bias in this classification. No single study achieved low bias across all criteria. However, three studies had low bias in three categories and an unclear status in the fourth.

4. Discussion

4.1. Results Interpretation

To the best of the authors’ knowledge, this is the first systematic review on the evaluation of AI in ERG interpretation. Additionally, this is corroborated by the PROSPERO database. In this systematic review, the machine learning tool, “artificial neural network” had the most impressive rates of accuracy compared to other machine learning and deep learning apparatus. Similar high-accuracy outputs by “artificial neural network” have been demonstrated in other fields of diagnostic medicine [36,37]. However, our findings indicate that there was a wide range of diagnostic accuracy in terms of AI interpretation of ERGs. This differs from some of the more established uses of AI in ophthalmic diagnostics, where results are more consistent. AI detection of diabetic retinopathy on fundus imaging has achieved an average accuracy of 93.6% across six countries [38]. To diagnose keratoconus, AI models have obtained precision between 91.9% and 98.9% [39]. Additionally, glaucoma detection had consistent results and accuracies of between 76% and 98.3% with machine learning tools such as “support vector machine”, “naïve Bayes” and “decision trees” [40].
Moreover, when deep learning AI has been used to interpret other electrophysiology tests, such as cardiac electrophysiology, it has achieved cardiologist-level performance with arrhythmia detection accuracy in over 95% [41]. Unlike electrocardiograms, however, ERGs are highly susceptible to noise interference making interpretation more challenging [6]. Due to the small amplitude received from the retinal and biologic artefact created by eyelid and facial muscles, noise is a significant limitation in the assessment of ERGs [42]. This could explain the current discrepancies in AI accuracy and encourage ERG de-noising techniques to improve reliability [6]. Alternatively, as seen in one of the included studies, AI can be trained using highly noise-contaminated ERGs to achieve lower error rates than human experts [26].
Electroencephalograms are analogous to ERGs in terms of their need for specialised interpretation and high susceptibility for noise artefact [43]. Despite this, an AI model has been able to correctly classify a large cohort of electroencephalograms with an accuracy of 88.3% [43]. Previous AI models in this field had insufficient precision for clinical implementation [44]. Translated to ERGs, this suggests that continued development of AI interpretation models and assessment of larger cohorts is necessary.
Implementation of AI in medical diagnostics has demonstrated increased efficiency, alleviated workforce burden and optimised patient health outcomes [45]. With development of AI models in future research, ERGs and electrophysiologists will likely reap the same benefits.

4.2. Strengths and Limitations

This systematic review was comprehensive, with a predefined search strategy across five international databases. While having a clearly outlined inclusion and exclusion criteria, it remained thorough by including all dates, all languages and human and non-human subjects. The search, study selection, data extraction and quality assessment were conducted independently by two reviewers, increasing the credibility of the findings.
There was significant heterogeneity across the included studies which limited this review. The populations varied between adults and children; additionally, two studies lacked information regarding the population. While broadly, machine learning was most frequently implemented, the subset AIs used were of wide variability and this is also reflected in the accuracy outcomes. As seen in other systematic reviews on the use of AI in ophthalmology, most studies contained patient sampling bias and were frequently unclear regarding blinding [46,47]. With substantial overlap in authors across studies, this increased the risk of publication bias and impaired the generalisability of results. Additionally, only one randomised controlled trial met criteria for the review, highlighting the limited quality of the studies included. The integrity of the review was influenced by small sample sizes, where most studies had less than 150 participants. These limitations highlight the need for further studies on this subject.

5. Conclusions

This review demonstrated promise in the use of AI as a diagnostic tool for interpreting ERGs. In comparison to other fields where AI has been successfully implemented, the studies in this review contain high variability. In ophthalmology, ERGs are an objective test that can make a diagnosis at the cellular level. Given the capacity of AI to achieve precision, reduce workforce burden and impact patient care, our recommendation is for further studies assessing the accuracy of AI in the interpretation of ERGs.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijms27083491/s1. Reference [17] is cited in the Supplementary Materials.

Author Contributions

Conceptualization, M.H. and S.S.; methodology, M.H. and A.T.; validation, M.H. and A.T.; formal analysis, M.H. and A.T.; investigation, M.H. and A.T.; resources, M.H. and A.T.; data curation, M.H. and A.T.; writing—original draft preparation, M.H. and A.T.; writing—review and editing, M.H., A.T. and S.S.; visualization, M.H.; supervision, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analysed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Complete Search Strategy Across Five Databases

  • PubMed:
  • ((electroretinography[MeSH Terms]) OR (electroretinographies[MeSH Terms]) OR (electroretinography[Title/Abstract]) OR (electroretinographies[Title/Abstract]) OR (electroretinogram[Title/Abstract])).
  • AND
  • ((ai artificial intelligence[MeSH Terms]) OR (artificial intelligence[Title/Abstract]) OR (AI[Title/Abstract]) OR (active machine learning[MeSH Terms]) OR (algorithm, machine learning[MeSH Terms]) OR (deep learning[MeSH Terms]) OR (computational neural network[MeSH Terms]) OR (computational neural networks[MeSH Terms]) OR (machine?learning[Title/Abstract]) OR (deep?learning[Title/Abstract])).
  • Medline:
  • (MH “Artificial Intelligence+”) OR (MH “Intelligent Systems”) OR (MH “Machine Learning+”) OR (MH “Deep Learning+”) OR (MH “Large Language Models”) OR (MH “Multifactor Dimensionality Reduction”) OR (MH “Ensemble Learning+”) OR (MH “Federated Learning”) OR (MH “Reinforcement Machine Learning”) OR (MH “Representation Machine Learning”) OR (MH “Supervised Machine Learning+”) OR (MH “Support Vector Machine”) OR (MH “Transfer Machine Learning”) OR (MH “Unsupervised Machine Learning+”) OR (MH “Particle Swarm Optimization”) OR (MH “Pattern Analysis, Machine”) OR (MH “Prediction Methods, Machine+”) OR (MH “Predictive Learning Models”) OR (MH “Sentiment Analysis”).
  • AND
  • (MH “Electroretinography”) OR “electroretinograph*” OR “electroretinogram*”.
  • Embase:
  • ‘electroretinography’:ti,ab,kw OR ‘electroretinogram’:ti,ab,kw OR’electroretinographies’:ti,ab,kw).
  • AND
  • (‘artificial intelligence’:ti,ab,kw OR ‘machinelearning’:ti,ab,kw OR ‘deep learning’:ti,ab,kw) OR (‘artificial intelligence’/exp OR ‘artificial intelligence’ OR ‘artificial intelligence-assisted technology’/exp OR ‘artificial intelligence-assisted technology’) AND(‘electroretinography’ OR ‘electroretinogram’/exp OR ‘electroretinogram’ OR’electroretinograph’/exp OR ‘electroretinograph’).
  • Scopus:
  • TITLE-ABS-KEY (electroretinography OR electroretinographies OR electroretinogram) AND TITLE-ABS-KEY (“artificial intelligence” OR “AI” OR “machine learning” OR “deep learning”) AND (LIMIT-TO (DOCTYPE, “ar”)).
  • Web of Science:
  • electroretinography OR electroretinographies OR electroretinogram (Topic) and artificial intelligence OR AI OR machine learning OR deep learning (Topic).

References

  1. Salmon, J. Kanski’s Clinical Ophthalmology: A Systematic Approach, 10th ed.; Elsevier: Amsterdam, The Netherlands, 2025. [Google Scholar]
  2. Shabto, J.; Tsang, S.; Sharma, T.; Diaconita, V. Electroretinography. In Atlas of Inherited Retinal Diseases; Tsang, S.H., Sharma, T., Diaconita, V., Eds.; Springer Nature: Cham, Switzerland, 2025; pp. 21–26. [Google Scholar]
  3. Cornish, E.E.; Vaze, A.; Jamieson, R.V.; Grigg, J.R. The electroretinogram in the genomics era: Outer retinal disorders. Eye 2021, 35, 2406–2418. [Google Scholar] [CrossRef]
  4. Yang, T.-H.; Kang, E.Y.-C.; Lin, P.-H.; Wu, P.-L.; Sachs, J.A.; Wang, N.-K. The Value of Electroretinography in Identifying Candidate Genes for Inherited Retinal Dystrophies: A Diagnostic Guide. Diagnostics 2023, 13, 3041. [Google Scholar] [CrossRef] [PubMed]
  5. Ichinose, T.; Habib, S. On and off signaling pathways in the retina and the visual system. Front. Ophthalmol. 2022, 2, 989002. [Google Scholar] [CrossRef]
  6. Angus, Z.G.; Sarossy, A.; Parihar, P.; Sarossy, M. Improving the reliability of the pattern electroretinogram with adaptive filtering. Sci. Rep. 2025, 16, 1197. [Google Scholar] [CrossRef]
  7. Lai, T.Y.; Chan, W.M.; Lai, R.Y.; Ngai, J.W.; Li, H.; Lam, D.S. The clinical applications of multifocal electroretinography: A systematic review. Surv. Ophthalmol. 2007, 52, 61–96. [Google Scholar] [CrossRef]
  8. Dettoraki, M.; Moschos, M.M. The Role of Multifocal Electroretinography in the Assessment of Drug-Induced Retinopathy: A Review of the Literature. Ophthalmic Res. 2016, 56, 169–177. [Google Scholar] [CrossRef]
  9. Tsang, S.H.; Sharma, T. Electroretinography. Adv. Exp. Med. Biol. 2018, 1085, 17–20. [Google Scholar] [CrossRef]
  10. Güven, A.; Karaman, B.; Öner, A.; Sinim Kahraman, N. Detection of retinitis pigmentosa stages with GAN and transfer learning in maps of MfERG P1 wave amplitudes. Signal Image Video Process. 2025, 19, 528. [Google Scholar] [CrossRef]
  11. Glinton, S.L.; Calcagni, A.; Lilaonitkul, W.; Pontikos, N.; Vermeirsch, S.; Zhang, G.; Arno, G.; Wagner, S.K.; Michaelides, M.; Keane, P.A.; et al. Phenotyping of ABCA4 Retinopathy by Machine Learning Analysis of Full-Field Electroretinography. Transl. Vis. Sci. Technol. 2022, 11, 34. [Google Scholar] [CrossRef] [PubMed]
  12. Aggarwal, R.; Sounderajah, V.; Martin, G.; Ting, D.S.W.; Karthikesalingam, A.; King, D.; Ashrafian, H.; Darzi, A. Diagnostic accuracy of deep learning in medical imaging: A systematic review and meta-analysis. npj Digit. Med. 2021, 4, 65. [Google Scholar] [CrossRef] [PubMed]
  13. Ting, D.S.W.; Pasquale, L.R.; Peng, L.; Campbell, J.P.; Lee, A.Y.; Raman, R.; Tan, G.S.W.; Schmetterer, L.; Keane, P.A.; Wong, T.Y. Artificial intelligence and deep learning in ophthalmology. Br. J. Ophthalmol. 2019, 103, 167–175. [Google Scholar] [CrossRef] [PubMed]
  14. Das, D.; Biswas, S.K.; Bandyopadhyay, S. Detection of Diabetic Retinopathy using Convolutional Neural Networks for Feature Extraction and Classification (DRFEC). Multimed. Tools Appl. 2023, 82, 29943–30001. [Google Scholar] [CrossRef] [PubMed]
  15. Haja, S.A.; Mahadevappa, V. Advancing glaucoma detection with convolutional neural networks: A paradigm shift in ophthalmology. Rom. J. Ophthalmol. 2023, 67, 222–237. [Google Scholar] [CrossRef]
  16. Abràmoff, M.D.; Lou, Y.; Erginay, A.; Clarida, W.; Amelon, R.; Folk, J.C.; Niemeijer, M. Improved Automated Detection of Diabetic Retinopathy on a Publicly Available Dataset Through Integration of Deep Learning. Investig. Ophthalmol. Vis. Sci. 2016, 57, 5200–5206. [Google Scholar] [CrossRef]
  17. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. Int. J. Surg. 2021, 88, 105906. [Google Scholar] [CrossRef]
  18. Whiting, P.F.; Rutjes, A.W.S.; Westwood, M.E.; Mallett, S.; Deeks, J.J.; Reitsma, J.B.; Leeflang, M.M.G.; Sterne, J.A.C.; Bossuyt, P.M.M. QUADAS-2: A Revised Tool for the Quality Assessment of Diagnostic Accuracy Studies. Ann. Intern. Med. 2011, 155, 529–536. [Google Scholar] [CrossRef]
  19. Kara, S.; Güven, A. Neural network-based diagnosing for optic nerve disease from visual-evoked potential. J. Med. Syst. 2007, 31, 391–396. [Google Scholar] [CrossRef]
  20. Bandhey, H.; Del Risco, A.; Gao, Q.; Rathinavelu, J.; Pajic, M.; Hadziahmetovic, M. Exploring Deep-Learning on SD-OCT images as an Effective Mitigation Technique for Hydroxychloroquine-Induced Retinal Toxicity. Investig. Ophthalmol. Vis. Sci. 2022, 63, 3330-F0139. [Google Scholar]
  21. De Roach, J.N. Neural networks--an artificial intelligence approach to the analysis of clinical data. Australas. Phys. Eng. Sci. Med. 1989, 12, 100–106. [Google Scholar]
  22. Giap, B.D.; Likosky, K.; Lustre, J.; Srinivasan, K.; Khan, N.; Nallasamy, N. A Deep Learning-Assisted Electroretinogram Analysis System for Automated Abnormality Detection in Patient Retinal Function. Investig. Ophthalmol. Vis. Sci. 2024, 65, 5667. [Google Scholar]
  23. Giap, B.D.; Likosky, K.; Srinivasan, K.; Khan, N.W.; Nallasamy, N. Automated Abnormality Detection in Patient Retinal Function: A Deep Learning-Powered Electroretinogram Analysis System. In Proceedings of the 2024 46th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 15–19 July 2024; pp. 1–4. [Google Scholar]
  24. Bagheri, A.; Persano Adorno, D.; Rizzo, P.; Barraco, R.; Bellomonte, L. Empirical mode decomposition and neural network for the classification of electroretinographic data. Med. Biol. Eng. Comput. 2014, 52, 619–628. [Google Scholar] [CrossRef]
  25. Diao, T.; Kushzad, F.; Patel, M.D.; Bindiganavale, M.P.; Wasi, M.; Kochenderfer, M.J.; Moss, H.E. Comparison of Machine Learning Approaches to Improve Diagnosis of Optic Neuropathy Using Photopic Negative Response Measured Using a Handheld Device. Front. Med. 2021, 8, 771713. [Google Scholar] [CrossRef] [PubMed]
  26. Fisher, A.C.; Hagan, R.P.; Brown, M.C. Automatic positioning of cursors in the transient pattern electroretinogram (PERG) with very poor SNR using an Expert System. Doc. Ophthalmol. 2007, 115, 61–68. [Google Scholar] [CrossRef]
  27. Habib, F.; Huang, H.; Gupta, A.; Wright, T. MERCI: A machine learning approach to identifying hydroxychloroquine retinopathy using mfERG. Doc. Ophthalmol. 2022, 145, 53–63. [Google Scholar] [CrossRef] [PubMed]
  28. Kara, S.; Güven, A. Training a learning vector quantization network using the pattern electroretinography signals. Comput. Biol. Med. 2007, 37, 77–82. [Google Scholar] [CrossRef]
  29. Kara, S.; Güven, A.; Oner, A.O. Utilization of artificial neural networks in the diagnosis of optic nerve diseases. Comput. Biol. Med. 2006, 36, 428–437. [Google Scholar] [CrossRef]
  30. Karaman, B.; Güven, A.; Öner, A.; Kahraman, N.S. Classification of Retinitis Pigmentosa Stages Based on Machine Learning by Fusion of Image Features of VF and MfERG Maps. Electronics 2025, 14, 1867. [Google Scholar] [CrossRef]
  31. Karaman, B.; Öner, A.; Güven, A. Early detection and staging of retinitis pigmentosa using multifocal electroretinogram parameters and machine learning algorithms. Phys. Eng. Sci. Med. 2025, 48, 1185–1205. [Google Scholar] [CrossRef]
  32. Kulyabin, M.; Zhdanov, A.; Dolganov, A.; Maier, A. Optimal Combination of Mother Wavelet and AI Model for Precise Classification of Pediatric Electroretinogram Signals. Sensors 2023, 23, 5813. [Google Scholar] [CrossRef] [PubMed]
  33. Kulyabin, M.; Zhdanov, A.; Dolganov, A.; Ronkin, M.; Borisov, V.; Maier, A. Enhancing Electroretinogram Classification with Multi-Wavelet Analysis and Visual Transformer. Sensors 2023, 23, 8727. [Google Scholar] [CrossRef]
  34. Zhdanov, A.; Dolganov, A.; Zanca, D.; Borisov, V.; Ronkin, M. Advanced Analysis of Electroretinograms Based on Wavelet Scalogram Processing. Appl. Sci. 2022, 12, 12365. [Google Scholar] [CrossRef]
  35. Zhdanov, A.E.; Dolganov, A.Y.; Zanca, D.; Borisov, V.I.; Luchian, E.; Dorosinsky, L.G. Evaluation of the effectiveness of the decision support algorithm for physicians in retinal dystrophy using machine learning methods. Comput. Opt. 2023, 47, 272–277. [Google Scholar] [CrossRef]
  36. Issaiy, M.; Zarei, D.; Saghazadeh, A. Artificial Intelligence and Acute Appendicitis: A Systematic Review of Diagnostic and Prognostic Models. World J. Emerg. Surg. 2023, 18, 59. [Google Scholar] [CrossRef] [PubMed]
  37. Nakajima, K.; Kudo, T.; Nakata, T.; Kiso, K.; Kasai, T.; Taniguchi, Y.; Matsuo, S.; Momose, M.; Nakagawa, M.; Sarai, M.; et al. Diagnostic accuracy of an artificial neural network compared with statistical quantitation of myocardial perfusion images: A Japanese multicenter study. Eur. J. Nucl. Med. Mol. Imaging 2017, 44, 2280–2289. [Google Scholar] [CrossRef]
  38. Ting, D.S.W.; Cheung, C.Y.-L.; Lim, G.; Tan, G.S.W.; Quang, N.D.; Gan, A.; Hamzah, H.; Garcia-Franco, R.; San Yeo, I.Y.; Lee, S.Y.; et al. Development and Validation of a Deep Learning System for Diabetic Retinopathy and Related Eye Diseases Using Retinal Images from Multiethnic Populations With Diabetes. JAMA 2017, 318, 2211–2223. [Google Scholar] [CrossRef] [PubMed]
  39. Tiong, E.W.W.; Liu, S.-H.; Ting, D.S.J. Cochrane corner: Artificial intelligence for keratoconus. Eye 2024, 38, 3406–3408. [Google Scholar] [CrossRef]
  40. Shahriari, M.H.; Asadi, F.; Moghaddasi, H.; Roshanpour, A.; Sharifipour, F.; Khorrami, Z. Applications of machine learning in glaucoma diagnosis based on tabular data: A systematic review. BMC Biomed. Eng. 2025, 7, 9. [Google Scholar] [CrossRef]
  41. Canino, G.; Di Costanzo, A.; Salerno, N.; Leo, I.; Cannataro, M.; Guzzi, P.H.; Veltri, P.; Sorrentino, S.; De Rosa, S.; Torella, D. Artificial Intelligence in Cardiac Electrophysiology: A Clinically Oriented Review with Engineering Primers. Bioengineering 2025, 12, 1102. [Google Scholar] [CrossRef]
  42. Yamashita, T.; Miki, A.; Tabuchi, A.; Funada, H.; Kondo, M. A novel method to reduce noise in electroretinography using skin electrodes: A study of noise level, inter-session variability, and reproducibility. Int. Ophthalmol. 2017, 37, 317–324. [Google Scholar] [CrossRef]
  43. Tveit, J.; Aurlien, H.; Plis, S.; Calhoun, V.D.; Tatum, W.O.; Schomer, D.L.; Arntsen, V.; Cox, F.; Fahoum, F.; Gallentine, W.B.; et al. Automated Interpretation of Clinical Electroencephalograms Using Artificial Intelligence. JAMA Neurol. 2023, 80, 805–812. [Google Scholar] [CrossRef]
  44. Kural, M.A.; Jing, J.; Fürbass, F.; Perko, H.; Qerama, E.; Johnsen, B.; Fuchs, S.; Westover, M.B.; Beniczky, S. Accurate identification of EEG recordings with interictal epileptiform discharges using a hybrid approach: Artificial intelligence supervised by human experts. Epilepsia 2022, 63, 1064–1073. [Google Scholar] [CrossRef] [PubMed]
  45. Jeong, J.; Kim, S.; Pan, L.; Hwang, D.; Kim, D.; Choi, J.; Kwon, Y.; Yi, P.; Jeong, J.; Yoo, S.J. Reducing the workload of medical diagnosis through artificial intelligence: A narrative review. Medicine 2025, 104, e41470. [Google Scholar] [CrossRef]
  46. Jin, K.; Yu, T.; Ying, G.S.; Ge, Z.; Li, K.Z.; Zhou, Y.; Shi, D.; Wang, M.; Goktas, P.; Grzybowski, A. A systematic review of vision and vision-language foundation models in ophthalmology. Adv. Ophthalmol. Pract. Res. 2026, 6, 8–19. [Google Scholar] [CrossRef] [PubMed]
  47. Sachdeva, K.; Butt, F.R.; Mihalache, A.; Nassrallah, G.; Muni, R.H.; Popovic, M.M. A systematic review of artificial intelligence models in ocular tumour diagnosis. Can. J. Ophthalmol. 2026; in press. [CrossRef] [PubMed]
Figure 1. PRISMA 2020 flow diagram of the study [17].
Figure 1. PRISMA 2020 flow diagram of the study [17].
Ijms 27 03491 g001
Figure 2. Risk of bias assessment of the studies [10,11,24,25,26,27,28,29,30,31,32,33,34,35] using QUADAS-2 tool [18]. Bias displayed as Unclear (yellow), Low (green) and High (red).
Figure 2. Risk of bias assessment of the studies [10,11,24,25,26,27,28,29,30,31,32,33,34,35] using QUADAS-2 tool [18]. Bias displayed as Unclear (yellow), Low (green) and High (red).
Ijms 27 03491 g002
Table 1. Summary of study characteristics.
Table 1. Summary of study characteristics.
First AuthorYearCountryStudy TypePopulationMean Age (Years)Disease for DiagnosisType of ERG ModalityType of
Artificial Intelligence
Key Findings
Bagheri et al. [24]2014Italy; United States of AmericaRetrospective cohort studyAdult participants
n = 94
Males = 42/94
Females = 52/94
47 +/− 5Achromatopsia; Congenital stationary night blindnessFull-field ERGMachine learningDiagnostic accuracy of the artificial neural network machine learning tool was 100%
Diao et al. [25]2021United States of AmericaCase-control studyAdult participants
n = 119
Males = 56/119
Females = 63/119
45.6 +/− 17.5Optic neuropathyFull-field ERGMachine learningHighest diagnostic accuracy was by time series forest machine learning at 74%
Fisher et al. [26]2007United KingdomCluster randomised controlled trialHuman participants
n = 10
UnstatedNormal physiologyPattern ERGMachine learningDiagnostic accuracy was achieved at a rate of 95% in two machine learning groups compared to 62% in the human group interpreting noisy ERGs
Glinton et al. [11]2022United KingdomRetrospective cohort studyHuman participants
n = 597
Group 1 = 344/597
Group 2 = 44/597
Group 3 = 209/597
Group 1 = 35
Group 2 = 35
Group 3 = 37
Group 1 = Macular dysfunction alone
Group 2 = Macular dysfunction + generalised cone dysfunction
Group 3 = cone and rod dysfunction
Full-field ERGMachine learningDiagnostic accuracy of support vector machine
Group 1 = 96.7%
Group 2 = 39.3%
Group 3 = 93.8%
Guven et al. [10]2025TurkeyCase-control studyAdult participants
n = 206
Males = 118/206
Females = 88/206
35.37 +/− 15.01Retinitis pigmentosaMultifocal ERGMachine learningHighest diagnostic accuracy was by resnet50 machine learning at 94.9%
Habib et al. [27]2022Canada; ChinaRetrospective cohort studyHuman participants
n = 748
UnstatedHydroxychloroquine retinopathyMultifocal ERGMachine learningDiagnostic accuracy of the support vector machine machine learning was 85.3%
Kara et al. [28]2007TurkeyProspective cohort studyAdult participants
n = 320
Males = 164/320
Females = 156/320
41.94Optic neuritisPattern ERGMachine learningDiagnostic accuracy of the artificial neural network machine learning tool was 92%
Kara et al. [29]2006TurkeyProspective cohort studyAdult participants
n = 256
Males = 117/256
Females = 119/256
Unspecified gender = 20/256
43.8Optic neuritisPattern ERGMachine learningDiagnostic accuracy of the artificial neural network machine learning tool was 94.2%
Karaman et al. [30]2025TurkeyRetrospective cohort studyAdult participants
n = 124
Males = 73/124
Females = 51/124
35.26 +/− 14.30Retinitis pigmentosaMultifocal ERGMachine learningHighest diagnostic accuracy by support vector machine machine learning at 87.1%
Karaman et al. [31]2025TurkeyRetrospective cohort studyHuman participants
n = 97
Males = 61/97
Females = 36/97
37.48 +/− 16.19Retinitis pigmentosaMultifocal ERGMachine learningHighest diagnostic accuracy was by naïve Bayes machine learning at 82.32%
Kulyabin et al. [32]2023Germany; RussiaRetrospective cohort studyPaediatric participants
n = unstated
UnstatedHealthy and unhealthy, unstated specific diagnosisFull-field ERGDeep learningHighest diagnostic accuracy was by vision transformer with ricker wavelet deep learning at 84.0% for maximum ERG response, 84.9% for scotopic ERG response, 87.5% for photopic ERG response
Kulyabin et al. [33]2023RussiaProspective cohort studyAdult and paediatric participants
n = 323
UnstatedHealthy and unhealthy, unstated specific diagnosisFull-field ERGDeep learningHighest diagnostic accuracy was by visual transformer small deep learning at 88.0% for maximum ERG response, 85.0% for scotopic ERG response, 91% for photopic ERG response
Zhdanov et al. [34]2022Russia; GermanyRetrospective cohort studyAdult and paediatric participants
n = 103
Adult = 38
Paediatric = 65
UnstatedHealthy and unhealthy, unstated specific diagnosisFull-field ERGMachine learningHighest diagnostic accuracy was by classical features + wavelet 1–4 machine learning at 83% in the adult group. High diagnostic accuracy was by “wavelet 1–2” machine learning at 70% in the paediatric group
Zhdanov et al. [35]2023Russia; Germany; RomaniaRetrospective cohort studyAdult and paediatric participants
n = unstated
UnstatedHealthy and unhealthy, unstated specific diagnosisFull-field ERGMachine learningDiagnostic accuracy of the “decision trees” machine learning tool was 52% in the adult group and 40% in the paediatric group
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hegde, M.; Thomis, A.; Shirke, S. Use of Artificial Intelligence in the Interpretation of Electroretinography (ERG) Studies. Int. J. Mol. Sci. 2026, 27, 3491. https://doi.org/10.3390/ijms27083491

AMA Style

Hegde M, Thomis A, Shirke S. Use of Artificial Intelligence in the Interpretation of Electroretinography (ERG) Studies. International Journal of Molecular Sciences. 2026; 27(8):3491. https://doi.org/10.3390/ijms27083491

Chicago/Turabian Style

Hegde, Manasi, Alexander Thomis, and Sheetal Shirke. 2026. "Use of Artificial Intelligence in the Interpretation of Electroretinography (ERG) Studies" International Journal of Molecular Sciences 27, no. 8: 3491. https://doi.org/10.3390/ijms27083491

APA Style

Hegde, M., Thomis, A., & Shirke, S. (2026). Use of Artificial Intelligence in the Interpretation of Electroretinography (ERG) Studies. International Journal of Molecular Sciences, 27(8), 3491. https://doi.org/10.3390/ijms27083491

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop