Next Article in Journal
Musical Novices Are Unable to Judge Musical Quality from Brief Video Clips: A Failed Replication of Tsay (2014)
Previous Article in Journal
Vergence and Accommodation Cues in Stereo-Localization during the Small-In Large-Out (SILO) Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessing the Performance of a Novel Bayesian Algorithm at Point of Care for Red Eye Complaints

1
Schulich School of Medicine and Dentistry, Western University, 1151 Richmond St., London, ON N6A 5C1, Canada
2
Department of Ophthalmology, Schulich School of Medicine and Dentistry, Western University, 1151 Richmond St., London, ON N6A 5C1, Canada
*
Author to whom correspondence should be addressed.
Vision 2022, 6(4), 64; https://doi.org/10.3390/vision6040064
Submission received: 11 August 2022 / Revised: 5 October 2022 / Accepted: 21 October 2022 / Published: 24 October 2022

Abstract

:
The current diagnostic aids for red eye are static flowcharts that do not provide dynamic, stepwise workups. The diagnostic accuracy of a novel dynamic Bayesian algorithm for red eye was tested. Fifty-seven patients with red eye were evaluated by an emergency medicine physician who completed a questionnaire about symptoms/findings (without requiring extensive slit lamp findings). An ophthalmologist then attributed an independent “gold-standard diagnosis”. The algorithm used questionnaire data to suggest a differential diagnosis. The referrer’s diagnostic accuracy was 70.2%, while the algorithm’s accuracy was 68.4%, increasing to 75.4% with the algorithm’s top two diagnoses included and 80.7% with the top three included. In urgent cases of red eye (n = 26), the referrer diagnostic accuracy was 76.9%, while the algorithm’s top diagnosis was 73.1% accurate, increasing to 84.6% (top two included) and 88.5% (top three included). The algorithm’s sensitivity for urgent cases was 76.9% (95% CI: 56–91%) using its top diagnosis, with a specificity of 93.6% (95% CI: 79–99%). This novel algorithm provides dynamic workups using clinical symptoms, and may be used as an adjunct to clinical judgement for triaging the urgency of ocular causes of red eye.

1. Introduction

Patients often present to healthcare providers with a red eye and no clear diagnosis [1]. This leads to delayed management and extraneous referrals or tests [2,3].
Clinical specialists’ diagnostic ability stems from Bayesian thinking. Their differential diagnosis is continuously adjusted at each step of decision-making by history or physical exam findings (pre-test odds), which inform their next clinical step [4]. However, clinical generalists (non-ophthalmologists) who consult resources such as UpToDate and clinical practice guidelines will find no Bayesian diagnostic aids. UpToDate provides encyclopedic, accurate documentation on clinical diseases, but does not provide easy, stepwise approaches [5]. Clinical practice guidelines use static algorithms (traditional flowcharts) that provide general approaches but do not account for pre-test probability when proposing the next steps [6]. They necessitate the same series and number of questions to yield a diagnosis, rendering them inflexible and inefficient [7,8].
Machine learning and artificial intelligence (AI) are increasingly being integrated into clinical settings for diagnostic aid. Within ophthalmology, they are helping to automate the detection or tracking of corneal ulcers, corneal ectasia, and iritis and to predict the prognosis of various diseases [9]. Improving diagnostic accuracy at the patient’s initial contact with the health care system can improve triaging [10]. However, current AI tools are largely image-based and track quantitative findings, and do not consider clinical symptoms or signs.
A robust, dynamic algorithm that streamlines a medical workup may fill the gap in available diagnostic aids at point of care. The author (AMD) developed Pickle, a novel app that employs Bayesian algorithms, to provide primary care practitioners with workups for common ophthalmic complaints. The algorithms use a dynamic Bayesian feedback process to continuously recreate themselves at each decision-making step.
This study tested the app’s red eye algorithm by comparing its diagnostic accuracy to that of referring physicians and ophthalmologists. The paper does not aim to present a detailed analysis of the algorithmic design itself, as this lies outside the medical scope.

2. Materials and Methods

2.1. Pickle App Design and Development

The author (AMD) developed Pickle, a novel mobile application with its user interface shown by Figure 1. Pickle first requires answers to 3–4 “Yes/No/I don’t know” questions to begin narrowing its differential diagnosis. It subsequently provides a differential diagnosis ranked by likelihood. Next, Pickle’s proprietary Bayesian-based artificial intelligence determines the most appropriate next workup step. Specifically, the differential diagnosis is parsed against the question bank to determine which questions can rule out diagnoses, significantly reduce their likelihood, or significantly increase their likelihood. Each question is stored with a table of values that represents its ability to stratify diseases by increasing or decreasing the level of suspicion depending on the answer. Employing Bayesian principles, a dynamic process ensues, whereby the question that garners the highest potential difference between pre-test and post-test probability is asked next to expedite decision-making. Further constraints are imposed on this dynamic order of questions, including the presence of prerequisite investigations and the elimination of questions that would add redundancy. For example, if the user answers “Yes” to “Is there corneal staining with fluorescein?”, the algorithm would prioritize the subsequent questions: “Is there contact lens use?” and “Is there a whitish corneal opacity?” to assess for a corneal ulcer. It would also eliminate the question “Is there chronic tearing?” from the question bank, as the answer would not add diagnostic value. Therefore, the algorithm’s novelty lies in its ability to adjust its order of questions to each case uniquely, as an ophthalmologist would, by balancing clinical suspicion with requisite investigations.
After each question, the algorithm re-sorts the differential by likelihood, displaying numerical and visual scores (Figure 1, Step 5) for each diagnosis that are indicative of its suspicion. The dynamic questioning is repeated until the user terminates the program or the level of suspicion for the top diagnosis heavily outweighs that of the others. Moreover, users can select a diagnosis within the differential, and this will be provided with the symptoms or history that support or refute it. The algorithm may also provide simple management steps depending on the responses provided, such as “Refer to Ophthalmology to rule out a corneal ulcer.”

2.2. Study Design and Sample

This prospective study tests whether a dynamic Bayesian algorithm can improve the diagnostic accuracy of red eye complaints when compared to that of primary care physicians, using ophthalmology residents’ diagnostic impression as a gold standard.
A questionnaire was developed with a list of all possible algorithm questions (Figure 2) which prompted for medical history and physical examination findings. It did not necessitate extensive slit lamp findings, prompting only for the presence of large corneal opacities or areas of fluorescein uptake.
The paper questionnaire was distributed to emergency departments who referred adult patients with red eye to an ophthalmology urgent eye clinic in London, Ontario, Canada. Questionnaires prompted for the referring physician’s diagnosis and patients’ medical history and physical exam findings without using checkboxes or multiple-choice options. Patients who presented with simultaneous new vision loss or diplopia were included. The main inclusion criteria were (i) adults (18 years or older) who (ii) presented with red eye, as (iii) assessed by the referring physician with (iv) no prior diagnosis by ophthalmology. Patients were seen by the ophthalmology clinic less than one day after referral, with no clinical interventions in the interim. A consecutive series sample of 57 patients was completed. This sample size allowed for all common causes of red eye to be represented. Data were collected between March and August 2021.
Referrers completed the questionnaire with their suspected diagnosis (the “referrer diagnosis”), which was then sent with a referral to ophthalmology. One of eight resident ophthalmologists then assessed the patient and attributed a gold-standard diagnosis after reviewing with a staff ophthalmologist. Since the resident received the referral, he or she was not blinded to the referrer diagnosis. After the first 21 patients, the questionnaire was updated to also prompt the ophthalmology resident for his or her impression of the medical history and physical exam findings, as the research team aimed to correlate these—which sometimes differed from the referring doctor’s findings—to the ultimate diagnosis. Post-visit, questionnaire data were entered into Pickle’s red eye algorithm, generating an “algorithm differential”. If a response was not documented by the referrer, it was entered as “I don’t know” on the algorithm. The ophthalmology resident was blinded to the algorithm differential.

2.3. Algorithm Diagnoses

The algorithm may make 15 possible diagnoses for red eye:
  • Eczema or Reaction to Topical Medications;
  • Blocked Tearduct;
  • Lid Abnormality (ectropion, entropion, or lagophthalmos);
  • Blepharitis;
  • Dry Eye;
  • Subconjunctival Blood;
  • AVM;
  • Conjunctivitis;
  • Episcleritis;
  • Scleritis;
  • Keratitis/Corneal Abrasion;
  • Corneal Ulcer;
  • Acute Angle-Closure Glaucoma;
  • Iritis;
  • Endophthalmitis/Severe Inflammation.
This list of possible algorithm diagnoses was not available to the resident and staff ophthalmologist. However, it was available to the investigators who entered the clinical data from referrers into the algorithm to assess for diagnostic accuracy. Certain external pathologies that cause red eye, including cellulitis (pre-septal and orbital), hordeolum, chalazion, pingueculitis, blunt ocular trauma, and globe-penetrating injuries, were not included in the algorithm’s diagnostic options as they are usually clinically apparent.
A list of abbreviations and their meanings are shown in Table 1.

2.4. Outcome Measures and Data Analysis

The referrer diagnostic accuracy was calculated as the concordance between the “referrer diagnosis” and the “gold-standard diagnosis”. If no diagnosis was attempted by the referring physician, it was considered an incorrect diagnosis. Algorithm diagnostic accuracy was calculated as the concordance between the “algorithm differential” and the “gold-standard diagnosis”. Given that the algorithm ranks its differential diagnoses by likelihood, algorithm accuracy was calculated using three divisions: the accuracy of the top-scoring diagnosis (Top 1), the top two diagnoses (Top 2), and the top three diagnoses (Top 3).
The referrer and algorithm diagnostic accuracies were also computed for a subset of urgent cases. This permitted the calculation of sensitivity and specificity for the algorithm’s ability to identify cases as urgent or non-urgent.

3. Results

Questionnaires were completed for 57 patients referred with red eye between March and August 2021, with all included for analysis.
Based on the gold-standard diagnosis, the causes of red eye were: eczema or reaction to topical medications (n = 1), blocked tearduct (n = 0), lid abnormality (n = 0), blepharitis (n = 0), dry eye (n = 2), subconjunctival blood (n = 1), AVM (n = 0), conjunctivitis (n = 4), episcleritis (n = 0), scleritis (n = 4), keratitis/corneal abrasion (n = 16), corneal ulcer (n = 11), acute ACG (n = 1), iritis (n = 13), and endophthalmitis/severe inflammation (n = 1). Three cases were not able to be classified under the algorithm’s diagnostic options: post-operative pain (n = 1), and pingueculitis (n = 2).

3.1. Referrer Diagnostic Accuracy

The referrer diagnosis was correct in 70.2% (40/57) of all cases (Figure 3). The most common referrer diagnoses were keratitis/corneal abrasion (n = 13), corneal ulcer (n = 10) ad iritis (n = 9). (Figure 3, Table 2).

3.2. Algorithm Diagnostic Accuracy

The algorithm’s diagnostic accuracy, using its top diagnosis, was 68.4% (39/57). When considering the top two diagnoses, accuracy was 75.4% (43/57). When considering the top three diagnoses, accuracy was 80.7% (46/57) (Figure 3). Table 2 displays the referrer and algorithm diagnostic accuracies for each independent diagnostic cluster.

3.3. Referrer and Algorithm Diagnostic Accuracy in Urgent Conditions

Twenty-six cases were deemed “urgent conditions” requiring rapid referral to exclude serious pathology. These comprised diagnoses of corneal ulcer, acute ACG, iritis, and endophthalmitis/severe inflammation. The referrer diagnostic accuracy for urgent conditions was 76.9% (20/26) (Figure 4). The algorithm diagnostic accuracy for these urgent cases was 73.1% (19/26) when considering the top diagnosis. This increased to 84.6% (22/26) with the top two diagnoses and 88.5% (23/26) with the top three diagnoses.
For non-urgent cases, the algorithm was 64.5% accurate (20/31). The referrer sensitivity for determining that a case was urgent was 84.6% (95% CI: 65–96%), regardless of the diagnosis. The referrer specificity for the same was 83.9% (95% CI: 66–95%). However, the algorithm sensitivity for the urgency of a case was 76.9% (95% CI: 56–91%) using the top diagnosis, with a specificity of 93.6% (95% CI: 79–99%).

4. Discussion

Overall, the accuracy of the referring providers was 70.2% for all cases, increasing to 76.9% for urgent cases. Given that questionnaires were hand-written, anonymous, and non-prompted (no checkboxes for diagnoses), they were designed to encourage referrers to provide unbiased diagnoses. A recent study found that referrals to a regional emergency eye clinic were 45% accurate overall, with 54% for optometrists, 39% for emergency physicians and 33% for primary care practitioners [11]. This reveals that diagnostic decision aids would benefit non-ophthalmologists in assessing red-eye complaints.
Given equivalent clinical data, the algorithm overall improved on the referrer diagnostic accuracy (70.2%) to a range of 68.4–80.7%. A comprehensive slit lamp exam requires significant practice and skill to be executed well [12]. The Pickle algorithm, however, does not prompt referrers for more advanced findings such as anterior chamber cells, inquiring instead about gross, more apparent findings such as whitish corneal opacities or areas of focal fluorescein uptake.
The algorithm was designed to be highly specific to non-urgent conditions (subconjunctival blood, AVM, conjunctivitis, etc.) to impart confidence in referring when an urgent condition is proposed by the tool. The algorithm’s specificity in urgent cases was 93.6%, compared to a referral specificity of 83.9%. While the algorithm’s sensitivity for urgent cases was slightly lower (76.9%) than that of referrers (84.6%), this is primarily due to differences in clinical findings noted by referrers compared to the gold-standard clinicians. For example, several cases ultimately diagnosed with an urgent condition were noted by referrers to have absent ‘red flag’ signs or symptoms—such as the presence of a whitish corneal opacity—that were noted to be present when evaluated by ophthalmology. This is further discussed in the analysis of incorrect diagnoses below.
In eighteen cases, Pickle’s top diagnosis was not correct, with eleven of these not containing the gold-standard diagnosis in the top three differential diagnoses. The latter cases were analyzed to determine if the algorithm had behaved appropriately, given the presented clinical data. Of the eleven cases, the algorithm displayed desired judgement in seven cases, performed moderately well in three, and failed outright in one.
First, a corneal ulcer (by gold-standard designation) was diagnosed by the algorithm as keratitis, a correct judgement given the presence of a whitish opacity in the absence of fluorescein uptake and foreign body sensation. However, it did textually prompt the user after the fourth question to rule out a corneal ulcer given the presence of an infiltrate. Second, a case of iritis was diagnosed as a corneal ulcer. The referrer noted a foreign body sensation, focal staining, and contact lens use, with ophthalmology later noting 0.5+ cells in the anterior chamber. These findings suggest a corneal ulcer rather than iritis, given the absence of moderate/severe pain or abnormal pupil. The cells are likely a secondary reaction. Third, conjunctivitis was diagnosed as keratitis. The patient presented with a foreign body sensation and diffuse redness without lid crusting or discharge, making conjunctivitis unlikely. Fourth, a subconjunctival bleed was diagnosed as keratitis. The referrer noted a history of trauma with moderate pain and fluorescein staining, attributing these to keratitis. The referring and algorithm diagnoses are appropriate, since a subconjunctival bleed does not cause corneal changes and is likely an accessory finding. Fifth, scleritis was diagnosed as keratitis. The referrer noted a foreign body sensation, moderate/severe pain, and corneal staining, leading to a diagnosis of keratitis appropriate (Figure 5). Scleritis does not usually present with a significant foreign body sensation.
Ophthalmology agreed with most findings but did not note staining. The algorithm was re-run with this change and proposed scleritis as its second diagnosis, after keratitis (Figure 6).
Sixth, a corneal abrasion was diagnosed as acute ACG. The referrer noted a firm eye with intraocular pressure greater than 40 mmHg and an abnormal pupil, consistent with acute ACG. Ophthalmology, however, noted normal pressures and a normal pupil, with which the algorithm was re-run and proposed a correct diagnosis of keratitis/corneal abrasion. Seventh, an acute ACG was diagnosed as iritis. The referrer noted an abnormal pupil, moderate/severe pain, and hazy cornea, with sectoral redness and normal pressures. This presentation is more consistent with iritis. Ophthalmology noted that the intraocular pressure increased from 26 mmHg in the emergency room to 55 mmHg at their assessment. When re-run, the algorithm’s top diagnosis was acute ACG. In these seven cases, therefore, the algorithm was effective at correlating the symptoms and history with a suitable diagnosis.
In two cases, pingueculitis was diagnosed by the algorithm as keratitis. Referrers noted corneal staining and moderate pain. Since pingueculitis was not included as a possible diagnostic option, keratitis is an appropriate candidate and should be ruled out. In another case, a dry eye was diagnosed as keratitis in a patient with lid crusting, foreign body sensation, and contact lens use. When assessed by ophthalmology, no crusting or foreign body sensation was noted, though the cornea did stain. This is consistent with keratitis, and it is unclear why a ‘dry eye’ diagnosis was attributed by ophthalmology. When re-run using the updated information, the algorithm proposed keratitis, dry eye, and a corneal ulcer as diagnoses. Pickle’s algorithm failed in one case of post-operative pain that it diagnosed as scleritis. The patient was only noted to have moderate pain. Given the absence of any positive symptoms or signs, the algorithm was unable to stratify diseases by their likelihood, and therefore provided an arbitrary, inappropriate diagnosis.

Limitations

This study may be limited by the small sample size of patients presenting with eczema/reaction to topical medications, scleritis, and endophthalmitis. No patients with a blocked tearduct, lid abnormality, blepharitis, AVM, or episcleritis were assessed. This is likely due to the clinical setting of the trial. An emergency department has a selection bias compared to overall primary care settings, whereby serious and acute pathology such as iritis or corneal ulcers present more commonly. Additionally, only patients who were referred to specialist care were included, although they may be a representative sample of the target population since most patients presenting with an acute red eye are referred, as recommended by the American Academy of Ophthalmology and the Canadian Ophthalmological Society [13,14].
The algorithm was noted to be limited by user error. While Pickle failed to include the gold-standard diagnosis in its ‘top 3’ differential in eleven cases, it made appropriate judgements in at least seven given the history and physical exam findings presented to it. In particular, the presence and absence of corneal staining contributed to incorrect analysis, as the referrers and ophthalmology differed in their answer to this question in several cases. This suggests that further emphasis on history and symptoms rather than physical exam findings may aid primary care physicians and Pickle’s algorithms in making more accurate diagnoses.
This red eye algorithm helps to provide a practical approach to red eye. Its novelty lies in its dynamic nature: it adjusts the sequence of workup steps as it receives clinical findings, mimicking a specialist’s approach. Additionally, the algorithm provides a differential instead of a single diagnosis, better reflecting the nuances of clinical practice. Importantly, it ensures the continued consideration of critical diagnoses, irrespective of likelihood.

5. Conclusions

Primary care settings are often the first triage point for patients with red eye. This study obtained a referring diagnostic accuracy of 70.2%. Using Pickle’s Bayesian algorithms, diagnostic accuracy was improved to a range of 68.4–80.7%, although the research may be biased toward urgent and serious pathology, as it was conducted with patients presenting to an Emergency Department. Therefore, a small sample sizes was attained for benign pathologies. However, the algorithm’s high specificity in urgent cases (93.6%) makes it useful for decreasing the suspicion of acute ACG, iritis, corneal ulcers, and endophthalmitis. Furthermore, it provides this benefit using only the tools available to non-ophthalmologists, without requiring slit lamp findings other than noting white opacities and fluorescein uptake. It may, therefore, be used in primary care settings as an adjunct to clinical judgement to optimize referral decisions for serious and urgent conditions. By improving diagnostic accuracy and patient triaging, it may reduce resource utilization and improve patient care.

Author Contributions

Conceptualization, A.M.D.; Data curation, A.M.D. and A.B.; Formal analysis, A.M.D., A.B. and C.M.H.; Investigation, A.M.D. and A.B.; Methodology, A.M.D. and A.B.; Project administration, A.M.D., A.B. and C.M.H.; Resources, A.M.D.; Supervision, C.M.H.; Validation, A.M.D., A.B. and C.M.H.; Visualization, A.M.D. and A.B.; Writing—Original draft, A.M.D. and A.B.; Writing—Review and editing, A.M.D., A.B. and C.M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable, as the institution, after consultation, determined that no IRB oversight was required.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sarkar, U.; Bonacum, D.; Strull, W.; Spitzmueller, C.; Jin, N.; López, A.; Giardina, T.D.; Meyer, A.N.; Singh, H. Challenges of making a diagnosis in the outpatient setting: A multi-site survey of primary care physicians. BMJ Qual. Saf. 2012, 21, 641–648. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Patel, M.P.; Schettini, P.; O’Leary, C.P.; Bosworth, H.B.; Anderson, J.B.; Shah, K.P. Closing the referral loop: An analysis of primary care referrals to specialists in a large health system. J. Gen. Intern. Med. 2018, 33, 715–721. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Neimanis, I.; Gaebel, K.; Dickson, R.; Levy, R.; Goebel, C.; Zizzo, A.; Woods, A.; Corsini, J. Referral processes and wait times in primary care. Can. Fam. Physician 2017, 63, 619–624. [Google Scholar] [PubMed]
  4. Gill, C.J.; Sabin, L.; Schmid, C.H. Why clinicians are natural Bayesians. BMJ 2005, 330, 1080–1083. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Fenton, S.H.; Badgett, R.G. A comparison of primary care information content in UpToDate and the National Guideline Clearinghouse. J. Med. Libr. Assoc. 2007, 95, 255. [Google Scholar] [CrossRef] [Green Version]
  6. Butler, L.; Yap, T.; Wright, M. The accuracy of the Edinburgh diplopia diagnostic algorithm. Eye 2016, 30, 812–816. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Timlin, H.; Butler, L.; Wright, M. The accuracy of the Edinburgh red eye diagnostic algorithm. Eye 2015, 29, 619–624. [Google Scholar] [CrossRef] [Green Version]
  8. Goudie, C.; Khan, A.; Lowe, C.; Wright, M. The accuracy of the Edinburgh visual loss diagnostic algorithm. Eye 2015, 29, 1483–1488. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Armstrong, G.W.; Lorch, A.C. A (eye): A review of current applications of artificial intelligence and machine learning in ophthalmology. Int. Ophthalmol. Clin. 2020, 60, 57–71. [Google Scholar] [CrossRef] [PubMed]
  10. Kostopoulou, O.; Porat, T.; Corrigan, D.; Mahmoud, S.; Delaney, B.C. Diagnostic accuracy of GPs when using an early-intervention decision support system: A high-fidelity simulation. Br. J. Gen. Pract. 2017, 67, e201–e208. [Google Scholar] [CrossRef]
  11. Nari, J.; Allen, L.H.; Bursztyn, L.L. Accuracy of referral diagnosis to an emergency eye clinic. Can. J. Ophthalmol. 2017, 52, 283–286. [Google Scholar] [CrossRef] [PubMed]
  12. Gibbons, H.; Masih, M. Focus on slit lamp examination. Int. J. Ophthalmic Pract. 2010, 1, 8–12. [Google Scholar] [CrossRef]
  13. C.P.G.E. Committee. Canadian Ophthalmological Society evidence-based clinical practice guidelines for the periodic eye examination in adults in Canada, Canadian journal of ophthalmology. J. Can. D’ophtalmologie 2007, 42, 39–163. [Google Scholar] [CrossRef]
  14. AAO Hoskins Center for Quality Eye Care. Referral of Persons with Possible Eye Diseases or Injury. Clin. Guidel. 2014. Available online: https://www.aao.org/clinical-statement/guidelines-appropriate-referral-of-persons-with-po (accessed on 10 August 2022).
Figure 1. Pickle app user interface.
Figure 1. Pickle app user interface.
Vision 06 00064 g001
Figure 2. Questionnaire including all possible questions of the Pickle red eye algorithm.
Figure 2. Questionnaire including all possible questions of the Pickle red eye algorithm.
Vision 06 00064 g002
Figure 3. Referrer and algorithm diagnostic accuracies.
Figure 3. Referrer and algorithm diagnostic accuracies.
Vision 06 00064 g003
Figure 4. Referrer and algorithm diagnostic accuracy for urgent conditions.
Figure 4. Referrer and algorithm diagnostic accuracy for urgent conditions.
Vision 06 00064 g004
Figure 5. “Scleritis” case, using data from referrers.
Figure 5. “Scleritis” case, using data from referrers.
Vision 06 00064 g005
Figure 6. “Scleritis” case, using data from Ophthalmology.
Figure 6. “Scleritis” case, using data from Ophthalmology.
Vision 06 00064 g006
Table 1. Abbreviations.
Table 1. Abbreviations.
AbreviationMeaning
AVMArterio-venous malformation
ACGAngle-closure Glaucoma
Table 2. Referrer and algorithm diagnostic accuracies.
Table 2. Referrer and algorithm diagnostic accuracies.
Gold Standard DiagnosisDiagnostic Accuracy
Algorithm Differential
Referrer DiagnosisTop DiagnosisTop 2 DiagnosesTop 3 Diagnoses
Eczema or Reaction to Topical Medications (n = 1)1/11/11/11/1
Dry Eye (n = 2)1/21/21/21/2
Subconjunctival Blood (n = 1)0/10/10/10/1
Conjunctivitis (n = 4)2/42/42/43/4
Scleritis (n = 4)2/41/42/43/4
Keratitis/Corneal Abrasion (n = 16)13/1615/1615/1615/16
Corneal Ulcer (n = 11)10/119/1110/1110/11
Acute ACG (n = 1)0/10/10/10/1
Iritis (n = 13)9/139/1311/1312/13
Endophthalmitis/Severe Inflammation (n = 1)1/11/11/11/1
Other (n = 3)1/30/30/30/3
Total40/5739/5743/5746/57
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Deans, A.M.; Basilious, A.; Hutnik, C.M. Assessing the Performance of a Novel Bayesian Algorithm at Point of Care for Red Eye Complaints. Vision 2022, 6, 64. https://doi.org/10.3390/vision6040064

AMA Style

Deans AM, Basilious A, Hutnik CM. Assessing the Performance of a Novel Bayesian Algorithm at Point of Care for Red Eye Complaints. Vision. 2022; 6(4):64. https://doi.org/10.3390/vision6040064

Chicago/Turabian Style

Deans, Alexander M., Amy Basilious, and Cindy M. Hutnik. 2022. "Assessing the Performance of a Novel Bayesian Algorithm at Point of Care for Red Eye Complaints" Vision 6, no. 4: 64. https://doi.org/10.3390/vision6040064

APA Style

Deans, A. M., Basilious, A., & Hutnik, C. M. (2022). Assessing the Performance of a Novel Bayesian Algorithm at Point of Care for Red Eye Complaints. Vision, 6(4), 64. https://doi.org/10.3390/vision6040064

Article Metrics

Back to TopTop