Next Article in Journal
The Short-Term Retention of Depth
Previous Article in Journal
The Functional Network of the Visual Cortex Is Altered in Migraine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Validation of Digital Applications for Evaluation of Visual Parameters: A Narrative Review

by
Kevin J. Mena-Guevara
1,2,
David P. Piñero
1,3,* and
Dolores de Fez
1
1
Group of Optics and Visual Perception, Department of Optics, Pharmacology and Anatomy, University of Alicante, San Vicente del Raspeig, 03690 Alicante, Spain
2
Department of Pathology, University Miguel Hernández, Sant Joan d’Alacant, 03550 Alicante, Spain
3
Department of Ophthalmology, Vithas Medimar International Hospital, 03016 Alicante, Spain
*
Author to whom correspondence should be addressed.
Vision 2021, 5(4), 58; https://doi.org/10.3390/vision5040058
Submission received: 17 July 2021 / Revised: 11 November 2021 / Accepted: 22 November 2021 / Published: 24 November 2021

Abstract

:
The current review aimed to collect and critically analyze the scientific peer-reviewed literature that is available about the use of digital applications for evaluation of visual parameters in electronic devices (tablets and smartphones), confirming if there are studies calibrating and validating each of these applications. Three bibliographic search engines (using the search equation described in the paper) and the Mendeley reference manager search engine were used to complete the analysis. Only articles written in English and that are evaluating the use of tests in healthy patients to measure or characterize any visual function aspects using tablets or smartphones were included. Articles using electronic visual tests to assess the results of surgical procedures or are conducted in pathological conditions were excluded. A total of 19 articles meeting these inclusion and exclusion criteria were finally analyzed. One critical point of all these studies is that there was no mention of the characterization (spatial and/or colorimetrical) of screens and the stimuli used in most of them. Only two studies described some level of calibration of the digital device before the beginning of the study. Most revised articles described non-controlled comparatives studies (73.7%), reporting some level of scientific evidence on the validation of tools, although more consistent studies are needed.

1. Introduction

The evaluation of visual function is crucial in the clinical practice of eye care professionals. This evaluation combines different tests characterizing the patient’s ability to perceive and integrate external light stimuli captured by sight, with proper coordination of both eyes’ visual systems [1,2]. In other words, the evaluation of visual function considers the optical eye system’s imaging process and brain processing of information. In the clinical setting, different aspects are evaluated in order to obtain complete information about the visual system status, such as visual acuity (VA), contrast sensitivity (CS), binocular vision (BV)—which in turn includes phorias, fusional vergences, or near point of convergence—accommodation, or color vision [3]. All these evaluations are performed by clinicians using different instruments and procedures that sometimes are time consuming and tiring for the patient.
In the current digital era, the use of computers, smartphones, tablets, and smartwatches is a typical daily practice, and there has been attempts to transfer theses usages to the eye care professional’s clinical practice [1,2]. Only in 2016 in Spain, the National Observatory of Telecommunications, and Information Society (ONTSI) reported that 85% of Spanish internet users with ages ranging from 16 to 65 used social media for an average of 1 h per day, and this included the usage of computers (91%), smartphones (95%), and tablets (48%) [4]. Likewise, digital applications (apps) for different purposes have increased exponentially in recent years [4]. Among these apps, those corresponding to the health field (e-health and m-health) are widely used [5]. Specifically, there are numerous apps for evaluating different aspects of visual function that can be easily accessed via the App Store (iOS) or Google Play store (android system). The use of these apps should be conducted with care as no information about the scientific validation of these tools is normally provided. Using non-validated apps in clinical settings to evaluate different aspects of the visual function may result in incorrect clinical decisions [6].
Furthermore, significant discrepancies in image reproduction are present among electronic devices. For example, previous research from our group has demonstrated large color reproduction differences between smartphones (Samsung Galaxy S4 and iPhone 4s), and tablets (Samsung Galaxy Tab 3 and iPad 4) [3]. Likewise, it has even been demonstrated that there are significant differences in digital reproduction of visual stimuli among different units of the same tablet model [7]. Therefore, it is worth asking to what extent is the use of applications available in digital stores correct with respect to evaluating visual function. Furthermore, in another recent study by our research group, the differences in luminance reproduction between 20 tablets (Samsung Galaxy Tab A, SM-T519, 2019 version), as well as their implications for contrast reproduction were evaluated, having as a result differences between the devices even if they are from the same manufacturing batch [8]. Although our study is based on applications for smartphones and tablets, it must be disregarded that validation tests are also being carried out on virtual reality devices, as reported by the study by Wroblewski et al. [9] that conducted a validation of the VirtualEye system (with its respective characterization of screens: luminance measurement).
It is necessary to define two concepts for a better understanding of the validation of the digital tools, the concepts of characterization and validation. Characterization is the procedure that allows us to determine peculiar attributes of an object in our study of a screen so that it is clearly distinguished from the others. Thus, the characterization ensures that the designed test is reproduced in each device, because each one can have a different reproduction. In contrast, validation is the procedure of providing firmness or certainty to an action or theory. In our study, clinical validation refers to confirming that the object of study (app) measures in a similar or comparable manner to the measures obtained with the gold standard (traditional test). Ideally, both concepts should be linked, but one does not exclude the other; they are complementary. If the screen of the device has not been previously characterized, how can you be ensuring that the comparison with the gold standard is reliable? At the very least, they cannot be considered to provide the same level of scientific evidence [8,10,11].
In addition, an important point to consider in this type of study is the privacy of the data, since these devices are using sensitive information of patients and must be properly treated, according to the European or American regulations’ General Data Protection Regulation (GDPR, 1995) and Protected Health Information (PHI) developed in the Health Insurance Portability and Accountability Act (HIPAA, 1996). This is clearly detailed in the scoping review conducted by Benjumea et al. [12] 2020 and Apple’s privacy section [13], where the regulations for the Health App are detailed as well.
The current review aimed to collect and analyze critically the scientific peer-reviewed literature that is available about the use of digital applications for the evaluation of visual parameters in electronic devices (tablets and smartphones), confirming if there are studies calibrating and validating each of these applications. To our knowledge, this is the first review on this crucial aspect: the validity of visual function evaluation using validated apps. It should be considered that some digital visual functions are being used in clinical investigations without confirming that they are adequate for such evaluations. Therefore, it is unknown if the validity of the results provided in these investigations are biased due to the use of non-validated digital clinical tools.

2. Materials and Methods

Three bibliographic search engines (the utilization of the search equation is described later in this paper) and the Mendeley reference manager search engine were used to complete the analysis. The following inclusion criteria were established in order to focus the search and to delimit the results:
  • Articles showing the use of tests in healthy patients to evaluate any aspect of the visual function using tablets or smartphones;
  • Articles written in English.
The exclusion criteria for the current review included articles using electronic visual tests to evaluate the results of surgical procedures or in pathological conditions, studies involving animals, and articles showing simulations or theoretical results. Before a more detailed analysis in pathological cases, we preferred to focus our analysis on healthy eyes, which is the most optimal situation as the potential bias of measurements may have a less relevant impact on clinical decisions. Future analysis of the literature should be performed in the future including pathological or post-surgical cases and considering the results of this previous analysis in healthy eyes, with the aim to compare possible differences.
The search equation used for this review was as follows.
(“visual function” OR “visual acuity”) AND (“iPad” OR “app”) NOT (“Mice model” OR “Animal model”) NOT (“Amblyopia treatment” OR “ocular diseases”)
The results obtained with this search equation was first filtered after reading the titles and abstracts and considering the inclusion and exclusion criteria defined. The articles’ complete text that passed this first filter was obtained and read to confirm their explicit inclusion or exclusion in the review. Finally, a qualitative assessment of the results obtained was performed after classifying them by subject.

3. Results

3.1. Search Results

After searching on different platforms, a total of 248 articles were found (Figure 1). Specifically, 54 potentially useful results were found in the first search in PubMed (performed on 26 June 2019). On the same search date, 97 studies were found in the rest of the search platforms used.
Of all the articles found, 180 articles were excluded after verifying the first analysis described above for they did not meet the inclusion criteria. In the second analysis of the full text, 49 studies were excluded from the potentially eligible 68 articles (those previously selected).

3.2. Analysis of the Articles Included and Excluded

Table 1 summarizes the most relevant information of the articles finally included in the current article [2,6,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30].
A total of 49 articles were excluded from the review. Most of them showed a comparison between digital measurements of parameters, such as visual acuity, contrast sensitivity, and reading speed, with traditional measurement methods, concluding that digital and traditional measures could not be used interchangeably.
The following causes of exclusion were found for the excluded articles: unhealthy patients and operated patients (40 works); analysis using animal models or human tissue (5 articles); articles not written in English; or use of tablet simulations (3 works). Specifically, of the 40 articles excluded due to unhealthy status or previous surgery, 25% were conducted in low vision patients, 17.5% in amblyopic patients, 12.5% in age-related macular degeneration (AMD) patients, and 7.5% in both diabetes and multiple sclerosis subjects. The rest of the excluded cases (2.5%) include other pathological conditions: maculopathy, stroke, retinoschisis, senile dementia, hemangiopericytoma, Parkinson, and albinism strabismus, blindness, or dry eye.

4. Discussion

Most of the articles finally included in this critical review of the existing scientific literature on the use of digital devices for evaluating visual function (57.9%) reported the use of evaluations of visual acuity: 10.5% for the assessment of reading speed, and 5.3% (each) for the evaluation of contrast sensitivity and color vision, respectively. Likewise, some articles report using digital applications to evaluate more than one aspect of the visual function, with 5.3% of studies studying the entire visual function, including healthy patients and those with ocular abnormalities. It is worth noting that, on average, the level of evidence reported in all of these investigations is limited, with 73.7% of them being comparatives studies, 5.3% being observational studies, and 5.3% being case reports. A cross-sectional study, a bibliographic review, and a clinical trial (blind examiner) were found and analyzed. More consistent studies should be designed and performed in order to validate the great variety of digital apps that are currently available for the evaluation of visual function. Comparative analyses in which the order of performance of tests (digital and traditional) is assigned randomly, with clear descriptions of the calibration process of screens and illumination conditions of the examination room, with an additional analysis of reliability, and with different examiners performing digital and traditional tests should be conducted. Indeed, at least one of these types of studies should have been conducted for any app that was released before indicating the possibility of its clinical use.
As mentioned, one critical point of all these studies is that there is no mention or partial description of the visualization conditions present during the measurements (observation distance, screen tilt, ambience illumination level, and screen brightness) or the characterization (spatial and/or colorimetrical) of screens and the stimuli used in most of them. Only the studies from de Fez et al. [2] and Rodriguez-Vallejo et al. [25] mentioned that a previous characterization of the digital device used had been performed. Likewise, it is not clear that digital apps evaluating contrast sensitivity have considered that the concept of contrast should be based on luminance and not on a concept of the difference of digital levels. The contrast defined considering the digital levels of a screen is not equivalent to the classical definition of contrast considering luminance, the definition used to calculate contrast sensitivity [3]. Therefore, in addition to improvement in the design of the clinical studies for evaluating the clinical usefulness of digital apps for assessing visual function, more information should be provided about the characterization of the screens in order to know if the differences obtained between digital and traditional tests may be due to problems of contrast and color reproduction of screens [3,7]. de Fez et al. [7] performed spatial and colorimetric characterization of different devices using different methods, demonstrating that mathematical adjustment methods such as gain-offset-gamma (GOG) adjustments provide worse results in color reproduction than methods based on 3D LUT tables. Furthermore, they find that another problem arises if a test designed colorimetrically for one device is presented on a different device. The digital levels of a stimulus calculated for a given device can produce different chromaticities when reproduced in another, because colorimetric characterization is device dependent. They found that on iPad devices, color reproduction errors are below the minimum level distinguishable by the human eye when using the device’s characterization data, but they can be more than six times higher if this is not performed. This fact implies that vision tests (color vision; CSF-Contrast Sensitivity Function) designed for a particular device can result in erroneous diagnoses when administered in other devices, even those of the same model. As an example, Black et al. [14] evaluated 85 patients in a blind examining clinical trial and found that the “Visual Acuity XL” application used on an iPad device provided significantly worse visual acuity measures (mean difference 0.18 logMAR) than those obtained with traditional methods. The authors did not mention that this difference may be attributed to screen reproduction errors or discrepancies in the stimuli’s conception and design.
Regarding the characterization of screens, it should be also mentioned that the reproduction of the chromaticity and/or the luminance of the stimulus may depend on its position due to the inhomogeneity of the screen (this is another part of spatial characterization). A full-screen stimulus may not be perceived as uniform due to this lack of homogeneity, and a stimulus at x degrees from the center may also be reproduced with different characteristics than the same stimulus at -x degrees. If only a small and constant area of the screen is to be used for the test in question, this part of the characterization can be omitted. In terms of color reproduction, colorimetry covers both chromaticity and luminance of color. A luminance-only characterization can be carried out if the stimulus is achromatic, but it should be considered that while a white stimulus has a single luminance value to be measured and characterized, an achromatic stimulus has gray levels. The use of a gamma curve is an approximation of luminance, which can be more or less reliable depending on the behavior of the particular device. For example, a gamma curve does not reproduce saturation at high power-on levels, as can happen on LCD and TFT screens. In contrast, the use of 3DLUT tables, which also approximates luminance values by interpolation can increase the reliability of the approximation [7,8].
As in the current review, Hogarty et al. [6] remarked in their bibliographic research the relevance of using properly validated digital applications for any type of praxis with patients, including the analysis of the entire visual function. This was a consistent conclusion from this previous review, as any of the digital applications revised in this literature search had proper validation. It should be noted that the use of non-validated digital apps for evaluating visual parameters in any research can result in inaccurate conclusions and subsequently to incorrect “scientific-based” clinical decisions. Indeed, new guidelines should be defined in the future in order to classify or identify, in Google play and Apple stores, whether an app can be used or not for clinical purposes according to scientific validations associated with this technology. A specific type of signaling or labelling in the store should be developed in order to avoid the inadequate use of apps in patients.
Despite the lack of information about the validation of apps and digital tests used in most of revised articles, some of them analyzed the clinical equivalences between digital and traditional tests, providing an evaluation or discussion of interchangeability between them. Azis et al. [27] found in a sample of 195 patients aged between 5 and 6 years that there was a good correlation between the results obtained using the “AAPOS Vision Screening” application and those obtained with conventional optotypes (ETDRS and Lea symbols). Specifically, the results obtained using the app (Lea symbols) to test visual acuity by parents as a potential screening tool were compared to gold standard vision testing by an optometrist using the Lea symbols chart. The authors concluded that the app evaluated could be considered as a promising tool for visual acuity screening among Malaysian preschool children. However, due to the specific type of population evaluated and the scarcity of information about the tablet used and its level of calibration, the results obtained cannot be extrapolated to the general population, but it can be considered as a first level of validation of this app. Concerning the measurement of visual acuity with digital devices, there is an additional aspect that must be considered. The minimum spatial detail available for some electronic devices given their pixel size does not allow measuring hyperacuity thresholds for close viewing distances. Therefore, many devices will not support visual acuity beyond 6/6 (1 min arc detail, e.g., VR goggles) and they will not allow measuring with reliability people with acuities down to 6/3 (0.5 min arc).
Rodríguez-Vallejo et al. [25] evaluated visual acuity and contrast sensitivity in 45 patients (Spanish adult university population) by employing a self-developed application that was previously validated [31]. These authors measured the chromaticity of an iPad with a retina display using the Spyder4Elite colorimeter and the luminosity of the room with theLX1330B luxmeter. Likewise, the brightness of the screen was set at the maximum. Concerning visual acuity, these authors compared the results of the visual acuity test of the app that was based on the ATS (Amblyopia Treatment Study) procedure and HOTV optotypes with the values obtained with the ETDRS chart projected on a screen (Optec 6500 system), obtaining a mean difference of approximately three letters on a logMAR chart with five letters per line. Moreover, the agreement between app and conventional test with respect to reliability was evaluated by citing 25 subjects a total of two or more sessions that are spaced a week apart. Coefficients of reliability were 0.15 logMAR for our method and 0.17 logMAR for the ETDRS testing protocol. In terms of contrast sensitivity, the mean differences between the app and the measures obtained with the Functional Acuity Contrast Test (FACT) were lower than 0.05 log units for all spatial frequencies. The limits of agreement between digital and traditional tests were higher for high spatial frequencies. According to these authors, this finding was explained by the lower repeatability of both tests for those spatial frequencies [25]. This study is one of the most complete validation of an app with respect to evaluating some visual function parameters and should be considered when trying to design a validation study for a specific app. They described the fact that the brightness of the screen was set on the maximum level (342 cd/m2) was a limitation of their study, which is over the recommended background luminance [25]. They describe as a potential solution for this the future development of a system for measuring environmental illumination and automatically setting up background luminance in accordance with the measured value. Likewise, it should be considered that different levels of brightness were used on the screens of Optec 6500 and the iPad. However, the contrast sensitivity results obtained were equivalent, and this can be explained because doubling the luminance level improves one letter on a test of five letters per row in the range of 40 cd/m2 to 600 cd/m2. Although the contrast values on screens of different brightness are similar, the contrast threshold values change with brightness are not similar depending on the spatial frequency [32]; therefore, it is risky to state that the results are completely equivalent. In another study, Kollbaum et al. [33] found results with an iPad-based letter contrast sensitivity test that agreed with those obtained with the Freiburg Acuity and Contrast Test, but it was higher than those measured with the Pelli–Robson Test. These results indicate that the app evaluated was an efficient alternative for clinical use. This parameter was evaluated in 40 subjects (20 with low vision and 20 healthy) in a monocular mode. Likewise, Habtamu et al. [34] developed a new tumbling-E smartphone-based contrast sensitivity test (Peek Contrast Sensitivity, PeekCS) and was compared with a tumbling-E Pelli–Robson contrast sensitivity test. These authors found highly comparable results with both tests.
de Fez et al. [2] used the iPad application “Optopad,” designed to detect color vision deficiencies in two different clinical studies in order to evaluate diagnostic precision. First, a comparison with the Ishihara test was performed in 341 patients (children). A second comparative study with the Farnsworth–Munsell test (FM 100 H) was conducted in 66 patients (university adult population). Previous characterization of the digital device used was made, and the colorimetric adjustments required in the test were performed using the MATLAB software (R2008a) [3,6]. No statistically significant differences were found between “Optopad” and Isihara in detecting protan-deutan defects. When FM 100 H and “Optopad” were compared, a clinically reasonable level of predictability of “Optopad” data from the FM results was observed. Multiple regression analysis provided a regression coefficient (R2) of 0.86, with 80% of cases with residuals less than 25 units [2]. This is the second study of validation of a visual function test that provides a complete comparison with conventional tests after a careful calibration of the screens used for the reproduction of stimuli.
Another aspect of the visual function that can be measured with digital applications is the reading speed. Kingsnorth et al. [16] compared reading speed measured by a mobile application and the Radner reading test’s printed version in 21 patients with their usual correction. These authors found that the measurement of reading speed using both methods was not interchangeable. However, the authors concluded that the use of mobile devices was reliable and fast to perform. Notwithstanding, this potential advantage is not altogether valid if the data obtained with this application are not accurate enough.
Finally, a new app has been recently developed for the measurement of the defocus curve with an iPad, which has become an essential clinical test for an adequate evaluation of visual performance with any type of presbyopic correction, including cataract surgery with the implantation of multifocal intraocular lenses [30]. These authors confirmed that the measurement of the visual acuity defocus curve (VADC) showed good agreement with the ETDRS test and good repeatability (two consecutive measurements) despite the short testing time, but it should be mentioned that this interchangeability analysis was performed for visual acuity measures for the level of defocus of 0 D but not for the rest. However, the repeatability of measurements of contrast sensitivity defocus curve (CSDC) was around three times poorer than that obtained along the VADC. The authors concluded that CSDC had to be optimized in the future in order to obtain more repeatable results. In addition to these other articles have been published in recent years, showing the results of the validity of some apps as screening tools used in the evaluation of the visual acuity and visual field in different pathological conditions [35,36,37], but a previous validation in healthy population is crucial.
As mentioned in the introduction part of this study, privacy is a fundamental aspect to consider in the development and operation of a ‘health application,’ since highly sensitive information from patients is used, which must be correctly treated following the current regulation in place. In this manner, the study carried out by Turpin et al. [38] indicated in 2014 that they were subject to American regulations (HIPAA 1996) and that their data will only be uploaded to the cloud if the application (PsyPad) has an internet connection, which shows consistency with the work developed by Benjumea et al. [12] in 2020.
Population selection bias is a critical factor for extrapolating the results of these investigations to the general population, including aspects such as setting (26.3% hospital [1,20,22,28,29], university 21.1% [2,14,18,25]), and ethnicity (26.3%) [15,23,24,26,27] and age (15.8%) [17,20,29]. Likewise, there is a need in some studies to provide separate and adequate description of materials (10.5%) [16,21]. Most of the studies did not report the electronic device model used (commercial model and version of the tablet or iPad). Only a certain level of characterization of the screen used for the generation of the stimuli was performed in two of the studies. However, despite these limitations and the minimal evidence of characterization of digital applications evaluating the visual function, they are widely used in clinical practice and investigations, with the potential of extracting, in some cases, inaccurate conclusions and decisions.

5. Conclusions

There are many apps that have been used and validated clinically for measuring different aspects of visual function, such as visual acuity, contrast sensitivity, or color vision. However, limited information is provided in the peer-reviewed literature about the characterization of screens used for displaying stimuli. Therefore, it is not only crucial to use an app validated clinically but also to provide information about the characterization of the devices used for displaying stimuli. One app can be very well designed and validated but used in an electronic device that cannot reproduce the stimuli reliably. For this reason, the technical requirements in terms of screen characteristics must be also provided, and information on how data privacy was obtained must be preserved according to national and international regulations. Thus, an app that is valid for clinical purposes should be validated, and the technical requirements for stimuli reproduction should be well established, allowing the differentiation of well-developed apps from the numerous applications that evaluate visual function and that are available in digital stores. As we are healthcare personnel dealing with patients, the use of previously validated tools (applications) should be mandatory in order to ensure that the results obtained are correct or comparable with conventional methods.

Author Contributions

Conceptualization, K.J.M.-G., D.P.P. and D.d.F.; methodology, K.J.M.-G., D.P.P. and D.d.F.; data analysis K.J.M.-G., D.P.P. and D.d.F.; data collection, K.J.M.-G., D.P.P. and D.d.F.; resources, D.P.P. and D.d.F.; writing—original draft preparation, K.J.M.-G.; writing—review and editing, D.P.P. and D.d.F.; visualization, K.J.M.-G., D.P.P. and D.d.F.; supervision, D.P.P. and D.d.F.; project administration, D.P.P. and D.d.F.; funding acquisition, D.P.P. and D.d.F. All authors have read and agreed to the published version of the manuscript.

Funding

The author David P. Piñero has been supported by the Ministry of Economy, Industry and Competitiveness of Spain within the program Ramón y Cajal, RYC-2016-20471.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. O’Neill, S.; McAndrew, D.J. The validity of visual acuity assessment using mobile technology devices in the primary care setting. Aust. Fam. Physician 2016, 45, 212–215. [Google Scholar]
  2. de Fez, D.; Luque, M.J.; Matea, L.; Piñero, D.P.; Camps, V.J. New iPAD-based test for the detection of color vision deficiencies. Graefe’s Arch. Clin. Exp. Ophthalmol. 2018, 256, 2349–2360. [Google Scholar] [CrossRef] [Green Version]
  3. de Fez, D.; Luque, M.J.; García-Domene, M.C.; Camps, V.; Piñero, D. Colorimetric characterization of mobile devices for vision applications. Optom. Vis. Sci. 2015, 93, 85–93. [Google Scholar] [CrossRef]
  4. Montanera, R.; Julia, V. Study of Social Networks. 2018. Available online: https://iabspain.es/wp-content/uploads/estudio-redes-sociales-2018_vreducida.pdf (accessed on 2 June 2020).
  5. Vázquez Martínez, R.; Martínez López, M. Citizens before e-Health. Opinions and Expectations of Citizens on the Use and Application of ICT in the Health Field. 2016. Available online: https://www.ontsi.red.es/ontsi/sites/ontsi/files/los_ciudadanos_ante_la_e-sanidad.pdf (accessed on 2 June 2020).
  6. Hogarty, D.T.; Hogarty, J.P.; Hewitt, A.W. Smartphone use in ophthalmology: What is their place in clinical practice? Surv. Ophthalmol. 2020, 65, 250–262. [Google Scholar] [CrossRef]
  7. de Fez, D.; Luque, M.J.; García-Domene, M.C.; Caballero, M.T.; Camps, V.J. Can applications designed to evaluate visual function be used in different iPads? Optom. Vis. Sci. 2018, 95, 1054–1063. [Google Scholar] [CrossRef] [Green Version]
  8. Molina-Martín, A.; Piñero, D.P.; Coco-Martín, M.B.; Leal-Vega, L.; de Fez, D. Reproduction between Electronic Devices for Visual Assessment: Clinical Implications. Technologies 2021, 9, 68. [Google Scholar] [CrossRef]
  9. Wroblewski, D.; Francis, B.A.; Sadun, A.; Vakili, G.; Chopra, V. Testing of visual field with virtual reality goggles in manual and visual grasp modes. BioMed Res. Int. 2014, 2014, 206082. [Google Scholar] [CrossRef] [Green Version]
  10. Livingstone, I.A.; Lok, A.S.; Tarbert, C. New mobile technologies and visual acuity. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2014, 2014, 2189–2192. [Google Scholar] [PubMed]
  11. Tahir, H.J.; Murray, I.J.; Parry, N.R.; Aslam, T.M. Optimisation and assessment of three modern touch screen tablet computers for clinical vision testing. PLoS ONE 2014, 9, e95074. [Google Scholar] [CrossRef] [PubMed]
  12. Benjumea, J.; Ropero, J.; Rivera-Romero, O.; Dorronzoro-Zubiete, E.; Carrasco, A. Privacy Assessment in Mobile Health Apps: Scoping Review. JMIR mHealth uHealth 2020, 8, e18868. [Google Scholar] [CrossRef]
  13. Health App & Privacy. Available online: https://www.apple.com/legal/privacy/data/en/health-app/ (accessed on 4 November 2021).
  14. Black, J.M.; Jacobs, R.J.; Phillips, G.; Chen, L.; Tan, E.; Tran, A.; Thompson, B. An assessment of the iPad as a testing platform for distance visual acuity in adults. BMJ Open 2013, 3, e002730. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Zhang, Z.T.; Zhang, S.C.; Huang, X.G.; Liang, L.Y. A pilot trial of the iPad tablet computer as a portable device for visual acuity testing. J. Telemed. Telecare 2013, 19, 55–59. [Google Scholar] [CrossRef] [PubMed]
  16. Kingsnorth, A.; Wolffsohn, J.S. Mobile app reading speed test. Br. J. Ophthalmol. 2015, 99, 536–539. [Google Scholar] [CrossRef]
  17. Norgett, Y.; Siderov, J. Foveal crowding differs in children and adults. J. Vis. 2014, 14, 23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Manzano, A.A.; Lagamayo, M.A.N. A comparison of distance visual acuity testing using a standard ETDRS chart and a tablet device. Philipp. J. Ophthalmol. 2015, 40, 88–92. [Google Scholar]
  19. Perera, C.; Chakrabarti, R.; Islam, F.M.A.; Crowston, J. The eye phone study: Reliability and accuracy of assessing Snellen visual acuity using smartphone technology. Eye 2015, 29, 888–894. [Google Scholar] [CrossRef]
  20. Tofigh, S.; Shortridge, E.; Elkeeb, A.; Godley, B.F. Effectiveness of a smartphone application for testing near visual acuity. Eye 2015, 29, 1464–1468. [Google Scholar] [CrossRef]
  21. Kingsnorth, A.; Drew, T.; Grewal, B.; Wolffsohn, J.S. Mobile app Aston contrast sensitivity test. Clin. Exp. Optom. 2016, 99, 350–355. [Google Scholar] [CrossRef] [Green Version]
  22. Pathipati, A.S.; Wood, E.H.; Lam, C.K.; Sales, C.S.; Moshfeghi, D.M. Visual acuity measured with a smartphone app is more accurate than Snellen testing by emergency department providers. Graefes Arch. Clin. Exp. Ophthalmol. 2016, 254, 1175–1180. [Google Scholar] [CrossRef]
  23. Phung, L.; Gregori, N.Z.; Ortiz, A.; Schiffman, J.C. Reproducibility and comparison of visual acuity obtained with sightbook mobile application to near card and Snellen chart. Retina 2016, 36, 1009–1020. [Google Scholar] [CrossRef]
  24. Rhiu, S.; Lee, H.J.; Goo, Y.S.; Cho, K.; Kim, J.H. Visual acuity testing using a random method visual acuity application. Telemed. J. e-Health 2016, 22, 232–237. [Google Scholar] [CrossRef] [PubMed]
  25. Rodríguez-Vallejo, M.; Llorens-Quintana, C.; Furlan, W.D.; Monsoriu, J.A. Visual acuity and contrast sensitivity screening with a new iPad application. Displays 2016, 44, 15–20. [Google Scholar] [CrossRef]
  26. Rhiu, S.; Kim, M.; Kim, J.H.; Lee, H.J.; Lim, T.H. Korean version self-testing application for reading speed. Korean J. Ophthalmol. 2017, 31, 202–208. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Bodduluri, L.; Boon, M.Y.; Ryan, M.; Dain, S.J. Normative values for a tablet computer-based application to assess chromatic contrast sensitivity. Behav. Res. Methods 2018, 50, 673–683. [Google Scholar] [CrossRef] [Green Version]
  28. Azis, N.N.N.; Chew, F.L.M.; Rosland, S.F.; Ramlee, A.; Che-Hamzah, J. Parents’ performance using the AAPOS vision screening app to test visual acuity in Malaysian preschoolers. JAAPOS J. Am. Assoc. Pediatr. Ophthalmol. Strabismus 2019, 23, 268.e1–268.e6. [Google Scholar] [CrossRef]
  29. Brucker, J.; Bhatia, V.; Sahel, J.A.; Girmens, J.F.; Mohand-Saïd, S. Odysight: A mobile medical application designed for remote monitoring—A prospective study comparison with standard clinical eye tests. Ophthalmol. Ther. 2019, 8, 461–476. [Google Scholar] [CrossRef] [Green Version]
  30. Fernández, J.; Rodríguez-Vallejo, M.; Tauste, A.; Albarrán, C.; Basterra, I.; Piñero, D. Fast measure of visual acuity and contrast sensitivity defocus curves with an iPad application. Open Ophthalmol. J. 2019, 13, 15–22. [Google Scholar] [CrossRef]
  31. Rodríguez-Vallejo, M.; Remón, L.; Monsoriu, J.A.; Furlan, W.D. Designing a new test for contrast sensitivity function measurement with iPad. J. Optom. 2015, 8, 101–108. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Van Nes, F.L.; Bouman, M.A. Spatial modulation transfer in the human eye. J. Opt. Soc. Am. 1967, 57, 401–406. [Google Scholar] [CrossRef]
  33. Kollbaum, P.S.; Jansen, M.E.; Kollbaum, E.J.; Bullimore, M.A. Validation of an iPad test of letter contrast sensitivity. Optom. Vis. Sci. 2014, 91, 291–296. [Google Scholar] [CrossRef]
  34. Habtamu, E.; Bastawrous, A.; Bolster, N.M.; Tadesse, Z.; Callahan, E.K.; Gashaw, B.; Macleod, D.; Burton, M.J. Development and validation of a smartphone-based contrast sensitivity test. Transl. Vis. Sci. Technol. 2019, 8, 13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Johnson, C.A.; Thapa, S.; Kong, Y.X.G.; Robin, A.L. Performance of an iPad application to detect moderate and advanced visual field loss in Nepal. Am. J. Ophthalmol. 2017, 182, 147–154. [Google Scholar] [CrossRef] [PubMed]
  36. Kergoat, H.; Law, C.; Chriqui, E.; Leclerc, B.S.; Kergoat, M.J. Tool for screening visual acuity in older individuals with dementia. Am. J. Alzheimers Dis. Other Dement. 2017, 32, 96–100. [Google Scholar] [CrossRef]
  37. Vingrys, A.J.; Healey, J.K.; Liew, S.; Saharinen, V.; Tran, M.; Wu, W.; Kong, G.Y.X. Validation of a tablet as a tangent perimeter. Transl. Vis. Sci. Technol. 2016, 5, 3. [Google Scholar] [CrossRef] [PubMed]
  38. Turpin, A.; Lawson, D.J.; McKendrick, A.M. PsyPad: A platform for visual psychophysics on the iPad. J. Vis. 2014, 14, 16. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Flowchart showing the procedure followed in the current bibliographic review.
Figure 1. Flowchart showing the procedure followed in the current bibliographic review.
Vision 05 00058 g001
Table 1. Summary of the most relevant information of the articles included in the current review.
Table 1. Summary of the most relevant information of the articles included in the current review.
AuthorsSampleDeviceApps or ApplicationsMeasuresMain Results and Conclusions
Black et al. [14] 2013n = 85 participantsiPadVisual Acuity XL appVA (Bailey Lovie and HOTV chart)
With external glare source reflecting over the iPad screen, the iPad (EDTRS chart) provided VA measurements that were significantly worse at an average of 0.18 logMAR
Zhang et al. [15] 2013n = 120 participantsiPad 2Eye Chart Pro appVA (E Snellen optotype)
VA median measurements taken with iPad (Eye Chart Pro): 0.40 logMAR
Kingsnorth et al. [16] 2014n = 21 participantsiPad 32 custom-made mobile app reading speed chartsReading speed (Radner reading chart)
ORS custom-made charts: 194 ± 29 wpm and 195 ± 25 wpm; ORS Radner chart: 166 ± 20 wpm; p < 0.001. Repeatability: app charts: 0.30 ± 22.5 wpm
Norgett et al. [17] 2014n = 89 participantsiPad 2Custom-designed visual acuity testVA (Sloan and ETDRS optotypes)
No statistically significant differences between digital and conventional measurements
VA mean unflanked: 0.0 logMAR
Manzanaro et al. [18] 2015n = 46 participantsiPad2020 Duo FLEX Visual Acuity ChartVA (ETDRS optotype)
Significant differences between far VA measured with iPad (2020 Duo FLEX Visual Acuity Chart) and measured with traditional tests
4 m VA mean; iPad app: 0.093, ETDRS: 0.165; p < 0.001; 2 m VA mean; iPad app: −0.089, ETDRS: −0.049; p = 0.016
Perera et al. [19] 2015n = 88 participantsiPhone 4“Snellen” DrBloggs Ltd. appVA (6 S VA chart)
VA mean difference: 0.02 logMAR (95 % limit of agreement)
Tofigh et al. [20] 2015n = 100 participantsiPhone 5EyeHand Book appVA (app vs. Rosenbaum near optotype)
Results could be overestimated as optotypes did not present random-memory effects
VA mean: EyeHand Book app: 0.1398 logMAR SD: 0.132; Rosenbaun optotype: 0.234 logMAR; SD: 0.186
Kingsnorth et al. [21] 2016n = 20 participantsiPadAston near app and Aston distance appCS (CSV-100 and Pelli–Robson tests)
Great repeatability of results obtained with the Aston near and Aston distance apps compared to the CSV-100 test but less than the Pelli–Robson test
Pathipati et al. [22] 2016n = 64 participantsiPhonePaxos Checkup appVA (Snellen optotype vs. app)
iPhone measurement was more accurate than non-ophthalmologist healthcare personnel measurement of VA
VA logMAR average: Snellen: 0.211 ± 0.35; p = 0.00003 vs. Paxos Checkup: 0.06 ± 0.40; p = 0.264)
Phung et al. [23] 2016n = 30 participantsiPadSightBook mobile appVA near and far (ETDRS optotype vs. app)
VA measured with both methods differed significantly: they could not be used interchangeably
SightBook VA mean difference: 5.4 letters (RE) and 6.1 letters (LE); Snellen VA mean difference: 7.7 (RE) and 7.9 (LE)
Rhiu et al. [24] 2016n = 43 participantsiPadiPad-based app: Snellen chart, Tumbling Echart, Landolt C chart, and Arabic figures chartVA (Snellen, Landolt C optotypes)
Significant correlation between both methods
New method not influenced by the memory effect
Mean logMAR differences: Snellen E: −0.004, Tumbing E: −0.03 and Landolt C: 0.04)
Rodriguez-Vallejo et al. [25] 2016n = 45 participantsiPadSelf-developed app for IOSVA and CS (ETDRS optotypes)
App comparable with results of OPTHEC6000
Very practical in clinical use for screening purposes
VA mean differences: 0.06 logMAR (p < 0.001) and CS mean differences: 0.05 log units (p > 0.05)
Rhiu et al. [26] 2017n = 65 participantsiPadKorean version reading speed chartReading speed
App easy to use, providing results reliable
First app with Korean optotypes to evaluate reading speed
Mean reading speed: 202.3 ± 84.4 wpm and mean reading and speaking speed 129.7 ± 25.9 wpm; p < 0.001
de Fez et al. [2] 2018n = 407 participantsiPadOptopadColor vision
Comparable diagnostic ability of color vision anomalies compared to Farnsworth–Munsell (FM 100 H) and Ishihara plates.
Bodduluri et al. [27] 2018n = 100
participants
iPadThree self-developed gamesChromatic contrast sensitivity
Games 1 and 2 and the Cambridge Colour Test (CCT): similar absolute thresholds and tolerance intervals
Game 3: significantly lower values than games 1, 2, and the CCT, due to visual task differences
Azis et al. [28] 2019n = 195 participantsiPadAAPOS Vision screening appVA (ETDRS and Sloan optotypes vs. app)
Good correlation between app and conventional optotypes
Brucker et al. [29] 2019n = 120 participantsiPadOdysight appVA and CS (app vs. ETDRS test)
Optimal VA measurements
CS results were not as reliable
Fernández et al. [30] 2019n = 127 participantsiPadDefocus curve app (version 1.0.8)VA (E Snellen optotype) and CS (CSF test)
Digital measurements quick to do in the clinical setting
Not interchangeable with traditional ones
VA logMAR mean: app: −0.04 ± 0.09 and ETDRS: −0.05 ± 0.08; p = 0.51. CS log units mean: app: 0.83 ± 0.23 and CSF: 1.74 ± 0.019; p < 0.001
Hogarti et al. [6] 2020-iPhone45 apps (Google play store) and 23 apps (Apple store)Visual function (mainly VA)
Australian bibliographic review concluding that there is a need to carry out app validations in order to corroborate their effectiveness
Abbreviations: VA, visual acuity; CS, contrast sensitivity; ETDRS, Early Treatment Diabetic Retinopathy Study; FM, Fansworth–Munsell.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mena-Guevara, K.J.; Piñero, D.P.; de Fez, D. Validation of Digital Applications for Evaluation of Visual Parameters: A Narrative Review. Vision 2021, 5, 58. https://doi.org/10.3390/vision5040058

AMA Style

Mena-Guevara KJ, Piñero DP, de Fez D. Validation of Digital Applications for Evaluation of Visual Parameters: A Narrative Review. Vision. 2021; 5(4):58. https://doi.org/10.3390/vision5040058

Chicago/Turabian Style

Mena-Guevara, Kevin J., David P. Piñero, and Dolores de Fez. 2021. "Validation of Digital Applications for Evaluation of Visual Parameters: A Narrative Review" Vision 5, no. 4: 58. https://doi.org/10.3390/vision5040058

APA Style

Mena-Guevara, K. J., Piñero, D. P., & de Fez, D. (2021). Validation of Digital Applications for Evaluation of Visual Parameters: A Narrative Review. Vision, 5(4), 58. https://doi.org/10.3390/vision5040058

Article Metrics

Back to TopTop