Next Article in Journal
Anxiety towards Statistics and Its Relationship with Students’ Attitudes and Learning Approach
Previous Article in Journal
Surgical Outcome in Extratemporal Epilepsies Based on Multimodal Pre-Surgical Evaluation and Sequential Intraoperative Electrocorticography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparing the Visual Perception According to the Performance Using the Eye-Tracking Technology in High-Fidelity Simulation Settings

1
Department of Anesthesiology, Faculty of Medicine, Montreal, Centre d’Apprentissage des Attitudes et Habiletés Cliniques (CAAHC), Université de Montréal, Montréal, QC H1T 2M4, Canada
2
Cognitive Computing Department, Faculty of Science, University of Quebec at Montreal, Montréal, QC H3C 3P8, Canada
3
Department of Emergency Medicine, Faculty of Medicine, Université de Montréal, Montréal, QC H1T 2M4, Canada
*
Author to whom correspondence should be addressed.
Behav. Sci. 2021, 11(3), 31; https://doi.org/10.3390/bs11030031
Submission received: 7 January 2021 / Revised: 26 February 2021 / Accepted: 1 March 2021 / Published: 5 March 2021
(This article belongs to the Section Cognition)

Abstract

:
Introduction: We used eye-tracking technology to explore the visual perception of clinicians during a high-fidelity simulation scenario. We hypothesized that physicians who were able to successfully manage a critical situation would have a different visual focus compared to those who failed. Methods: A convenience sample of 18 first-year emergency medicine residents were enrolled voluntarily to participate in a high-fidelity scenario involving a patient in shock with a 3rd degree atrioventricular block. Their performance was rated as pass or fail and depended on the proper use of the pacing unit. Participants were wearing pre-calibrated eye-tracking glasses throughout the 9-min scenario and infrared (IR) markers installed in the simulator were used to define various Areas of Interest (AOI). Total View Duration (TVD) and Time to First Fixation (TFF) by the participants were recorded for each AOI and the results were used to produce heat maps. Results: Twelve residents succeeded while six failed the scenario. The TVD for the AOI containing the pacing unit was significantly shorter (median [quartile]) for those who succeeded compared to the ones who failed (42 [31–52] sec vs. 70 [61–90] sec, p = 0.0097). The TFF for the AOI containing the ECG and vital signs monitor was also shorter for the participants who succeeded than for those who failed (22 [6–28] sec vs. 30 [27–77] sec, p = 0.0182). Discussion: There seemed to be a connection between the gaze pattern of residents in a high-fidelity bradycardia simulation and their performance. The participants who succeeded looked at the monitor earlier (diagnosis). They also spent less time fixating the pacing unit, using it promptly to address the bradycardia. This study suggests that eye-tracking technology could be used to explore how visual perception, a key information-gathering element, is tied to decision-making and clinical performance.

1. Introduction

Visual perception is a complex psychological operation used to understand our environment [1]. It is linked to higher cognitive function and highly affects the decision-making processes. The eye-tracking technology aims to report on some key elements of visual perception. It is commonly used in many fields such as marketing, neuroscience and industrial engineering [2,3,4]. Such uses rely on the suggestion that an individual’s gaze pattern is influenced by many things, including its knowledge level or expertise regarding a particular situation [5]. Several studies have linked the visual gaze pattern to expertise in the medical field. Katsuru et al. used eye-tracking technology to discriminate the differences in eye gaze between medical students and experienced physicians [6]. Literature research performed by Van De Luecht et al. to examine the visual processes of radiologists during lung nodule detection concluded that the eye-tracking technology could be a useful tool to inform advancements in lung screening [7]. Schultz et al. assessed workload in an anesthesia induction scenario in high-fidelity simulation. They concluded that pupil size and heart rate were not good comparison factors between individuals. The results were not statistically significant, but there was an inverse correlation between fixation and workload [5]. Additionally, Capogna et al. established, using eye-tracking technology, the importance of the distribution of the leaders’ gaze across all members of their team and the establishment of direct eye contact [8]. The association between the eye gaze pattern and the performance was also tested in other research fields such as aviation. Di-Flumeri et al. assessed, in real-time, the vigilance level of an Air Traffic Controller by measuring the time-to-first fixation as an indicator of a high level of vigilance [9]. Consequently, there seems to be a difference in gaze depending on the workload [10], and the experience of individuals. This difference in visual perception according to expertise is now well established. It incorporates mechanisms that capture the practice-related increase in the selective use of information [11], which leads to faster and accurate diagnoses. It is also based on a global-initial-focal search pattern, followed by a detailed, focal search-to-find mode [12]. The expert decision-making processes lead him to exhibit the more visual fixations in his relevant environment areas [13]. It goes without saying that a precise and rapid cognitive process should lead to a better performance. However, the current literature does offer only little evidence linking visual perception to performance, and basically compares the performance of the expert to that of the novice while monitoring their visual perception [14,15]. To our knowledge, no research compared gaze pattern and performance in a clinical simulation in a prospective and hypothesis-based manner. This exploratory work aimed to examine the visual perception of clinicians using eye-tracking technology during a high-fidelity simulation scenario and to look for differences between participants who succeeded and those who failed, apart from their expertise, when managing the critically ill simulated patient. The rationale of our hypothesis was to show that, with similar expertise, the performance of a physician during his practice depends on his situational awareness and understanding of his clinical environment. We hypothesized that the gaze patterns of physicians who managed the scenario successfully would be different from those who failed. As part of situational awareness building, we hypothesized that the practitioner would spend more time looking at the elements of his clinical practice setting that are unknown to him, but which would be crucial for the care of the patient.

2. Methods

The protocol was registered on ClinicalTrials.gov under the identifier NCT03049098, 9 February 2017. The institutional review board approved the research (Comité d’Éthique de la Recherche en Santé de l’Université de Montréal CÉRES, December 2017, #16-082-CERES-D-1) and written informed consents were obtained from all participants. A convenience sample of 21 first-year residents in their first three weeks of training from the emergency medicine department were enrolled voluntarily to be part of this exploratory study. We excluded those for whom it was not their first residency or those who are already licensed physicians in another country. Participants had to go through a 9-min high-fidelity simulation scenario in which a patient presented unstable bradycardia that needed to be externally paced. Throughout the scenario, individuals were wearing Tobii Pro Glasses®, an eye-tracking system that aggregates gaze data (https://www.tobiipro.com/fr/produits/tobii-pro-glasses-2/) (accessed on 23 January 2018). Glasses were fit to each resident’s face and calibrated following the manufacturer’s instructions. Prescription lenses were kept for calibration and during the experiment. During the scenario, the participant was allowed to move around and take care of the simulated patient as if he was in real clinical practice. In a time-wise fashion, we performed 3 calibration attempts to reach at least 75% accuracy and tracking scores, as it is recommended by the manufacturer. To do so, the participant stood one meter away from a whiteboard (a square of 1-m side) on which four infrared sensors were hung. After wearing the glasses (over his prescription lenses if necessary), the participant had to follow, only with his eyes, without moving his head, a 5th central sensor that the investigator made move following a Z-shape from bottom to top of the board, then from right to left. No calibration was needed after the experiment. Once calibrated, the squared IR AOIs can become distorted with peripheral vision, but the distortions do not affect the recognition of fixation in that AOI. The study took place at the University of Montreal high-fidelity Simulation Centre from 14–18 July 2017.

2.1. Scenario

This study was done using the two-mannequin-model (TMM) that was developed to improve our ability to use simulation to assess the clinical competency of the residents and the attending staff to treat severe bradycardia with hemodynamic instability. A high-fidelity mannequin (METI’s Human Patient Simulator, HPS, CAE Healthcare, Montreal, Canada) (https://caehealthcare.com/patient-simulation/hps/) (accessed on 23 January, 2018) on which blood pressure, electrocardiogram, saturation, respiratory rhythm, heart rate, and the pacing was controlled provided all physical findings and vital signs except for the ECG signal while another mannequin (SimMan 3G: https://laerdal.com/ca/products/simulation-training/emergency-care-trauma/simman-3g/) (accessed on 23 January 2018), located in an adjacent room, generated the ECG signal seen by the participants, which is sent to the modified HPS using a custom-made cable [16].
Individual data were obtained from pretests and general explanations about the simulation were given to participants. The scenario consisted of an 80-year-old patient hospitalized for pneumonia. The ECG signal showed a 3rd-degree atrioventricular block, a heart rhythm disorder that causes an extreme slowdown in the heart rate and results in a hemodynamic shock. Based on a recent publication addressing this specific scenario [16] successful management was defined by the participant properly installing the cardiac pacing material and achieving electrical and mechanical pacing of pulse in 9 min or less. Based on Ahn et al. [17], in a recently published list of items, clinical competency in transcutaneous pacing requires recognizing when TCP is indicated, properly applying the multifunction electrode pads and connecting them to a functioning pacing unit, properly applying the pacemaker’s (PM) electrocardiography (ECG) leads to the patient, setting the appropriate pacing mode, pacing rate and verifying ventricular capture. All other treatments, such as atropine were ineffective or declined. Unless pacing was set correctly, the patient continued to have nausea and to be hypotensive. A debriefing followed every simulation.

2.2. Data Collection

Infrared (IR) markers were installed in the simulator and used to frame predefined Area of Interest (AOI). Four IR markers needed to be placed squarely to produce an AOI and allow data extraction. Three AOIs were created (Figure 1).
The first one featured the head of the patient and the second one, the ECG and vital signs monitor. The defibrillator/pacing unit, the key device needed for successful management of the bradycardia, was featured in the third AOI. Tobii Studio® software (Tobii Technology, Inc. Reston, VA, United States) was used to edit and aggregate gaze data and link them to each AOI. Information from IR markers was also used to build heat maps. Data collected included Total View Duration (TVD) or the time, in seconds, spent by a participant looking at an AOI during the whole simulation and Time to First Fixation (TFF) defined as the interval needed by the resident to take his first look at the AOI.

2.3. Heat Map and Gaze-Trajectories Analysis

Heatmaps relative to each AOI were extracted. A heat map uses different colors depending on the number and on the duration of fixations that participants compiled in certain areas of the screen [18]. The heat map is a function of the TVD and the number of fixations.

2.4. Statistical Methods

Our sample size was a convenient one. Data are expressed as median and 25–75% quartile. Due to the limited sample size and the non-normal distribution of the data (D’Agostino and Pearson normality test), the Mann-Whitney U test was used to search for differences in TFF and TVD between the success and fail groups. A p-value under 0.05 was considered significant. To assess, in retrospect, whether our sample size was adequate, using the t value and the degrees-of-freedom (df) of the unpaired t-test, we calculated the effect size (r) to characterize the magnitude of the difference between the groups. An r value of 0.2 was considered a “weak” effect, 0.5 “medium,” and 0.8 “strong” [19]. Statistical analysis was performed using Prism 9.0.1 for Mac OS X (GraphPad Software, La Jolla, CA, USA).

3. Results

Twenty-one residents were enrolled. They were chosen following their registration order. Amongst them, three (17%) had some previous experience managing unstable bradycardia and one (5.5%) had placed pacing pads in a clinical setting. However, before simulation, most participants felt confident (61%) that they could manage unstable bradycardia.
Three of the 21 participants were excluded; two due to technical difficulties during recording and the other because of failed calibration (10% accuracy), leaving 18 residents whose data were analyzed. Twelve participants managed the scenario successfully while six failed to do so. The TVD spent looking at the pacing unit was shorter for the residents who managed the simulation successfully compared to the ones who failed (42 [31–52] sec vs. 70 [61–90] sec, p = 0.0097). The group who succeeded also showed a shorter TFF before looking at the monitor compared with those who failed (22 [6–28] sec vs. 30 [27–77] sec, p = 0.0182) (Figure 2 and Figure 3). The effect size r was respectively 0.44 and 0.45 for the TVD of the defibrillator and the TFF of the monitor. Although it was significant, this indicates a “medium” effect between the groups.
TFF and TVD of the AOI head of the patient were similar between the groups (TFF: 9 [4–17] sec vs. 4 [3–12 sec, p = 0.2908] and TVD: 17 [5–34] sec vs. 10 [4–26] sec, p = 0.6820).
We made comparisons of the TVD between the three AOIs, among participants who passed the scenario and among participants who did not. The results showed a similar gaze distribution profile among AOIs regardless of the outcome of the scenario. All participants spent significantly more time looking at the defibrillator than at the patient’s head or the monitor. They also spent a similar amount of time looking at the monitor and the patient’s head (Table 1).

4. Discussion

4.1. Key Results

Results showed that the residents who managed the scenario successfully looked at the ECG monitor, the only place where bradycardia was shown, earlier than those who failed. Additionally, participants succeeded looked at the pacing unit for a shorter time than their colleagues who failed. To our knowledge, this is one of the few studies that collected gaze patterns in participants having roughly the same expertise and looked for a link with their performance. By cons, several studies found an association between expertise and visual perception. In radiology, Cooper et al. used eye-tracking during the showing of CT-scan slices displaying different pathologies to participants with various levels of medical training. They found differences in visual search patterns between novices and experienced radiologists [20]. Turgeon et al. compared search patterns of fourth-year dental students and licensed oral and maxillofacial radiologists that were shown normal and pathologic panoramic radiographic images. Although all participants gazed longer at normal images, experts observed the normal images for a longer time and the pathologic images for a shorter time than dental students did [21]. Wilson et al. used a laparoscopic simulator linked to eye-tracking technology to compare experienced and novice surgeons. In this experiment, participants needed to touch flashing balls with laparoscopic tools. Results showed that experts watched the tools for a shorter time and the ball for a longer time than novices did [22]. The eye-tracking technology allowed the distinctions in visual perception of the experts and novices [23,24,25,26]. Typically, novice clinicians tend to unnecessarily activate data-collection approach to validate their hypothesis before arriving at a final diagnosis while experts make appropriate assumptions more quickly based on past similar experiences and accurately perceive more precisely the relative importance of information such as case progression, vital signs, and symptoms [27]. Experts frequently exhibit ‘‘verification-driven visual perception’’; they will typically display wide and few saccades with less erratic eye movements. On the contrary, novices typically display narrow and more saccades with more erratic eye movements [28,29].
Research in the field of visual perceptions also suggests a relationship between workload increase and the decrease in fixation duration [30]. In the aviation field, Thomas et al. observed a different gaze pattern between pilots who succeeded during an off-normal flight and those who failed. Pilots who detected the problem scanned the cockpit dashboard more often, whereas the pilots who failed ignored information [31]. This is consistent with our research. Indeed, we noticed that participants who succeed the patient’s management stared at the monitor more quickly. The rapid fixation of the monitor could allow to discover the bradycardia and to initiate the management more promptly. De Rivecourt et al. showed that with increasing workload, trained pilots tend to decrease fixation duration [32]. Their results perfectly support our research, as we demonstrated that amongst untrained medical students unfamiliar with bradycardia, the ones who fixated the pacing unit longer were mostly unsuccessful.
There is also a difference in gaze patterns depending on the knowledge level of the participant [33,34] as demonstrated in radiology, nursing, and dental education [35,36,37]. Cooper et al. used eye-tracking technology in radiology where they showed CT-scan slices displaying different pathologies to participants at different levels of medical training. They have shown differences in visual search patterns mainly between novices and experienced radiologists [38]. Turgeon et al. compared search patterns of fourth-year dental students and certified oral and maxillofacial radiologists that were shown normal and pathologic panoramic radiography. All participants gazed longer at normal images [21]. Certified oral and maxillofacial radiologists observed the normal images for a longer time and the pathologic images for a shorter time than dental students. Wilson et al. [22] used a laparoscopic simulator linked to eye-tracking technology to compare experienced and novice surgeons. In this experiment, participants needed to touch flashing balls with laparoscopy instruments. Results showed that experts watched the tools for a shorter time and the ball for a longer time than novices.
Combining wearable EEG and eye-tracking makes it possible to refine the analysis of the gaze and to take one more step to integrate information processing with visual perception [39,40,41,42].

4.2. Limitations

There are several limitations to this exploratory study. An obvious limitation is its small sample size. Due to the participants’ busy schedules and our concern about the confidentiality of the scenario, we could only enroll 21 participants of which 3 were excluded because of technical problems. Despite the limited number of participants, the calculation of the effect size showed a “medium” effect, which would suggest that the number of additional participants would not be major. Our observational study will undoubtedly, as a seed study, serve to calculate the sample size of other similar ones.
Naturally, the external validity of the results remains unknown. We can only speculate whether the findings would have been different had we opted for another scenario. Additionally, the AOIs were chosen because they contained elements that were deemed to be the most relevant to the situation such as the monitor [5,43] or the mannequin’s head [44]. Nonetheless, these were arbitrary decisions, and we cannot exclude that other visual elements could have been of interest to the participants. Ending the scenario after 9 minutes was also an arbitrary decision although previous work conducted at our center [16] suggests that allowing participants who fail this type of scenario more times is unlikely to bring more understanding.

5. Conclusions

In a high-fidelity simulation scenario, there were significant differences between the environmental visual interests of the first-year residents who performed successfully and those who failed. This suggests that the eye-tracking technology could be used to look into the relationship between key situational awareness elements like visual perception and clinical performance. Other research methods such as the think-aloud protocol, the simulation with iterative discussions [45] or the debriefing on-demand need to be performed before concluding the relationship between clinical reasoning, visual perception, and clinical management.

Author Contributions

Conceptualization, I.T. and K.S.; methodology: I.T., K.S. and M.T.; investigation: I.T., M.T., K.S., R.P., M.-È.B., C.R., J.M. and M.-R.P.; resources: R.P. and K.S.; writing and original draft preparation: I.T., M.T. and R.P.; writing and review and editing I.T., P.D., A.R. and M.G. supervision and project administration: I.T. and K.S. All authors have read and agreed to the published version of the manuscript.

Funding

Support was provided solely from institutional and simulation center sources.

Institutional Review Board Statement

The protocol was registered on ClinicalTrials.gov under the identifier NCT03049098, 9 February 2017. The institutional review board approved the research (Comité d’éthique de la Recherche en Santé de l’Université de Montréal CÉRES, December 2017, #16-082-CERES-D-1). The study was conducted according to the guidelines of the Declaration of Helsinki.

Informed Consent Statement

Written informed consents were obtained from all participants.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gibson, J.J. The Perception of the Visual World; Houghton Mifflin: Boston, MA, USA, 1950. [Google Scholar]
  2. Duchowski, A.T. A breadth-first survey of eye-tracking applications. Behav. Res. Methods Instrum. Comput. 2002, 34, 455–470. [Google Scholar] [CrossRef]
  3. Valtakari, N.V.; Hooge, I.T.C.; Viktorsson, C.; Nystrom, P.; Falck-Ytter, T.; Hessels, R.S. Eye tracking in human interaction: Possibilities and limitations. Behav. Res. Methods 2021, 1–17. [Google Scholar] [CrossRef]
  4. Hussain, Z.; Simonovic, B.; Stupple, E.J.N.; Austin, M. Using Eye Tracking to Explore Facebook Use and Associations with Facebook Addiction, Mental Well-being, and Personality. Behav. Sci. 2019, 9, 19. [Google Scholar] [CrossRef] [Green Version]
  5. Schulz, C.M.; Schneider, E.; Fritz, L.; Vockeroth, J.; Hapfelmeier, A.; Wasmaier, M.; Kochs, E.F.; Schneider, G. Eye tracking for assessment of workload: A pilot study in an anaesthesia simulator environment. Br. J. Anaesth 2011, 106, 44–50. [Google Scholar] [CrossRef] [Green Version]
  6. Tatsuru, K.; Keisuke, Y.; Shun, O.; Mayu, M.; Ayaka, N.; Masakazu, M.; Koshiro, S.; Toshio, H.; Koji, Y.; Waka, Y.; et al. The evaluation of eye gaze using an eye tracking system in simulation training of real-time ultrasound-guided venipuncture. J. Vasc. Access 2021, 2021, 1129729820987362. [Google Scholar]
  7. Van De Luecht, M.R.; Reed, W.M. The cognitive and perceptual processes that affect observer performance in lung cancer detection: A scoping review. J. Med. Radiat. Sci. 2021. [Google Scholar] [CrossRef] [PubMed]
  8. Capogna, E.; Capogna, G.; Raccis, D.; Salvi, F.; Velardo, M.; Del Vecchio, A. Eye tracking metrics and leader’s behavioral performance during a post-partum hemorrhage high-fidelity simulated scenario. Adv. Simul. 2021, 6, 4. [Google Scholar] [CrossRef] [PubMed]
  9. Di Flumeri, G.; De Crescenzio, F.; Berberian, B.; Ohneiser, O.; Kramer, J.; Arico, P.; Borghini, G.; Babiloni, F.; Bagassi, S.; Piastra, S. Brain-Computer Interface-Based Adaptive Automation to Prevent Out-Of-The-Loop Phenomenon in Air Traffic Controllers Dealing With Highly Automated Systems. Front. Hum. Neurosci. 2019, 13, 296. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Castner, N.; Appel, T.; Eder, T.; Richter, J.; Scheiter, K.; Keutel, C.; Huttig, F.; Duchowski, A.; Kasneci, E. Pupil diameter differentiates expertise in dental radiography visual search. PLoS ONE 2020, 15, e0223941. [Google Scholar]
  11. Haider, H.; Frensch, P. Eye Movement During Skill Acquisition: More Evidence for the Information-Reduction Hypothesis. J. Exp. Psychol. 1999, 25, 172–190. [Google Scholar] [CrossRef]
  12. Van der Gijp, A.; Ravesloot, C.J.; Jarodzka, H.; van der Schaaf, M.F.; van der Schaaf, I.C.; van Schaik, J.P.J.; Ten Cate, T.J. How visual search relates to visual diagnostic performance: A narrative systematic review of eye-tracking research in radiology. Adv. Health Sci. Educ. Theory Pract. 2017, 22, 765–787. [Google Scholar] [CrossRef] [Green Version]
  13. Eivazi, S.; Hafez, A.; Fuhl, W.; Afkari, H.; Kasneci, E.; Lehecka, M.; Bednarik, R. Optimal eye movement strategies: A comparison of neurosurgeons gaze patterns when using a surgical microscope. Acta Neurochir. 2017, 159, 959–966. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Gegenfurtner, A.; Lehtinen, E.; Säljö, R. Expertise Differences in the Comprehension of Visualizations: A Meta-Analysis of Eye-Tracking Research in Professional Domains. Educ. Psychol. Rev. 2011, 23, 523–552. [Google Scholar] [CrossRef]
  15. Tien, T.; Pucher, P.H.; Sodergren, M.H.; Sriskandarajah, K.; Yang, G.Z.; Darzi, A. Differences in gaze behaviour of expert and junior surgeons performing open inguinal hernia repair. Surg. Endosc. 2015, 29, 405–413. [Google Scholar] [CrossRef]
  16. Robitaille, A.; Perron, R.; Germain, J.F.; Tanoubi, I.; Georgescu, M. High-fidelity simulation of transcutaneous cardiac pacing: Characteristics and limitations of available high-fidelity simulators, and description of an alternative two-mannequin model. Simul. Healthc. 2015, 10, 122–127. [Google Scholar] [CrossRef]
  17. Ahn, J.; Kharasch, M.; Aronwald, R.; Menon, S.; Du, H.; Calabrese, N.; Gitelis, M.; Aitchson, P.; Wang, E. Assessing the accreditation council for graduate medical education requirement for temporary cardiac pacing procedural competency through simulation. Simul. Healthc. 2013, 8, 78–83. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, X.; Ley, A.; Koch, S.; Hays, J.; Holmqvist, K.; Alexa, M. Computational discrimination between natural images based on gaze during mental imagery. Sci. Rep. 2020, 10, 13035. [Google Scholar] [CrossRef] [PubMed]
  19. Effect Size Calculators. Available online: https://lbecker.uccs.edu (accessed on 15 January 2018).
  20. Cooper, L.; Gale, L.; Darker, L.; Toms, A.; Saada, J. Radiology image perception and observer performance: How does expertise and clinical information alter interpretation? Stroke detection explored through eye-tracking. In Proceedings of the Medical Imaging 2009: Image Perception, Observer Performance, and Technology Assessment 2009, Lake Buena Vista (Orlando Area), FL, USA, 7–12 February 2009; Volume 7263. [Google Scholar]
  21. Turgeon, D.P.; Lam, E.W. Influence of Experience and Training on Dental Students’ Examination Performance Regarding Panoramic Images. J. Dent. Educ. 2016, 80, 156–164. [Google Scholar] [CrossRef] [PubMed]
  22. Wilson, M.; McGrath, J.; Vine, S.; Brewer, J.; Defriend, D.; Masters, R. Psychomotor control in a virtual laparoscopic surgery training environment: Gaze control parameters differentiate novices from experts. Surg. Endosc. 2010, 24, 2458–2464. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Capogna, E.; Salvi, F.; Delvino, L.; Di Giacinto, A.; Velardo, M. Novice and Expert Anesthesiologists’ Eye-Tracking Metrics During Simulated Epidural Block: A Preliminary, Brief Observational Report. Local Reg. Anesth. 2020, 13, 105–109. [Google Scholar] [CrossRef]
  24. Shojima, M.; Okamoto, Y.; Niizuma, K.; Ohta, M.; Ishikawa, O.; Fujisawa, A.; Tsukihara, H.; Sakai, N.; Tominaga, T. Preliminary study of eye tracking to investigate the differences in gaze behaviors depending on the experience of neuroendovascular therapy. Surg. Neurol. Int. 2020, 11, 351. [Google Scholar] [CrossRef] [PubMed]
  25. Sahraian, S.; Yousem, D.; Beheshtian, E.; Jalilianhasanpour, R.; Morales, R.E.; Krupinski, E.A.; Zhan, H. Improving Radiology Trainees’ Perception Using Where’s Waldo? Acad. Radiol. 2020. [Google Scholar] [CrossRef]
  26. Emhardt, S.N.; Kok, E.M.; Jarodzka, H.; Brand-Gruwel, S.; Drumm, C.; van Gog, T. How Experts Adapt Their Gaze Behavior When Modeling a Task to Novices. Cognit. Sci. 2020, 44, e12893. [Google Scholar] [CrossRef]
  27. Tourangeau, M.; Tanoubi, I.; Perron, R.; Expert, R.T.M.S.T.; Belanger, M.E.; Sodoke, K.; Drolet, P.; Robitaille, A.; Georgescu, L.M. Comparing the visual perception of novice and expert clinicians: An exploratory simulation-based study. Anaesth. Crit. Care Pain Med. 2020, 39, 247–249. [Google Scholar] [CrossRef] [PubMed]
  28. Dreyfus, H.; Dreyfus, H.; Athanasiou, T. Mind over Machine; The Free Press: New York, NY, USA, 1986. [Google Scholar]
  29. Arocha, J.F.; Patel, V.L.; Patel, Y.C. Hypothesis generation and the coordination of theory and evidence in novice diagnostic reasoning. Med. Decis. Mak. 1993, 13, 198–211. [Google Scholar] [CrossRef] [PubMed]
  30. Wu, W.; Hall, A.K.; Braund, H.; Bell, C.R.; Szulewski, A. The Development of Visual Expertise in ECG Interpretation: An Eye-Tracking: Augmented Re Situ Interview Approach. Teach. Learn. Med. 2020, 2020, 1–20. [Google Scholar] [CrossRef]
  31. Thomas, L.; Wickens, C. Eye-tracking and individual differences in off-normal event detection with flying with a synthetic vision system display. In Proceedings of the Human Factors and Ergonomics Society 48th Annual Meeting, New Orleans, LA, USA, 20–24 September 2004; pp. 223–228. [Google Scholar]
  32. De Rivecourt, M.; Kuperus, M.N.; Post, W.J.; Mulder, L.J. Cardiovascular and eye activity measures as indices for momentary changes in mental effort during simulated flight. Ergonomics 2008, 51, 1295–1319. [Google Scholar] [CrossRef] [Green Version]
  33. Thompson, V.A. Eye-tracking IQ: Cognitive capacity and strategy use on a ratio-bias task. Cognition 2020, 208, 104523. [Google Scholar] [CrossRef] [PubMed]
  34. Liao, D.; Ishii, M.; Ishii, L.E. Where Do We Look? Assessing Gaze Patterns in Cosmetic Face-Lift Surgery with Eye Tracking Technology. Plast. Reconstr. Surg. 2020, 146, 820e–821e. [Google Scholar] [CrossRef]
  35. Hofmaenner, D.A.; Herling, A.; Klinzing, S.; Wegner, S.; Lohmeyer, Q.; Schuepbach, R.A.; Buehler, P.K. Use of eye tracking in analyzing distribution of visual attention among critical care nurses in daily professional life: An observational study. J. Clin. Monit. Comput. 2020, 1–8. [Google Scholar] [CrossRef]
  36. Richter, J.; Scheiter, K.; Eder, T.F.; Huettig, F.; Keutel, C. How massed practice improves visual expertise in reading panoramic radiographs in dental students: An eye tracking study. PLoS ONE 2020, 15, e0243060. [Google Scholar] [CrossRef]
  37. Zhang, T.; Yang, J.; Liang, N.; Pitts, B.J.; Prakah-Asante, K.O.; Curry, R.; Duerstock, B.S.; Wachs, J.P.; Yu, D. Physiological Measurements of Situation Awareness: A Systematic Review. Hum. Factors 2020, 2020, 18720820969071. [Google Scholar]
  38. Auffermann, W.F.; Drew, T.; Krupinski, E.A. Special Section Guest Editorial: Medical Image Perception and Observer Performance. J. Med. Imaging (Bellingham) 2020, 7, 022401. [Google Scholar] [CrossRef] [Green Version]
  39. Di Flumeri, G.; Arico, P.; Borghini, G.; Sciaraffa, N.; Di Florio, A.; Babiloni, F. The Dry Revolution: Evaluation of Three Different EEG Dry Electrode Types in Terms of Signal Spectral Features, Mental States Classification and Usability. Sensors 2019, 19, 1365. [Google Scholar] [CrossRef] [Green Version]
  40. Dimigen, O.; Sommer, W.; Hohlfeld, A.; Jacobs, A.M.; Kliegl, R. Coregistration of eye movements and EEG in natural reading: Analyses and review. J. Exp. Psychol. Gen. 2011, 140, 552–572. [Google Scholar] [CrossRef] [Green Version]
  41. Kamienkowski, J.E.; Ison, M.J.; Quiroga, R.Q.; Sigman, M. Fixation-related potentials in visual search: A combined EEG and eye tracking study. J. Vis. 2012, 12, 4. [Google Scholar] [CrossRef] [PubMed]
  42. Borghini, G.; Arico, P.; Di Flumeri, G.; Sciaraffa, N.; Colosimo, A.; Herrero, M.T.; Bezerianos, A.; Thakor, N.V.; Babiloni, F. A New Perspective for the Training Assessment: Machine Learning-Based Neurometric for Augmented User’s Evaluation. Front. Neurosci. 2017, 11, 325. [Google Scholar] [CrossRef]
  43. Seagull, J.; Xiao, Y.; MacKenzie, C.; Jaberi, M.; Dutton, R. Monitoring Behavior: A Pilot Study Using an Ambulatory Eye-Tracker in Surgical Operating Rooms. In Proceedings of the Human Factors and Ergonomics Society 43rd Annual Meeting, Houston, TX, USA, 27 September–1 October 1999; Volume 43, pp. 850–854. [Google Scholar]
  44. Browning, M.; Cooper, S.; Cant, R.; Sparkes, L.; Bogossian, F.; Williams, B.; O’Meara, P.; Ross, L.; Munro, G.; Black, B. The use and limits of eye-tracking in high-fidelity clinical scenarios: A pilot study. Int. Emerg. Nurs. 2016, 25, 43–47. [Google Scholar] [CrossRef] [PubMed]
  45. Pennaforte, T.; Moussa, A.; Loye, N.; Charlin, B.; Audetat, M.C. Exploring a New Simulation Approach to Improve Clinical Reasoning Teaching and Assessment: Randomized Trial Protocol. JMIR Res. Protoc. 2016, 5, e26. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Infrared markers (red oval) framing three different Areas of interest (AOI): (A) Head of the patient, (B) Monitor, (C) Defibrillator/pacing unit.
Figure 1. Infrared markers (red oval) framing three different Areas of interest (AOI): (A) Head of the patient, (B) Monitor, (C) Defibrillator/pacing unit.
Behavsci 11 00031 g001
Figure 2. The data showed that residents who succeeded had a significantly shorter median [IQR] TVD at the pacing unit than the ones who failed (42 [31–52] sec vs. 70 [61–90] sec, p = 0.0097). The residents who succeeded also had a shorter median [IQR] TFF at the monitor compared with those who failed (28 [6–48] sec vs. 30 [25–77] sec, p = 0.5532). The boxplots show the median and IQR.
Figure 2. The data showed that residents who succeeded had a significantly shorter median [IQR] TVD at the pacing unit than the ones who failed (42 [31–52] sec vs. 70 [61–90] sec, p = 0.0097). The residents who succeeded also had a shorter median [IQR] TFF at the monitor compared with those who failed (28 [6–48] sec vs. 30 [25–77] sec, p = 0.5532). The boxplots show the median and IQR.
Behavsci 11 00031 g002
Figure 3. A heat map showing that the pacing unit exhibits fewer fixations (green colored) for the residents who succeeded than for those who failed (red-colored) A: Defibrillator/Pacing Unit B: Infrared markers.
Figure 3. A heat map showing that the pacing unit exhibits fewer fixations (green colored) for the residents who succeeded than for those who failed (red-colored) A: Defibrillator/Pacing Unit B: Infrared markers.
Behavsci 11 00031 g003
Table 1. Comparisons of the TVD between the three AOIs, among participants who passed the scenario and among participants who did not. Kruskal-Wallis comparison test with Dunn’s multiple comparisons test. AOI: Area of Interest.
Table 1. Comparisons of the TVD between the three AOIs, among participants who passed the scenario and among participants who did not. Kruskal-Wallis comparison test with Dunn’s multiple comparisons test. AOI: Area of Interest.
Compared AOIsMean First AOIMean Second AOIMean Diff.Adjusted p-Value
Participants who passed the scenario
Monitor vs. Defibrillator (sec)11.0028.42−17.420.0002
Monitor vs. Head of the patient (sec)11.0016.08−5.080.7118
Defibrillator vs. Head of the patient (sec)28.4216.0812.330.0124
Participants who did not pass the scenario
Monitor vs. Defibrillator (sec)5.3315.50−10.170.0029
Monitor vs. Head of the patient (sec)5.337.66−2.33>0.9999
Defibrillator vs. Head of the patient (sec)15.507.667.830.0331
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tanoubi, I.; Tourangeau, M.; Sodoké, K.; Perron, R.; Drolet, P.; Bélanger, M.-È.; Morris, J.; Ranger, C.; Paradis, M.-R.; Robitaille, A.; et al. Comparing the Visual Perception According to the Performance Using the Eye-Tracking Technology in High-Fidelity Simulation Settings. Behav. Sci. 2021, 11, 31. https://doi.org/10.3390/bs11030031

AMA Style

Tanoubi I, Tourangeau M, Sodoké K, Perron R, Drolet P, Bélanger M-È, Morris J, Ranger C, Paradis M-R, Robitaille A, et al. Comparing the Visual Perception According to the Performance Using the Eye-Tracking Technology in High-Fidelity Simulation Settings. Behavioral Sciences. 2021; 11(3):31. https://doi.org/10.3390/bs11030031

Chicago/Turabian Style

Tanoubi, Issam, Mathieu Tourangeau, Komi Sodoké, Roger Perron, Pierre Drolet, Marie-Ève Bélanger, Judy Morris, Caroline Ranger, Marie-Rose Paradis, Arnaud Robitaille, and et al. 2021. "Comparing the Visual Perception According to the Performance Using the Eye-Tracking Technology in High-Fidelity Simulation Settings" Behavioral Sciences 11, no. 3: 31. https://doi.org/10.3390/bs11030031

APA Style

Tanoubi, I., Tourangeau, M., Sodoké, K., Perron, R., Drolet, P., Bélanger, M. -È., Morris, J., Ranger, C., Paradis, M. -R., Robitaille, A., & Georgescu, M. (2021). Comparing the Visual Perception According to the Performance Using the Eye-Tracking Technology in High-Fidelity Simulation Settings. Behavioral Sciences, 11(3), 31. https://doi.org/10.3390/bs11030031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop