Next Article in Journal
Eye Tracking in Optometry: A Systematic Review
Previous Article in Journal
The Influence of Eye Model Parameter Variations on Simulated Eye-Tracking Outcomes
 
 
Journal of Eye Movement Research is published by MDPI from Volume 18 Issue 1 (2025). Previous articles were published by another publisher in Open Access under a CC-BY (or CC-BY-NC-ND) licence, and they are hosted by MDPI on mdpi.com as a courtesy and upon agreement with Bern Open Publishing (BOP).
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Eye Tracking Based Framework for Safety Improvement of Offshore Operations

by
Muhammad A. Raza
1,
Raj Kiran
2,
Saima Ghazal
3,
Ziho Kang
4,
Saeed Salehi
2,
Edward Cokely
5 and
Jiwon Jeon
4
1
University of Engineering & Technology, Lahore, Pakistan
2
School of Petroleum and Geological Eng., University of Oklahoma, USA
3
Inst. of Applied Psychology, University of the Punjab, Lahore, Pakistan
4
Department of Industrial & Systems Eng., University of Oklahoma, USA
5
Department of Psychology, University of Oklahoma, USA
J. Eye Mov. Res. 2023, 16(3), 1-17; https://doi.org/10.16910/jemr.16.3.2
Submission received: 28 December 2022 / Published: 10 August 2023

Abstract

:
Offshore drilling operations consist of complex and high-risk processes. Lack of situational awareness in drilling operations has become an important human factor issue that causes safety accidents. Prolonged work shifts and fatigue are some of the crucial issues that impact performance. Eye tracking technology can be used to distinguish the degree of awareness or alertness of participants that might be related to fatigue or onsite distractions. Oculomotor activity can be used to obtain visual cues that can quantify the drilling operators’ situational awareness that might enable us to develop warning alarms to alert the driller. Such systems can help reduce accidents and save non-productive time. In this paper, eye movement characteristics were investigated to differentiate the situational awareness between a representative expert and a group of novices using a scenario-based Virtual Reality Drilling Simulator. Significant visual oculomotor activity differences were identified between the expert and the novices that indicate an eye-tracking based system can detect the distraction and alertness exhibited by the workers. Results show promise on developing a framework which implements a real-time eye tracking technology in various drilling operations at drilling rigs and Real Time Operation Centers to improve process safety.

Introduction

Offshore energy extraction processes are overly complex systems that involve risky operations. Real-time data monitoring is a unique and highly discussed concept which has caught the imagination of every industry and academia in recent days. Dynamic data monitoring and feedback systems have taken a leap with the technological advancements in cloud computing and data mining techniques. In a typical application, the real-time system receives the onsite data from sensors, stores it in a database, processes the data using sophisticated algorithms, and finally provides the output in the form of alerts or feedback to the users. These feedback systems are now being used in many industries such as meteorology, hydrology, and the health sector. The real-time monitoring systems in the petroleum industry have also been consistently on the upsurge. The concept was laid back in the 1980s, but due to lack of technological capabilities and wide acceptance, it was not at the focal point of industrial operation. The breakthroughs in real-time data technology have allowed the improvement in the quality of remote monitoring infrastructure and now several remote centers such as Drilling Command and Control System, Drilling Data Center, Real Time Operations Center (RTOC) have been established to improve the safety and efficiency of the onsite operations (Booth & Hebert, 1989; Foreman, 1985; Ursem, Williams, Pellerin, & Kaminski, 2003). In the early days, the focus was to monitor the operation and simultaneously provide expert assistance to onsite personnel. However, with time the application has spread further and included informed decision making by remote monitoring and assessment.
Despite the success of existing approaches, current efforts only focus on the feedback system for the data obtained, especially in the petroleum industry. However, if there is a human system interface, it becomes significantly important to obtain an idea about the human performance and establish a feedback system to caution the operators’ dealing if it is deviating from the standard procedures.
The prospect of real-time data monitoring relies on the performance of the involved personnel. The success of such a system in a complex and highly interactive system like oil and gas relies on the dynamic mental and physical performance of the personal and several related human factors. The behavior is directly implicated by the situation prevailing on the site as well as the personal or social circumstances. The in-situ distractions during real-time data monitoring, especially in the well logs and well operations are eminent at the rig site.
In this context, human errors become the focal point of every accident. Internal data review and analysis by Bureau of Safety and Environmental Enforcement (BSEE) suggest that human engineering problems (e.g., humanmachine interface, poor working environments, system complexity, and non-fault-tolerant systems) and problems in work direction (e.g., poor planning, site preparation, selection of workers, and supervision) are the primary drivers of the rate of injuries (BSEE, 2016). The injury rates for the calendar year 2018 are depicted in Figure 1 (BSEE, 2018). These factors need to be addressed through industry vigilance and strong regulatory oversight to avoid continuous injuries or accidents. With regards to this, feedback systems are highly recommended to supplement human performance.
The underlying human-centered problems during the real-time operation can be eliminated by empowering the system element with capability to alert in case of near-miss environments. The eye tracking system has shown huge promise in several sectors and has become readily available and affordable in the form of portable devices. The oculomotor activity of the participant/onsite personnel can be registered in real-time. This will make the use of this device in the realm of human interactions with the system as well as other fellow professionals in virtual reality setting possible. Analysis of the real-time ocular movement data can hugely complement the human system interface performance. In high-risk environments, the alert system and dynamic feedback system will empower the struggling professionals or trainees to discharge their duty efficiently and accurately.
This paper attempts to bring forth a framework to implement such technology in the oil and gas operation. Before delving into the framework details, it is important to understand the nuances of operations and human factors depicted under process safety and situational awareness.

Process Safety

Offshore energy extraction operations are complex in nature and consist of several interlinked complex processes. The offshore operations include a complex environment composed of a diverse workforce with different backgrounds. To prevent and minimize accidents, it becomes vital to select and implement a holistic process safety system in the daily drilling operation. Real-time feedback on the crew response to an ongoing activity can help to minimize potential damage. Detecting a deviation of crew attention as early as possible can help the crew to focus on implementing well control procedures before it escalates into a catastrophic event. Review of loss of well control (LOWC) incidents and blowouts while drilling and tripping showed common systemic risks in offshore drilling especially in deep water environments where the late response can escalate to deadly blowouts. Recent ExproSoft report to Board of Safety and Environmental Enforcement (BSEE), summarizing the study of LOWC in the Gulf of Mexico, revealed that the frequency of LOWC incidents is more than that of the workover/intervention operations (Holland, 2017). A robust process safety system can be implemented by integrating both technical and non-technical skills (NTSs). Technical skills are addressed on a regular basis by industry and academicians in depth.
However, the NTSs have only caught the attention of researchers and industry professionals significantly after the catastrophic incident of Deepwater Horizon. The NTSs including communication, situational awareness, risk perception, teamwork, and decision making, play a vital role to avoid the breakdown in the operation process. For example, workers need to communicate with people from different companies as well as their supervisors and management team. Often the professionals make quick decisions especially on LOWC accounts where any loss of attention span increases risks and may cause disasters. No advancement in technological aspects of well control, drilling tools, and machinery can replace a timely expert decision-making process during an event of critical LOWC and blowout. Communication barriers between the operators, contractors and service company’s personnel increase the odds of dangerous and costly decisions. The offshore workforce mix is diverse, and their communication styles differ extensively. Many technical and non-technical issues can bias a person’s situation awareness (J. L. Thorogood, 2015). This can be a challenge to the necessary process safety system which can break down due to lack of understanding.
Human factors play a prominent role in the performance and operational non-productive time (NPT) in drilling operations (J. Thorogood, Crichton, & Bahamondes, 2014). J. Thorogood and Crichton (2014) suggested that important characteristics of high-reliability organizations’ achievements in understanding non-technical aspects from other industries, can be applied to the oil and gas industry. Crew resource management was the type of training adopted from Aerospace industry in late 1970’s by looking at Pilot’s non-technical incidents and the role it plays in airline incidents (O'Connor & Flin, 2003). Recently, crew resource management concepts have frequently been applied in the oil and gas industry. However, the effectiveness has not been established full-fledged. These limitations lead to the problem of quantifying these aspects of NTSs objectively.
Fletcher, McGeorge, Flin, Glavin, and Maran (2002) suggested that a comparative study of expert and novices’ behavior can provide a significant impetus to understand non-technical skills. Situational awareness of experts and novices differs in terms of their ability to perceive and interpret the information available in a given situation. Novices, because of their lesser experience and domain knowledge, may exhibit limited ability to detect and interpret relevant cues available in the environment and hence show lower SA. In contrast, the more experience and domain knowledge enables experts to assess situations and make decisions quickly and accurately. Naqvi, Raza, Ybarra, Salehi, and Teodoriu (2018) noticed some important differences between the psychological mind states in communication between expert and novice workers. The comparative study relies on the objective measurement of any process safety aspect. Eye tracking technology has a great opportunity in this arena since it can provide quantitative cues on cognitive responses of the participants. These cognitive responses help in understanding the situational awareness of the participants.

Situational Awareness

Human factors have been attributed to the primary cause of major accidents in high-reliability organizations. Situation or situational awareness (SA) is an important aspect of human factors. Situational awareness is defined as the sum of operator perception and comprehension of associated information, and the ability to effectively project future states of the system using it (Kaber & Endsley, 2004). One of the most popular frameworks for situational awareness is developed by Mica R. Endsley (1995) which comprised of three levels: perception, comprehension, and projection. The first level perception includes scanning, gathering the information, and detecting the underlying cues about the prevailing situation in the spatial and temporal domain. The next level, comprehension, consists of the understanding of gathered information and connecting the dots to create mental models. Finally, the third level includes the projection of future circumstances which accompanies the ability to make the final decisions.
The process control system including offshore operation share persuasion of multiple goals by the operators simultaneously, multi-tasking competing for the attention, and performance under severely stressful environment. Failure in the exhibition of these characteristics can be due to difficulties of the acting operators’ SA. Poor situational awareness can slow down the process of problem identification, diagnosis, and reaching a solution which might result in an untimely response (Mica R Endsley, 2016). Hence, the failure at the level of human performance can be attributed to inadequate situational awareness. In this context, the major challenge is to develop tools that can measure the situational awareness and cognitive state of the participant in real-time.
Several subjective approaches including memory probe techniques and questionnaire analysis had been adopted to measure the responses in the past and some of the important techniques include Situation Awareness Global Assessment Technique (SAGAT) (Mica R Endsley, 1988) and Situation Awareness Rating Technique (SART) (Taylor, 1990). Raza et al. (2019) proposed a simulation-based training framework for offshore drilling operations. However, the use of subjective measures is insufficient and will be impossible to connect with RTOC. RTOC can only rely on objective measures. One of the promising technologies regarding objective measurement of situational awareness is eye tracking.

Eye Tracking

Eye tracking method has been widely used in several industries in the commercial and academic settings to assess the movement of eye activity and patterns of visual search (Helbren et al., 2014). These visual search patterns associated with static and dynamic images have been deemed useful in understanding human behavior, thinking, and decision-making process of the observers. The visual activity is measured using the commercially available eye trackers such as the products sold by Tobii, IMotions, and SR-Research.
The eye trackers use Pupil Center Corneal Reflection technique in which an infrared light source is directed towards the eye and reflections from eyes are recorded by a camera. A specialized camera collects these reflection patterns with added visual information. Then Image processing algorithms are used to generate features such as eyes and corneal reflection patterns. Finally, the gaze point on the screen is calculated using the spatial position of each eyeball (Tobii, 2018). There are usually two types of eye tracker available: static monitors, and dynamic eye tracking glasses.
These eye tracking systems provide different types of data including eye fixation duration, fixation count, fixation rate, scan paths, and saccade duration of individuals or group of individuals in a fixed or dynamic area of interest (AOI). An AOI on a static screen-type monitor for eyetracking is represented in Figure 2, where the center of a circles indicates the eye fixation location and the numbers inside circles indicate the time order of the eye fixations. The size of the circles indicates the eye fixation durations. These circles are connected through straight lines. The straight lines indicate the fast saccades among the eye fixations. The sequence of the eye fixation data constitutes the scan paths which indicate the movement of eye fixations of the participants from one position to the other. These scan paths are useful to investigate the visual search behaviors of the participants. The number of fixations per second is accounted as fixation rate. The saccade duration is a transition time between successive fixations. This data can be further used to establish the cognitive behavioral pattern and identify the objective measures of situation awareness of the participants (Mica R Endsley, 2016). In addition, metrics-based systems can also be implemented to compare the performance of different participants. Atik, (2019) and Atik & Arslan, (2020) explored the use and effectiveness of eye tracking technology to assess situational awareness in two different maritime training contexts. In their studies expert participants demonstrated higher situational awareness and more efficient and effective use of electronic navigation aids than novices, as evidenced by their longer fixation durations on relevant AOIs and shorter fixation durations on these aids compared to novices.
Another important aspect of eye tracking data analysis is based on pupillometry, which involves the analysis of the pupil diameter data of both the eyes of the participants. Pupil size is considered as another robust measure of cognitive load and dynamic measure of the processing by the participants while tracking the real-time data. Several studies suggest that pupil size is directly related to stimuli and brain activation distinct with the level of cognitive processing (Beatty & Wagoner, 1978). Pupillary dilation has been used to assess the cognitive load (Klingner, Tversky, & Hanrahan, 2011). Pupil dilation measures have also been used in the medical industry to track the memory load and cognitive response of patients (Fotiou et al., 2009). Satterthwaite et al. (2007) implemented pupillometry to investigate human decision making. Based on the evidence, we propose that pupillogram may be used to assess participants’ cognitive state during oil well drilling operation, and that may further be used to design a feedback system to enhance process safety in high-risk offshore environments.

Industrial Applications of Eye Tracking

Eye tracking technology has been used to understand the mental processes in several industries. Study based on flight simulation validated that the dispersion of eye fixation can act as an indicator of mental workload (Camilli, Terenzi, & Di Nocera, 2007). The frequency of more extended fixation is shown to be directly proportional to the workload (Van Orden, Limbert, Makeig, & Jung, 2001). Research also showed that the participants’ ocular behavior becomes structured in higher difficulty level workload while it stays random in lower difficulty level (Dick, 1980; Simon, Rousseau, & Angue, 1993). By analyzing the visual scanning behaviors during an incident, it is possible to identify when and how human factor issues begin to emerge, whether specific visual search behaviors affected their reactions to the incident, and whether crucial data were overlooked.
One of the important avenues in which eye tracking has been widely implemented is air traffic management systems. The analysis of the participants’ ocular activity is used to get insight into the cognitive processes in the flight simulation environment. Several experiments have been conducted involving the scanning of aircraft, identifying and solving the conflicts in the spatial framework (Metzger & Parasuraman, 2005). The simulated experiments aid the understanding of underlying cognitive processes of controllers and pilots. These understandings led to concepts that can envisage the decision-making strategies between controllers and pilots and pave the path for automated decision aids. Kang and Landry (2014) developed an algorithm to interrogate dynamic multiple element objects in the air traffic control display to address the issue of eye fixation mapping in the overlapping AOIs. These techniques can help in automation of eye tracking data processing and develop visualization techniques. Eye tracking has shown to have significant potential in training and selection of pilots (Babu et al., 2019; Robinski & Stein, 2013; Vlačić, Knežević, Mandal, Rođenkov, & Vitsas, 2020).
In past decades, eye tracking analysis has also established its application in weather forecasting systems. These studies helped in understanding the underlying cognitive processes of the weather forecasters while detecting the hazards and can be used to improve the situation awareness in the future participants. The reader is encouraged to visit the website for Human Factors & Simulation laboratory at the University of Oklahoma (https://humanfactors.oucreate.com/research.html) for details of various eye-tracking applications in different industries.
Other avenues where eye tracking analyses have been popular are the medical industry. Surgery is a complicated process in the medical profession where effective teamwork becomes the yardsticks for success. The ocular movement analysis of the personnel in the surgery room has been implemented to understand the complex cognitive processes and their complexity during surgery (Topalli & Cagiltay, 2018; Wilson et al., 2011). Besides that, eye tracking analyses have also been implemented recently in Human Factors course at the University of Oklahoma, where the real-time scenarios were simulated for well control training (Dadmohammadi et al., 2017).
The use of eye tracking for performance enhancement is increasing across different sectors. Eye tracking has long been utilized in driving and navigation training simulators. (Groner & Kasneci, 2021). Recently, there is evidence of using eye-tracking for expertise assessment and expertnovice differences in visual arts, poetry, and music (Fokin, Blohm, & Riekhakainen, 2022; Kolodziej, Majkowski, Francuz, Rak, & Augustynowicz, 2018; Perra, Latimier, Poulin-Charronnat, Baccino, & Drai-Zerbib, 2022).

Application of Eye Tracking System from Offshore Rigs to Real Time Operation Centers

Oil and gas drilling process includes highly sophisticated and physically and mentally exhausting operations. Simultaneous running of several procedures and tasks makes it even more challenging. In a typical offshore drilling cabin where the driller continually monitors different gauges and must process huge amounts of information. Most of the routine operations require the driller to be alert of changes in the indicators (such as torque, hook loads), operational parameters (such as weight on bit (WOB), revolutions per minute (RPM)), and more importantly primary kick indicators (such as pit levels and flow). The repetitive and similar set of such data induces boredom in the operators and can act as distracting agent. There is a greater likelihood of an error in such intensive environments. The role of situational awareness becomes crucial for the success of a daily drilling operation (Bullemer & Reising, 2013).
Although alarms are currently employed on offshore rigs to alert drilling crew for potential hazards and they are an essential operator support system for managing abnormal situations, however, they are not very efficient. The status of alarming systems reveals that crews are inundated with several different/types of alarms. Therefore, it is challenging to prioritize them in a way that is taken seriously. The Deepwater Horizon oil spill incident provides an excellent example of an application related to alarm deficiencies. The alarm system guidelines suggest lesser than one alarm per minute as very likely to be unacceptable and one alarm per ten minutes as very likely to be acceptable (Norwegian Petroleum Directorate, 2001). In a typical drilling operation, as presented in Figure 3(a), the drillers are inundated with around 26,000 alarms over a week (Goetz, 2016), and most of them can be ignored. This is highly entangled with the human capability to interact with alarm systems and represents a great human-factors related challenge in control rooms (Bjerkebaek & Eskedal, 2004). Ikuma, Harvey, Taylor, and Handal (2014) suggested a guideline for assessing the performance of petrochemical control room operators. Kodappully, Srinivasan, and Srinivasan (2016) used eye fixations to gain insights into the cognitive behavior of control room operators that could be used to proactively tackle human errors.
Figure 3b shows a case study of loss of well control (LOWC) incident from South Timbalier area of Louisiana happened in 2013 (BSEE, 2015). A panel investigation concluded that one of the causes of 2013 LOWC and fire in South Timbalier area was: “Failure of the Rig-floor Crew to recognize the loss of well control in a timely manner made it impossible to follow the well control procedures which called for stabbing the safety valve on top of the work string as an initial step.” “The initiation of the emergency procedure sequence to activate blowout preventer (BOP) elements was delayed because of the Rigfloor Crew’s failure to recognize the LOWC at an early stage.” (BSEE, 2015). This case clearly reveals that human errors and situation awareness/alertness are crucial in such high-risk operations. Recently, digital data has become an integral part of the petroleum industry. With the advancement of sensors, an enormous amount of quality data is generated and processed within a fraction of a second. In such an environment, eye tracking can be a useful tool to improve the efficiency and accuracy of such an operation. In case the driller is tired and loses the attention span from the monitors, warning alarms can be implemented using real-time eye tracking to alert the driller.

Application for Real Time Operation Center

Remote real-time monitoring is an interesting and useful technological advancement which is used to look at offshore operations remotely. This is common in the offshore oil and gas industry, particularly for monitoring higherrisk drilling and production operations. This technology is typically used to monitor all aspects of drilling and production operations, including downhole parameters measured while drilling, and readings from the blowout preventer, and the drilling fluid handling system. Shore-based monitoring facilities are often used to analyze data streamed from the higher-risk offshore drilling and production platforms, especially floating platforms in the Deepwater regions of the Outer Continental Shelf. These monitoring facilities are often staffed with highly trained and experienced drilling and production experts, who can view the same data that is being monitored in the Control Room of the offshore platform. The staff at the monitoring facilities provide “a second set of eyes” for observing the offshore operations and can also assist by providing technical and troubleshooting support.
Figure 4 presents a typical offshore RTOC equipped with an eye tracking system. The utilization of multiple screens might distract the operator from effectively performing multiple tasks simultaneously. The eye tracking data depicts the attention span of the operator at the terminal. In Figure 4, the green circular block represents the spatial domain where the operator attentions are missing. The red circular block represents the distraction during the operation. It might be possible that the attention of the operator might have shifted from green circled area to red one. Further, the distribution of ocular movement data is nonuniform. Having such distinct features of this eye fixation data can provide cues which can act as a reliable input for state-of-the-art algorithms to cater with the on-site distractions.
The mental states of an alert and tired person are entirely different. A tired person could feel sleepy and may close his eyes. The basic premise of the eye-tracking technology relies on recording of light reflected from the cornea. The sleepy person will frequently close its eyes, and the eye tracking camera will not be able to capture the fixation data or pupil size. This very concept can be utilized in the realization of the difference between the scanning behavior of alert and non-alert person. The eye fixation count and duration will be significantly missed during such conditions. The comparison of eye-fixation count and duration of the subject and standardized expert data over a particular real-time monitoring operation can provide a cue towards the state of the operators. The decrease in pupil size in a real-time system can also provide cues on alertness of the subjects. If the ocular activity of the onsite participant is below a satisfactory level, the system can be triggered. The triggered event can alert the participant to be cautious about the AOI which lacked the focus of the participant. Keeping this in mind, establishing the working procedure of eye tracking technology with regards to oil and gas operation with different mental state of the participants becomes highly desirable. Hence in the next section, an experimental investigation with data analysis is presented to extract the difference in situational awareness quotient of different participants.

Methods

This study used a controlled simulated environment consisting of eye trackers, display screen, and well-log data to measure SA. The participants that we define as novice and an expert were presented a scenario of a real-time well-log data of an oil well on the display screen, and were tasked to monitor the real-time data, interpret the data, explain their understanding, and point out any abnormalities while their eye movements were being recorded using an eye-tracker. The experiment was conducted using the Virtual Reality Drilling Simulator (VRDS) available at the University of Oklahoma.
Some AOIs were defined on the well-log data. Then, the eye fixation durations and counts were collected during the simulation for all the participants. The data were processed to obtain the percentage of eye fixation durations and counts for further analysis. “Percent eye fixation duration” represents the percentage of the eye fixation span of participants at an AOI over the total fixation duration considering the whole log as an AOI. Similarly, “percent eye fixation count” refers to the percentage of fixation counts in an AOI with respect to total counts in the whole log. The percentages of fixation duration and fixation count were used to normalize the difference in the plot area of the considered AOI. Then a statistical study was conducted to obtain cues on the cognitive response of the participants.
The participants were given a basic knowledge of drilling scenario and human factors. The novices were trained for a theoretical understanding of influx indicators and power loss on the rig. Before simulation, a pre-briefing was conducted where the details of the scenario, execution procedure, and the expected responses were explained. The scenario was limited to five minutes. During the simulation, the participants’ gaze movements were captured using eye trackers in real-time on the pre-defined AOIs. We hypothesized to observe differences in the ocular activity of the expert and the novice.

Participants

For this research, we recruited twenty-three undergraduate and graduate students of the Petroleum engineering department at the University of Oklahoma as novices. These participants were given classroom training about the drilling operations and well control procedures prior to the experimentation. In addition, we had an expert participant from the industry who had more than 30 years of field experience of working on oil and gas rigs. In total, this study involved twenty-four participants.
In this paper, we compared the ocular activity of a novice (one of the twenty-three novices) and an expert as indicators of situational awareness to show the potential of using these differences for a feedback system at an RTOC.

Materials

The study was conducted using the VRDS facility at the University of Oklahoma. The VRDS room is equipped with the National Oilwell Varco (NOV) offshore drilling simulator, and it closely mimics the offshore operational environment. This multimillion-dollar facility simulates an offshore environment of a Drillship with three cyber base chairs as shown in Figure 5. The cyber chairs have two display monitors (one on each arm), keypads, roller-mouse, and joysticks to control the equipment and perform common surface operations on a rig such as Tripping-in, Tripping-out, Top-drive to Pipe connection, and Drilling. The facility also has joystick module to operate pipe and riser catwalk. The display wall is used for visualization of an oil rig platform. For this simulation and eye tracking, a Visual Basic program was written to display the data in real-time on the monitor screen which included the different parameters as exhibited in the scenario plot shown in Figure 6. All the data was displayed in the form of lines with different colors, and a legend was provided to depict the characteristics of the lines.
The eye tracking was captured using Tobii TX 300 eye tracker with a sampling rate of 300 Hz and a visual angle accuracy of 0.4 degrees. The eye tracker was placed approximately three feet from the participants’ eyes. A 23 inches monitor was fitted above the eye tracker to display the log of the scenario. The data was recorded and digitized using Tobii Studio 4 software and then later analyzed using R software.

Procedure

Prior to this simulation study, the novices were given training on reading and interpretation of well logs and to identify the abnormalities. For this scenario, the protocols for pre-briefing, simulation, and debriefing were developed. The pre-briefing included training on different aspects of well log parameters and simulation practice. After that, the participants were briefed about the procedures and standard practices of the scenario simulation.
The participants’ ocular activity was monitored using Tobii eye tracking monitor, and the data was collected. The participants were instructed to interpret the displayed dynamic well data. The primary task was to identify the abnormality in parameters, in-situ conditions, and forecast the consequences. The objective of this scenario was to assess the participant’s cognitive behavior during the realtime data monitoring of well-logs.

Data Collection

Eye movement data for eye fixation duration, eye fixation count, and pupil size during the scenario were collected using Tobii eye trackers in a single trial. We defined three AOIs to capture eye movements data as are shown in Figure 6. The full log is considered as the AOI I less the other two AOIs. Apart from this, two spatially fixed AOIs were designed to correspond to the real-time monitoring of participants’ response. These two AOIs are represented as AOI II and AOI III. The AOIs were selected based on their support as indicators for the interpretation. The major indicators include rate of penetration (ROP), flow out (Flow out %), gas penetration (Gas %), and active pit volume (Active Pits). The ROP, which represents the speed of penetration of the drill bit into the rock to deepen it, is the most important indicator as it would increase in possible situations of a kick (a rapid increase in oil well pressure or gas influx). However, this alone cannot be used to confirm the kick as it could be due to softer strata other than gas influx. Therefore, the operator needs to confirm it through Flow out %, Gas %, and Active Pits. AOI II covers sudden picks in the Flow out %. AOI III covers Gas % and Active Pits.
The eye movement data was registered and stored in Tobii Studio software which included the spatial fixation and 0.42 pixels/milliseconds velocity threshold (Kang and Landry, 2014). The analysis of these AOIs was conducted using simulation output and eye tracking output.

Results and Discussions

The sample eye fixation data for the expert is shown in Figure 7, whereas the data of the novice is shown in Figure 8. The data consists of various sizes of circles with different numbers inside it. The size of these circles represents the amount of time span of the participant at a particular spatial position during the simulation. The visual search/scan behavior differs from participant to participant. To account for that, the eye tracking software implements the timestamping of each fixation. Corresponding to the time stamp, the software assigns the numeric order. The number inside the circle represents this sequential numbering of visual search. In this study, only the fixation spans and counts are accounted to conduct the assessment of participants’ attention while scanning the plot in realtime.
The eye fixation data in Figure 7(b) and Figure 8(b) was collected for a 50-second duration, during a critical time period from 2:00 to 2:50 seconds, where an anomaly (a kick) was present, and the participants were required to properly identify it and take appropriate actions to verify and control it.
The preliminary investigation of the overall distribution of scanning behaviors among participants shows a significant difference from one sector of the plot to the other. The data shows that the expert correctly detected the anomaly and is providing proper inputs to mitigate the anomaly, whereas the novice continues observing the graph as the eye fixation data is all over the graph. During the critical 50 second duration, the expert correctly creates eye fixations on the anomaly that required attention (i.e., AOI III), whereas the novice does not with minimal eye fixation in the AOI III.
Further, it is seen that during the critical 50-second interval, the expert is well aware of the situation, as evidenced by the eye tracking data (i.e., eye fixation in AOI III), whereas the novice seems to be too much occupied on the left end of the graph (continuously observing ROP) and not confirming the anomaly through other indicators in AOI II and AOI III which can be inferred as low situation awareness of the novice. From the full-duration eye fixation data, it seems that the expert knows it is meaningless to continuously observe further (which accords with why we chose such AOIs for analyses) and concentration the anomaly detected in AOI III, whereas the novice does not really know what is going on.
It is worth noting that a real-time alarm/feedback system, which can caution the participant about this deficiency, will improve the performance of the subject. We hypothesized that the situation awareness will be different between the novice and the expert based on the assumption that the situation awareness is correlated with eye fixation data which is well supported in the literature. In detail, the hypothesis assumes that the participant with greater situational awareness will have greater fixation data on certain AOIs. This hypothesis is also substantiated by the high correlation and similar trends observed in the two categories with respect to eye fixation duration and count data.
There is a considerable amount of scientific evidence to support the hypothesis that situation awareness is correlated with eye fixation data and that expert participants may exhibit greater situational awareness than novice participants. has been supported by several studies in literature. Studies have shown that eye tracking data can be used to assess situational awareness and training effectiveness in various industries, including aviation, healthcare, military settings, and oil and gas operations. For example, a study by Endsley and Garland (2000) found that expert pilots had greater situational awareness than novice pilots, and that this was reflected in differences in their eye gaze patterns. Similarly, Salehi et al. (2018) developed a crossdisciplinary, scenario-based training approach integrated with eye tracking data collection to enhance situational awareness in offshore oil and gas operations. The authors found that eye tracking data could be used to identify areas where situational awareness was lacking and to tailor training interventions accordingly. Similarly, Kang, Jeon, and Salehi (2020) used eye tracking analytics to evaluate the effectiveness of virtual reality training for situational awareness. The authors found that eye tracking data could be used to measure trainees' visual attention and identify areas for improvement in the training program. Hermens, Flin, and Ahmed (2013) provide additional evidence to support the idea that there are differences in situational awareness between novices and experts and that experts tend to have more focused and efficient eye movements compared to novices.
The oculomotor data for the mean percent eye fixation duration and count for the novice and the expert are provided in Figure 9. The comparative study of participants’ response between two AOIs suggested that participants tended to focus more on AOI III than the other AOIs. Figure 9(a) depicts the visual representation of participants’ data for mean percent eye fixation duration. The significant difference in response between AOI II and AOI III can be explained by the deviation of the log parameters in both AOIs. The gas percent fluctuation is more prominent in AOI III than other parameters of the log which might have caught the attention of the participants. This pattern suggests that the participants’ visual attention can be correlated with their situation awareness.
Similar observations were made for the percent mean eye fixation count data as shown in Figure 9(b). The novice was found to be on the lower side compared to the expert fixation counts in case of AOI II and AOI III. This trend can be attributed to distraction exhibited by the less situationally aware participant. The significant differences of scanning behaviors between novice and expert on the important AOI III is significantly lower than that of the novice’s duration indicating that the expert was able to quickly assess the situation than the novice.
Apart from visually observing the data, a statistical study was conducted on the visual data where we compared the mean of all novices’ data with the expert to understand the underlying. To find out the significance of the difference in fixation duration and fixation counts across three AOIs, non-parametric Kruskal-Wallis test was performed. The results indicated significant difference across three AOIs in terms of fixation duration (H (2, 66) = 51.24, p < .01) and fixation counts (H (2, 66) = 55.39, p < .01). Further, pairwise comparison indicated significant difference in fixation duration of AOI I with AOI II and III, and of AOI II with AOI III (see Table 1a). The same was observed for fixation counts, i.e., significant difference in fixation counts of AOI I with AOI II and III and of AOI II with AOI III (see Table 1b), indicating that there is significant difference in terms of fixation duration and fixation counts on all three AOIs.
While comparing novices’ data it was observed that novices spent more time and had more counts on AOI I which is of lesser importance for anomaly detection and confirmation (see Figure 10a & Figure 10b). In comparison, the expert had much lesser counts in this AOI, clearly showing that he knew where to look in case of an anomaly detection.
For the AOI II and III, the overall time frame comparisons between novices and the expert did not clearly show significance. This could be due to multiple reasons such as the expert might have immediately noticed the anomaly, acted, then ended (no need to see more), whereas a portion of the novices immediately noticed but took more time by looking it longer, and another portion of the novices did not notice, so averaging them would provide unexpected results. This needs further experimentation and is planned as future research.
Further, the novices might have different levels of situation awareness meaning that some were able to quickly notice the issue whereas some were not. The differences in situational awareness of participants can be due to several factors such as domain specific knowledge, prior experiences, mood, alertness, fatigue, distractions etc. The representative example shows a case when a novice clearly lacked awareness (see Figure 7). The analysis of each participant's awareness level has not been analyzed in this paper, and an in-depth and systematic method is being devised to better classify various awareness levels as future research to corroborate.

Pupillogram Analysis

Apart from the fixation data, pupil dilation of the participants was also studied. The main steps involved in this data analysis are noise filtering from the pupil size data, averaging both (right and left) pupil size values, baseline adjustment, and then normalizing the data. First, one of the signal smoothing and noise filtering techniques, SavitzkyGolay filter was implemented to minimize the error for excessive sampling at the onset or offset of gaps. SavitzkyGolay filter is one of commonly used low pass filtering technique in signal processing (Chen et al., 2004; Luo, Ying, He, & Bai, 2005). In addition to this filtering technique used for removing the noise from the data, linear interpolation was applied to calculate the missing values of the pupil size. For interpolation, the slope was calculated using a window of 21 data points (10 data points above the missing data and 10 below the missing data) (Jackson & Sirois, 2009). Furthermore, the average value of pupil size of both eyes was calculated. To compare the responses between the participants, the data was normalized after baseline adjustment.
Figure 11 shows the dynamic pupil size distribution for the novice and expert. On a closer look at the plot, it is evident that both participants have similar change of pupil size in the beginning. However, the novice exhibited less dilation in pupil size during 100-200 seconds of the experiment. Again, in the range of 200-250 seconds both participants’ responses were almost similar. However, in the final 50 seconds the dilation of the novice was lower than that of the expert. As explained in the previous section, pupil dilation measure shows better cognitive response from the expert which can be inferred as better alertness and awareness. In addition, when the data of AOI II and III were combined for both participants, the percentage of fixation exhibited similar characteristics suggesting a high correlation between the eye fixation data and the pupillogram.
Overall, we can infer that the explicit distinctions between the eye movements of experts and novices can be used to set up a framework that can characterize the alertness and awareness of the participants. The eye tracking data can also be used to provide feedback while the users are tracking the data in the RTOC. For example, if the eye movement characteristics deviate from those of a representative expert, then an alert can be issued, or scaffolding materials can be provided in a timely manner.

Conclusion

A study of the real-time monitoring system with regards to the oil and gas industry has been conducted. In addition, the use of eye tracking technology in various applications, such as aviation, weather forecasting, and medical industry, were investigated that can be adapted to offshore oil drilling operations. Based on the theoretical concepts, the role of eye tracking in the real-time monitoring system was explained and resulted in establishing a framework for applying the oculomotor data to better assess operators’ cognitive states. Finally, an experiment using the scenario-based offshore operation simulation was conducted using eye tracking technology.
The simulated experiments exhibited the non-uniform distribution of eye fixation data over the dynamically displayed well-logging data which can be correlated with the situational awareness of the participants. In addition, eye fixation data can be used to quantify the scanning behavior. The study showed promising results which can be further linked to individual situational awareness and be processed in real-time to establish a feedback/alert system. In addition, a pupillometry analysis was conducted and the pupil dilation effect clearly distinguished the cognitive responses of the participants and showed high correlation with the other eye tracking data such as fixation counts and durations.
For this study, we were fortunate to have an experienced field expert with 30 years of experience. It is planned to recruit more experts for future studies. Further, the ranking of participants in terms of performance may be possible with more data. This is planned studies as well. Furthermore, upcoming studies aim to modify advanced techniques for analyzing eye movements, which encompass visual entropy (Naeeri, Kang, & Palma Fraga, 2022), a multimodal approach to analyzing fatigue and eye movements (Naeeri, Kang, Mandal, & Kim, 2021), a framework for analyzing eye movements in real-time (Palma Fraga, Kang, Lee, & Crutchfield, 2021), and machine learning (Palma Fraga, Kang, & Axthelm, under review). These techniques can be utilized to formulate real-time safety assessment algorithms for various drilling operations at drilling rigs and Real Time Operation Centers.

Ethics and Conflict of Interest

The author(s) declare(s) that the contents of the article are in agreement with the ethics described in http://bib-lio.unibe.ch/portale/elibrary/BOP/jemr/ethics.html and that there is no conflict of interest regarding the publication of this paper.

Acknowledgments

This material was based upon work supported by the National Academy of Sciences, Engineering, and Medicine under the Gulf Research Program (Grant #2000007356). The opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Academies of Sciences.
The authors would like to thank Mr. Stan Christman who helped in the review of scenario experiments.

Abbreviations and Acronyms

AOIArea of Interest
BOPBlowout Preventer
BSEEBureau of Safety and Environmental Enforcement
CRMCrew Resource Management
LOWCLoss of Well Control
NOVNational Oilwell Varco
NPTNon-productive Time
NTSsNon-technical Skills
ROPRate of Penetration
RTOCReal Time Operations Center
SASituation/Situational Awareness
SPPStandpipe Pressure
VRDSVirtual Reality Drilling Simulator
WOBWeight on Bit

References

  1. Atik, O., and O. Arslan. 2019. Use of eye tracking for assessment of electronic navigation competency in maritime training. Journal of Eye Movement Research 12, 3. [Google Scholar] [CrossRef]
  2. Atik, O. 2020. Eye tracking for assessment of situational awareness in bridge resource management training. Journal of Eye Movement Research 12, 3. [Google Scholar] [CrossRef]
  3. Babu, M. D., D. V. JeevithaShree, G. Prabhakar, K. P. S. Saluja, A. Pashilkar, and P. Biswas. 2019. Estimating pilots’ cognitive load from ocular parameters through simulation and in-flight studies. Journal of Eye Movement Research 12, 3. [Google Scholar] [CrossRef]
  4. Beatty, J., and B. L. Wagoner. 1978. Pupillometric signs of brain activation vary with level of cognitive processing. Science 199, 4334: 1216–1218. [Google Scholar] [CrossRef] [PubMed]
  5. Bjerkebaek, E., and T. S. Eskedal. 2004. Safety Assessment of Alarm Systems on Offshore Oil and Gas Production Installations in Norway. Paper presented at the SPE International Conference on Health, Safety, and Environment in Oil and Gas Exploration and Production, Calgary, Alberta, Canada. [Google Scholar] [CrossRef]
  6. Booth, J., and J. Hebert. 1989. Support of Drilling Operations Using a Central Computer and Communications Facility With Real-Time MWD Capability and Networked Personal Computers. Paper presented at the Petroleum Computer Conference. [Google Scholar]
  7. BSEE. 2015. Investigation of Loss of Well Control and Fire South Timbalier Area Block 220, Well No. A-3 OCS-G 24980. 23 July 2013. Translated by D. o. Interior. Bureau of Safety and Environmental Enforcement. US Department of Interior: p. 43. [Google Scholar]
  8. BSEE. 2016. Annual Report 2016; Bureau of Safety and Environmental Enforcement. Available online: https://www.bsee.gov/sites/bsee.gov/files/bsee_2016_annual_report_v5_final_january2017.pdf.
  9. BSEE. 2018. Bureau of Safety and Environmental Enforcement. Available online: https://www.bsee.gov/ (accessed on 25 May 2018).
  10. Bullemer, P., and D. Reising. 2013. Improving the Operations Team Situation Awareness: Lessons Learned from Major Process Industry Incidents. Paper presented at the American Fuel & Petrochemical Manufacturers Annual Meeting. [Google Scholar]
  11. Camilli, M., M. Terenzi, and F. Di Nocera. 2007. Concurrent validity of an ocular measure of mental workload. Human factors issues in complex system performance, 117–129. [Google Scholar]
  12. Chen, J., P. Jönsson, M. Tamura, Z. Gu, B. Matsushita, and L. Eklundh. 2004. A simple method for reconstructing a high-quality NDVI time-series data set based on the Savitzky–Golay filter. Remote sensing of Environment 91, 3-4: 332–344. [Google Scholar]
  13. Dadmohammadi, Y., S. Salehi, R. Kiran, J. Jeon, Z. Kang, E. T. Cokely, and V. Ybarra. 2017. Integrating Human Factors into Petroleum Engineering's Curriculum: Essential Training for Students. Paper presented at the SPE Annual Technical Conference and Exhibition, San Antonio, Texas, USA. [Google Scholar]
  14. Dick, A. 1980. Instrument scanning and controlling: using eye movement data to understand pilot behavior and strategies. NASA Contractor Report 3306. [Google Scholar]
  15. Endsley, M. R. 1988. Situation awareness global assessment technique (SAGAT). Paper presented at the Aerospace and Electronics Conference, 1988. NAECON 1988., Proceedings of the IEEE 1988 National. [Google Scholar]
  16. Endsley, M. R. 1995. Toward a Theory of Situation Awareness in Dynamic Systems. Human Factors 37, 1: 32–64. [Google Scholar] [CrossRef]
  17. Endsley, M. R. 2016. Designing for situation awareness: An approach to user-centered design. CRC press. [Google Scholar]
  18. Endsley, M. R., and D. J. Garland. 2000. Pilot situation awareness training in general aviation. Paper presented at the Proceedings of the Human Factors and Ergonomics Society Annual Meeting. [Google Scholar]
  19. Fletcher, G., P. McGeorge, R. H. Flin, R. J. Glavin, and N. J. Maran. 2002. The role of non-technical skills in anaesthesia: a review of current literature. British Journal of Anaesthesia 88, 3: 418–429. [Google Scholar]
  20. Fokin, D., S. Blohm, and E. Riekhakainen. 2022. Reading Russian poetry: An expert–novice study. Journal of Eye Movement Research 13, 3. [Google Scholar] [CrossRef]
  21. Foreman, R. 1985. The drilling command and control system. Paper presented at the SPE Annual Technical Conference and Exhibition. [Google Scholar]
  22. Fotiou, D., V. Stergiou, D. Tsiptsios, C. Lithari, M. Nakou, and A. Karlovasitou. 2009. Cholinergic deficiency in Alzheimer's and Parkinson's disease: evaluation with pupillometry. International Journal of Psychophysiology 73, 2: 143–149. [Google Scholar]
  23. Goetz, C. 2016. Control System Alarms: From Inundation to Pragmatic Use. Paper presented at the 2016 IADC Human Factors Conference & Exhibition. [Google Scholar]
  24. Groner, R., and E. Kasneci. 2021. Eye movements in real and simulated driving and navigation control Foreword to the Special Issue. Journal of Eye Movement Research 12, 3. [Google Scholar] [CrossRef] [PubMed]
  25. Helbren, E., S. Halligan, P. Phillips, D. Boone, T. R. Fanshawe, S. A. Taylor, and S. Mallett. 2014. Towards a framework for analysis of eye-tracking studies in the three dimensional environment: a study of visual search by experienced readers of endoluminal CT colonography. The British Journal of Radiology 87, 1037: 20130614. [Google Scholar] [PubMed]
  26. Hermens, F., R. Flin, and I. Ahmed. 2013. Eye movements in surgery: A literature review. Journal of Eye Movement Research 6, 4. [Google Scholar]
  27. Holland, P. 2017. Loss of Well Control Occurrence and Size Estimators, Phase I and II. BSEE Report no. ES201471/2. [Google Scholar]
  28. Ikuma, L. H., C. Harvey, C. F. Taylor, and C. Handal. 2014. A guide for assessing control room operator performance using speed and accuracy, perceived workload, situation awareness, and eye tracking. Journal of Loss Prevention in the Process Industries 32: 454–465. [Google Scholar] [CrossRef]
  29. Jackson, I., and S. Sirois. 2009. Infant cognition: going full factorial with pupil dilation. Developmental Science 12, 4: 670–679. [Google Scholar]
  30. Kaber, D. B., and M. R. Endsley. 2004. The effects of level of automation and adaptive automation on human performance, situation awareness and workload in a dynamic control task. Theoretical Issues in Ergonomics Science 5, 2: 113–153. [Google Scholar]
  31. Kang, Z., J. Jeon, and S. Salehi. 2020. Eye tracking data analytics in virtual reality training: Application in Deepwater Horizon oil drilling operation. Paper presented at the Proceedings of the Human Factors and Ergonomics Society Annual Meeting. [Google Scholar]
  32. Kang, Z., and S. Landry. 2014. An eye movement analysis algorithm for a multielement target tracking task: Maximum transition-based agglomerative hierarchical clustering. IEEE Transactions on Human-Machine Systems 45, 1: 13–24. [Google Scholar]
  33. Klingner, J., B. Tversky, and P. Hanrahan. 2011. Effects of visual and verbal presentation on cognitive load in vigilance, memory, and arithmetic tasks. Psychophysiology 48, 3: 323–332. [Google Scholar]
  34. Kodappully, M., B. Srinivasan, and R. Srinivasan. 2016. Towards predicting human error: Eye gaze analysis for identification of cognitive steps performed by control room operators. Journal of Loss Prevention in the Process Industries 42: 35–46. [Google Scholar] [CrossRef]
  35. Kolodziej, M., A. Majkowski, P. Francuz, R. J. Rak, and P. Augustynowicz. 2018. Identifying experts in the field of visual arts using oculomotor signals. Journal of Eye Movement Research 11, 3. [Google Scholar] [CrossRef]
  36. Luo, J., K. Ying, P. He, and J. Bai. 2005. Properties of Savitzky–Golay digital differentiators. Digital Signal Processing 15, 2: 122–136. [Google Scholar] [CrossRef]
  37. Metzger, U., and R. Parasuraman. 2005. Automation in future air traffic management: Effects of decision aid reliability on controller performance and mental workload. Human Factors 47, 1: 35–49. [Google Scholar] [CrossRef]
  38. Naeeri, S., Z. Kang, S. Mandal, and K. Kim. 2021. Multimodal analysis of eye movements and fatigue in a simulated glass cockpit environment. Aerospace 8, 10: 283. [Google Scholar] [CrossRef]
  39. Naeeri, S., Z. Kang, and R. Palma Fraga. 2022. Investigation of Pilots' Visual Entropy and Eye Fixations for Simulated Flights Consisted of Multiple Take-Offs and Landings. Journal of Aviation/Aerospace Education & Research 31, 2: 2. [Google Scholar]
  40. Naqvi, S. A. M., M. Raza, V. T. Ybarra, S. Salehi, and C. Teodoriu. 2018. Using Content Analysis through Simulation-Based Training for Offshore Drilling Operations: Implications for Process Safety. Process Safety and Environmental Protection. [Google Scholar]
  41. NorskOlje, Gass. 2017. Sharing to be Better #1. Well control incident in 9,5 inch section: The Norwegian Oil and Gas Association. Norwegian Petroleum Directorate. (2001). Principles for alarm system design. [Google Scholar]
  42. O'Connor, P., and R. Flin. 2003. Crew resource management training for offshore oil production teams. Safety Science 41, 7: 591–609. [Google Scholar] [CrossRef]
  43. Palma Fraga, R., Z. Kang, and C. M. Axthelm. under review. Effect of machine learning cross-validation algorithms considering human participants and time-series: A case study using biometric data obtained from a virtual reality experiment. Paper presented at the 67th Annual Meeting of Human Factors and Ergonomics Society, Washington D.C, Oct. 23 - 27. [Google Scholar]
  44. Palma Fraga, R., Z. Kang, J. Lee, and J. M. Crutchfield. 2021. Real-time eye tracking analysis for training in a dynamic task. Paper presented at the 2021 4th International Conference on Bio-Engineering for Smart Technologies (BioSMART), Paris, France, Dec. 8-10. [Google Scholar]
  45. Perra, J., A. Latimier, B. Poulin-Charronnat, T. Baccino, and V. Drai-Zerbib. 2022. A meta-analysis on the effect of expertise on eye movements during music reading. Journal of Eye Movement Research 15, 4. [Google Scholar] [CrossRef]
  46. Raza, M. A., S. Salehi, S. Ghazal, V. T. Ybarra, S. A. Mehdi Naqvi, E. T. Cokely, and C. Teodoriu. 2019. Situational awareness measurement in a simulationbased training framework for offshore well control operations. Journal of Loss Prevention in the Process Industries 62: 103921. [Google Scholar] [CrossRef]
  47. Robinski, M., and M. Stein. 2013. Tracking Visual Scanning Techniques in Training Simulation for Helicopter Landing. Journal of Eye Movement Research 6, 2. [Google Scholar] [CrossRef]
  48. Salehi, S., R. Kiran, J. Jeon, Z. Kang, E. T. Cokely, and V. Ybarra. 2018. Developing a cross-disciplinary, scenario-based training approach integrated with eye tracking data collection to enhance situational awareness in offshore oil and gas operations. Journal of Loss Prevention in the Process Industries 56: 78–94. [Google Scholar] [CrossRef]
  49. Satterthwaite, T. D., L. Green, J. Myerson, J. Parker, M. Ramaratnam, and R. L. Buckner. 2007. Dissociable but inter-related systems of cognitive control and reward during decision making: evidence from pupillometry and event-related fMRI. Neuroimage 37, 3: 1017–1031. [Google Scholar] [CrossRef]
  50. Simon, P., F. Rousseau, and J.-C. Angue. 1993. Quantitative analysis of mental workload influence on eye scanning movements. Paper presented at the Proceedings of IEEE Systems Man and Cybernetics Conference-SMC. [Google Scholar]
  51. Taylor, R. M. 1990. Situation awareness rating technique (SART): the development of a tool for aircrew systems design. In Situational Awareness in Aerospace Operations. France: Neuilly sur-Seine, NATOAGARD-CP-478. [Google Scholar]
  52. Thorogood, J., M. Crichton, and A. Bahamondes. 2014. Case Study of Weak Signals and Confirmation Bias in Drilling Operations. SPE Drilling Completion 29, 03: 304–310. [Google Scholar] [CrossRef]
  53. Thorogood, J., and M. T. Crichton. 2014. Threat-and-error management: the connection between process safety and practical action at the worksite. SPE Drilling Completion 29, 04: 465–472. [Google Scholar]
  54. Thorogood, J. L. 2015. The Macondo Inflow Test Decision: Implications for Well Control and Non-technical Skills Training. Paper presented at the SPE/IADC Drilling Conference and Exhibition. [Google Scholar]
  55. Tobii. 2018. Tobii Pro: Learn and Support. Available online: https://www.tobiipro.com/learn-and-sup-port/learn/eye-tracking-essentials/.
  56. Topalli, D., and N. E. Cagiltay. 2018. Eye-hand coordination patterns of intermediate and novice surgeons in a simulation-based endoscopic surgery training environment. Journal of Eye Movement Research 11, 6. [Google Scholar] [CrossRef]
  57. Ursem, L.-J., J. H. Williams, N. M. Pellerin, and D. H. Kaminski. 2003. Real time operations centers; the people aspects of drilling decision making. Paper presented at the SPE/IADC Drilling Conference. [Google Scholar]
  58. Van Orden, K. F., W. Limbert, S. Makeig, and T.-P. Jung. 2001. Eye activity correlates of workload during a visuospatial memory task. Human Factors 43, 1: 111–121. [Google Scholar] [CrossRef] [PubMed]
  59. Vlačić, S. I., A. Z. Knežević, S. Mandal, S. Rođenkov, and P. Vitsas. 2020. Improving the pilot selection process by using eye-tracking tools. Journal of Eye Movement Research 12, 3. [Google Scholar] [CrossRef]
  60. Wilson, M. R., J. S. McGrath, S. J. Vine, J. Brewer, D. Defriend, and R. S. Masters. 2011. Perceptual impairment and psychomotor control in virtual laparoscopic surgery. Surgical Endoscopy 25, 7: 2268–2274. [Google Scholar]
Figure 1. Offshore incidents in the Gulf of Mexico; Major = More than 3 days away from work or more than 3 days of restricted work or job transfer (collectively referred to as DART); Minor = 1-3 DART; and Other = Injuries that resulted in less than one DART (or those that required evacuation to offshore or to another offshore facility for medical treatment but did not result in any DART). (BSEE, 2016).
Figure 1. Offshore incidents in the Gulf of Mexico; Major = More than 3 days away from work or more than 3 days of restricted work or job transfer (collectively referred to as DART); Minor = 1-3 DART; and Other = Injuries that resulted in less than one DART (or those that required evacuation to offshore or to another offshore facility for medical treatment but did not result in any DART). (BSEE, 2016).
Jemr 16 00014 g001
Figure 2. Eye tracking during drilling simulation at the Virtual Reality Drilling Simulator facility at the University of Oklahoma. Visual scan path data overlaid onto an operator’s view of sight: The center of the circles indicate the eye fixation location, the numbers in the circles indicate the time order of the eye fixations, the size of the circles indicate the eye fixation durations, and the linear lines indicate the fast saccades among the eye fixations. Visual search behaviors can be investigated using visual scan paths.
Figure 2. Eye tracking during drilling simulation at the Virtual Reality Drilling Simulator facility at the University of Oklahoma. Visual scan path data overlaid onto an operator’s view of sight: The center of the circles indicate the eye fixation location, the numbers in the circles indicate the time order of the eye fixations, the size of the circles indicate the eye fixation durations, and the linear lines indicate the fast saccades among the eye fixations. Visual search behaviors can be investigated using visual scan paths.
Jemr 16 00014 g002
Figure 3. (3a) The alarms received on a drilling rig in a week. The drillers can be inundated with this much information (After Goetz, 2016); and (3b) A case study showing the parameters during LOWC and fire in South Timbalier area-2013 (After BSEE, 2015).
Figure 3. (3a) The alarms received on a drilling rig in a week. The drillers can be inundated with this much information (After Goetz, 2016); and (3b) A case study showing the parameters during LOWC and fire in South Timbalier area-2013 (After BSEE, 2015).
Jemr 16 00014 g003
Figure 4. Eye tracking data at real-time operation center (RTOC).
Figure 4. Eye tracking data at real-time operation center (RTOC).
Jemr 16 00014 g004
Figure 5. Virtual Reality Drilling Simulator (VRDS) at the University of Oklahoma.
Figure 5. Virtual Reality Drilling Simulator (VRDS) at the University of Oklahoma.
Jemr 16 00014 g005
Figure 6. Area of Interests on the dynamically displayed well logging data (after NorskOlje&Gass (2017)).
Figure 6. Area of Interests on the dynamically displayed well logging data (after NorskOlje&Gass (2017)).
Jemr 16 00014 g006
Figure 7. Ocular data distribution of the expert during the full duration of the simulation 7(a) and during 2:00 to 2:50 second interval 7(b).
Figure 7. Ocular data distribution of the expert during the full duration of the simulation 7(a) and during 2:00 to 2:50 second interval 7(b).
Jemr 16 00014 g007
Figure 8. Ocular data distribution of the novice during the full duration of the simulation 8(a) and during 2:00 to 2:50 second interval 8(b).
Figure 8. Ocular data distribution of the novice during the full duration of the simulation 8(a) and during 2:00 to 2:50 second interval 8(b).
Jemr 16 00014 g008
Figure 9. Percent mean eye fixation duration 9(a) and count for the novice and the expert 9(b).
Figure 9. Percent mean eye fixation duration 9(a) and count for the novice and the expert 9(b).
Jemr 16 00014 g009
Figure 10a. Mean fixation duration on all three AOIs by Novices, with expert’s values imposed in red.
Figure 10a. Mean fixation duration on all three AOIs by Novices, with expert’s values imposed in red.
Jemr 16 00014 g010
Figure 10b. Mean fixation counts on all three AOIs by Novices, with expert’s values imposed in red.
Figure 10b. Mean fixation counts on all three AOIs by Novices, with expert’s values imposed in red.
Jemr 16 00014 g011
Figure 11. Average change in pupil diameter (millimeter) of novice and expert while monitoring the real-time data during scenario execution.
Figure 11. Average change in pupil diameter (millimeter) of novice and expert while monitoring the real-time data during scenario execution.
Jemr 16 00014 g012
Table 1a. Pairwise Comparisons of three AOIs on fixation duration.
Table 1a. Pairwise Comparisons of three AOIs on fixation duration.
Jemr 16 00014 i001
Table 1b. Pairwise Comparisons of three AOIs on fixation counts.
Table 1b. Pairwise Comparisons of three AOIs on fixation counts.
Jemr 16 00014 i002

Share and Cite

MDPI and ACS Style

Raza, M.A.; Kiran, R.; Ghazal, S.; Kang, Z.; Salehi, S.; Cokely, E.; Jeon, J. An Eye Tracking Based Framework for Safety Improvement of Offshore Operations. J. Eye Mov. Res. 2023, 16, 1-17. https://doi.org/10.16910/jemr.16.3.2

AMA Style

Raza MA, Kiran R, Ghazal S, Kang Z, Salehi S, Cokely E, Jeon J. An Eye Tracking Based Framework for Safety Improvement of Offshore Operations. Journal of Eye Movement Research. 2023; 16(3):1-17. https://doi.org/10.16910/jemr.16.3.2

Chicago/Turabian Style

Raza, Muhammad A., Raj Kiran, Saima Ghazal, Ziho Kang, Saeed Salehi, Edward Cokely, and Jiwon Jeon. 2023. "An Eye Tracking Based Framework for Safety Improvement of Offshore Operations" Journal of Eye Movement Research 16, no. 3: 1-17. https://doi.org/10.16910/jemr.16.3.2

APA Style

Raza, M. A., Kiran, R., Ghazal, S., Kang, Z., Salehi, S., Cokely, E., & Jeon, J. (2023). An Eye Tracking Based Framework for Safety Improvement of Offshore Operations. Journal of Eye Movement Research, 16(3), 1-17. https://doi.org/10.16910/jemr.16.3.2

Article Metrics

Back to TopTop