Next Article in Journal
Eye Tracking for Assessment of Situational Awareness in Bridge Resource Management Training
Previous Article in Journal
Eye Tracking Use in Researching Driver Distraction: A Scientometric and Qualitative Literature Review Approach
 
 
Journal of Eye Movement Research is published by MDPI from Volume 18 Issue 1 (2025). Previous articles were published by another publisher in Open Access under a CC-BY (or CC-BY-NC-ND) licence, and they are hosted by MDPI on mdpi.com as a courtesy and upon agreement with Bern Open Publishing (BOP).
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analysis and Evaluation of Eye Behavior for Marine Operation Training—A Pilot Study

by
Runze Mao
,
Guoyuan Li
*,
Hans Petter Hildre
and
Houxiang Zhang
Norwegian University of Science and Technology, Alesund, Norway
*
Author to whom correspondence should be addressed.
J. Eye Mov. Res. 2019, 12(3), 1-14; https://doi.org/10.16910/jemr.12.3.6
Submission received: 2 August 2019 / Published: 6 December 2019

Abstract

:
This paper presents a new analysis approach for evaluating situation awareness in marine operation training. Taking advantage of eye tracking technology, the situation awareness reflected by visual attention can be visualized and analyzed. A scanpath similarity comparison method that allows group-wise comparisons is proposed. The term ‘Expert zone’ is introduced to evaluate the performance of novice operator based on expert operators’ eye movement. It is used to evaluate performance of novice operators in groups in certain segment of marine operation. A pilot study of crane lifting experiment was carried out. Two target stages of operation for the load descending until total immersion to the seabed were selected and analyzed for both novice and expert operators. The group-wise evaluation method is proven to be able to access the performance of the operator. Besides that, from data analysis of fixation-related source and scanpath, the similarities and dissimilarities of eye behavior between novice and expert is concluded with the scanpath mode in target segment.

Introduction

Marine operations have become more complex and diverse in recent years due to the development of technology in the marine industry. At the same time, it is reported that marine accidents such as offshore oil rig mishaps, crane mishaps and the grounding of ships, are usually followed by the development of new technologies. Human error has been established as one of the main threats to marine safety. Four-fifths of major offshore accidents are caused by human error (Luo & Shin, 2016). Therefore, from safety to economy, it is important to understand what causes human error and learn from these failures. Human error has been verified to be associated with situational awareness (Jones & Endsley, 1996). Effectively allocating attention to maintain higher situational awareness plays a key role in detecting hazards and avoiding potential accidents. Although briefing and debriefing are introduced in marine training, evaluation of situational awareness is seldom seen. It cannot clearly and effectively reflect the potential mistake during the training and thus would lay danger in real practice.
To detect situational awareness of human, human eyes can be a reasonable option. If correctly recorded, human eyes are responsible for storing 70% of human information, along with rational analysis and visualization methods, can bring huge progress for engineering improvement. To fulfill this idea, eye tracking technology emerges at the right moment. Eye tracking, is a technique that can measure eye movements to understand the human visual focus. Because eye movements play an important part in consuming attention, the researchers invented eye trackers to objectively study the human error issue in a new way in order to prevent potential accidents. Eye tracking is implemented in many research areas with respect to training. For example, Kasarskis et al. (2001), Sadasivan et al. (2005), Sarter et al. (2007), Yang et al. (2013), Kang & Landry (2014), Dehais et al. (2015) and Wickens et al. (2017) implemented eye tracking techniques to study flight training with a focus on fixation number, mean fixation duration, mean dwell duration and scanpath; Balk et al. (2006), Fisher et al. (2007), Palinko et al. (2010), Paeglis et al. (2011), and Xu et al. (2018) studied driving training by analyzing mean fixation duration and its standard deviation, saccade size, pupil diameter and scanpath; Miall & Tchalenko (2001), and Tchalenko (2009) researched drawing issues introducing dwell time; Law et al. (2004), and Chetwood et al. (2012) carried out surgery training; Kovácsová et al. (2018) researched cyclists’ eye movements by introducing fixation duration; Although many successes of eye tracking applications validate the feasibility in human training, it is seldom to see applying eye tracking technology to marine operation training (Li et al., 2017), much less researches on analyzing and evaluating eye behavior during the training.
It is therefore believed that using eye tracker for marine operation training to study human error is reasonable and feasible. The present research is a pilot study with an aim to improve training quality of marine operation through the analysis of eye behaviors of marine operator in simulator. Considering the complexity of marine operation and the operation differences between experts and novices, how to efficiently evaluate the eye behavior is challenging. To this end, we propose a group-wise comparison method for eye behavior comparison in marine training. The paper’s main contributions can be two folds:
  • A scanpath similarity comparison method which allows group-wise comparisons is proposed to compare and evaluate the performance of novice;
  • A complete eye behavior evaluation in real marine crane lifting operation is carried out, together with a pattern extraction by experts’ operation.

Related Work

Eye movement data type analysis

During years of researches on eye-tracking area, researchers found that there are many different eye movement types which share different eye behaviors, and according to the characteristics of these eye movement, they can be briefly separated into gaze, fixation, saccade, pupil data, blink and scanpath. Analysis and evaluation of eye tracking data has been a hot topic in recent years. They can be separated simply into performance assessment, fatigue, cognitive load, risk perception, experience and expertise. In our study of marine training in this paper, it belongs to the issue of experience and expertise.
Experience and expertise related issue is a commonly discussed topic. It is believed that operators’ experience is important, but it is abstract and hard to measure. The eye-tracking technique offers a suitable method to measure and evaluate experience of operators. Currently, there are only a few researches of eye behavior for marine operations. For example, Atik & Arslan (2019) captured eye movement data from ship officers in simulation exercises and proposed a method for assessing competency between novice and expert operators in navigation training; Hareide & Ostnes (2017) collected eye tracking data from both simulator and field studies for analysis and presented a recommended visual scan path for the maritime navigators. In contrast, there are many studies about experience and expertise issue using eye trackers regarding in automobile and aviation areas. Konstantopoulos et al. (2010) and Paeglis et al. (2011) studied the experience and expertise of drivers considering mainly on fixation related data such as mean fixation duration, mean fixation number, total fixation percentage on AOIs (area of interest), fixation percentage on specific AOIs. In this topic, scanpath comparison method is also utilized to distinguish between novice and expert. Scanpath is considered another option for eye behavior analysis. For example, Bellenkes et al. (1997), Kasarskis et al. (2001), Schriver et al. (2008), Sullivan et al. (2011) compared the novice and expert pilots by analyzing fixation related data and scanpath. Based on existing researches, it is agreeable to introduce fixation related data, such as mean fixation number, duration and fixation percentage, and scanpath to study experience and expertise related issue.

Scanpath comparison method

‘Scanpath’, which was first mentioned by Noton & Stark (1971), can also refer to a ‘scan pattern’, ‘scan sequence’ or ‘gaze sequence’, and is an eye movement sequence that consists of fixations and saccades.
There are several methods for evaluating scanpaths depending on the user’s research emphasis. The Levenshtein distance method, for example, is one useful metric to indicate spatial difference of scanpaths. It converts sequence of points of fixation into a string and measures the minimum number of operations needed to transform the string into the other (Biswas & Robinson, 2009). Another common method is the attention map, which calculates on the basis of a set of points (raw data sample or fixations) with no order. Theoretically, it is the superposition of a Gaussian function centered on these points. The result is a Gaussian landscape, or a ‘heat map’ if represented in color. Since each scanpath can form its own attention map, it is possible to compare scanpaths by overlapping pairwise or group-wise attention maps. Comparing the scanpaths with attention map was first described by Pomplun et al. (1996) when they researched disambiguating complex visual information. The method was popularized by Wooding (2002b) who used the attention map method to examine the large number of eye-movement traces and discussed their application to the quantification of trace similarities.
To date, there have been a large number of methods to calculate the similarity between part-wise and groupwise attention maps. For instance, Wooding (2002a) designed the simple subtraction method to quantify the similarities between part-wise attention maps. With a critical threshold value of d, the sum coverage would be the similarity. Ouerhani et al. (2004) proposed a correlation coefficient to calculate the attention map similarity and also publicized the algorithm that could transform the fixation points to a human attention map. The study showed the relationship between fixation distribution patterns and common visual search behaviors. Besides that, Rajashekar et al. (2004) and Nyström et al. (2004) raised the idea of KullbackLeibler divergence to compare scanpath similarities; Caldara & Miellet (2011) proposed a new method named the iMap statistical fixation mapping of eye movement data and comparison of scanpaths; Grindinger et al. (2010) proposed a method on group-wise scanpath comparison based on expert scanpath and can be visualized by heat map. Inspired by Grindinger’s method (Grindinger et al., 2010), a new scanpath comparison method is proposed that can effectively evaluate novice performance in group-wise. It is introduced in detail in Section 3.

Methodology

According to Section 2.1, several fixation metrics including fixation percentage, mean fixation number and duration are thus applied for analysis of eye behavior in marine operation training. In addition, a new scanpath similarity comparison method inspired by Grindinger’s method (Grindinger et al., 2010) is proposed to fit marine operation training. The goal of this method is to evaluate the performance of novice operators based on the expert operators’ eye-movement data.
The expert zone, can be easily understood as the intersection areas of two or more expert operators’ fixations. Basically, each expert’s fixations are first clustered with clear boundary according to its spatial distribution. Then the intersected area of two or more expert fixation boundaries is extracted and named the expert zone. Figure 1 shows a schematic example of the establishment of expert zone in three-expert condition. The boundary of each expert’s fixations may be depicted as a complex polygon. Here region 1 with blue boundary is the expert zone. Besides expert zone region 1 in Figure 1, it is also emphasized that the two-expert areas from regions 2 to 5; the single expert area from region 6 to region 10. These areas are also critical when evaluating novice operator performance.
Once experts’ fixated boundary has been determined, the distribution of novices’ fixations, i.e., which regions the fixations belong to, can be identified. In this way, the percentage of novices’ fixations in the ‘Expert zone’ and expert fixated areas can be calculated for eye behavior evaluation of the novice operators. To study this issue in a general way, when there are N experts as reference to evaluate the performance of novice operators, each fixation from novice in the expert zone is given 1 point, each fixation in n-expert (n < N) intersected areas but not in expert zone is given n/N points, then summarizing all the points, dividing by total fixations of novices, their performance score is calculated. Table 1 summarizes points in different regions.
Based on the explanation above, the score formula for novice operators is proposed:
Jemr 12 00019 i004
where m is the total AOI number; N is the total expert number; n is a positive integer always less than or equal to N; Pni are fixations in n-expert intersected areas but not in the expert zone of a novice in AOIi; P are total fixations of a novice in AOIi; ωi are weight number in AOIi.
Weight number is considered in (1) since there are different priorities for fixations in different AOIs. The fixations in critical AOIs are more valuable than those in other AOIs. In such a way, one can evaluate each novice operator’s performance in all defined AOIs. Furthermore, it is possible to extract critical information from the analyzed data to figure out the pattern modes of eye behavior in marine operation training.

Experiment

A marine crane-lifting experiment in marine operation simulator was conducted using eye trackers to study the difference in eye behavior between expert and novice marine operators using the aforementioned methods.

Goal

For this marine crane-lifting experiment, there are two main goals. The first is to study the visual behaviors of operators and conclude the rules behind them by comparing similar and dissimilar features between novice and expert operators. The second is using the aforementioned evaluation method to prove its feasibility in real cases. The marine crane lifting experiment is an example to study the implementation of this methodology.

Participants

Ten male marine crane operators participated in this study. All of them had experience in marine operation and were divided into two groups. The novice group had six participants with experience ranging from 4 to 6 years (m = 5.16, SD = 0.98). The expert group had four participants who had experience ranging from 8 to 10 years (m = 9.00, SD = 1.15). They had the same schedules for this experiment except that some participants from the novice group did some pre-practice in the marine simulator. The aim is to let participants get familiar with the devices. Since the pre-practice scenario is totally different from the experiment, it will not affect the experimental result. Considering the incompleteness of data samples for some participants, the main experimental data is composed of the data from four novice participants and two expert participants.

Apparatus

To measure the eye movements of the participants, we used Tobii Pro 2 glasses to collect data. To analyze the data, we mainly utilized Tobii studio analysis software and MATLAB.

Scenario

An offshore construction vessel with a knuckle boom crane is used in the experiment. The scenario is to lift a suction anchor (weight 80 tons, height 20 meters, diameter 5.3 meters) from the deck of the vessel to the seabed at a depth of 100m. The whole operation took each participant around 30 minutes.

Questionnaire

After finishing the experiment, every participant operator was required to fill in a questionnaire to give feedback on the finished operation. This questionnaire is of importance to define the size and shape of the AOIs, and also for deciding the weight number proposed in the evaluation method.
The questionnaire includes the following questions:
  • Where are the crucial areas during the experiment?
  • Where is your visual focus when the load is in the air?

AOI definition

In this experiment, to analyze and study with high efficiency, two segments were selected. Target stage one can be described as ‘descending the load until total immersion’, as shown in Figure 2; target stage two can be described as ‘total underwater period of the load’, as illustrated in Figure 3.
Then the attention maps of all participant operators using the software Tobii Pro Lab are plotted, and the main focus was given to the attention map of expert. After superposing attention maps of all experts in the target stages, the peak of this Gaussian landscape model was cut manually by referring to the questionnaires done by the participants to decide the AOIs. To decide where to cut the peak of the Gaussian model, the questionnaires written by the participant operators were taken as the reference for deciding the AOIs. Figure 4 shows the results of AOI definition for target stage one and two, in which ‘Monitor 1’ displays the sensor data about the underwater situation, such as rope paying out and tension force; ‘Monitor 2’ is from the camera mounted on the crane-tip, which visualizes the tension of crane rope during operation; ‘Swing of load and sinking process’ covers the potential area with high risk of collision before it totally submerges into the water; and ‘Crane tip’ is the stress concentration area that determines operation safety in this experiment.

Results

To study the experience and expertise, following by related works as mentioned above, fixation related data and scanpath are mainly measured and analyzed. For fixation related data, mean fixation number, total fixation percentage on AOIs and fixation percentage on specific AOIs are introduced.

Mean fixation number and total fixation percentage on AOIs

Mean fixation number is the parameter that can indicate how many fixation times one has over a defined period of time. The total fixation percentage on defined AOIs is also introduced, which is the parameter that indicates the percentage of ‘effective’ fixation through the whole operation.
For target stage one, Figure 5a shows two of the novice operators fixated most among all participants with mean fixation number 1.85 and 1.81, respectively; the rest two novice operators have a lower number of 0.83 and 0.62, respectively. The expert operators have the mean fixation number with 1.69 and 1.56 where these numbers are in the middle level among all participants. For fixation percentage on AOIs, all of the participants performed high percentage on AOIs ranging from 83% to 90% except for ‘Novice 2’ with only 50% fixation percentage. Expert operators and novice operators share similar percentage in this stage.
For target stage two, Figure 5b shows that both the experts obtained high mean fixation number. The novices except ‘Novice 3’ obtained relatively low mean fixation number ranging from 0.52 to 0.97. In addition, both the experts and the novices except ‘Novice 1’ performed high fixation percentage on defined AOIs.
In these two stages, it can be seen that although the experts do not always have the highest mean fixation number and fixation percentage compared to novice operators, they have a relatively stable eye movement in terms of mean fixation number and fixation percentage.

Fixation percentage on specific AOIs

In this part, fixation percentages on specific AOIs is calculated and analyzed. Different from the total fixation percentage mentioned above, studying the fixation percentage on specific AOIs can better indicate how the operator allocates their spatial fixation on each AOI.
Figure 6a shows the fixation percentage in each AOI for target stage one. It can be seen that ‘Swing of load and sinking process area’ has the highest fixation percentage for all operators. Even there are three novices obtained higher fixation percentage than that of the experts. This is reasonable because from the questionnaire, all participants wrote that the swing of load is quite worth fixating in this target stage. All participants except ‘Novice 1’ fixated on ‘Monitor 1’ area in this stage and ‘Expert 1’ obtained the highest fixation percentage up to 35%. Four of the participants fixated on ‘Crane-tip’ area but with a relatively low fixation percentage. From Figure 6a, the experts tend to allocate fixations among the three AOIs; while the novices pay less attention on ‘Monitor 1’ and ‘Crane-tip’ but fixate more on ‘Swing of load’ area.
Figure 6b illustrates the result of fixation percentage on ‘Monitor 1’ and ‘Monitor 2’ in target stage two. All participants put more effort on ‘Monitor 1’ and only ‘Novice 2’ allocates fixations evenly between the two AOIs. In fact, after the load fully sinking into the sea, there are no target on the sea surface. The two AOIs play important roles in this case, where ‘Monitor 1’ displays the sensor data about the underwater situation and ‘Monitor 2’ visualizes the tension of crane rope during operation. Therefore, it is reasonable to fixate most of time on ‘Monitor 1’ and only keep necessary focus on ‘Monitor 2’ as well.

Mean fixation duration and its standard deviation

In this section, mean fixation duration with its standard deviation of participant operators in target stages are measured and visualized. Normal distribution of the mean fixation duration is used to better understand the eye behaviors differences between experts and novices.
Figure 7a depicts that the mean fixation duration ranges from 226 ms to 471 ms in target stage one. The experts tend to perform a longer fixation duration with a wide standard deviation. The novices, however, have totally different distributions. In particular, ‘Novice 1’ and ‘Novice 3’ perform a shorter fixation duration with a narrow standard deviation, which shows different pattern compared to that of the experts.
The mean fixation duration for target stage two is shown in Figure 7b, ‘Novice 1’ and ‘Expert 1’ perform the shortest and longest mean fixation duration for 340 ms to 838 ms in this stage, respectively. The average duration is longer compared to that in target stage one. The reason is that in target stage two, the task is relatively easy, and it is unnecessary for the participants to switch between the two AOIs. The standard deviation ranges from 388 ms to 1816 ms, which is also larger than target stage one. It is hard to distinguish the experts and the novices from Figure 7b, because of the huge difference in standard deviation between each participant in this stage.

Scanpath analysis

In this part, scanpath analysis is conducted. Considering full time period scanpath may cause complex visualization data, we use ‘critical area’ to simplify the scanpath from both spatial and temporal aspects. The ‘critical area’ is based on the questionnaires. As a result, ‘Swing of load’ is selected as the ‘critical area’ for target stage one, while for target stage two, the ‘critical area’ is the oneminute time period when the load reach the seabed.
From Figure 8a, by combining with the ‘critical area’ in target stage one, it would be seen that experts fixate mostly in ‘Swing of load’ and move with bottom of the load as it descends. In contrast to experts’ scanpaths, Figure 8b shows that some novices tend to fixate more on ‘Crane tip’ while some novices care more on ‘Swing of load’; the scanpaths of novices are more scattered, which implies the novices have their own visual focus and strategies.
Regarding scanpath in target stage two, Figure 9a shows that experts allocate most of their fixations between ‘Monitor 1’ and ‘Monitor 2’. It is observed that they also focus on the sinking point of the load at the center of the screen. For novices, from Figure 9b, they also fixate on the two monitors and the sinking point back and forth. However, because some novices have focused on the ‘Crane tip’ that the experts do not pay attention to in target stage two, it is considered a slight behavioral difference compared to the experts.

Scanpath evaluation

The proposed evaluation method in Section 3 is used here to evaluate and compare eye behavior of the novices and experts during the training.
The fixation points within the defined AOIs are considered. In order to establish expert zone, the boundary of each expert’s fixation around the defined AOIs is calculated, followed by an exclusion of non-overlapping area. Figure 10 shows the expert zones for the two of the target stages. It is obvious that the generated expert zones are irregular polygons, and smaller than the areas of the defined AOIs.
Next, as far as the distribution of novices’ fixations is known, the percentage of their fixations in ‘Expert zone’ are calculated. In addition, the fixations located within the boundary of a single expert’s fixations but outside the ‘Expert zone’ are also evaluated. All of the fixations are associated with different weight numbers that are determined on the basis of the questionnaires by participants. In this experiment, the weight numbers for ‘Swing of load’, ‘Crane tip’ and ‘Monitor 1’ are set to 0.7, 0.2 and 0.1, respectively, in target stage one; the weight numbers for ‘Monitor 1’ and ‘Monitor 2’ are assigned to 0.7 and 0.3 in target stage two. Eventually, according to Equation (1), score of novice operators can be calculated.
Table 2 lists the score of the novices’ eye behavior in target stage one. ‘Novice 1’ obtains the highest score of 0.77 and the rest novices have a similar result around 0.66. It is revealed from Table 2 that ‘Novice 1’ has fixations on ‘Crane tip’ and thus obtains extra score. This indicates the eye behavior of ‘Novice 1’ is more similar to that of experts shown in Figure 8a. In addition, it is noted that the majority of the scores comes from the ‘Swing of load’ AOI. From the score distribution in ‘Expert zone’, all novices have relatively high percentage of fixation on this AOI with the highest weight number of 0.7 in this stage. This implies setting the high weight number on this AOI is reasonable.
The scores for novices’ eye behavior in target stage two is listed in Table 3. There is large variation among them that the highest score is up to 0.88 while the lowest one is only 0.54. It is interesting to see that the majority of their scores derive from ‘Monitor 1’ area; the fixation percentage outside the expert zone of ‘Monitor 1’ is high and contribute most of the scores.

Discussion

This section discusses the relationship between operation time, experience and task difficulty, summarizes eye behavior statistics for both experts and novices, extracts scan modes of experts in the two target stages, and prospects how the proposed method could be applied in training programs.
From Figure 4, we can see that target stage one is much more difficult compared to target stage two, since more areas need to be focused during the operation. The expert group performed more efficiently than the novice group (178 ± 16 s vs. 204 ± 41 s) in this stage. While in target stage two, both groups have a similar operation time (986 ± 136 s vs. 1020 ± 95 s). The result implies the operation time depends not only on the degree of familiarity of the device but also on the difficulty level of the task.
The expert group in the experiment have a stable scan mode. Regarding the mean fixation number in Figure 5, one can see from the plot of the target stages, experts do not always have the highest mean fixation number in every stage; whereas they are moderately stable and high in the two target stages. In addition, experts have relatively high and stable total fixation percentage on the AOIs of the two target stages, as well as similar fixation percentages on specific AOIs. This indicates the correctness of the defined AOIs in Figure 4. When considering mean fixation duration and its standard derivation, experts do not have obvious features that can separate from novices. Different operators have certain strategies, which leads to similar mean fixation duration and standard derivation features in the two target stages. However, in terms of novices, they share an unstable mean fixation number distribution in target stages and so do the total fixation percentage on the AOIs and fixation percentages on specific AOIs.
Regarding the scanpaths of all the participants, we summarize the scan modes of experts in the two target stages. The scanpaths of experts are distinguished from these of novices in that experts fixate more frequently on the ‘Swing of load’, ‘Monitor 1’ and ‘Monitor 2’ areas. More specifically, for target stage one, experts conduct a wider range of observations, that is, to scan left and right, to up and down in ‘Swing of load’ AOI to ensure the safety of the load, as shown in Figure 11a; for target stage two, Figure 11b depicts that experts switch focus between ‘Monitor 1’ and ‘Monitor 2’ areas frequently, and occasionally keep eyes on the water surface to observe the rope tension.
The eye behavior for novices in the experiment varies from person to person. Some novices, such as ‘Novice 4’, can perform similar eye behavior such as long mean fixation duration and standard deviation to that of ‘Expert 1’ in the two target stages in Figure 7; whereas some novices like ‘Novice 3’, have obvious difference compared to ‘Expert 1’. From Table 2 and Table 3, it is concluded that using the statistical metrics on eye tracking data indeed distinguishes the eye behavior to some extent, even though different participants have different scan strategies during marine operation training.
The proposed eye behavior evaluation method is also promising for improving the training program of marine operations in two aspects. First, the AOIs, the ‘critical area’ and the extracted scan patterns could be emphasized in the briefing phase, to guild the trainees to grasp the key skills of the operation. Second, in debriefing phase, a comprehensive analysis of the evaluation results, especially the fixation percentage, the mean fixation number and duration, and scanpath, will be served as the basis to help the trainees to reflect on why the deviations happen and how to avoid them. As a result, the method will accelerate the training progress without losing quality.

Conclusion

This paper presents a new analysis approach for evaluating situation awareness in marine operation training. A term ‘Expert zone’ representing the similarity of fixations among experts in a form of overlapping AOI is introduced. According to the questionnaire from experts, the AOIs and their weight numbers are recognized. A scanpath evaluation for novices is thus established to quantify how similar the eye behaviors of novices compared to these of experts.
A pilot study of crane lifting experiment was carried out in simulator. Two target stages of the operation were selected. A comprehensive analysis on different eye tracking metrics together with the scanpath evaluation is conducted between both novices and experts. The results show the effectiveness of the proposed method in assessing eye behavior for marine operation training.
For future work, efforts will be put into (1) involving more participants in the same experiment and applying statistical hypothesis tests for better understanding of eye behaviors between novices and experts in marine operation training; (2) refining the scanpath comparison method by integrating more eye tracking metrics; and (3) conducting other types of marine operation experiments to verify the scope of adaptation of the proposed method.

Acknowledgments

The research is supported in part by a grant from RFF Midt-Norge “On-Board Augmented Simulation” (Project nr: 283856), and in part by a grant from the Research Based Innovation “SFI Marine Operation in Virtual Environment (SFI-MOVE)” (Project nr: 237929) in Norway.

Ethics and Conflict of Interest

The authors declare that the contents of the article are in agreement with the ethics described in http://biblio.unibe.ch/portale/elibrary/BOP/jemr/ethics.html and that there is no conflict of interest regarding the publication of this paper.

References

  1. Atik, O., and O. Arslan. 2019. Use of eye tracking for assessment of electronic navigation competency in maritime training. [Google Scholar] [CrossRef]
  2. Balk, S. A., K. S. Moore, J. E. Steele, W. Spearman, and A. Duchowski. 2006. Mobile phone use in a driving simulation task: Differences in eye movements. J Vis 6, 6: 872. [Google Scholar] [CrossRef]
  3. Bellenkes, A. H., C. D. Wickens, and A. F. Kramer. 1997. Visual scanning and pilot expertise: The role of attentional flexibility and mental model development. Aviation, space, and environmental medicine 68, 7: 569–579. [Google Scholar] [PubMed]
  4. Biswas, P., and P. Robinson. 2009. Modelling perception using image processing algorithms. In Proceedings of the 23rd British HCI Group Annual Conference on People and Computers: Celebrating People and Technology. British Computer Society, September, pp. 494–503. [Google Scholar]
  5. Caldara, R., and S. Miellet. 2011. iMap: A novel method for statistical fixation mapping of eye movement data. Behavior research methods 43, 3: 864–878. [Google Scholar] [CrossRef] [PubMed]
  6. Chetwood, A. S., K. W. Kwok, L. W. Sun, G. P. Mylonas, J. Clark, A. Darzi, and G. Z. Yang. 2012. Collaborative eye tracking: A potential training tool in laparoscopic surgery. Surgical endoscopy 26, 7: 2003–2009. [Google Scholar] [CrossRef]
  7. Dehais, F., V. Peysakhovich, S. Scannella, J. Fongue, and T. Gateau. 2015. Automation surprise in aviation: Real-time solutions. In Proceedings of the 33rd annual ACM conference on Human Factors in Computing Systems. ACM, April, pp. 2525–2534. [Google Scholar] [CrossRef]
  8. Fisher, D. L., A. K. Pradhan, A. Pollatsek, and M. A. Knodler, Jr. 2007. Empirical evaluation of hazard anticipation behaviors in the field and on driving simulator using eye tracker. Transportation Research Record 2018, 1: 80–86. [Google Scholar] [CrossRef]
  9. Grindinger, T., A. T. Duchowski, and M. Sawyer. 2010. Group-wise similarity and classification of aggregate scanpaths. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications. ACM, March, pp. 101–104. [Google Scholar] [CrossRef]
  10. Hareide, O. S., and R. Ostnes. 2017. Scan pattern for the maritime navigator. TransNav: International Journal on Marine Navigation and Safety of Sea Transportation 11. [Google Scholar] [CrossRef]
  11. Jones, D. G., and M. R. Endsley. 1996. Sources of situation awareness errors in aviation. Aviation, space, and environmental medicine. [Google Scholar]
  12. Kang, Z., and S. J. Landry. 2014. Using scanpaths as a learning method for a conflict detection task of multiple target tracking. Human factors 56, 6: 1150–1162. [Google Scholar] [CrossRef]
  13. Kasarskis, P., J. Stehwien, J. Hickox, A. Aretz, and C. Wickens. 2001. Comparison of expert and novice scan behaviors during VFR flight. In Proceedings of the 11th international symposium on aviation psychology. March. Vol. 6. [Google Scholar]
  14. Konstantopoulos, P., P. Chapman, and D. Crundall. 2010. Driver’s visual attention as a function of driving experience and visibility. Using a driving simulator to explore drivers’ eye movements in day, night and rain driving. Accident Analysis & Prevention 42, 3: 827–834. [Google Scholar] [CrossRef]
  15. Kovácsová, N., C. D. D. Cabrall, S. J. Antonisse, T. De Haan, R. van Namen, J. L. Nooren, R. Schreurs, M. P. Hagenzieker, and J. C. F. de Winter. 2018. Cyclists’ eye movements and crossing judgments at uncontrolled intersections: An eyetracking study using animated video clips. Accident Analysis & Prevention 120: 270–280. [Google Scholar] [CrossRef]
  16. Law, B., M. S. Atkins, A. E. Kirkpatrick, and A. J. Lomax. 2004. Eye gaze patterns differentiate novice and experts in a virtual laparoscopic surgery training environment. In Proceedings of the 2004 symposium on Eye tracking research & applications. ACM, March, pp. 41–48. [Google Scholar] [CrossRef]
  17. Li, G., E. Homlong, and H. Zhang. 2017. Integration of visual focus into marine operation simulator for behavior observation and analysis. In OCEANS 2017-Aberdeen. IEEE, June, pp. 1–7. [Google Scholar] [CrossRef]
  18. Luo, M., and S. H. Shin. 2016. Half-century research developments in maritime accidents: Future directions. Accident Analysis & Prevention. [Google Scholar] [CrossRef]
  19. Miall, R. C., and J. Tchalenko. 2001. A painter’s eye movements: A study of eye and hand movement during portrait drawing. Leonardo 34, 1: 35–40. [Google Scholar] [CrossRef]
  20. Noton, D., and L. Stark. 1971. Scanpaths in eye movements during pattern perception. Science 171, 3968: 308–311. [Google Scholar] [CrossRef]
  21. Nyström, M., M. Novak, and K. Holmqvist. 2004. A novel approach to image coding using off-line foveation controlled by multiple eye-tracking measurement. Proc. Picture Coding Symp. December. Available online: http://lup.lub.lu.se/record/614983.
  22. Ouerhani, N., R. Von Wartburg, H. Hugli, and R. Müri. 2004. Empirical validation of the saliency-based model of visual attention. ELCVIA: Electronic letters on computer vision and image analysis 3, 1: 13–24. [Google Scholar] [CrossRef]
  23. Paeglis, R., K. Bluss, and A. Atvars. 2011. Driving experience and special skills reflected in eye movements. In Infrared Sensors, Devices, and Applications; and Single Photon Imaging II. International Society for Optics and Photonics, September, Vol. 8155, p. 815516. [Google Scholar] [CrossRef]
  24. Palinko, O., A. L. Kun, A. Shyrokov, and P. Heeman. 2010. Estimating cognitive load using remote eye tracking in a driving simulator. In Proceedings of the 2010 symposium on eye-tracking research & applications. ACM, March, pp. 141–144. [Google Scholar] [CrossRef]
  25. Pomplun, M., H. Ritter, and B. Velichkovsky. 1996. Disambiguating complex visual information: Towards communication of personal views of a scene. Perception 25, 8: 931–948. [Google Scholar] [CrossRef]
  26. Rajashekar, U., L. K. Cormack, and A. C. Bovik. 2004. Point-of-gaze analysis reveals visual search strategies. In Human vision and electronic imaging IX. International Society for Optics and Photonics, June, Vol. 5292, pp. 296–306. [Google Scholar] [CrossRef]
  27. Sadasivan, S., J. S. Greenstein, A. K. Gramopadhye, and A. T. Duchowski. 2005. Use of eye movements as feedforward training for a synthetic aircraft inspection task. In Proceedings of the SIGCHI conference on Human factors in computing systems. ACM, April, pp. 141–149. [Google Scholar] [CrossRef]
  28. Sarter, N. B., R. J. Mumaw, and C. D. Wickens. 2007. Pilots’ monitoring strategies and performance on automated flight decks: An empirical study combining behavioral and eye-tracking data. Human factors 49, 3: 347–357. [Google Scholar] [CrossRef]
  29. Schriver, A. T., D. G. Morrow, C. D. Wickens, and D. A. Talleur. 2008. Expertise differences in attentional strategies related to pilot decision making. Human Factors 50, 6: 864–878. [Google Scholar] [CrossRef]
  30. Sullivan, J., J. H. Yang, M. Day, and Q. Kennedy. 2011. Training simulation for helicopter navigation by characterizing visual scan patterns. Aviation, space, and environmental medicine 82, 9: 871–878. [Google Scholar] [CrossRef]
  31. Tchalenko, J. 2009. Segmentation and accuracy in copying and drawing: Experts and beginners. Vision research 49, 8: 791–800. [Google Scholar] [CrossRef] [PubMed]
  32. Wickens, C. D., J. Goh, J. Helleberg, W. J. Horrey, and D. A. Talleur. 2017. Attentional models of multitask pilot performance using advanced display technology. In Human Error in Aviation. Routledge: pp. 155–175. [Google Scholar] [CrossRef]
  33. Wooding, D. S. 2002a. Eye movements of large populations: II. Deriving regions of interest, coverage, and similarity using fixation maps. Behavior Research Methods, Instruments, & Computers 34, 4: 518–528. [Google Scholar] [CrossRef]
  34. Wooding, D. S. 2002b. Fixation maps: Quantifying eyemovement traces. In Proceedings of the 2002 symposium on Eye tracking research & applications. ACM: pp. 31–36. [Google Scholar] [CrossRef]
  35. Xu, Y., S. Li, S. Gao, D. Tan, D. Guo, and Y. Wang. 2018. Recognition method of construction conflict based on driver’s eye movement. Accident Analysis & Prevention 113: 193–201. [Google Scholar] [CrossRef]
  36. Yang, J. H., Q. Kennedy, J. Sullivan, and R. D. Fricker. 2013. Pilot performance: Assessing how scan patterns & navigational assessments vary by flight expertise. Aviation, space, and environmental medicine 84, 2: 116–124. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of how expert zone established in three-expert condition.
Figure 1. Example of how expert zone established in three-expert condition.
Jemr 12 00019 g001
Figure 2. Target stage one: load descending from air to total immerse.
Figure 2. Target stage one: load descending from air to total immerse.
Jemr 12 00019 g002
Figure 3. Target stage two: load descending to seabed and back to water surface.
Figure 3. Target stage two: load descending to seabed and back to water surface.
Jemr 12 00019 g003
Figure 4. AOI definition for (a) target stage one, and (b) target stage two.
Figure 4. AOI definition for (a) target stage one, and (b) target stage two.
Jemr 12 00019 g004
Figure 5. Mean fixation number and total fixation percentage on AOIs at (a) target stage one, and (b) target stage two.
Figure 5. Mean fixation number and total fixation percentage on AOIs at (a) target stage one, and (b) target stage two.
Jemr 12 00019 g005
Figure 6. Fixation percentage on AOIs at (a) target stage one, and (b) target stage two.
Figure 6. Fixation percentage on AOIs at (a) target stage one, and (b) target stage two.
Jemr 12 00019 g006
Figure 7. Mean fixation duration with standard deviation at (a) target stage one, and (b) target stage two.
Figure 7. Mean fixation duration with standard deviation at (a) target stage one, and (b) target stage two.
Jemr 12 00019 g007
Figure 8. Scanpath for (a) experts and (b) novices in target stage one.
Figure 8. Scanpath for (a) experts and (b) novices in target stage one.
Jemr 12 00019 g008
Figure 9. Scanpath for (a) experts and (b) novices in target stage two.
Figure 9. Scanpath for (a) experts and (b) novices in target stage two.
Jemr 12 00019 g009
Figure 10. Defined ‘Expert zone’ in (a) target stage one, and (b) target stage two.
Figure 10. Defined ‘Expert zone’ in (a) target stage one, and (b) target stage two.
Jemr 12 00019 g010
Figure 11. Extracted scanpath mode in (a) target stage one, and (b) target stage two.
Figure 11. Extracted scanpath mode in (a) target stage one, and (b) target stage two.
Jemr 12 00019 g011
Table 1. Points in different regions.
Table 1. Points in different regions.
Jemr 12 00019 i001
Table 2. Scores of eye behavior for novices in target stage one.
Table 2. Scores of eye behavior for novices in target stage one.
Jemr 12 00019 i002
Table 3. Scores of eye behavior for novices in target stage two.
Table 3. Scores of eye behavior for novices in target stage two.
Jemr 12 00019 i003

Share and Cite

MDPI and ACS Style

Mao, R.; Li, G.; Hildre, H.P.; Zhang, H. Analysis and Evaluation of Eye Behavior for Marine Operation Training—A Pilot Study. J. Eye Mov. Res. 2019, 12, 1-14. https://doi.org/10.16910/jemr.12.3.6

AMA Style

Mao R, Li G, Hildre HP, Zhang H. Analysis and Evaluation of Eye Behavior for Marine Operation Training—A Pilot Study. Journal of Eye Movement Research. 2019; 12(3):1-14. https://doi.org/10.16910/jemr.12.3.6

Chicago/Turabian Style

Mao, Runze, Guoyuan Li, Hans Petter Hildre, and Houxiang Zhang. 2019. "Analysis and Evaluation of Eye Behavior for Marine Operation Training—A Pilot Study" Journal of Eye Movement Research 12, no. 3: 1-14. https://doi.org/10.16910/jemr.12.3.6

APA Style

Mao, R., Li, G., Hildre, H. P., & Zhang, H. (2019). Analysis and Evaluation of Eye Behavior for Marine Operation Training—A Pilot Study. Journal of Eye Movement Research, 12(3), 1-14. https://doi.org/10.16910/jemr.12.3.6

Article Metrics

Back to TopTop