Next Article in Journal
Free-Base Carboxyphenyl Porphyrin Films Using a TiO2 Columnar Matrix: Characterization and Application as NO2 Sensors
Next Article in Special Issue
Gaze-Assisted User Intention Prediction for Initial Delay Reduction in Web Video Access
Previous Article in Journal
MMP-2/9-Specific Activatable Lifetime Imaging Agent
Previous Article in Special Issue
An Informationally Structured Room for Robotic Assistance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessing Visual Attention Using Eye Tracking Sensors in Intelligent Cognitive Therapies Based on Serious Games

by
Maite Frutos-Pascual
* and
Begonya Garcia-Zapirain
DeustoTech Life [eVIDA] Faculty of Engineering University of Deusto, Avda de las Universidades 24, Bilbao 48015, Spain
*
Author to whom correspondence should be addressed.
Sensors 2015, 15(5), 11092-11117; https://doi.org/10.3390/s150511092
Submission received: 24 February 2015 / Revised: 22 April 2015 / Accepted: 27 April 2015 / Published: 12 May 2015
(This article belongs to the Special Issue HCI In Smart Environments)

Abstract

: This study examines the use of eye tracking sensors as a means to identify children's behavior in attention-enhancement therapies. For this purpose, a set of data collected from 32 children with different attention skills is analyzed during their interaction with a set of puzzle games. The authors of this study hypothesize that participants with better performance may have quantifiably different eye-movement patterns from users with poorer results. The use of eye trackers outside the research community may help to extend their potential with available intelligent therapies, bringing state-of-the-art technologies to users. The use of gaze data constitutes a new information source in intelligent therapies that may help to build new approaches that are fully-customized to final users' needs. This may be achieved by implementing machine learning algorithms for classification. The initial study of the dataset has proven a 0.88 (±0.11) classification accuracy with a random forest classifier, using cross-validation and hierarchical tree-based feature selection. Further approaches need to be examined in order to establish more detailed attention behaviors and patterns among children with and without attention problems.

1. Introduction

In recent years, the usage of video game-related content in areas, such as education, therapies and training, has risen sharply. Several studies suggest that the future of pedagogy will inevitably be linked to the proposal of combined play and learning in order to promote creativity in future generations [1]. The boom in serious games brings together the potential available invideo games, devoting it fully to the enhancement of specific abilities, skills and aptitudes in children and adults.

Moreover, the design and development of new adaptive serious games whose content changes based on user interaction make therapies, training and education more customized. These techniques provide systems with an efficient way of learning based on the users themselves, providing them with customized and personal experiences, which may increase their potential effects [2].

One of the most widely-used forms of adaptive intervention consists of helping students to complete some educational activities when they have specific difficulties proceeding on their own [3].

The purpose of this study is to explore the use of eye tracking sensors to evaluate the behavior of children in attention-related cognitive therapies based on serious games to determine the utility of eye-related data as an input biofeedback signal for attention improvement therapies.

Eye movements are a natural information source for proactive systems that analyze user behavior, where the goal is to infer implicit relevance feedback from gaze [4]. Moreover, following the eye-mind hypothesis put forth by Carpenter in 1980, there is a close link between the direction of the human gaze and the focus of attention [5], provided that the visual environment in front of the eyes is pertinent to the task that we want to study [6]. Eye tracking sensors collect information about the location and duration of an eye fixation within a specific area on a computer monitor.

In this study, normal developing children aged between eight and 12 years and with different attention skills are asked to solve a set of puzzles while their gaze patterns and interaction are recorded using an eye-tracking sensor. The recorded eye information includes the location of gaze fixation on the computer screen, the duration of fixations and saccades (the path of the eye movements), along with interaction information regarding performance during the exercise. We hypothesize that participants with better performance in the proposed exercises would demonstrate patterns of eye-movements that are quantifiably different from individuals with a weaker performance. Identification of these differences would be especially advantageous for teachers and psychologists, as this study may provide new insight into the strategies for the improvement of attention skills. Moreover, the authors would like to study the relation between gaze patterns and the degree of expertise. This will be done by determining if there are any differences between the first approach to an exercise and the subsequent ones.

This article is outlined as follows: First, the use of eye tracking sensors in the field of serious games will be studied and placed in context. Subsequently, the Materials and Methods Section will be introduced, in which the authors discuss the form and function of the data collected from the eye tracking sensor. Next, a discussion of the collected data and the approaches to data analysis are examined. Finally, the manuscript concludes with a discussion of possibilities for further research into the uses of eye tracking sensor and data as a biofeedback input to intelligent therapies.

2. Literature Review

The observation of eye-movements is not a new area of research within psychology-related fields, having been studied in depth over the last few decades [79].

Research using eye tracking sensors affords a unique opportunity to test aspects of theories about multimedia learning concerning processing during learning [10]. Moreover, the use of this approach may help in understanding where players focus their attention during game play [11], as well as how they confront unfamiliar games and software [12].

However, it was not until recently that researchers began to analyze and introduce eye tracking sensors and techniques in serious games and computer games [1315]. Games that can be controlled solely through eye movement would be accessible to persons with decreased mobility or control. Moreover, the use of eye tracking data can change the interaction with games, producing new input experiences based on visual attention [15].

Eye tracking devices have been used in the design of educational games, in terms of assessing usability based on user gaze behaviors when interacting with the game [16,17]. El-Nasr and Yan used eye tracker sensors to analyze attention patterns within an interactive 3D game environment, so as to improve game level design and graphics [18].

Kickmeier-Rust et al. focused on assessing the effectiveness and efficiency of serious games. For this purpose, they assessed these variables with gaze data and gaze paths, in order to obtain interaction strategies in specific game situations [19]. Sennersten and Lindley also evaluated the effectiveness of virtual environments in games through the analysis of visual attention using eye tracking data [20]. Johansen et al. discussed the efficiency of eye tracker sensors in assessing users' behavior during game play [21].

Józsa and Hamornik used recorded eye tracking data to evaluate learning curves in university students while using a seven hidden differences puzzle game. They used this data to assess similarities and differences in information acquisition strategies considering gender- and education-dependent characteristics [22]. Dorr et al. conducted a similar study concluding that expert and novice players use different eye movement strategies [23]. Muir et al. used eye-tracking data to capture user attention patterns and to present results on how those patterns were affected by existing user knowledge, attitude towards getting help and performance while using the educational game, Prime Club [3].

Radoslaw et al. used eye tracker sensors for assessing render quality in games. They argued that gaze-dependent rendering was especially important when immersed in serious games, where players in virtual environments played a primary role [24]. Smith and Graham and Hillaire et al. concluded that use of an eye tracker increases video game immersion, altering the game play experience [25,26].

Chang et al. developed the game WAYLAas a means to evaluate the potential to offer new interaction experiences based on eye tracking and visual attention. These authors took advantage of the popularity and arrival of more affordable eye tracker sensors [27].

Li and Zhang used eye-movement analysis to assess patients' mental engagement in a rehabilitation game. Therapists use this feedback to adjust rehab exercises to users' needs [28]. Continuing with the health-related field, Lin et al. developed an eye-tracking system for eye motion disability rehabilitation as a joystick-controlled game [29]. Vickers et al. developed a framework that integrated automatic modification of game tasks, interaction techniques and input devices according to a user ability profile [30].

Walber et al. presented EyeGrab, a game for image classification controlled by the players' gaze. The main purpose of this game was to collect eye tracking data to enrich image context information [31].

Other studies, such as those conducted by Nacke et al., evaluated the use of eye tracker sensors as an alternative way of controlling interaction with games, obtaining favorable outcomes where this challenge results in positive affection and feelings of flow and immersion [32]. Ekman et al. goes one step further, discussing the limitations of using pupil-based interaction and providing suggestions for using pupil size as an input modality [33].

Table 1 shows the experimental conditions for the most relevant articles included in this section.

3. Materials and Methods

This section presents the methodology used in this study along with participants' characteristics and the selection procedure.

3.1. Participants

The process for assessing attention was performed with a group of typically developing children. This process relies on data recorded with an eye tracking sensor. Participants were aged between 8 and 12 years, with an average age of 10.0 (SD = 1.34). Thirty-two randomly-selected participants (13 girls and 19 boys) were selected from a group of 83 volunteers by their teachers. This sample size was considered adequate for the purpose of the outlined pilot study [34].

These children live in the Basque Country, Spain, have not been diagnosed with any attention-related disorder and speak Spanish as their mother tongue. All of the participants were recruited from the Colegio Vizcaya School.

Since they were mature minors, the approval of parents or guardians was requested prior to conducting the study. This approval consisted of an informed consent following receipt of a detailed description of the study, distributed via the school's regular newsletter.

3.2. Materials

All participants in the study completed the same assessment, which consisted of a puzzle exercise with four different levels of difficulty. Users have to connect each of the four slices presented in the exercise with its corresponding part in the main image. As Figure 1 displays, all of the participants were presented with the same image for each level, and all of the elements in the user interface appeared in the same part of the screen at each level.

The main image and the slices appeared in the middle of the screen, occupying the whole display from left to right. The question stem appeared in the upper middle part of the screen. The button to advance to the next level appeared at the lower middle part of each screen. The consistent layout of the screen was intended to minimize wide eye movements.

Different levels' settings are outlined in Table 2. All of the users had a maximum pre-set time of 50 seconds to complete each of the levels. However, if they finished the level before the time ended, they could go on to the next exercise. Depending on the level, the displayed image was labeled as easy, medium or hard. Only Level 1 is displayed in color. Hard images have very similar slices and are more complicated to complete. Table 2 displays the different levels’ settings.

3.3. Devices and Technologies

All of the data for this study were collected on the same device, which was located at the children’s school outside the laboratory environment. These conditions were considered appropriate due to the nature of the system.

The set of puzzles was developed in Python [35]. The results obtained and the necessary parameters were stored in a SQLitedatabase [36]. The user interface and user interaction were developed using PyQT4 [37]. Fixation heat maps were produced based on the implementation developed by jjguy [38]. The classification process was implemented using the Scikit-learn library for machine learning in Python [39].

The puzzles were displayed on a 19-inch Lenovo monitor interface with an Acer Aspire Timeline X laptop running on Ubuntu 12.04. All of the text in the different exercises was displayed as black text against a light-grey background following normal grammatical conventions in Spanish. Images were inserted as JPEG digital pictures scaled from their original versions. Response selection and any changes were stored by monitoring the user interaction and recording eye movements with a Tobii X1 Light eye tracker sensor. Figure 2 shows the study setting while one of the participants was interacting with the system.

The eye tracker is a non-invasive sensor with remote function. Participants were not required to remove their glasses or contact lenses during the tests. Accuracy under ideal conditions is 0.5 deg of the visual angle, while the sampling rate in this study was typically 28–32 Hz. As Figure 2 displays, the Tobii X1 light sensor was located beneath the computer monitor with the headrest fastened to the front edge of the desk, monitoring the participant's head. The laptop was located behind the monitor, without interfering with the participants' field of vision.

A typical experimental trial including calibration lasted less than 20 min for each participant.

3.4. Experimental Procedure

Prior to this study, participants' teachers responded autonomously to the EDAH scale for the evaluation of ADHD in the questionnaire on children between 6 and 12 years old [40]. Farré and Narbona designed this scale based on their experience with the adapted Conners questionnaire [41]. The EDAH measures the main characteristics of ADHD and the behavioral problems that may coexist with attentional deficit. This questionnaire was used to ensure that participants did not exhibit any ADHD-related behavior.

After completing the exercises, participants themselves were asked to fill in a usability questionnaire. The usability of the system was evaluated by a user satisfaction test based on the System Usability Scale [42]. This questionnaire consists of 10 items, which were evaluated by using a Likert scale ranging from 1, strongly agree, to 5, strongly disagree. Through feedback from this questionnaire, researchers will be able to continue to adapt the system to users’ final needs.

Before completing the usability questionnaire, participants were seated in front of the eye tracking sensor to permit data collection. Users were seated opposite the center of the monitor, after adjusting the seating position to their height. Once they were aligned with the screen, the calibration process started, which took between 2 and 5 min per child. This calibration entails a visual target that moves around the screen. Participants were asked to follow this target with their gaze for a period of time. The target consists of a calibration grid with 5 positions, one on each corner of the screen and the last one right in the screen’s center. The target consists of different calibration bullet points that appeared one after the other in the same order for all participants, starting from the top left corner.

Prior to the start of the exercises, participants were told in which kind of tasks they were taking part. They were also introduced to the eye tracking technology, and the sensor functionality was explained.

Participants used the system and filled in the questionnaire in a controlled environment, with a researcher observing and keeping track of all of the behavioral aspects of the study, but not interfering in the experimental setting.

3.5. Data Analysis, Processing and Classification

Recorded gaze data during the exercises has been processed, analyzed and used in order to identify the set of features that may help to build a classifier, as shown in Figure 3.

This section will explain in detail the different steps involved in the data analysis and feature identification process, so as to contribute to the core of intelligent therapies based on visual attention and user interaction.

3.5.1. Eye Fixation Parameters

The analysis of fixations and saccadic movements during the performance of certain tasks is related to attention in various ways. Several studies support this hypothesis [43-45], concluding that oculomotor mechanisms rely on attention for some aspects of eye movement control [46].

During the performance of the study, raw gaze data were recorded with the eye tracking sensor. These raw gaze data were stored as .xml files in the system, with information related to the level of the exercise that was currently running.

Listing 1 shows the stored gaze data for each participant and exercise. These data consist of the (x, y) coordinates recorded by the eye tracking sensor, the timestamp in which they were perceived, the pupil size for each eye and the exercise; the level and the mode the coordinates belong to were also stored for matching the raw gaze data with other interaction recordings.

These raw data were used for analysis and processing so as to obtain meaningful information about eye fixation locations, fixation durations, saccades and saccadic durations. Fixations are the period of time when the eyes remain fairly still and new information is acquired from the visual array [9], while saccades are the eye movements themselves. During saccades, no information is retrieved by the brain, since vision is suppressed under most normal circumstances [47].

Listing 1: Raw gaze data example

1<user date = "2014–05–22" id = " 12" sessionid = "899" time = "09:20:15" >
2 <exercise id = "puzzle" level = "Level5" mode="performance">
3  <eyedata>
4   <timestamp time = " 1167610729217997 ">
5    <left_eye pupil_diam = "3.0435333252" validity = "0" x=" 134.831732304" y=" 64.8299084174"/>
6    <right_eye pupil_diam = "2.89215087891" validity = "0" x=" 69.9537996816" y=" 59.426052033"/>
7   </timestamp>
8   …
9  </ eyedata>
10 </ exercise>
11</user>

In order to detect the saccades and fixations, some processing techniques need to be applied to the raw data file. These steps are based on the Tobii I-VTfixation filter algorithm [48], have all been implemented in the Python programming language and are outlined in Figure 4.

As Figure 4 shows, the first step in the processing algorithm is to apply the gap fill-in interpolation function. This step consists of filling in data where data are missing due to tracking problems that are not related to participants’ behavior (such as blinks or when the user looks away from the screen). In order to distinguish between tracking problems and users’ behavior, a max gap length is set, which limits the maximum length of the gap to be filled in. Following Tobii’s white paper for the I-VT fixation filter and the value used by Komogortsev, this value was set at 75 ms [48,49].

After the gaps are filled in, the noise reduction function is applied. This function is based on a low-pass filter, which aims to smooth out the noise. The third step is the velocity calculator, which relates each sample with its velocity, in terms of visual angle (degrees per second). In order to reduce the impact of noise, the velocity for each sample is calculated as the average velocity of a period of time, taking as the central data input the current sample. This is done using a window length of 20 ms, which, according to the literature, has been found to handle a reasonable level of noise without distorting the signal [48].

The I-VT classifier applied to the signal is based on the one described by Komogortsev et al. [49] and outlined in the Tobii white paper [48]. The classifier determines which samples belong to a saccade, fixation or gap, based on a velocity threshold and the angle velocities calculated in the previous step. It also groups together consecutive samples using the same classification. The velocity threshold is set to 30 deg/s [48,50].

The merge fixations function aims to merge adjacent fixations that have been split up. This is done taking into account two different thresholds, the max-time between fixations, which is set to 75 ms [48], that is lower than the normal blink duration [49,51,52], and the max-angle between fixations, which is set at 0.5 deg [48,49,5355]

Once all of the fixations have been identified, the shorter ones are removed. For the purposes of this analysis, 100 ms was set as the lower limit for fixation duration. This value was chosen based on the work of McConkie et al., who concluded that 60 ms must pass before current visual information becomes available to the visual cortex for processing [56]. R. Tai et al. arrived at the lower limit of 100 ms by adding 30 ms, which is the time that elapses, at the end of a fixation, between when a command to move the eyes is sent and the onset of that saccade is reported. They allowed also 10 ms for the processing of any currently-observed stimuli, arriving at the 100-ms threshold [57].

After all of the processing functions have been applied to the current data, a new gaze data file is created with all of the fixations for the current exercise and participant. As shown in Listing 2, fixation data have a similar structure to raw data.

Listing 2: Fixation data example

1<user date="2014–05–22" id="12" sessionid="897" time="09 :15:27 ">
2 <exercise id=" puzzle " level=" Level1 " mode=" performance ">
3  <fixationData>
4   <f ixation >
5    <time duration =" 212.356933594 " end_time=" 1.1676104137 e+12" start_time =" 1.16761041349 e+12">
6     < position x=" 54.2251062717 " y=" 130.508713537 " />
7    </ time>
8   </ fixation>
9   …
10  </ fixationData>
11 </ exercise >
12</ user>

The stored fixation data save all of the fixations recorded during the exercise, along with the current activity information, user data and the duration, start time, end time and position of each fixation.

3.5.2. Outlier Detection Process

Once the processing stage is over, the fixation data are used to determine the outliers among the recorded data. This process is outlined in Figure 5.

As Figure 5 shows, a fixation heat map is created for each file. The fixation count heat map shows the accumulated number of fixations for each puzzle level and for each participant. Each fixation made adds a value to the color map at the location of the fixation [58].

The alpha layers of the images stored are then analyzed as a measure to identify the location and amount of fixations and saccades. All of the images are the same size and dimensions.

The alpha information per image is stored in order to be processed by the median absolute deviation (MAD) algorithm for outlier detection implemented in Python.

The median deviation is a measure of scale based on the median of the absolute deviations from the median of the distribution [59]. The formula is shown in Equation (1).

MAD = median i ( | X i median j ( X j ) | )

Moreover, the heat maps were analyzed taking into account users’ overall performance during the entire study, so as to have another feature to determine outlier detection.

3.5.3. Classification

This section outlines the first steps taken in the classification process. The aim of this part is to assess the feasibility of using a set of combined features to evaluate user performance. These features are related to user interaction, timing and visual attention, as well as image-related data obtained directly from the heat maps.

This part explains the theoretical insights taken in this process. Please refer to the same section in the Results part for the mathematical outcomes of this process.

3.5.3.1. Feature Identification

Feature selection is a determining factor when classifying patterns. Features need to be insensitive to noise and separated from each other. Their main purpose is to objectively describe certain aspects, in this case of the attention and performance process in intelligent therapies aimed at children.

A collection of 34 features was selected based on image characteristics and user performance related to the current exercise. Features were selected based on the recorded data. The authors, in conjunction with the multidisciplinary team taking part in this project, took into consideration performance variables, as well as gaze pattern recordings. The subset of selected features for analysis from the pilot phase is outlined in Table 3.

Heat maps were divided into 9 quadrants in order to obtain detailed data about the location and density of fixations per participant and level.

The selected features were chosen for further analysis and consideration, so as to determine if they are suitable for use in an automatic classifier, capable of discerning the users' performance based on their interaction and gaze patterns.

3.5.3.2. Feature Selection

Feature selection creates a subset of features, improving their predictive performance and constructing patterns more efficiently. This helps to avoid multidimensionality, which may otherwise have an adverse effect on the decision making process [60].

Several techniques were used in this process. In order to assess the success rate of the classifier while obtaining the most accurate set of features, a set of different ensemble classifiers was used and compared with a traditional decision tree classifier.

  • Sequential search: This process works by selecting the best features based on univariate statistical tests [39]. Inside this topic, the select k-best feature selection algorithm was applied. This process removes all but the k highest scoring features.

  • L1-based feature selection: This was applied to assess the feasibility of discarding the zero coefficients. This is a means of reducing the dimensionality of data [39].

  • Hierarchical feature selection: In these feature selection processes, the set of features is divided into smaller subsets until only one remains in each node [61]. Tree-based estimators were applied to compute feature importance, so as to discard the irrelevant ones [39].

3.5.3.3. Classifier Performance Analysis

Ensemble learning algorithms works by running a base learning algorithm multiple times, voting out the resulting hypotheses [62]. Ensemble learning has received an increasing interest recently, since it is more accurate and robust to noise than single classifiers [63,64].

This article compares the performance capabilities of 3 different ensemble algorithms when they are applied to the real dataset recorded in this study. The aim of this experiment is to assess the feasibility of building a classifier able to determine user performance using an adequate set of features of a different nature recorded during the therapy.

All of the classifiers were evaluated using cross-validation. The studied classifiers were:

  • Random forest: This classifier is defined as a combination of tree predictors. Each tree depends on the values of a random vector sampled independently and with the same distribution for all trees [65]. Using the random selection of features yields error rates that compare favorably to AdaBoost [66], but are more robust with noise handling [65].

  • Extremely randomized trees: A tree-based ensemble method for supervised classification and regression. It is a strongly randomized attribute selection method. This algorithm is accurate and computationally efficient [67].

  • AdaBoost: This algorithm is an iterative procedure that tries to approximate the Bayes classifier by combining several weaker classifiers. A score is assigned to each classifier, and the final classifier is defined as the linear combination of the classifiers from each stage [68].

Moreover, a regular decision tree classifier was applied in order to assess the potential and improvement in accuracy, if any, of the previously mentioned tree-based ensemble methods.

4. Results

The recordings for the results explained in this section were taken during the month of May, 2014, at the Colegio Vizcaya school in Biscay, Spain.

4.1. Analysis of User Performance: Outcome Scores and Response Times

Although the present study is focused on the use of gaze data to analyze performance in attention-related cognitive therapies, we feel that it is also important to address commonly-used measurements to categorize user performance in this type of exercise: outcome scores and response times. These measures might be quite general in some cases where they show only a vague impression of the user's performance.

Participants' responses were recorded through the system implemented in Python. Their overall number of correct responses, as well as their number of correct responses per level are shown in Table 4.

The overall mean of correct responses is 11.937 (SD = 2.20) out of a possible score of 16. When the results are examined by levels, there are some differences in performance between the first two levels, which participants considered much easier, and the last two, which they found more difficult.

Users had a maximum of 50 seconds to complete each exercise. However, they were able to finish the level before time ran out. Considering the response times, i.e., total time spent on test questions, the data show that the majority of participants took most of the entire time available at all of the levels.

Levels can be segmented into two groups, according to difficulty. The first two are considered the easiest ones, while the last two are trickier. There is a tendency between the two groups; users tend to perform slower with Levels 1 and 3 than with Levels 2 and 4. This may be because they tend to be more careful with novelty exercises or when the difficulty suddenly changes.

Figure 6 shows the overall performance of participants, regarding total time versus correct answers. As is displayed in Figure 6, users tend to respond correctly to more than half of the possible answers, while using 75% or more than the available time. When analyzing the group with the weakest performance, with a number of total correct answers below 10, it is clear that 75% of the participants in this group have a higher performance time.

Further analysis of user performance will be outlined in the following sections. The correct items' mean (11.93 out of 16) and standard deviation (SD = 2.20) values were used for obtaining the threshold for the weakest performers. This results in the value 9.73; since the study needs an entire threshold, this value was rounded up to 10. Participants with scores lower than this threshold were classified as the weakest performers. A total of four participants matched this criteria, so they were paired with the four best performers to obtain two balanced groups for further analysis. In order to address the research question stated in the Introduction, the four best performers (users with IDs 20, 25, 28 and 43) and the weakest four (users with Is 15, 16, 29 and 36) will be analyzed.

4.2. Fixation Heat Maps

Fixations were analyzed for each of the participants. Fixations were defined as a gaze longer than 100 ms. In order to address the research question stated in the Introduction, the most accurate and the weakest performers were selected for further analysis.

Fixations were displayed as heat maps, which were created based on the entire time participants took for each level. Red spots indicate higher levels of fixation, with yellow and green indicating decreasing amounts of fixations. Areas without color were not fixated upon. The most accurate performers are displayed in Figure 7, while the four with the weakest performance are displayed in Figure 8.

When comparing the heat maps of both groups, there are some differences between the number, density and clustering of fixations. In Figure 7, where the total score results of the participants are 15 correct answers out of 16 possible ones for every case, the number of fixations is lower than for the participants with a weaker performance. Not only is it lower among participants, it also seems to decrease when analyzing the intra-level gaze behavior for each of them.

It is important to bear in mind that an overall lower number of total fixations suggests less time spent viewing specific areas of the assessment item.

Regarding Figure 8, where the total score for these participants ranges between six and nine correct answers out of 16 possible ones, the fixation density is higher for all the cases, except for the participant with ID 29.

This hypothesis agrees with R. Tai et al., who found an inverse relation between the fixation and saccade amount and the degree of expertise of the participants [57].

4.3. Quantitative Analysis

This section includes a quantitative analysis of the data regarding various features, such as the number of fixations per level, their average duration and the gender and age of the selected subgroups of participants, in order to analyze the feasibility of establishing some behavioral patterns.

Table 5 displays the four participants with the best results. When processing the number of fixations, we observe that they decrease in number with the progression of the levels for all of the users, as displayed in Figure 9a. Since the exercise has the same visual layout for every level, this may be related to their having achieved a certain degree of expertise with each new level.

Table 6 displays the four participants with the weakest performance results. When processing the number of fixations, we find no specific relationship among them either, as displayed in Figure 9b, which may be related to the lack of appropriate techniques for solving the puzzle task.

Further analysis of the results was made on the best vs. weakest performers' data. Due to the number of users that were used for the further analysis of the results, a Mann–Whitney non-parametric test was applied. The results of the test are outlined in Table 7.

Table 7 shows that there are some significant differences (p ≤ 0.05) in performance between groups. These differences appeared in the number of fixations in Levels 2 and 4 and globally. Moreover, there are other significant differences for fixation average time (global), time (Level 4) and the number of correct answers (Levels 3 and 4). However, these results may not be enough to conclude that there are consistent differences regarding the level of expertise of the participants.

R. Tai et al. [57] and Chi et al. [69] hypothesized that fixation duration data did not produce clear and consistent differences regarding the level of expertise of the participants, which agrees with the results obtained in this section.

4.4. Classification

This section outlines the first steps taken in the classification process. The aim of this part is to assess the feasibility of using a set of combined features to evaluate user performance. These features are related to user interaction, timing and visual attention, as well as image-related data obtained directly from the heat maps.

This part explains the mathematical outcomes of this process. Please refer to Section 3.5.3 for the theoretical insights.

In order to further assess the number of optimal features for the classification part, a recursive feature elimination process with cross-validation was applied. Table 8 displays the existing relation between the number of features and the classifier's accuracy. The number of features depends on the feature selection algorithm applied. These algorithms were outlined in Section 3.5.3.2.

The accuracy results displayed in Table 8 were obtained by applying a cross-validation process of 100 iterations to all of the available data. These user data were divided as follows: 60% of the data for training and 40% for testing the classifier inside the cross-validation process.

With this setting, feature selection seems to be beneficial for building any type of analyzed classifier. However, when employing all of the available features, the accuracy rate falls below 0.80 for decision trees and AdaBoost classifiers. The select K-best features algorithm improves the classifiers' accuracy, especially when using 22 features. The L1-based algorithm displays good accuracy results for all of the ensemble methods and falls below 0.80 for the decision tree classifier. The tree-based hierarchical algorithm employed gives good results in accuracy with a limited number of features that range between 10 and 14, depending on the classifier employed.

The authors compared the accuracy performance of the selected ensemble classifiers with the overall performance of the a decision tree classifier. Since the data did not follow a normal distribution, a Mann–Whitney analysis was used. The results of comparing the performance of every ensemble classifier (with all features) with the decision tree classifier (with all features) is displayed in Table 9.

As is displayed in Table 9, the use of ensemble classifier methods significantly improves the overall performance of the classifier, regardless of the number of features employed. In the case of using all of the available features, the best classifier for the recorded data is the random forest.

Analyzing the difference in intra-classifier performance, Table 10 displays the Mann–Whitney analysis of the different feature-selection algorithms, comparing their performance with the accuracy obtained with the all features approach.

Table 10 illustrates that for almost all of the analyzed settings in this article, the use of a smaller set of features significantly improves the overall accuracy of all of the ensemble classifiers and the decision tree.

After carrying out all of the detailed experimental tests based on the recorded data, it can be concluded that accurate classification of different user performance according to their interaction and visual attention is possible.

5. Discussion and Conclusions

In the Conclusion, we intend to give an answer to the research questions outlined in the Introduction, as well as put forth new thoughts and trends about the present and future of assessing visual attention using eye tracker sensors in serious games.

According to the literature, there are several theories that link eye-movements with attentional processes [5,6], linking eye movements with cognitive processes, such as reading, visual search and scene perception. However, regarding intelligent therapies, eye movements do not always tell the whole story about the attentional process [70]. These resources should be complemented with other interaction records, as well as with relevant data about the participant. The higher the system information, the more accurate its customization to users' final needs.

In the Introduction, we hypothesized that participants with better performance may demonstrate patterns of eye-movements quantifiably different from individuals with weaker performance. Although some differences were found during the exercises, it is necessary to extend the study or to replicate it, in order to make stronger assumptions.

A comparison of the fixation duration data did not produce clear and consistent differences corresponding to the level of performance. These results corresponded with those related to the expertise level found by R. Tai et al. [57] and Chi et al. [69].

Regarding Figure 8, which shows the fixation heat maps for the weaker performers, fixation density is higher for all of the cases, except for the participant with ID 29. Moreover, the fixation density in Figure 7 decreases with the performance of new levels. These findings agree with R. Tai et al., who found an inverse relationship between the fixation and saccade amount and the participants' degree of expertise [57].

When analyzing performance data, there are some differences between the two groups for which the puzzle levels are classified into according to difficulty. Table 4 shows the performance results. When changing the exercise type or level of challenge, users tend to spend more time and perform the exercise with taking more time to think. When the tasks are repeated, the ability level increases and the time to complete them drops. This may be related to the acquisition of specific problem-solving skills, which become more accurate with repetition. Further studies need to be carried out about the users' ability and performance capabilities in repetitive tasks.

Intelligent therapies that dynamically adapt themselves to users' needs and performance based on their interaction with the system have been proven to be efficient in terms of improvement comparisons [71]. A good set of collected data may provide improved means for obtaining adapted and efficient intelligent cognitive data. Researchers should be very careful with the selected and recorded features. Several different approaches need to be followed in order to obtain the most accurate set of performance data.

Moreover, a deeper analysis of timing per exercise may also prove to be interesting for study. As a future approach, the reading instructions stage will be separated from the performance of the exercise, so that we can obtain explicit performance timing, with and without the reading stage. This could give further information about whether there are any differences between the first performance of an exercise and the subsequent ones. This new approach may also help in further assessment of attention in the performance and instruction reading stages.

Reviewing the literature, there are several studies published linking the size of the pupils with cognitive processes [7274]. Although, this response in the pupils is slow [75]. Current eye trackers measure pupil size and give it as another parameter, so it is easy to analyze this feature during the performance of tasks. This parameter was not analyzed in this study, and it may be an interesting additional feature in future research about this topic.

In recent years, the popularity of eye trackers has increased, and there are some open-source projects offering tools for gaze data analysis [7680], while some manufacturers offer low-cost devices, such as the EyeTribe [81]. There are also several DIY approaches for building custom eye trackers [8284]. The accuracy of these systems may sometimes be slightly inferior to high-end eye trackers, but they may be a viable solution for use outside the laboratory setting [85]. The use of eye trackers outside the research community may help to extend its potential with available intelligent therapies, bringing state-of-the-art technologies to users.

This study may expand in future directions, such as the design and development of the system, so that the tool includes new skills that continue along the lines of the current tool, for work on new capabilities, such as working-memory or processing speed.

Moreover, future lines should include the design and development of a robust classifier, with the selected features outlined in Section 3.5.3.1. The initial study of the classifier capabilities of ensemble methods with the available user data has produced positive results, especially when implementing a feature selection algorithm beforehand (see Section 4.4 for further information about the ensemble classifiers performance). Other classifiers need to be studied and tested, in order to consider others that may be more accurate, alone or in combination with others. This approach will help to create an autonomous system able to discern user implication based on visual attention and performance records.

Finally, some directions for the future are to replicate this study:

-

with a greater number of users;

-

with users with and without attention-related problems;

-

developing a bilingual or trilingual tool that allows the study to be replicated in other areas in Spain and abroad where reported diagnosis of attention-related problems are significantly different from the Basque Country, Spain.

The use of gaze data constitutes a new information source in intelligent therapies that may help to build new approaches that are completely customized to final users' needs. Further studies need to be carried out in order to establish more detailed attention behaviors and patterns among children with and without attention problems. The replication of this study, along with the extension of the current system with new exercises, may help to build personalized performance profiles per user. These profiles may help in creating new customized therapies, while providing a new degree of information to the children themselves, therapists, psychologists, teachers and family.

Acknowledgements

The authors would like to express their appreciation to DEIKER, the Agency for Research Management and Promotion at the University of Deusto. They would also like to thank Patricia Clemente and Mari Jose Pecharroman and their team at Colegio Vizcaya School for their support and collaboration during the study and the AHIDA association for their advice in defining the given puzzle. The authors would like to thank professors Quasim Mehdi and Georgios Paltoglou at the University of Wolverhampton for their support and advice and researcher Philip Heaton at Deustotech-Life for his help with revising this article.

Author Contributions

Maite Frutos-Pascual is a PhD candidate whose research area is intelligent cognitive therapies based on biofeedback and serious games. She designed and developed the system and carried out the data analysis. Begonya Garcia-Zapirain is her PhD supervisor, she helped with the design of the experiment and she were on charge of participants recruitment. All the authors were responsible of the final version of this manuscript and all of them contributed extensively to this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Samuelsson, I.P.; Carlsson, M.A. The playing learning child: Towards a pedagogy of early childhood. Scand. J. Educ. Res. 2008, 52, 623–641. [Google Scholar]
  2. Tobail, A.; Crowe, J.; Arisha, A. Learning by gaming: Supply chain application. Proceedings of the IEEE Proceedings of the 2011 Winter Simulation Conference (WSC), Phoenix, AZ, USA, 11–14 December 2011; pp. 3935–3946.
  3. Muir, M.; Conati, C. An analysis of attention to student—Adaptive hints in an educational game. In Intelligent Tutoring Systems; Springer: Crete, Greece, 2012; pp. 112–122. [Google Scholar]
  4. Klami, A. Inferring task-relevant image regions from gaze data. Proceedings of the 2010 IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Kittilä, Finland, 29 August–1 September 2010; pp. 101–106.
  5. Just, M.A.; Carpenter, P.A. A theory of reading: From eye fixations to comprehension. Psychol. Rev. 1980, 87, 329–354. [Google Scholar]
  6. Hyönä, J. The use of eye movements in the study of multimedia learning. Learn. Instr. 2010, 20, 172–176. [Google Scholar]
  7. Rayner, K. Eye movements in reading and information processing. Psychol. Bull. 1978, 85, 618–660. [Google Scholar]
  8. Rayner, K. Eye movements in reading and information processing: 20 years of research. Psychol. Bull. 1998, 124, 372–422. [Google Scholar]
  9. Rayner, K. Eye movements and attention in reading, scene perception, and visual search. Q. J. Exp. Psychol. 2009, 62, 1457–1506. [Google Scholar]
  10. Mayer, R.E. Unique contributions of eye-tracking research to the study of learning with graphics. Learn. Instr. 2010, 20, 167–171. [Google Scholar]
  11. Sundstedt, V.; Stavrakis, E.; Wimmer, M.; Reinhard, E. A psychophysical study of fixation behavior in a computer game. Proceedings of the 5th symposium on Applied Perception in Graphics and Visualization, Los Angeles, CA, USA, 9–10 August 2008; pp. 43–50.
  12. Pretorius, M.; Gelderblom, H.; Chimbo, B. Using eye tracking to compare how adults and children learn to use an unfamiliar computer game. Proceedings of the 2010 Annual Research Conference of the South African Institute of Computer Scientists and Information Technologists, Bela Bela, South Africa, 11–13 October 2010; pp. 275–283.
  13. Deng, S.; Kirkby, J.A.; Chang, J.; Zhang, J.J. Multimodality with Eye tracking and Haptics: A New Horizon for Serious Games? Int. J. Serious Games 2014, 1, 17–33. [Google Scholar]
  14. Almeida, S.; Veloso, A.; Roque, L.; Mealha, O. The eyes and games: A survey of visual attention and eye tracking input in video games. Proceedings of the SBGames 2011—X Brazilian Symposium on Computer Games and Digital Entertainment—Arts & Design Track, Salvador, Brazil, 7–9 November 2011.
  15. Isokoski, P.; Joos, M.; Spakov, O.; Martin, B. Gaze controlled games. Univ. Access Inf. Soc. 2009, 8, 323–337. [Google Scholar]
  16. Zain, N.H.M.; Razak, F.H.A.; Jaafar, A.; Zulkipli, M.F. Eye tracking in educational games environment: evaluating user interface design through eye tracking patterns. In Visual Informatics: Sustaining Research and Innovations; Springer: Selangor, Malaysia, 2011; pp. 64–73. [Google Scholar]
  17. Kiili, K.; Ketamo, H.; Kickmeier-Rust, M.D. Eye Tracking in Game-based Learning Research and Game Design. Int. J. Serious Games 2014, 1, 51–65. [Google Scholar]
  18. El-Nasr, M.S.; Yan, S. Visual attention in 3D video games. Proceedings of the 2006 ACM SIGCHI International Conference On Advances in Computer Entertainment Technology, Hollywood, CA, USA, 14–16 June 2006; p. 22.
  19. Kickmeier-Rust, M.D.; Hillemann, E.; Albert, D. Tracking the UFO's Paths: Using Eye-Tracking for the Evaluation of Serious Games. In Virtual and Mixed Reality-New Trends; Springer: Orlando, FL, USA, 2011; pp. 315–324. [Google Scholar]
  20. Sennersten, C.C.; Lindley, C.A. Real time eye gaze logging in a 3D game/simulation world. Key Eng. Mat. 2010, 437, 555–559. [Google Scholar]
  21. Johansen, S.A.; Noergaard, M.; Rau, J. Can eye tracking boost usability evaluation of computer games. Proceedings of the CHI, Florence, Italy, 5 April 2008.
  22. Józsa, E.; Hámornik, B. Find the difference: Eye tracking study on information seeking behavior using an online game. J. Eye Track. Vis. Cognit. Emot. 2011, 2, 27–35. [Google Scholar]
  23. Dorr, M.; Böhme, M.; Martinetz, T.; Barth, E. Gaze beats mouse: A case study. Proceedings of the 2nd Conference on Communication by Gaze Interaction, Turin, Italy, 4 September 2006; pp. 16–19.
  24. Turner, J.; Velloso, E.; Gellersen, H.; Sundstedt, V. EyePlay: applications for gaze in games. Proceedings of the First ACM SIGCHI Annual Symposium On Computer-Human Interaction in Play, London, UK, 19 October 2014; pp. 465–468.
  25. Smith, J.D.; Graham, T. Use of eye movements for video game control. Proceedings of the 2006 ACM SIGCHI International Conference On Advances in Computer Entertainment Technology, Montreal, QC, Canada, 22 April 2006; p. 20.
  26. Hillaire, S.; Lécuyer, A.; Cozot, R.; Casiez, G. Using an eye-tracking system to improve camera motions and depth-of-field blur effects in virtual environments. Proceedings of the IEEE Virtual Reality Conference (VR'08), Reno, NE, USA, 8 March 2008; pp. 47–50.
  27. Chang, W.; Shen, P.A.; Ponnam, K.; Barbosa, H.; Chen, M.; Bermudez, S. WAYLA: Novel gaming experience through unique gaze interaction. Proceedings of the ACM SIGGRAPH 2013 Emerging Technologies, San Francisco, CA, USA, 23 July 2013; p. 16.
  28. Li, S.; Zhang, X. Eye-movement-based objective real-time quantification of patient's mental engagement in rehabilitation: A preliminary study. Proceedings of the 2014 IEEE International Conference on Mechatronics and Automation (ICMA), Tianjin, China, 3–6 August 2014; pp. 180–185.
  29. Lin, C.S.; Huan, C.C.; Chan, C.N.; Yeh, M.S.; Chiu, C.C. Design of a computer game using an eye-tracking device for eye's activity rehabilitation. Opt. Lasers Eng. 2004, 42, 91–108. [Google Scholar]
  30. Vickers, S.; Istance, H.; Heron, M.J. Accessible gaming for people with physical and cognitive disabilities: a framework for dynamic adaptation. Proceedings of the CHI'13 Extended Abstracts on Human Factors in Computing Systems, Paris, France, 27 April 2013; pp. 19–24.
  31. Walber, T.; Neuhaus, C.; Scherp, A. EyeGrab: A Gaze-based Game with a Purpose to Enrich Image Context Information. Proceedings of the EuroHCIR, Nijmegen, Netherlands, 24 August 2012; pp. 63–66.
  32. Nacke, L.E.; Stellmach, S.; Sasse, D.; Lindley, C.A. Gameplay experience in a gaze interaction game. Proceedings of the 5th Conference on Communication by Gaze Interaction—COGAIN 2009, Lyngby, Denmark, 26 May 2009; pp. 49–54.
  33. Ekman, I.M.; Poikola, A.W.; Mäkäräinen, M.K. Invisible eni: Using gaze and pupil size to control a game. Proceedings of the CHI'08 Extended Abstracts on Human Factors in Computing Systems, Florence, Italy, 5 April 2008; pp. 3135–3140.
  34. Hertzog, M.A. Considerations in determining sample size for pilot studies. Res. Nurs. Health 2008, 31, 180–191. [Google Scholar]
  35. Python Programming Language. Available online: https://www.python.org/ (accessed on 9 January 2015).
  36. SQLite Database Engine. Available online: http://www.sqlite.org/ (accessed on 9 January 2015).
  37. PyQT4. Available online: http://www.riverbankcomputing.com/software/pyqt/ (accessed on 9 January 2015).
  38. Heatmaps - jjguy. Available online: http://jjguy.com/heatmap/ (accessed on 2 May 2015).
  39. Scikit-Learn. Available online: http://scikit-learn.org/stable/index.html (accessed on 3 February 2015).
  40. Farré, A.; Narbona, J.; EDAH. Escalas para la evaluación del trastorno por déficit de atención con hiperactividad. In Madrid: TEA Ediciones; TEA Ediciones: Madrid, Spain, 2001. [Google Scholar]
  41. Conners, C.K.; Sitarenios, G.; Parker, J.D.; Epstein, J.N. The revised Conners' Parent Rating Scale (CPRS-R): Factor structure, reliability, and criterion validity. J. Abnorm. Child Psychol. 1998, 26, 257–268. [Google Scholar]
  42. Brooke, J. SUS-A quick and dirty usability scale. Usabil. Eval. Ind. 1996, 189–194. [Google Scholar]
  43. Kowler, E.; Zingale, C. Smooth eye movements as indicators of selective attention. Atten. Perform. 1985, 11, 285–300. [Google Scholar]
  44. Schneider, W.X.; Deubel, H. Visual attention and saccadic eye movements: Evidence for obligatory and selective spatial coupling. Stud. Vis. Inf. Process. 1995, 6, 317–324. [Google Scholar]
  45. Klein, R.; Kingstone, A.; Pontefract, A. Orienting of visual attention. In Eye Movements and Visual Cognition; Springer: New York, NY, USA, 1992; pp. 46–65. [Google Scholar]
  46. Hoffman, J.E.; Subramaniam, B. The role of visual attention in saccadic eye movements. Percept. Psychophys. 1995, 57, 787–795. [Google Scholar]
  47. Matin, E. Saccadic suppression: A review and an analysis. Psychol. Bull. 1974, 81, 899–917. [Google Scholar]
  48. Olsen, A. The Tobii I-VT Fixation Filter; Tobii AB: Danderyd, Sweden; 20; March; 2012. [Google Scholar]
  49. Komogortsev, O.V.; Gobert, D.V.; Jayarathna, S.; Koh, D.H.; Gowda, S.M. Standardization of automated analyzes of oculomotor fixation and saccadic behaviors. IEEE Trans. Biomed. Eng. 2010, 57, 2635–2645. [Google Scholar]
  50. Olsen, A.; Matos, R. Identifying parameter values for an I-VT fixation filter suitable for handling data sampled with various sampling frequencies. Proceedings of the Symposium on Eye Tracking Research and Applications, Santa Barbara, CA, USA, 28–30 March 2012; pp. 317–320.
  51. Ingre, M.; ÅKerstedt, T.; Peters, B.; Anund, A.; Kecklund, G. Subjective sleepiness, simulated driving performance and blink duration: Examining individual differences. J. Sleep Res. 2006, 15, 47–53. [Google Scholar]
  52. Volkmann, F.C.; Riggs, L.A.; Moore, R.K. Eyeblinks and visual suppression. Science 1980, 207, 900–902. [Google Scholar]
  53. Najemnik, J.; Geisler, W.S. Optimal eye movement strategies in visual search. Nature 2005, 434, 387–391. [Google Scholar]
  54. Nakatani, C.; van Leeuwen, C. A pragmatic approach to multi-modality and non-normality in fixation duration studies of cognitive processes. J. Eye Mov. Res. 2008, 1, 1–12. [Google Scholar]
  55. Kliegl, R.; Grabner, E.; Rolfs, M.; Engbert, R. Length, frequency, and predictability effects of words on eye movements in reading. Eur. J. Cogn. Psychol. 2004, 16, 262–284. [Google Scholar]
  56. McConkie, G.W.; Underwood, N.R.; Zola, D.; Wolverton, G. Some temporal characteristics of processing during reading. J. Exp. Psychol. Hum. Percept. Perform. 1985, 11, 168–186. [Google Scholar]
  57. Tai, R.H.; Loehr, J.F.; Brigham, F.J. An exploration of the use of eye-gaze tracking to study problem-solving on standardized science assessments. Int. J. Res. Method Educ. 2006, 29, 185–208. [Google Scholar]
  58. Bojko, A.A. Informative or misleading? Heat maps deconstructed. In Human-Computer Interaction. New Trends; Springer: Berlin, Germany, 2009; pp. 30–39. [Google Scholar]
  59. Howell, D.C. Median Absolute Deviation. In Encyclopedia of Statistics in Behavioral Science; Taylor & Francisr: Hoboken, NJ, USA, 2005. [Google Scholar]
  60. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  61. Freeman, C.; Kulic, D.; Basir, O. Feature-Selected Tree-Based Classification. IEEE Trans. Cybern. 2013, 43, 1990–2004. [Google Scholar]
  62. Dietterich, T.G. Ensemble learning. In The Handbook of Brain Theory and Neural Networks; MIT Press: Cambridge, MA, USA, 2002; pp. 110–125. [Google Scholar]
  63. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar]
  64. Rodriguez-Galiano, V.; Ghimire, B.; Rogan, J.; Chica-Olmo, M.; Rigol-Sanchez, J. An assessment of the effectiveness of a random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 2012, 67, 93–104. [Google Scholar]
  65. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar]
  66. Schapire, R.E.; Freund, Y.; Bartlett, P.; Lee, W.S. Boosting the margin: A new explanation for the effectiveness of voting methods. Ann. Stat. 1998, 26, 1651–1686. [Google Scholar]
  67. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar]
  68. Zhu, J.; Zou, H.; Rosset, S.; Hastie, T. Multi-class adaboost. Stat. Its Interface 2009, 2, 349–360. [Google Scholar]
  69. Chi, M.T.H.; Glaser, R.; Rees, E. Expertise in problem solving. In Advances in the Psychology of Human Intelligence; Sternberg, R.J., Ed.; Lawrence Erlbaum Associates: Hillsdale, NJ, USA, 1982; pp. 1–75. [Google Scholar]
  70. Borji, A.; Itti, L. State-of-the-art in visual attention modeling. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 185–207. [Google Scholar]
  71. Solana, J.; Cáceres, C.; García-Molina, A.; Chausa, P.; Opisso, E.; Roig-Rovira, T.; Menasalvas, E.; Tormos-Muñoz, J.M.; Gómez, E.J. Intelligent Therapy Assistant (ITA) for cognitive rehabilitation in patients with acquired brain injury. BMC Med. Inform. Decis. Mak. 2014, 14. [Google Scholar] [CrossRef]
  72. Binda, P.; Pereverzeva, M.; Murray, S.O. Pupil size reflects the focus of feature-based attention. J. Neurophysiol. 2014, 112, 3046–3052. [Google Scholar]
  73. Sebastián-Gallés, N. Eyes wide shut: Linking brain and pupil in bilingual and monolingual toddlers. Trends Cogn. Sci. 2013, 17, 197–198. [Google Scholar]
  74. Gabay, S.; Pertzov, Y.; Henik, A. Orienting of attention, pupil size, and the norepinephrine system. Atten. Percept. Psychophys. 2011, 73, 123–129. [Google Scholar]
  75. Wierda, S.M.; van Rijn, H.; Taatgen, N.A.; Martens, S. Pupil dilation deconvolution reveals the dynamics of attention at high temporal resolution. Proc. Natl. Acad. Sci. USA 2012, 109, 8456–8460. [Google Scholar]
  76. OGAMA—Open Gaze and Mouse Analyzer. Available online: http://www.ogama.net/ (accessed on 23 January 2015).
  77. PyGaze—The Open-Source Toolbox for Eye Tracking. Available online: http://www.pygaze.org/ (accessed on 23 January 2015).
  78. OpenEyes—Open-Source Open-Hardware Toolkit for Low-Cost Real-Time Eye Tracking. Available online: http://thirtysixthspan.com/openEyes/ (accessed on 23 January 2015).
  79. OpenGazer—Open-Source Gaze Tracker for Ordinary Webcams. Available online: http://www.inference.phy.cam.ac.uk/opengazer/ (accessed on 23 January 2015).
  80. Pupil Labs—Open-Source Mobile Eye Tracking Platform. Available online: http://pupil-labs.com/pupil/ (accessed on 23 January 2015).
  81. EyeTribe—Ther World's First 99 Dollar Eye Tracker. Available online: http://www.theeyetribe.com/ (accessed on 23 January 2015).
  82. Mantiuk, R.; Kowalik, M.; Nowosielski, A.; Bazyluk, B. Do-It-Yourself Eye Tracker: Low-Cost Pupil-Based Eye Tracker For Computer Graphics Applications; Springer: Klagenfurt, Austria, 2012. [Google Scholar]
  83. Huang, C.W.; Jiang, Z.S.; Kao, W.F.; Huang, Y.L. Low-Cost and High-Speed Eye Tracker. In Intelligent Technologies and Engineering Systems; Springer: New York, NY, USA, 2013; pp. 421–427. [Google Scholar]
  84. Ho, H.F. Low cost and better accuracy eye tracker. Proceedings of the IEEE 2014 International Symposium on Next-Generation Electronics (ISNE), Kwei-Shan Tao-Yuan, Taiwan, 7 May 2014; pp. 1–2.
  85. Johansen, S.A.; San Agustin, J.; Skovsgaard, H.; Hansen, J.P.; Tall, M. Low cost vs. high-end eye tracking for usability testing. Proceedings of the CHI'11 Extended Abstracts on Human Factors in Computing Systems, Vancouver, BC, Canada, 7 May 2011; pp. 1177–1182.
Figure 1. Different levels of the task.
Figure 1. Different levels of the task.
Sensors 15 11092f1 1024
Figure 2. Participant using the system while his gaze is being recorded.
Figure 2. Participant using the system while his gaze is being recorded.
Sensors 15 11092f2 1024
Figure 3. Raw data processing.
Figure 3. Raw data processing.
Sensors 15 11092f3 1024
Figure 4. Raw data processing [48].
Figure 4. Raw data processing [48].
Sensors 15 11092f4 1024
Figure 5. Outlier detection process.
Figure 5. Outlier detection process.
Sensors 15 11092f5 1024
Figure 6. Time vs. correct answers: user performance.
Figure 6. Time vs. correct answers: user performance.
Sensors 15 11092f6 1024
Figure 7. Participants with the best performance results.
Figure 7. Participants with the best performance results.
Sensors 15 11092f7 1024
Figure 8. Participants with the worst performance results.
Figure 8. Participants with the worst performance results.
Sensors 15 11092f8 1024
Figure 9. Fixation data: best and weakest performers. (a) No. of fixations vs. fixation avg. duration, best performers; (b) No. of fixations vs. fixation avg. duration, weaker performers.
Figure 9. Fixation data: best and weakest performers. (a) No. of fixations vs. fixation avg. duration, best performers; (b) No. of fixations vs. fixation avg. duration, weaker performers.
Sensors 15 11092f9 1024
Table 1. Literature review: Experimental conditions.
Table 1. Literature review: Experimental conditions.
AuthorsCitationYearCountryGameDeviceDisplayParticipants

TotalFemaleMaleAge (SD)
Kiili et al.[17]2014FIAnimal ClassTobii T6017"8267–13 years
Chang et al.[27]2013PTWaylaTobii REX
Muir et al.[3]2012USPrime ClubTobii T12017"126610–12 years
Walber et al.[31]2012DEEyeGrab2471715–32 years
Kickmeier et al.[19]2011AT80 DaysTobii 175094513 (1.61)
Józsa and Hammornik[22]2011HU7 Hidden DifferencesTobii T12017"43142919–26 years
Pretorious et al.[12]2010ZATimez AttackTobii 175017" (1280 × 1024)8449–12 years over 40
Sennersten and Lindley[20]2010SEFPS computer gameTobii 1750
Nacke et al.[32]2010CAHalf-Life 2Tobii T1203022818.67 (4.26)
Hillaire et al.[26]2008FRQuake IIIASL6000Cylindrical Screen (1280 × 1025)80825.8 (4.3)
Dorr et al.[23]2007DEBreakout GameSensoMotoric IViewX Hi-Speed20"9
El Nasr et al.[18]2006USGame Soul CaliberISCAN ETL-500 (head-mounted)620–30 years
Smith and Graham[25]2006CACustom build sceneRED25022" (1680 × 1050)2112021–24 years
Table 2. Different levels’ settings.
Table 2. Different levels’ settings.
Time (s)Grid SizeNo. of SlicesDisplayImage Level


ColorGreyscaleEasyMediumHard
Level 1503×39XX
Level 2503×412XX
Level 3504×416XX
Level 4505×420XX
Table 3. Selected features for analysis.
Table 3. Selected features for analysis.
FeatureData TypeFeatureData TypeFeatureData TypeFeatureData Type
Classificationoutlier/normalFixation avgdurationmsFixation No. in A1–C3 (9 features)IntegerTime per level (4 features)seconds
sFixation No. in A1–C3 (9 features)IntegerTime per level (4 features)seconds
Global alphapercentageFixation max durationmsFixation avg duration in A1–C3 (9 features)msTotal correct answersinteger
sTotal correct answersinteger
sFixation avg duration in A1–C3 (9 features)msTotal correct answersinteger
sTotal correct answersinteger
Fixation total No.IntegerFixation min durationmsTotal time in exercisesecondsCorrect answers per level (4 features)integer
Table 4. Participants outcome scores and response times for each level.
Table 4. Participants outcome scores and response times for each level.
Participant ID (n = 32)Level 1Level 2Level 3Level 4Total (16 max.)

Time (s)Items CorrectTime (s)Items CorrectTime (s)Items CorrectTime (s)Items CorrectTime (s)Items Correct
1240.58434.62447.94248.133171.2813
1536.79436.83248.66049.033171.339
1334.76448.28448.29348.333179.6714
1440.87430.72448.06247.293166.9513
1644.99448.77148.35047.531189.666
1747.48448.08348.05448.262191.8913
1848.81347.71347.82238.032182.3910
1942.01433.93447.95224.864148.7714
2043.62440.49447.99342.854174.9615
2129.76324.35448.12221.474123.7213
2231.21433.36443.60240.812148.9812
2348.71333.31448.09446.193176.3214
2439.25441.33448.24139.532168.3711
2525.63429.38448.03429.733132.7915
2630.11448.50348.28139.943166.8411
2748.90344.77448.07247.923189.6712
2843.09441.57447.94335.004167.6115
2948.53448.12347.82048.072192.569
3048.65437.18041.56439.364166.7712
3139.92433.26347.76144.942165.9110
3248.72446.81348.09248.252191.8911
3342.77446.85447.90237.313174.8513
3446.62431.48447.93343.862169.9213
3548.60348.95448.42248.792194.7811
3648.69346.78348.29148.571192.348
3749.08448.84348.52147.882194.3410
3835.96446.81446.56348.623177.9714
3946.30439.33347.39146.842179.8810
4040.34436.38348.07348.201172.9911
4138.17445.31348.08239.874171.4513
4248.50427.59447.95248.022172.0712
4348.03323.81438.89438.894149.6315

Average (SD)42.36 (6.70)3.75 (0.42)39.80 (7.97)3.37 (0.94)47.40 (2.09)2.12 (1.18)42.89 (7.16)2.65 (0.94)172.45 (17.18)11.93 (2.20)
Table 5. Participants gender, age, number of fixations and average duration of these per level: best performers.
Table 5. Participants gender, age, number of fixations and average duration of these per level: best performers.
Participant IDAgeGenderLevel 1Level 2Level 3Level 4Total

No. Fix.Fix. Avg Duration (ms)SD (ms)No. Fix.Fix. Avg Duration (ms)SD (ms)No. Fix.Fix. Avg Duration (ms)SD (ms)No. Fix.Fix. Avg Duration (ms)SD (ms)Time (s)Correct Ans
209Female206172.7267.0331193.1074.8974188.5968.6593188.0867.93174.9615
2512Male46153.8558.5329160.4970.6222128.4037.6430142.6672.09132.7915
2811Male43162.4259.0411167.0757.249121.989.8219158.4255.27167.6115
4311Female112182.6150.9051167.7090.6684165.4378.3521133.1539.96149.6315
Table 6. Participants gender, age, number of fixations and average duration of them per level: weakest performers.
Table 6. Participants gender, age, number of fixations and average duration of them per level: weakest performers.
Participant IDAgeGenderLevel 1Level 2Level 3Level 4Total

No. Fix.Fix. Avg Duration (ms)SD (ms)No. Fix.Fix. Avg Duration (ms)SD (ms)No. Fix.Fix. Avg Duration (ms)SD (ms)No. Fix.Fix. Avg Duration (ms)SD (ms)Time (s)Correct Ans
1511Female125169.9180.86180167.9574.16209168.8571.15245167.8269.23171.339
169Female138237.16128.3535243.4494.4150240.2690.6898235.1584.44189.666
2912Male91174.3876.5838185.2991.7917179.0170.6736176.0870.79192.569
368Male87163.6169.03142159.4160.23164165.2571.36186167.9976.02192.348
Table 7. Best vs. weakest performers by level: Mann–Whitney analysis of the results.
Table 7. Best vs. weakest performers by level: Mann–Whitney analysis of the results.
Level 1Level 2Level 3Level 4Global

pMann–WhitneypMann–WhitneypMann–WhitneypMann–WhitneypMann–Whitney
U-ValueU-ValueU-ValueU-ValueU-Value
Fix Number0.336.00.052.00.235.00.031.00.031.0
Fix Avg. Time0.235.00.333.00.154.00.093.00.010.0
Time0.154.00.093.00.093.00.010.00.093.0
Correct Answers0.428.00.438.00.010.00.015.00.093.0
Table 8. Performance comparison of feature selection algorithms using selected classifiers.
Table 8. Performance comparison of feature selection algorithms using selected classifiers.
Feature Selection Algorithms

Select k-bestL1-BasedHierarchical
No. of FeaturesAll221612647Btw10–14
Decision Tree0.76 (±0.21)0.80 (±0.18)0.80 (±0.14)0.81 (±0.15)0.80 (±0.14)0.80 (±0.15)0.79 (±0.15)0.82 (±0.17)
Random Forest0.84 (±0.17)0.87 (±0.11)0.86 (±0.12)0.87 (±0.11)0.84 (±0.11)0.83 (±0.14)0.86 (±0.11)0.88 (±0.11)
Extra Tree0.80 (±0.18)0.85 (±0.12)0.82 (±0.14)0.82 (±0.14)0.81 (±0.14)0.80 (±0.14)0.83 (±0.14)0.84 (±0.14)
AdaBoost0.78 (±0.21)0.85 (±0.14)0.85 (±0.15)0.84 (±0.15)0.82 (±0.14)0.81 (±0.15)0.85 (±0.14)0.86 (±0.13)
Table 9. Ensemble methods vs. decision trees: Mann–Whitney analysis of their accuracy.
Table 9. Ensemble methods vs. decision trees: Mann–Whitney analysis of their accuracy.
Ensemble Method(accuracy)Decision Tree (accuracy)pMann–Whitney U-Value
Random Forest0.84<0.0012843.5
Extra Trees0.80.760.0033919.5
AdaBoost0.780.024206
Table 10. All features vs. feature selection algorithms: Mann–Whitney analysis of their accuracy.
Table 10. All features vs. feature selection algorithms: Mann–Whitney analysis of their accuracy.
All Features vs. Selected K best (22 features)All Features vs. L1-based (7 features)All Features vs. Hierarchical (btw. 10–14 features)



pMann–WhitneypMann–WhitneypMann–Whitney
U-ValueU-ValueU-Value
Decision Trees0.0013550.50.0023979<0.0013134.5
Random Forest0.00073798.50.1624201.50.0033901.5
Extra Trees<0.0013228.50.00033638.5<0.0013308
AdaBoost<0.0013096.5<0.0013075<0.0012740.5

Share and Cite

MDPI and ACS Style

Frutos-Pascual, M.; Garcia-Zapirain, B. Assessing Visual Attention Using Eye Tracking Sensors in Intelligent Cognitive Therapies Based on Serious Games. Sensors 2015, 15, 11092-11117. https://doi.org/10.3390/s150511092

AMA Style

Frutos-Pascual M, Garcia-Zapirain B. Assessing Visual Attention Using Eye Tracking Sensors in Intelligent Cognitive Therapies Based on Serious Games. Sensors. 2015; 15(5):11092-11117. https://doi.org/10.3390/s150511092

Chicago/Turabian Style

Frutos-Pascual, Maite, and Begonya Garcia-Zapirain. 2015. "Assessing Visual Attention Using Eye Tracking Sensors in Intelligent Cognitive Therapies Based on Serious Games" Sensors 15, no. 5: 11092-11117. https://doi.org/10.3390/s150511092

Article Metrics

Back to TopTop