A study on the use of eye tracking to adapt gameplay and procedural content generation in first-person shooter games

This paper studies the use of eye tracking in a First-Person Shooter (FPS) game as a mechanism to (1) control the attention of the player's avatar according to the attention deployed by the player and (2) guide the gameplay and game's procedural content generation, accordingly. This results in a more natural use of eye tracking in comparison to a use in which the eye tracker directly substitutes control input devices, such as gamepads. The study was conducted on a custom endless runner FPS, Zombie Runner, using an affordable eye tracker. Evaluation sessions showed that the proposed use of eye tracking provides a more challenging and immersive experience to the player, when compared to its absence. However, a strong correlation between eye tracker calibration problems and player's overall experience was found. This means that eye tracking technology still needs to evolve but also means that once technology gets mature enough players are expected to benefit greatly from the inclusion of eye tracking in their gaming experience.


I. INTRODUCTION
With eye tracking dating back from the XVIII century [2], many have been the applications for this technology, such as medicine [3]), robotics [4], [5]), advertising [6]) and, more recently, computer games [7]. In the past decade, research on computer games tried to compare traditional input (e.g., mouse, keyboard, gamepad) with eye tracking input, in terms of action accuracy and responsiveness [8], [9], [7], [10], [11]. These comparisons were often made after asking players to compete against each other or asking users to complete a given task, using the different input methods. Overall, these comparisons provided mixed results, with some studies claiming that the use of eye tracking contributed to better task completion [11], while others claiming that traditional input devices provided better overall results [8]. Hence, these previous studies have provided contradictory results regarding the effectiveness of the use of eye tracking in computer games as a direct control input, which may be an indication that eye tracking is not best suited to direct input control.
Bearing the limitations of eye tracking as a simple direct control input in mind, in this paper, we propose to use it in order to control the attention of the player's avatar and the game's procedural content generation. This use of eye tracking is focused in mapping the mental state of the player and her/his avatar, which we believe to be much more  natural and useful than directly controlling a pointer with the eyes, which has no actual mapping to real life. That is, instead of replacing the traditional input for an eye tracker, we are more interested in studying meaningful ways about how eye tracking can improve gameplay, which includes coexistence with traditional inputs. This also means that, instead of analysing objective performance-based data, as in previous studies, we are more interested in a subjective analysis of how eye tracking improves enjoyability and how well the player adapts to the technology.
To demonstrate the value of the herein proposed alternative uses of eye tracking in computer games, we developed and tested our own endless runner First-Person Shooter (FPS), Zombie Runner (see Fig. 1). In this game, shot accuracy, automatic obstacle avoidance, and procedural obstacle spawn probability are controlled as a function of the avatar's attention model, which, in turn, operates according to the player's gaze estimated with an affordable eye tracker. The goal is to better represent the player's actions in the game, thus contributing to a more immersive experience. Based on a set of testing sessions, we conclude that the use of eye tracking provides a more challenging and immersive experience to the player. Participants reported better levels of satisfaction while playing the game with the gaze tracking turned on. However, a strong correlation between eye tracker calibration problems and player's overall experience was found. This means that eye tracking technology still needs to evolve but also means that once technology gets mature enough players are expected to benefit greatly from the inclusion of eye tracking in their gaming experience. This paper is organised as follows. Section 2 presents an overview of previous and related work. Then, in Section 3, Zombie Runner is described and its implementation detailed. Section 4 describes the experimental setup and analysis the obtained results. Finally, Section 5 presents some conclusions and provides some future work directions.

II. RELATED WORK
Eye tracking is the process of estimating one's gaze direction, identifying the object in which the the subject is focused [12], [13]. Eye tracking dates from the XVIII century, in which persistent images to describe the human eye movements were used [14]. In the XX century, the first eye movement measures through a non-intrusive method using photographs and light reflections were made [15]. In the 80s, with the evolution of computing capacity, it became possible to perform real-time eye tracking with access to video, what opened the possibility of human-machine interaction [16]. With increasingly more accessible prices [17], [18], the use of eye trackers increased in several areas, such as marketing [6], psychology [3] and, more recently, computer games [7].
Eye tracking has been explored in computer games as an alternative to traditional input methods, such as mouse or keyboard [7], [10]. By testing gaze input versus mouse input in three different computer games, Smith and Graham [7] concluded that the use of eye tracking can provide a more immersive experience to the player. Isokoski and Martin [9] conducted a preliminary study on the use of eye tracker in First Person Shooters. Each participant in the test was asked to play the same game using three different input method schemes: (1) mouse, keyboard, and eye tracker; (2) only mouse and keyboard; or (3) a console gamepad. The conclusions were not exactly encouraging, suggesting that the performance with the eye tracker was quite inferior to the other two. Still, Isokoski and Martin attributed these results to the players' greater knowledge and contact with the traditional input methods, suggesting that this scenario could change with more training.
Other studies achieved similar conclusions. Leyba and Malcolm [8] created a simple test in which the player was asked to eliminate twenty-five balls that moved around the screen at different velocities. The player would move the pointer using the mouse or the eye tracker and would eliminate the balls by clicking on them with the mouse. Two conditions were tested: with and without time limit to complete the task. The results showed that, without time limit, precision and time to complete the task was worse while using the eye tracker than when using a mouse. The same results were obtained for the no time limit condition, in which performance was based on percentage of balls eliminated by the player. Michael Dorr et al. [11] achieved totally opposite results. After creating and adapting a clone of the classic game Breakout, twenty players were asked to participate in a tournament. Players were separated in pairs. The two players of each pair played against each other, one using the mouse and the other using an eye tracker. The control inputs were swapped between rounds. The results showed that the players who used the eye tracker achieved higher scores and won more rounds. The players also stated that using the eye tracker was highly enjoyable.
These discrepant results between studies suggest that the type of game and development method of the same game are key elements to achieve a satisfying final result.
Bearing the limitations of using eye tracking as a simple direct control input in mind, we propose to use it in order to control the attention of the player's avatar and the game's procedural content generation. This use of eye tracking is focused in mapping the mental state of the player and her/his avatar, which we believe to be much more natural and useful than controlling a pointer with the eyes, which has no mapping to real life. Another alternative and interesting use of eye tracking, not tackled in this paper, is to know when to actively redirect the player's attention [19].
Procedural Content Generation (PCG) concerns all creation of game content (e.g., sounds, levels, objects, characters, textures) with algorithms, with limited or indirect stimuli from the user [20]. PCG is a way of coping with the daunting task of manually creating and populating massive open worlds. Besides this more traditional use of PCG, some researchers proposed that, through the analysis of the interaction between the player and the game, PCG could be used to create a playing experience that adapts itself to the player [21], [22], [23], improving the game's replay value. This new approach is known as Experience-Driven Procedural Content Generation (EDPCG) [24].
Within the EDPCG framework, it is possible to generate levels that are adapted to the strengths and limitations of the player, in an attempt to maximize the fun factor. Many studies propose that the fun and challenge factors are directly linked [25], [26], [27], [28], [29], [30], [31], which means that to tune the fun factor one often has to tune the challenge factor. In this line, in this paper, we propose the integration of PGC with gaze tracking so as to adapt the challenge the player as to face according to his/her form of playing and, as a result, to improve the fun factor.

III. Zombie Runner
The game herein presented, Zombie Runner, was purposely designed to integrate gaze tracking at its core mechanic as a way of mapping both player's and avatar's visual attention, instead of as a way of directly controlling the avatar's motor actions. The latter is done via a traditional input, a gamepad. This way gaze and hand orthogonally and, thus, more naturally, control their virtual counterparts. To better analyse the advantages and disadvantages of the use of eye tracking in games, focus should be given to game genres where the player's visual attention has to be shared between different elements in the scene under tight temporal restrictions, where the eye tracking use is more central and challenging. For this reason, Zombie Runner, has been designed and implemented as the common genre (of high impact) FPS. The character's running action is fully automated so as to reduce the number of actions the player had to memorize and control and, hence, reduce variability in the game evaluation phase.

A. System Configuration
The game has been with the following hardware configuration in mind (see Fig. 2): a Gazepoint GP3 eye tracker; a Microsoft Xbox gamepad; and a 32 inch computer screen. The player sits in front of the computer screen, with the gamepad in hand. The eye tracker sits bellow the computer screen. Behind it, a laptop running the game is available for the research team during the evaluation sessions. Eye tracking was done using the Gazepoint Control software [32] and the game was developed on Unreal Engine.

B. Game Rules and Mechanics
Zombie Runner follows the following set of rules. The main objective of the game is to ensure that the avatar survives for as long as possible while running along a corridor with a non controllable constant forward movement. The player can achieve this by killing enemies and by noticing elements in the scene (enemies and obstacles). The player can kill enemies by aiming and sooting the avatar's gun via the gamepad. The player can notice the several elements in the scene by actively looking at them on the screen for a sufficient amount of time. The gaze of the player is estimated with the eye tracker. An element noticed by the player is also noticed by the avatar.
If the avatar approaches a previously noticed obstacle, it will automatically avoid that obstacle (i.e., jump over a rock or dodge an hanging tree branch). If this obstacle was not noticed early enough, it will not be seen by the avatar, which will result in a collision and subsequent avatar's health decrement. A shot enemy that was noticed dies instantly, whereas it will only be hurt if unnoticed. This intends to simulate the lack of shot accuracy resulting from an insufficient focus on the enemy. If a noticed enemy approaches the avatar, it will attack, causing the avatar to lose health. If the enemy was not noticed early enough, its attack will cause instant avatar's death. The idea is that the avatar is able to dodge the attack provided the attacker had been noticed. The avatar dies when its health reaches zero, leading to game over. Fig. 3 depicts the state diagrams associated to these game rules.

C. Game Implementation
Zombie Runner was developed on Unreal Engine. The programming of Zombie Runner was split between C++ and Blueprints, the Unreal's visual scripting language. C++ was used for the core algorithms while Blueprints was used to program actor behaviour, such as enemies and obstacles. The game is based on the Unreal's FPS template, which already implements the expected behaviour for a generic game of the genre, including shooting, walking, and aiming mechanisms. The corridor in which the game takes place is made of an unidimensional array of 3D tiles (see Fig. III-C), which are procedurally generated and removed from the scene once the avatar passes by the them. Tiles are comprised of one plane for the floor, two planes for the side walls, an array of marker points distributed randomly on the sides of the floor (not in the region in which the avatar will be crossing), and an array of tree models. Every time a tile is spawned, the array of marker points is traversed. For each marker, there is a 66 % probability of spawning a tree (non-obstacle) in its position, plus a small random offset. This tree is spawned with a random rotation. The tree's height is also randomly set between reasonable values. The result are sufficiently different tiles that when placed in succession give the feeling of a dense and varied forest on each side of the player's sight. A set of 15 tiles are initially generated. After the 8 th tile is spawned, obstacles or enemies can start being spawned with them. This value was achieved by trial and error and produces an initial tile generation that gives enough time for the player to settle in and prepare for the incoming obstacles and enemies. After this initial generation, all subsequent tile generation is recursively and procedurally handled by the tiles themselves. An obstacle (i.e., a rock or an hanging tree branch) or an enemy has a 33 % probability of being procedurally spawn in each new tile. If it does occur, there's a 55 % probability for either spawning an obstacle or an enemy. A spawned enemy has a 20 % probability of being spawned as  a runner, which has double the speed of a walker, the default behaviour of an enemy. These percentages have been tuned according to a set of informal tests so as to ensure that the game was playable without training.
An obstacle or an enemy can be spawned in three different regions in the tile (see Fig. 5): central region, left region, or right region. If the previous spawned obstacle or enemy was spawned in either the left or the right regions, the new one will be spawned in the central region. This adds variety to the game and avoids objects of the same type to be spawned close to each other. If the previous spawned obstacle or enemy was spawned in the central region, the new one will be spawned in either left or right regions, depending on the player's gaze. Concretely, the obstacle or enemy is spawn in the region of the tile least gazed by the player during the time spent running over the two previous tiles. This forces the attention of the player to be often alternating between regions, thus increasing the challenge. When the obstacle is spawned in the central, its asset is a rock, being a tree branch when spawned on one of the side regions. Tree branches are implemented as laying small trees (see Fig. ).
Zombie Runners relies on ray casting to determine which objects are being attended by the player, given the gaze position in the screen. A given object is labelled as noticed if the gaze of the player intersects that same object over an accumulated period of 0.5 seconds. This period was obtained from a set of informal tests as the one that produced the least false positives and negatives when classifying obstacles as noticed/unnoticed. To speed up computation, ray-object intersections are carried out with accessory bounding boxes, rather than with the objects themselves (see Fig. 6). These bounding boxes are also used to determine object-object  intersections, such as to determine when the avatar collides against an obstacle (see Fig. 7). The sizes of these bounding boxes had to be adjusted so as to accommodate the inaccuracy of the eye tracker. These adjustments were done during a set of informal tests.
All assets (e.g., obstacles) were freely obtained from the Unreal's dedicated store. The rigged 3D model and basic animations for the enemies were obtained from Adobe Mixamo. An Unreal Animation Blueprint was built as a state machine which defining the different animation states and transitions between them (see Fig. 8). To provide the player with compelling visual feedback, the appearance of the objects changes when their are first gazed by the player and also later on when they become actually labelled as noticed. The first change consists in rendering the object in wireframe with shades of purple. The second change, more dramatic, consists in turning the object in light blue and in adding a particle effect representing a blue shock wave. Fig. 9 shows the evolution of the materials and effects used on a rock obstacle through the process of being noticed.

IV. EVALUATION AND DISCUSSION
As mentioned, several informal tests were run in order to guide the development and tune the game's parameters. Then, the whole game was played by a set of ten people in formal game evaluation sessions in order to systematically assess what eye tracking technology brings to the game in terms of overall enjoyability.

A. Evaluation Method
Test sessions were carried out privately in a room without the presence of anyone but the participant and the research team. An email-based call for participants that considered themselves gamers was issued to an universe of 170 people. The first ten that responded to the email and complied with the requirements were selected for the test sessions. The participants had no previous knowledge about the game and experience. The ages of the ten participants spanned from 25 to 49 years old (see Table I), with different occupations such as software developer, quality assurance tester, and student. All participants were male, with no female subjects volunteering for the experience.  In addition to bottom-line data stored during the test sessions, we also enquired the participants with Game Experience Questionnaires (GEQ) [33], which has been widely applied in previous studies related to control inputs in computer games [34], [35]. By using GEQ, we intend to evaluate if the use of eye tracker is enjoyable and comfortable for the player. We also intend to pinpoint possible advantages and disadvantages of the technology in the way it impacts the overall experience of the player and its relationship with the game, comparing the player experience with and without eye tracking.
1) Test Sessions: Each test session started with a brief questionnaire the participant had to fill in, regarding personal details, such as age and occupation, if the player had some degree of visual impairment, as well as classifying their experience as video game players, with the use of eye trackers, and with the use of gamepads in FPS games. As Tables II, III, and IV show, the average participant had a considerable amount of gaming experience, little to no previous exposition to eye tracking technology, and was moderately experienced using gamepads in FPS games.
To reduce any pleasing biases the participants might have, we conveyed the message that we are agnostic regarding the use of eye tracking in games by means of a brief paragraph on the top of the questionnaire with "the advent of eye tracking technology some developers are including this technology in games" and "this study aims at providing scientific validation of the advantages and disadvantages of such inclusion in terms of gameplay and comfort to the user". Then, a brief explanation of the game's rules and objectives was given, followed by a briefing on the eye tracker calibration process. This process was ran as many times as deemed necessary until the calibration was acceptable. When this acceptable calibration was achieved, the participants were asked to state their feelings towards the process and if they would imagine themselves doing it at home before playing a game.
To get the participant acquainted with the game, two separate runs were done, with each one only having an input form enabled (gamepad versus eye tracker). These runs had no time limit and did not count for the evaluation. On the first run, only the eye tracker was enabled and no enemies were spawned. The result was a play session with only obstacles being spawned. This allowed the participants to freely use their eyes to notice the obstacles in front of them and get used to this mechanic, without having to worry about the gamepad. The participants were also asked to state any false positives or obstacles that they had noticed but were not tagged as such by the game. On the second run, no obstacles or enemies were spawned. This allowed the participants to get acquainted with aiming and shooting with the gamepad, adapt to its sensibility and button scheme. The main objective was for the participant to feel comfortable with the different inputs and the test session would only advance when the participants confirmed they felt acquainted with the controls.
The participant was then asked to play the game in its original form, as described in Section 3, for three sessions of 2 minutes each. For each session, the ratio between enemies killed and spawned, the ratio between overall (enemies and obstacles) noticed count and overall spawned count, as well as the number of deaths experienced by the player, were registered. After these three sessions, the player was asked to fill the core and post-game modules of a Game Experience Questionnaire (GEQ) [36]. This questionnaire is filled by answering a set of questions whose answers are provided on a Likert-type scale scored from 0 to 4. Afterwards, the participant was asked to play another set of three sessions of 2 minutes, but this time with eye tracking disabled, which means that all obstacles and enemies were automatically labelled as noticed, with the participant only being required to shoot the enemies. With these three sessions over, another GEQ was filled by the player. Then, the calibration process was performed again and the player was asked to play one 2 minute session of the game but this time with the visual effects, that occur when the obstacle or enemy is noticed, disabled.
Finally, an end questionnaire was then handed to the participant. This questionnaire was more subjective and consisted on three questions: (1) what do you think of the visual effects used on the first sessions, comparing with the last session you played? (2) How was the overall experience of playing the game? (3) Would you consider an eye tracker as part of your gaming setup and why?

B. Experimental Results
All test sessions were concluded with success, in the sense that all test subjects were able to perform the tasks required from them.

1) Eye Tracker Calibration Process:
The calibration process revealed itself as the most challenging step of the testing session. A calibration was deemed good enough if the accuracy in the areas where the player mostly interacts with in the game, i.e., the whole screen except the top corners (see Fig. 10), was above a given threshold. Two participants had to remove their glasses so that their gaze could be properly detected. Table V shows the distribution of participants according to the number of calibration tries each participant required to obtain a good enough calibration. As the table shows, a single calibration process was enough for half the participants.
After achieving a calibration deemed successful, the participants were asked about their feelings towards it and if they would see themselves repeating this process at home before playing a game. From all the participants, 8 said that they would see themselves doing it, stating that a calibration process also had to be performed with other controllers. Three of these participants reported that the process was easy and fast, while others stated that it should be easier, but it is still acceptable. The other 2 participants did not see themselves calibrating an eye tracker each time they wanted to play, expressing that they wished the process was easier.
2) Play Sessions: The three 2 minute play sessions produced results that suggest an approximation to the ideal game flow, with the participant being able to progressively learn how to use play more competently. This can be observed in Tables VI, VII, and VIII, which show how the averages of all participants for the ratio of enemies killed and noticed obstacles and enemies, along with the number of deaths, evolved along the play sessions. The two ratios evolved positively, with the player improving in the tasks of killing enemies and noticing them, along with obstacles. The number of deaths went down abruptly from the first to the second session and then went up slightly on the third one, but to a value close to the lowest one obtained in the second session. We suspect this slight increase in deaths can be a consequence of the better results achieved by the participants in terms of killing enemies and noticing objects. To excel in these tasks, participants had to better coordinate the two input forms, which lead to a greater risk of being killed.
The GEQ was filled out by the participants after these three play sessions. As mentioned, after these sessions the participants were asked to play an additional set of three play sessions, this time without using eye tracker, i.e., with all obstacles and enemies automatically labelled as noticed. Table IX compares, for each of the components of GEQ core module (competence, sensory and imaginary Immersion, flow, tension/annoyance, negative affect, positive affect), the sessions with and without eye tracking. In the Competence component, the participants felt more competent and able to complete tasks without the eye tracker than with it, which can be explained by the higher degree of difficulty of the game when eye tracking is on. In the Sensory and imaginative immersion component, the participants reported higher scores with eye tracking, which means the feeling of immersion was felt at a greater level with the full game experience. The Flow component was also higher with eye tracking on, which reveals the participants felt a stronger feeling of game flow, a better balance within the game's Challenge, another

Play Session
Number of deaths . Both positive and negative affects are slightly higher when eye tracking was on, suggesting that players were more emotionally invested in that condition. In both conditions, the positive affect is considerably higher than the negative one, suggesting that the game provides an overall positive experience.
Although GEQ scores are used in this study to compare two designs, we can also analyse the absolute meaning of the obtained scores. In this regard, although the GEQ scores are low when compared to the maximum of the scale, these are not substantially different from the ones obtained for full-fledged commercial FPS games, reported in a previous study [34]. The lower scoring can be explained by the awareness players have of the current state of the art in commercial games.
The results were treated in the same way for the GEQ post-game module, whose results are summarised in Table X. The differences in levels of positive and negative experience are too small (≈ 0.05) to allow any analysis from them. The more significant difference in Tiredness and Returning to reality components provide additional support to the idea that eye tracking renders the experience more immersive and engaging. The similar values obtained with and without the eye tracker suggest that the participant's perception of the overall experience was more shaped by the game itself than by the use or lack of the eye tracker. Still, these results become more meaningful and easier to understand when compared with the answers the test subjects gave to the last set of informal questions (see below). The main goal of these questions was to extract from the participants more subjective perceptions they had from the game, which could help us understand the GEQ scores.
3) Informal Questions: The first question asked about the participant's opinion about the used visual effects signalling which objects were labelled as noticed. Nine of the ten participants stated the importance of the effects as a means to give feedback to the player, with some pointing out that without effects the player may be forced to look more than needed as the player never knows if the obstacle or enemy was actually tagged as noticed. The use of effects was also pointed out as more rewarding to the players actions. From all the participants, seven reported that without the visual effects, the game was more immersive and the experience more realistic, making the way the player looks at things more natural. This suggests that being more immersive does not mean that a game is necessarily more rewarding. In order for it to be both rewarding and immersive, a different, more subtle feedback mechanism should be implemented as to avoid breaking the suspension of disbelief. Some participants, to whom the eye tracker calibration was not fully successful, stated that the effects induced frustration on them as they could see the discrepancy between where the eye tracker thought they were looking at and where they were actually looking at. This suggests that for this kind of implementation in a computer game to be successful, we need better eye trackers, capable of finer calibration.
To the second question about the overall experience of playing the game, three out of the 10 participants complained about having to be as static as possible to avoid the eye tracking decalibration. Two of the participants reported problems with the game registering when they looked at enemies and rock obstacles. Some participants complained about hardware problems and the frustration of going to the process of calibration and then the technology still not working right. A participant that had problems with getting the eye tracker to work while wearing glasses wished that the technology was more prepared for people with glasses. Regarding the game, a participant stated that it was well designed, also expressing appreciation for the feedback given when the player loses life. Another participant said that the mechanic of noticing things was fun, but that the game lacked progression, having nothing new after the first minute. This suggests that a more complex game could have been a better fit for this test session, as it would avoid frustration resulting from lack of novelty. One participant felt that the time to set an obstacle or enemy as noticed was too long. The experience was classified as immersive by three participants, with one of them stating enthusiastically that eye tracking was amazing.
To the last question regarding if the participants would consider the inclusion of an eye tracking camera in their gaming setup, the responses were mixed. From the entire group, two of the participants said they would not do it, with reasons such as it being another piece of hardware that has little application and would be quickly abandoned after the novelty effect wore off. Other three participants stated that they would not in the current state but that they could try it in the future, stating that depending on the game it could facilitate precision tasks such as aiming in FPS games or passing the ball to another player in a football game. The reasons these participants presented as to holding off in the adoption of the technology regarded its poor performance while wearing glasses, and the fact that it was not stable or precise, which allied with the calibration process ruined the experience. The other five participants said they would adopt the technology, stating that it was a new form of interaction, that opened new possibilities and created more immersive experiences.
These answers showed, along with the previous questionnaires, that the use of eye tracking in games has both pros and cons. On the positive side, the use of gaze-oriented gameplay provided a more immersive and richer experience, providing a better game flow. On the negative side, the technology's limitations raised feelings on the participants that were not desired. The calibration process and its results, along with some disbelief that eye tracking could have an important role in a computer game, are the main reasons for the participants being adamant about adopting the technology.

V. CONCLUSIONS AND FUTURE WORK
This paper presented Zombie Runner, an endless runner FPS game, whose core mechanics and procedurally generated content are modulated by the player's gaze estimated with an affordable eye tracker. A set of testing sessions were carried out in order to assess the impact of eye-tracking in the player's satisfaction when playing Zombie Runner. The results obtained from testing sessions show that the use of eye tracking provides a more challenging and immersive experience to the player. These results complement previous studies, which were mostly focused on performance-based metrics when using eye tracking as a direct control input. Conversely, Zombie Runner exploits eye tracking as a mechanism to match the avatar's attention model with the one of the player. Moreover, Zombie Runner also exploits eye tracking to guide the procedural content generation of the game's environment, contributing in an original way to the emerging field of Experience-Driven Procedural Content Generation (EDPCG) [24].
During the evaluation, a strong correlation between problems that surfaced with the eye tracker calibration and participants' overall experience was observed. Participants for whom the hardware worked without major flaws reported better levels of satisfaction when contrasting with participants for whom the calibration process was not perfect or took a longer time. Among the participants' complaints about the eye tracking technology, many were related the need to have the head mostly static during the play session, the calibration process requiring repetitions to meet the required accuracy, and overall eye tracker's lack of precision. These complaints show that the affordable eye tracking technology still has to grow and develop untill it is in a state that can be accepted by the overall gaming community. However, positive player's experience when calibration was easily attained is a sign that the method will be a valuable asset for the game designer as soon as the eye tracking technology matures.
As future work, we intend to validate the use of eye tracking in other types of video games and allow for free avatar's movement. We also intend to extend this testing framework, including the philosophy behind the way the player's attention is integrated in the core gameplay, to games with other types of camera perspective, such as third-person games. We pretend to perform a more intensive set of tests, enlarging the participant population, in order to obtain more robust statistics.