Next Article in Journal
Fabrication and Characterization of a Micromachined Swirl-Shaped Ionic Polymer Metal Composite Actuator with Electrodes Exhibiting Asymmetric Resistance
Next Article in Special Issue
Surveying Wearable Human Assistive Technology for Life and Safety Critical Applications: Standards, Challenges and Opportunities
Previous Article in Journal
Calorimetric Thermoelectric Gas Sensor for the Detection of Hydrogen, Methane and Mixed Gases
Previous Article in Special Issue
Seamless Tracing of Human Behavior Using Complementary Wearable and House-Embedded Sensors

Sensors 2014, 14(5), 8363-8379; doi:10.3390/s140508363

Article
An Investigation on the Feasibility of Uncalibrated and Unconstrained Gaze Tracking for Human Assistive Applications by Using Head Pose Estimation
Dario Cazzato 1,*, Marco Leo 2 and Cosimo Distante 2
1
Faculty of Engineering, University of Salento, Lecce 73100, Italy
2
National Research Council of Italy—Institute of Optics, Arnesano (LE) 73010, Italy; E-Mails: marco.leo@cnr.it (M.L.); cosimo.distante@cnr.it (C.D.)
Author Contributions: Dario Cazzato conceived the proposed approach, designed the architecture and prepared the manuscript. Marco Leo carried out the experiments, processed and interpreted results and contributed to the figures and manuscript preparation. Cosimo Distante co-designed the experimental strategies, contributed to the interpretation and discussion of results at all stages, and critically edited the manuscript. They all read and approved the final draft.
*
Author to whom correspondence should be addressed; E-Mail: dario.cazzato@unisalento.it; Tel.: +39-083-232-1816.
Received: 4 March 2014; in revised form: 11 April 2014 / Accepted: 4 May 2014 /
Published: 12 May 2014

Abstract

: This paper investigates the possibility of accurately detecting and tracking human gaze by using an unconstrained and noninvasive approach based on the head pose information extracted by an RGB-D device. The main advantages of the proposed solution are that it can operate in a totally unconstrained environment, it does not require any initial calibration and it can work in real-time. These features make it suitable for being used to assist human in everyday life (e.g., remote device control) or in specific actions (e.g., rehabilitation), and in general in all those applications where it is not possible to ask for user cooperation (e.g., when users with neurological impairments are involved). To evaluate gaze estimation accuracy, the proposed approach has been largely tested and results are then compared with the leading methods in the state of the art, which, in general, make use of strong constraints on the people movements, invasive/additional hardware and supervised pattern recognition modules. Experimental tests demonstrated that, in most cases, the errors in gaze estimation are comparable to the state of the art methods, although it works without additional constraints, calibration and supervised learning.
Keywords:
gaze estimation; head pose estimation; gaze tracking; human-computer interaction

1. Introduction

Gaze tracking plays a fundamental role in understanding human attention, feelings and desires [1]. Automatic gaze tracking provides several application in the fields of human-computer interaction (HCI) and human behavior analysis, therefore several techniques and methods have been investigated in recent years. When a person is in the field of view of a static camera, gaze can give information about the focus of attention of the subject, allowing for gaze-controlled interfaces for disabled people [2], driver attention monitoring [3], pilot training [4], provision of virtual eye contact in conferences [5] or marketing strategies analysis [6].

A survey of existing works and a detailed classification of extant methods can be viewed in [7]. Most gaze tracking methods are based on Pupil Center Corneal Reflection (PCCR) technique, [814]. They obtain the pose of the eye using the center of pupil contour and corneal reflections (glint) on the corneal surface from point light sources, usually one or multiple infrared (IR) lights. Normally this kind of approaches are not quite appropriate for generic interactive applications since a high-resolution camera is needed and a careful calibration is required for coupling IR lights and camera. Less invasive solutions that do not use IR are also available. The work of [15] proposes a method to achieve gaze estimation from multimodal Kinect data that is invariant to head pose, but it needs a learned person-specific 3D mesh model. In [16], after a one-time personal calibration, facial features are tracked and then used to estimate the 3D visual axis, proposing a 3D geometrical model of the eye. The method requires to accurately detect eye corners in order to create a complete 3D eye model. In [17] gaze tracking is performed using a stereo approach to detect the position and the orientation of the pupil in 3D space. A calibration procedure must be provided and, moreover, the reconstruction of the elliptic eye model cannot be well defined for all the gaze orientations. A low-cost system for 2D eye gaze estimation with low-resolution webcam images is presented in [18]: binary deformable eyeball template is modeled and 2D gaze estimation is performed depending on the displacement in eye movements and after a rigid calibration procedure. A method that, after an initial calibration, enables tracking motion of user's eye and gaze by using a single webcam—in a simplified, special case, i.e., when the face is still while the eyes move—is instead proposed in [19]. Valenti et al. [20] combine head pose and eye location informations to accurately estimate gaze track. Their results are suitable for several applications but, unfortunately, a calibration phase that makes use of a target plane is needed in order to get reference positions to extract eye gaze directionality.

All the aforesaid methods operate in constrained condition (e.g., a very short range of head pose variations) and they need a learning phase by which manually labeled data are used to train one or more classifiers. Furthermore, a calibration phase is often needed to set up the parameters performing the mapping between the real word and the computational models embedded in the algorithmic procedures. The predominant idea behind the works in this research field was that head pose estimation information can supply only a rough estimation of the human gaze. In other words, only the area of interest of the person can be retrieved and, therefore it is inadequate, if considered alone, to obtain accurate estimation of the gaze direction and to allow applications such as remote device control or in rehabilitation scenarios. For example, in [21] authors assert that the head pose contributes to about 70% of the visual gaze, and focus of attention estimation based on head orientation alone can get an average accuracy of 88.7% in a meeting application scenario. The influence that head orientation exerts on the perception of the eye-gaze direction is investigated in [22], where the authors conclude that an image-based mechanism is responsible for the influence of head profile on gaze perception, whereas the analysis of nose angle involves the processing of face features. In [23,24] head (and eventually body) pose information is used for estimating where a person is looking at. In [25,26] the visual focus of attention is recognized by evaluating only head pose information. More recently, the authors in [27] introduced a scene-specific gaze estimator for visual surveillance: it models the interactions between head motion, walking direction and appearance in order to recover gaze directions. Anyway, in the last years this point of view is being changing due to the new perspectives emerged from the exploitation of the most advanced sensorial technologies that allow the pose estimation to become more and more accurate. For example, in [28] the pioneering attempt to use head pose information extracted by a complex supervised algorithm working on 2D images in order to control a mouse is performed: the experimental results qualitatively showed the promise of the algorithm. To overcome the drawbacks of the related works, an early study on the estimation of visual focus of attention using fuzzy fusion of head rotations and eye gaze directions has been recently introduced in [29]. Two novel techniques to estimate head rotations, based on local and appearance information, are introduced and then fused in a common framework. Anyway, this framework does not focus on inferring exact gaze estimation but, rather, it detects degrees of confidence, through fuzzy logic, regarding hypotheses that a person is looking towards a specific point.

Unfortunately, no studies have still been performed on the feasibility of an accurate gaze estimator based only on head pose information. To fill this gap, in the proposed work, an innovative approach to achieve the exact position of gaze tracking ray from data acquired from a low cost depth sensor device is introduced. The proposed solution estimates the head pose of a subject freely moving on the environment, requiring only the presence of his head in the field of view of the sensor, in order to directly derive his 3D gaze ray. Neither training nor calibration phase are required to accomplish the gaze estimation task. This is another important contribution of this paper with respect to leading approaches in the state of the art. In our work, quantitative evaluations of the gaze estimation accuracy have been achieved by a large experimental phase, with several different distance ranges and different people with diverse levels of knowledge of the performing task. The remaining sections of the paper are organized as follows: Section 2 discusses the methodological steps of the proposed approach whereas the experimental setup is introduced in Section 3. Finally, the experimental outcomes are reported in Section 4, while their discussion is reported in Section 5.

2. Overview of the Proposed System

The proposed solution works as follows: the input data are acquired from a commercial depth sensor providing as output both RGB and depth data, which are the input for the following algorithmic steps. First of all, face detection is performed on the RGB image by matching appearance with predetermined models. The detected faces are then tracked over time using local features and topological information. The available depth information is then used to iteratively match the tracked points with a 3D point cloud representing the face geometry. The 3D head pose is finally estimated in terms of yaw, pitch and roll angles, and the gaze vector is computed as the vector having the origin in the average point between the two detected 3D positions of the eyes and direction according to the estimated head pose. A block diagram of the overall system is showed in Figure 1.

2.1. Face Tracking and 3D Head Pose Estimation

The first algorithmic step performed on the acquired RGB images is aimed at detecting the human faces. This is done by an approach that consists of three steps [30]. A linear pre-filter is at firstly used to increase the detection speed. Then, a boosting chain [31] is applied to remove most of the non-faces from the candidate list and, finally, a color filter and a SVM filter are used to further reduce false alarms. When the system detects a face, characteristic points are identified and a parameterized face mask is automatically overlapped on the human face. After the first detection, it is then possible to track the detected face over time, reducing this way the computational load needed to process the input images. Detection of the characteristic points of the face and their temporal tracking are based on the Active Appearance Model (AAM) [32]. AAM contains a statistical model of the shape and its representation as grey-level appearance. The core of the algorithm is the matching procedure that involves finding the model parameters which minimize the difference between the given appearance and the synthesized model example, projected into the image. In order to improve tracking performances, temporal matching constraint and color information are included in the model, as suggested in [33]. The next step consists in building a 3D model of the detected face. This is done by the Iterative Closest Point (ICP) [34] technique by which a 3D point cloud model is iteratively aligned with the available 2D facial features (target). The algorithm revises the transformation, i.e., combination of translation and rotation, needed to minimize the distance between the model and the target. The used 3D face model is the Candide-3 [35], a parameterized mask specifically developed for model-based coding of human faces. It allows fast reconstruction with small computing overhead. It is invariant to operating conditions and it does not depend on a specific person. This model is based on 121 linked feature points which are stored in a vector g containing their (x, y, z) coordinates. The model is reshaped by the equation:

g t + 1 ( σ , α ) = g t + S σ + A σ
where g t + 1 is the updated vector, S and A are the Shape and Animation Units and σ, α contain shape and animation parameters.

When the distance between the Candide-3 model and the target face is minimized the depth information of the 121 feature points is extracted from the available depth map and it will represent the input of the head pose estimation block.

The head pose estimation supplies the information about rotation angles in terms of yaw, pitch and roll, and translations, in meters, that in this paper are assumed to be expressed considering as reference point the center of the sensor. Head pose estimation is a problem with 6 Degrees of Freedom (DoF), and it can be represented with the parameter vector p = [ ω x , ω y , ω x , t x , t y , t z ], where ωx, ωy, ωz are the rotation parameters and tx, ty, tz are the translation parameters. They define the 3-DoF rotation matrixes R3×3 as:

R = [ 1 ω z ω y ω z 1 ω x ω y ω x 1 ]
and the 3-DoF translation vector T3×1 as:
T = [ t x t y t z ]

The rigid motion of a head point X = [ x , y , z ] T between time t and time t + 1 is: X ( t + 1 ) = M × X ( t ), where M is defined in [36] as:

M = [ R T 0 1 ]

Let point X ( t ) be projected on the image plane in u = [ u x u y ] T. The explicit representation of the perspective projection function in terms of the rigid motion vector parameters and the coordinates of the point at t + 1 is:

u ( t + 1 ) = [ x y ω z + z ω y + t x x ω z + y z ω x + t y ] f L x ω y + + y ω x + z + t z ( t )
where fL is the focal length.

In order to fuse rotation and translation information into the Candide-3 model, Equation (1) is modified as:

g ( σ , α ) = R s ( g + S σ + A α ) + T
where s represents the scale. Thus, vector p has now components:
p = [ ω x , ω y , ω z , s , t x , t y , t z , σ , α ]

At the end, the position of the user's head is expressed in world coordinate X, Y, and Z which are reported based on a right-handed coordinate system with the origin at the sensor, Z pointed towards the user and Y pointed up.

Figure 2 shows the 3D mask overlapped to the 2D facial image in three different frames. From the figure it is possible to observe that the face tracker works also in presence of non frontal views.

2.2. Gaze Estimation

Gaze estimation step is based on the 3D model coming from the previous blocks. It geometrically models the gaze ray direction with regard to the 3D position of the sensor and thus it can be categorized as model based method. First of all, the 3D positions of the eye centers are extracted from the actual position of the overlapped 3D face model. After that, in order to define a point on the face from which the computed gaze vector takes its origin, a conventional point in the middle of the segment connecting the 3D eye center positions is taken. This point approximatively corresponds to the nose septum and it is used as the origin of the gaze track. Note that small occlusions are handled by the system, and that the eye center point is always estimated when the overlapping with the face successes. Moreover, in this way it is not necessary to use a precise pupil detector, since the computed point on the face is enough to completely solve the geometric problem. At this point, exploiting available head pose information, the direction of the gaze track is derived from the angles ωx and ωy, corresponding respectively to pitch and yaw of the vector p . Then the intersection of the gaze track with a vertical plane, parallel to the image plane of the sensor, is computed. Actually, the intersection point is computed separately for the x and y axes. In Figure 3 the procedure along the x axis is shown and it is described in the followings. The depth sensor is able to give the information about the length of the segment A B ¯ as the component tz of the translation vector T. It follows that, knowing a side and an angle, we can completely solve the right-angled triangle ABC ̂ In particular A B ¯ = A C ¯ cos ω y and B C ¯ = A B ¯ 2 A C ¯ 2. Using the same coordinate system, it is possible to compute also the cartesian equation of the gaze ray as the straight line passing for points A = (xa, ya, za) and B = (xB, yB, zb) expressed as:

r : { x x A x B x A = y y A y B y A y y A y B y A = z z A z B z A
with za = 0 for the particular plane under consideration.

In case of translations on the x and y axes, the vector can be algebraical summed up with the computed value, in order to translate the gaze vector to the right position. Finally, in order to represent on a monitor the actual intersection point between the gaze vector and the plane, the world coordinates are normalized to image plane according to:

x = X L R L I x y = Y T T B I y
where: (X, Y) and (x, y) are the world and image plane coordinates, respectively; L, R, T, B are the left, right, top and bottom bounds of the considered user's field of view; Ix and Iy are the width and height of the displaying area on the monitor (in pixels).

3. Experimental Setup

The experimental setup was defined as follows: a Microsoft Kinect device was used as depth sensor and it was positioned at a height of 150 cm from the ground. Behind the sensor a square panel (2 m per side) was positioned and 15 circular markers were stuck on it. The markers were distributed on three rows, 5 markers on each row, with a distance of 50 cm from each other. Markers were divided into subsets as showed in Figure 4 in order to group together points that presented the same distance from the sensor in terms of x,y or both axes, from P1 to P5, while P0 corresponds to the depth sensor position. For example, P4 are the points with a distance of 1 meter from the sensor along the x axes and aligned along y axes, and so on. The depth sensor was placed in correspondence of the marker at the centre of the panel.

Figure 5 shows one quarter of the panel, exactly the upper-leftmost. The 3D positions of the markers represented the ground truth information for the following experimental phase (Section 4) is devoted to the evaluation of the system's capability to estimate the points of regard of the users.

Figure 6 shows instead a picture of the monitor where gaze hits can be drawn, making use of the Equation (9): the red circle represents the point of the panel where the user is meanwhile looking at. Green circles represent instead the projection onto the monitor of the markers placed on the panel surface. The sensor's working ranges are 43° vertical, 57° horizontal and [40 cm, 300 cm] in depth [37].

Three different distances between the user and the panel were considered in the experiments (i.e., 70 cm, 150 cm and 250 cm) (The exact distance between the user and the panel was retrieved by adding 4 cm to the computed depth map in order to take into account the displacement of the physical position of the sensor with respect the plane of the panel.): in this operating conditions the head pose angles range in the intervals [−56.0 °, +56.0 ° ] for the yaw, [−35.5 °, +35.5 ° ] for the pitch and [−90.0 °, +90.0 ° ] for the roll.

A scheme of the experimental setup illustrating the three different users positions is shown in Figure 7. Notice that, the above operating ranges allow the user to freely move the head in all the directions and in particular, despite of most of the state-of-the-art methods making use of the Viola-Jones face detector [31], also large rolling movements can be handled.

4. Experimental Results

The proposed method was tested with nine different persons. According to different similar works in this research field (for example in [38]), in order to get a comprehensive study, persons were divided into three groups, three persons for each group. The first group was composed by experienced persons, i.e., persons that knew how the system works and that already had tried the system before the test session. The second group was composed by persons that were trying the system for the first time but that had been informed about how the system works. In light of that knowledge, it was more probable that they would have moved the head even in case of sequential pointing of close markers onto the panel. Finally, in the last group there were unaware people who were just placed in front of the sensor and they were asked to point towards the markers. No constraints were given to the participants in terms of eyeglasses, beard or hairstyle and, in order to allow for wild settings, no panel or uniform background color were put behind the participants. These three experimental benchmarks permitted to verify the system's accuracy in relation to different levels of awareness of the users, which may be encountered in different applicative contexts.

The experiment was made as follows for all three groups: persons were asked to look at each of the markers onto the panel, in a predefined order. The gaze direction relative to a given marker, was that one estimated by the system when the person confirmed, by an oral feedback, that the marker was its point of regard. The estimated gaze direction was projected onto the panel and then compared with manually computed ground truth data. Therefore the errors were measured as the distances between the estimated intersection points and the ground truth data. The errors were expressed both in centimeters, as well as by the difference between the angles described by the estimated and ground truth rays.

The outcomes of the experimental tests are shown in Tables 13. The first column reports the labels of the markers under consideration (see Section 3 for label assignment) whereas the second column shows the tested distances, i.e., 70 cm, 150 cm and 250 cm. Errors were computed separately for each group and for yaw and pitch angles. Also error standard deviation, reported in degrees, was taken into account and reported in tables. Note that “n.a.” stands for not available data, corresponding to missing overlapping of the Candide-3 model on the current face.

From Table 1, it can be observed that the results for the first group of persons were very accurate: the average error was at maximum about 3 degrees, encountered when the persons stand at 70 cm from the panel, with a standard deviation of the error that was about 1.5 degrees in both directions. This is the prove that the proposed system is well suited for those application contexts where the users can be learned to exploit at best its functionalities: e.g., remote control of the device's cursor in cases of physical impairments. Table 2 demonstrates that the system reported encouraging results also on the second benchmark of persons, i.e., the informed ones: the average errors slightly increased, compared with those achieved for the group 1, to about 4.5 degrees and some peaks in the standard deviation were accounted due to the slowness of the corresponding person to become familiar with the system. The results reported on this group demonstrated that the proposed system can be exploited in all those application contexts where users can be preliminarily informed about the right modalities to interact (e.g., for gaming purposes, where a quick information session generally precedes the start of the game).

Finally, from Table 3, reporting the results on the third group of persons , it is possible to derive that the errors in accuracy still remained under 12 degrees (often around 5 degrees) with a standard deviation ranging from 2 to 6 degrees. In our opinion, this is a very interesting result considering that this group of persons are completely unaware about the system. This demonstrates that the proposed system can be exploited also in those applications where it is not possible to constraint the user behavior, e.g., in assistive applications involving persons with neurological impairments or audience measurements.

Notice also that the gaze estimation accuracy, independently from the benchmark under investigation, remained satisfactory even when the user's distance increased from the depth sensor (in particular the errors encountered at a distance of 150 cm were very encouraging). By our knowledge, this paper represents the first investigation of the performance of a gaze estimation in those challenging conditions: all the state-of-the-art methods, in fact, reports the accuracy achieved at a distance less than 1 m between user and target. In light of this, in order to evaluate the accuracy of the proposed method with respect the leading approaches in the literature, a further experiment was performed. In this further experiment the testing scenario mostly used in the promising works in the literature was set up. In that scenario the users are put in front of a screen, at a distance in the range [54 cm, 67 cm], and they are asked to observe a list of point on the screen.

In Table 4 the errors in estimating the gaze direction, of the proposed approach and of the leading methods mentioned in Section 1, are reported. Here we point out again that all the compared approaches make use of supervised algorithms and/or invasive device and/or a calibration and/or they limit the head movements.

From the table, it can be seen that in the case of experienced or informed users, only the state-of-the-art methods making use of additional training and/or calibration phases outperform in accuracy the proposed approach. In the case of unaware users, the experienced accuracy is a little bit lower than that of some comparing approaches but, it remains well suited for most of the main attractive application domains for gaze tracking systems (e.g., remote rehabilitation, healthcare, monitoring, audience measurement, etc.). This is a very encouraging results that open a new way to deal with the gaze estimation issue: it demonstrated that new sensorial technologies combined with robust head pose estimation algorithms can bring to a relaxation of the environmental constraints and to a simplification of the algorithmic steps involved in the gaze estimation approaches.

During the above experimental phases a further evaluation was done on the users in groups 1 and 2 in order to check the actual possibility of using the proposed system to remotely control a device. After each experimental session, each person was asked to look at a screen and to try to control the mouse pointer by using his gaze and finally to give a feedback about usability and familiarity. All of the participant feel comfortable and able to easily use the system as a control device.

Concerning computational remarks, the involved algorithms have been implemented using Microsoft Visual C++ developing environment and, running on an Ultrabook Intel i3 CPU @ 1.8 GHz with 4 GB of RAM, with RGB and depth images taken at a resolution of 640 × 480, 30 fps, during the experimental phase the system was able to work in real-time.

5. Conclusions

This work presented an investigation on the feasibility of a gaze estimation system working in an unconstrained and noninvasive environment and that does not require any additional hardware (like IR light sources, wearable devices, etc.). The proposed solution makes use of a low cost commercial depth sensor and it estimates head pose information by combining RGB and depth data. This method has then been tested with both trained and untrained persons in an unconstrained setting, and errors have been quantitatively measured. In addition, it has been also carefully compared with the leading approaches in the state of the art, showing that their errors are comparable even if the considered approaches work well only under particular conditions and/or if specialized hardware (sometime invasive) is available. For this reason, the proposed approach is more suitable for most of the main attractive application domains concerning unconstrained gaze tracking systems like remote rehabilitation, therapy supply or ambient assistive living. Another advantage of the proposed solution is that it makes use of commercial hardware and no calibration phase is required. This could make it exploitable also from non experts user and then it can become a technological support in related research fields, for example for studying social human behaviors or being used in socially assistive robotics during human-robot interaction. Furthermore, we are aware that in particular application domains, e.g., for studying particular neurological diseases or to realize hand-free control of mobile devices, is also indispensable to combine head pose information with an accurate pupil center locator and this investigation will be the subject of our future research works.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jellinger, K. Cognitive Processes in Eye Guidance. Eur. J Neurol. 2006, 13. [Google Scholar] [CrossRef]
  2. Mateo, J.C.; San Agustin, J.; Hansen, J.P. Gaze beats mouse: Hands-free selection by combining gaze and emg. Proceedings of the CHI'08 Extended Abstracts on Human Factors in Computing Systems, Florence, Italy, 5–10 April 2008; pp. 3039–3044.
  3. Doshi, A.; Trivedi, M.M. Attention estimation by simultaneous observation of viewer and view. Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), San Francisco, CA, USA, 13–18 June 2010; pp. 21–27.
  4. Wetzel, P.A.; Krueger-Anderson, G.; Poprik, C.; Bascom, P. An Eye Tracking System for Analysis of Pilots' Scan Paths; Human Resources Directorate, Aircrew Training Research Division, U.S. Air Force, Armstrong Laboratory: Brooks City-Base, TX, USA, 1997. [Google Scholar]
  5. Waizenegger, W.; Atzpadin, N.; Schreer, O.; Feldmann, I.; Eisert, P. Model based 3D gaze estimation for provision of virtual eye contact. Proceedings of the 2012 19th IEEE International Conference on Image Processing (ICIP), Orlando, FL, USA, 30 September–3 October 2012; pp. 1973–1976.
  6. Pieters, R. A review of eye-tracking research in marketing. Rev Market. Res. 2008, 4, 123–147. [Google Scholar]
  7. Hansen, D.; Ji, Q. In the eye of the beholder: A survey of models for eyes and gaze. IEEE Pattern Anal. Mach. Intell. 2012. doi: 10.1109/TPAMI.2009.30. [Google Scholar]
  8. Morimoto, C.H.; Mimica, M.R. Eye gaze tracking techniques for interactive applications. Comput. Vis. Image Underst. 2005, 98, 4–24. [Google Scholar]
  9. Guestrin, E.D.; Eizenman, M. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 2006, 53, 1124–1133. [Google Scholar]
  10. Nakazawa, A.; Nitschke, C. Point of Gaze Estimation through Corneal Surface Reflection in an Active Illumination Environment. In Computer Vision–ECCV 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 159–172. [Google Scholar]
  11. Villanueva, A.; Cabeza, R. A novel gaze estimation system with one calibration point. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2008, 38, 1123–1138. [Google Scholar]
  12. Zhu, Z.; Ji, Q. Novel eye gaze tracking techniques under natural head movement. IEEE Trans. Biomed. Eng. 2007, 54, 2246–2260. [Google Scholar]
  13. Yoo, D.H.; Chung, M.J. A novel non-intrusive eye gaze estimation using cross-ratio under large head motion. Comput. Vis. Image Underst. 2005, 98, 25–51. [Google Scholar]
  14. Noureddin, B.; Lawrence, P.; Man, C. A non-contact device for tracking gaze in a human computer interface. Comput. Vis. Image Underst. 2005, 98, 52–82. [Google Scholar]
  15. Funes Mora, K.; Odobez, J.M. Gaze estimation from multimodal Kinect data. Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Providence, RI, USA, 16–21 June 2012; pp. 25–30.
  16. Chen, J.; Ji, Q. 3D gaze estimation with a single camera without IR illumination. Proceedings of the 19th International Conference on IEEE Pattern Recognition (ICPR 2008), Tampa, FL, USA, 8–11 December 2008; pp. 1–4.
  17. Kohlbecher, S.; Bardinst, S.; Bartl, K.; Schneider, E.; Poitschke, T.; Ablassmeier, M. Calibration-free eye tracking by reconstruction of the pupil ellipse in 3D space. In Proceedings of the 2008 Symposium on Eye Tracking Research & Applications, Savannah, GA, USA, 26–28 March 2008; pp. 135–138.
  18. Ince, I.F.; Kim, J.W. A 2D eye gaze estimation system with low-resolution webcam images. EURASIP J. Adv. Signal Process. 2011, 2011, 1–11. [Google Scholar]
  19. Janko, Z.; Hajder, L. Improving Human-Computer Interaction by gaze tracking. Proceedings of the 2012 IEEE 3rd International Conference on IEEE Cognitive Infocommunications (CogInfoCom), Kosice, Slovakia, 2–5 December 2012; pp. 155–160.
  20. Valenti, R.; Sebe, N.; Gevers, T. Combining head pose and eye location information for gaze estimation. IEEE Trans. Image Process. 2012, 21, 802–815. [Google Scholar]
  21. Stiefelhagen, R.; Zhu, J. Head orientation and gaze direction in meetings. Proceedings of the CHI'02 Extended Abstracts on Human Factors in Computing Systems, Minneapolis, MN, USA, 20–25 April 2002; pp. 858–859.
  22. Langton, S.R.; Honeyman, H.; Tessler, E. The influence of head contour and nose angle on the perception of eye-gaze direction. Percept. Psychophys. 2004, 66, 752–771. [Google Scholar]
  23. Brolly, X.L.; Stratelos, C.; Mulligan, J.B. Model-based head pose estimation for air-traffic controllers. Proceedings of the 2003 International Conference on IEEE Image Processing (ICIP 2003), Barcelona, Spain, 14–17 September 2003; Volume 2. doi:10.1109/ICIP.2003.1246629.
  24. Robertson, N.; Reid, I.; Brady, J. What are you looking at? gaze estimation in medium-scale images. Proceedings of the HAREM Workshop (in assoc. with BMVC), Oxford, UK, 9 September 2005.
  25. Ba, S.O.; Odobez, J.M. Recognizing visual focus of attention from head pose in natural meetings. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2009, 39, 16–33. [Google Scholar]
  26. Zhang, W.Z.; Wang, Z.C.; Xu, J.K.; Cong, X.Y. A Method of Gaze Direction Estimation Considering Head Posture. Int. J. Signal Process. Image Process Pattern Recognit. 2013, 6, 103–111. [Google Scholar]
  27. Benfold, B.; Reid, I. Unsupervised learning of a scene-specific coarse gaze estimator. Proceedings of the 2011 IEEE International Conference on IEEE Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2344–2351.
  28. Nabati, M.; Behrad, A. Robust facial 2D motion model estimation for 3D head pose extraction and automatic camera mouse implementation. Proceedings of the 2010 5th International Symposium on IEEE Telecommunications (IST), Tehran, Iran, 4–6 December 2010; pp. 817–824.
  29. Asteriadis, S.; Karpouzis, K.; Kollias, S. Visual Focus of Attention in Non-calibrated Environments using Gaze Estimation. Int. J. Comput. Vis. 2013, 107, 1–24. [Google Scholar]
  30. Xiao, R.; Li, M.J.; Zhang, H.J. Robust multipose face detection in images. IEEE Trans. Circuits Syst. Video Technol. 2004, 14, 31–41. [Google Scholar]
  31. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. Proceedings of the 2001 IEEE Computer Society Conference on IEEE Computer Vision and Pattern Recognition (CVPR 2001), Kauai, HI, USA, 8–14 December 2001; Volume 1, pp. 511–518.
  32. Cootes, T.; Edwards, G.; Taylor, C. Active appearance models. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 681–685. [Google Scholar]
  33. Zhou, M.; Liang, L.; Sun, J.; Wang, Y. AAM based face tracking with temporal matching and face segmentation. Proceedings of the 2010 IEEE Conference on IEEE Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 701–708.
  34. Besl, P.; McKay, N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar]
  35. Ahlberg, J. Candide-3-an Updated Parameterised Face; Report No. LiTH-ISY-R-2326. Department of Electrical Engineering, Linkping University: Linkping, Sweden, 2001. [Google Scholar]
  36. Li, Z.; Sastry, S.S.; Murray, R. A Mathematical Introduction to Robotic Manipulation; CRC Press: Boca Raton, FL, USA, 1994. [Google Scholar]
  37. MSDN-Microsoft. Available online: http://msdn.microsoft.com/en-us/library/jj131033.aspx (accessed on 1 April 2014).
  38. Lu, F.; Okabe, T.; Sugano, Y.; Sato, Y. A Head Pose-free Approach for Appearance-Based Gaze Estimation. Proceedings of the BMVC, Dundee Scotland, 29 August–2 September 2011; pp. 1–11.
  39. Lu, F.; Okabe, T.; Sugano, Y.; Sato, Y. Learning gaze biases with head motion for head pose-free gaze estimation. Image Vis. Comput. 2014, 32, 169–179. [Google Scholar]
  40. Sugano, Y; Matsushita, Y.; Sato, Y.; Koike, H. An Incremental Learning Method for Unconstrained Gaze Estimation. In Computer Vision–ECCV 2008; Springer: Berlin/Heidelberg: Germany, 2008; pp. 656–667. [Google Scholar]
  41. Guestrin, E.; Eizenman, E. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 2006, 53, 1124–1133. [Google Scholar]
Sensors 14 08363f1 1024
Figure 1. A block diagram of the proposed method.

Click here to enlarge figure

Figure 1. A block diagram of the proposed method.
Sensors 14 08363f1 1024
Sensors 14 08363f2 1024
Figure 2. Three different snapshots of the face tracking module.

Click here to enlarge figure

Figure 2. Three different snapshots of the face tracking module.
Sensors 14 08363f2 1024
Sensors 14 08363f3 1024
Figure 3. A scheme of the gaze estimation solution.

Click here to enlarge figure

Figure 3. A scheme of the gaze estimation solution.
Sensors 14 08363f3 1024
Sensors 14 08363f4 1024
Figure 4. The used grouping scheme for target points during tests.

Click here to enlarge figure

Figure 4. The used grouping scheme for target points during tests.
Sensors 14 08363f4 1024
Sensors 14 08363f5 1024
Figure 5. A portion of the panel used in the experimental phase.

Click here to enlarge figure

Figure 5. A portion of the panel used in the experimental phase.
Sensors 14 08363f5 1024
Sensors 14 08363f6 1024
Figure 6. A picture of the monitor where gaze hits can be drawn.

Click here to enlarge figure

Figure 6. A picture of the monitor where gaze hits can be drawn.
Sensors 14 08363f6 1024
Sensors 14 08363f7 1024
Figure 7. The different users positions for the experimental phase.

Click here to enlarge figure

Figure 7. The different users positions for the experimental phase.
Sensors 14 08363f7 1024
Table Table 1. Experiments with the first group of experienced persons: they knew how the system works and that already tried the system before the test session.

Click here to display table

Table 1. Experiments with the first group of experienced persons: they knew how the system works and that already tried the system before the test session.
Errors
x (cm)x (deg)y (cm)y (deg)Std x (deg)Std y (deg)
P070 cm1.501.222.662.180.340.87
150 cm3.501.334.831.840.540.93
250 cm6.001.378.501.940.670.85
P170 cmn.a.n.a.n.a.n.a.n.an.a
150 cm6.001.618.002.790.931.36
250 cm2.600.524.000.880.200.37
P270 cm8.775.037.664.372.362.09
150 cm0.160.0511.834.150.011.66
250 cm4.330.955.831.290.380.52
P370 cm5.614.583.501.942.331.06
150 cm6.832.608.833.080.981.20
250 cm4.661.061.830.400.441.50
P470 cmn.a.n.a.n.a.n.a.n.a.n.a.
150 cm0.500.139.333.560.071.95
250 cm0.660.1315.663.470.491.97
P570 cm3.662.033.833.131.111.58
150 cm0.330.114.161.590.040.60
250 cm4.330.958.161.870.360.97
Total Averages70 cm4.883.224.412.901.531.40
150 cm2.880.977.832.830.421.28
250 cm3.770.837.251.640.421.03
Table Table 2. Experiments with the second group persons that were trying the system for the first time but that have been informed how it works.

Click here to display table

Table 2. Experiments with the second group persons that were trying the system for the first time but that have been informed how it works.
Errors
x (cm)x (deg)y (cm)y (deg)Std x (deg)Std y (deg)
P070 cm2.502.042.331.900.530.57
150 cm6.242.387.412.830.951.09
250 cm28.56.50132.971.951.18
P170 cmn.a.n.a.n.a.n.a.n.a.n.a.
150 cm2.000.5321.587.700.303.61
250 cm19.003.8427.006.052.302.96
P270 cm10.165.8917.1610.402.595.30
150 cm5.831.8919.086.780.923.45
250 cm3.001.500.650.330.790.33
P370 cm2.832.318.334.771.012.85
150 cm15.836.0213.334.693.992.83
250 cm18.54.2320.54.581.982.12
P470 cmn.a.n.a.n.a.n.a.n.a.n.a.
150 cm4.751.2715.085.740.772.91
250 cm33.006.7912.502.863.662.01
P570 cm11.160.836.510.680.360.29
150 cm7.502.533.001.141.100.59
250 cm22.505.0319.004.342.731.52
Total Averages70 cm6.664.197.164.441.122.25
150 cm6.982.4413.254.811.332.41
250 cm20.754.5115.583.522.231.68
Table Table 3. Experiments with the third group of persons that were totally unaware of how the system works.

Click here to display table

Table 3. Experiments with the third group of persons that were totally unaware of how the system works.
Errors
x (cm)x (deg)y (cm)y (deg)Std x (deg)Std y (deg)
P070 cm4.003.270.000.001.920.00
150 cm15.005.7110.003.813.012.19
250 cm71.0012.0015.852.748.862.03
P170 cmn.a.n.a.n.a.n.a.n.a.n.a.
150 cm17.004.7329.0010.463.157.61
250 cm62.0013.1534.007.647.964.19
P270 cm24.9923.1334.6123.1312.1714.01
150 cm1.780.6110.903.820.491.38
250 cm27.006.0517.003.793.421.83
P370 cm24.6119.3710.105.1110.332.91
150 cm33.2012.4817.106.067.083.20
250 cm90.2319.8412.62.8010.051.18
P470 cmn.a.n.a.n.a.n.a.n.a.n.a.
150 cm3.901.014.801.830.730.81
250 cm12.542.5224.675.631.893.21
P570 cm19.804.1012.203.351.871.79
150 cm2.780.9617.386.600.494.03
250 cm4.440.9824.245.530.413.31
Total Averages70 cm18.3512.6712.207.906.574.67
150 cm12.274.2514.865.432.493.20
250 cm44.619.7520.754.695.432.62
Table Table 4. Head motion ranges (in world coordinate system) used in the final evaluation.

Click here to display table

Table 4. Head motion ranges (in world coordinate system) used in the final evaluation.
MethodCategoryReported ErrorCamera(s)Additional Requirements
Proposed (experienced users)Model3.1°Depth SensorNone
Proposed (informed users)Model3.6°Depth SensorNone
Proposed (unaware users)Model6.9°Depth SensorNone
Lu et al. [39]Appearance2–3°1Capture video
Sugano et al. [40]Appearance4–5°1≈ 103 training samples
Nakazawa and Nitschke [10]Model0.9°1 IRIR LEDs & projector
Villanueva and Cabeza [11]Model1 IR2–4 IR LEDs
Zhu and Ji [12]Model2 IRn IR LEDs
Guestrin and Eizenman [41]Model1–3°1 IR2 IR LEDs
Yoo and Chung [13]Model1–2.5°2 IR5 IR LEDs
Noureddin et al. [14]Model1–3°2–4 IRIR LEDs + Mirrors
Sensors EISSN 1424-8220 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert