Robust Camera-Based Eye-Tracking Method Allowing Head Movements and Its Application in User Experience Research
Highlights
- This paper proposes a robust camera-based eye-tracking method that effectively accommodates head movements.
- Comprehensive evaluations through a gaze task confirm that the proposed approach achieves higher accuracy than most existing methods, both with and without head motion. Further validation via a smooth pursuit task demonstrates its efficacy in tracking dynamic visual targets.
- A case study in user experience research illustrates the practical value of the method, showcasing its potential for real-world setting.
Abstract
1. Introduction
- We proposed a robust camera-based eye-tracking method that allows head movements.
- A fixation task was performed to test the accuracy. The results showed that our method outperforms most existing methods under head fixation and head motion.
- A smooth pursuit task was performed to verify the performance of our method when gazing at a moving target.
- We also conducted a case study to demonstrate the effectiveness of our method in user experience research.
2. Related Work
2.1. User Experience Analysis by Eye-Tracking
2.2. Camera-Based Eye-Tracking Methods
3. Methodology
3.1. Features Extraction
3.2. Calibration Phase
3.3. Eye-Tracking Phase
3.3.1. Head State Detection
3.3.2. Head-Pointing Computation
3.3.3. Benchmark Features Updating
3.3.4. Gaze Point Prediction
4. Experiment
4.1. Participants
4.2. Task 1. Fixation Task
4.2.1. Experimental Procedure
4.2.2. Experimental Results
4.3. Task 2. Smooth Pursuit
4.3.1. Experimental Procedure
4.3.2. Experimental Results
5. Case Study
5.1. Participants
5.2. Experimental Material
5.3. Experimental Procedure
5.4. Experimental Results
5.4.1. Results of Subjective Experiences and Game Score
5.4.2. Results of Eye-Tracking Data
5.4.3. Analysis of Subjective Experiences and Objective Eye-Tracking Results
6. Discussion
6.1. Comparison Against CNN Models
6.2. Comparison Against Professional Eye Trackers
6.3. Limitations
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Mock, J.; Huber, S.; Klein, E.; Moeller, K. Insights into numerical cognition: Considering eye-fixations in number processing and arithmetic. Psychol. Res. 2016, 80, 334–359. [Google Scholar] [CrossRef]
- Peake, C.; Moscoso-Mellado, J.; Guerra, E. First fixation duration as a bottom-up measure during symbolic and non-symbolic numerical comparisons. Stud. Psychol. 2020, 41, 563–579. [Google Scholar]
- Naeini, A.B.; Mahdipour, A.G.; Dorri, R. Using eye tracking to measure overall usability of online grocery shopping websites. Int. J. Mob. Comput. Multimed. Commun. (IJMCMC) 2023, 14, 1–24. [Google Scholar] [CrossRef]
- Modi, N.; Singh, J. Real-time camera-based eye gaze tracking using convolutional neural network: A case study on social media website. Virtual Real. 2022, 26, 1489–1506. [Google Scholar] [CrossRef]
- Zhang, H.; Yin, L.; Zhang, H. A real-time camera-based gaze-tracking system involving dual interactive modes and its application in gaming. Multimed. Syst. 2024, 30, 15. [Google Scholar] [CrossRef]
- Yang, X.; Krajbich, I. Webcam-based online eye-tracking for behavioral research. Judgm. Decis. Mak. 2021, 16, 1485–1505. [Google Scholar] [CrossRef]
- Modi, N.; Singh, J. Understanding online consumer behavior at e-commerce portals using eye-gaze tracking. Int. J. Hum.-Interact. 2023, 39, 721–742. [Google Scholar] [CrossRef]
- Wang, X.; Zhang, J.; Zhang, H.; Zhao, S.; Liu, H. Vision-based gaze estimation: A review. IEEE Trans. Cogn. Dev. Syst. 2021, 14, 316–332. [Google Scholar] [CrossRef]
- Guo, Z.; Zhou, Q.; Liu, Z. Appearance-based gaze estimation under slight head motion. Multimed. Tools Appl. 2017, 76, 2203–2222. [Google Scholar] [CrossRef]
- Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.L.; Yong, M.G.; Lee, J.; et al. Mediapipe: A framework for building perception pipelines. arXiv 2019, arXiv:1906.08172. [Google Scholar] [CrossRef]
- Liu, W.; Cao, Y.; Proctor, R.W. How do app icon color and border shape influence visual search efficiency and user experience? Evidence from an eye-tracking study. Int. J. Ind. Ergon. 2021, 84, 103160. [Google Scholar] [CrossRef]
- Deng, L.; Zhang, Z.; Zhou, F.; Liu, R. Effects of app icon border form and interface background color saturation on user visual experience and search performance. Adv. Multimed. 2022, 2022, 1166656. [Google Scholar] [CrossRef]
- Chen, T.; Samaranayake, P.; Cen, X.; Qi, M.; Lan, Y.C. The impact of online reviews on consumers’ purchasing decisions: Evidence from an eye-tracking study. Front. Psychol. 2022, 13, 865702. [Google Scholar] [CrossRef]
- AlSalem, T.S.; AlShamari, M.A. Assessing Interactive Web-Based Systems Using Behavioral Measurement Techniques. Future Internet 2023, 15, 365. [Google Scholar] [CrossRef]
- Li, D.; Zhou, H.; Zhou, S.; Huang, G.; Ma, X.; Zhao, Y.; Wang, W.; Ng, S.T. An eye-tracking-based approach to evaluate the usability of government portal websites in pilot smart cities. Eng. Constr. Archit. Manag. 2025, 32, 2369–2396. [Google Scholar] [CrossRef]
- Zammarchi, G.; Frigau, L.; Mola, F. Markov chain to analyze web usability of a university website using eye tracking data. Stat. Anal. Data Min. ASA Data Sci. J. 2021, 14, 331–341. [Google Scholar] [CrossRef]
- Shi, R.; Wang, M.; Qiao, T.; Shang, J. The effects of live streamer’s facial attractiveness and product type on consumer purchase intention: An exploratory study with eye tracking technology. Behav. Sci. 2024, 14, 375. [Google Scholar] [CrossRef] [PubMed]
- He, S.j.; Zhi, J.y.; Du, Y.; Feng, S. Effects of design features of train electronic guidance interface on visual search, behavior, and usability. Int. J. Ind. Ergon. 2023, 93, 103387. [Google Scholar] [CrossRef]
- Guo, F.; Chen, J.; Li, M.; Lyu, W.; Zhang, J. Effects of visual complexity on user search behavior and satisfaction: An eye-tracking study of mobile news apps. Univers. Access Inf. Soc. 2022, 21, 795–808. [Google Scholar] [CrossRef]
- Guo, F.; Ding, Y.; Liu, W.; Liu, C.; Zhang, X. Can eye-tracking data be measured to assess product design?: Visual attention mechanism should be considered. Int. J. Ind. Ergon. 2016, 53, 229–235. [Google Scholar] [CrossRef]
- Cybulski, P.; Horbiński, T. User experience in using graphical user interfaces of web maps. ISPRS Int. J. Geo-Inf. 2020, 9, 412. [Google Scholar] [CrossRef]
- Jiang, J.Y.; Guo, F.; Chen, J.H.; Tian, X.H.; Lv, W. Applying eye-tracking technology to measure interactive experience toward the navigation interface of mobile games considering different visual attention mechanisms. Appl. Sci. 2019, 9, 3242. [Google Scholar] [CrossRef]
- Lan, H.; Liao, S.; Kruger, J.L.; Richardson, M.J. Processing Written Language in Video Games: An Eye-Tracking Study on Subtitled Instructions. J. Eye Mov. Res. 2025, 18, 44. [Google Scholar] [CrossRef]
- Krebs, C.; Falkner, M.; Niklaus, J.; Persello, L.; Klöppel, S.; Nef, T.; Urwyler, P. Application of eye tracking in puzzle games for adjunct cognitive markers: Pilot observational study in older adults. JMIR Serious Games 2021, 9, e24151. [Google Scholar] [CrossRef]
- Wang, H.; Yang, J.; Hu, M.; Tang, J.; Yu, W. A comparative analysis for eye movement characteristics between professional and non-professional players in FIFA eSports game. Displays 2024, 81, 102599. [Google Scholar]
- Valenti, R.; Sebe, N.; Gevers, T. Combining head pose and eye location information for gaze estimation. IEEE Trans. Image Process. 2011, 21, 802–815. [Google Scholar] [CrossRef]
- Markuš, N.; Frljak, M.; Pandžić, I.S.; Ahlberg, J.; Forchheimer, R. Eye pupil localization with an ensemble of randomized trees. Pattern Recognit. 2014, 47, 578–587. [Google Scholar] [CrossRef]
- Valenti, R.; Gevers, T. Accurate eye center location through invariant isocentric patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 1785–1798. [Google Scholar] [CrossRef] [PubMed]
- Shih, S.W.; Liu, J. A novel approach to 3-D gaze tracking using stereo cameras. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2004, 34, 234–245. [Google Scholar] [CrossRef]
- Lee, J.M.; Lee, H.C.; Gwon, S.Y.; Jung, D.; Pan, W.; Cho, C.W.; Park, K.R.; Kim, H.C.; Cha, J. A new gaze estimation method considering external light. Sensors 2015, 15, 5935–5981. [Google Scholar] [CrossRef] [PubMed]
- Papoutsaki, A.; Sangkloy, P.; Laskey, J.; Daskalova, N.; Huang, J.; Hays, J. WebGazer: Scalable Webcam Eye Tracking Using User Interactions. In Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI-16); AAAI: New York, NY, USA, 2016; pp. 3839–3845. [Google Scholar]
- Blignaut, P. Mapping the pupil-glint vector to gaze coordinates in a simple video-based eye tracker. J. Eye Mov. Res. 2014, 7, 1–11. [Google Scholar] [CrossRef]
- Chhimpa, G.; Loura, A.; Garhwal, S.; Sangwan, D. Development of a real-time eye movement-based computer interface for communication with improved accuracy for disabled people under natural head movements. J. Real-Time Image Process. 2023, 20, 81. [Google Scholar] [CrossRef]
- Xia, L.; Sheng, B.; Wu, W.; Ma, L.; Li, P. Accurate gaze tracking from single camera using gabor corner detector. Multimed. Tools Appl. 2014, 75, 221–239. [Google Scholar] [CrossRef]
- Ohno, T.; Mukawa, N. A free-head, simple calibration, gaze tracking system that enables gaze-based interaction. In Proceedings of the 2004 Symposium on Eye Tracking Research & Applications, San Antonio, TX, USA, 22–24 March 2004; pp. 115–122. [Google Scholar]
- Park, K.R. A real-time gaze position estimation method based on a 3-D eye model. IEEE Trans. Syst. Man, Cybern. Part B Cybern. 2007, 37, 199–212. [Google Scholar] [CrossRef]
- Zhu, Z.; Ji, Q. Eye gaze tracking under natural head movements. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–26 June 2005; IEEE: New York, NY, USA, 2005; Volume 1, pp. 918–923. [Google Scholar]
- Sj, D.; Chauhan, S.S.; Shekhawat, B.S.; Kumar, L.; Ghosh, S. Real-time eye tracking using representation learning and regression. In Proceedings of the 7th Joint International Conference on Data Science & Management of Data (11th ACM IKDD CODS and 29th COMAD), Bangalore, India, 4–7 January 2024; pp. 298–306. [Google Scholar]
- Falch, L.; Lohan, K.S. Webcam-based gaze estimation for computer screen interaction. Front. Robot. AI 2024, 11, 1369566. [Google Scholar] [CrossRef]
- Liu, J.; Wu, J.; Yang, H.; Chi, J. An Improved Cross-Ratio Based Gaze Estimation Method Using Weighted Average and Polynomial Compensation. IEEE Access 2023, 11, 2410–2423. [Google Scholar] [CrossRef]
- Banks, A.; Eldin Abdelaal, A.; Salcudean, S. Head motion-corrected eye gaze tracking with the da Vinci surgical system. Int. J. Comput. Assist. Radiol. Surg. 2024, 19, 1459–1467. [Google Scholar] [CrossRef]
- Kartynnik, Y.; Ablavatski, A.; Grishchenko, I.; Grundmann, M. Real-time facial surface geometry from monocular video on mobile GPUs. arXiv 2019, arXiv:1907.06724. [Google Scholar] [CrossRef]
- Sánchez-Brizuela, G.; Cisnal, A.; de la Fuente-López, E.; Fraile, J.C.; Pérez-Turiel, J. Lightweight real-time hand segmentation leveraging MediaPipe landmark detection. Virtual Real. 2023, 27, 3125–3132. [Google Scholar]
- Latreche, A.; Kelaiaia, R.; Chemori, A.; Kerboua, A. Reliability and validity analysis of MediaPipe-based measurement system for some human rehabilitation motions. Measurement 2023, 214, 112826. [Google Scholar] [CrossRef]
- Berglund, E.; Jedel, I.; Berglund, A. Using MediaPipe machine learning to design casual exertion games to interrupt prolonged sedentary lifestyle. In HCI International 2023–Late Breaking Papers, Proceedings of the 25th International Conference on Human-Computer Interaction, HCII 2023, Copenhagen, Denmark, 23–28 July 2023; Springer: Cham, Switzerland, 2023; pp. 237–251. [Google Scholar]
- Ariz, M.; Villanueva, A.; Cabeza, R. Robust and accurate 2D-tracking-based 3D positioning method: Application to head pose estimation. Comput. Vis. Image Underst. 2019, 180, 13–22. [Google Scholar] [CrossRef]
- Hu, D.; Qin, H.; Liu, H.; Zhang, S. Gaze tracking algorithm based on projective mapping correction and gaze point compensation in natural light. In Proceedings of the 2019 IEEE 15th International Conference on Control and Automation (ICCA), Edinburgh, UK, 16–19 July 2019; IEEE: New York, NY, USA, 2019; pp. 1150–1155. [Google Scholar]
- Sasaki, M.; Nagamatsu, T.; Takemura, K. Screen corner detection using polarization camera for cross-ratio based gaze estimation. In Proceedings of the 11th ACM Symposium on Eye Tracking Research & Applications, Denver, CO, USA, 25–28 June 2019; pp. 1–9. [Google Scholar]
- Li, X.; Zhang, D.; Li, M.; Lee, D.J. Accurate Head Pose Estimation Using Image Rectification and a Lightweight Convolutional Neural Network. IEEE Trans. Multimed. 2023, 25, 2239–2251. [Google Scholar] [CrossRef]
- Hempel, T.; Abdelrahman, A.A.; Al-Hamadi, A. 6d Rotation Representation For Unconstrained Head Pose Estimation. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 2496–2500. [Google Scholar]
- Cheung, Y.m.; Peng, Q. Eye gaze tracking with a web camera in a desktop environment. IEEE Trans. Hum.-Mach. Syst. 2015, 45, 419–430. [Google Scholar] [CrossRef]
- Ansari, M.F.; Kasprowski, P.; Peer, P. Person-specific gaze estimation from low-quality webcam images. Sensors 2023, 23, 4138. [Google Scholar] [CrossRef] [PubMed]
- Koshikawa, K.; Nagamatsu, T.; Takemura, K. Model-based Gaze Estimation with Transparent Markers on Large Screens. Proc. ACM Hum.-Comput. Interact. 2022, 6, 1–16. [Google Scholar]
- Li, Z.; Tong, I.; Metcalf, L.; Hennessey, C.; Salcudean, S.E. Free head movement eye gaze contingent ultrasound interfaces for the da Vinci surgical system. IEEE Robot. Autom. Lett. 2018, 3, 2137–2143. [Google Scholar]
- Lu, F.; Sugano, Y.; Okabe, T.; Sato, Y. Gaze estimation from eye appearance: A head pose-free method via eye image synthesis. IEEE Trans. Image Process. 2015, 24, 3680–3693. [Google Scholar] [CrossRef] [PubMed]
- Zhou, X.; Cai, H.; Li, Y.; Liu, H. Two-eye model-based gaze estimation from a Kinect sensor. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May 29–3 June 2017; IEEE: New York, NY, USA, 2017; pp. 1646–1653. [Google Scholar][Green Version]
- Millington, I.; Funge, J. Artificial Intelligence for Games; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar] [CrossRef]
- Salvucci, D.D.; Goldberg, J.H. Identifying fixations and saccades in eye-tracking protocols. In Proceedings of the 2000 Symposium on Eye Tracking Research & Applications, Palm Beach Gardens, FL, USA, 6–8 November 2000; pp. 71–78. [Google Scholar]
- Birawo, B.; Kasprowski, P. Review and evaluation of eye movement event detection algorithms. Sensors 2022, 22, 8810. [Google Scholar] [CrossRef]
- Olsen, A. The Tobii I-VT fixation filter. Tobii Technol. 2012, 21, 4–19. [Google Scholar]
- Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. Appearance-based gaze estimation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4511–4520. [Google Scholar]
- Kellnhofer, P.; Recasens, A.; Stent, S.; Matusik, W.; Torralba, A. Gaze360: Physically Unconstrained Gaze Estimation in the Wild. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Repulic of Korea, 27 October–2 November 2019; pp. 6911–6920. [Google Scholar] [CrossRef]
- Zhang, X.; Park, S.; Beeler, T.; Bradley, D.; Tang, S.; Hilliges, O. Eth-xgaze: A large scale dataset for gaze estimation under extreme head pose and gaze variation. In Computer Vision–ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 365–381. [Google Scholar]
- Park, S.; Mello, S.D.; Molchanov, P.; Iqbal, U.; Hilliges, O.; Kautz, J. Few-shot adaptive gaze estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Repulic of Korea, 27 October–2 November 2019; pp. 9368–9377. [Google Scholar]
- Saxena, S.; Fink, L.K.; Lange, E.B. Deep learning models for webcam eye tracking in online experiments. Behav. Res. Methods 2024, 56, 3487–3503. [Google Scholar] [CrossRef]
- Onkhar, V.; Dodou, D.; De Winter, J. Evaluating the Tobii Pro Glasses 2 and 3 in static and dynamic conditions. Behav. Res. Methods 2024, 56, 4221–4238. [Google Scholar] [CrossRef] [PubMed]
- Housholder, A.; Reaban, J.; Peregrino, A.; Votta, G.; Mohd, T.K. Evaluating accuracy of the Tobii eye tracker 5. In Intelligent Human Computer Interaction, Proceedings of 13th International Conference, IHCI 2021, Kent, OH, USA, 20–22 December 2021; Springer: Cham, Switzerland, 2021; pp. 379–390. [Google Scholar]
- Ehinger, B.; Groß, K.; Ibs, I.; König, P. A new comprehensive eye-tracking test battery concurrently evaluating the Pupil Labs glasses and the EyeLink 1000. PeerJ 2019, 7, e7086. [Google Scholar] [CrossRef] [PubMed]











| Approaches | Head Fixation | Head Motion |
|---|---|---|
| Sasaki et al. [48] | 1.46 | - |
| Ansari et al. [52] | 1.98 | - |
| Koshikawa et al. [53] | 2.1 | - |
| Hu et al. [47] | - | 2.67 |
| Guo et al. [9] | - | 1.2 |
| Li et al. [54] | - | 2.0 |
| Lu et al. [55] | - | 2.5 |
| Cheung et al. [51] | 1.28 | 2.27 |
| Banks et al. [41] | - | 2.67 |
| Falch et al. [39] | 3.2 | - |
| Liu et al. [40] | - | 1.33 |
| Zhou et al. [56] | - | 1.99 |
| Proposed | 1.13 | 1.37 |
| Mode | Rectangular | Circluar |
|---|---|---|
| Head fixation | 1.45 (0.15) | 1.34 (0.16) |
| Head motion | 1.83 (0.14) | 1.74 (0.14) |
| Mode | Rectangular | Circluar | ||
|---|---|---|---|---|
| rmse_x | rmse_y | rmse_x | rmse_y | |
| Head fixation | 56.1 (9.0) | 57.8 (7.6) | 50.3 (9.6) | 56.6 (7.6) |
| Head motion | 77.4 (9.0) | 65.1 (11.3) | 68.8 (7.3) | 67.4 (11.6) |
| Game Mode | Number of the Bead Colors | Speed of the Bead |
|---|---|---|
| Mode 1 | 4 | 1.2 |
| Mode 2 | 6 | 1.5 |
| Mode 3 | 8 | 1.6 |
| Metrics | User Type | Mode 1 | Mode 2 | Mode 3 |
|---|---|---|---|---|
| Difficulty | inexperienced | 1.75 | 3.13 | 4.13 |
| experienced | 1.63 | 2.63 | 4.50 | |
| Preference | inexperienced | 3.25 | 3.5 | 3.0 |
| experienced | 3.25 | 3.38 | 3.63 | |
| Game score | inexperienced | 50.71 | 29.25 | 21.75 |
| experienced | 58.63 | 41.75 | 30.50 |
| Metric | User Type | Mode 1 | Mode 2 | Mode 3 |
|---|---|---|---|---|
| fixation count | inexperienced | 91.63 | 83.88 | 76.13 |
| experienced | 77.50 | 77.63 | 74.25 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, H.; Yin, L. Robust Camera-Based Eye-Tracking Method Allowing Head Movements and Its Application in User Experience Research. J. Eye Mov. Res. 2025, 18, 71. https://doi.org/10.3390/jemr18060071
Zhang H, Yin L. Robust Camera-Based Eye-Tracking Method Allowing Head Movements and Its Application in User Experience Research. Journal of Eye Movement Research. 2025; 18(6):71. https://doi.org/10.3390/jemr18060071
Chicago/Turabian StyleZhang, He, and Lu Yin. 2025. "Robust Camera-Based Eye-Tracking Method Allowing Head Movements and Its Application in User Experience Research" Journal of Eye Movement Research 18, no. 6: 71. https://doi.org/10.3390/jemr18060071
APA StyleZhang, H., & Yin, L. (2025). Robust Camera-Based Eye-Tracking Method Allowing Head Movements and Its Application in User Experience Research. Journal of Eye Movement Research, 18(6), 71. https://doi.org/10.3390/jemr18060071
