Next Article in Journal
Transformer–CNN Hybrid Framework for Pavement Pothole Segmentation
Previous Article in Journal
Drowsiness Classification in Young Drivers Based on Facial Near-Infrared Images Using a Convolutional Neural Network: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Electro-Oculography and Proprioceptive Calibration Enable Horizontal and Vertical Gaze Estimation, Even with Eyes Closed

by
Xin Wei
,
Felix Dollack
,
Kiyoshi Kiyokawa
and
Monica Perusquía-Hernández
*
Graduate School of Science and Technology, Nara Institute of Science and Technology (NAIST), Ikoma 630-0192, Japan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(21), 6754; https://doi.org/10.3390/s25216754 (registering DOI)
Submission received: 1 September 2025 / Revised: 20 October 2025 / Accepted: 27 October 2025 / Published: 4 November 2025
(This article belongs to the Section Biomedical Sensors)

Abstract

Eye movement is an important tool used to investigate cognition. It also serves as input in human–computer interfaces for assistive technology. It can be measured with camera-based eye tracking and electro-oculography (EOG). EOG does not rely on eye visibility and can be measured even when the eyes are closed. We investigated the feasibility of detecting the gaze direction using EOG while having the eyes closed. A total of 15 participants performed a proprioceptive calibration task with open and closed eyes, while their eye movement was recorded with a camera-based eye tracker and with EOG. The calibration was guided by the participants’ hand motions following a pattern of felt dots on cardboard. Our cross-correlation analysis revealed reliable temporal synchronization between gaze-related signals and the instructed trajectory across all conditions. Statistical comparison tests and equivalence tests demonstrated that EOG tracking was statistically equivalent to the camera-based eye tracker gaze direction during the eyes-open condition. The camera-based eye-tracking glasses do not support tracking with closed eyes. Therefore, we evaluated the EOG-based gaze estimates during the eyes-closed trials by comparing them to the instructed trajectory. The results showed that EOG signals, guided by proprioceptive cues, followed the instructed path and achieved a significantly greater accuracy than shuffled control data, which represented a chance-level performance. This demonstrates the advantage of EOG when camera-based eye tracking is infeasible, and it paves the way for the development of eye-movement input interfaces for blind people, research on eye movement direction when the eyes are closed, and the early detection of diseases.
Keywords: electro-oculography; signal processing; eyes closed; gaze direction estimation electro-oculography; signal processing; eyes closed; gaze direction estimation

Share and Cite

MDPI and ACS Style

Wei, X.; Dollack, F.; Kiyokawa, K.; Perusquía-Hernández, M. Electro-Oculography and Proprioceptive Calibration Enable Horizontal and Vertical Gaze Estimation, Even with Eyes Closed. Sensors 2025, 25, 6754. https://doi.org/10.3390/s25216754

AMA Style

Wei X, Dollack F, Kiyokawa K, Perusquía-Hernández M. Electro-Oculography and Proprioceptive Calibration Enable Horizontal and Vertical Gaze Estimation, Even with Eyes Closed. Sensors. 2025; 25(21):6754. https://doi.org/10.3390/s25216754

Chicago/Turabian Style

Wei, Xin, Felix Dollack, Kiyoshi Kiyokawa, and Monica Perusquía-Hernández. 2025. "Electro-Oculography and Proprioceptive Calibration Enable Horizontal and Vertical Gaze Estimation, Even with Eyes Closed" Sensors 25, no. 21: 6754. https://doi.org/10.3390/s25216754

APA Style

Wei, X., Dollack, F., Kiyokawa, K., & Perusquía-Hernández, M. (2025). Electro-Oculography and Proprioceptive Calibration Enable Horizontal and Vertical Gaze Estimation, Even with Eyes Closed. Sensors, 25(21), 6754. https://doi.org/10.3390/s25216754

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop