You are currently viewing a new version of our website. To view the old version click .
Sensors
  • This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
  • Article
  • Open Access

17 November 2025

Low-Cost Eye-Tracking Fixation Analysis for Driver Monitoring Systems Using Kalman Filtering and OPTICS Clustering

,
and
1
Faculty of Mechanical Engineering, Institute for Vehicle Systems Engineering, Karlsruhe Institute of Technology (KIT), 76131 Karlsruhe, Germany
2
Department of Complete Vehicle Data Driven Testing, Analysis and Vehicle Functions, Porsche Engineering Services GmbH, 74321 Bietigheim-Bissingen, Germany
*
Author to whom correspondence should be addressed.
Sensors2025, 25(22), 7028;https://doi.org/10.3390/s25227028 
(registering DOI)
This article belongs to the Section Sensing and Imaging

Abstract

Driver monitoring systems benefit from fixation-related eye-tracking features, yet dedicated eye-tracking hardware is costly and difficult to integrate at scale. This study presents a practical software pipeline that extracts fixation-related features from conventional RGB video. Facial and pupil landmarks obtained with MediaPipe are denoised using a Kalman filter, fixation centers are identified with the OPTICS algorithm within a sliding window, and an affine normalization compensates for head motion and camera geometry. Fixation segments are derived from smoothed velocity profiles based on a moving average. Experiments with laptop camera recordings show that the combined Kalman and OPTICS pipeline reduces landmark jitter and yields more stable fixation centroids, while the affine normalization further improves referential pupil stability. The pipeline operates with minimal computational overhead and can be implemented as a software update in existing driver monitoring or advanced driver assistance systems. This work is a proof of concept that demonstrates feasibility in a low-cost RGB setting with a limited evaluation scope. Remaining challenges include sensitivity to lighting conditions and head motion that future work may address through near-infrared sensing, adaptive calibration, and broader validation across subjects, environments, and cameras. The extracted features are relevant for future studies on cognitive load and attention, although cognitive state inference is not validated here.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.