Next Article in Journal
Incentive Mechanism for Privacy-Preserving Collaborative Routing Using Secure Multi-Party Computation and Blockchain
Next Article in Special Issue
Mechanistic Assessment of Cardiovascular State Informed by Vibroacoustic Sensors
Previous Article in Journal
Reliability and Validity of Shore Hardness in Plantar Soft Tissue Biomechanics
Previous Article in Special Issue
A Robust Approach Assisted by Signal Quality Assessment for Fetal Heart Rate Estimation from Doppler Ultrasound Signal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fusion Algorithm Based on a Constant Velocity Model for Improving the Measurement of Saccade Parameters with Electrooculography

by
Palpolage Don Shehan Hiroshan Gunawardane
1,
Raymond Robert MacNeil
2,
Leo Zhao
1,
James Theodore Enns
1,*,
Clarence Wilfred de Silva
1 and
Mu Chiao
2
1
Department of Mechanical Engineering, The University of British Columbia, Vancouver, BC V6T 1Z4, Canada
2
Department of Psychology, The University of British Columbia, Vancouver, BC V6T 1Z4, Canada
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(2), 540; https://doi.org/10.3390/s24020540
Submission received: 17 November 2023 / Revised: 3 January 2024 / Accepted: 11 January 2024 / Published: 15 January 2024

Abstract

:
Electrooculography (EOG) serves as a widely employed technique for tracking saccadic eye movements in a diverse array of applications. These encompass the identification of various medical conditions and the development of interfaces facilitating human–computer interaction. Nonetheless, EOG signals are often met with skepticism due to the presence of multiple sources of noise interference. These sources include electroencephalography, electromyography linked to facial and extraocular muscle activity, electrical noise, signal artifacts, skin-electrode drifts, impedance fluctuations over time, and a host of associated challenges. Traditional methods of addressing these issues, such as bandpass filtering, have been frequently utilized to overcome these challenges but have the associated drawback of altering the inherent characteristics of EOG signals, encompassing their shape, magnitude, peak velocity, and duration, all of which are pivotal parameters in research studies. In prior work, several model-based adaptive denoising strategies have been introduced, incorporating mechanical and electrical model-based state estimators. However, these approaches are really complex and rely on brain and neural control models that have difficulty processing EOG signals in real time. In this present investigation, we introduce a real-time denoising method grounded in a constant velocity model, adopting a physics-based model-oriented approach. This approach is underpinned by the assumption that there exists a consistent rate of change in the cornea-retinal potential during saccadic movements. Empirical findings reveal that this approach remarkably preserves EOG saccade signals, resulting in a substantial enhancement of up to 29% in signal preservation during the denoising process when compared to alternative techniques, such as bandpass filters, constant acceleration models, and model-based fusion methods.

1. Introduction

Ocular motion sensing, also referred to as eye tracking, is utilized in both medical and engineering applications [1,2,3,4,5]. These systems leverage various types of eye movements, encompassing saccades, smooth pursuit, vergences, and vestibular–ocular reflexes [6,7]. Among the array of technologies available for sensing eye movement, electrooculography (EOG) and videooculography (VOG) stand as the predominant techniques [8]. EOG relies on the deployment of electrodes to measure the relative changes in the cornea-retinal potential that manifest during the simultaneous rotation of the eye and the cornea-retinal potential’s field vector [9]. On the other hand, VOG harnesses the infrared signals’ reflection on the corneal–retinal surface to gauge the eye’s angular displacement [8]. VOG systems are equipped with high-speed video cameras, surpassing EOG systems in accuracy. Notably, commercially available VOG systems, such as the EyeLink 1000 eye-tracker (EL), possess the capability to capture eye movements at a remarkable sampling rate of 2000 Hz , with an accuracy reaching an impressive 0.001 degrees of visual angle (DVA) [10]. However, it is essential to recognize that VOG systems do entail several drawbacks. They tend to be bulky, come with substantial costs, necessitate stringent or controlled lighting conditions, and fall short in their capacity to monitor closed-eye movements. These inherent limitations render EOG systems an indispensable choice for tracking eye movements across diverse scenarios where the use of VOG systems may be unfeasible. Notably, in contexts such as sleep disorder monitoring, EOG becomes indispensable, enabling the measurement of eye movements even when the eyes are closed [2,11].
Among the different types of eye movements, saccades hold paramount significance in medical and engineering applications. A saccade is defined as a swift and ballistic eye motion that redirects the point of fixation from one relatively stable position to another [5]. Both EOG and VOG systems possess the capability to capture this eye movement. Nevertheless, EOG systems suffer from a notable drawback—the presence of a substantial volume of noise and artifacts that detrimentally impact the fidelity of the recorded EOG eye movements. A typical EOG signal exhibits the intrusion of diverse artifacts, including but not limited to electroencephalography (EEG) artifacts, electromyography (EMG) artifacts, blink artifacts, and an assortment of noise sources [12,13]. These artifacts coexist within similar frequency spectra, complicating the denoising procedure [14,15]. Consequently, the extraction of saccades and the associated attributes, such as saccade amplitude, velocity, and latency [5,10], from an EOG recording becomes an onerous and time-intensive endeavor [16,17]. This onerous process exerts a detrimental influence on the accurate identification and classification of saccades [14].
Contemporary traditional saccade detection and classification methods, which rely on current threshold-based techniques, necessitate EOG signals characterized by minimal noise levels to effectively discern abrupt changes in amplitude and velocity profiles [18]. To attain this objective, conventional signal processing techniques such as bandpass filters [18,19], wavelet transforms [20,21,22], and smoothing filters [14,15], as well as specialized filters like morphological filters [23] and dynamic time-warping filters [24], have been employed to denoise EOG signals. Nonetheless, these traditional filtering methods often introduce distortions by inadvertently diminishing peak velocities and extending saccade duration during the denoising process [14,15,18]. Furthermore, they tend to compromise the preservation of EOG saccades and result in a significant deviation from the ground truth [10]. In response to these limitations, adaptive filtering approaches have been developed to effectively filter EOG signals while retaining the fidelity of EOG saccades [25]. Notably, Kalman filters (KFs) [26] have been introduced as a means to fuse the measured EOG signals with values estimated using a predefined mathematical model [26,27,28]. This approach has shown substantial improvements in maintaining the shape and integrity of EOG saccades [10]. Comparative studies have corroborated the superiority of KFs over traditional filtering methods in denoising EOG signals [25]. Nevertheless, it is worth noting that the accuracy of a KF is contingent on the precision of the underlying state estimator [10].
Diverse KF state estimators have been developed, drawing upon the mechanical [10,20,29], electrical [30,31], and parametric [15,32] attributes of the eye and its ocular movements. Notably, lumped-element-based dynamic models have served as foundational state estimators in this context [10,20,29]. Within these state estimators, the intricate agonist–antagonist dynamics of the extraocular muscles are characterized by their electromechanical properties to accurately estimate saccades, and these estimations are effectively integrated with EOG signals, leading to substantial enhancements in EOG saccade fidelity [10].
While most of the model-based techniques have demonstrated proficiency in addressing issues such as the elimination of eye blinks, offset correction, and signal denoising, they have typically necessitated real-time operation facilitated by a brain or neural controller [10,29]. Similarly, electrical models rooted in Coulomb’s law have been developed to rectify the baseline drift in EOG signals [25,30], but they encounter comparable challenges in preserving the integrity of EOG saccades during the denoising process. In VOG systems, KF fusion-based methodologies have been employed to denoise ocular motion signals derived from pupil reflections [33]. Nevertheless, these model-based denoising approaches [34,35], which utilize acceleration-based fusion algorithms [25], exhibit a notable disparity in accuracy when compared to ground truth measurements.
The present study explores a novel model-based technique aimed at enhancing the denoising of EOG signals while concurrently preserving the fidelity of EOG saccades. Specifically, we introduce a constant velocity-based model that considers the relationship between the peak velocity and saccade amplitude in the human eye, serving as a state estimator for the KF. We evaluate the effectiveness of this model in retaining EOG saccades throughout the denoising process and compare its performance to several traditional and adaptive denoising methods. This paper is structured as follows: Section 2 provides the mathematical foundation of the model-based fusion algorithms, experimental procedures employed for data acquisition and the algorithmic techniques employed for saccade identification and measurement. Section 3 and Section 4 present the outcomes of our analyses and the ensuing conclusions, respectively.

2. Materials and Methods

In this section, we introduce the constant velocity-based (CVM) KF method, where d E e d t is held at a constant value, and we present other established techniques for performance comparison. Well-established KF-related approaches, including bandpass filters (BP) [18,19], Brownian model-based (BM) KFs [17], constant acceleration model-based (CAM) KFs [25] (where d 2 E e d t 2 is kept constant), and linear reciprocal (LR) model-based filters [10], are employed as benchmarks to assess and evaluate the efficacy of the CVM.

2.1. Theoretical Models

2.1.1. Saccadic Eye Movements and EOG Signals

EOG is a method for measuring the corneal–retinal potential of the eye. This potential exhibits linearity with regard to angular displacements, spanning from 35 ° to 35 ° for horizontal movements and 10 ° to 10 ° for vertical movements [1,9]. The amplitude of EOG signals ranges from 50 to 3500 μ V for both horizontal and vertical gaze movements and falls within the frequency band of 1 to 35 Hz . Notably, the voltage sensitivity to changes during horizontal and vertical movements (saccades) is approximately 16 μ V and 14 μ V DVA, respectively [1,9]. In the design of adaptive filters, effective noise models are crucial, and the modeling and analysis of noise in biomedical systems have been elaborated upon in a previous work [36]. For the purposes of this study, an additive white noise model is employed to represent the noise inherent in EOG signals. This assumption has been effectively utilized by previous researchers in the development of their denoising methods [10,20,32].
A raw EOG signal comprises a combination of the corneo-retinal potential of the eye, noise, and artifacts. The induced voltage in the EOG electrode, denoted as E ( t ) , can be described as follows:
E ( t ) = E e ( t ) + φ + Ω
where E e ( t ) = corneal–retinal potential, φ = artifacts (eye muscles, eyelids, blinks, etc.), and Ω = noise (electromechanical noise). However, E e ( t ) is linearly proportional to the rotation of the eyeball. Therefore,
E e ( t ) = K θ a x i s ( t )
where θ a x i s ( t ) is the angular displacement around the eyeball axis parallel to the location of the electrode, and K is the calibration factor [10].

2.1.2. Noise Model

The noise inherent in EOG can be effectively characterized, as illustrated in Figure 1, with reference to prior studies [10,20,32,36]. In this depiction, U ( t ) serves as the input to the deterministic plant, providing the visual cues necessary to initiate saccades, resulting in the generation of an output denoted as E e ( t ) . A noise signal e ( t ) introduces stochasticity to the process and gives rise to the additive noise component Ω in the signal. Additionally, U * ( t ) is responsible for triggering artifacts, such as blinks, in the system. The ultimate measurement received by the electrode, E ( t ) , represents a composite of all these contributing factors and artifacts. The accurate identification of the deterministic system is of paramount importance to mitigate the adverse effects of the stochastic component. In this context, a variety of corneal–retinal potential models, serving as state estimators, are explored within the framework of this model (as discussed in Section 2.1.3). In the present study, the stochastic system is described using a Gaussian noise model, while the deterministic system is represented by the CVM.

2.1.3. Corneo-Retinal Potential Models for Saccades

The sensor fusion approach introduced in this paper is based on a KF, which employs linear quadratic estimation to filter a sequence of measurements collected over time. In this context, both process uncertainties and uncertainties in the measured values are treated as sources of noise. The process can be mathematically represented in a discretized form as follows:
E e _ k = A E e _ ( k 1 ) + B U k + Ω k
In this context, E e _ k corresponds to the discretized states of the system, which can be expressed as a 3 × 1 column vector. Here, s k , v k , and a k denote the biosignal, its rate of change, and the second rate of change, respectively, for the k t h sample. The process noise, denoted as Ω k , is assumed to follow an additive Gaussian white noise distribution with a covariance matrix Q, i.e., Ω k N ( 0 , Q ) . Additionally, the state transition matrix, denoted as A, characterizes the dynamics of the system, reflecting the state transition relation. The specific form of this matrix is contingent on the time-domain process model of the system. For this study, three saccadic eye models are considered: Brownian, constant velocity, and constant acceleration. Notably, the input gain matrix B is omitted, as this work operates as a “free” model and does not involve an input variable U k . The measurement, denoted as E e _ k (EOG), is subject to measurement noise, which is assumed to follow an additive white Gaussian noise distribution with a covariance matrix R, i.e., ϵ k N ( 0 , R ) . This assumption is made under the premise of no correlation between Ω k and ϵ k . The output model is represented as follows:
E k = H E e _ k + ϵ k
where H represents the measurement matrix. The CVM is employed to describe E e ( t ) , which is discretized as E e _ k . Subsequently, for the purpose of assessing the performance of the CVM in conjunction with state-of-the-art KFs, H is substituted with the BM, CAM, and LR, as elucidated in subsection D.

Constant Velocity Model

The analysis of the peak velocity in saccades has been a common practice in studies related to ocular motion [15,32]. Controlled experiments involving saccades have revealed that the peak velocity conforms to a nonlinear function characterized by its amplitude, as depicted in Equation (5). Furthermore, the product of the peak velocity and saccade duration, as described by Equation (6), exhibits a linear correlation with the amplitude [32].
V p = a 1 ( 1 e θ / a 2 )
V p × D = θ
Here, in Equations (5) and (6), V p denotes the peak velocity, a 1 and a 2 represent arbitrary constants, D signifies the duration of the saccade, and θ stands for the amplitude of the saccade. Drawing upon the findings presented in [17,32], Equation (6) can be construed as a representation of the saccade amplitude achieved by maintaining a constant velocity for a given duration. Based on this observation, we propose a linear constant velocity state estimator model for the Kalman filter (KF), under the assumption that the peak velocity remains approximately constant for higher saccade amplitudes. This assumption further implies that d E e d t , representing the peak rate of change of the differential potential, remains relatively constant, as we are linearly associating changes in potential with the saccade amplitude. The state-space representation of this corresponding process model, in discrete form, can be expressed as follows:
E e _ ( k + 1 ) = A E e _ k + K E ˙ e _ k
E ˙ e _ ( k + 1 ) = E ˙ e _ k
Therefore, the state-space form of the model is
E e _ ( k + 1 ) E ˙ e _ ( k + 1 ) = 1 k 0 1 E e _ ( k ) E ˙ e _ ( k ) + N o i s e

Kalman Filter Model

The state estimation uses the following state estimation scheme of prediction and correction.
Prediction:
E ^ e _ k = E ^ e _ ( k 1 ) + B U k + R P k = A P k 1 A T + Q
Correction:
K k = P k H T ( H P k H T + R ) 1 P k = ( I K k H ) P k E ^ e _ k = E ^ k + K k ( Z k H E ^ e _ k )
where E ^ e _ k = estimate, P k = covariance, and K k = Kalman gain.

2.1.4. Other State Estimator Models

Brownian Motion

In the context of the BM model, Brownian motion serves as the foundational concept from which the state estimator is derived, as referenced in [10,17]. The state-space representation of the corresponding process model, when expressed discretely, can be written as follows:
E e _ ( k + 1 ) = A E e _ k + N o i s e
where A = system matrix, which is an identity matrix, in the present work.

Constant Acceleration Model

The CAM considers ( d 2 E e d t 2 ), the 2nd rate of change of the differential potential, to be a constant [10,17,25]. The state-space representation of the corresponding process model, in the discrete form, can be written as
E e _ ( k + 1 ) = E e _ k + k E ˙ e _ k + 1 2 k 2 E ¨ e _ k
E ˙ e _ ( k + 1 ) = E ˙ e _ k + k E ¨ e _ k
E e _ ( k + 1 ) = E ¨ e _ k
Therefore, the state-space form of the model is
E e _ ( k + 1 ) E ¨ e _ ( k + 1 ) E e _ ( k + 1 ) = 1 k 1 2 k 2 0 1 k 0 0 1 E e _ ( k ) E ¨ e _ ( k ) E ¨ e _ ( k ) + N o i s e

Linear Reciprocal Model

The state estimator of the LR model-based KF is inspired by the agonist–antagonist EOM model of the human eye. The development of the LR model-based KF is detailed in [10]. The state-space representation of the corresponding process model, in the discrete form, can be written as
X 1 ˙ X 2 ˙ X 3 ˙ X 4 ˙ = 0 1 0 0 0 0 1 0 0 0 0 1 C 0 C 1 C 2 C 3 X 1 X 2 X 3 X 4 + 0 0 0 1 δ U + n o i s e
where X i = E, C i are constants derived from the parameters that govern eye movements as detailed in [10].

Bandpass Filtering

The employed BP filter within the scope of this study is as mentioned in [1,9,16]. The FIR process involves the utilization of a bandpass filter with a bandwidth ranging from 0.5 to 35 Hz , a drift removal step, the incorporation of a notch filter set at 60 Hz , and the application of a Savitzky–Golay filter with a fifth-order polynomial and a frame length of 111.

Sensor Fusion

EOG dry electrodes are employed for the assessment of the corneal–retinal potential of the eyeball. The experimentally obtained signals are integrated with the saccade models delineated in this section. An overview of the comprehensive fusion algorithm is provided in Figure 2. The initiation of saccades is prompted by visual cues ( U k ), and their measurement is conducted via the raw EOG signals ( E k ) within the time interval from t = 0 to t = t e n d . Subsequently, distinct state estimators inspired by the BM, CVM, CAM, and LR models are applied to compute E e k and enhance the signal quality using the KF.

2.2. Experimental Description

2.2.1. Participants

The primary objective of this human study is to assess the efficacy of the proposed denoising process in preserving EOG saccades by comparing them with saccades recorded by the EyeLink 1000, considered the gold standard in saccade measurement. This direct comparison of EOG and VOG saccades necessitates the simultaneous recording of saccades by both devices, as elaborated in Section 2.2.2. Procedure: In this experiment, any inherent variability in individual eye movements is mitigated, as both devices capture identical eye movements concurrently. This study involved thirteen healthy adults, aged 18 to 29 years, all of whom possessed normal or corrected visual acuity. The participants who regularly used eyeglasses or contact lenses to achieve 20/20 vision in their daily lives also wore these corrective lenses during the experiments. Ethical approval for conducting experiments on human subjects was granted by the University of British Columbia’s Behavioral Research Ethics Board (Approval No. H18-03792). Prior to participating in the study, written informed consent was obtained from all the participants. All the testing sessions took place at the UBC Vision Lab in Canada. The dataset reported in this study excluded data collected from participants with known eye disorders or individuals with significant optical corrections. However, two participants were excluded from the subsequent analysis and results: one participant did not meet the calibration requirements for pupil–corneal reflection (PCR) eye tracking, and another reported a diagnosis of Leber’s hereditary optic nephropathy, which is a congenital eye disorder associated with oculomotor abnormalities.

2.2.2. Apparatus

To minimize any potential interference caused by head movements during data acquisition, the participants were instructed to maintain the utmost stillness by utilizing a chin-and-forehead rest. Monocular eye movement recordings were performed on the right eye at a sampling rate of 250 Hz , employing both the OpenBCI Cyton board for acquiring the EOG signal and the EyeLink 1000, resulting in the simultaneous generation of two parallel data streams. For EOG data collection, pre-gelled Skintact electrodes were securely positioned at the participant’s outer canthus and forehead. The EyeLink system tracked the gaze position based on the PCR, which was captured using its built-in infrared camera. Owing to the superior reliability of the EyeLink system, the data recorded by it were later employed for the calibration and validation of the EOG measurements (as illustrated in the experimental setup depicted in Figure 3). The stimulus presentation was under the precise control of the custom software developed using the PsychoPy3 package in Python. Visual cues were displayed on a 121.9 c m × 71.1 c m LCD monitor with a resolution of 1920 × 1080 pixels and a refresh rate of 30 Hz . Audio cues were emitted through standard computer speakers. A keyboard, placed in front of the participants, allowed them to initiate the next trial in the queue by pressing a key. The event data, including keyboard inputs, EOG signals, and EyeLink measurements, were synchronized in real time with a shared computer clock utilizing the Lab Streaming Layer (LSL) software. Although all the data samples were recorded using the same clock reference, some random time lags were observed in certain samples, as exemplified in Figure 4. These lags were unsystematic and sporadic, occurring only in the data collected from specific participants. These discrepancies were not considered in the qualitative analysis, as our calculations were independent of the temporal data provided by the time stamps.

2.2.3. Procedure

The display featured a central marker in the form of a small circle, serving as the “home” fixation point. The magnitude of a saccade was quantified in relation to this home position, which was aligned with the participants’ midsagittal plane. Four target locations, labeled “A”, “B”, “C”, and “D”, were presented at angular deviations of 22 ° , 11 ° , 11°, and 22° from the visual angle, respectively. Notably, “A” and “B” targets were situated to the left of the home position, while “C” and “D” were positioned to the right. The investigation exclusively focused on horizontal saccades, and as such, all the target points were aligned along a horizontal axis traversing the center of the display. The initiation of each trial was contingent upon a keypress executed by the participant. Following trial initiation, an auditory cue was delivered, directing the participant to execute a saccade toward the designated target location. Subsequently, a second auditory cue, signified by a “ping” sound, prompted the participant to return their gaze to the home position, thus concluding the trial. The administration of the trials followed a pseudo-random order.

2.2.4. Calibration

Before each testing session, a calibration procedure was conducted, comprising five consecutive trials for each of the four target locations, namely, A, B, C, and D. Subsequently, the raw EOG signal, measured in microvolts, was transformed into the eyeball angle through the application of linear parametric regression. Calibration points, denoted as E p e a k a v e r a g e , were derived based on signal peak averages ( E e p e a k ) and peak counts ( N p c ), following the formula:
E ( p e a k _ a v e r a g e ) ( θ ) = 1 N p c n = 1 n = N p c [ min max t h r e s h ( E [ n ] ) ]

2.2.5. Kalman Filter

In KFs, a higher value of Q representing the model noise leads to an elevated gain and imparts greater significance to the measurement, albeit at the cost of introducing a time lag. Conversely, a lower Q value enhances the accuracy but may result in a delayed response. Therefore, the selection of an appropriate Q value necessitates a trade-off between the time response and accuracy, as discussed in references [10,26,37]. In this study, the Q values were determined using an iterative trial-and-error methodology. The values of R (representing sensor noise) were computed based on the mean of the standard deviations of the normalized errors observed in the calibration trial data.

2.2.6. Study Parameters

Throughout the data processing phase, the trials were delineated by recorded event markers, which corresponded to audio cues and keyboard inputs denoting the commencement and conclusion of each trial. The saccades were subsequently identified by detecting the two change points employing MATLAB’s “findchangepts” function, a tool designed to pinpoint abrupt shifts within a signal. The peaks situated between these two change points were extracted using MATLAB’s “findpeaks” function, and their average values were employed to define the saccade amplitude, corresponding to the angular displacement of the eyeball for both the EOG ( θ E O G ) and EyeLink ( θ E L ) data. An additional key feature calculated during the analysis was eye movement latency, denoted as t l a t E O G and t l a t E L , representing the time interval between the audio cue signaling the commencement of the trial and the first identified change point. Numerical differentiation was applied to the EOG and EL signals to derive the peak velocities, defined as V E O G and V E L , denoting the maximum slope of the saccade response.
Furthermore, two definitions of error were considered: the EOG error relative to EL ( E E O G E L = | θ E O G θ E L | ) and the measurement error (absolute error in EOG or EL, E E O G / E L = | θ E O G / E L θ | , where θ denotes the physical angle in the experimental setup). Additionally, for both EL and EOG, the accuracy and precision were assessed by examining the mean value and variance of the percentage of the normalized error (measurement error or absolute error, expressed as a percentage of the normalized error). To compare the noise power and signal power of the filtered signals, the signal-to-noise ratio (SNR) was employed. In this context, the SNR was defined as follows:
S N R ( d B ) = 10 log [ n = 1 n = k E k ¯ n = 1 n = k [ E k E k ] ]
Furthermore, the analysis involves the consideration of the computation time ( t c ) and nonlinear curve fitting of the peak velocity, utilizing Equation (5). A comprehensive overview of the parameters employed in this study is provided in Table 1.

3. Results and Discussion

The proper calibration of the KF is of utmost importance for optimizing the performance of each of these filters. The determination of the R values was carried out based on the results of the individual calibration trials, while the Q values remained consistent across various methods, with Q being determined using a trial-and-error approach as elaborated in Section 2.2.5. The values of both R and Q employed in each trial are documented in Table 2. Each experimental session consisted of two distinct types of trials, wherein the participants first engaged in a series of calibration trials followed by a sequence of experimental trials. The data recorded during the calibration trials were employed in conjunction with linear regression techniques to calibrate the subsequent experimental trials. The calibration factors generated through this process are presented in Table 3. It is worth noting that both the EOG and EL recordings underwent the same calibration process. However, it is important to emphasize that the recorded EL data, due to its inherent high quality, did not require denoising and were considered as the ground truth for this experiment. To process the recorded signals, a polynomial piece-wise detrend function in MATLAB was employed. This function was utilized to rectify any drift between the trial markers, specifically the “start” and “end” events, and to normalize the isolated EOG saccades to their baseline value, effectively removing the systematic baseline drift. Outliers within the data were identified and removed using the interquartile range method, which entails removing data points that fall outside a range of 1.5 times the interquartile range from the first and third quartiles, thereby ensuring a dataset devoid of bias. The application of different denoising methods yielded varying results in terms of signal denoising and the reduction in outliers in the EOG saccades. Table 2 provides a summary of the percentage of outliers removed after each denoising method was applied (namely, 13.48% for BP, 10.90% for BM, 12.05% for CVM, 12.02% for CAM, and 12.2% for LR). On average, all the model-based filters exhibited a 12.5% improvement in comparison to the bandpass filter method. Figure 5 illustrates an example of the recorded data extracted between the “start” and “end” event markers for participant P023 during the execution of a C trial (i.e., a saccade of 11 ). Subfigure (a) displays the raw recorded signal (comprising both the EOG and EL data), while subfigures (b) to (e) depict the outcomes following the application of various denoising methods. A thorough examination of these figures reveals that CVM has the capability to preserve EOG saccades more effectively when compared to other filters, as the overall signal shape closely resembles the EL records. In order to further assess and quantify the performance of the CVM, a comprehensive analysis was conducted, encompassing a correlation study, a numerical feature analysis of the EOG parameters, and an error analysis, utilizing data from all the participants.
Given the utilization of two distinct recording devices in this study, an examination was conducted to assess the feature correlation between different features and the applied denoising methods, thereby establishing a relationship between each filter and the features derived from the two devices. Specifically, the θ , E E O G / E L , V p , and t s of EOG and EL were computed for each denoising method and are comprehensively presented in Table 4. Notably, all these methods exhibited a robust correlation, with the KF-based filter techniques demonstrating an enhanced correlation of approximately 0.8% in amplitude. It is pertinent to note that θ , representing the average peak values of the saccade at fixation, is a key feature influencing the shape of the signal. Furthermore, the SNRs were computed and their results are displayed in Table 5 to affirm the efficacy of all the employed denoising methods in generating a potent output signal. As anticipated, the average t c (also showcased in Table 5) was higher for the model-based KF in comparison to the BP method, aligning with expectations.
To conduct a more comprehensive examination of the capacity to retain EOG saccades within the proposed denoising procedure, four key parameters were subjected to analysis. Figure 6 provides a visualization of the average values of these primary parameters, considering data from all the participants (encompassing both EOG and EL saccades) throughout the study, subsequent to the application of various denoising methods. In Figure 6a,b, the behavior of θ following the application of each denoising method is depicted. As summarized in Table 6, the utilization of the CVM has demonstrated a notable enhancement of the EOG signal by 28.7% in comparison to the BP method, resulting in a reduction in the E E O G . Furthermore, when compared to the EL signal, the overall EOG signal has experienced a 22.3% improvement ( E E O G E L ). Figure 6c illustrates the t s of the saccades, revealing that the application of these denoising methods does not have a significant impact on the EOG saccade t s . In Figure 6d, the relationship between the peak velocity ( V p ) and magnitude after the application of various denoising methods is elucidated. The filtered saccades were fitted to Equation (5) and compared with the data from reference [32]. The resulting fitted curves produced values of 62.13 for CVM, 110.63 for BP, 83.86 for CAM, 18.78 for BM, and 16.11 for LR, all fitted with p < 0.001 . Although BM and LR exhibited low RMSE values, their V p was diminished post-filtering. In contrast, the CVM, BP, and CAM yielded far more realistic V p values compared to the peak velocity versus amplitude data published in reference [32]. Among these methods, CVM displayed the lowest RMSE, indicating the closest fit to the V p formula. This analysis underscores that the CVM filter has not only effectively denoised the signal but has also preserved the EOG saccades during the denoising process. Representative examples of randomly selected trials are provided in Appendix A for further illustration.
To conduct a more in-depth performance analysis of the CVM in comparison to other denoising methods, we examined the kernel density of the probability distribution of the percentage of the normalized error, as depicted in Figure 7. This was computed by evaluating the (measured value − true value)/true value after the removal of outliers using 1.5 times the IQR criterion. The calculated mean values ( μ : EL = 0.18, CVM = 0.28, CAM = 0.3, BP = 0.38, and BM = 0.55) of this distribution serve as an indicator of the accuracy of each technique concerning EL. It is evident from the analysis that the CVM exhibits an improvement in accuracy of 2%, 10%, and 27% when compared to the CAM, BP, and Brownian techniques relative to EL. Furthermore, two-sample t-tests were conducted against EL, and all the filtering techniques demonstrated the rejection of the null hypothesis, thereby establishing the significance of the difference between the recorded percentages of the normalized error values (p values: CVM— 1.65 × 10 6 , CAM— 1.59 × 10 7 , BP— 3.94 × 10 16 , and BM— 2.15 × 10 54 . Consequently, the CVM showcases superior accuracy in comparison to other techniques concerning EL. The variance ( σ 2 : EL = 0.03, CVM = 0.06, CAM = 0.07, BP = 0.08, and Brownian = 0.08) of the percentages of the normalized error values also reveals that the CVM demonstrates a 1%, 2%, and 2% enhancement in precision relative to the CAM, BP, and Brownian techniques concerning EL.

4. Conclusions

In this paper, we introduce an adaptive filtering method grounded in the assumption of constant velocity, which entails considering the rate of change in the corneal–retinal potential as constant. We then conduct a comparative assessment of this CVM against established techniques, such as BP, BM, CAM, and LR-based methods. To rigorously evaluate these methods, we employ controlled experiments wherein saccades are concurrently measured using an OpenBCI Cyton Biosensing board (for EOG signal acquisition) and an EyeLink 1000 eye-tracker. We extract essential parameters including θ , E E O G / E L , V p , and t s from the signals and employ them to assess the performance of each denoising method, with particular emphasis on characterizing the efficacy of the recently introduced CVM approach. The CVM exhibited the most superior performance, manifesting in the filtered data by achieving the lowest errors ( E a b s E O G = 9.52 ° and E E O G E L = 10.63 ° ). The outcomes reveal that, in comparison to the physical size of the saccade, the CVM enhanced the EOG signal by approximately 29%, and in relation to the EyeLink (EL) recordings, it improved it by over 22% when contrasted with the BP filter. Furthermore, compared to the physical size of the saccade, the CVM outperformed the BP filter by approximately 2% and surpassed it by over 21% when evaluated against the EL recordings. The kernel density distribution of the normalized error percentages indicates that the introduction of sound mathematical models can augment the filtering capabilities of Kalman filters (KFs). As demonstrated by the results, both the CVM and CAM substantially elevated the accuracy and precision of the eye movement recordings. Specifically, the CVM exhibited an average improvement of 13% in accuracy and 3% in precision when compared to other denoising methods.
In Figure 7, we present the computed means derived from the entire dataset, characterizing the four key saccade features: amplitude (Figure 6a), error (Figure 6b), peak velocity (Figure 6c), and latency (Figure 6d). A comparative analysis is performed between the means obtained from the EyeLink and the EOG signals, utilizing various filtering methods, including the BP, Brownian, CAM, and CVM. The accompanying error bars depict the standard deviation. It is important to note that the EyeLink signal is regarded as the primary reference point for comparison, given its recognized accuracy in measurements. Upon examining Figure 6a,b, a discernible pattern emerges, wherein saccades with smaller amplitudes exhibit superior accuracy in the EOG signal when compared to saccades with relatively larger amplitudes. Specifically, targets A and D correspond to the most peripheral goal locations, entailing saccade amplitudes of ± 22 concerning the home position, where the induced charge approximates zero. This peripheral positioning may introduce nonlinearity into the signal, potentially diminishing the accuracy of the recording. A similar trend is observed in Figure 6b, where the saccades directed toward targets B and C display reduced errors in comparison to those aimed at targets A and D. It is noteworthy that the application of the filtering methods has a negligible impact on the latency of the signal. This outcome aligns with expectations, as these filters are not anticipated to influence the temporal resolution of the signal.
The CVM-based KF employed fixed values for Q and R , as outlined in Section 2.2.5. These values were established at the outset of each session and remained unaltered throughout the signal processing phase. It is plausible that this static configuration may have constrained the algorithm’s overall performance. Thus, it is posited that the implementation of a methodology to dynamically compute real-time adjustments for the Q and R values could enhance the efficacy of the signal filtering process. As indicated in Figure 6d, saccades demonstrate a constant velocity behavior for only high degrees and speeds. Consequently, there may be a necessity to modify this method for effective application in very short range saccades and low-speed scenarios, such as those encountered in smooth pursuit eye movements. Furthermore, it is important to note that the method does not encapsulate any mechanical or electrical characteristics of the eyeball or eye muscles, unlike models such as LR which incorporate electromechanical properties. Therefore, in instances of extreme conditions, such as individuals with disorders affecting the eyeball or eye muscles, the method, in its current state, may encounter challenges in accurately enhancing signals. Furthermore, the amalgamation of sensor data from both EyeLink (EL) and EOG sources has the potential to yield highly accurate EOG saccades. However, such an approach may impose limitations on the practical applicability of this technology, particularly in scenarios requiring the detection of closed-eye movements. One notable limitation in the present study pertains to the collection of EOG data under conditions of consistent illumination. Given the sensitivity of the corneal–retinal potential to variations in lighting conditions, this factor assumes significance in real-world applications where illumination levels are subject to change. Future research endeavors could extend upon this foundation by exploring the filtering capabilities of this algorithm when applied to EOG records acquired under varying lighting conditions. Additionally, there exists the opportunity to investigate novel machine learning-based techniques for EOG denoising that offer comparable capabilities. The current investigation is specifically concentrated on saccadic eye movements, given their prevalence and frequent occurrence in daily human activities. Subsequent research endeavors are anticipated to extend the scope of these methods to encompass less common eye movement types, such as smooth pursuit movements and vestibular–ocular reflexes.

Author Contributions

Conceptualization, P.D.S.H.G. and R.R.M.; methodology, P.D.S.H.G. and R.R.M.; software, P.D.S.H.G., R.R.M. and L.Z.; validation, P.D.S.H.G. and R.R.M.; formal analysis, P.D.S.H.G. and R.R.M.; investigation, P.D.S.H.G. and R.R.M.; writing—original draft preparation, P.D.S.H.G., R.R.M., J.T.E. and M.C.; writing—review and editing, J.T.E., C.W.d.S. and M.C.; supervision, J.T.E. and M.C.; project administration, J.T.E. and M.C.; funding acquisition, J.T.E., M.C. and C.W.d.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded by the grants from the Natural Sciences and Engineering Research Council (NSERC) of Canada, specifically the following: (1). A Strategic Partnership Grants project (STPGP 493908) entitled “Research in Sensory Information Technologies and Implementation in Sleep Disorder Monitoring”, which was awarded to CWdS/MC, (2). A Discovery Grant (RGPIN-2018-03741) bestowed upon JTE, and (3). A Doctoral Scholarship (PGSD3-547166-2020) granted to R.R.M.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration 502 of Helsinki and approved by the University of British Columbia’s Behavioral Research Ethics Board (Approval No. H18-03792).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The code and data can be downloaded from the following source: https://github.com/HiroshanOU/EOG_CVM (accessed on 30 December 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. Randomly selected trials. (a,c,e,g) are EL vs. raw EOG and (b,d,f,h) are raw EOG vs. CVM.
Figure A1. Randomly selected trials. (a,c,e,g) are EL vs. raw EOG and (b,d,f,h) are raw EOG vs. CVM.
Sensors 24 00540 g0a1aSensors 24 00540 g0a1b

References

  1. Heide, W.; Koenig, E.; Trillenberg, P.; Kömpf, D.; Zee, D. Electrooculography: Technical standards and applications. Electroencephalogr. Clin. Neurophysiol. Suppl. 1999, 52, 223–240. [Google Scholar] [PubMed]
  2. Dong, H.; Supratak, A.; Pan, W.; Wu, C.; Matthews, P.M.; Guo, Y. Mixed neural network approach for temporal sleep stage classification. IEEE Trans. Neural Syst. Rehab. Eng. 2017, 26, 324–333. [Google Scholar] [CrossRef]
  3. Maremmani, C.; Monastero, R.; Orlandi, G.; Salvadori, S.; Pieroni, A.; Baschi, R.; Pecori, A.; Dolciotti, C.; Berchina, G.; Rovini, E.; et al. Objective assessment of blinking and facial expressions in Parkinson’s disease using a vertical electro-oculogram and facial surface electromyography. Physiol. Meas. 2019, 40, 065005. [Google Scholar] [CrossRef] [PubMed]
  4. Gallicchio, G.; Cooke, A.; Ring, C. Assessing ocular activity during performance of motor skills using electrooculography. Psychophysiology 2018, 55, e13070. [Google Scholar] [PubMed]
  5. Mengoudi, K.; Ravi, D.; Yong, K.X.; Primativo, S.; Pavisic, I.M.; Brotherhood, E.; Lu, K.; Schott, J.M.; Crutch, S.J.; Alexander, D.C. Augmenting dementia cognitive assessment with instruction-less eye-tracking tests. IEEE J. Biomed. Health Inform. 2020, 24, 3066–3075. [Google Scholar] [CrossRef] [PubMed]
  6. Iskander, J.; Hossny, M.; Nahavandi, S. A review on ocular biomechanic models for assessing visual fatigue in virtual reality. IEEE Access 2018, 6, 19345–19361. [Google Scholar] [CrossRef]
  7. Ranjbaran, M.; Smith, H.L.; Galiana, H.L. Automatic classification of the vestibulo-ocular reflex nystagmus: Integration of data clustering and system identification. IEEE Trans. Biomed. Eng. 2015, 63, 850–858. [Google Scholar] [CrossRef]
  8. Duchowski, A.T. Eye tracking methodology. Theory Pract. 2007, 328, 2–3. [Google Scholar]
  9. Constable, P.A.; Bach, M.; Frishman, L.J.; Jeffrey, B.G.; Robson, A.G.; International Society for Clinical Electrophysiology of Vision. ISCEV Standard for clinical electro-oculography (2017 update). Doc. Ophthalmol. 2017, 134, 1–9. [Google Scholar] [CrossRef] [PubMed]
  10. Gunawardane, P.D.S.H.; Macneil, R.; Zhao, L.; Enns, J.; De Silva, C.; Chiao, M. A Fusion Algorithm for Saccade Eye Movement Enhancement with EOG and Lumped-element Models. IEEE Trans. Biomed. Eng. 2021, 68, 3048–3058. [Google Scholar] [CrossRef] [PubMed]
  11. Sharma, M.; Yadav, A.; Tiwari, J.; Karabatak, M.; Yildirim, O.; Acharya, U.R. An automated wavelet-based sleep scoring model using eeg, emg, and eog signals with more than 8000 subjects. Int. J. Environ. Res. Public Health 2022, 19, 7176. [Google Scholar] [CrossRef] [PubMed]
  12. Martínez-Cerveró, J.; Ardali, M.K.; Jaramillo-Gonzalez, A.; Wu, S.; Tonin, A.; Birbaumer, N.; Chaudhary, U. Open Software/Hardware Platform for Human-Computer Interface Based on Electrooculography (EOG) Signal Classification. Sensors 2020, 20, 2443. [Google Scholar] [CrossRef] [PubMed]
  13. Pleshkov, M.; Zaitsev, V.; Starkov, D.; Demkin, V.; Kingma, H.; van de Berg, R. Comparison of EOG and VOG obtained eye movements during horizontal head impulse testing. Front. Neurol. 2022, 13, 917413. [Google Scholar] [PubMed]
  14. Diaz, G.; Cooper, J.; Kit, D.; Hayhoe, M. Real-time recording and classification of eye movements in an immersive virtual environment. J. Vis. 2013, 13, 5. [Google Scholar]
  15. Nyström, M.; Holmqvist, K. An adaptive algorithm for fixation, saccade, and glissade detection in eyetracking data. Behav. Res. Methods 2010, 42, 188–204. [Google Scholar] [PubMed]
  16. Gunawardane, P.; de Silva, C.; Chiao, M. An Oculomotor Sensing Technique for Saccade Isolation of Eye Movements using OpenBCI. In Proceedings of the 2019 IEEE SENSORS, Montreal, QC, Canada, 27–30 October 2019; pp. 1–4. [Google Scholar]
  17. Gunawardane, P.D.S.H. A Model-Based Fusion Technique for Ocular Motion Sensing. Master’s Thesis, The University of British Columbia, Vancouver, BC, Canada, 2019. [Google Scholar]
  18. Stuart, S.; Hickey, A.; Vitorio, R.; Welman, K.; Foo, S.; Keen, D.; Godfrey, A. Eye-tracker algorithms to detect saccades during static and dynamic tasks: A structured review. Physiol. Meas. 2019, 40, 02TR01. [Google Scholar]
  19. Jia, Y.; Tyler, C.W. Measurement of saccadic eye movements by electrooculography for simultaneous EEG recording. Behav. Res. Methods 2019, 51, 2139–2151. [Google Scholar]
  20. Wadehn, F.; Weber, T.; Mack, D.J.; Heldt, T.; Loeliger, H.A. Model-based separation, detection, and classification of eye movements. IEEE Trans. Biomed. Eng. 2019, 67, 588–600. [Google Scholar] [CrossRef]
  21. Dasgupta, A.; Routray, A. A New Multi-resolution Analysis Method for Electrooculography Signals. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 1. [Google Scholar] [CrossRef]
  22. Yan, B.; Pei, T.; Wang, X. Wavelet method for automatic detection of eye-movement behaviors. IEEE Sens. J. 2018, 19, 3085–3091. [Google Scholar]
  23. Jiang, P.; Zhou, R. De-noising and recognition of EOG signal based on mathematical morphology. In Proceedings of the 2013 Sixth International Symposium on Computational Intelligence and Design, Hangzhou, China, 28–29 October 2013; Volume 2, pp. 351–354. [Google Scholar]
  24. Dasgupta, A.; Chakraborty, S.; Mondal, P.; Routray, A. Identification of eye saccadic signatures in electrooculography data using time-series motifs. In Proceedings of the 2016 IEEE Annual India Conference (INDICON), Bangalore, India, 16–18 December 2016; pp. 1–5. [Google Scholar]
  25. Dasgupta, A.; Chakraborty, S.; Routray, A. A two-stage framework for denoising electrooculography signals. Biomed. Signal Process. Control 2017, 31, 231–237. [Google Scholar]
  26. Bulut, Y.; Vines-Cavanaugh, D.; Bernal, D. Process and estimation for Kalman filtering. In Structural Dynamics; Springer: Berlin/Heidelberg, Germany, 2011; Volume 3, pp. 375–386. [Google Scholar]
  27. Acharya, S.; Mongan, W.M.; Rasheed, I.; Liu, Y.; Anday, E.; Dion, G.; Fontecchio, A.; Kurzweg, T.; Dandekar, K.R. Ensemble learning approach via Kalman filtering for a passive wearable respiratory monitor. IEEE J. Biomed. Health Inform. 2018, 23, 1022–1031. [Google Scholar] [CrossRef] [PubMed]
  28. Ranjbaran, M.; Jalaleddini, K.; Lopez, D.G.; Kearney, R.E.; Galiana, H.L. Analysis and modeling of noise in biomedical systems. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 997–1000. [Google Scholar]
  29. Komogortsev, O.V.; Khan, J.I. Eye movement prediction by oculomotor plant Kalman filter with brainstem control. J. Control Theory Appl. 2009, 7, 14–22. [Google Scholar] [CrossRef]
  30. Barbara, N.; Camilleri, T.A.; Camilleri, K.P. EOG-Based Ocular and Gaze Angle Estimation Using an Extended Kalman Filter. In Proceedings of the ACM Symposium on Eye Tracking Research and Applications, Stuttgart, Germany, 2–5 June 2020; pp. 1–5. [Google Scholar]
  31. Barbara, N.; Camilleri, T.; Camilleri, K.P. A comparison of EOG baseline drift mitigation techniques. Biomed. Signal Process. Control 2020, 57, 101738. [Google Scholar] [CrossRef]
  32. Dai, W.; Selesnick, I.; Rizzo, J.R.; Rucker, J.; Hudson, T. Detection of normal and slow saccades using implicit piecewise polynomial approximation. J. Vis. 2021, 21, 8. [Google Scholar]
  33. Toivanen, M. An advanced Kalman filter for gaze tracking signal. Biomed. Signal Process. Control 2016, 25, 150–158. [Google Scholar] [CrossRef]
  34. Barbara, N.; Camilleri, T.A.; Camilleri, K.P. Real-time continuous EOG-based gaze angle estimation with baseline drift compensation under stationary head conditions. Biomed. Signal Process. Control 2023, 86, 105282. [Google Scholar] [CrossRef]
  35. Barbara, N.; Camilleri, T.A.; Camilleri, K.P. TEMoD: Target-Enabled Model-Based De-Drifting of the EOG Signal Baseline using a Battery Model of the Eye. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 1–5 November 2021; pp. 562–565. [Google Scholar]
  36. Verhaegen, M.; Verdult, V. Filtering and System Identification: A Least Squares Approach; Cambridge University Press: Cambridge, UK, 2007. [Google Scholar]
  37. Park, S.; Gil, M.S.; Im, H.; Moon, Y.S. Measurement noise recommendation for efficient Kalman filtering over a large amount of sensor data. Sensors 2019, 19, 1168. [Google Scholar] [PubMed]
Figure 1. Noise model for a raw EOG signal. U ( t ) : input to the deterministic systems (visual cue), U * ( t ) : trigger signal for the artifacts, E e ( t ) : output of the deterministic signal (corneo-retinal potential), e ( t ) : white noise signal to the stochastic system, Ω : additive noise signal, E ( t ) : raw EOG signal, and φ : artifacts [10].
Figure 1. Noise model for a raw EOG signal. U ( t ) : input to the deterministic systems (visual cue), U * ( t ) : trigger signal for the artifacts, E e ( t ) : output of the deterministic signal (corneo-retinal potential), e ( t ) : white noise signal to the stochastic system, Ω : additive noise signal, E ( t ) : raw EOG signal, and φ : artifacts [10].
Sensors 24 00540 g001
Figure 2. Schematic diagram of the model-based fusion algorithm. The saccades are initiated by visual cues leading to estimation and measurement. The velocity model is used to estimate the states of the eye and the Kalman filter is used to fuse the estimation with the measurement to generate the final output. To compare the velocity model with other model-based approaches, the state estimator was modified accordingly (e.g., with a constant acceleration model).
Figure 2. Schematic diagram of the model-based fusion algorithm. The saccades are initiated by visual cues leading to estimation and measurement. The velocity model is used to estimate the states of the eye and the Kalman filter is used to fuse the estimation with the measurement to generate the final output. To compare the velocity model with other model-based approaches, the state estimator was modified accordingly (e.g., with a constant acceleration model).
Sensors 24 00540 g002
Figure 3. Experimentation arrangement of EyeLink 1000 tracker, OpenBCI device, and participant.
Figure 3. Experimentation arrangement of EyeLink 1000 tracker, OpenBCI device, and participant.
Sensors 24 00540 g003
Figure 4. Recording of EyeLink 1000 tracker signals and OpenBCI EOG signals simultaneously using Lab Streaming Layer. The corneo-retinal potential is recorded by the electrodes placed on the outer canthus with respect to the electrode placed on the forehead. Horizontal saccades are directed to targets presented at −12, −11, 11, and 22 degrees of visual angle on the LCD screen.
Figure 4. Recording of EyeLink 1000 tracker signals and OpenBCI EOG signals simultaneously using Lab Streaming Layer. The corneo-retinal potential is recorded by the electrodes placed on the outer canthus with respect to the electrode placed on the forehead. Horizontal saccades are directed to targets presented at −12, −11, 11, and 22 degrees of visual angle on the LCD screen.
Sensors 24 00540 g004
Figure 5. The responses of different filters, when applying on raw EOG signals. The comparison against (a), raw EOG signal and EL signal demonstrate the contribution of each filter. Note: in (a), left Y axis and blue line represent the raw EOG signal and right green axis represents the EL signal. From (b) to (f), the left Y axis and blue line are the same and the right green axis represents the filtered signal.
Figure 5. The responses of different filters, when applying on raw EOG signals. The comparison against (a), raw EOG signal and EL signal demonstrate the contribution of each filter. Note: in (a), left Y axis and blue line represent the raw EOG signal and right green axis represents the EL signal. From (b) to (f), the left Y axis and blue line are the same and the right green axis represents the filtered signal.
Sensors 24 00540 g005
Figure 6. Extracted saccade characteristics from EyeLink and filtered EOG signals, grouped by target locations A, B, C, and D, averaged for all participants; Dai et al. [32].
Figure 6. Extracted saccade characteristics from EyeLink and filtered EOG signals, grouped by target locations A, B, C, and D, averaged for all participants; Dai et al. [32].
Sensors 24 00540 g006
Figure 7. Probabilistic density distribution (kernel density) of % normalized error. Mean values marked as dash lines.
Figure 7. Probabilistic density distribution (kernel density) of % normalized error. Mean values marked as dash lines.
Sensors 24 00540 g007
Table 1. Definitions of study parameters.
Table 1. Definitions of study parameters.
Study ParameterSymbolDefinition
Amplitude θ Angle of the landing position of eye from the center
Peak velocity V p Average of the maximum peaks at the fixation
Latency t s Time taken to initiate the saccade after receiving cue
Abs. relative error E E O G E L Absolute of difference between EOG and EL mag.
Abs. EyeLink error E E L Absolute difference between EL and physical mag.
Abs. relative error EOG E E O G Absolute of difference between EOG and physical mag.
Normalized error E N Normalized absolute relative error
Computation time t c Time taken to process the data between markers
Signal-to-noise ratio S N R Ratio of the signal power to the noise power
Table 2. Percentage of outliers (Out.%) excluded, Q, and R values for each participant.
Table 2. Percentage of outliers (Out.%) excluded, Q, and R values for each participant.
BPBMCVM CAM LR *
PIDOut. %Out. %QOut. %QOut. %QRQROut. %
P2014.0810.060.018.660.511.570.510.530.51113
P229.957.520.018.760.57.190.515.560.51611.1
P2315.2515.160.0115.700.516.190.58.060.5812.8
P246.318.290.014.710.56.580.515.550.5167.3
P2514.1317.880.0117.490.513.590.516.640.51718
P2621.249.260.018.500.511.090.55.550.5610
P2712.5211.540.019.840.59.150.5150.400.515013.9
P2816.0314.150.0114.490.514.510.510.140.51014.8
P2914.0811.370.0115.720.514.420.521.710.52212.1
P3217.1019.240.0119.200.517.430.522.850.51315
P337.636.880.019.420.510.490.5173.310.51736.4
Note: Data points were excluded as outliers if they fell outside the interquartile range at the participant level. BP = bandpass, BM = Brownian motion model, EL = EyeLink 1000, CAM = constant acceleration model, CVM = constant velocity model, PID = Participant ID. * The results for LR model were extracted from [10].
Table 3. Calibration factors for EOG based on each filtering technique ( m V deg 1 ) and EL (px deg 1 ).
Table 3. Calibration factors for EOG based on each filtering technique ( m V deg 1 ) and EL (px deg 1 ).
PIDBPBMCAMLR *ELCVM
P200.250.280.290.030.070.25
P220.210.230.200.010.080.18
P230.480.400.520.060.070.47
P240.290.470.410.020.070.31
P250.120.080.090.030.080.09
P260.150.140.490.030.070.36
P270.500.990.570.040.070.42
P280.230.240.300.040.070.27
P290.271.110.420.020.080.37
P320.200.190.260.030.070.24
P330.170.370.250.020.080.25
Note: PID = Participant ID. * The results for LR model were extracted from [10].
Table 4. Feature correlations between the filtered EOG signal and EL.
Table 4. Feature correlations between the filtered EOG signal and EL.
FeatureBPBMCAMLR *CVM
Amplitude0.9910.9990.9990.9970.999
Error0.6540.1390.5400.7320.671
Peak Velocity0.9760.9350.1570.4850.967
Latency0.9990.6490.5760.3360.583
Improvement (%)887.58
Note: Correlations were computed using Pearson’s method. * The results for LR model were extracted from [10].
Table 5. Signal-to-noise ratio and computation time.
Table 5. Signal-to-noise ratio and computation time.
MeasurementBPBMCAMLRCVM
SNR10.7212.211.469.0111.48
t c 0.480.220.940.720.98
% improvement in t c with respect to BP−13.4795.3649.19104.43
Note: BM has a lower t c and other methods have a higher t c than BP.
Table 6. Mean errors (DVA) and improvement (%) with respect to bandpass for model-based KFs.
Table 6. Mean errors (DVA) and improvement (%) with respect to bandpass for model-based KFs.
Parameters BPBMCAMCVMLR *
E _ a b s _ E L Mean2.852.852.852.852.85
SD0.420.420.420.420.42
E _ a b s _ E O G Mean6.805.494.934.856.75
SD1.040.950.850.841.23
Improv. %-19.3027.5028.701
E _ E O G _ E L Mean7.576.766.185.887.83
SD Improv. %1.131.160.970.931.28
Improv. %-10.7018.4022.303.43
Note: * The results for LR model were extracted from [10] and E _ a b s _ E L mean and SD remain unchanged, as the filters were exclusively applied to EOG signals, with EL considered as the gold standard.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gunawardane, P.D.S.H.; MacNeil, R.R.; Zhao, L.; Enns, J.T.; de Silva, C.W.; Chiao, M. A Fusion Algorithm Based on a Constant Velocity Model for Improving the Measurement of Saccade Parameters with Electrooculography. Sensors 2024, 24, 540. https://doi.org/10.3390/s24020540

AMA Style

Gunawardane PDSH, MacNeil RR, Zhao L, Enns JT, de Silva CW, Chiao M. A Fusion Algorithm Based on a Constant Velocity Model for Improving the Measurement of Saccade Parameters with Electrooculography. Sensors. 2024; 24(2):540. https://doi.org/10.3390/s24020540

Chicago/Turabian Style

Gunawardane, Palpolage Don Shehan Hiroshan, Raymond Robert MacNeil, Leo Zhao, James Theodore Enns, Clarence Wilfred de Silva, and Mu Chiao. 2024. "A Fusion Algorithm Based on a Constant Velocity Model for Improving the Measurement of Saccade Parameters with Electrooculography" Sensors 24, no. 2: 540. https://doi.org/10.3390/s24020540

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop