Next Article in Journal
A New Contact Structure and Dielectric Recovery Characteristics of the Fast DC Current-Limiting Circuit Breaker
Previous Article in Journal
Intelligent Energy Efficiency Maximization for Wirelessly-Powered UAV-Assisted Secure Sensor Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination

1
Department of Mechanical and Mechatronics Engineering, University of Waterloo, 200 University Ave W, Waterloo, ON N2L 3G1, Canada
2
Mechanical Engineering Department, University of Alberta, Edmonton, AB T6G 2R3, Canada
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(5), 1537; https://doi.org/10.3390/s25051537
Submission received: 23 January 2025 / Revised: 24 February 2025 / Accepted: 25 February 2025 / Published: 1 March 2025
(This article belongs to the Section Physical Sensors)

Abstract

:
Inertial navigation systems augmented with visual and wheel odometry measurements have emerged as a robust solution to address uncertainties in robot localization and odometry. This paper introduces a novel data-driven approach to compensate for wheel slippage in visual-inertial-wheel odometry (VIWO). The proposed method leverages Gaussian process regression (GPR) with deep kernel design and long short-term memory (LSTM) layers to model and mitigate slippage-induced errors effectively. Furthermore, a feature confidence estimator is incorporated to address the impact of dynamic feature points on visual measurements, ensuring reliable data integration. By refining these measurements, the system utilizes a multi-state constraint Kalman filter (MSCKF) to achieve accurate state estimation and enhanced navigation performance. The effectiveness of the proposed approach is demonstrated through extensive simulations and experimental validations using real-world datasets. The results highlight the ability of the method to handle challenging terrains and dynamic environments by compensating for wheel slippage and mitigating the influence of dynamic objects. Compared to conventional VIWO systems, the integration of GPR and LSTM layers significantly improves localization accuracy and robustness. This work paves the way for deploying VIWO systems in diverse and unpredictable environments, contributing to advancements in autonomous navigation and multi-sensor fusion technologies.

1. Introduction

In wheeled mobile robot navigation, visual inertial navigation system (VINS) algorithms are considered state of the art for localization in GPS-denied environments [1,2,3]. The initialization procedure of VINS algorithms significantly influence the overall accuracy and robustness of the algorithm [4,5]. The initialization methods can be categorized as joint versus disjoint ones [6]. Joint methods estimate both the camera trajectory and inertial parameters simultaneously within a tightly coupled framework. This approach ensures global consistency in the estimates, reducing error propagation. However, joint methods often require higher computational resources and are more complex to implement, making them less practical for real-time applications. On the other hand, the disjoint methods assume that an accurate camera trajectory can be derived through monocular vision, and the inertial parameters are then estimated by using this trajectory.
Many of these algorithms rely on the use of monocular vision due to its affordability. However, they present inherent challenges, particularly with degenerate motions that introduce unobservable directions in the estimation process [7]. These motions, often uniform linear or circular, make scale and orientation unobservable, compromising the system’s robustness and accuracy [8]. To address these issues, researchers have integrated wheel odometry (WO) measurements into VINS, resulting in robust initialization and improved performance [9,10,11]. The existing VIWO schemes mainly focus on fusing WO information in the backend while ignoring it in the front end [12] and often assume little to no wheel slippage, which does not hold in practice [13]. Furthermore, side-slip effects pose challenges for tracking cornering trajectories, which require additional adaptive control techniques, observer design, and trajectory planning schemes [14,15]. Hence, when augmenting a VINS with a WO one, the longitudinal and lateral errors due to wheel slippage need to be taken into account and compensated.
Many studies in the literature on slip compensation rely on accurate slip ratio models [16], which may not hold under varying surface conditions. As an alternative, Gaussian process (GP) modeling can encode prior distributions over functions, allowing updates to derive posterior distributions based on new test data. Leveraging this feature of GP modeling, regression tasks can be carried out where estimates of a function model are obtained given observed input–output data [17]. Notably, in the recent past, GP regression (GPR) has been used for learning the inverse dynamics of systems [18,19]. Unlike these model-based approaches, Liu et al. address slip effects by detecting and omitting updates during slip events using a threshold-based method, ensuring real-time performance but leaving the underlying slip dynamics unmodeled [20]. In contrast, our work utilizes GPR alongside other tools to estimate wheel slippage continuously, providing a more robust and adaptive compensation strategy.
In visual odometry (VO) estimations using sparse features, random sample consensus (RANSAC) is used as the preferred outlier handling mechanism [21]. However, RANSAC’s performance is inadequate when a significant portion of feature points are in motion or when dynamic objects obstruct the camera’s field of view (FOV) [22]. To address this issue, additional sensory data, such as those obtained from IMUs, are utilized to evaluate the reliability of matched features and eliminate dynamic feature points [23], which represent a preliminary version of the results of this article.
Recent state-of-the-art visual-inertial-wheel odometry (VIWO) algorithms have focused on addressing key challenges such as long-term drift, sensor calibration, and multi-modal data integration. The robust neural gyroscope calibration VIWO (RNGC-VIWO) system by Zhi et al. reduces gyroscope drift and improves localization accuracy through neural gyroscope calibration, while the approach by Lee et al. performs online calibration of wheel encoders’ intrinsic and extrinsic parameters to maintain consistent performance across various terrains [24]. Additionally, the partial invariant extended Kalman filter VIWO (PIEKF-VIWO) algorithm by Hua et al. enhances positioning accuracy by leveraging a partial invariant extended Kalman filter to optimally fuse wheel measurements with kinematic constraints [25]. Similarly, Zhou et al. propose the real-time appearance-based mapping VIWO (RTABMAP-VIWO) method, which uniquely incorporates wheel odometry and IMU data into RTABMAP using an extended Kalman filter (EKF) to reduce local cumulative errors in simultaneous localization and mapping (SLAM) systems [26]. However, despite these advancements, a common area for improvement remains in accurately compensating for wheel slippage, which can introduce significant errors in localization, particularly in challenging terrains or during dynamic maneuvers. Most existing methods do not fully account for the longitudinal and lateral slip effects, leading to reduced accuracy in real-world applications.
Two other effective technical tools used in the literature for improving VINS and VIWO results via modeling and mitigating slippage-induced errors are recurrent neural network (RNN) and long short-term memory (LSTM) [27]. These tools are particularly useful in learning sequential correlations of a sensor’s error propagation with time to more accurately position a vehicle. LSTM networks have also been employed in global navigation satellite system (GNSS)-free navigation systems to correct drift in velocity estimates derived from optical flow measurements [28]. In such drift correction, optical odometry and radar height estimates are fused with LSTM-based velocity corrections to mitigate the effects of GNSS signal loss. Similarly, deep-learning-based visual odometry (VO) methods, such as recurrent convolutional neural networks (RCNNs) and bi-directional LSTM (Bi-LSTM) architectures, have been used to improve pose estimation accuracy by capturing both spatial and temporal dependencies [29,30]. These methods, including frameworks like MagicVO, leverage CNNs for feature extraction and Bi-LSTM for sequential modeling, outperforming traditional VO systems in pose estimation accuracy and generalization ability.
A preliminary version of the results of this article has appeared in [23]. In contrast to [23], this work introduces a data-driven methodology employing GPR and LSTM networks to address wheel slippage, which is not previously addressed in [23]. This integration of GPR and LSTM enhances the prior approach substantially via effectively mitigating errors induced by slippage; thus, it significantly improves the precision of wheel odometry measurements. Furthermore, this work advances our earlier methodology in [23] by integrating a feature confidence estimator that utilizes additional sensory data. This ensures that only reliable sensor data influences the system’s state estimation. The advancements presented in this work are rigorously validated through extensive experimental trials using real-world datasets, demonstrating the methodologies’ capability to effectively manage challenging terrains and dynamic environments. In addition, a GPR [31] scheme is designed using a deep kernel approach with LSTM layers that are known to capture spatio-temporal variations [32]. The WO error residuals are learned between the physical prediction and the predicted ground truth data obtained using a differential drive kinematic model and the history of control input values. The refined WO measurements are then fused in a multi-state constraint Kalman filter (MSCKF) [33] framework for state estimation. Comparative simulations based on real-world data and indoor experiments reveal the performance improvement using the proposed approach.
In summary, this paper introduces a novel data-driven methodology to enhance visual-inertial-wheel odometry (VIWO) by addressing wheel slippage and improving localization accuracy. The key contributions of this work are
(i)
The use of a data-driven methodology employing GPR and LSTM networks to address wheel slippage. This integration of GPR and LSTM presents a substantial enhancement over prior approach by effectively mitigating errors induced by slippage; thus, it significantly improves the precision of wheel odometry measurements.
(ii)
The incorporation of a feature confidence estimator that utilizes additional sensory data. This ensures that only reliable sensor data influence the system’s state estimation.
A comprehensive comparative analysis of the key features of our work with respect to some related state-of-art works in the literature on VIWO/VINS and some recently developed deep-learning-based VO algorithms are presented in Table A1 and Table A2 in the Appendix A. The remainder of this article is structured as follows: Section 2 outlines the system description and preliminary details. Section 3 introduces the proposed method for compensating wheel slippage and other inherent errors associated with WO. The experimental results are presented and analyzed in Section 6. Finally, Section 7 concludes the paper and discusses future research directions.

2. The System Description and Preliminaries

In this section, we briefly describe the key components of the tightly coupled visual inertial wheel odometry (VIWO) system. To describe the system, we first define the frame notations. The notations ( · ) b , ( · ) w , ( · ) c , ( · ) o denote the body-fixed IMU {b}, world {w}, camera {c}, and WO {o} frames, respectively. The origin of frame {o} is at the center of the axle connecting the two wheels of the robot. Furthermore, R represents a rotation matrix, and R o k represents the rotation in {o} frame when the k th image is taken. The following sections describe the key measurements of the system. The system receives measurements from camera, IMU, and wheel encoders. Figure 1 represents the key overall system components of our proposed method.

2.1. IMU Kinematics Model

An IMU provides angular velocity ω m and linear acceleration a m , which can be measured as [34]
ω m = ω + b g + n g ,
S a m = a + R G I g + b a + S n a ,
where ω R 3 and a R 3 are the true angular velocity and acceleration, g is a vector of gravitational acceleration, S is a symmetric matrix representing the scale factors, and n g and n a are the considered zero-mean Gaussian noises with E [ n g ] = 0 , E [ n a ] = 0 .
Both scale factors and biases can be determined and integrated into the IMU signal processing prior to their application in the multi state constraint Kalman (MSCKF) measurement update. The scale factor of the IMU is determined based on the methodology outlined in reference [5,35]. The IMU biases b a and b g are computed through a least squares optimization procedure. To determine the gyroscope bias, the optimization aims to minimize the difference between the rotation calculated from the pre-integrated IMU data and the one derived from WO [36]. Similarly, the accelerometer bias is estimated using the WO translation.
Once the biases are estimated or known, one can compensate for the bias and determine the translation of the robot with respect to the world frame p b w and the rotation q of the robot q w b k , which can be determined by solving the following differential kinematics equations [37]:
d d t ( q w b ) = 1 2 Ω ( ω m b g ) q w b
d d t ( p b w ) = v b w
d d t ( v b w ) = S a m R G I g b a
where Ω ( · ) is the skew-symmetric operator on a 3-dimensional vector. These three equations serve as references for the discrete-time process model of the proposed MSCKF.

2.2. Camera Measurement Model

The visual measurements corresponding to feature i measured by the camera projection function π ( [ x , y , z ] ) = x z , y z , are given as
z m ( i ) = π ( p f ( i ) c ) + n i ,
p f ( i ) c = R b c R w b ( p f ( i ) w p b w ) + p b c ,
where p f ( i ) w is the 3D feature position, n i is the measurement noise, R b c and p b c are the camera to IMU extrinsic information. Equation (6) is primarily used as a reference for a measurement model on refined feature points after static feature point selection. As stated in [23], nullspace marginalization can be applied to (6) to obtain a new measurement equation that is independent of feature positions, resulting in a substantial reduction in computation and the need for storing feature positions in the state vector of MSCKF.
Filtered measurement estimates z ^ m ( i ) and the state estimate x ^ are obtained by Kalman filtering based on the residuals z ˜ m ( i ) = z m ( i ) z ^ m ( i ) , x ˜ = x x ^ and the measurement model
z ˜ m ( i ) = H x , i x ˜ + H f , i p ˜ f ( i ) w + n i
where H x , i is the Jacobian related to the robot state, and H f , i is the Jacobian related to feature position. Stacking the residuals z ˜ m ( i ) , k for all feature points i and computing the nullspace marginalization as in (8), a camera measurement residual process model independent of the feature positions is obtained as
z ˜ m ( i ) = H x x ˜ + n i ,
which is directly used in the multi state constraint Kalman filter (MSCKF) design to estimate the robot states, as in [11]. Since the feature positions are instantly marginalized from the state vector, this improves the computational efficiency of the filter, making it suitable for resource constrained platforms and applications.

3. Overall Methodology

The overall structure of the proposed multi-state constraint Kalman filter (MSCKF) is discussed in this section. Figure 1 shows the block diagram with key components of the proposed system. The MSCKF takes three sets of measurements: camera images, 6-axis IMU measurements, and wheel encoder measurements. Each set of raw data measurements is processed in an interconnected manner before being fed through the measurement update of the filter.

3.1. Robot State Vector

For any k th iteration of sampling, the computation or sampling is performed at a sampled time t k . At each sampled time, the state vector x k = x I k x C k x O k of the robot consists of the current inertial state,
x I k = q ¯ w b k p b k w v b k w b g b a ,
and the lumped vector of the past n world to inertial frame transformations
x C k = q ¯ w b k 1 p b k 1 w q ¯ w b k n p b k n w ,
where q ¯ w b j denotes the quaternion corresponding to the rotation R w b j and a translation p b j w at time step j, for j = k n , , k . p b k w and v b k w are the position and velocity vectors, and b g and b a are the gyroscope and accelerometer biases, respectively. Furthermore, x O k = q ¯ o k w p o k w consists of the rotation and position information obtained using wheel encoder data.

3.2. Process Model

The proposed MSCKF uses the generic IMU nonlinear kinematics process model to propagate the state x by numerically integrating Equations (3)–(5) [38]. The discretized form of the integration can be described as
x k + 1 = f ( x k , a m , ω m , n I ) ,
where n I denotes zero-mean white Gaussian IMU measurement noise. The predicted state, x , in the prediction step is generated as
x ˇ k = f ( x ^ k , a m , ω m , 0 ) .
The associated predicted state covariance matrix, P ˇ , is propagated via
P ˇ k = Φ k P ^ k Φ k + Q k ,
where P ^ k is the estimated state covariance matrix from the previous iteration, and Φ k and Q k are the system Jacobian and noise covariance matrices [39].

3.3. Dynamic Feature Elimination

The computed translation and rotation from the process model can also be used for calculating the corresponding fundamental matrix F using the camera-IMU extrinsic parameters, which then be used in the dynamic point elimination method [23]. A fundamental matrix F can also be calculated using the image data. With both fundamental matrices, a distance measure between the epipolar line and the corresponding features using both matrices can be computed and denoted as d k , i 1 and d k , i 2 , respectively, via
[ x , y , z ] = F p f ( i ) c k , [ x , y , z ] = F p f ( i ) c k + 1
d k , i 1 = | p f ( i ) c k + 1 F p f ( i ) c k | | | x | | 2 + | | y | | 2
d k , i 2 = | p f ( i ) c k F p f ( i ) c k + 1 | | | x | | 2 + | | y | | 2
A threshold ρ is used to determine whether each matched feature point is static or dynamic via
ϕ k , i = 1 , | d k , i 1 d k , i 2 | ρ 0 , | d k , i 1 d k , i 2 | < ρ ,
where ϕ k , i indicates the consistency between the i th feature point of the k th frame and the fundamental matrix. ϕ k , i is set to 0 if the feature point is consistent and 1 if otherwise.
Additionally, there can be a case when IMU data is invalid for the elimination method. To combat this issue, we use an additional threshold μ to keep track of match count.
ψ k = 0 , i = 1 m ϕ k , i < μ 1 , i = 1 m ϕ k , i μ .
The dynamic point elimination will only use the IMU data if and only if the data are valid (i.e., ψ k = 0 ) and trustworthy. Alternatively, if the confidence in the IMU-based dynamic point elimination is insufficient, the system switches to a RANSAC-based outlier rejection scheme.

3.4. Measurement Update

After deriving the camera measurement functions from the provided measurement model in Equation (9) and wheel odometry from the wheel-slip compensation, the estimation of the robot’s pose x is executed through an extended Kalman filter (EKF) process. The bearing measurement model in Equation (6) is reformulated in the following manner.
z m , k = h ( x ˇ k ) + n m , k
where the measurement noise is assumed to be n m , k N ( 0 , R m , k ) . (20) is then linearized as
z ˜ m , k = H k x ˜ k
where H k is the measurement Jacobian. Using this linearized measurement model, the EKF update equations are set as
x ^ k + 1 = x ˇ k + K k z ˜ m , k ,
P ^ k + 1 = P ˇ k K k H k P ˇ k ,
K k = P ˇ k H k H k P ˇ k H k + R m , k 1 ,
where K k is the Kalman gain.
In WO pre-integration, one common approach is to integrate the velocity kinematics directly. However, in this research, the proposed scheme uses the odometry increments estimated from the wheel slip compensation algorithm, which is described in details in the following section.

4. Slip Compensation

Instead of directly integrating the twist of a robot for the fusion with MSCKF state estimator, the proposed algorithm utilizes slip-compensated estimation to obtain a more accurate robot pose instead. In this section, we describe the process to estimate the wheel slippage using GPR as well as the methodology used to construct the kernel to encapsulate a deep architecture. Our slip compensation design, as seen in Figure 2, involves a learning-based estimation of the (error) residuals between ground truth and physical WO model considering a planar motion differential drive robot. The goal of this scheme is to obtain more accurate robot poses based on a sequence of given velocity inputs from the encoders and the fiber optic gyroscope (FOG).
Instead of a full 6-DOF robot pose, a simplified pose (see Figure 3) is represented by the state vector
ξ k = ξ 1 k ξ 2 k ξ 3 k = x k y k ψ k ,
where ( x k , y k ) denotes the position of the robot, and ψ k denotes yaw angle of the robot at training sample k for the residual estimator design in the sequel, respectively.
The velocity of both left and right wheels, v l and v r , and the yaw angle that is converted from the pre-integrated IMU data, ψ , are used to obtain the robot’s differential odometry, u , based on the forward kinematics of the differential drive mobile robot,
u k = v k Δ ψ k = 1 2 ( v l k + v r k ) ψ k + 1 ψ k ,
where v k and Δ ψ k are the predicted linear and angular velocity of the robot based on the encoders and the FOG. The obtained differential odometry can be transformed and integrated into the corresponding positional state increment Δ ξ using the discrete-time dynamic system model [40]
Δ ξ k = ξ k + 1 ξ k = u ¯ k + w k , u ¯ k = u ¯ 1 k u ¯ 2 k u ¯ 3 k = f o ( u k ) = v k cos ( Δ ψ k ) Δ k v k sin ( Δ ψ k ) Δ k Δ ψ k , w k = w 1 k w 2 k w 3 k ,
where w k denotes the Gaussian process noise such that w i k N ( 0 , σ w i k 2 ) and Δ k denotes the sampling time.
In training of this learning-based estimator design utilizing GPR, which is formulated for predictive correction of the odometry increments Δ ξ k governed by (27), consider a sequence of inputs (from training data sequences) collected in the individual input vector U ¯ i = [ u i k , , u i k n ] R n and the corresponding sequence of state increments collected in the vector Δ X i = [ Δ ξ i k , , Δ ξ i k n ] R n .
For predictions Δ X i * = [ Δ ξ i k * , , Δ ξ i k m * ] on new data U ¯ i * R m (test sequences), the predictive inference is given by
p ( Δ X i * | U ¯ i , Δ X i , U ¯ i * ) N ( Δ X ¯ i * , cov ( Δ X i * ) ) ,
where
Δ X ¯ i * = K ( U ¯ i * , U ¯ i ) [ K ( U ¯ i , U ¯ i ) + σ w i 2 I ] 1 Δ X i
is the mean prediction, and
cov ( Δ X ¯ i * ) = K ( U ¯ * , U ¯ * ) K ( U ¯ * , U ¯ ) [ K ( U ¯ , U ¯ ) + σ w i 2 I ] 1 K ( U ¯ , U ¯ * )
is the prediction covariance; K ( U ¯ * , U ¯ * ) R m × m , K ( U ¯ * , U ¯ ) R m × n , K ( U ¯ , U ¯ * ) R n × m , and K ( U ¯ , U ¯ ) R n × n are the kernel matrices where the element (i,j) of k i j = k ( u i , u j ) . k ( · , · ) is the kernel function, which consists of the Matern kernel [41] as the base kernel k ¯ combined with the non-linear mapping given by the neural network (NN) architecture g f ( · ) . As shown in Figure 4 the kernel function can be transformed into a deep kernel architecture. Hence, the inputs are transformed as [42]
k ( u ¯ i , u ¯ j | Θ ) = k ¯ ( g f ( u ¯ i , w ) , g f ( u ¯ j , w ) ) ,
with the Matern kernel function k ¯ ( · , · ) given as
k ¯ ( u ¯ i , u ¯ j | Θ ) = σ k 2 1 + 5 | u ¯ i u ¯ j | l + 5 | u ¯ i u ¯ j | 2 3 l 2 E x p ( 5 | u ¯ i u ¯ j | l )
where σ k is the magnitude parameter and l > 0 is the scale parameter, which are hyperparameters of the kernel. The input to the kernel function is warped by the non-linear mapping g f ( u ¯ , w ) , which includes ReLU activation functions where the NN architecture is parameterized by the weights w . We construct two deep kernel architectures utilizing convolutional neural networks (CNNs) and recurrent neural networks (RNNs). The CNN-based architecture follows two layers of 1D CNN layers followed by a fully connected layer. The LSTM-based architecture follows two layers of LSTM followed by a fully connected layer.
In the training process, the hyperparameters γ = { Θ and w } , which include the weights w of the NN network and Θ = { σ k , l } the parameters of the base kernel k ¯ ( · , · ) , are learned and optimized. The magnitude parameter σ k dictates the vertical span of the kernel function in (32), and the scale parameter l dictates the rate at which the correlation between two points decreases with increasing distance. Table 1 provides the pre-defined learning parameters used. These parameters are initially set by trial and error and this library [43] was used to develop the training function that optimized the hyperparameters. As shown in Figure 4, we use the ReLU activation function for CNN-based deep kernel architecture. In addition, sigmoid and tanh activation functions are used for RNN-based deep kernel architecture. The computational bottleneck of the system for inference is the computation of [ K ( U ¯ , U ¯ ) + σ w i 2 I ] 1 . Due to the large number of inputs in the data-set, we use a sparse GPR with approximate kernel learning based on induced points. Hence, instead of having to invert an R n × n matrix, we make a low rank approximation and invert an R m × m matrix instead, where m is the number of inducing points. Using the optimized parameters, the corrected state vector increment Δ ξ i q for i = 1 , 2 , 3 , and q = 1 , , m is obtained. Once the refined compensated set of odometry increments Δ ξ q are obtained for a test sequence, they can be used by the MSCKF state estimator for updating the state.

5. Real-World Data Simulation

In this section, we present experimental findings utilizing the publicly available KAIST dataset [44]. Gathered in an urban environment populated by dynamic objects such as vehicles, pedestrians, and bicycles, the dataset was selected due to its incorporation of wheel encoder data, which are a feature missing in many existing visual inertial datasets. The data were collected using a robotic platform equipped with a camera, an IMU, and wheel encoders. Table 2 details the sensor specifications.
For training the GPR for slip compensation, the parameters employed were as follows: the optimizer was ‘Adam’ [45], which is an inbuilt optimizer in pytorch for stochastic optimization; the learning rate was set to 0.01; the process was trained over 100 epochs; and the sample rate was 100 Hz.
The total lengths of the two test sequences were 11.19 km and 7.6 km, respectively. The paths for both test sequences are shown in Figure 5 and Figure 6, respectively. To quantify the enhancement provided by our proposed method, we measured the mean absolute trajectory error (MATE) and mean absolute rotation error (MRTE) against the ground truth trajectory. We examined three variations of multi-state constraint Kalman filters (MSCKFs): a standard MSCKF VIWO, a MSCKF VIWO with dynamic feature point elimination, and a MSCKF VIWO incorporating both dynamic feature point elimination and wheel slip compensation. Table 3 demonstrates the superior alignment of the ground truth with the third approach, which combined slip compensation and dynamic point removal.
In the case of VIWO without the incorporation of slip compensation and dynamic feature point elimination, the MATE/MRTE recorded was 10.825 m. However, this error reduced to 6.121 m with the integration of these features, indicating a substantial decrease in both MATE and MRTE, which are shown in Figure 7.

6. Indoor Experimentation

We validate our proposed methodology by conducting experiments in an indoor setup. We implement and test the scheme using Clearpath Robotics’ Jackal unmanned ground vehicle (UGV) (Clearpath Robotics, Kitchener, ON, Canada), as shown in Figure 8, equipped with readily available sensors such as wheel encoders and an inertial measurement unit (IMU). We also attach an Intel RealSense T265 tracking camera (Intel Corporation, Santa Clara, CA, USA) to obtain visual data. The Jackal UGV is a skid-steer differential-drive robot that consists of four wheels, where each pair of wheels on the left and right sides is actuated by a single motor. Hence, for modeling, it can be represented as a standard two-wheel differential-drive system. The sensor specifications and associated measurements can be found in Table 4. The data from the Jackal UGV are obtained via Robot Operating System (ROS2 Humble) [46]. To validate the experimental results, we compare them with data obtained from the OptiTrack motion capture system (NaturalPoint, Inc., Corvallis, OR, USA), which serves as the ground truth reference.
Figure 9 illustrates the difference in wheel speeds when turning and traversing in an arc and the associated slippage.
With the indoor set-up, we tested the proposed algorithm under the same metrics as in Section 5. Figure 10 and Figure 11 depict the robot trajectories for the cases, which include slip compensation plotted against the ground truth trajectory obtained with the optitrack motion capture system. As seen in Figure 10 and Figure 11, which displays the barplots for RMSE(m), the trajectory obtained via augmented slip compensation and dynamic feature point elimination had the lowest RMSE, as shown in Figure 12. In summary, our methodology was able to increase the accuracy of localization for the outdoor real-world data, as well as the indoor experiments conducted using a UGV.
To further test our methodology, we used gravel and sand on the test floor to determine the effectiveness of the proposed system. Figure 13 shows the robot trajectories for when the navigating surface is layered with gravel and sand. Our proposed scheme was able to improve the localization accuracy through aiding in slippage compensation.

7. Conclusions

In this study, we introduced a data-driven Gaussian process regression (GPR) approach for wheel slip compensation, paired with a dynamic feature elimination mechanism to enhance visual-inertial-wheel odometry (VIWO)-aided navigation systems. The proposed slip compensation scheme incorporates advanced deep kernel designs and long short-term memory (LSTM) network layers, which are adept at capturing spatiotemporal variations in odometry errors. A core component of the scheme is the feature point confidence estimator, which plays a pivotal role in mitigating the adverse effects of dynamic feature points on visual measurements. By refining these measurements, the system achieves improved pose estimation accuracy. Furthermore, inertial measurements are corrected and fused with wheel encoder data through the multi-state constraint Kalman filter (MSCKF), ensuring robust state estimation. The effectiveness of the approach was validated through extensive real-world data simulations and controlled indoor experiments, demonstrating its superiority in reducing errors and enhancing localization accuracy under various operational conditions.
Extensive real-world outdoor experiments using the KAIST dataset and controlled indoor tests with a Clearpath Jackal UGV validated the effectiveness of our approach. In outdoor experiments, our method reduced the mean absolute trajectory error (MATE) from 11.45 m to 5.45 m and the mean absolute rotation error (MRTE) from 2.526° to 1.079°. Similarly, in indoor trials using OptiTrack as ground truth, the proposed scheme significantly reduced RMSE, particularly on challenging terrains such as gravel and sand, demonstrating its robustness in varying conditions.
Looking ahead, the proposed framework sets the foundation for several promising research directions. One key area of improvement is enhancing the neural network architecture used for slip compensation. Future work will explore more advanced deep learning models, such as transformer-based architectures and convolutional-recurrent hybrid networks, to improve the system’s ability to capture complex temporal dependencies in slip dynamics. Additionally, integrating unsupervised and self-supervised learning techniques into the slip compensation process could enable the system to adapt autonomously to new environments and conditions, further improving its robustness without the need for extensive labeled data. Furthermore, refining the system architecture to incorporate terrain-aware modeling using deep learning techniques could allow real-time adjustments based on surface characteristics such as friction, inclination, and deformability. This would improve adaptability in dynamic and challenging environments, making the system less sensitive to shifts in slip dynamics. Such advancements would broaden the applicability of the system to diverse scenarios, including off-road navigation, disaster response, and industrial automation. The insights gained from this research highlight the potential of combining advanced neural-network-based slip compensation with multi-sensor fusion, paving the way for more reliable and efficient autonomous navigation systems.

Author Contributions

Conceptualization, N.R., B.F. and E.H.; Methodology, N.R., B.F. and E.H.; Validation, N.R.; Formal analysis, N.R.; Investigation, O.A.-B. and T.C.; Writing—original draft, N.R.; Writing—review & editing, O.A.-B., T.C., B.F. and E.H.; Supervision, B.F. and E.H.; Project administration, O.A.-B. and B.F.; Mentorship, O.A.-B.; Funding acquisition, B.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by Canadian Mitacs Accelerate Project IT16435 and Avidbots Corp, Kitchener, ON, Canada.

Data Availability Statement

The data supporting the findings of this study are openly available at https://github.com/UW-CAMS-Lab/VIWO-SLIP accessed on 23 January 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Comparative analysis of the proposed approach’s key features with respect to some related VIWO/VINS designs in the literature.
Table A1. Comparative analysis of the proposed approach’s key features with respect to some related VIWO/VINS designs in the literature.
CategoryProposed ApproachVINS with WO [1,6,24]Filter-Based VIWO [5,25,35]VIWO with Slip Detection [20]
ObjectiveCompensate for wheel slippage and improve VIWO accuracy.Refine inertial measurements and improve the accuracy of VIWO.Fuse WO and VINS for improved localization.Detect and mitigate slip in real time.
MethodologyGPR and LSTM to model and correct slip errors; integrates a feature confidence estimator.Neural gyroscope calibration to reduce drift in localization.Partial invariant EKF (PIEKF) to optimally fuse WO and VINS.Threshold-based slip detection and rejection.
Key Innovation and LimitationsIntegrates adaptive slip correction with dynamic feature elimination for robust VIWO.Uses deep learning for enhanced gyroscope drift compensation. Affected by wheel slip and inherent IMU drifts, leading to localization errors. Ref. [24] has focused on reducing the gyroscope drift.Leverages kinematic constraints to refine odometry accuracy but assumes minimal slip, causing errors in dynamic conditions.Ensures real-time slip detection but does not model slip dynamics, leading to accumulated drift over time.
Table A2. Comparative feature summary of the proposed approach with respect to some recently developed deep-learning-based VO algorithms in the literature.
Table A2. Comparative feature summary of the proposed approach with respect to some recently developed deep-learning-based VO algorithms in the literature.
FeatureProposedGNSS-Free UAV Navigation [28]VO Using RCNNs [29]VO Using CNN and Bi-LSTM [30]
ObjectiveUses GPR and LSTM to compensate for wheel slippage, enhancing VO accuracy.LSTM-based GNSS-free UAV navigation using optical odometry and radar.Monocular VO with deep RCNNs estimating poses from raw RGB images.Monocular VO with CNN and Bi-LSTM for 6-DoF pose estimation.
MethodologyGPR-LSTM fusion to correct slippage-induced errors in VO.LSTM for drift correction and radar for scale estimation.CNN for feature extraction, RNN for sequential motion modeling.Hybrid CNN for features, Bi-LSTM for sequence learning.
Key Innovation and LimitationsCombines GPR and LSTM for adaptive slippage compensation, improving localization.Uses LSTM for velocity correction and radar for scale resolution.End-to-end learning without explicit VO components.Bi-LSTM utilizes past and future frames for improved estimation.
Main ContributionsEnhances wheel odometry with ML and sensor fusion.Reliable UAV navigation in GNSS-denied areas.VO without feature engineering or scale priors.End-to-end VO without geometric constraints.
Technological FocusGround navigation with: (1) Slippage compensation, (2) Robust localization, (3) Improved odometry.UAV navigation, GNSS signal loss mitigation.Learning-based feature extraction and motion estimation.Enhancing feature extraction and motion estimation.
ValidationTested on KAIST dataset, showing robust performance in dynamic terrains.Flight-tested, reducing velocity errors in GNSS outages.Validated on KITTI dataset, competitive with traditional VO.Tested on KITTI, ETH-asl-cla, suited for structured driving.
AdvantagesAccurate slippage compensation, improving localization in uneven terrains.Reliable UAV navigation under GNSS loss.Adaptable VO without manual tuning.No manual feature extraction, robust across environments.

References

  1. Mur-Artal, R.; Tardós, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  2. Brasch, N.; Bozic, A.; Lallemand, J.; Tombari, F. Semantic Monocular SLAM for Highly Dynamic Environments. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 393–400. [Google Scholar] [CrossRef]
  3. Huang, G.P. Visual-Inertial Navigation: A Concise Review. In Proceedings of the International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 9572–9582. [Google Scholar]
  4. Jung, J.H.; Heo, S.; Park, C. Observability Analysis of IMU Intrinsic Parameters in Stereo Visual–Inertial Odometry. IEEE Trans. Instrum. Meas. 2020, 69, 7530–7541. [Google Scholar] [CrossRef]
  5. Zhang, Z.; Scaramuzza, D. A Tutorial on Quantitative Trajectory Evaluation for Visual(-Inertial) Odometry. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 7244–7251. [Google Scholar] [CrossRef]
  6. Campos, C.; Montiel, J.M.; Tardós, J.D. Inertial-only optimization for visual-inertial initialization. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 51–57. [Google Scholar]
  7. Yang, Y.; Geneva, P.; Eckenhoff, K.; Huang, G. Degenerate motion analysis for aided INS with online spatial and temporal sensor calibration. IEEE Robot. Autom. Lett. 2019, 4, 2070–2077. [Google Scholar] [CrossRef]
  8. Wu, K.; Guo, C.; Georgiou, G.; Roumeliotis, S. VINS on wheels. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 5155–5162. [Google Scholar]
  9. Liu, J.; Gao, W.; Hu, Z. Visual-Inertial Odometry Tightly Coupled with Wheel Encoder Adopting Robust Initialization and Online Extrinsic Calibration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Macau, China, 3–8 November 2019; pp. 5391–5397. [Google Scholar]
  10. Kang, R.; Xiong, L.; Xu, M.; Zhao, J.; Zhang, P. VINS-Vehicle: A Tightly-Coupled Vehicle Dynamics Extension to Visual-Inertial State Estimator. In Proceedings of the IEEE Intelligent Transportation Systems Conference, Auckland, New Zealand, 27–30 October 2019; pp. 3593–3600. [Google Scholar]
  11. Lee, W.; Eckenhoff, K.; Yang, Y.; Geneva, P.; Huang, G. Visual-Inertial-Wheel Odometry with Online Calibration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Las Vegas, NV, USA, 25–29 October 2020; pp. 4559–4566. [Google Scholar]
  12. Peng, G.; Lu, Z.; Chen, S.; He, D.; Xinde, L. Pose Estimation Based on Wheel Speed Anomaly Detection in Monocular Visual-Inertial SLAM. IEEE Sens. J. 2021, 21, 11692–11703. [Google Scholar] [CrossRef]
  13. Kim, J.; Kim, B.K. Cornering Trajectory Planning Avoiding Slip for Differential-Wheeled Mobile Robots. IEEE Trans. Ind. Electron. 2020, 67, 6698–6708. [Google Scholar] [CrossRef]
  14. Liu, W.; Xia, X.; Xiong, L.; Lu, Y.; Gao, L.; Yu, Z. Automated Vehicle Sideslip Angle Estimation Considering Signal Measurement Characteristic. IEEE Sens. J. 2021, 21, 21675–21687. [Google Scholar] [CrossRef]
  15. Cheng, S.; Li, L.; Chen, J. Fusion Algorithm Design Based on Adaptive SCKF and Integral Correction for Side-Slip Angle Observation. IEEE Trans. Ind. Electron. 2018, 65, 5754–5763. [Google Scholar] [CrossRef]
  16. Kim, C.H.; Cho, D.I. DNN-Based Slip Ratio Estimator for Lugged-Wheel Robot Localization in Rough Deformable Terrains. IEEE Access 2023, 11, 53468–53484. [Google Scholar] [CrossRef]
  17. Sun, X.; Quan, Z.; Cai, Y.; Chen, L.; Li, B. Intelligent Tire Slip Angle Estimation Based on Sensor Targeting Configuration. IEEE Trans. Instrum. Meas. 2024, 73, 9503016. [Google Scholar] [CrossRef]
  18. Tang, T.Y.; Yoon, D.J.; Pomerleau, F.; Barfoot, T.D. Learning a Bias Correction for Lidar-Only Motion Estimation. In Proceedings of the 2018 15th Conference on Computer and Robot Vision, Toronto, ON, Canada, 8–10 May 2018; pp. 166–173. [Google Scholar]
  19. Yao, S.; Li, H.; Wang, K.; Zhang, X. High-Accuracy Calibration of a Visual Motion Measurement System for Planar 3-DOF Robots Using Gaussian Process. IEEE Sens. J. 2019, 19, 7659–7667. [Google Scholar] [CrossRef]
  20. Liu, Y.; Zhao, C.; Ren, M. An Enhanced Hybrid Visual–Inertial Odometry System for Indoor Mobile Robot. Sensors 2022, 22, 2930. [Google Scholar] [CrossRef] [PubMed]
  21. Sun, C.; Wu, X.; Sun, J.; Qiao, N.; Sun, C. Multi-Stage Refinement Feature Matching Using Adaptive ORB Features for Robotic Vision Navigation. IEEE Sens. J. 2022, 22, 2603–2617. [Google Scholar] [CrossRef]
  22. Li, A.; Wang, J.; Xu, M.; Chen, Z. DP-SLAM: A visual SLAM with moving probability towards dynamic environments. Inf. Sci. 2021, 556, 128–142. [Google Scholar] [CrossRef]
  23. Reginald, N.; Al-Buraiki, O.; Fidan, B.; Hashemi, E. Confidence Estimator Design for Dynamic Feature Point Removal in Robot Visual-Inertial Odometry. In Proceedings of the 48th Annual Conf. of the IEEE Industrial Electronics Society, Brussels, Belgium, 17–20 October 2022; pp. 1–6. [Google Scholar]
  24. Zhi, M.; Deng, C.; Zhang, H.; Tang, H.; Wu, J.; Li, B. RNGC-VIWO: Robust Neural Gyroscope Calibration Aided Visual-Inertial-Wheel Odometry for Autonomous Vehicle. Remote Sens. 2023, 15, 4292. [Google Scholar] [CrossRef]
  25. Hua, T.; Li, T.; Pei, L. PIEKF-VIWO: Visual-Inertial-Wheel Odometry using Partial Invariant Extended Kalman Filter. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 2083–2090. [Google Scholar] [CrossRef]
  26. Zhou, S.; Li, Z.; Lv, Z.; Zhou, C.; Wu, P.; Zhu, C.; Liu, W. Research on Positioning Accuracy of Mobile Robot in Indoor Environment Based on Improved RTABMAP Algorithm. Sensors 2023, 23, 9468. [Google Scholar] [CrossRef]
  27. Onyekpe, U.; Palade, V.; Kanarachos, S.; Christopoulos, S.R.G. Learning Uncertainties in Wheel Odometry for Vehicular Localisation in GNSS Deprived Environments. In Proceedings of the 19th IEEE International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 14–17 December 2020; pp. 741–746. [Google Scholar] [CrossRef]
  28. Deraz, A.A.; Badawy, O.; Elhosseini, M.A.; Mostafa, M.; Ali, H.A.; El-Desouky, A.I. Deep learning based on LSTM model for enhanced visual odometry navigation system. Ain Shams Eng. J. 2023, 14, 102050. [Google Scholar] [CrossRef]
  29. Wang, S.; Clark, R.; Wen, H.; Trigoni, N. Deepvo: Towards end-to-end visual odometry with deep recurrent convolutional neural networks. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 2043–2050. [Google Scholar]
  30. Jiao, J.; Jiao, J.; Mo, Y.; Liu, W.; Deng, Z. MagicVO: An end-to-end hybrid CNN and bi-LSTM method for monocular visual odometry. IEEE Access 2019, 7, 94118–94127. [Google Scholar] [CrossRef]
  31. Sheng, H.; Xiao, J.; Cheng, Y.; Ni, Q.; Wang, S. Short-Term Solar Power Forecasting Based on Weighted Gaussian Process Regression. IEEE Trans. Ind. Electron. 2018, 65, 300–308. [Google Scholar] [CrossRef]
  32. Saleem, H.; Malekian, R.; Munir, H. Neural Network-Based Recent Research Developments in SLAM for Autonomous Ground Vehicles: A Review. IEEE Sens. J. 2023, 23, 13829–13858. [Google Scholar] [CrossRef]
  33. Sun, W.; Li, Y.; Ding, W.; Zhao, J. A Novel Visual Inertial Odometry Based on Interactive Multiple Model and Multistate Constrained Kalman Filter. IEEE Trans. Instrum. Meas. 2024, 73, 5000110. [Google Scholar] [CrossRef]
  34. Castro-Toscano, M.J.; Rodríguez-Quiñonez, J.C.; Hernández-Balbuena, D.; Lindner, L.; Sergiyenko, O.; Rivas-Lopez, M.; Flores-Fuentes, W. A methodological use of inertial navigation systems for strapdown navigation task. In Proceedings of the IEEE 26th International Symposium on Industrial Electronics, Edinburgh, UK, 19–21 June 2017; pp. 1589–1595. [Google Scholar]
  35. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  36. Quan, M.; Piao, S.; Tan, M.; Huang, S.S. Tightly-Coupled Monocular Visual-Odometric SLAM Using Wheels and a MEMS Gyroscope. IEEE Access 2019, 7, 97374–97389. [Google Scholar] [CrossRef]
  37. Li, M.; Mourikis, A.I. Improving the accuracy of EKF-based visual-inertial odometry. In Proceedings of the IEEE International Conference on Robotics and Automation (TCRA), Saint Paul, MN, USA, 14–18 May 2012; pp. 828–835. [Google Scholar] [CrossRef]
  38. Chatfield, A.B. Fundamentals of High Accuracy Inertial Navigation; American Institute of Aeronautics and Astronautics: Reston, VA, USA, 1997. [Google Scholar]
  39. Geneva, P.; Eckenhoff, K.; Lee, W.; Yang, Y.; Huang, G. OpenVINS: A Research Platform for Visual-Inertial Estimation. In Proceedings of the IEEE International Conference on Robotics and Automation, Paris, France, 31 May–31 August 2020; pp. 4666–4672. [Google Scholar]
  40. Carlevaris-Bianco, N.; Ushani, A.K.; Eustice, R.M. University of Michigan North Campus long-term vision and lidar dataset. Int. J. Robot. Res. 2016, 35, 1023–1035. [Google Scholar] [CrossRef]
  41. Tronarp, F.; Karvonen, T.; Särkkä, S. Mixture representation of the Matern class with applications in state space approximations and Bayesian quadrature. In Proceedings of the IEEE 28th International Workshop on Machine Learning for Signal Processing, Aalborg, Denmark, 17–20 September 2018; pp. 1–6. [Google Scholar]
  42. Wilson, A.G.; Hu, Z.; Salakhutdinov, R.; Xing, E.P. Deep Kernel Learning. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics, Cadiz, Spain, 9–11 May 2016; Gretton, A., Robert, C.C., Eds.; Proceedings of Machine Learning Research. Volume 51, pp. 370–378.
  43. Uber AI Labs. Pyro: Deep Universal Probabilistic Programming; Uber AI Labs: San Francisco, CA, USA, 2023. [Google Scholar]
  44. Jeong, J.; Cho, Y.; Shin, Y.S.; Roh, H.; Kim, A. Complex Urban Dataset with Multi-level Sensors from Highly Diverse Urban Environments. Int. J. Robot. Res. 2019, 38, 642–657. [Google Scholar] [CrossRef]
  45. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; Bengio, Y., LeCun, Y., Eds.; [Google Scholar]
  46. Macenski, S.; Foote, T.; Gerkey, B.; Lalancette, C.; Woodall, W. Robot Operating System 2: Design, architecture, and uses in the wild. Sci. Robot. 2022, 7, eabm6074. [Google Scholar] [CrossRef]
Figure 1. Overall structure of the proposed VIWO scheme.
Figure 1. Overall structure of the proposed VIWO scheme.
Sensors 25 01537 g001
Figure 2. Slip compensation scheme.
Figure 2. Slip compensation scheme.
Sensors 25 01537 g002
Figure 3. Robot kinematic schematic.
Figure 3. Robot kinematic schematic.
Sensors 25 01537 g003
Figure 4. Deep kernel architecture using CNN.
Figure 4. Deep kernel architecture using CNN.
Sensors 25 01537 g004
Figure 5. Test sequence 1 odometry trajectories comparison. VIWO: visual-inertial-wheel odometry, DPE: dynamic point elimination, WSC with CNN: wheel-slip compensation with CNN for kernel design, WSC with RNN: wheel-slip compensation with RNN for kernel design.
Figure 5. Test sequence 1 odometry trajectories comparison. VIWO: visual-inertial-wheel odometry, DPE: dynamic point elimination, WSC with CNN: wheel-slip compensation with CNN for kernel design, WSC with RNN: wheel-slip compensation with RNN for kernel design.
Sensors 25 01537 g005
Figure 6. Test sequence 2 odometry trajectories comparison.
Figure 6. Test sequence 2 odometry trajectories comparison.
Sensors 25 01537 g006
Figure 7. MATE error for test sequences.
Figure 7. MATE error for test sequences.
Sensors 25 01537 g007
Figure 8. Experimental setup.
Figure 8. Experimental setup.
Sensors 25 01537 g008
Figure 9. Comparison of wheel speeds of the robot.
Figure 9. Comparison of wheel speeds of the robot.
Sensors 25 01537 g009
Figure 10. Experiment sequence 1 odometry trajectory comparison.
Figure 10. Experiment sequence 1 odometry trajectory comparison.
Sensors 25 01537 g010
Figure 11. Experiment sequence 2 odometry trajectory comparison.
Figure 11. Experiment sequence 2 odometry trajectory comparison.
Sensors 25 01537 g011
Figure 12. RMSE for the experimental trajectories.
Figure 12. RMSE for the experimental trajectories.
Sensors 25 01537 g012
Figure 13. Experiment sequence 3 odometry trajectories comparison in gravel/sand-based terrain.
Figure 13. Experiment sequence 3 odometry trajectories comparison in gravel/sand-based terrain.
Sensors 25 01537 g013
Table 1. Parameters for training GP.
Table 1. Parameters for training GP.
Training ParameterValue
OptimizerAdam
Learning rate0.01
No. of epochs100
sample rate100 Hz
Kernel length parameter1
Table 2. Sensor specifications for KAIST dataset.
Table 2. Sensor specifications for KAIST dataset.
SensorManufacturerModelDescriptionFreq. (Hz)
CameraPointgrayFlea31600 × 1200 color, 59 FPS10
IMUXsensMTi-300Enhanced AHRS Gyro100
wheel encoderRLSLM13Magnetic rotary encoder100
Table 3. MATE/MRTE KAIST dataset.
Table 3. MATE/MRTE KAIST dataset.
MethodMATE [m]MRTE [Degree]
VIWO uncorrected11.452.526
VIWO with DPE9.131.912
VIWO with DPE and CNN GPR6.041.279
VIWO with DPE and RNN GPR5.451.079
Table 4. Jackal Sensor specifications.
Table 4. Jackal Sensor specifications.
SensorDescriptionFrequency
CameraIntel realsense T26510 fps
IMU 400 Hz
Wheel encoders78,000 pulses/m quadrature100 Hz
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Reginald, N.; Al-Buraiki, O.; Choopojcharoen, T.; Fidan, B.; Hashemi, E. Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination. Sensors 2025, 25, 1537. https://doi.org/10.3390/s25051537

AMA Style

Reginald N, Al-Buraiki O, Choopojcharoen T, Fidan B, Hashemi E. Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination. Sensors. 2025; 25(5):1537. https://doi.org/10.3390/s25051537

Chicago/Turabian Style

Reginald, Niraj, Omar Al-Buraiki, Thanacha Choopojcharoen, Baris Fidan, and Ehsan Hashemi. 2025. "Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination" Sensors 25, no. 5: 1537. https://doi.org/10.3390/s25051537

APA Style

Reginald, N., Al-Buraiki, O., Choopojcharoen, T., Fidan, B., & Hashemi, E. (2025). Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination. Sensors, 25(5), 1537. https://doi.org/10.3390/s25051537

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop