Next Article in Journal
Sensing and Rendering Method of 2-Dimensional Haptic Texture
Next Article in Special Issue
Biosignal-Based Human–Machine Interfaces for Assistance and Rehabilitation: A Survey
Previous Article in Journal
Recovering the Free-Field Acoustic Characteristics of a Vibrating Structure from Bounded Noisy Underwater Environments
Previous Article in Special Issue
Optimal Deployment of Charging Stations for Aerial Surveillance by UAVs with the Assistance of Public Transportation Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Tightly Coupled Pose Measurement Based on Multi-Sensor Fusion in Mobile Robot System

1
Key Laboratory of Image Processing and Intelligent Control, Ministry of Education, Wuhan 430070, China
2
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430070, China
3
School of Automation, Southeast University, Nanjing 210096, China
4
Shantui Construction Machinery Co., Ltd., Jining 272000, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5522; https://doi.org/10.3390/s21165522
Submission received: 7 June 2021 / Revised: 26 July 2021 / Accepted: 13 August 2021 / Published: 17 August 2021

Abstract

:
Currently, simultaneous localization and mapping (SLAM) is one of the main research topics in the robotics field. Visual-inertia SLAM, which consists of a camera and an inertial measurement unit (IMU), can significantly improve robustness and enable scale weak-visibility, whereas monocular visual SLAM is scale-invisible. For ground mobile robots, the introduction of a wheel speed sensor can solve the scale weak-visibility problem and improve robustness under abnormal conditions. In this paper, a multi-sensor fusion SLAM algorithm using monocular vision, inertia, and wheel speed measurements is proposed. The sensor measurements are combined in a tightly coupled manner, and a nonlinear optimization method is used to maximize the posterior probability to solve the optimal state estimation. Loop detection and back-end optimization are added to help reduce or even eliminate the cumulative error of the estimated poses, thus ensuring global consistency of the trajectory and map. The outstanding contribution of this paper is that the wheel odometer pre-integration algorithm, which combines the chassis speed and IMU angular speed, can avoid the repeated integration caused by linearization point changes during iterative optimization; state initialization based on the wheel odometer and IMU enables a quick and reliable calculation of the initial state values required by the state estimator in both stationary and moving states. Comparative experiments were conducted in room-scale scenes, building scale scenes, and visual loss scenarios. The results showed that the proposed algorithm is highly accurate—2.2 m of cumulative error after moving 812 m (0.28%, loopback optimization disabled)—robust, and has an effective localization capability even in the event of sensor loss, including visual loss. The accuracy and robustness of the proposed method are superior to those of monocular visual inertia SLAM and traditional wheel odometers.

1. Introduction

At present, the influential simultaneous localization and mapping (SLAM) systems with monocular vision include ORB-SLAM2 [1], LSD-SLAM [2], DSO [3], and SVO [4]. Among them, ORB-SLAM2 [4] has high adaptability and strong robustness to indoor and outdoor environments. However, the algorithm itself is based on pure vision. When there are no visual features, the accuracy and robustness of the pose estimation will decrease rapidly, and the algorithm may fail. Therefore, in the follow-up development of visual SLAM, in order to overcome the shortcomings of pure vision, the strategy of multisensor fusion is adopted.
Accordingly, sensors with scale measurement capabilities and monocular vision sensors are used to perform fusion vision SLAM to increase accuracy and robustness. Relatively stable and reliable solutions can be obtained with lasers [5]; however, at present, this method is generally only suitable for large-scale scenarios, such as unmanned driving, and is unsuitable for applications with limited costs. The inertial measurement unit (IMU) has become a generally accepted option. However, it exhibits a non-negligible cumulative error if run for a long period of time [6], especially in visually restricted conditions without texture or under weak illumination, in which case the visual mile cannot be used to correct the IMU error. In [7], the scale observability of a monocular visual inertial odometer on a ground mobile robot was analyzed in detail. When the robot moved at a constant speed, due to the lack of acceleration excitation, the constraint on the scale was lost, resulting in a gradual increase in the scale uncertainty and positioning error.
A wheel speed inertial odometer was integrated with a monocular visual odometer using the extended Kalman filter (EKF) [8,9], assuming the robot to be running on an ideal plane, and 3-DOF pose measurement was performed. The wheel speed inertial odometer uses wheel speed measurement for EKF status prediction; the angular speed measurement is integrated for dead reckoning; and the visual odometer method is used for an EKF measurement update. In the above-mentioned EKF-based loose coupling method, when the visual odometer cannot accurately calculate the pose due to insufficient visual characteristics, some proportion of its output in the filter will fall, resulting in ineffective visual observation. Consequently, the accuracy is reduced. VINS-Mono used tightly coupled nonlinear optimization [10,11]. By combining the IMU pre-integral measurement and visual measurement in a tightly coupled form to achieve the maximum posterior probability, we use the nonlinear optimization method to reach the optimal state. An open-source visual inertia SLAM algorithm VINS-Fusion [12,13] was developed based on VINS-Mono, supporting multiple sensor combinations (binocular camera + IMU; monocular camera + IMU; binocular camera only); GPS is used to improve the accuracy of the global path. In pursuit of better performance, wheel speed is used in SLAM system [14]—the wheel speed sensor and the visual odometer were integrated tightly. In this way, they make least-squares rather than optimization. However, they do not consider the unreliability of wheel speed. When a robot moves on uneven surfaces, or a wheel slip occurs, an incorrect wheel speed measurement will significantly affect the scale accuracy, and can potentially lead to system failure. In [15], a batch optimization framework with formulated edges was proposed. Moreover, tightly coupled nonlinear optimization methods were used to integrate vision, IMU, and wheel speed, and to perform pose. The error cost function was composed of vision errors, inertial measurement errors, and wheel speed sensor dead reckoning errors. In addition, assuming that the vehicle was moving on an approximate plane, a “soft” plane constraint term was added to the error cost function. While the robot is running with fixed acceleration or going straight, we do not know how long it moves in real time or what the direction is. Therefore, the encoder and soft plane constraints will make the task easier.
According to whether the visual feature points are added to the state vector of the filter or the optimization equation, the visual inertial odometry method based on filtering can be divided into two types: loose coupling and tight coupling. Further, the loose coupling method is divided into two parts: the visual odometer and state estimator. The visual odometer tracks the feature points and calculates the current camera pose, and the state estimator uses the IMU to measure and adjust the camera pose obtained by the visual odometer. As the visual features are invisible to the state estimator, they cannot be adjusted based on the IMU measurement, and the poor position accuracy of the visual features causes the accuracy of the visual odometer to decline. As a result, the correlation between measurements is not fully utilized for pose measurement, thereby reducing the accuracy. In the tight coupling method, the carrier pose and position of the feature point are used as variables to be estimated, and the filter method or optimized method is used for estimation.
Thus far, for the existing visual inertia SLAM algorithm, when the robot is moving at a constant speed, or purely rotating, and encounters scenes with insufficient visual features, the problems of low accuracy and poor robustness arise. Tightly coupled wheel speed information can help solve this problem. However, there are few studies on multi-sensor fusion SLAM for wheeled mobile robots based on vision, inertia, and wheel speed measurements that are tightly coupled and optimized. There is no complete solution similar to ORB-SLAM. Therefore, research on tightly coupled monocular visual odometer combined with wheel speed measurement is significant. Therefore, we use IMU and wheel odometry to make a monocular visual odometry system to gain the scale reconstruction.
The main novelty and contributions of this paper include:
  • A multi-sensor fusion SLAM algorithm using monocular vision, inertial measurement, and wheel speed measurement is proposed. The wheel speed and IMU angular velocity pre-integration of the wheel odometer can avoid the repeated integration caused by the linearization point change in the iterative optimization process.
  • Based on the state initialization of the wheel odometer and IMU, the initial state value required by the state estimator can be quickly and reliably calculated in both stationary and moving states.
The method in this paper solves the problem of poor pose estimation accuracy caused by the weak observability of monocular visual inertial SLAM, and further improves the robustness of pose estimation.

2. Multi-Sensor State Estimation Based on Tight Coupling Optimization

2.1. SLAM Based on Multi-Sensor Fusion

The multi-sensor fusion state estimator in this study uses monocular vision, IMU, and wheel odometer measurements based on feature point optical flow tracking [16]. None of these sensors can measure the absolute pose. Therefore, the multi-sensor fusion state estimator, as an odometer algorithm, has an unavoidable cumulative error. Therefore, we used key frame selection, loop detection [17], and back-end optimization algorithms based on the VINS-Mono framework and applied them to the multi-sensor fusion state estimator to form a complete SLAM system. Figure 1 shows a system block diagram of the SLAM method.
In a multi-sensor state estimation process, the main data processing and analysis processes include raw sensor input, data pre-processing (including calibration compensation, time alignment, and pre-fused wheel odometer), state initialization, state estimation problem solving, loopback detection, and backend optimization. The processes of optimal state estimator, wheel odometer pre-integration, and initialization reflect the contributions and innovations of this paper. The rest is our flexible use of existing methods. Section 2.2, Section 2.3, Section 2.4, Section 2.5, Section 2.6 and Section 2.7 analyze the core multi-sensor fusion state estimation problem and the pre-integration process of IMU and wheel speed. Section 2.8 analyzes the initialization process of the state.

2.2. Variables to Be Estimated and Observation

Considering that the IMU zero offset always exists, when defining the variable X to be estimated, the IMU zero offset of each key frame needs to be included. Therefore, in the k key frame, we have:
X = { x k , λ l } k K , l x k = [ p B k W , v B k W , q B k W , b a k , b g k ]
For the IMU, x k R 16 × 1 is the state, p B W R 3 × 1 denotes the position, q B W R 4 × 1 is the attitude (quaternion form), v B W R 3 × 1 denotes the speed, and b a R 3 × 1 and b g R 3 × 1 are the accelerometer and gyroscope bias, respectively. K is the key frames and is the feature points. Further, λ l is the inverse depth of (the inverse of the Z-axis coordinate).
Additionally, pre-fusion wheel odometer observations are added based on VINS-Mono. The original wheel odometer is defined as:
m ^ o d o m k [ p x k p y k θ k ] T
where { p x k   p y k } is the coordinate of the wheel odometer relative to the starting point and θ k is the angle of rotation relative to the starting direction. Therefore, the observation Z used to constrain the X :
Z = { Z C i , i j , O i j } ( i , j ) K
where visual feature point observation Z C i = { z ^ i l } l i , containing all feature points i observed under the i key frame; i j = { a ^ t , ω ^ t } t i t t j is IMU pre-integration of accelerated speed and angular speed; and O i j = { Δ m ^ o d o m t , a ^ a v g t , ω ^ a v g t } t i t t j is pre-fusion wheel odometer.

2.3. Optimal Estimation Problem

The maximum posterior estimation and Bayes’ theorem shows:
X = argmax X p ( Z | X ) p ( X )
Here, p ( Z | X ) is the conditional probability of observing the occurrence of Z under the existed state X , which can be obtained by the observation equation and the covariance. p ( X ) is the prior probability (edge probability) of X . According to formula (3), we have:
p ( Z | X ) p ( X ) =   p ( X ) ( i , j ) K p ( Z C i , i j , O i j   |   X ) =   p ( X ) i K l p ( z ^ i l | x i , λ l ) ( i , j ) K p ( { a ^ t , ω ^ t } t i t t j | x i , x j ) ×   ( i , j ) K p ( { Δ m ^ o d o m t , a ^ a v g t , ω ^ a v g t } t i t t j | x i , x j )
We use the least-squares method to find the maximum posterior estimate, and use Mahalanobis distance to measure the difference between the residual and the covariance matrix.
X =   argmin X { r p H p X 2 + i K l ρ ( r C ( z ^ i l , x i , λ l ) Σ C i l 2 ) +   ( i , j ) K r ( { a ^ t , ω ^ t } t i t t j , x i , x j ) Σ i j 2 +   ( i , j ) K ρ ( r O ( { Δ m ^ o d o m t , a ^ a v g t , ω ^ a v g t } t i t t j | x i , x j ) Σ O i j 2 ) }
Here, r ( · ) Σ 2 is the Mahalanobis distance and Σ is the covariance matrix. { r p , H p } is the prior information from marginalization; r C R is the visual residual; r O R 3 × 1 is the wheel odometer residual; and r R 15 × 1 is the IMU residual. In addition, the Huber loss function ρ [18] is used to improve the robustness, where ρ is
ρ ( s ) = { 1   s 1 2 s 1   s < 1

2.4. Visual Measurement Constraints

The specific process is as follows: while the feature point l is first observed in the key frame i, it is recorded and tracked. Its spatial pose is defined as a function of the key frame i pose ( p C i , q C i ) and the inverse depth λ l of the feature point. While the feature point l is observed again in the key frame j, a visual residual term is generated. The residual term r C j l represents the error of the feature point l in the position of j. It is also called the re-projection error, which is a function of the key frame i pose ( p C i , q C i ) .
r C ( z ^ j l , X ) = r C j l ( p B i W , p B j W , q B i W , q B j W , λ l ) = [ b 1 b 2 ] T · ( P ^ l C j P l C j P l C j )
Here,
P ^ l C j = π c 1 ( z ^ i l ) = π c 1 ( [ u ^ l c j v ^ l c j ] ) P l C j = ( R B C ( R W B j ( R B i W ( R C B · 1 λ l · π c 1 ( [ u l c u v l c u ] ) + p C B ) + p B i W p B j W ) p C B )
In the formula, R 2 1 S O ( 3 ) represents the rotation matrix from 2 coordinate system to 1 coordinate system; P ^ l c j R 2 × 1 is the position where the feature point l is projected onto the unit ball in the key frame j; π c 1 is the back projection function, which can project the pixel coordinates into the camera coordinate system C j ;   P l C j R 2 × 1 is the position projected on the unit ball in the key frame j; to compare the error with P l C j , it needs to be transformed into the camera coordinate system Cj of the key frame j; and [ b 1 b 2 ] are the two orthogonal base vectors on the tangent plane of the unit ball and the projection lines with the feature points in the orthogonal direction.

2.5. IMU Constraints

In the visual odometer method based on bundle adjustment, the state of the carrier is optimized and visual measurement is used to constrain. The IMU measurement between frames is added as a constraint on the optimization framework, thereby improving robustness.

2.5.1. IMU Pre-Integration

To reduce the complicated operation caused by reintegration, we used the IMU pre-integration method [19] to fuse IMU measurements between two consecutive key frames.

2.5.2. Residual Term

The IMU pre-integration processes the IMU measurement for a continuous period based on the given IMU zero offset and obtains the relative pose constraint between the initial and end states of the time period. The IMU pre-integration residual term is defined as:
r ( { a ^ t , ω ^ t } t k t t k + 1 , x k , x k + 1 ) = [ δ α B k + 1 B k δ β B k + 1 B k δ θ B k + 1 B k δ b a B k + 1 B k δ b g B k + 1 B k ] = [ R { q B k W } T ( p B k + 1 W p B k W v B k W Δ t + 1 2 g Δ t 2 ) α ^ B k + 1 B k R { q B k W } T ( v B k + 1 W v B k W + g Δ t ) β ^ B k + 1 B k 2 { q ^ B k + 1 B k q B k W q B k + 1 W } x y z b a B k + 1 b a B k b g B k + 1 b g B k ]
The random distribution of the residual term r conforms to N ( 0 , Σ B ) and Σ B is obtained by updating the covariance equation. δ θ B k + 1 B k R 3 × 1 is three-dimensional small perturbation and δ represents the error term. The IMU pre-integration provides constraints on the variables to be optimized contained in the two key frames before and after. In the process of nonlinear optimization, the essence of the constraint is to provide the direction and gradient of the variable to be optimized by calculating the Jacobian matrix of the residual. As the direction of gravity is obtained during the initialization of the visual inertial odometer, the gravity acceleration g W is not used as a variable to be optimized.

2.6. Wheeled Odometer Constraints

On ground mobile robots, a wheel speed meter is typically used to perform dead reckoning to obtain continuous relative poses of the robot. The continuous position and reliable scale estimation of the wheel odometer make it suitable for tasks such as path planning and navigation.

2.6.1. Two-Dimensional Wheel Odometer Algorithm

The two-dimensional wheel odometer has an unavoidable cumulative, error but can provide a continuous carrier trajectory. As the wheel speed meter measures the average wheel speed over a period of time, the chassis speed measurement m ^ b a s e measures the average speed during this time. The position and attitude update methods of the wheel speed odometer include Euler points, median points, and higher-order Runge–Kutta methods. As the sampling speed of the wheel speed meter is high (1 kHz), the Euler integration method is used to reduce the calculation time of the main control microcontroller. This is achieved while assuming that the chassis moves with fixed direction and speed in the original direction during the period and rotates to a new direction at the end of the time period.
The initial state of the wheel odometer is m ^ o d o m 0 = [ p x 0   p y 0   θ 0 ] T = 0 . Given the previous state m ^ o d o m k 1 = [ p x k 1   p y k 1   θ k 1 ] T of the wheel odometer, the current chassis speed measurement v ^ b a s e k = [ v x k   v y k   ω k ] T , and the time difference d t k = t k t k 1 , we can obtain the new wheel odometer state m ^ o d o m k as:
m ^ o d o m k = [ p x k   p y k   θ k ] T
[ p x k p y k ] = [ p x k 1 p y k 1 ] + [ c o s θ k 1 s i n θ k 1 s i n θ k 1 c o s θ k 1 ] [ v x k v y k ] d t k
θ k = θ k 1 + ω k d t k

2.6.2. Wheel Odometer Pre-Integration

In this paper, we propose a pre-integration method for wheel odometer. We use IMU and wheel speed to measure the relative pose between two key frames. The data of the two sensors makes pre-fusion m ^ f u s e d   o d o m . Thereafter, only the m ^ f u s e d   o d o m is measured, according to the wheel odometer kinematics equation, and a continuous calculation and integration is made to find the displacement.
The incremental update equation of the wheel odometer is:
{ p O i + 1 O k = p O i O k + R { q O i O k } Δ p O i + 1 O i = p O i O k + R { q O i O k } ( Δ p ^ O i + 1 O i + D O k + 1 O k η o s ) q O i + 1 O k = q O i O k q O i + 1 O i = q O i O k q { R B O ( ω ^ a v g i + 1 b g η g ) Δ t } b g i + 1 = b g i + η b g Δ t
where Δ p ^ O i + 1 O i [ Δ p ^ x i + 1   Δ p ^ y i + 1   0 ] T denotes wheel odometer with noise measurement; the initial state value p 0 = 0 ,   q 0 = [ 1   0   0   0 ] T . x = [ p O k   q O k   b g ] T is the pre-credit term.
The nominal weight of the wheel odometer pre-integration item can be incrementally updated based on the pre-fusion wheel odometer measurement:
{ p ^ O i + 1 O k = p ^ O i O k + R { q ^ O i O k } Δ p ^ O i + 1 O i q ^ O i + 1 O k = q ^ O i O k q ^ O i + 1 O i = q ^ O i O k q ^ { R B O ( ω ^ a v g i + 1 b g ) Δ t } b ^ g i + 1 = b ^ g i
The initial value of nominal weight: p ^ 0 = 0 ,   q ^ 0 = [ 1   0   0   0 ] T .
According to the definition of the error amount of the pre-integration term, an incremental update formula of the error amount of the pre-integration term is:
{ δ p O i + 1 O k = δ p O i O k + R { q O i O k } D O k + 1 O k η o s δ θ O i + 1 O k = δ θ O i O k ( ω ^ a v g i + 1 b g ) ^ δ θ O i O k Δ t δ b g Δ t + η g Δ t } δ b g i + 1 = δ b g i + η b g Δ t
( · ) ^ is 3 × 3 anti-symmetric matrix of Lie algebra SO (3). The nominal value of the wheel odometer pre-integration term depends on the pre-fusion wheel odometer measurement and the gyroscope zero offset. For the variable to be optimized, the zero bias of the gyroscope needs to be continuously adjusted in the pose measurement process to reduce the residual error. Therefore, in the optimization process, the partial derivatives J b g p and J b g θ of the nominal value of the pre-integral term of the wheel odometer with respect to the zero offset of the gyroscope need to be used.
According to the incremental update of the error value of the pre-integration term of the wheel odometer, the Jacobian matrix of the error value between the two frames before and after can be obtained as:
J δ x i δ x i + 1 = δ x i + 1 δ x i = I + F i
According to the definition of the nominal value of the pre-integration item of the wheel odometer, the Jacobian matrix of the error value is the nominal Jacobian matrix of J x ^ i x ^ i + 1 = J δ x i δ x i + 1 . Therefore, according to the chain-derivation rule, J x ^ 0 x ^ i + 1 = J x ^ i x ^ i + 1 J x ^ 0 x ^ i , the update equation of the nominal value of the pre-integration term of the wheel odometer with respect to the zero-biased Jacobian matrix is:
J x ^ 0 x ^ i + 1 = ( I + F i ) J x ^ 0 x ^ i
The initial value of the Jacobian matrix: J 0 = J x ^ 0 x ^ 0 = d i a g ( 9 ) .

2.6.3. Residual Term

Definition: In the least-squares problem of robot pose measurement, the wheel odometer residual term r O represents the error distance between the frame-to-frame relative displacement p ^ O k + 1 O k and the key frame displacement p O k + 1 O k in the variable to be optimized, where p ^ O k + 1 O k is the observation, and p O k + 1 O k is the estimator.
r O ( { Δ m ^ o d o m t , ω ^ a v g t } t k t t k + 1 , x k , x k + 1 ) = [ δ p O k + 1 O k ] = p O k + 1 O k p ^ O k + 1 O k
The wheel odometer residual does not include the errors δ θ O k + 1 O k   and δ b g O k + 1 O k with respect to the rotation and gyro zero offset. This is as these terms are already defined in the residual term of the IMU pre-integration. The IMU pre-integration uses the original IMU measurement as the angular velocity input, which provides higher rotational integration accuracy than the wheel odometer pre-integration measured with a lower frequency pre-fused wheel odometer.
To use the variable xk, xk+1 to represent the wheel odometer pre-integration residual term, p O k + 1 O k needs to be transformed:
p O k + 1 O k = R W O k ( p O k + 1 W p O k W ) = R O B 1 R { q B k W 1 } ( p B k + 1 W p B k W ) + R O B 1 R { q B k W 1 } R { q B k + 1 W } p O B R O B 1 p O B
We obtain the residual term expressed using only the variables to be optimized and the wheel odometer pre-integration:
r O = R O B 1 R { q B k W 1 } ( p B k + 1 W p B k W ) + R O B 1 R { q B k W 1 } R { q B k + 1 W } p O B R O B 1 p O B p ^ O k + 1 O k
Here, R O B and p O B are known constants.
As a maximum posterior problem, the robot pose measurement is the same as least-squares by introducing a covariance matrix of the residuals to transform the residuals with dimensions into a unified probability representation. The wheeled odometer residual r O obeys the covariance matrix Σ O of the wheeled odometer pre-integration, r O ~ N ( 0 ,   [ Σ O ] p ) . Here, [ Σ O ] p represents the displacement covariance in the wheel odometer pre-integration covariance matrix Σ O , [ Σ O ] p = [ Σ O ] l e f t   r i g h t   3 × 3 .
Regarding the Jacobian matrix, according to the definition of the wheel odometer residual r O , in the optimization process, the residual value r O will change with the adjustment of the previous key frame poses p B k W and q B k W , and the poses p B k + 1 W and q B k + 1 W of the next key frame, and the gyroscope zero offset b g i of the previous frame. To provide the necessary gradient direction for optimization, the system needs to be linearized in the current state x k , x k + 1 and the ratio between the increment of the residual r O and the increment of the variable to be optimized is calculated. Thus, the Jacobian matrix J x k , x k + 1 r O is defined:
r O = J x k , x k + 1 r O [ x k x k + 1 ]
[ δ p O k + 1 O k ] = J x k , x k + 1 r O [ δ p B k W δ v B k W δ θ B k W δ b a B k δ b g B k δ p B k W δ v B k W δ θ B k W δ b a B k δ b g B k ]
Here, J x k , x k + 1 r O R 3 × 30 . As the increment is small, using the quaternion definition will produce additional degrees of freedom. The increment for the rotation state in the formula is defined as the shaft angle representation.
As the wheel odometer residual is only related to some variables in the previous key frame state x k and next key frame state x k + 1 , the value of the Jacobian matrix J x k , x k + 1 r O is:
J x k , x k + 1 r O = [ δ p O k + 1 O k δ p B k W   0   δ p O k + 1 O k δ θ B k W   0   δ p O k + 1 O k δ b g i   δ p O k + 1 O k δ p B k + 1 W   0   δ p O k + 1 O k δ θ B k + 1 W   0   0 ]
δ p O k + 1 O k δ p B k W = R O B 1 R { q B k W 1 } δ p O k + 1 O k δ θ B k W = R O B 1 [ R { q B k W 1 } ( p B k + 1 W p B k W + R { q B k + 1 W } p O B ) ] ^ δ p O k + 1 O k δ b g i = J b g p δ p O k + 1 O k δ p B k + 1 W = R O B 1 R { q B k W 1 } δ p O k + 1 O k δ θ B k + 1 W = R O B 1 R { q B k W 1 } R { q B k + 1 W } ( p O B ) ^

2.7. Marginalization and Prior Constraints

The state of the key frame and its related observations are constantly removed from the optimization equation. If all observations related to the removed key frames are directly discarded, the constraints of state estimation will be reduced, and the loss of valid information will lead to a decrease in accuracy. Here, a marginalization algorithm is used, while removing the key frames, retaining the removed observations to constrain the optimization variables. According to [6], the use of the Gauss–Newton method can be understood as adding an increment to the variable to be optimized; the objective function is the smallest.
If the residual function r ( x ) is linearized at x, and the Jacobian matrix J x r is obtained, the nonlinear least-squares problem becomes a linear least-squares problem:
min δ x r ( x + δ x ) 2 min δ x r ( x ) + J x r δ x 2  
Here, r ( x ) + J x r δ x 2 = [ r ( x ) + J x r δ x ] T [ r ( x ) + J x r δ x ] . Taking the derivative of this formula with respect to δ x be 0, we obtain:
J x r T J x r δ x = J x r T r ( x )
Let H = J x r T J x r ,     b = J x r T r ( x ) ; thus, we obtain the incremental equation H δ x = b , where H is called the Hessian matrix.
Divide the variable x to be optimized into the part x a , which needs to be removed, and the part x b , δ x = [ δ x a   δ x b ] T , which needs to be retained, then the incremental equation becomes:
[ H a H b H b T H c ] [ δ x a δ x b ] = [ b a b b ]
The Schur method is used to eliminate the element to obtain the solution of δ x b :
[ I 0 H b T H a 1 I ] [ H a H b H b T H c ] [ δ x a δ x b ] = [ I 0 H b T H a 1 I ] [ b a b b ]
[ H a H b 0 H c H b T H a 1 H b ] [ δ x a δ x b ] = [ b a b b H b T H a 1 b a ]
Intercepting the second row of the above matrix, we obtain:
( H c H b T H a 1 H b ) δ x b = b b H b T H a 1 b a
In the above formula, only x b is unknown, and no information in H and b is lost. This process removes the rows and columns related to x a from the incremental equation, marginalizes the state x a that needs to be removed, and retains the historical observation constraints on the state x b . When the next image frame arrives, the prior information in the above Formula (31) will be used as a prior constraint term to construct a nonlinear least-squares problem.

2.8. State Initialization

VINS-Mono uses multiple steps to initialize the state: gyroscope zero offset correction, initializing gravity, speed, and scale coefficients, and modifying the direction of gravity. The disadvantage of this method is that it depends on sufficient visual measurement of parallax and sufficient acceleration excitation. When there is no abnormal situation, such as skidding, the wheel odometer has better accuracy and reliability over a short distance and a short duration. Compared with monocular vision, there is no scale uncertainty, and it is easier to initialize the keyframe pose, velocity, and gravity directions.

2.8.1. Gyro Zero Offset Initialization

As the gyroscope and wheel odometer measurements are on the same rigid body, the rotations of the two are the same. The relative rotation between the two key frames can be obtained through IMU pre-integration and wheel odometer pre-integration, respectively: q B k + 1 B k and q O k + 1 O k . The rotation term of the pre-integration of the wheeled odometer above is also obtained through the gyro integration and has no reference value. Therefore, during the initialization process, the gyroscope pre-integration will use the heading angle for rotation integration. Rotation term q B k + 1 B k is a function of the gyroscope bias q B k + 1 B k . If the error between q B k + 1 B k and q O k + 1 O k is used as a constraint, the gyroscope bias b g can be estimated. Assuming that the gyro bias b g k of each key frame during the initialization process is the same b g k = b g , the construction of the least-squares problem is as follows:
b g = argmin b g i = 1 k 1 q O B q ^ O i + 1 O i 1 q B O q B k + 1 B k 2
Linearizing the rotation transform at q ^ B k + 1 B k , we obtain:
q B k + 1 B k = q ^ B k + 1 B k q { J b g q b g } q ^ B k + 1 B k [ 1 1 2 J b g q b g ]
Here, J b g q is the partial derivative of the inter-frame rotation q B k + 1 B k . b g is the gyroscope zero bias. The objective function of the least-squares problem is written as:
q O B q ^ O i + 1 O i 1 q B O q B k + 1 B k = [ 1 0 ] q ^ B k + 1 B k [ 1 1 2 J b g q b g ] = q O B q ^ O i + 1 O i q B O [ 1 1 2 J b g q b g ] = q ^ B k + 1 B k 1 q O B q ^ O i + 1 O i q B O
Considering only the imaginary part of the quaternion, we obtain:
J b g q b g = 2 { q ^ B k + 1 B k 1 q O B q ^ O i + 1 O i q B O } x y z  
The above Formula (35) conforms to the format of Hx = b , and the Cholesky decomposition can be used to find the least-squares solution:
J b g q T J b g q b g = 2 J b g q T { q ^ B k + 1 B k 1 q O B q ^ O i + 1 O i q B O } x y z  
If the robot rotates rapidly during the gyro work offset initialization process, the elasticity of the wheel, the rigid connection between the wheel and the IMU, the misalignment of the wheel odometer clock and the IMU clock, and a calibration error of the wheel odometer rotation scale factor may lead to poor gyro work offset initialization results.

2.8.2. Initialization of Key Frame Speed and Gravity

As the Mecanum wheel will tremble during movement, and the wheeled odometer algorithm can only obtain the heading angle information, it is difficult to obtain accurate relative rotation between key frames through wheeled odometer integration. In the previous step, the zero offset of the gyroscope has been initialized, and the relative rotation between all key frames can be obtained through IMU pre-integration. As the rotation is known, the key frame speed and gravity can be calculated by solving linear equations. Decomposing the position term α B k + 1 B k and speed term β B k + 1 B k in the IMU pre-integration and transforming it into the form of matrix multiplication z ^ B k + 1 B k = H B k + 1 B k x B k + 1 B k , we obtain:
[ α B k + 1 B k p O B + R B k + 1 B k p O B β B k + 1 B k ] = [ I Δ t 0 1 2 R B 0 B k Δ t 2 R O B p ^ O k + 1 O k I R B k + 1 B k R B 0 B k Δ t   0   ] [ v B k B k v B k + 1 B k + 1 g B 0 s ]
Here, z ^ B k + 1 B k is the IMU pre-integration measurement between the key frames B k   and B k + 1 . The variable to be estimated related to the key frames B k   and B k + 1 is defined as x B k + 1 B k , and H B k + 1 B k   represents the constraint between the measurement z ^ B k + 1 B k   and the x B k + 1 B k . The s in the x B k + 1 B k represents the distance of the wheel odometer with respect to the actual distance, that is, the X-axis and Y-axis scale factors of the wheel odometer S x , S y . If the IMU excitation is sufficient, it can be used to calibrate the scale factor. S = 1 is defined here for the reliability of initialization.
To reduce initialization errors and improve reliability, multiple key frame measurements need to be used as constraints to calculate the key frame speed and gravity direction. By combining the multiple linear equations above, we can obtain the least-squares problem:
x = argmin x k F r a m e s z ^ B k + 1 B k H B k + 1 B k x B k + 1 B k 2 , x = [ v B 0 B 0 v B 1 B 1 v B k B k g B 0 s ]
Here, x is the variable to be estimated, and x is the optimal estimated value of x. The program uses Cholesky decomposition to solve the least-squares problem:
H T H x = H T z ^
In the formula, the matrix H is obtained by inserting all H B k + 1 B k   into empty columns at the corresponding positions of the unrelated variables and summing them, and z ^ is obtained by combining all z ^ B k + 1 B k .

3. Experiments

In this paper, a mobile robot platform is built, using a dual-processor architecture. The high-performance PC upper computer is used for data processing of each sensor and the SLAM algorithm operation, and the embedded controller lower computer is used for the motion control of the mobile robot chassis. The parameters of the high-performance PC are Intel Core i7-9750H CPU @ 2.6 GHz, 16 GB RAM. The camera is IntelRealSense ZR300.

3.1. Accuracy Verification Experiment

3.1.1. Room-Scale Pose Measurement Experiment

Experimental conditions: In a laboratory where objects are placed in a complex environment, as shown in Figure 2, the control robot walked through all the channels. The channel width is narrow, and the width at the narrowest point is less than 1 m; the movement speed was maintained at approximately 0.5 m/s; a large number of fixed marking points are arranged inside the room to obtain the real position of the robot in order to analyze the positioning error of the robot. Abnormal conditions during the experiment included:
3.
magnetic guide bars with a height of approximately 0.5 cm were fixed on the ground, and the wheels slipped slightly as they passed;
4.
due to turning too close to the weakly textured wall surface, the visual tracking was completely lost several times.
The RViz interface at the end of the test data playback is shown in Figure 3, which includes the raw picture, feature points, matched feature points, and the path of pose measurement.
The path diagrams of pose measurement, as shown in Figure 4, start along the positive X-axis. Therefore, the distance between the start point and end point shows the experimental accuracy. Throughout the experiment, the car traveled for 184.3 s, the average speed was 0.264 m/s, the maximum speed was 0.591 m/s, the cumulative translation was 51.321 m, the cumulative rotation was 3428.318°, and the chassis had no abnormalities. As this experiment was performed in an actual scene, the experimental system did not have the equipment to measure ground truth; therefore, it could only measure the distance error of the car passing several fixed road marking points. In each experiment, the robot was controlled to accurately pass the road marking point and keep the rotation direction consistent. Therefore, the pose error in Table 1 is average position error and yaw angle error among several road marking points. When the robot passes through each road marking point, a timestamp is marked. Calculate the coordinate error (∆X and ∆Y), position error (∆P(m), absolute value), and yaw angle error (∆A) between the algorithm trajectory pose and the road marking point when it is at the time stamp. For each algorithm, take the average of the errors generated by all road marking points as the final experimental result. The position error rate (∆P(%)) is the ratio of the position error to the cumulative translation. The table shows our algorithm performed best, with and without closed loops. The position error was 0.206 m and 0.015 m, respectively. As shown in Figure 4, we can clearly see the error of the different methods.
As listed in Table 1, we compared the error of the X-axis, Y-axis, position, and heading angle. According to the data presented in Table 1, the accuracy of posture estimation using the monocular vision inertial wheel odometer algorithm was better than that of the VINS-Mono algorithm, and the accumulated position error was only approximately 0.2 m after 51 m. We divided the cumulative position error by the cumulative translation distance to obtain the cumulative position error rate. The position error rate of the proposed algorithm was only 0.4%, which was lower than that of VINS-Mono (1%). This experiment verified that a multi-sensor fusion odometer algorithm could perform high-precision positioning in a room-scale indoor environment, and the result was better than that of the monocular and wheel separately. By further combining complete SLAM system, ours almost have no bias.

3.1.2. Floor Scale Experiment

Experimental conditions included:
On the first floor of the building where the laboratory was located, the floor area was approximately 250 × 100 m;
The robot ran in the central hall with 1 m/s (the area of the hall was approximately 15 × 15 m);
A large number of fixed marking points are arranged inside the corridor to obtain the real position of the robot in order to analyze the positioning error of the robot.
Anomalies during the experiment included:
There was a considerable amount of dust on the ground, which decreased the friction between the wheels, leading to a slight slip during rapid turns and a significant slip during left-to-right translation;
Due to the fast movement of the robot, the picture of the rolling shutter camera continued to exhibit disturbances;
There were cable manhole covers in numerous places in the corridor, the ground surface was uneven, and there were 2–3 cm step-like undulations. The robot vibrated significantly while passing over these obstacles;
The corridor contained semi-open areas and closed areas, the environment brightness changed significantly, and some areas were almost completely dark;
The walls around the robot in some areas were covered with tiles and had reflections;
The corridor and hall scenes had a high degree of similarity, lacking special landmarks. Loop detection was successfully performed only when the robot passed the hall halfway and finally returned to the hall;
During the experiment, pedestrians appeared several times.
The RViz interface at the end of test data playback is shown in Figure 5, which includes the raw picture, feature points, matched feature points, and the path.
Throughout the experiment, the car traveled for 889.2 s, the average speed was 0.87 m/s, the maximum speed was 1.36 m/s, the cumulative translation was 809.27 m, the cumulative rotation was 15,502.1°, and the chassis abnormal time was 26.983 s. As shown in Figure 6 and Table 2, the uneven ground made the wheel slip, so the error was very high. As shown in the floor-scale scenario, both the VINS-Mono algorithm with loop detection disabled and ours performed better; however, there was still a cumulative positioning error that could not be neglected. Compared with the complete SLAM system, VINS-Mono and ours both significantly improved the accuracy. Additionally, ours was better than that of VINS-Mono.

3.2. Robustness Verification Experiment

This time, we created experiments with sensor measurement errors or even loss. Experimental conditions included intentionally blocking the camera, resulting in the visual signal being lost for approximately 15 s. This was undertaken in the laboratory, during the movement of the robot, during which the robot kept moving and turning.
Throughout the experiment, the car traveled for 104.4 s, the average speed was 0.142 m/s, the maximum speed was 0.803 m/s, the cumulative translation was 17.809 m, the cumulative rotation was 1110.32°, and the chassis had no abnormalities. In the process of visual measurement loss, upon which the VINS-Mono is based, the state estimator was downgraded to inertial navigation dead reckoning. The error increased rapidly, thereby reducing the final positioning accuracy. As shown in Figure 7 and Table 3 during the loss of visual features, the path of the SLAM algorithm was the same as that measured by the wheel odometer. This shows that, although the positioning accuracy was affected by the loss of visual measurement, the pre-integration constraint can still provide absolute speed measurement; thus, the final positioning error was less than that of VINS-Mono.

4. Discussion

Aiming to solve the problems of low accuracy and poor robustness of the visual inertial SLAM algorithm, we designed and implemented a tightly coupled monocular visual inertial pose estimation algorithm that integrates wheel speed information. The state estimator, based on tightly coupled multi-sensor information, is used as the core of the algorithm, and the wheel odometer pre-integration that integrates wheel speed and IMU angular velocity avoids repeated integration in the iterative optimization process. The pose of the robot in the sliding window and the visual feature points are the state to be estimated. Then, the state is initialized based on the wheel odometer and IMU, so that the initial value of the state can be calculated quickly and reliably in both stationary and moving states. Finally, the residual constraint is added to the state estimator, and the optimal robot pose is solved through nonlinear optimization. Compared with the existing SLAM algorithm for many experiments, the SLAM algorithm in this paper can achieve higher pose accuracy and robustness in environments with more extreme situations. Experimental comparison results show that the SLAM algorithm proposed in this paper is feasible and effective.
In future research, our algorithm can be improved in the following ways. First, a motion capture system or D-GPS can be employed to obtain ground truth for quantitative pose comparison analysis. Second multiple monocular cameras with non-common view relationships can be used, which can improve robustness. When some cameras are blocked, the other cameras can still provide reliable visual measurement. Further, based on the robust robot pose measurement, using binocular or depth cameras to build a dense map of the environment can meet the requirements of robot navigation and obstacle avoidance.

Author Contributions

Conceptualization, G.P.; writing—original draft, Z.L.; writing—review and editing, J.P.; validation, D.H.; methodology, X.L.; resources, B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number No. 61672244, No. 91748106, and Hubei Province Natural Science Foundation of China, grant number No. 2019CFB526, and Shandong Province Key Research and Development Project of China, grant number 2019JZZY010443.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the data may involve confidential information of our research institution.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

SymbolsMeaning
X variables to be estimated
i , j , k key frame
K set of key frames
l feature point
set of feature points
x k state under key frame k
λ l inverse depth of feature point l
p B W position(displacement coordinate of B relative to W)
q B W attitude(quaternion form)
v B W speed
b a accelerometer bias
b g gyroscope bias
m ^ o d o m k wheel odometer under key frame k
m ^ means term m is the observation
defined as
{ p x k p y k }coordinate of the wheel odometer under key frame k
θ k angle of rotation of the wheel odometer under key frame k
Z observation
Z C i visual feature point observation under key frame i
z ^ i l feature points observed under key frame i
i j IMU pre-integration between key frame i and key frame j
ttime
a ^ t pre-integration of accelerated speed at time t
ω ^ t pre-integration of angular speed at time t
O i j pre-fusion wheel odometer between key frame i and key frame j
X maximum posterior estimation of variables
p ( Z | X ) conditional probability of observing the occurrence of Z under the existed state X
p ( X ) prior probability (edge probability)
r ( · ) Σ 2 Mahalanobis distance
Σ covariance matrix
{ r p , H p } prior information from marginalization
r C visual residual
r O wheel odometer residual
r IMU residual
ρ (·)Huber loss function
( p C i , q C i ) spatial pose under key frame i
r C j l error of the feature point l in the position of key frame j
R 2 1 rotation matrix from 2 coordinate system to 1 coordinate system
P ^ l c j position where the feature point l is projected onto the unit ball in the key frame j
π c 1 back projection function
C j camera coordinate system of the key frame j
P ^ l C j position projected on the unit ball in the key frame j
[ b 1 b 2 ] two orthogonal base vectors on the tangent plane of the unit ball and the projection lines with the feature points in the orthogonal direction
α ^ B k + 1 B k , β ^ B k + 1 B k ,   q ^ B k + 1 B k pre-integrated IMU measurement terms, observation
α is position term; β is velocity term; q is rotation term
δ represents the error term
δ θ B k + 1 B k three-dimensional small perturbation
p O i O k , p ^ O i O k displacement between key frame k and key frame i(the term with “^” is the observation, without “^” is the estimate)
η o s position measurement noise, Gaussian
D O k + 1 O k noise variance coefficient between key frame k and key frame k+1
η g angular velocity white noise
η b g angular velocity zero error
( · ) ^ 3×3 anti-symmetric matrix of Lie algebra
J δ x i δ x i + 1 Jacobian matrix of the error value between the two frames before and after
J x ^ i x ^ i + 1 Jacobian matrix of the nominal
R O B known constant
z ^ B k + 1 B k IMU pre-integration measurement between the key frames B k and B k + 1
x B k + 1 B k variable to be estimated related to the key frames B k and B k + 1
H B k + 1 B k constraint between the measurement z ^ B k + 1 B k and the x B k + 1 B k
S x , S y X-axis and Y-axis scale factors of the wheel odometer

References

  1. Mur-Artal, R.; Montiel, J.M.; Tardos, J.D. ORB-SLAM: A versatile and accurate monocular SLAM system. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  2. Engel, J.; Schöps, T.; Cremers, D. LSD-SLAM: Large-scale direct monocular SLAM. In European Conference on Computer Vision; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Zurich, Switzerland, 2014; pp. 834–849. [Google Scholar]
  3. Engel, J.; Koltun, V.; Cremers, D. Direct sparse odometry. IEEE Trans. Pattern Anal. Mach Intell. 2018, 40, 611–625. [Google Scholar] [CrossRef] [PubMed]
  4. Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014; pp. 15–22. [Google Scholar]
  5. Valente, M.; Joly, C.; de la Fortelle, A. Evidential SLAM fusing 2D laser scanner and stereo camera. Unmanned Syst. 2019, 7, 149–159. [Google Scholar] [CrossRef]
  6. Unsal, D.; Demirbas, K. Estimation of deterministic and stochastic IMU error parameters. In Proceedings of the 2012 IEEE/ION Position, Location and Navigation Symposium, Myrtle Beach, SC, USA, 23–26 April 2012; pp. 862–868. [Google Scholar]
  7. Wu, K.J.; Guo, C.X.; Georgiou, G.; Roumeliotis, S.I. VINS on wheels. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation, Marina Bay Sands, Singapore, 29 May–3 June 2017; pp. 5155–5162. [Google Scholar]
  8. Karamat, T.B.; Lins, R.G.; Givigi, S.N.; Noureldin, A. Novel EKF-Based Vision/Inertial System Integration for Improved Navigation. IEEE Trans. Instrum. Meas. 2017, 67, 116–125. [Google Scholar] [CrossRef]
  9. Qian, J.; Zi, B.; Wang, D.; Ma, Y.; Zhang, D. The design and development of an omni-directional mobile robot oriented to an intelligent manufacturing system. Sensors 2017, 17, 2073. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Zhang, C.; Chen, L.; Yuan, S. ST-VIO: Visual Inertial Odometry Combined with Image Segmentation and Tracking. IEEE Trans. Instrum. Meas. 2020, 69, 8562–8570. [Google Scholar] [CrossRef]
  11. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef] [Green Version]
  12. Qin, T.; Cao, S.; Pan, J.; Shen, S. A general optimization-based framework for global pose estimation with multiple sensors. arXiv 2019, arXiv:1901.03642. [Google Scholar]
  13. Qin, T.; Pan, J.; Cao, S.; Shen, S. A general optimization-based framework for local odometry estimation with multiple sensors. arXiv 2019, arXiv:1901.03638. [Google Scholar]
  14. Wang, J.; Shi, Z.; Zhong, Y. Visual SLAM incorporating wheel odometer for indoor robots. In Proceedings of the 2017 36th Chinese Control Conference (CCC), Dalian, China, 26–28 July 2017; pp. 5167–5172. [Google Scholar]
  15. Zheng, F.; Tang, H.; Liu, Y.-H. Odometry-Vision-Based Ground Vehicle Motion Estimation with SE(2)-Constrained SE(3) Poses. IEEE Trans. Cybern. 2018, 49, 2652–2663. [Google Scholar] [CrossRef] [PubMed]
  16. Yuan, W.; Li, Z.; Su, C.-Y. Multisensor-Based Navigation and Control of a Mobile Service Robot. IEEE Trans. Syst. Man. Cybern. Syst. 2019, 49, 1–11. [Google Scholar] [CrossRef]
  17. Galvez-López, D.; Tardos, J. Bags of Binary Words for Fast Place Recognition in Image Sequences. IEEE Trans. Robot. 2012, 28, 1188–1197. [Google Scholar] [CrossRef]
  18. Huber, P.J. Robust Estimation of a Location Parameter; Springer: New York, NY, USA, 1992. [Google Scholar]
  19. Forster, C.; Carlone, L.; Dellaert, F.; Scaramuzza, D. IMU preintegration on manifold for efficient visual-inertial maximum-a-posteriori estimation. In 2015 Robotics Science and Systems; Georgia Institute of Technology: Rome, Italy, 2015; pp. 1–20. [Google Scholar]
Figure 1. Block diagram of monocular vision inertial SLAM.
Figure 1. Block diagram of monocular vision inertial SLAM.
Sensors 21 05522 g001
Figure 2. Room-scale experimental environment.
Figure 2. Room-scale experimental environment.
Sensors 21 05522 g002
Figure 3. RViz interface at the end of test data playback.
Figure 3. RViz interface at the end of test data playback.
Sensors 21 05522 g003
Figure 4. Path of pose measurement at room scale.
Figure 4. Path of pose measurement at room scale.
Sensors 21 05522 g004
Figure 5. RViz interface at the end of test data playback.
Figure 5. RViz interface at the end of test data playback.
Sensors 21 05522 g005
Figure 6. (a) Path of floor-scale without loop and (b) path of floor-scale with loop.
Figure 6. (a) Path of floor-scale without loop and (b) path of floor-scale with loop.
Sensors 21 05522 g006
Figure 7. Path while visual tracking is failed.
Figure 7. Path while visual tracking is failed.
Sensors 21 05522 g007
Table 1. Room-scale pose measurement results.
Table 1. Room-scale pose measurement results.
Algorithm∆X (m)∆Y (m)∆P (m)∆P (%)∆A (°)
wheel−0.785−0.3890.871.711.907
wheel inertial−0.905−0.0400.9061.77−0.507
VINS-Mono0.851−0.2200.8791.71−0.574
VINS-Mono (loop)0.0090.0200.0220.04−0.530
ours0.148−0.1430.2060.40−0.213
ours (loop)−0.0030.0150.0150.03−0.443
Table 2. Floor-scale pose measurements results.
Table 2. Floor-scale pose measurements results.
Algorithm∆X (m)∆Y (m)∆P (m)∆P (%) ∆A (°)
wheel−83.7680.29583.76810.3132.868
wheel inertial−52.108−0.82652.1156.426.404
VINS-Mono10.692−31.42233.1914.094.080
VINS-Mono (loop)0.138−5.2815.2810.653.086
ours−2.0250.9472.2350.283.661
ours (loop)−0.2950.1650.3380.043.172
Table 3. Pose measurement results when visual tracking was lost.
Table 3. Pose measurement results when visual tracking was lost.
Algorithm∆X (m)∆Y (m)∆P (m)∆P (%) ∆A (°)
wheel−0.0420.1060.1140.64−1.078
wheel inertial−0.137−0.0140.1370.77−2.811
VINS-Mono−0.4420.8900.9945.58−2.057
VINS-Mono (loop)−0.0300.0150.0250.19−1.599
ours0.2950.0110.2951.66−2.883
ours (loop)0.020−0.0140.0230.14−1.718
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Peng, G.; Lu, Z.; Peng, J.; He, D.; Li, X.; Hu, B. Robust Tightly Coupled Pose Measurement Based on Multi-Sensor Fusion in Mobile Robot System. Sensors 2021, 21, 5522. https://doi.org/10.3390/s21165522

AMA Style

Peng G, Lu Z, Peng J, He D, Li X, Hu B. Robust Tightly Coupled Pose Measurement Based on Multi-Sensor Fusion in Mobile Robot System. Sensors. 2021; 21(16):5522. https://doi.org/10.3390/s21165522

Chicago/Turabian Style

Peng, Gang, Zezao Lu, Jiaxi Peng, Dingxin He, Xinde Li, and Bin Hu. 2021. "Robust Tightly Coupled Pose Measurement Based on Multi-Sensor Fusion in Mobile Robot System" Sensors 21, no. 16: 5522. https://doi.org/10.3390/s21165522

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop