Next Article in Journal
Output Feedback Integrated Guidance and Control Design for Autonomous Underwater Vehicles Against Maneuvering Targets
Previous Article in Journal
Introducing a Quality-Driven Approach for Federated Learning
Previous Article in Special Issue
A Novel Technique for Drone Path Planning Based on a Neighborhood Dragonfly Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Loosely Coupled Collaborative Localization Method Utilizing Integrated IMU-Aided Cameras for Multiple Autonomous Robots

1
State Key Laboratory of Explosion Science and Safety Protection, Beijing Institute of Technology, Beijing 100081, China
2
Chongqing Innovation Center, Beijing Institute of Technology, Chongqing 401100, China
3
School of Electrical Engineering, Liaoning University of Technology, Jinzhou 121000, China
4
School of Information Science and Engineering, Chongqing Jiaotong University, Chongqing 401100, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(10), 3086; https://doi.org/10.3390/s25103086
Submission received: 7 April 2025 / Revised: 1 May 2025 / Accepted: 10 May 2025 / Published: 13 May 2025

Abstract

:
IMUs (inertial measurement units) and cameras are popular sensors for autonomous localization due to their convenient integration. This article proposes a collaborative localization method, the CICEKF (collaborative IMU-aided camera extended Kalman filter), with a loosely coupled and two-step structure for the autonomous locomotion estimation of collaborative robots. The first step is for single-robot localization estimation, fusing and connecting the IMU and visual measurement data on the velocity level, which can improve the robustness and adaptability of different visual measurement approaches without redesigning the visual optimization process. The second step is for estimating the relative configuration of multiple robots, which further fuses the individual motion information to estimate the relative translation and rotation reliably. The simulation and experiment demonstrate that both steps of the filter are capable of accomplishing locomotion estimation missions, standalone or collaboratively.

1. Introduction

1.1. Motivation

The autonomous localization of multiple robots working on ground and aerial space is well-explored and is achieved with different sensor setups such as visual measurement [1], GNSS (global navigation satellite system) [2], and visual–inertial equipment [3]. However, for search and rescue robots facing emergent or indoor situations [4,5], collaborative autonomous navigation under unknown and GNSS-denied environments [6] aiming for group closed-loop motion control [7,8] is still challenging due to the need for accuracy against environmental uncertainty. Visual–inertial approaches are popular for their robustness and real-time performance against interferences [9,10] to solve localization problems for both individual and group robots. Although the vision-based odometry and SLAM (simultaneous localization and mapping) methods have achieved great accuracy for their batch-optimization capability [11,12,13], the vision-only approaches are genuinely weak in solving the collaborative localization problem owing to the lack of a common world coordinate frame as an anchorage [14,15]. On the other hand, IMUs have good real-time responsiveness, while the inherent drift affects the long-term performance.
Therefore, it is necessary to explore approaches with an open and flexible data fusion framework to improve the accuracy of cooperative group localization, which can balance the instantaneity of the IMU sensors and the accuracy of visual measurement for collaborative agents under GNSS-restricted situations.

1.2. Related Work

The research on autonomous localization based on the data fusion of the IMU and visual measurement mainly focuses on tightly and loosely coupled approaches.
The tightly coupled methods are typically biased towards visual measurement when integrating the IMU’s real-time output from the gyroscope and accelerometer into the batch-optimization process. The state-of-the-art single-robot visual–inertial localization methods, such as ORB-SLAM v3 [16] and MSCKF 2.0 [17], have shown great robustness and accuracy. TH Nguyena et al. [18] proposed a two-step optimization-based collaborative visual–inertial range localization method that separately estimates the relative transformation and single-robot localization. Tian et al. [19] proposed a distributed multi-robot system for robust and dense metric semantic SLAM that detects loop closures and performs distributed trajectory estimation when communication among individuals is available. Though achieving great performance in the tests with the benefits of post-optimization, such tightly coupled methods usually lack both openness due to the fixed optimization structure and low real-time performance due to the high computing power demand. These result in inefficiency in dealing with uncertain topology and different mounted sensors of multi-agents.
Loosely coupled methods are naturally fit to perform relative localization due to their flexible algorithm structure and timeliness by emphasizing the IMU update [20]. The main weakness is their unsatisfactory accuracy owing to the inevitable drift of the IMU measurement, which should be carefully handled.
Kelly and Sukhatme [21] achieved the self-calibration of a monocular visual–inertial system with a specifically designed UKF (unscented Kalman filter). Following a similar approach, Weiss and Siegwart [22] introduced the estimation of the world coordinate frame drift of the visual measurement into a multi-level EKF (extended Kalman filter) for a better trajectory estimation for a single drone. Based on further research, Achtelik and Weiss [23] achieved the estimation of the relative configuration of two drones by utilizing a multi-level EKF to handle the loosely coupled visual–inertial data fusion. Researchers have introduced new filter techniques, such as the invariant EKF based on Lie algebra [24], the multi-state constraint Kalman filter [25], and the equivariant filter [26,27], into the trajectory estimation with visual–inertial equipment.
These loosely coupled methods usually treat visual measurement as a black box, thus achieving a flexible structure with great portability across different sensor setups. However, owing to only considering the update with IMU data on the acceleration and velocity levels, the final correction on the position level with the visual measurement result may lose accuracy when a sudden visual measurement instability is encountered.

1.3. Our Approach

This article presents a loosely coupled filter-based approach, the CICEKF, with two steps aiming for both the autonomous localization of a single robot and the collaborative estimation of relative multi-robot localization. The time complexity of the method is O n . The SR-CICEKF (single-robot CICEKF) step focuses on the fusion of the data on the stereo vision and IMU velocity levels from the integrated sensors for a robust locomotive estimation. The MR-CICEKF (multi-robot CICEKF) step fuses outcomes of the SR-CICEKF from multiple robots to estimate the real-time relative position and attitude from each slave robot to the master robot. The two steps can run separately as standalone filters or serially as a complete CICEKF. Furthermore, to demonstrate the feasibility of the CICEKF, the observability is proven.
This article is organized as follows: The design of the state vectors of the CICEKF is described in Section 2, a detailed description of the propagation and update processes of the filter is provided in Section 3, the local weak observability of the CICEKF is demonstrated in Section 4, simulation with quintic curves is presented in the Section 5, and real-world experiments were carried out and are analyzed in Section 6. The conclusions of this study are provided in Section 7.

2. Design of the State Vector

2.1. Definition of Variables

To focus the discussion on the filter-based data fusion process, the stereo vision is treated as black-box autonomous locomotion measurement equipment. It is assumed that each robot is equipped with an IMU-aided stereo-vision sensor suite, of which the relative position and attitude of the IMU and the cameras are already calibrated without further drifting. The locomotive variables and relative transformations are defined as shown in Figure 1, and the coordinate frames and the notation of variables used in this study are provided in Table 1.
The fixed index of the master robot is set as 0, and the indexes of the slave robots are numbered from 1 to n . p c i n and q ¯ c i n represent the calibrated relative position and attitude, respectively, between the IMU and the stereo-vision coordinate frames of the n -th robot. The n -th robot’s linear translation p w c n and the rotation q ¯ w c n obtained from the black-box stereo vision are measured by referring to the world coordinate frame. p w i n and q ¯ w i n are the linear translation and the rotation of the IMU inspected in the world coordinate frame, respectively. The coupled position and attitude fusion results of each IMU–camera suite, i.e., p w i c n and q ¯ w i c n , are attached to its IMU coordinate frame to simplify the filtering process. The raw measurement outputs of the n -th IMU are the linear acceleration a m i n and angular velocity ω m i n , which are measured in its body-following coordinate frame. The relative configuration consists of the relative position and attitude between the n -th slave robot and the master robot, which are p r 0 n and q ¯ r 0 n , respectively. The units used in this article are described in the SI. This design of the master/slave relationship among multiple agents allows for convenient modification of the topology under practical situations.
In this study, the drift of the vision’s world coordinate frame is not considered since it is slow enough to be treated as noise. Furthermore, all noise variables are modeled as a random walk·. The small-angle assumption is applied, which means letting the rotation angle be converted from q ¯ be θ , and the error of q ¯ is written as δ q ¯ = q 0 , δ q T T 1 , 1 2 δ θ T T when θ 1 [28].

2.2. Construction of the SR-CICEKF State Vector

From this point onwards, for the filter construction considering the IMU-aided stereo-vision suite of any single robot, the left corner markers are omitted for simplification, such as p w i 0 = p w i .
Thus, the vector with 29 elements representing the state of the SR-CICEKF is defined as follows:
X = p w i c T   v w c T   v w i T   q ¯ w i c T   q ¯ w i T   ω c c T   ω i i T   b a i T   b ω i T T ,
where v w c is the derivative of p w c ; v w i is integrated from a m i , which possesses a bias b a i ; ω c c is the angular velocity of the stereo vision body, which is inspected in the IMU coordinate frame rather than the vision coordinate frame; and ω i i is the angular velocity of the IMU body in its own coordinate frame with a bias b ω i . Additionally, there are v ˙ w i = a w i g = R w i a m i b a i n a m i g and ω i i = ω m i b ω i n ω m i , where n a m i and n ω m i are the measurement noises, and g is the gravity vector with respect to the world coordinate frame.

2.2.1. Coupling Process for the SR-CICEKF

The independent coupling coefficients of the linear and angular velocities are μ v [ 0 , 1 ] and μ ω [ 0 , 1 ] , respectively, ensuring adjustability while enhancing motion estimation accuracy without affecting observability.
The linear velocity coupling is straightforward and written as follows:
v w i c = μ v v w i + ( 1 μ v ) R c i v w c .
The coupling process of angular velocities cannot be conducted directly since the rotations of the IMU and vision body are inspected in different coordinate frames. Assuming q ¯ w i c = q ¯ w c q ¯ w i , the following equations are obtained referring to [29]:
d d t q ¯ w i μ ω = 1 2 q ¯ w i μ ω μ ω ω ¯ i i ,
d d t q ¯ w c 1 μ ω = 1 2 q ¯ w c 1 μ ω 1 μ ω ω ¯ c c ,
where q ¯ w c is achieved by inspecting the rotation of the stereo vision body in the IMU coordinate frame, and q ¯ μ is the μ -th power of q ¯ , which denotes a unit quaternion scaling the rotation angle around the virtual axis with μ 0 , 1 [30].
According to Appendix A, Equations (3) and (4), q ¯ w i c can thus be derived as follows:
q ¯ ˙ w i c = d d t q ¯ w i μ ω q ¯ w c 1 μ ω + q ¯ w i μ ω d d t q ¯ w c 1 μ ω = μ ω 2 q ¯ w i μ ω ω ¯ i i q ¯ w c 1 μ ω + 1 - μ ω 2 q ¯ w i c ω ¯ c c ,
where ω ¯ = [ 0 , ω ] .

2.2.2. Simplification of the SR-CICEKF State Vector

To simplify the further discussion, the SR-CICEKF state vector in Equation (1) is rewritten as follows:
X = p i c   T v c   T v i   T q ¯ i c   T q ¯ i ω c T ω i T   T b a T b ω T T .
The corresponding derivatives are as follows:
p ˙ i c = v i c = μ v v i + ( 1 μ v ) R c i v c ,   v ˙ c = n v c ,   v ˙ i = a i g = R i a m i b a i n a m i g ,   q ¯ ˙ i c = μ ω 2 q ¯ i μ ω ω ¯ i q ¯ c + 1 - μ ω 1 - μ ω 2 q ¯ i c ω ¯ c ,   q ¯ ˙ i = 1 2 q ¯ i ω ¯ i ,   ω ˙ c = n ω c ,   ω ˙ i = d d t ω m i b ω i n ω m i ,   b ˙ a i = n b a i ,   b ˙ ω i = n b ω i ,
where q ¯ c = q ¯ w c q ¯ c i q ¯ w c .

2.2.3. Error of the SR-CICEKF State Vector

The error state contributes to the propagation matrices and is essential for the filtering process. X ˜ contains 27 elements, written as follows:
X ˜ = Δ p i c   T Δ v c   T Δ v i   T δ θ i c   T δ θ i   T Δ ω c   T Δ ω i   T Δ b a i Δ T b ω i T T .
The update rate of this filter-based approach is determined by the IMU refresh rate. Thus, small high-order terms in the process can be omitted, such as δ q Δ ω , δ q δ q , and δ q n since the IMU can usually run faster than 100 Hz.
To build the connection from the state vector to the propagation process, the derivative form of X ˜ should be inspected.
The derivative of the coupled translation error can be obtained directly as follows:
Δ p i c . = μ v Δ v i + ( 1 μ v ) R c i Δ v c .
Let a ^ i = a m i b ^ a i and Δ b a i = b a i b ^ a i , and according to R i = R q ¯ i R ^ i I 3 + δ θ i × achieved by following Appendix A, the equations are as follows:
Δ v i . = a i a ^ i = R i ( a m i b a i n a m i ) g R ^ i ( a m i b ^ a i ) + g R ^ i a ^ i × δ θ i R ^ i Δ b a i n a m i ,
where the higher-order terms are omitted.
According to Appendix A, the equations are as follows:
q ¯ i c * = q ¯ c 1 μ ω * q ¯ i μ ω * q ¯ i μ ω * q ¯ i μ ω = 1 q ¯ ^ c 1 μ ω * p ¯ i c q ¯ ^ c 1 μ ω R ^ q ¯ ^ c 1 μ ω T p i c = R ^ μ ω c p i c T .
By subjecting Equation (11) to Equation (5), there is obtained as follows:
δ q ¯ i c . μ ω 2 0 2 R μ ω c ω ^ i T × δ q i c 0 q 0 R μ ω c Δ T ω i 1 - μ ω 2 0 2 ω ^ c × δ q i c 0 q 0 Δ ω c ,
where Δ ω ¯ = ω ¯ ω ¯ ^ is applied, and the high-order terms are disregarded.
Following the small-angle assumption, let δ q ¯ i c = q 0 , δ q i c T T 1 , 1 2 δ θ i c T T , and there is as follows:
δ θ i c . = μ ω R μ ω c ω ^ i T × δ θ i c R μ ω c Δ T ω i 1 μ ω ω ^ c × δ θ i c Δ ω c .
Similarly, there is as follows:
δ q ¯ i . 0 ω ^ i × δ q i 0 1 2 q i 0 Δ ω i ,
where ω ^ i = ω m i b ^ ω i and Δ b ω i = b ω i b ^ ω i are applied.
Following the small-angle assumption, let δ q ¯ i = q 0 , δ q i T T 1 , 1 2 δ θ i T T , and there is as follows:
δ θ i . = ω ^ i × δ θ i Δ ω i = ω ^ i × δ θ i Δ b ω i n ω m i .
For other terms in the error state, there is no change compared to the state vector since they are obtained directly from the measurement. Therefore, there is as follows:
Δ v . c = n v c , Δ ω c . = n ω c , Δ ω i . = d d t Δ b ω i + n ω m i = n ω i , Δ b a i . = n b a i , Δ b ω i = . n b ω i .

2.3. Construction of the MR-CICEKF State Vector

One primary purpose of this study is to achieve a relative configuration between two robots. Therefore, the IMU coordinate frame, which is inherent and fixed after calibration, is chosen as the anchorage to establish the filter. For the conditions with multiple robots running simultaneously, the relative configuration chain can be easily developed with a convenient adjustment to the topology.
The construction process of the MR-CICEKF state equations is partially similar to SR-CICEKF. The input of the MR-CICEKF is obtained from the output of SR-CICEKF, and the state vector with 19 elements is written as follows:
Y = p r 0   0 T v w i   1 T v w i   0 T q ¯ r   0 T ω i i   1 T ω i i T T ,
where p r T 0 and q ¯ r T 0 have been described previously, v w i 0 and ω i i 0 are the IMU’s linear velocity and angular velocity obtained after applying the SR-CICEKF to the master robot’s sensor suite, respectively; and v w i 1 and ω i i 1 are the outputs of the slave robot.
The derivative form of Y connects the variables in the state vector. The derivative of the relative translation is obtained as follows:
p w i 1 0 p w i = R q ¯ r 0 p r 0 d d t p w i 1 0 p w i = R r ω i i 0 × p 0 + r R r p ˙ r 0 p ˙ r 0 = v w i r ω i i 0 × p r 0 ,
where v w i r = R r T v w i 1 v w i 0 is the relative linear velocity of the robots.
According to q ¯ r 0 = 0 q ¯ w i * 1 q ¯ w i , the derivative of the relative rotation is written as follows:
q ¯ ˙ r 0 = d d t q ¯ w i * 0 1 q ¯ w i + 0 q ¯ w i * d d t q ¯ w i 1 = 1 2 q ¯ r 0 1 ω i i 0 ω i i 0 q ¯ r .

2.3.1. Simplification of the MR-CICEKF State Vector

To simplify the further discussion, the MR-CICEKF state vector is rewritten as follows:
Y = p r   T v i 0   T v i 1   T q ¯ r   T ω i 0   T ω i 1 T T .
Additionally, there is v w i r = v r . The derivatives of the relative motion variables are written as follows:
p ˙ r = v r = R r T v i 1 v i 0 ω i 0 × p r ,   q ¯ ˙ r = 1 2 q ¯ r ω i 1 ω i 0 q ¯ r ,   v ˙ i 0 = n v i 0 ,   v ˙ i 1 = n v i 1 ,   ω ˙ i 0 = n ω i 0 ,   ω ˙ i 1 = n ω i 1 .

2.3.2. Error of the MR-CICEKF State Vector

Y ˜ contains 18 elements and is written as follows:
Y ˜ = Δ p r   T Δ v i 0   T Δ v i 1   T δ θ r   T Δ ω i 0   T Δ ω i 1 T T .
Following Appendix A, R r T = I 3 δ θ r × R ^ r T holds. According to p r = p ^ r + Δ p r , by omitting high-order terms, there is as follows:
Δ p ˙ r = R r v i 1 v i 0 T ω i 0 × p r R ^ r v ^ i 1 v ^ i 0 T ω ^ i 0 × p ^ r R ^ r Δ T v i 1 R ^ r Δ T v i 0 + R ^ r v i 1 v i 0 T × δ θ r + p ^ r × Δ ω i 0 ω ^ i 0 × Δ p r ,
where ω i 0 = ω ^ i 0 + Δ ω i 0 , v i 0 = v ^ i 0 + Δ v i 0 and v i 1 = v ^ i 1 + Δ v i 1 are applied.
Resembling the deduction process of Equations (11) and (12), the error of the relative rotation can be obtained as follows:
δ q ¯ ˙ r = 0 ω ^ i 1 × δ q r + 0 1 2 Δ ω i 1 0 1 2 R ^ r Δ T ω i 0 ,
After applying the small-angle assumption, the result is as follows:
δ θ ˙ r = ω ^ i 1 × δ θ r + Δ ω i 1 R ^ r Δ T ω i 0 ,
Since the linear and angular velocities are generated from the SR-CICEKF process, which can be seen as a measurement process, there is as follows:
Δ v i 0 = n v i 0 , Δ v i 1 = n v i 1 , Δ ω i 0 = n ω i 0 , Δ ω i 1 = n ω i 1 .

2.4. Relationships Among the Variables in the CICEKF

By analyzing the construction of the state vectors, it can be concluded that the SR-CICEKF and the MR-CICEKF are tightly connected in the inheritance of variables and the synchronous update, which are assembled as an entire CICEKF process. The relationships among the variables in the CICEKF states are shown in Figure 2. The whole filter is divided into three parts: data preparation, filter process, and data output. The data preparation contains the data input and fusion for both the master and slave robots as individuals. The filter process is separated into two steps, which are SR-CICEKF and MR-CICEKF.

3. Propagation and Update of the CICEKF

The propagation process is key to connecting different layers of the filter. The update process introduces measurement data to correct the entire data fusion process.

3.1. Propagation and Measurement of the SR-CICEKF

3.1.1. Propagation of the SR-CICEKF

Let the continuous state transition matrix of the SR-CICEKF be F c and the continuous noise transition matrix be G c , which are constant over each integration time step. By linearizing the continuous-time errors of a CICEKF state, there is as follows [31]:
X ˜ ˙ = F c X ˜ + G c n X ,
where the state noise vector is n X = [ n ω c T   n a m i T   n ω c T   n ω m i T   n b a i T   n b ω i T ] T .
By applying Taylor’s formula, the discrete form of Equation (27) is as follows [31]:
F d = exp ( F c Δ t ) = I d + F c Δ t + 1 2 ! F c 2 Δ t 2 + Q d = t t + Δ t F d ( τ ) G c Q c G c T F d ( τ ) T d τ ,
where F d is the discrete form of F c , and the continuous noise covariance matrix Q c = d i a g ( σ n v c 2 , σ n a i 2 , σ n ω c 2 , σ n ω i 2 , σ n b a i 2 , σ n b ω i 2 ) is a diagonal matrix with Q d being its discrete form.
By considering the first-order expansion, after computing the Jacobians X ˜ ˙ X ˜ to obtain F c , the expression of F d can be written as follows:
F d = I 3 ( 1 μ v ) R c i Δ t μ v I 3 Δ t 0 3 × 18 0 3 I 3 0 3 × 21 0 3 0 3 I 3 0 3 R ^ i a ^ i × Δ t 0 3 0 3 R ^ i Δ t 0 3 0 3 0 3 0 3 I 3 A 0 3 1 μ ω Δ t I 3 μ ω R μ ω c Δ T t 0 3 0 3 0 3 0 3 0 3 0 3 I 3 ω ^ i × Δ t 0 3 0 3 0 3 Δ t I 3 0 12 × 15 I 12 × 12 27 * 27 ,
where A = I 3 μ ω R μ ω c ω ^ i T × + 1 μ ω ω ^ c × Δ t .
Following Equations (28) and (29), Q d can be obtained from G c , which is derived according to X ˜ ˙ n X and written as follows:
G c = 0 3 × 18 I 3 0 3 0 3 0 3 0 3 0 3 0 3 R ^ i 0 3 0 3 0 3 0 3 0 3 × 18 0 3 0 3 0 3 I 3 0 3 0 3 0 3 0 3 I 3 0 3 0 3 0 3 0 3 0 3 0 3 0 3 I 3 0 3 0 6 × 18 27 * 18 .

3.1.2. Measurement of the SR-CICEKF

For an autonomous robot, the keyframes of the visual measurement can possess high localization accuracy for its post-batch optimization with visual features. Let the measurement vector of the SR-CICEKF at the k -th step be z k with its error form as z ˜ k = z ˜ p T z ¯ ˜ q T z ˜ v T z ˜ ω T T , where z ˜ P is the error of the linear position, the quaternion z ¯ ˜ q is the error of the rotation, z ˜ v is the error of the linear velocity, and z ˜ ω is the error of the angular velocity. The measurement noise of the variables is simplified as n m = n p T n θ T n v T n ω T T .
Following x ˜ = x x ^ and δ q ¯ = q ¯ ^ q ¯ , the entire description of z ˜ P , z ¯ ˜ q , and z ˜ v based on variables in the error state vector of the SR-CICEKF can be deducted as follows:
z ˜ p = Δ p + R ^ i c p i c × δ θ i c + n p z ¯ ˜ q = q ¯ ^ i c * q ¯ c i q ¯ c = δ q ¯ i c 1 1 2 δ θ i c + n θ z ˜ v = μ v Δ t R ^ i a ^ i × δ θ i μ v Δ t R ^ i Δ b a + ( 1 μ v ) R c i Δ v c + n v .
In particular, the angular velocities of both sensors cannot be directly subtracted since they do not share the same coordinate frame. Therefore, z ˜ ω is obtained indirectly from Equation (15) as follows:
z ˜ ω = δ θ i c . Δ t + n ω = Δ t μ ω R μ ω c ω ^ i T × + 1 μ ω ω ^ c × δ θ i c + 1 μ ω Δ t Δ ω c + μ ω Δ t R μ ω c Δ T ω i + n ω .
The measurement matrix H k can be recovered from z ˜ k H k X ˜ k + n m , which is written as follows:
H k = I 3 × 3 0 3 × 6 R ^ i c p i c × 0 3 × 15 0 3 × 9 1 2 I 3 × 3 0 3 × 15 0 3 × 3 Δ t ( 1 μ v ) R c i 0 3 × 6 μ v Δ t R ^ i a ^ i × 0 3 × 6 μ v Δ t R ^ i 0 3 × 3 0 3 × 9 B 0 3 × 3 1 μ ω Δ t I 3 × 3 μ ω Δ t R μ ω c T 0 3 × 6 12 * 27 ,
where B = Δ t μ ω R μ ω c ω ^ i T × + 1 μ ω ω ^ c × .

3.2. Propagation and Measurement of the MR-CICEKF

The processes of the propagation and measurement of the MR-CICEKF are similar to those of the SR-CICEKF and focus on relative motion estimation.
Following the deduction of the propagation matrix from Equation (27) to (28), let the noise vector of the state variables of the MR-CICEKF be n r = n v i 0 T n v i 1 T n ω i 0 T n ω i 1 T T . And the discrete relative state transition matrix is written as follows:
F d r = I 3 ω ^ i 0 × R ^ r T R ^ r T R ^ r v i 1 v i 0 T × p ^ r × 0 3 0 3 I 3 0 3 × 12 0 3 0 3 I 3 0 3 × 9 0 3 × 9 I 3 ω ^ i 1 × R ^ r T I 3 0 3 × 12 I 3 0 3 0 3 × 15 I 3 18 × 18 .
The discrete noise covariance matrix is Q d r = t t + Δ t F d r ( τ ) G c r Q c r G c r F d r T ( τ ) T d τ , where there is as follows:
G c r = 0 3 0 3   0 3   0 3 I 3 0 3   0 3   0 3 0 3 I 3   0 3   0 3 0 3 0 3   0 3   0 3 0 3 0 3 I 3   0 3 0 3 0 3   0 3 I 3 18 × 12 .
By defining the measurement noise vector of the MR-CICEKF as n r m = n r p T n r θ T T , let the measurement error of the MR-CICEKF at the k -th step be z ˜ r k = z ˜ r p T z ˜ r q T T , then, there is as follows:
z ˜ r p = p c 1 p c 0 p ^ r z ˜ r q = δ q ¯ r = q ¯ c * 0 q ¯ c 1 q ¯ ^ r 1 1 2 δ θ r + n r θ .
After inspecting the relationship between the measurement and the desired output of the filter through z ˜ k r H k r Y ˜ k + n r m , the measurement matrix H k r can be deducted as follows:
H k r = I 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 1 2 I 3 × 3 0 3 × 3 0 3 × 3 6 × 18 .

3.3. Entire CICEKF Process

From the analysis above, it can be concluded that the CICEKF has two major EKF-based loops to update and correct the relevant states. The first loop, i.e., the SR-CICEKF, is shown in Algorithm 1, and the second loop, i.e., the MR-CICEKF, is shown in Algorithm 2, where R k and R k r are the measurement noise matrices pre-established as described in [32,33], K k and K k r are the Kalman gains, P k + 1 | k and P k + 1 | k r are the prior covariance matrices of error, and P k + 1 | k + 1 and P k + 1 | k + 1 r are the posterior covariance matrices of the error.
The first loop governs the propagation, measurement, and update of a single robot’s motion state by fusing the measurement data of the stereo vision and the IMU on the velocity level. The second loop is focused on the estimation of the relative positions and rotations of multiple slave robots with respect to the master robot. This dual-loop design keeps the independence of either algorithm for application in different motion estimation situations. When the two algorithms work serially to yield X ^ k + 1 and Y ^ k + 1 successively, the entire algorithm complexity can be restricted to O n .
Algorithm 1: SR-CICEKF
Input:  X ˙ ˜ k , X ^ k , H k ;
1While true do
2Update F d , Q d according to X ˙ ˜ and X ^ k ; ( Propagation process in Section 3.1.1)
3Update P k + 1 | k = F d P k | k F d + Q d ;
4Update K k = P k + 1 | k H k T H k P k + 1 | k H k T + R k 1 ;
5Calculate z ˜ k according to H k ; ( Measurement process in Section 3.1.2)
6Update the current state X ^ k + 1 | k + 1 = X ^ k + 1 | k + X ˜ ^ k ;
7Update P k + 1 | k + 1 = ( I 27 × 27 K k H k ) P k + 1 | k ( I 27 × 27 K k H k ) T + K k R k K k T ;
8 k = k + 1 ;
9Output:  X ^ k + 1 .
Algorithm 2: MR-CICEKF
Input:  Y ˙ ˜ k , Y ^ k , H k r , X ^ k 0 , X ^ k 1 ;
1While true do
2Update Y ^ k according to X ^ k 0 and X ^ k 1 ;
3Update F d r , Q d r according to Y ˙ ˜ k and Y ^ k ; ( Propagation process in Section 3.2)
4Update P k + 1 | k r = F d r P k | k r F d r + Q d r ;
5Update K k r = P k + 1 | k r H k r H k r P k + 1 | k r H k r + T R k r 1 T ;
6Calculate z ˜ r k according to H k r ; ( Measurement process in Section 3.2)
7Update the current state Y ^ k + 1 | k + 1 = Y ^ k + 1 | k + Y ˜ ^ k ;
8Update P k + 1 | k + 1 r = ( I 18 × 18 K k r H k r ) P k + 1 | k r ( I 18 × 18 K k r H k r ) T + K k r R k r K k r T ;
9 k = k + 1 ;
10Output:  Y ^ k + 1 .
When updating each current state, the quaternion error is recovered from the angular errors using the following equations [28]:
δ q ^ k 1 2 δ θ ^ k q ¯ ^ k + 1 = 1 δ q ^ k δ T q ^ k , δ q ^ k T T   if   δ q ^ k δ T q ^ k 1     1 1 + δ q ^ k δ T q ^ k 1 , δ q ^ k T T   if   δ q ^ k δ T q ^ k > 1   .

4. Nonlinear Observability Analysis

4.1. Observability Analysis of the SR-CICEKF

When treating the filter as a nonlinear multi-input and multi-output system, its prerequisite for the filter to achieve correct estimation is that the system is observable. Researchers have proven that a filter possessing local weak observability can function properly, which is equivalent to the constructed observability matrix having full rank [21,34]. To simplify the discussion, a virtual measured angle velocity ω m i c and its bias b ω i c are assumed and partially coupled based on the IMU and visual measurement. Due to q ¯ i c being in the filter state, and thus treated as paired with ω m i c , this assumption does not affect the observability analysis. Thus, following Appendix A, the nonlinear measurement process of the SR-CICEKF can be expressed as follows:
f X = X ˙ = f 0 + f 1 ω m i + f 2 ω m i c + f 3 a m i ,
where for a general unit quaternion q ¯ , there is Ξ q ¯ = q T q 0 I 3 × 3 q × T T [21], and
f 0 = μ v v i + ( 1 μ v ) R c i v c T   0 1 × 3   R i b a i g T 1 2 Ξ q ¯ i c b ω i c T   1 2 Ξ q ¯ ˙ i b ω i T   0 1 × 12 T f 1 = 0 3 × 13   1 2 Ξ q ¯ i T   0 3 × 12 T f 2 = 0 3 × 9   1 2 Ξ q ¯ i c T 0 3 × 16   T f 3 = 0 3 × 6   R i 0 3 × 20 T T .
The measurement functions for the SR-CICEKF are designed as h X ˙ = h 1 , T , h 8 T T , with h 1 = ( p i c R i c p i c ) λ , h 2 = q ¯ i c , h 3 = q ¯ i , h 4 = q ¯ i c q ¯ i c T , h 5 = q ¯ i q ¯ i T , h 6 = v c , h 7 = ω c , and h 8 = ω i .
Following the Lie derivative rules in Appendix A and in [21], the observability matrix is as follows:
Ω = L 0 h 1 L 0 h 2 L 0 h 3 L 0 h 4 L 0 h 5 L 0 h 6 L 0 h 7 L 0 h 8 L f 0 1 h 1 L f 0 1 h 3 L f 0 2 h 1 = I 3 × 3 p i c 0 3 × 3 v c 0 3 × 3 v i G [ 1 , 4 ] q ¯ i c 0 3 × 4 q ¯ i 0 3 × 3 ω c 0 3 × 3 ω i 0 3 × 3 b a 0 3 × 3 b ω 0 4 × 3 0 4 × 3 0 4 × 3 I 4 × 4 0 4 × 4 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 4 I 4 × 4 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 3 0 1 × 3 0 1 × 3 0 1 × 3 G [ 4 , 4 ] 0 1 × 4 0 1 × 3 0 1 × 3 0 1 × 3 0 1 × 3 0 1 × 3 0 1 × 3 0 1 × 3 0 1 × 4 G [ 5 , 4 ] 0 1 × 3 0 1 × 3 0 1 × 3 0 1 × 3 0 3 × 3 I 3 × 3 0 3 × 3 0 3 × 4 0 3 × 4 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 4 0 3 × 4 I 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 4 0 3 × 4 0 3 × 3 I 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 G [ 9 , 2 ] μ v I 3 × 3 G [ 9 , 4 ] 0 3 × 4 0 3 × 3 0 3 × 3 G [ 9 , 8 ] 0 3 × 3 0 4 × 3 0 4 × 3 0 4 × 3 0 4 × 4 0 4 × 4 G [ 10 , 6 ] 0 4 × 3 0 4 × 3 0.5 Ξ q ¯ i 0 3 × 3 0 3 × 3 0 3 × 3 G [ 11 , 4 ] G [ 11 , 5 ] 0 3 × 3 0 3 × 3 μ v R i G [ 11 , 9 ] 31 × 29 ,
where the matrices G with the row and column indexes as subscripts do not contribute to the rank analysis.
To calculate the column rank, block Gaussian elimination can be applied. Once the corresponding block is an identity matrix, the relative column can be eliminated. Thus, only the last two columns are needed for the analysis and can be simplified as follows:
Ω = 0 0.5 Ξ q ¯ ˙ i μ v R i G [ 11 , 9 ] .
Therefore, if μ v were non-zero, and any axis of the IMU was excited in any direction, Ω has full column rank, and Ω , thus, has full column rank. This means the SR-CICEKF has local weak observability [21,34]. The above conditions of achieving full rank are easily fulfilled during the application.

4.2. Observability Analysis of the MR-CICEKF

Resembling the previous analysis, the nonlinear system representing the measurement results of the MR-CICEKF can be expressed as follows:
f Y = Y ˙ = p ˙ r   T 0 3 × 3   0 3 × 3   q ¯ ˙ T   0 3 × 3   0 3 × 3 T .
The measurement functions for MR-CICEKF are designed as h r Y ˙ = h r 1 , T , h r 6 T T , with h r 1 = p r , h r 2 = q ¯ r , h r 3 = v i 0 , h r 4 = v i 1 , h r 5 = ω i 0 , and h r 6 = ω i 1 .
After applying Lie derivative rules, the observability matrix of the MR-CICEKF system is constructed as follows:
Θ = L 0 h r 1 L 0 h r 2 L 0 h r 3 L 0 h r 4 L 0 h r 5 L 0 h r 6 = I 3 × 3 p r 0 3 × 3 v i 0 0 3 × 3 v i 1 0 3 × 4 q ¯ r 0 3 × 3 ω i 0 0 3 × 3 ω i 1 0 4 × 3 0 4 × 3 0 4 × 3 I 4 × 4 0 4 × 3 0 4 × 3 0 3 × 3 I 3 × 3 0 3 × 3 0 3 × 4 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 I 3 × 3 0 3 × 4 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 4 I 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 4 0 3 × 3 I 3 × 3 19 × 19 .
All the block columns in Θ have full column rank when the single-robot measurement yields non-zero outcomes, which means that the system has local weak observability [21,34].

5. Data Test

Two sets of motion simulations were performed to test the feasibility of the CICEKF algorithm. The motion curve was chosen as quintic, which is convenient for calculating the linear accelerations.

5.1. SR-CICEKF Simulation Test

To simplify the simulation for an individual sensor suite, the coordinate frames of the IMU and the stereo vision are assumed to coincide, and the biases are omitted. Once the parameters of the quintic curve are confirmed, the position coordinates of the curve can be treated as the position ground truth. By analyzing the projections of the curve on the three standard planes, the roll, pitch, and yaw orientations of the adjacent points can be achieved by allocating the starting and ending points, which yields the ground truth of the orientation. Thus, the ideal linear acceleration, linear velocity, and angular velocity data can be differentiated from the position and orientations, which are utilized as the filter input after being corrupted with Gaussian noise.
The position and orientation curves of the generated quintic trajectory across all three axes are shown in Figure 3 and Figure 4, respectively. It should be noted that the change in the trajectory is designed in a short time span of 10 s with a 1000 Hz update rate, which means the change in simulation data is quite fast; for example, the maximum angular velocity is over 6 rad/s, and the max linear acceleration is over 11 m2/s.
P k | k can be asymptotically stable after a complete run with the randomized initial values. By reusing the stabilized P k | k in other runs, the initial convergence of the filter process can be greatly boosted. R k is designed as a skew-symmetric matrix containing small random values. The coefficients are set as μ v = 0.5 and μ ω = 0.5 . The initial position and orientation are set as [0, 0, 0].
The three-dimensional position curves of the SR-CICEKF simulation are shown in Figure 5, with the circle mark as the starting point and the star markers as the ending points. The errors between the virtual measurement values and the filtered values of the position and orientation are shown in Figure 6 and Figure 7, respectively. The statistical results of RMSE, mean errors, and STD of the norm errors involving the three axes of the position and the orientation are listed in Table 2.
By inspecting the curves and the statistical values, it can be concluded that the SR-CICEKF quickly converges under severe data changes, while it does not need careful parameter tuning for the first estimation of P k | k . This indicates that the filter is robust, and the observability of the fusion process is identified indirectly.

5.2. MR-CICEKF Simulation

The goal of the simulation is to estimate the relative translation and rotation between the master and slave robots. The MR-CICEKF simulation utilizes the same quintic data as the target trajectories of both the master and slave robots. The absolute relative translation of the two robots is set as [1, 1, 1], and the relative rotation is set as [0, 0, 0]. The MR-CICEKF is tested independently of the SR-CICEKF. Thus, the performance of the MR-CICEKF is not affected by the possible inaccuracy produced by the SR-CICEKF process. The configurations of P k | k r and R k r are similar to the SR-CICEKF simulation, and the simulation input is corrupted with Gaussian noise.
Figure 8 depicts the three-dimensional position curves of the MR-CICEKF simulation, containing the ground truth of the master and slave robots’ trajectories, and the filtered slave robot trajectory, which is achieved by adding the estimated p r and q ¯ r to the master trajectory. The circle marks are the starting points, and the star markers are the ending points. The statistical values of RMSE, mean errors, and STD of the norm errors involving the three axes of the position and the orientation obtained after comparing the ground truth with the filtered slave robot trajectories are listed in Table 3.
The statistical values indicate that the final error is similar to the small white noise added. Thus, it can be concluded that the MR-CICEKF possesses good estimation accuracy, while the main error is caused by eliminating high-order terms during the state transition. This feature can significantly simplify further application since the MR-CICEKF contributes little error to the entire CICEKF process.

5.3. Dataset Test

To further analyze the performance of the CICEKF in real-world scenarios, a test was performed with the EuRoC dataset. The SR-CICEKF part is tested solely since the MR-CICEKF already showed great performance. The state-of-the-art visual SLAM method, ORB-SLAM V3 [16], was utilized as the visual black box measurement algorithm. The SR-CICEKF runs at 200 Hz, which is synchronous with the IMU update rate of the dataset. By employing EVO tools [35], the RPE features of the trajectories can be directly digitized.
The stereo vision of ORB-SLAM V3 runs at 20 Hz, and all measurement frames were treated as keyframes. The coefficients were set as μ v = 0.9 and μ ω = 0.5 . The initial covariance matrix P k | k was tuned following the aforementioned strategy. The SR-CICEKF, stereo ORB-SLAM V3, and its variant, along with the stereo MSCKF [17], were carried out with the room dataset of EuRoC under the same computer configuration. Figure 9 shows the comparison of the estimation trajectories of the SR-CICEKF and the ground truth. Figure 10 shows the dataset test scenarios of the algorithms separately. The RPE comparison results between each algorithm with the ground truth achieved by utilizing the EVO tools are listed in Table 4.
From the comparison, it can be concluded that by fusing IMU information and the pure visual black box measurement with the proposed SR-CICEKF algorithm, the accuracy has been greatly improved. Although the accuracy of the approach does not reach that of the optimization-based method, the fusion computational complexity is as low as O n . This can benefit the fusion with other visual measurement algorithms to improve the performance of trajectory estimation without affecting the computing efficiency. Furthermore, it can be combined with MR-CICEKF for the simultaneous estimation of multiple agents’ trajectories and collaborative configurations.

6. Experiments

Two experiments were designed to focus the discussion on the feasibility of the proposed method. Firstly, the Intel RealSense D435i, which contains a stereo set of cameras and an IMU, was mounted on a wheeled robot that performs locomotion. The SR-CICEKF estimated the locomotive trajectory while the ground truth was measured by the NOKOV motion capture system. Then, the D435i camera data were utilized for collaborative loco-motion estimation. The experimental scenarios (Supplementary Materials) are shown in Figure 11.
For both parts of the CICEKF, the black-box stereo visual algorithm is ORB-SLAM V3 operating at 30 Hz, while the IMU update rate is set to 200 Hz. The calibration of the cameras and IMUs was conducted using Kalibr tools [36] to obtain the intrinsic and extrinsic parameters and the fixed configuration of the integrated sensors.
During the experiment, we noticed that ORB-SLAM V3 and its IMU variant might not be stable under our experimental conditions due to the calibration accuracy and the IMU lacking sufficient excitation, since we used a slow ground vehicle as the motion platform. Therefore, to demonstrate the performance of the SR-CICEKF, the estimated trajectory was compared with that of stereo ORB-SLAM V3 and stereo MSCKF. The coefficients were set as μ v = 0.5 and μ ω = 0.5 . The comparison of the measured trajectory and the ground truth is shown in Figure 12. The RPE comparison results between each algorithm with the ground truth in a real scenario are listed in Table 5.
For the MR-CICEKF experiment, the two D435i cameras were calibrated with Kalibr tools to achieve a fixed relative configuration as both the ground truth and the initial guess of the filter. The relative configuration estimation of the MR-CICEKF was compared with the ground truth. The norm errors of the relative configuration estimation are listed in Table 6.
From the experimental results, it can be concluded that the proposed algorithm can not only improve the accuracy of pure visual measurement algorithms but also reliably estimate the relative configuration of multiple robots.

7. Conclusions

This study introduces a novel motion estimation approach with a loosely coupled and dual-step structure for the autonomous and collaborative localization of multiple locomotive robots. Both steps of the algorithm can run standalone, serially, or in parallel while maintaining a computational complexity of O n , which allows for convenient improvement to pure vision measurement techniques after introducing an IMU. Both the simulations and experiments demonstrate that fusing data on the velocity level without post-optimization greatly contributes to the robot trajectory estimation accuracy. Due to current experimental limitations, further research will be focused on systematically examining how the coupling coefficients affect filter performance and stability across various integrated sensors and high-speed motion units.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s25103086/s1.

Author Contributions

Conceptualization, Z.L. and S.L.; methodology, C.L.; software, Z.L. and T.W.; validation, C.L.; formal analysis, C.L.; investigation, Z.L. and S.L.; resources, Z.L. and S.L.; data curation, C.L. and T.W.; writing—original draft preparation, Z.L. and C.L.; writing—review and editing, T.W. and S.L.; visualization, Z.L. and P.T.; supervision, S.L.; project administration, Z.L. and S.L.; funding acquisition, T.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was sponsored by the National Key Research and Development Program of China [Grant number 2022YFC3320503] and the Foundation for Innovative Research Groups of the National Natural Science Foundation of China [Grant number 12221002], funded by Tao Wang.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source code presented in this study is available upon request from the corresponding author.

Acknowledgments

The authors acknowledge the anonymous reviewers for their helpful comments on the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Assuming that the rotation q ¯ is accomplished during a certain period, and the corresponding body angular velocity is ω , there is q ¯ ˙ = 1 2 q ¯ ω ¯ with ω ¯ = [ 0 , ω ] [28]. The μ -th power of q ¯ is written as q ¯ μ , which is a unit quaternion that denotes scaling the rotation angle around the virtual axis with μ 0 , 1 [30], and there is d d t q ¯ μ = 1 2 q ¯ μ μ ω ¯ [29].
Rotating a three-element vector p with a rotation q ¯ is written as q ¯ p ¯ q ¯ * R q ¯ p with p ¯ = 0 , p T T .
For general quaternions, the error between the measurement and the expectation is defined as δ q ¯ from q ¯ = q ¯ ^ δ q ¯ . The derivative form is δ q ¯ ˙ = q ¯ ^ * q ¯ ˙ q ¯ ^ ˙ δ q ¯ [23].
According to R δ q ¯ I 3 + δ θ × [28], there is R q ¯ = R ^ q ¯ ^ R δ q ¯ R ^ q ¯ ^ I 3 + δ θ × .
Assuming that there are two quaternions, q ¯ = q 0 , q T T and p ¯ = p 0 , p T T , the quaternion multiplication can be written as follows [28]:
q ¯ p ¯ = q 0 q T q q 0 I 3 + q × p ¯ = p 0 p T p p 0 I 3 + p × q ¯ .
The Lie derivative of the measurement function h x with respect to the vector f x is written as follows [21]:
L f h x = f h x = h x x f x .
The k -th-order Lie derivative of h x with respect to f x is written as follows:
L f k h x = L f k 1 h x x f x .
In particular, the zeroth-order Lie derivative of h x is the measurement function itself, i.e., L 0 h x h x .

References

  1. Servières, M.; Renaudin, V.; Dupuis, A.; Antigny, N. Visual and Visual-Inertial SLAM: State of the Art, Classification, and Experimental Benchmarking. J. Sens. 2021, 2021, 2054828. [Google Scholar] [CrossRef]
  2. Sun, Z.; Gao, W.; Tao, X.; Pan, S.; Wu, P.; Huang, H. Semi-tightly coupled robust model for GNSS/UWB/INS integrated positioning in challenging environments. Remote Sens. 2024, 16, 2108. [Google Scholar] [CrossRef]
  3. He, M.; Zhu, C.; Huang, Q.; Ren, B.; Liu, J. A review of monocular visual odometry. Vis. Comput. 2020, 36, 1053–1065. [Google Scholar] [CrossRef]
  4. Elamin, A.; El-Rabbany, A.; Jacob, S. Event-Based Visual/Inertial Odometry for UAV Indoor Navigation. Sensors 2024, 25, 61. [Google Scholar] [CrossRef] [PubMed]
  5. Hu, C.; Huang, P.; Wang, W. Tightly coupled visual-inertial-UWB indoor localization system with multiple position-unknown anchors. IEEE Robot. Autom. Lett. 2023, 9, 351–358. [Google Scholar] [CrossRef]
  6. Tonini, A.; Castelli, M.; Bates, J.S.; Lin, N.N.N.; Painho, M. Visual-Inertial Method for Localizing Aerial Vehicles in GNSS-Denied Environments. Appl. Sci. 2024, 14, 9493. [Google Scholar] [CrossRef]
  7. Nemec, D.; Šimák, V.; Janota, A.; Hruboš, M.; Bubeníková, E. Precise localization of the mobile wheeled robot using sensor fusion of odometry, visual artificial landmarks and inertial sensors. Robot. Auton. Syst. 2019, 112, 168–177. [Google Scholar] [CrossRef]
  8. Li, Z.; You, B.; Ding, L.; Gao, H.; Huang, F. Trajectory Tracking Control for WMRs with the Time-Varying Longitudinal Slippage Based on a New Adaptive SMC Method. Int. J. Aerosp. Eng. 2019, 2019, 4951538. [Google Scholar] [CrossRef]
  9. Tschopp, F.; Riner, M.; Fehr, M.; Bernreiter, L.; Furrer, F.; Novkovic, T.; Pfrunder, A.; Cadena, C.; Siegwart, R.; Nieto, J. Versavis—An open versatile multi-camera visual-inertial sensor suite. Sensors 2020, 20, 1439. [Google Scholar] [CrossRef]
  10. Reginald, N.; Al-Buraiki, O.; Choopojcharoen, T.; Fidan, B.; Hashemi, E. Visual-Inertial-Wheel Odometry with Slip Compensation and Dynamic Feature Elimination. Sensors 2025, 25, 1537. [Google Scholar] [CrossRef]
  11. Sun, X.; Zhang, C.; Zou, L.; Li, S. Real-Time Optimal States Estimation with Inertial and Delayed Visual Measurements for Unmanned Aerial Vehicles. Sensors 2023, 23, 9074. [Google Scholar] [CrossRef] [PubMed]
  12. Lajoie, P.-Y.; Beltrame, G. Swarm-slam: Sparse decentralized collaborative simultaneous localization and mapping framework for multi-robot systems. IEEE Robot. Autom. Lett. 2023, 9, 475–482. [Google Scholar] [CrossRef]
  13. Wang, S.; Wang, Y.; Li, D.; Zhao, Q. Distributed relative localization algorithms for multi-robot networks: A survey. Sensors 2023, 23, 2399. [Google Scholar] [CrossRef]
  14. Cai, Y.; Shen, Y. An integrated localization and control framework for multi-agent formation. IEEE Trans. Signal Process. 2019, 67, 1941–1956. [Google Scholar] [CrossRef]
  15. Yan, Y.; Zhang, B.; Zhou, J.; Zhang, Y.; Liu, X. Real-time localization and mapping utilizing multi-sensor fusion and visual–IMU–wheel odometry for agricultural robots in unstructured, dynamic and GPS-denied greenhouse environments. Agronomy 2022, 12, 1740. [Google Scholar] [CrossRef]
  16. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.; Tardós, J.D. Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  17. Li, M.; Mourikis, A.I. High-precision, consistent EKF-based visual-inertial odometry. Int. J. Robot. Res. 2013, 32, 690–711. [Google Scholar] [CrossRef]
  18. Nguyen, T.H.; Nguyen, T.-M.; Xie, L. Flexible and resource-efficient multi-robot collaborative visual-inertial-range localization. IEEE Robot. Autom. Lett. 2021, 7, 928–935. [Google Scholar] [CrossRef]
  19. Tian, Y.; Chang, Y.; Arias, F.H.; Nieto-Granda, C.; How, J.P.; Carlone, L. Kimera-multi: Robust, distributed, dense metric-semantic slam for multi-robot systems. IEEE Trans. Robot. 2022, 38, 2022–2038. [Google Scholar] [CrossRef]
  20. Zhang, Z.; Zhao, J.; Huang, C.; Li, L. Learning visual semantic map-matching for loosely multi-sensor fusion localization of autonomous vehicles. IEEE Trans. Intell. Veh. 2022, 8, 358–367. [Google Scholar] [CrossRef]
  21. Kelly, J.; Sukhatme, G.S. Visual-inertial sensor fusion: Localization, mapping and sensor-to-sensor self-calibration. Int. J. Robot. Res. 2011, 30, 56–79. [Google Scholar] [CrossRef]
  22. Weiss, S.; Siegwart, R. Real-time metric state estimation for modular vision-inertial systems. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 4531–4537. [Google Scholar]
  23. Achtelik, M.W.; Weiss, S.; Chli, M.; Dellaerty, F.; Siegwart, R. Collaborative stereo. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 2242–2248. [Google Scholar]
  24. Brossard, M.; Bonnabel, S.; Barrau, A. Invariant Kalman filtering for visual inertial SLAM. In Proceedings of the 2018 21st International Conference on Information Fusion (FUSION), Cambridge, UK, 10–13 July 2018; pp. 2021–2028. [Google Scholar]
  25. Sun, W.; Li, Y.; Ding, W.; Zhao, J. A novel visual inertial odometry based on interactive multiple model and multistate constrained Kalman filter. IEEE Trans. Instrum. Meas. 2023, 73, 5000110. [Google Scholar] [CrossRef]
  26. Fornasier, A.; Ng, Y.; Mahony, R.; Weiss, S. Equivariant filter design for inertial navigation systems with input measurement biases. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 4333–4339. [Google Scholar]
  27. van Goor, P.; Mahony, R. An equivariant filter for visual inertial odometry. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 14432–14438. [Google Scholar]
  28. Trawny, N.; Roumeliotis, S.I. Indirect Kalman Filter for 3D Attitude Estimation; University of Minnesota, Department of Computer Science & Engineering, Technical Report: Minneapolis, MN, USA, 2005; Volume 2. [Google Scholar]
  29. Liu, C.; Wang, T.; Li, Z.; Tian, P. A Novel Real-Time Autonomous Localization Algorithm Based on Weighted Loosely Coupled Visual–Inertial Data of the Velocity Layer. Appl. Sci. 2025, 15, 989. [Google Scholar] [CrossRef]
  30. Dam, E.B.; Koch, M.; Lillholm, M. Quaternions, Interpolation and Animation; Datalogisk Institut, Københavns Universitet: Copenhagen, Denmark, 1998; Volume 2. [Google Scholar]
  31. Maybeck, P. Stochastic Models, Estimation, and Control; Academic Press: Cambridge, MA, USA, 1982. [Google Scholar]
  32. Beder, C.; Steffen, R. Determining an initial image pair for fixing the scale of a 3d reconstruction from an image sequence. In Proceedings of the Joint Pattern Recognition Symposium, Hong Kong, China, 17–19 August 2006; pp. 657–666. [Google Scholar]
  33. Eudes, A.; Lhuillier, M. Error propagations for local bundle adjustment. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 2411–2418. [Google Scholar]
  34. Hermann, R.; Krener, A. Nonlinear controllability and observability. IEEE Trans. Autom. Control 1977, 22, 728–740. [Google Scholar] [CrossRef]
  35. Grupp, M. evo: Python Package for the Evaluation of Odometry and SLAM. Available online: https://github.com/MichaelGrupp/evo (accessed on 5 April 2025).
  36. Rehder, J.; Nikolic, J.; Schneider, T.; Hinzmann, T.; Siegwart, R. Extending kalibr: Calibrating the extrinsics of multiple IMUs and of individual axes. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 4304–4311. [Google Scholar]
Figure 1. Relationship of multi-robot coordinate frames.
Figure 1. Relationship of multi-robot coordinate frames.
Sensors 25 03086 g001
Figure 2. Relationships among the variables in the CICEKF state vector.
Figure 2. Relationships among the variables in the CICEKF state vector.
Sensors 25 03086 g002
Figure 3. Position curves of the generated quintic trajectory.
Figure 3. Position curves of the generated quintic trajectory.
Sensors 25 03086 g003
Figure 4. Orientation curves of the generated quintic trajectory.
Figure 4. Orientation curves of the generated quintic trajectory.
Sensors 25 03086 g004
Figure 5. Three-dimensional position curves of the SR-CICEKF simulation.
Figure 5. Three-dimensional position curves of the SR-CICEKF simulation.
Sensors 25 03086 g005
Figure 6. Position error curves of the SR-CICEKF simulation.
Figure 6. Position error curves of the SR-CICEKF simulation.
Sensors 25 03086 g006
Figure 7. Orientation error curves of the SR-CICEKF simulation.
Figure 7. Orientation error curves of the SR-CICEKF simulation.
Sensors 25 03086 g007
Figure 8. Three-dimensional position curves of the MR-CICEKF simulation.
Figure 8. Three-dimensional position curves of the MR-CICEKF simulation.
Sensors 25 03086 g008
Figure 9. Comparison of the estimation trajectories of the SR-CICEKF and the ground truth.
Figure 9. Comparison of the estimation trajectories of the SR-CICEKF and the ground truth.
Sensors 25 03086 g009
Figure 10. Dataset test scenarios: (a) stereo ORB-SLAM V3, (b) stereo ORB-SLAM V3 with the IMU, and (c) stereo MSCKF.
Figure 10. Dataset test scenarios: (a) stereo ORB-SLAM V3, (b) stereo ORB-SLAM V3 with the IMU, and (c) stereo MSCKF.
Sensors 25 03086 g010
Figure 11. Experimental scenarios.
Figure 11. Experimental scenarios.
Sensors 25 03086 g011
Figure 12. Comparison of the estimated trajectory and the ground truth of the SR-CICEKF.
Figure 12. Comparison of the estimated trajectory and the ground truth of the SR-CICEKF.
Sensors 25 03086 g012
Table 1. Coordinate frames and notations.
Table 1. Coordinate frames and notations.
SymbolDescription
wCoordinate frame of the fixed world
iCoordinate frame attached to the IMU’s rigid body
cCoordinate frame attached to the rigid body of the dual-camera stereo vision
icCoordinate frame attached to the virtual rigid body of the IMU-aided camera system
x A B x represents a general viable vector, A and B are the reference coordinate frames, e.g., p A B denotes the linear translation in the coordinate frame B measured with respect to the coordinate frame A
x C D x represents a general viable vector, C and D are the robot indexes, e.g., p C D denotes the linear translation of the D -th robot measured with respect to the C -th robot, particularly, x C = x C C .
x × Skew-symmetric matrix of x , and x × y = x y × [28] holds
p The linear translation vector of rigid bodies along 3 axes, of which the quasi-quaternion description is p ¯ = [ 0 , p T ] T
q ¯ The unit quaternion following the Hamilton notation [21], written as q ¯ = [ q 0 , q 1 , q 2 , q 3 ] T = [ q 0 , q T ] T
q ¯ * The conjugate form of q ¯ , and q ¯ q ¯ = 1 holds
b The uncertain bias of the measurement result
R The equivalent rotation matrix of the quaternion q ¯ , e.g., R w i = R q ¯ w i
n White Gaussian noise vector with zero mean and covariance σ 2
x ˙ , x ^ The first-order time derivative form and the estimated form of the vector x , respectively
x ˜ The error form of the vector x , which is defined as x ˜ = x x ^
δ q ¯ The error of the quaternion q ¯ , and δ q ¯ = q ¯ ^ q ¯ holds
Table 2. The norm errors between the ground truth and the SR-CICEKF simulation results.
Table 2. The norm errors between the ground truth and the SR-CICEKF simulation results.
Translation RMSETranslation Mean ErrorTranslation STDOrientation RMSEOrientation Mean ErrorOrientation STD
0.1593 m0.0429 m0.1535 m0.106 rad0.0187 rad0.1044 rad
Table 3. The norm errors between the ground truth and the filtered slave robot trajectories.
Table 3. The norm errors between the ground truth and the filtered slave robot trajectories.
Translation RMSETranslation Mean ErrorTranslation STDOrientation RMSEOrientation Mean ErrorOrientation STD
0.01872 m5.286 × 10−6 m0.0187 m0.0016 rad4.95 × 10−7 rad0.0016 rad
Table 4. Norm errors of the translations between the ground truth and the results of the dataset tests.
Table 4. Norm errors of the translations between the ground truth and the results of the dataset tests.
Translation RMSETranslation Mean ErrorTranslation STD
SR-CICEKF0.004211 m0.003982 m0.001371 m
Stereo ORB-SLAM V30.03844 m0.03428 m0.01739 m
Stereo ORB-SLAM V3 with the IMU0.003236 m0.002785 m0.001648 m
Stereo MSCKF0.0556 m0.048994 m0.02629 m
Table 5. Norm errors of the translations between the ground truth and the experimental results.
Table 5. Norm errors of the translations between the ground truth and the experimental results.
Translation RMSETranslation Mean ErrorTranslation STD
SR-CICEKF0.00459 m0.006448 m0.004528 m
Stereo ORB-SLAM V30.01723 m0.01275 m0.01159 m
Stereo MSCKF0.01745 m0.01292 m0.01173 m
Table 6. Norm errors of the relative configuration estimation.
Table 6. Norm errors of the relative configuration estimation.
Translation RMSETranslation Mean ErrorTranslation STDOrientation RMSEOrientation Mean ErrorOrientation STD
0.009398 m0.0001 m0.0094 m0.032 rad0.00602 rad0.03145 rad
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, C.; Wang, T.; Li, Z.; Li, S.; Tian, P. A Novel Loosely Coupled Collaborative Localization Method Utilizing Integrated IMU-Aided Cameras for Multiple Autonomous Robots. Sensors 2025, 25, 3086. https://doi.org/10.3390/s25103086

AMA Style

Liu C, Wang T, Li Z, Li S, Tian P. A Novel Loosely Coupled Collaborative Localization Method Utilizing Integrated IMU-Aided Cameras for Multiple Autonomous Robots. Sensors. 2025; 25(10):3086. https://doi.org/10.3390/s25103086

Chicago/Turabian Style

Liu, Cheng, Tao Wang, Zhi Li, Shu Li, and Peng Tian. 2025. "A Novel Loosely Coupled Collaborative Localization Method Utilizing Integrated IMU-Aided Cameras for Multiple Autonomous Robots" Sensors 25, no. 10: 3086. https://doi.org/10.3390/s25103086

APA Style

Liu, C., Wang, T., Li, Z., Li, S., & Tian, P. (2025). A Novel Loosely Coupled Collaborative Localization Method Utilizing Integrated IMU-Aided Cameras for Multiple Autonomous Robots. Sensors, 25(10), 3086. https://doi.org/10.3390/s25103086

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop