Next Article in Journal
Multirobot Heterogeneous Control Considering Secondary Objectives
Previous Article in Journal
L-VTP: Long-Term Vessel Trajectory Prediction Based on Multi-Source Data Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unique 4-DOF Relative Pose Estimation with Six Distances for UWB/V-SLAM-Based Devices

by
Francisco Molina Martel
1,2,*,
Juri Sidorenko
1,2,
Christoph Bodensteiner
1,
Michael Arens
1 and
Urs Hugentobler
2
1
Fraunhofer Institute of Optronics, System Technologies and Image Exploitation IOSB, Gutleuthausstraße 1, 76275 Ettlingen, Germany
2
Institute of Astronomical and Physical Geodesy, Technical University of Munich, Arcisstrasse 21, 80333 Munich, Germany
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(20), 4366; https://doi.org/10.3390/s19204366
Submission received: 26 August 2019 / Revised: 1 October 2019 / Accepted: 4 October 2019 / Published: 9 October 2019
(This article belongs to the Section Sensor Networks)

Abstract

:
In this work we introduce a relative localization method that estimates the coordinate frame transformation between two devices based on distance measurements. We present a linear algorithm that calculates the relative pose in 2D or 3D with four degrees of freedom (4-DOF). This algorithm needs a minimum of five or six distance measurements, respectively, to estimate the relative pose uniquely. We use the linear algorithm in conjunction with outlier detection algorithms and as a good initial estimate for iterative least squares refinement. The proposed method outperforms other related linear methods in terms of distance measurements needed and in terms of accuracy. In comparison with a related linear algorithm in 2D, we can reduce 10% of the translation error. In contrast to the more general 6-DOF linear algorithm, our 4-DOF method reduces the minimum distances needed from ten to six and the rotation error by a factor of four at the standard deviation of our ultra-wideband (UWB) transponders. When using the same amount of measurements the orientation error and translation error are approximately reduced to a factor of ten. We validate our method with simulations and an experimental setup, where we integrate ultra-wideband (UWB) technology into simultaneous localization and mapping (SLAM)-based devices. The presented relative pose estimation method is intended for use in augmented reality applications for cooperative localization with head-mounted displays. We foresee practical use cases of this method in cooperative SLAM, where map merging is performed in the most proactive manner.

1. Introduction

A considerable amount of research and progress has been made in the last few decades in the field of simultaneous localization and mapping (SLAM) [1,2]. SLAM methods concurrently map the surrounding three-dimensional space and estimate the own pose (position and orientation) relative to the constructed map. Augmented reality (AR) has notably benefited from advances in this field.
Cooperative SLAM (C-SLAM) is a related important topic with special relevance in robotics and it adds an additional layer of complexity, because it requires prior knowledge of the map or the relative poses (coordinate frame transformations). The relative poses are calculated either by using common information of local maps or by meeting (rendezvous) and performing relative line-of-sight (LOS) measurements. The need to scan common areas of the environment or to meet, especially if the relative poses are not known, results in a time loss, which is unacceptable in time-critical applications such as emergency rescue operations.
We have integrated an ultra-wideband (UWB) transceiver in a SLAM-based AR test device (Microsoft HoloLens), extending its functionality. The UWB transceiver uses the time of arrival (TOA) technique to measure the distance to other UWB-transceivers. UWB technology is ideal for indoor and GPS-denied applications due to its reduced sensitivity against multipath effects. It further allows the resilient localization of several SLAM-based devices in the absence of a loop closure or under harsh conditions (low lighting, smoke, dust), where cooperative visual SLAM would fail. In this work, we address the relative localization problem of two UWB/SLAM-based devices with distance measurements only.
Our AR test device has four external cameras, an inertial measurement unit (IMU) and a TOF camera. The last two sensors allow metric localization and mapping. The IMU can estimate the gravitation vector with high accuracy [3]. We restrict the six-dimensional relative pose estimation problem to four degrees of freedom (4-DOF) using this information. This decreases the minimum number of distances needed and improves the accuracy of the 3D relative pose estimation. Our method is well suited in conjunction with outlier detection algorithms such as RANSAC, since minimizing the subset sample data considerably reduces their computational complexity.

1.1. Related Work

Early works in C-SLAM assume that the initial poses of the robots are known [4]. More recent publications [5,6,7] use different approaches but they all rely on common map information and/or robot encounters with LOS.
In a review of C-SLAM [8], estimating the relative poses and their respective uncertainty are considered two major problems. The importance of estimating the relative transforms is also emphasized in [9,10] and [11].
Our method can be seen as an extension of rendezvous [12]. Because we use radio frequency (RF), it is possible to use LOS and non-line-of-sight (NLOS) measurements to estimate the relative poses and we estimate it with minimal data communication. In this manner a global map can be created and loop closures can be predicted without overlaps between the local maps.
Several cooperative localization (CL) methods fuse odometry readings with different types of relative measurements using a particle filter [13], a maximum likelihood estimation (MLE) [14] or an extended Kalman filter (EKF) [15] approach in order to estimate the current poses of the robots. The authors in [16] investigate the CL problem of two UAVs equipped with inertial sensors and a camera performing relative bearing measurements only. They apply an EKF to estimate the relative position, orientation and velocity between both UAVs.
The authors in [17,18] and [19] have extensively researched the problem of determining the robot-to-robot relative pose using only range measurements and local positions. They compute an initial estimate of the relative pose transformation with non-iterative algorithms and refine the solution using weighted least squares. They present solutions for the 2D and 3D case.
The work in [18] provides non-iterative algorithms for computing the initial relative pose in 2D. The authors describe a linear algorithm to estimate the solution with five or more distance measurements. They provide experimental results using wheel odometry and adding noise to the ground truth distances measured by an external camera. However, it would be interesting to test these methods under field conditions.
In [19], a linear solution with ten distance measurements is presented for the 3D problem with six degrees of freedom. In [20], the same authors provide algorithms based on polynomial system solving for the minimal problem with six distance measurements. The presented methods can have up to 40 real solutions.
In [21] several two-dimensional localization problems are addressed: localizing three agents with inter-agent range measurements and relative bearing measurements to two anchors, and the problem of relative pose estimation (relative frame determination). From graph rigidity theory they prove for the relative frame determination problem, that in the absence of noise and for generic positions, if four or more distance measurements are available, there is one unique solution. To solve these problems the authors propose a sum of squares (SOS) relaxation approach. They compare their solution with the two-dimensional relative pose determination in [18]. This non-linear method results in a lower error, however the linear method is approximately a factor 1000 faster and is accordingly very suitable for real-time computations.
Prorok and Martinoli [22] use a particle filter to fuse TDOA measurements with relative range and bearing measurements. They use multimodal probability distributions to model both LOS and NLOS UWB distance measurements. An experimental setup with a group of 10 robots is shown. Each robot is equipped with wheel odometry, a relative range and bearing module and an UWB emitter which communicates with four UWB base stations.
The authors in [23] address a 2D cooperative localization problem using an EKF in order to account for the displacement noise. With the assumption of known orientation, they provide a method to localize a device with an embedded IMU using distances to a single UWB beacon. The authors also introduce a relative localization (RL) method based on relative distance measurements between the devices. They use the first described method with a static beacon for the initialization of this RL method. The authors show the results of simulations and an experimental setup based on three unmanned aerial vehicles (UAVs), where one of them is hovering in the air, acting as a base station.
To the best of our knowledge, we have not found any experiments in the literature that integrate RF in visual-SLAM devices to estimate the relative poses without the use of external beacons.

1.2. Contributions

We present a relative localization method based on distance measurements between two UWB/SLAM-based devices and a novel linear algorithm for estimating the unique 4-DOF relative pose with a minimum of six distance measurements. In comparison, the linear algorithm proposed in [19] needs ten distances and a minimal solver obtains up to 40 solutions for the more general 6-DOF case [20].
The presented method can be also applied for the 2D case and presents an alternative linear method to [18], since we reduce 10% of the translation error. It estimates the relative pose uniquely between two devices with a minimum of five distances for the 2D case.
Both of the proposed 2D and 3D linear algorithms support more than the minimum required distance measurements. Using these algorithms in an outlier-free overdetermined system provides very accurate initial estimates for iterative refinement, reducing both the number of iterations and the probability of converging to a local minima.
An experimental analysis with data measured by our UWB/visual (V)-SLAM-based systems validates the effectiveness of our method.

1.3. Outline

Section 2 describes the distance-based relative pose estimation problem and the proposed linear closed form solution in 2D and 3D for the special case of 4-DOF. Section 3 presents simulation results for both cases and compares the error with other related linear methods [18,19]. Section 4 describes the experimental setup and shows the results with real data. Section 5 concludes with an interpretation of the achieved results.

2. Methods

2.1. Problem Formulation and Notation

We define k p i ( j ) as the set of positions, where the positions are indexed with i, measured by device j and expressed in coordinate frame k. Figure 1 illustrates the geometric problem in 2D. Each SLAM-based device will measure a set of positions p i ( 1 ) and p i ( 2 ) . If not stated otherwise, they are expressed in their own coordinate systems: Coordinate frame {1} for device 1 and coordinate frame {2} for device 2. While moving, they perform distance measurements d i between each other. The goal is to find the relative pose between both by estimating the transform that converts the positions measured by the second device p i ( 2 ) into the positions 1 p i ( 2 ) expresed in coordinate frame {1} of the first device. The transform is given by a translation ε and a rotation angle α . Assuming that both coordinate system origins lie at the position of the first distance measurement, we see that the Euclidian norm of the translation ε 2 is equal to this first distance. This fact has been exploited in all mentioned related works. Our method can be decoupled from any related assumptions or constraints. We analyse the effect of fixing the norm of the translation to the first distance measurement in Section 3.
We formulate solutions to the problem in 2D and 3D with 4-DOF using the vector notation in Table 1. In the 2D case, we only use the x- and y-axis. In 3D, we assume that the gravity vector is known, since most mobile devices integrate an inertial measurement unit. R α is the rotation matrix around the vertical axis with angle α .

2.2. Cooperative Localization in 2D

2.2.1. Solution of an Overdetermined System

We generate a system of linear equations following Equation (1) with d i being the measured distances and p i ( 1 ) , p i ( 2 ) the measured positions of each device.
d i 2 = p i ( 1 ) 1 p i ( 2 ) 2 2 = A i X i 2 + B i Y i 2 ,
with
X i = x i cos α y i sin α + u
Y i = x i sin α + y i cos α + v .
Inserting Equations (2) and (3) into Equation (1) leads to the following system of equations: b i = 2 A i u 2 B i v + ( 2 A i x i 2 B i y i ) cos α +
( 2 A i y i 2 B i x i ) sin α + 2 x i L 1 + 2 y i L 2 + L 3 ,
with b i : = d i 2 A i 2 B i 2 x i 2 y i 2 being known. The auxiliary variables L 1 , L 2 and L 3 substitute the non-linear terms.
L 1 : = u cos α + v sin α
L 2 : = v cos α u sin α
L 3 : = u 2 + v 2 .
We formulate Equation (4) in the linear matrix form
M x = b ,
where M is the coefficient matrix and b is the column vector of the terms b i . The solution vector x R 7 is defined as:
x : = u v cos α sin α L 1 L 2 L 3 T .
Note that we do not make any assumptions of the translation length ε 2 while calculating the solution. This allows us to shift the origin of the second coordinate system previous to relative pose calculation. We lay it at the center of mass (mean) of the positions p i ( 2 ) measured by the second device. Given the calculated rotation and the translation to the center of mass of the second coordinate system, we can calculate the translation vector ε . Originating from p 1 ( 1 ) , the translation vector should ideally have the length of the first distance measurement. We optionally constrain the Euclidian norm of the translation with ε 2 = d 1 .

2.2.2. Unique Solution with Five Distance Measurements

We cannot solve Equation (8) directly with five distance measurements since r a n k ( M ) < = 5 < d i m ( R 7 ) . We calculate the nullspace of the augmented matrix H = M b with M and b from Equation (8). The nullspace k e r ( H ) is the solution subspace where the homogeneous Equation (9) holds, where x ^ : = x T 1 T .
H x ^ = 0
Assuming that the rows in H are linearly independent, then r a n k ( H ) = 5 . Due to the rank-nullity theorem we can state that the dimension of the nullspace N u l l i t y ( H ) = d i m ( R 8 ) r a n k ( H ) = 3 . The three vectors r , s and t span the nullspace so that
x ^ = λ 1 r + λ 2 s + λ 3 t .
From Equation (10) and the definition of x ^ it is evident that
λ 1 r 8 + λ 2 s 8 + λ 3 t 8 = 1 .
where r 8 , s 8 and t 8 are the eighth components of the respective vectors. We use Equation (11) to eliminate λ 3 and use Equations (5), (6), (7), (12) and (13) as constraints,
cos 2 α + sin 2 α = 1
L 1 2 + L 2 2 = u 2 + v 2 ,
to calculate the parameters λ 1 , λ 2 and λ 3 using an approach similar to [18]. We substitute the corresponding elements from x ^ in Equation (10) into the constraint equations. We then reformulate these equations in a linear matrix form (8) with the unknown vector
x : = λ 1 λ 2 λ 1 λ 2 λ 1 2 λ 2 2 T ,
in order to determine these parameters uniquely.
With known parameters λ 1 , λ 2 and λ 3 we apply Equation (10) to obtain the solution x ^ . We finally use the first distance measurement as the Euclidian norm of the translation.

2.3. Cooperative Localization in 3D

2.3.1. Solution of an Overdetermined System

We assume that our test device is sufficiently precise for a correct vertical axis alignment with a common gravity vector and a common metric scale of the two different coordinate frames. The coordinate transformation is restricted to 4-DOF with a single elementary rotation with angle α around the vertical axis. The vector of unknowns has now an additional element w for the vertical component of the translation.
x : = u v w cos α sin α L 1 L 2 L 3 T
Following the same procedure as in Equation (1) we arrive at the 3D linear system of equations
b i = β i 1 u + β i 2 v + β i 3 w + β i 4 cos α + β i 5 sin α + β i 6 L 1 + β i 7 L 2 + L 3 ,
where b i : = d i 2 A i 2 B i 2 C i 2 x i 2 y i 2 z i 2 + 2 C i z i and the respective derived coefficients β i j are defined as β i 1 : = 2 A i , β i 2 : = 2 B i , β i 3 : = 2 z i 2 C i , β i 4 : = 2 A i x i 2 B i y i , β i 5 : = 2 A i y i 2 B i x i , β i 6 : = 2 x i and β i 7 : = 2 y i .
The auxiliary variable L 3 is now given by
L 3 : = u 2 + v 2 + w 2 .

2.3.2. Unique Solution with Six Distance Measurements

We use the same approach explained previously in the 2D case and calculate the nullspace of the matrix H . If the rows are linearly independent, r a n k ( H ) = 6 and the dimension of the nullspace stays the same as in Section 2.2.2. The only constraints that change are L 3 with Equation (15) and Equation (11) is replaced by
λ 1 r 9 + λ 2 s 9 + λ 3 t 9 = 1 .

3. Simulation Results

3.1. Cooperative Localization in 2D

We implement a MATLAB simulation in order to evaluate the different implemented methods. In this simulation, two devices move randomly for one to two meters between consecutive positions. At each new position they perform a distance measurement which is affected by Gaussian noise. We assume precise measurements of the local positions given by SLAM. The illustrated results are based on the average of 5000 trials for each setting on the horizontal axis. Figure 2 shows the mean absolute error of the translation for different numbers of distance measurements N, with the standard deviation of our UWB-based distance measurement noise σ d = 2 cm. As expected, the error decays with an increasing number of measurements. Note that for seven distances the error increases given that we solve directly for the solution with Equation (8) for N 7 . We can observe that fixing the Euclidean norm of the translation ε 2 to the first distance measurement d 1 reduces the error of the translation when the number of measured distances is small. For a larger set of distances, the error induced by fixing the translation length outweighs the additional information provided by this constraint.
Applying the method described in Section 2.2.2 with five distances and fixing the translation length to the first distance measurement delivers the same result as in [18]. As our method is decoupled from the first distance measurement, we calculate the solution at the center of mass of the second coordinate system. This reduces approximately 10% of the translation estimate error. In Figure 3, we illustrate the mean absolute errors of the 2D relative pose estimation as a function of the distance measurement noise. In Figure 4, the errors are given as a function of the displacement noise. A two-dimensional Gaussian noise is added to each new movement in between distance measurements. The standard deviation of the distance measurement noise is σ d = 2 cm. We can observe the performance limitation induced by the distance measurement noise, when the displacement noise is small. We remark a similar improvement of the translation error in Figure 3b and in Figure 4b, when taking displacement noise into account.

3.2. Cooperative Localization in 3D

We implement a similar simulation to the previous subsection including the three-dimensional movement. We put our method into perspective with [19], which handles the more general case of relative pose estimation with 6-DOF. This linear method has a total of six unknowns and requires a minimum of ten distances. Our method only takes the yaw rotation under consideration and needs a minimum of six distances. In Figure 5, we compare the mean absolute error of the calculated translation and yaw rotation of both methods as a function of the distance measurement noise. For a fair comparison, we always apply the first distance measurement as the fixed norm of the translation. This improves the accuracy of the algorithms for small sets as shown previously. We observe that our 4-DOF method outperforms the linear 6-DOF algorithm when both using the minimum required distances. Our method shows a reduced rotation error by a factor of four at the measured distance standard deviation σ d = 2 cm of our UWB transponders. The translation error is also reduced when the standard deviation of the distance measurements is kept low. When using ten distances, our method clearly improves accuracy, reducing the orientation error by a factor of 17 and the translation error by a factor of seven at the measured distance standard deviation σ d = 2 cm of our UWB transponders.
Similarly to the previous section, we show the mean absolute error as a function of the displacement noise in Figure 6. At each new position of each device, a Gaussian three-dimensional noise with standard deviation σ p is added. Thus both devices drift away from their ground truth positions. Then again, we use the measured distance standard deviation σ d = 2 cm.
We also show the 3D error dependency of the number of measurements and the effect of fixing the length of the distance in Figure 7 for the sake of completeness.

Tilt Error Effect on the 4-DOF Relative Pose Estimation

Motivated by the limited accuracy that noisy one-dimensional range measurements provide to estimate the six-dimensional relative pose, we use the IMU sensor information in order to reduce the problem to four dimensions. IMU sensors are ubiquitous today and can be used to estimate the gravity vector using long term accelerometer information and the angular position, which is mainly calculated from the gyroscope data. While the gyroscope data suffers from drift over time, the gravity vector can be estimated with a tilt error well below 5° [3]. It is worth noting that the gyroscope drift can be easily mitigated using additional sensors, such as a camera in conjunction with image processing algorithms.
To study the effect of the the gravity vector tilt error on the 4-DOF relative pose estimation performance, we perform a simulation with tilted coordinate frames and compare the magnitude of the error with the related 6-DOF linear method. In Figure 8, we show the mean absolute error of the relative pose estimation as a function of the tilt error of the gravity vector. As the tilt error increases, the performance of the 4-DOF relative pose estimation degrades. The 6-DOF method implicitly estimates the relative tilt between the coordinate frames, thus its performance is barely affected. For precise range measurements and an inaccurate estimate of the gravity vector, it may be preferable to compute the full 6-DOF relative pose. The range precision of UWB however, lies in between centimeter and decimeter level. Figure 8 indicates that a reduction to 4-DOF yields a better relative pose estimation using UWB, even when the gravity vector is estimated with an IMU only.

4. Experimental Results

We use our test devices to perform an experimental validation of our method. Two SLAM-based AR devices integrate each an UWB transceiver (DW1000 IC evaluation board) and are shown in Figure 9. The devices take a random path in all three dimensions inside of our office, which measures 8 m in width, 4.1 m in length and 3.5 m in height. The length of the path taken in this experiment for each device is greater than 40 m. Approximately 4500 distances are measured in LOS conditions while both systems are moving. The local position data of both devices and respective distance measurements are sent to a server where the relative pose is calculated.
We take random samples out of the data measured in this experiment and perform the proposed 4-DOF linear method described in Section 2.3. The randomly selected local position data of each device lies between one and two m apart of the previous picked position, imitating the geometrical conditions of the simulation in Section 3.2. We perform 5000 calculations with random subsets of six and ten distances of the measured data. Unfavorable configurations are detected by checking the rank of M in Equation (8). The ground thruth relative pose transform is estimated using MeshLab’s iterative closest point (ICP) algorithm on both local maps.
We illustrate this process in Figure 10 by showing a random picked subset of six distance measurements from the experimental data and the respective path extract of each device. In Figure 10a, the ground truth path of each device is shown from a top view, where the random picked up positions p i ( 1 ) and 1 p i ( 2 ) and the distance measurements d i are displayed by markers in the paths and black dashed lines, respectively. We show the estimated relative localization with these six distances in Figure 10b.
The resulting mean absolute errors are listed in Table 2. Overall, we observe an improvement in accuracy over the linear 6-DOF method [19] in our experimental validation. When using the same number of distances, the relative pose estimation is improved by approximately a factor of ten.
The server is able to visualize the aligment of local maps. Figure 11 shows an example of map merging performed with distance-based relative pose estimation. Two local maps (green and violet) are correctly overlaid after applying the estimated transform. They show the office where the experiments were performed. One of the devices had previously visited the hallway of our institute building (illustrated in violet colour).

5. Conclusions

We presented a linear method for solving the relative pose between two SLAM-based devices with distance measurements only. The method is applicable to 2D and to 3D problems for the special case of 4-DOF. The 2D algorithm represents an alternate method to the linear solution proposed in [18].
The 3D algorithm calculates the 4-DOF relative pose with only six distances, in contrast to 10 distances in the 6-DOF method presented in [19]. Using six distance measurements with our method results in a more accurate solution as the 6-DOF linear method with 10 distances. Our proposed method has a rotational mean absolute error that is always smaller and the translation error is also smaller if the standard deviation of the distance measurements is kept in the centimeter level or below. This holds true in our application, as the measured standard deviation of our UWB devices using time of arrival is σ d ≈ 2 cm. When using the same amount of distances, our 3D method is clearly superior. The proposed linear method imposes the advantage over minimal 3D methods [20] that it can calculate the unique solution directly with the same amount of distances and avoids the post processing phase needed to discard all inconsistent solutions.
The proposed method decouples dependency of a fixed translation length to the problem and we show improved results for larger sets. Decoupling also allows to estimate the frame transformation to the center of mass of the second coordinate system, improving the results of the translation estimate.
Both 2D and 3D proposed algorithms can be used with any number of measurements larger than the minimal unique solution, providing a very accurate estimate of the relative pose solution for larger outlier-free sets. This estimate lies near the global minimum and we further refine it with an iterative least squares method.
We have implemented simulations of the proposed methods and performed first experiments with two UWB/SLAM-based AR test devices for cooperative localization and mapping.

6. Patents

Patent pending status.

Author Contributions

F.M.M. and J.S. developed the idea of combining V-SLAM and UWB technology to perform relative localization. The co-authors have helped to improve the paper with their suggestions in their areas of expertise and reviewing the first drafts of this publication. Literature review, methodology, implementation of own methods and related methods in the literature, method validation, visualization and writing of this paper was done by the first author F.M.M. Software of the experimental setup in Section 4 and a real-time demonstrator of the methods presented in this work, including Augmented Reality head-mounted display and server visualization, was implemented by F.M.M., J.S. helped collecting the measurement data of Section 4 and working on the UWB side. Mentoring and supervision was performed by C.B. and U.H. Funding acquisition and project administration was done by M.A.

Funding

This research was funded by Fraunhofer Institute of Optronics, System Technologies and Image Exploitation IOSB.

Acknowledgments

Thanks to Dimitri Bulatov from Fraunhofer IOSB for reading the paper and providing wise suggestions to improve the content of this publication. We would also like to thank the reviewers for their valuable remarks.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARAugmented Reality
DOFdegrees of freedom
EKFextended Kalman filter
ICintegrated circuit
LOSline-of-sight
NLOSnon-line-of-sight
MLEmaximum likelihood estimation
RFradio frequency
SLAMSimultaneous Localization and Mapping
C-SLAMcooperative SLAM
V-SLAMvisual SLAM
TOAtime of arrival (TOA)
UWBultra-wideband (UWB)
UAVunmanned aerial vehicle

References

  1. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics (Intelligent Robotics and Autonomous Agents); The MIT Press: Cambridge, MA, USA, 2005; ISBN 0262201623. [Google Scholar]
  2. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, present, and future of simultaneous localization and mapping: Toward the robust-perception age. IEEE Trans. Rob. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  3. LaValle, S.M.; Yershova, A.; Katsev, M.; Antonov, M. Head tracking for the oculus rift. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 187–194. [Google Scholar]
  4. Thrun, S. A probabilistic on-line mapping algorithm for teams of mobile robots. Int. J. Rob. Res. 2001, 20, 335–363. [Google Scholar] [CrossRef]
  5. Kim, B.; Kaess, M.; Fletcher, L.; Leonard, J.; Bachrach, A.; Roy, N.; Teller, S. Multiple relative pose graphs for robust cooperative mapping. In Proceedings of the 2010 IEEE International Conference on Robotics and Automation (ICRA), Anchorage, AK, USA, 3–7 May 2010; pp. 3185–3192. [Google Scholar]
  6. Fox, D.; Ko, J.; Konolige, K.; Limketkai, B.; Schulz, D.; Stewart, B. Distributed multirobot exploration and mapping. Proc. IEEE 2006, 94, 1325–1339. [Google Scholar] [CrossRef]
  7. Howard, A. Multi-robot simultaneous localization and mapping using particle filters. Int. J. Rob. Res. 2006, 25, 1243–1256. [Google Scholar] [CrossRef]
  8. Saeedi, S.; Trentini, M.; Seto, M.; Li, H. Multiplerobot simultaneous localization and mapping: A review. J. Field Rob. 2016, 33, 33–46. [Google Scholar] [CrossRef]
  9. Kuzmin, M. Review, Classification and Comparison of the Existing SLAM Methods for Groups of Robots. In Proceedings of the 22nd Conference of Open Innovations Association (FRUCT), Jyväskylä, Finland, 15–18 May 2018; pp. 115–120. [Google Scholar]
  10. Lee, H.C.; Lee, S.H.; Lee, T.S.; Kim, D.J.; Lee, B.H. A survey of map merging techniques for cooperative-SLAM. In Proceedings of the 9th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Daejeon, Korea, 26–28 November 2012; pp. 285–287. [Google Scholar]
  11. Rone, W.; Ben-Tzvi, P. Mapping, localization and motion planning in mobile multi-robotic systems. Robotica 2013, 31, 1–23. [Google Scholar] [CrossRef]
  12. Zhou, X.S.; Roumeliotis, S.I. Multi-robot slam with unknown initial correspondence: The robot rendezvous case. In Proceedings of the 2006 IEEE/RSJ international conference on intelligent robots and systems (IROS), Beijing, China, 9–15 October 2006; pp. 1785–1792. [Google Scholar]
  13. Rekleitis, I.M.; Dudek, G.; Milios, E.E. Multi-robot cooperative localization: A study of trade-offs between efficiency and accuracy. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, 30 September–4 October 2002; pp. 2690–2695. [Google Scholar]
  14. Howard, A.; Matark, M.J.; Sukhatme, G.S. Localization for mobile robot teams using maximum likelihood estimation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, 30 September–4 October 2002; pp. 434–439. [Google Scholar]
  15. Martinelli, A.; Pont, F.; Siegwart, R. Multi-robot localization using relative observations. In Proceedings of the 2005 IEEE International Conference on Robotics and Automation (ICRA), Barcelona, Spain, 18–22 April 2005; pp. 2797–2802. [Google Scholar]
  16. Martinelli, A.; Renzaglia, A. Cooperative visual-inertial sensor fusion: Fundamental equations. In Proceedings of the 2017 International Symposium on Multi-Robot and Multi-Agent Systems (MRS), Los Angeles, CA, USA, 4–5 December 2017; pp. 24–31. [Google Scholar]
  17. Zhou, X.S.; Roumeliotis, S.I. Determining the robot-to-robot relative pose using range-only measurements. In Proceedings of the 2007 IEEE International Conference on Robotics and Automation (ICRA), Roma, Italy, 10–14 April 2007; pp. 4025–4031. [Google Scholar]
  18. Zhou, X.S.; Roumeliotis, S.I. Robot-to-robot relative pose estimation from range measurements. IEEE Trans. Rob. 2008, 6, 1379–1393. [Google Scholar] [CrossRef]
  19. Trawny, N.; Zhou, X.S.; Zhou, K.X.; Roumeliotis, S.I. 3D relative pose estimation from distance-only measurements. In Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), San Diego, CA, USA, 29 October–2 November 2007; pp. 1071–1078. [Google Scholar]
  20. Trawny, N.; Zhou, X.S.; Roumeliotis, S.I. 3D Relative Pose Estimation from Six Distances. Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.148.4386&rep=rep1&type=pdf (accessed on 7 October 2019).
  21. Shames, I.; Fidan, B.; Anderson, B.D.; Hmam, H. Cooperative self-localization of mobile agents. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 1926–1947. [Google Scholar] [CrossRef]
  22. Prorok, A.; Martinoli, A. Accurate indoor localization with ultra-wideband using spatial models and collaboration. Int. J. Rob. Res. 2014, 33, 547–568. [Google Scholar] [CrossRef]
  23. Guo, K.; Qiu, Z.; Meng, W.; Xie, L.; Teo, R. Ultrawideband based cooperative relative localization algorithm and experiments for multiple unmanned aerial vehicles in gps denied environments. Int. J. Micro Air Veh. 2017, 9, 169–186. [Google Scholar] [CrossRef]
Figure 1. Illustration of the 2D relative pose problem.
Figure 1. Illustration of the 2D relative pose problem.
Sensors 19 04366 g001
Figure 2. Mean absolute error (MAE) of the translation in 2D as a function of the number of distance measurements with standard deviation σ d = 2 cm.
Figure 2. Mean absolute error (MAE) of the translation in 2D as a function of the number of distance measurements with standard deviation σ d = 2 cm.
Sensors 19 04366 g002
Figure 3. Mean absolute error (MAE) as a function of the distance standard deviation σ d in 2D. (a) MAE of the rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Figure 3. Mean absolute error (MAE) as a function of the distance standard deviation σ d in 2D. (a) MAE of the rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Sensors 19 04366 g003
Figure 4. Mean absolute error (MAE) as a function of the standard deviation of the displacement noise σ p and a distance standard deviation σ d = 2 cm. (a) MAE of the rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Figure 4. Mean absolute error (MAE) as a function of the standard deviation of the displacement noise σ p and a distance standard deviation σ d = 2 cm. (a) MAE of the rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Sensors 19 04366 g004
Figure 5. Mean absolute error (MAE) as a function of the distance standard deviation σ d in 3D. (a) MAE of the yaw rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Figure 5. Mean absolute error (MAE) as a function of the distance standard deviation σ d in 3D. (a) MAE of the yaw rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Sensors 19 04366 g005
Figure 6. Mean absolute error (MAE) as a function of the standard deviation of the displacement noise σ p and a distance standard deviation σ d = 2 cm. (a) MAE of the rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Figure 6. Mean absolute error (MAE) as a function of the standard deviation of the displacement noise σ p and a distance standard deviation σ d = 2 cm. (a) MAE of the rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Sensors 19 04366 g006
Figure 7. Mean absolute error (MAE) of the translation in 3D as a function of the number of distance measurements with standard deviation σ d = 2 cm.
Figure 7. Mean absolute error (MAE) of the translation in 3D as a function of the number of distance measurements with standard deviation σ d = 2 cm.
Sensors 19 04366 g007
Figure 8. Mean absolute error (MAE) as a function of the standard deviation of the gravity vector tilt error and with distance standard deviation σ d = 2 cm. (a) MAE of the yaw rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Figure 8. Mean absolute error (MAE) as a function of the standard deviation of the gravity vector tilt error and with distance standard deviation σ d = 2 cm. (a) MAE of the yaw rotation with respect to the ground truth. (b) MAE of the translation with respect to the ground truth.
Sensors 19 04366 g008
Figure 9. Simultaneous localization and mapping (SLAM)-based augmented reality (AR) devices with integrated ultra-wideband (UWB) transceivers.
Figure 9. Simultaneous localization and mapping (SLAM)-based augmented reality (AR) devices with integrated ultra-wideband (UWB) transceivers.
Sensors 19 04366 g009
Figure 10. Several subsets of the measured data are selected in order to validate our method. (a) Ground truth paths of each device and the selected distance measurements. (b) Comparisson of the estimated path of device 2 with the ground truth.
Figure 10. Several subsets of the measured data are selected in order to validate our method. (a) Ground truth paths of each device and the selected distance measurements. (b) Comparisson of the estimated path of device 2 with the ground truth.
Sensors 19 04366 g010
Figure 11. Map merging with UWB/visual (V)-SLAM-based relative pose estimation.
Figure 11. Map merging with UWB/visual (V)-SLAM-based relative pose estimation.
Sensors 19 04366 g011
Table 1. Vector notation.
Table 1. Vector notation.
Vector/Pointsx-Axisy-Axisz-Axis
p i ( 1 ) A i B i C i
p i ( 2 ) x i y i z i
1 p i ( 2 ) = R α p i ( 2 ) + ε X i Y i Z i
ε uvw
Table 2. Experimental mean absolute errors.
Table 2. Experimental mean absolute errors.
MethodTranslation MAERotation MAE
Trawny et al, 10 distances3.42 m27.2°
Own method, 6 distances1.0 m15.94°
Own method, 10 distances0.26 m2.68°

Share and Cite

MDPI and ACS Style

Molina Martel, F.; Sidorenko, J.; Bodensteiner, C.; Arens, M.; Hugentobler, U. Unique 4-DOF Relative Pose Estimation with Six Distances for UWB/V-SLAM-Based Devices. Sensors 2019, 19, 4366. https://doi.org/10.3390/s19204366

AMA Style

Molina Martel F, Sidorenko J, Bodensteiner C, Arens M, Hugentobler U. Unique 4-DOF Relative Pose Estimation with Six Distances for UWB/V-SLAM-Based Devices. Sensors. 2019; 19(20):4366. https://doi.org/10.3390/s19204366

Chicago/Turabian Style

Molina Martel, Francisco, Juri Sidorenko, Christoph Bodensteiner, Michael Arens, and Urs Hugentobler. 2019. "Unique 4-DOF Relative Pose Estimation with Six Distances for UWB/V-SLAM-Based Devices" Sensors 19, no. 20: 4366. https://doi.org/10.3390/s19204366

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop