Next Article in Journal
Tsukamoto Fuzzy Logic Controller for Motion Control Applications: Assessment of Energy Performance
Next Article in Special Issue
Web System for Solving the Inverse Kinematics of 6DoF Robotic Arm Using Deep Learning Models: CNN and LSTM
Previous Article in Journal
The Remote Sensing Data Transmission Problem in Communication Constellations: Shop Scheduling-Based Model and Algorithm
Previous Article in Special Issue
A Multi-Scale Cross-Layer Fusion Method for Robotic Grasping Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Localization Meets Uncertainty: Uncertainty-Aware Multi-Modal Localization

by
Hye-Min Won
1,†,
Jieun Lee
2,† and
Jiyong Oh
1,*
1
Daegu-Gyeongbuk Research Division, Electronics and Telecommunications Research Institute, Daegu 42994, Republic of Korea
2
Polaris3D, Pohang 37684, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Technologies 2025, 13(9), 386; https://doi.org/10.3390/technologies13090386
Submission received: 25 June 2025 / Revised: 31 July 2025 / Accepted: 13 August 2025 / Published: 1 September 2025
(This article belongs to the Special Issue AI Robotics Technologies and Their Applications)

Abstract

Reliable localization is critical for robot navigation in complex indoor environments. In this paper, we propose an uncertainty-aware localization method that enhances the reliability of localization outputs without modifying the prediction model itself. This study introduces a percentile-based rejection strategy that filters out unreliable 3-degree-of-freedom pose predictions based on aleatoric and epistemic uncertainties the network estimates. We apply this approach to a multi-modal end-to-end localization that fuses RGB images and 2D LiDAR data, and we evaluate it across three real-world datasets collected using a commercialized serving robot. Experimental results show that applying stricter uncertainty thresholds consistently improves pose accuracy. Specifically, the mean position error, calculated as the average Euclidean distance between the predicted and ground-truth (x, y) coordinates, is reduced by 41.0%, 56.7%, and 69.4%, and the mean orientation error, representing the average angular deviation between the predicted and ground-truth yaw angles, is reduced by 55.6%, 65.7%, and 73.3%, when percentile thresholds of 90%, 80%, and 70% are applied, respectively. Furthermore, the rejection strategy effectively removes extreme outliers, resulting in better alignment with ground truth trajectories. To the best of our knowledge, this is the first study to quantitatively demonstrate the benefits of percentile-based uncertainty rejection in multi-modal and end-to-end localization tasks. Our approach provides a practical means to enhance the reliability and accuracy of localization systems in real-world deployments.

1. Introduction

Localization is a classical problem in the robotics literature. With simultaneous localization and mapping (SLAM) [1,2] and its related technologies [3,4], localization techniques have been constantly developed, and the advances have led to some commercial services using mobile robots and self-driving cars. In particular, service robots have been widely used in indoor environments such as hotels, hospitals, and restaurants for tasks like food delivery or guest assistance in recent years. In such dynamic environments, robots frequently encounter localization challenges such as occlusions, dynamic obstacles, or map drift. These issues can lead to localization failure, causing the robot to become lost or deliver items to incorrect locations. One of the most well-known failure scenarios is the kidnapped robot problem, where a robot is unexpectedly displaced without any sensor trace, making recovery difficult for conventional SLAM-based approaches. To ensure robust and uninterrupted operation, especially after initialization or recovery from failure, global localization—where a robot must estimate its pose from scratch without prior information—plays a critical role in these scenarios. End-to-end localization based on a deep neural network can be a promising solution to global localization.
PoseNet [5] is the first study on the deep learning-based end-to-end localization method. It estimates a six-dimensional pose directly from sensor data (image) using neural networks in an end-to-end manner. Since PoseNet, the following studies have been conducted based on images [6,7,8], 3D point clouds [9,10,11,12], inertial information [13], and the fusion of images and 2D point clouds [14]. These end-to-end localization methods are known to be more robust against sensor data variations such as noise and illumination. However, it is difficult to use them solely because they generally provide relatively higher localization errors compared to the matching-based localization methods. Meanwhile, Jo and Kim [15] utilized PoseNet to obtain an initial pose for a particle filter-based localization. However, they overlooked how much confidence we can have in the output of PoseNet. If a localization result with a large error is served as the initial pose, the navigation system may fail to operate properly. In detail, due to factors such as complex environments, sensor noise, or data inconsistency, the estimated pose can suddenly jump or become unstable. Such issues may cause the localization failure, causing a serious risk to safe robot operation.
In this paper, we propose an uncertainty-aware multi-modal end-to-end network that can quantitatively estimate both the pose and its associated uncertainty, along with an uncertainty-based percentile rejection mechanism. The proposed network is built on the FusionLoc architecture, extending it to estimate both aleatoric and epistemic uncertainty. Specifically, the regression branch is augmented to predict the variance representing aleatoric uncertainty, while Monte Carlo (MC) dropout [16] is employed to approximate the model posteriors for estimating epistemic uncertainty. Furthermore, going beyond the findings of Li et al. [12], which demonstrated a strong correlation between the measured uncertainty and localization error, we apply a percentile rejection method based on the predicted uncertainty. This approach effectively reduces localization errors and enhances the reliability of localization results by filtering out unreliable localization results with large errors. As a result, it can improve safety and practical utility in safety-critical scenarios such as adaptive Monte Carlo localization (AMCL) initialization or recovery from localization failures. Experiments using our datasets collected by a 2D LiDAR and a camera demonstrate that the uncertainty-based rejection effectively reduces position and orientation errors. More specifically, our uncertainty-based rejection method reduces the position error by up to 69.4% and the orientation error by 73.3%. In particular, experimental results show that our strategy can exclude the outputs with significant position and orientation errors. This means that the passed outputs are reliable enough for the results of global localization. To the best of our knowledge, this is the first study that systematically utilizes uncertainty-based thresholds to reject unreliable localization results predicted in an end-to-end manner and improves the accuracy and confidence of pose estimates. The main contributions of this paper are as follows:
  • A novel uncertainty-aware multi-modal end-to-end localization framework that not only estimates the robot’s 3-DoF pose but also quantifies uncertainty associated with the predictions. The original FusionLoc architecture is extended to output uncertainty values by integrating MC dropout for epistemic uncertainty and direct variance regression for aleatoric uncertainty, enabling a more reliable and informative localization result.
  • We introduce a percentile-based rejection method using the thresholds (e.g., 70%, 80%, and 90%) to reject low-confidence pose estimations, thereby improving localization robustness and safety in uncertain environments.
  • The proposed approach is validated on three indoor datasets under various environmental conditions, demonstrating the effectiveness of uncertainty-aware localization over standard approaches.
The remainder of this paper is organized as follows. Section 2 reviews related works on end-to-end localization and uncertainty quantification. Section 3 describes the proposed method, including the original FusionLoc architecture and the extensions introduced for measuring uncertainty. Section 4 presents the datasets used and the experimental results. Finally, Section 5 concludes the paper and provides future work.

2. Related Works

2.1. End-to-End Localization

End-to-end localization, also known as absolute pose regression, directly predicts the pose of a robot or a sensor from its data using deep neural networks, without conventional procedures such as feature detection and matching. It can serve as a solution for global localization, particularly in environments where GPS is unavailable. Depending on the type of sensors used for localization, it can be categorized into camera, LiDAR, and multi-modal localizations.
Visual localization estimates the current pose using only images captured by a camera in indoor or outdoor environments. Initially, convolutional neural networks (CNNs) are primarily leveraged to extract salient features. However, recent studies have introduced various techniques in their models. Kendall et al. [5] developed a CNN-based 6-DoF pose regression model that allows localization without relying on feature matching or keyframes. Meanwhile, Wang et al. [6] employed the self-attention technique [17] to improve the accuracy of the end-to-end localization. Moreover, Transformer architectures [17] have also been utilized for end-to-end camera localization [7,8]. Also, Wang et al. [18] integrated CNNs, self-attention, and long short-term memory (LSTM) modules in a unified architecture to extract static features, which can lead to more effective 6-DoF pose estimation compared to using dynamic features. Recently, Zheng et al. [19] introduced a more advanced algorithm than the end-to-end localization methods, called scene-agnostic pose regression. It is noteworthy that it does not require the collection of new datasets or retraining for deployment in previously unseen environments.
LiDAR-based end-to-end localization leverages 3D structural information, making it robust in textureless environments. Wang et al. [9] introduced the first LiDAR-based 6-DoF pose regression model, enhancing feature learning with the self-attention mechanism. Yu et al. [10] proposed a deep neural network for pose regression that consists of two modules: a universal encoder for scene feature extraction and a regressor for pose estimation. They also demonstrated the relationship between the regression capability and the number of hidden units in the regression module. Yu et al. [20] introduced additional classification headers alongside the original regression headers, together with a feature aggregation module based on temporal attention for spatial and temporal constraints. Ibrahim et al. [21] presented a self-supervised learning approach utilizing a Transformer-based backbone for LiDAR-based end-to-end localization. Also, SGLoc [11] enhanced pose regression accuracy by incorporating scene geometry encoding. Recently, Li et al. [12] proposed measuring uncertainty from the outputs of their proposed method, DiffLoc, a diffusion model for localization to estimate the 6-DoF pose of a 3D LiDAR sensor from a point cloud. Their approach further improves accuracy by applying an iterative denoising process based on a diffusion model to the pose regression. Yu et al. [22] proposed a 3D LiDAR localization method inspired by neurobiological mechanisms of three different cells in mammalian brains. Yang et al. [23] utilized semantic awareness to mitigate negative effects caused by dynamic objects and repeating structures so that they improved the localization accuracy and robustness.
Some studies have combined complementary information from multiple modalities to improve the robustness of localization. For instance, Lai et al. [24] proposed leveraging both visual and LiDAR features to achieve more accurate and robust place recognition. Wang et al. [25] developed a vision-assisted LiDAR localization method that effectively utilizes visual information to address issues related to 2D LiDAR-based localization drift. Additionally, Nakamura et al. [26] incorporated a fisheye camera together with a 2D LiDAR system to enhance localization fault detection. However, the methods mentioned above do not fall under the category of end-to-end localization techniques. FusionLoc [14] is an end-to-end localization method that utilizes multi-modality. It was demonstrated that taking both RGB images and 2D LiDAR point clouds as input could provide superior performance compared to a mono-modal input, either image or 2D LiDAR point cloud, by effectively leveraging complementary information from each modality. Table 1 provides a brief comparison of the previous studies mentioned above. In this study, we present a FusionLoc-based approach to make localization more reliable by rejecting network outputs with significant errors.

2.2. Uncertainty Quantification

Uncertainty quantification is a well-established topic in pattern recognition and machine learning. While it did not receive much attention during the early stages of the deep learning revolution—especially in comparison to efforts to enhance the accuracy of deep learning algorithms—its importance is becoming increasingly recognized, particularly in safety-critical applications. A Bayesian approach is one of the most comprehensive frameworks for managing uncertainty. However, developing and implementing a Bayesian deep neural network for regression tasks is very challenging because it is often impractical to determine posterior probabilities accurately. Fortunately, Gal and Ghahramani [16] presented dropout as an alternative to Bayesian approximation. Additionally, Kendall and Gal [27] proposed using the MC dropout to quantify both aleatoric and epistemic uncertainties in regression tasks, such as pixel-wise depth estimation.
After the groundbreaking studies, the following researchers focused on network calibration, which aims to align estimated uncertainty values with empirical results. Kuleshov et al. [28] introduced a simple, algorithm-agnostic method inspired by Platt scaling [29]. Cui et al. [30] utilized the maximum mean discrepancy, viewing calibration as a form of distribution matching. Similarly, Bhatt et al. [31] applied the f-divergence with the same perspective. In another approach, Yu et al. [32] proposed an auxiliary network branch to estimate uncertainty alongside the main branch used for the original regression task. This method is similar to the work of [33], which also employs additional network branches to estimate uncertainty or confidence in classification problems. For more details on uncertainty quantification in deep neural networks, refer to [34,35].
However, none of the studies mentioned above addressed the localization problem. Chen et al. [36] recently proposed a method for quantifying uncertainty in visual localization. However, their approach estimates the pose of a query image through keypoint matching rather than in an end-to-end manner. Unfortunately, the matching-based approaches are known to be more sensitive to variations such as illumination change and dynamic objects compared to the end-to-end approaches of the proposed method. In contrast, Li et al. [12] suggested an end-to-end localization approach using a diffusion model. However, their work primarily focused on the relationship between quantified uncertainties (variance) and positional errors, lacking qualitative experimental results. Unlike [12,36], this study aims to quantify uncertainty in the results of an end-to-end localization method. We will also demonstrate how this quantification can effectively reject network outputs with significant errors, ultimately improving the localization performance.
Table 1. Summary of related works on end-to-end localization.
Table 1. Summary of related works on end-to-end localization.
ReferenceSensor ModalityLocalizationUncertainty
Kendall et al. [5]RGB Camera
Wang et al. [6]RGB Camera
Li and Ling [7]Multi-view Camera
Qiao et al. [8]RGB Camera
Wang et al. [18]RGB Camera
Wang et al. [9]3D LiDAR
Yu et al. [10]3D LiDAR
Yu et al. [20]3D LiDAR
Ibrahim et al. [21]3D LiDAR
Li et al. [11]3D LiDAR
Lai et al. [24]RGB Camera + 3D LiDAR
Wang et al. [25]RGB Camera + 2D LiDAR
Nakamura et al. [26]Fisheye RGB Camera + 2D LiDAR
Lee et al. [14]RGB Camera + 2D LiDAR
Chen et al. [36]RGB Camera
Li et al. [12]3D LiDAR
OursRGB Camera + 2D LiDAR

3. Method

3.1. Measuring Uncertainty

Deep learning has shown remarkable performance on various complex tasks, primarily focused on enhancing predictive accuracy. However, real-world applications often face uncertainty due to some factors, such as incomplete information and ambiguities. This complexity makes it difficult to assess the performance of the model solely on the basis of accuracy [30]. Therefore, quantifying uncertainty is crucial to improve prediction reliability, improve model robustness, and ensure safety.
Uncertainty can be divided into two categories: epistemic uncertainty and aleatoric uncertainty [27]. Epistemic uncertainty arises from limitations in the model’s knowledge or training process, typically due to insufficient data. This uncertainty can be reduced by incorporating additional training data or enhancing the model architecture. In contrast, aleatoric uncertainty stems from sensor noise, measurement errors, or inherent randomness in the data collection procedure. Unlike epistemic uncertainty, aleatoric uncertainty cannot be eliminated through additional training, as it originates from data sensing. Both types of uncertainty can be estimated using Bayesian neural networks (BNNs). BNNs treat model weights as probabilistic distributions to quantify uncertainty. However, computing the exact posterior distribution in high-dimensional spaces is almost impractical. In this study, we use MC dropout [27] to approximate Bayesian inference and provide uncertainty estimation.
Let us consider a regression task with N data pairs of input x and output y, i.e., { x i , y i } i = 1 N . To quantify aleatoric and epistemic uncertainties in this task, we consider a BNN model f to infer the posterior distribution. From an input x , it provides a model output y ^ together with a variance σ ^ 2 of the aleatoric uncertainty. In contrast, to estimate the epistemic uncertainty, we employ the MC dropout to approximate the posterior over the model. By representing the model weights as W ^ from the approximate posterior, the model provides both the predictive mean and the variance, i.e., [ y ^ , σ ^ 2 ] = f W ^ ( x ) . The objective function of learning the model can be defined without the regularization term as follows:
1 N i = 1 N 1 2 σ ^ i 2 y i y ^ i 2 + 1 2 log σ ^ i 2 = 1 2 N i = 1 N exp ( s i ) y i y ^ i 2 + s i ,
where s i = log σ ^ i 2 . Here, the second equation is more numerically stable for the case of division by zero. After training, we can estimate the uncertainty of an output y ^ by T times multiple inferences for a given input as follows:
1 T t = 1 T y ^ t 2 1 T t = 1 T y ^ t 2 + 1 T t = 1 T σ ^ t 2 ,
where ( y ^ t , σ ^ t 2 ) is the t-th output of the model based on randomly determined weights by dropout. Note that the first and the second terms correspond to the epistemic and the aleatoric uncertainties of the output, respectively. More details are referred to as [27].

3.2. Uncertainty-Aware Localization Based on FusionLoc

In this section, we describe our uncertainty-aware localization method based on the fusion of RGB image and 2D LiDAR data captured from a commercialized serving robot. As shown in Figure 1, our approach builds upon FusionLoc [14], a deep learning-based framework that estimates a robot’s 3-DoF pose from RGB image and 2D LiDAR data. The proposed method extends FusionLoc by quantifying localization uncertainty and applying a percentile rejection method to improve safety and reliability in scenarios with large localization errors.
To better understand the underlying mechanism, we first describe the FusionLoc network architecture and its pose estimation process. FusionLoc takes the input data, consisting of an RGB image I and 2D LiDAR data S , as shown in the input data part of Figure 1, and predicts the robot’s position p ^ and orientation q ^ , which are the outputs of the original FusionLoc, as follows:
[ p ^ , q ^ ] = f ( I , S ) ,
where p ^ = [ x ^ , y ^ ] is the 2D coordinates of the robot position, and q ^ = [ cos θ ^ , sin θ ^ ] corresponds to the robot orientation. The method computes an image feature from the input image using a feature extractor from AtLoc [6], while it calculates point features from the input LiDAR data using a different feature extractor from PointLoc [9]. To enhance the interaction between these two modalities, multi-head self-attention [17] is employed. This approach enables more effective multi-modality fusion than traditional methods such as concatenation or addition of the image and point features. Lastly, the output of the multi-head self-attention block is passed through the MLP layers, which consist of two branches in the regression part of Figure 1, each responsible for estimating the position and orientation. These branches predict the position p ^ and the orientation q ^ .
Although the predicted pose obtained through this process generally reduces localization errors compared to mono-modal approaches, significant errors still occasionally occur. These large errors can cause the robot to navigate to incorrect locations, thereby compromising its operational safety and reliability. To address this issue, we assess the reliability of each estimated pose by quantifying the uncertainty of the estimated pose. To perform the uncertainty-aware localization, we modify the FusionLoc [14] architecture such that it has two more output nodes σ ^ p 2 and σ ^ q 2 to measure the aleatoric uncertainty. Thus, our model can be represented as the following:
[ p ^ , σ ^ p 2 , q ^ , σ ^ q 2 ] = f W ^ ( I , S ) .
As mentioned above, we replace σ ^ p 2 and σ ^ q 2 by s p = log σ ^ p 2 and s q = log σ ^ q 2 for computational stability. Given N training samples { ( I i , S i , p i , q i ) } i = 1 N , the loss function can be defined as the following:
1 2 N i = 1 N exp ( s p i ) p i p ^ i 2 + s p i + 1 2 N i = 1 N exp ( s q i ) q i q ^ i 2 + s q i .
Another difference between the proposed method and the original FusionLoc is the dropout, which is applied to the output of the self-attention block in the original FusionLoc, as shown in Figure 1. It is important to note that the dropout is applied only in inference to measure the epistemic uncertainty. After finishing training process, we can predict the position p * and orientation q * of the robot by performing the inference T times using a pair of ( I , S ) as the following:
p * = 1 T t = 1 T p ^ t , q * = 1 T t = 1 T q ^ t ,
where p ^ t and q ^ t are the t-th position and orientation outputs obtained using dropout in the trained model. Due to the dropout, the network can output different values of ( p ^ t , s p , q ^ t , s q ) even if the input is the same for all t. And, their corresponding uncertainties u p and u q are computed as follows:
u p = 1 T t = 1 T p ^ t 2 ( p * ) 2 + 1 T t = 1 T σ ^ p t 2 , u q = 1 T t = 1 T q ^ t 2 ( q * ) 2 + 1 T t = 1 T σ ^ q t 2 ,
where the first and the second terms of u p and u q correspond to the epistemic and the aleatoric uncertainties of the output, respectively, as mentioned in the previous subsection.
Note that we cannot predict how much error the network output has, which may often lead to a serious problem in safety. Under the assumption that a network output with high uncertainty has a large localization error, we can select reliable outputs based on the uncertainty values computed as (2) as follows:
u p < η p and u q < η q ,
where η p and η q are the rejection thresholds for the position and orientation, respectively. If one of u p and u q is greater than the corresponding threshold, the output ( p * , q * ) is rejected. By discarding results with high uncertainty, we ensure that the remaining outputs have comparatively lower errors.
Our approach utilizes a percentile-based thresholding method that rejects a portion of the network outputs that exceed predefined uncertainty thresholds. In this situation, it is crucial to determine the appropriate threshold value. We experiment with different percentile thresholds (100%, 90%, 80%, and 70%), progressively filtering the top 0%, 10%, 20%, and 30% of the most uncertain predictions. Although this thresholding approach is straightforward, it effectively rejects the network results corresponding to outliers, leading to a more reliable and precise localization system.
In the next section, we will present a detailed analysis of how the uncertainty-based rejection strategy improves localization performance.

4. Experiments

4.1. Datasets

For our experiments, we constructed three datasets from indoor environments named TheGardenParty, ETRI, and SusungHotel. For multi-modality, sensor data such as RGB images and 2D LiDAR scans were collected using a commercialized serving robot, Polaris3D Ereon (Polaris3D, Pohang, Republic of Korea), as shown in Figure 2. This robot is equipped with two cameras and a 2D LiDAR sensor. For our purposes, we utilized a lower-mounted camera to capture RGB images and the LiDAR sensor to gather 2D range data. We utilized an Intel RealSense D435 camera (Intel Corporation, Santa Clara, CA, USA) for collecting the TheGardenParty dataset and an Astra Stereo SU3 camera (Orbbec 3D, Troy, MI, USA) for the ETRI and SusungHotel datasets. The LiDAR sensor is the SLAMTEC RPLiDAR A1M8 (Shanghai SLAMTEC Co., Ltd., Shanghai, China), which operates at 8 Hz with a maximum range of 12 m and an angular resolution of 0.313 ° . It performs 360 ° scans, generating up to 1150 2D points per scan. To effectively train and evaluate relocalization algorithms, it is essential to obtain robot poses aligned with sensor data such as images and point clouds. For this purpose, we first generated a map of our test environment using a 2D LiDAR-based SLAM approach. Then, we acquired the robot poses by running the SLAM system in localization mode. The collected datasets consist of RGB images that provide visual context, 2D LiDAR scans that capture structural and geometric information, and 3-DoF poses derived from these scans. We aimed to synchronize the images, 2D LiDAR data, and poses as closely in time as possible. These datasets were utilized for training and evaluating deep learning models focused on robot localization.
Figure 3 presents ground truth trajectories collected from three different datasets. In the figure, each color represents a different sequence, while the robot’s start and end positions are marked with the gold and downward stars, respectively. Table 2 compares the key characteristics of the three datasets and summarizes their features, including the number of samples and sequences used for training, evaluation, and validation. The TheGardenParty dataset provides images at a resolution of 320 × 240 pixels, and it was collected in a restaurant environment with predefined paths. This dataset consists of fragmented and partial sequences rather than full traversals of the environment. The dataset comprises 13,326 data tuples across 35 sequences, with 24 sequences used for training, 6 for evaluation, and 5 for validation. The ETRI dataset supports 640 × 480 pixels and it represents a more complex navigation environment, featuring a wide lobby space with three pillars where the robot navigates along several paths, including exploration and obstacle avoidance. Each sequence of this dataset covers the entire environment and was collected in a well-structured setting with minimal obstacles. As shown in Figure 4, the ETRI dataset features four distinct movement patterns:
  • Straight corridor navigation: In this pattern, the robot navigates the entire space in a continuous loop before returning to its starting point.
  • Zigzag movement: Here, the robot moves in a zigzag manner, weaving between obstacles.
  • Repetitive back-and-forth motion: This pattern involves the robot moving back and forth within a confined space before proceeding with further exploration.
  • Rotational maneuvers: In this last pattern, the robot performs in-place rotations at specific locations before retracing the same trajectory as in the first pattern.
The dataset comprises 19,014 tuples and 100 sequences, with 66 sequences designated for training, 16 for evaluation, and 16 for validation. The SusungHotel dataset also provides high-resolution images with a resolution of 640 × 480 pixels. It consists of a total of 9625 data tuples, collected from 20 distinct sequences. These sequences are categorized into 16 for training, 2 for validation, and 2 for evaluation. Notably, among the three datasets, the SusungHotel dataset features the longest continuous trajectories captured in a single recording session. This dataset contains sequences that cover the entire environment. However, the environment itself is narrow and takes the form of a long corridor. Furthermore, it contains some variations such as unpredictable dynamic obstacles, lighting conditions, and random perturbations to the robot’s movements.
Figure 3. Robot trajectories in each dataset. Each color represents a distinct sequence of data.
Figure 3. Robot trajectories in each dataset. Each color represents a distinct sequence of data.
Technologies 13 00386 g003
Table 2. Summary of each dataset’s characteristics.
Table 2. Summary of each dataset’s characteristics.
AttributeTheGardenPartyETRISusungHotel
Image resolution (pixels)320 × 240640 × 480640 × 480
Robot movement typePredefined4 patternsPredefined + random perturbation
Environment typeRestaurantWide lobbyLong and narrow curved corridor
Sequence typeFragmentFullFull
# Training tuples (# Sequences)7848 (24)12,688 (66)7625 (16)
# Validation tuples (# Sequences)2294 (5)2964 (16)1258 (2)
# Test tuples (# Sequences)3184 (6)2794 (16)1276 (2)
Total tuples (# Sequences)13,326 (35)19,014 (100)9625 (20)
Figure 4. Visualization of robot trajectories in different scenarios on the ETRI dataset. (a) Full-loop trajectory. (b) Zigzag navigation. (c) Localized back-and-forth motion. (d) In-place rotations at specific locations.
Figure 4. Visualization of robot trajectories in different scenarios on the ETRI dataset. (a) Full-loop trajectory. (b) Zigzag navigation. (c) Localized back-and-forth motion. (d) In-place rotations at specific locations.
Technologies 13 00386 g004

4.2. Training and Inference Details

We implemented, trained, and inferred the network on a single NVIDIA GeForce RTX 4090 GPU (NVIDIA Corporation, Santa Clara, CA, USA) environment under the settings below. The input data consist of RGB images, 2D point clouds, and ground-truth trajectory information. The RGB images were resized to a resolution of 320 × 240 and normalized using the mean and standard deviation computed during training. The point cloud was randomly sampled to a fixed number of points (e.g., 850), with each point represented by 2D coordinates ( x , y ) . The position vectors in the trajectory information were normalized using the mean and standard deviation of the training data and then fed into the model. The predicted outputs were subsequently un-normalized to recover real-world coordinates in meters. To keep the continuity of the rotational angle, the orientation was regressed in the form of [ cos θ ^ , sin θ ^ ] T instead of θ ^ . Finally, the predicted yaw angle was reconstructed using the equation θ ^ = tan 1 ( sin θ ^ , cos θ ^ ) .
The model was optimized using the Adam optimizer, and we set both the learning rate and weight decay to 0.0001. Training was conducted with a batch size of 128 for up to 1000 epochs. Model performance was evaluated at each epoch using a validation set, and only the model corresponding to the lowest validation loss was retained. Additionally, a dropout probability of 0.5 was applied during training to prevent overfitting. To enable the model to estimate predictive confidence alongside the output, we incorporated the aleatoric uncertainty into the regression loss function as in (1). In contrast, the epistemic uncertainty based on a BNN framework was not utilized during training. Instead, it was estimated at inference time by performing multiple forward passes using MC dropout and computing the variance of the predictions. Furthermore, the training loss function incorporates an L 2 loss weighted by aleatoric uncertainty through learnable log-variance terms. This approach allows the model to dynamically adjust the contribution of the position and orientation errors by scaling the loss according to the predicted uncertainty. As a result, the impact of high-error samples is mitigated, and the model is trained to jointly estimate both the target values and their associated confidence levels.
In this study, to evaluate the generalization performance of the proposed method, we designed a single-sequence inference procedure incorporating an uncertainty estimation mechanism based on MC dropout. During inference, the trained model parameters were loaded, and both an RGB image and a 2D LiDAR point cloud were provided as inputs to jointly predict the position and orientation. We performed inference applying the MC dropout technique using a dropout probability of 0.2. Multiple forward passes were conducted to compute the mean prediction and variance. In this context, the epistemic uncertainty is interpreted as the prediction variance estimated via the MC dropout, while the aleatoric uncertainty is directly obtained from the learned log-variance values. Both types of uncertainty were output alongside the predicted pose and were used to quantitatively assess the confidence level of the prediction results. For all inference samples, the epistemic uncertainty and aleatoric uncertainty were collected separately for both position and orientation. These two uncertainty measures were then combined to form a unified uncertainty score, which serves as a criterion for rejection. Specifically, for each sample, the epistemic and aleatoric uncertainties were summed as in (2) to compute a total uncertainty score for both position and orientation. Based on this score, only the samples falling below a specified percentile threshold are retained as in (3), allowing the evaluation to focus on high-confidence predictions.

4.3. Evaluation

In this subsection, we evaluate the performance of our localization method by applying the uncertainty-based rejection approach. To achieve this, we utilized the model mentioned in Section 3 and measured the epistemic and aleatoric uncertainties in the model’s predictions for position and orientation. We demonstrate the improvement in the reliability of position and orientation predictions by applying percentile-based thresholds using uncertainty values and discarding outputs that exceed these thresholds. Specifically, experiments were conducted using 100%, 90%, 80%, and 70% as thresholds, progressively rejecting the top 0%, 10%, 20%, and 30% of the outputs with the highest uncertainty. This rejection method retains only the reliable results, minimizing the influence of extreme outliers with high uncertainty. As a result, this approach finally leads to a more robust evaluation of the model’s performance.
To evaluate the impact of the rejection approach on localization performance, we compared the median and mean errors in position and orientation before and after applying the rejection based on uncertainty thresholding. Table 3 illustrates the changes in position and orientation errors under varying uncertainty thresholds applied to the TheGardenParty, ETRI, and SusungHotel datasets. The results demonstrate consistent reductions in mean position and orientation errors across all rejection thresholds (90%, 80%, and 70%). In the TheGardenParty dataset, applying a 70% uncertainty threshold led to a reduction in the mean position error by as much as 12.1% (from 0.776 m to 0.682 m) and a decrease in the mean orientation error by up to 36.0% (from 4.880 ° to 3.126 ° ). For the ETRI dataset under the same conditions, the position error slightly increased by approximately 0.9% (from 0.324 m to 0.327 m), while the orientation error decreased by 29.2% (from 1.990° to 1.409°). Notably, the effectiveness of the uncertainty-based rejection strategy was also observed in the SusungHotel datasets. For the dataset, the mean position error decreased from 4.262 m to 1.575 m, while the mean orientation error was reduced from 13.711 ° to 4.329 ° , reflecting improvements of approximately 63.0% and 68.4%, respectively. These findings indicate that the TheGardenParty and SusungHotel datasets have more outliers and noise, which leads to more significant performance improvements with the uncertainty-based rejection strategy. In contrast, the ETRI dataset, likely collected in a relatively simple environment, shows lower errors even without applying the rejection method, resulting in relatively less impact from this approach. The differences in performance improvement across the datasets may be attributed to their spatial characteristics. The TheGardenParty dataset was collected from a restaurant with many tables and chairs, while the SusungHotel dataset was gathered from a long and curved corridor, which presents challenges for robot localization. In contrast, the environment in which the ETRI dataset was collected is relatively straightforward and typical. Additionally, the size of the datasets may have played a role in the observed performance differences. The ETRI dataset contains about 1.6 times more training data than the other datasets. It is also noticeable that, in the TheGardenParty dataset, the maximum position error decreased from 5.559 m to 4.192 m, and the maximum orientation error significantly dropped from 72.738 ° to 48.397 ° . This demonstrates the effectiveness of the rejection method in addressing extreme localization errors. These results indicate our end-to-end localization with the uncertainty-based rejection method can be utilized as a global localization solution, which can also provide a reliable initial pose to conventional localization modules, e.g., AMCL.
Figure 5 illustrates the distribution of position and orientation errors after applying the uncertainty-based rejection approach. Each scatter plot shows error values on the x-axis and their corresponding uncertainties on the y-axis. The outputs from the network are color-coded to distinguish between low-uncertainty (more reliable) and high-uncertainty (potentially erroneous) outputs. In the figure, red dots represent outputs with low uncertainty (below the 70% threshold), which are considered reliable predictions. They remain after applying all thresholds (100%, 90%, 80%, and 70%). On the other hand, blue dots indicate high-uncertainty outputs that are rejected when the 90% threshold is applied, meaning a higher likelihood of being erroneous. Additionally, green and orange dots represent outputs with moderate uncertainty, positioned between the red and blue dots. The black dashed lines in each plot indicate the rejection thresholds applied at different percentages. As the threshold decreases (from 100% to 70%), the number of low-uncertainty outputs (red) increases, while high-uncertainty outputs (blue) are progressively discarded. On the right side of each plot, the first number represents the uncertainty threshold applied at that level, while the percentages and corresponding values indicate the number of the remaining outputs. Overall, we can see that this strategy can effectively enhance the reliability of the network outputs by rejecting those with high uncertainty. In Figure 5, we also observe that applying a 70% threshold led to the removal of 955 outputs from the TheGardenParty dataset, 838 outputs from the ETRI dataset, and 383 outputs from the SusungHotel dataset. This indicates that excessive rejection can result in data loss and require multiple inferences to provide a non-rejected output, while the rejection approach effectively reduces errors on average. Thus, it is crucial to determine an appropriate threshold. Experimental results indicate that the 70% threshold achieves a satisfactory rejection while maintaining a sufficient number of network outputs. However, in a specific application, the rejection ratio should be carefully adjusted to balance performance improvement and multiple inferences. Therefore, selecting an optimal threshold is essential to maximizing performance while preserving sufficient data for model learning.
To further illustrate the impact of uncertainty-based thresholding, we visualize the predicted trajectories with and without filtering in various test sequences. Figure 6 shows the predicted trajectories overlaid with ground truth paths, highlighting the effect of applying thresholds at 90%, 80%, and 70%. In these plots, the reduction of noisy predictions and the improved alignment with the ground truth after filtering are clearly observable.

5. Conclusions and Future Work

This study experimentally demonstrated that our uncertainty-based rejection can effectively enhance robot localization performance. By applying different rejection thresholds (90%, 80%, and 70%), we confirmed that discarding network outputs with high uncertainty reduces both positional and orientation errors, thereby improving the reliability of model evaluation. Unlike conventional end-to-end localization methods that treat all evaluations equally, the proposed approach improves the reliability of network outputs by selectively rejecting those with high uncertainty. This method can be applied to other localization techniques based on deep neural networks. However, it is essential to determine the appropriate value for the rejection threshold. Thus, dynamically adjusting the rejection threshold based on dataset characteristics is necessary.
Future work includes optimizing the uncertainty-based rejection method for real-time robotic applications and verifying its practicality through experiments on actual robots. In addition, we aim to enhance the correlation between prediction error and uncertainty through calibration techniques such as post-processing [28] and distribution matching [30,31]. It is also important to improve the relatively slow inference speed observed in the current model. These efforts are expected to further increase the reliability and accuracy of robot localization systems in real-world environments.

Author Contributions

Conceptualization, J.L. and J.O.; methodology, J.L. and J.O.; software, H.-M.W. and J.L.; validation, H.-M.W. and J.L.; data curation, H.-M.W. and J.L.; writing—original draft preparation, H.-M.W., J.L., and J.O.; writing—review and editing, H.-M.W., J.L., and J.O.; visualization, H.-M.W. and J.L.; supervision, J.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Electronic and Telecommunications Research Institute (ETRI) grant funded by the Korean government [25ZD1130, Development of ICT Convergence Technology for Daegu-Gyeongbuk Regional Industry (Robots)]. This work was also supported by the Commercialization Promotion Agency for R&D Outcomes (COMPA), granted by the Korean Government (Ministry of Science and ICT, RS-2023-00304776).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated and/or analyzed during the current study are not publicly available due to security restrictions and the protection of personal data.

Acknowledgments

The authors especially thank Hakjun Lee for his assistance with dataset collection.

Conflicts of Interest

Author Jieun Lee was employed by the company Polaris3D. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
  2. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  3. Lowry, S.; Sünderhauf, N.; Newman, P.; Leonard, J.J.; Cox, D.; Corke, P.; Milford, M.J. Visual Place Recognition: A Survey. IEEE Trans. Robot. 2016, 32, 1–19. [Google Scholar] [CrossRef]
  4. Zhang, Y.; Shi, P.; Li, J. Lidar-Based Place Recognition for Autonomous Driving: A Survey. ACM Comput. Surv. 2024, 57, 1–36. [Google Scholar] [CrossRef]
  5. Kendall, A.; Grimes, M.; Cipolla, R. Posenet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 2938–2946. [Google Scholar]
  6. Wang, B.; Chen, C.; Lu, C.X.; Zhao, P.; Trigoni, N.; Markham, A. AtLoc: Attention Guided Camera Localization. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10393–10401. [Google Scholar] [CrossRef]
  7. Li, X.; Ling, H. GTCaR: Graph Transformer for Camera Re-localization. In Computer Vision—ECCV 2022, Proceedings of the 17th European Conference on Computer Vision (ECCV 2022), Tel-Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer: Cham, Switzerland, 2022; pp. 229–246. [Google Scholar]
  8. Qiao, C.; Xiang, Z.; Fan, Y.; Bai, T.; Zhao, X.; Fu, J. TransAPR: Absolute Camera Pose Regression With Spatial and Temporal Attention. IEEE Robot. Autom. Lett. 2023, 8, 4633–4640. [Google Scholar] [CrossRef]
  9. Wang, W.; Wang, B.; Zhao, P.; Chen, C.; Clark, R.; Yang, B.; Markham, A.; Trigoni, N. PointLoc: Deep Pose Regressor for LiDAR Point Cloud Localization. IEEE Sens. J. 2022, 22, 959–968. [Google Scholar] [CrossRef]
  10. Yu, S.; Wang, C.; Wen, C.; Cheng, M.; Liu, M.; Zhang, Z.; Li, X. LiDAR-based localization using universal encoding and memory-aware regression. Pattern Recognit. 2022, 128, 108685. [Google Scholar] [CrossRef]
  11. Li, W.; Yu, S.; Wang, C.; Hu, G.; Shen, S.; Wen, C. SGLoc: Scene Geometry Encoding for Outdoor LiDAR Localization. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 9286–9295. [Google Scholar] [CrossRef]
  12. Li, W.; Yang, Y.; Yu, S.; Hu, G.; Wen, C.; Cheng, M.; Wang, C. DiffLoc: Diffusion Model for Outdoor LiDAR Localization. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 15045–15054. [Google Scholar] [CrossRef]
  13. Herath, S.; Caruso, D.; Liu, C.; Chen, Y.; Furukawa, Y. Neural Inertial Localization. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  14. Lee, J.; Lee, H.; Oh, J. FusionLoc: Camera-2D LiDAR Fusion Using Multi-Head Self-Attention for End-to-End Serving Robot Relocalization. IEEE Access 2023, 11, 75121–75133. [Google Scholar] [CrossRef]
  15. Jo, H.; Kim, E. New Monte Carlo Localization Using Deep Initialization: A Three-Dimensional LiDAR and a Camera Fusion Approach. IEEE Access 2020, 8, 74485–74496. [Google Scholar] [CrossRef]
  16. Gal, Y.; Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 1050–1059. [Google Scholar]
  17. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All you Need. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  18. Wang, J.; Yu, H.; Lin, X.; Li, Z.; Sun, W.; Akhtar, N. EFRNet-VL: An end-to-end feature refinement network for monocular visual localization in dynamic environments. Expert Syst. Appl. 2024, 243, 122755. [Google Scholar] [CrossRef]
  19. Zheng, J.; Liu, R.; Chen, Y.; Chen, Z.; Yang, K.; Zhang, J.; Stiefelhagen, R. Scene-agnostic Pose Regression for Visual Localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 11–15 June 2025. [Google Scholar]
  20. Yu, S.; Wang, C.; Lin, Y.; Wen, C.; Cheng, M.; Hu, G. STCLoc: Deep LiDAR Localization With Spatio-Temporal Constraints. IEEE Trans. Intell. Transp. Syst. 2023, 24, 489–500. [Google Scholar] [CrossRef]
  21. Ibrahim, M.; Akhtar, N.; Anwar, S.; Wise, M.; Mian, A. Slice Transformer and Self-supervised Learning for 6DoF Localization in 3D Point Cloud Maps. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 11763–11770. [Google Scholar] [CrossRef]
  22. Yu, S.; Sun, X.; Li, W.; Wen, C.; Yang, Y.; Si, B.; Hu, G.; Wang, C. NIDALoc: Neurobiologically Inspired Deep LiDAR Localization. IEEE Trans. Intell. Transp. Syst. 2024, 25, 4278–4289. [Google Scholar] [CrossRef]
  23. Yang, B.; Li, Z.; Lil, W.; Cai, Z.; Wen, C.; Zang, Y.; Muller, M.; Wang, C. LiSA: LiDAR Localization with Semantic Awareness. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 15271–15280. [Google Scholar] [CrossRef]
  24. Lai, H.; Yin, P.; Scherer, S. AdaFusion: Visual-LiDAR Fusion With Adaptive Weights for Place Recognition. IEEE Robot. Autom. Lett. 2022, 7, 12038–12045. [Google Scholar] [CrossRef]
  25. Wang, E.; Chen, D.; Fu, T.; Ma, L. A Robot Relocalization Method Based on Laser and Visual Features. In Proceedings of the 2022 IEEE 11th Data Driven Control and Learning Systems Conference (DDCLS), Emei, China, 3–5 August 2022; pp. 519–524. [Google Scholar]
  26. Nakamura, Y.; Sasaki, A.; Toda, Y.; Kubota, N. Localization Fault Detection Method using 2D LiDAR and Fisheye Camera for an Autonomous Mobile Robot Control. In Proceedings of the 2024 SICE International Symposium on Control Systems (SICE ISCS), Higashi-Hiroshima, Japan, 18–20 March 2024; pp. 32–39. [Google Scholar]
  27. Kendall, A.; Gal, Y. What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  28. Kuleshov, V.; Fenner, N.; Ermon, S. Accurate Uncertainties for Deep Learning Using Calibrated Regression. In Proceedings of Machine Learning Research, Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Dy, J., Krause, A., Eds.; JMLR: Cambridge, MA, USA, 2018; Volume 80, pp. 2796–2804. [Google Scholar]
  29. Platt, J.C. Probabilistic Outputs for Support Vector Machines and Comparisons to Regularized Likelihood Methods. In Advances in Large Margin Classifiers; MIT Press: Cambridge, MA, USA, 1999; pp. 61–74. [Google Scholar]
  30. Cui, P.; Hu, W.; Zhu, J. Calibrated Reliable Regression using Maximum Mean Discrepancy. Adv. Neural Inf. Process. Syst. 2020, 33, 17164–17175. [Google Scholar]
  31. Bhatt, D.; Mani, K.; Bansal, D.; Murthy, K.; Lee, H.; Paull, L. f-Cal: Aleatoric uncertainty quantification for robot perception via calibrated neural regression. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 6533–6539. [Google Scholar] [CrossRef]
  32. Yu, X.; Franchi, G.; Aldea, E. SLURP: Side Learning Uncertainty for Regression Problems. In Proceedings of the 32nd British Machine Vision Conference, BMVC, Virtual, 22–25 November 2021. [Google Scholar]
  33. Corbière, C.; Thome, N.; Saporta, A.; Vu, T.H.; Cord, M.; Pérez, P. Confidence Estimation via Auxiliary Models. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 6043–6055. [Google Scholar] [CrossRef] [PubMed]
  34. Abdar, M.; Pourpanah, F.; Hussain, S.; Rezazadegan, D.; Liu, L.; Ghavamzadeh, M.; Fieguth, P.; Cao, X.; Khosravi, A.; Acharya, U.R.; et al. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Inf. Fusion 2021, 76, 243–297. [Google Scholar] [CrossRef]
  35. Gawlikowski, J.; Tassi, C.R.N.; Ali, M.; Lee, J.; Humt, M.; Feng, J.; Kruspe, A.; Triebel, R.; Jung, P.; Roscher, R.; et al. A survey of uncertainty in deep neural networks. Artif. Intell. Rev. 2023, 56, 1513–1589. [Google Scholar] [CrossRef]
  36. Chen, J.; Monica, J.; Chao, W.L.; Campbell, M. Probabilistic Uncertainty Quantification of Prediction Models with Application to Visual Localization. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 4178–4184. [Google Scholar] [CrossRef]
Figure 1. Our pipeline for uncertainty-aware end-to-end localization.
Figure 1. Our pipeline for uncertainty-aware end-to-end localization.
Technologies 13 00386 g001
Figure 2. A serving robot used in this study: The blue box represents a SLAMTEC RPLiDAR A1M8, while the red boxes indicate Intel RealSense D435 cameras.
Figure 2. A serving robot used in this study: The blue box represents a SLAMTEC RPLiDAR A1M8, while the red boxes indicate Intel RealSense D435 cameras.
Technologies 13 00386 g002
Figure 5. Comparison of uncertainty-based rejection results with varying thresholds. The uncertainty values are scaled between zero and one for visualization. The upper and lower scatter plots of each dataset represent the uncertainty for position and orientation errors, respectively.
Figure 5. Comparison of uncertainty-based rejection results with varying thresholds. The uncertainty values are scaled between zero and one for visualization. The upper and lower scatter plots of each dataset represent the uncertainty for position and orientation errors, respectively.
Technologies 13 00386 g005aTechnologies 13 00386 g005b
Figure 6. Trajectory visualization with different uncertainty thresholds (100%, 90%, 80%, 70%). (a) TheGardenParty. (b) ETRI: Full-loop trajectory. (c) ETRI: Zigzag navigation. (d) ETRI: Localized back-and-forth motion. (e) ETRI: In-place rotations at specific locations. (f) SusungHotel.
Figure 6. Trajectory visualization with different uncertainty thresholds (100%, 90%, 80%, 70%). (a) TheGardenParty. (b) ETRI: Full-loop trajectory. (c) ETRI: Zigzag navigation. (d) ETRI: Localized back-and-forth motion. (e) ETRI: In-place rotations at specific locations. (f) SusungHotel.
Technologies 13 00386 g006
Table 3. Comparison of position and orientation metrics under different uncertainty thresholds across all datasets.
Table 3. Comparison of position and orientation metrics under different uncertainty thresholds across all datasets.
DatasetMetricThreshold100% th.90% th.80% th.70% th.
TheGardenPartyPosition (m)Min0.0720.0740.0690.098
Median0.5730.5670.5520.523
Max5.5595.2884.3174.192
Mean0.7760.7380.7120.682
Orientation ( ° )Min0.02040.0170.0030.015
Median2.6722.5442.5122.299
Max72.73854.50849.38448.397
Mean4.8803.7833.3583.126
ETRIPosition (m)Min0.0230.0250.0260.031
Median0.2890.2980.3000.298
Max1.6101.6131.2871.490
Mean0.3240.3260.3270.327
Orientation ( ° )Min0.0130.0120.0080.010
Median1.1271.0520.9780.902
Max26.22416.18215.56911.024
Mean1.9901.6611.5271.409
SusungHotelPosition (m)Min0.0220.0280.0310.019
Median0.6590.5610.4770.424
Max66.39266.27959.82959.753
Mean4.2622.7882.2111.575
Orientation ( ° )Min0.0040.0080.0090.008
Median2.7322.4662.1682.020
Max179.385174.738173.323127.828
Mean13.7119.0085.8614.329
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Won, H.-M.; Lee, J.; Oh, J. Localization Meets Uncertainty: Uncertainty-Aware Multi-Modal Localization. Technologies 2025, 13, 386. https://doi.org/10.3390/technologies13090386

AMA Style

Won H-M, Lee J, Oh J. Localization Meets Uncertainty: Uncertainty-Aware Multi-Modal Localization. Technologies. 2025; 13(9):386. https://doi.org/10.3390/technologies13090386

Chicago/Turabian Style

Won, Hye-Min, Jieun Lee, and Jiyong Oh. 2025. "Localization Meets Uncertainty: Uncertainty-Aware Multi-Modal Localization" Technologies 13, no. 9: 386. https://doi.org/10.3390/technologies13090386

APA Style

Won, H.-M., Lee, J., & Oh, J. (2025). Localization Meets Uncertainty: Uncertainty-Aware Multi-Modal Localization. Technologies, 13(9), 386. https://doi.org/10.3390/technologies13090386

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop