Next Article in Journal
A Multi-Input Multi-Output Considering Correlation and Hysteresis Prediction Method for Gravity Dam Displacement with Interpretative Functions
Previous Article in Journal
Evaluation of the Barriers to Maintenance 4.0 for the Textile Industry via Pythagorean Fuzzy SWARA
Previous Article in Special Issue
Predicting Post-Liposuction Body Shape Using RGB Image-to-Image Translation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Precise Feature Removal Method Based on Semantic and Geometric Dual Masks in Dynamic SLAM

1
School of Computer and Electronic Information, Guangxi University, Nanning 530004, China
2
Nanning Huishi Technology Co., Ltd., Nanning 530000, China
3
Guangxi Intelligent Digital Services Research Center of Engineering Technology, Nanning 530004, China
4
Key Laboratory of Parallel, Distributed and Intelligent Computing (Guangxi University), Education Department of Guangxi Zhuang Autonomous Region, Nanning 530004, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(13), 7095; https://doi.org/10.3390/app15137095
Submission received: 15 May 2025 / Revised: 13 June 2025 / Accepted: 19 June 2025 / Published: 24 June 2025

Abstract

In visual Simultaneous Localization and Mapping (SLAM) systems, dynamic elements in the environment pose significant challenges that complicate reliable feature matching and accurate pose estimation. To address the issue of unstable feature points within dynamic regions, this study proposes a robust dual-mask filtering strategy that synergistically integrates semantic segmentation information with geometric outlier detection techniques. The proposed method first identifies outlier feature points through rigorous geometric consistency checks, then employs morphological dilation to expand the initially detected dynamic regions. Subsequently, the expanded mask is intersected with instance-level semantic segmentation results to precisely delineate dynamic areas, effectively constraining the search space for feature matching and reducing interference caused by dynamic objects. A key innovation of this approach is the incorporation of a Perspective-n-Point (PnP)-based optimization module. This module dynamically updates the outlier set on a per-frame basis, enabling continuous monitoring and selective removal of dynamic features. Extensive experiments conducted on benchmark datasets demonstrate that the proposed method achieves average accuracy improvements of 3.43% and 11.42% on the KITTI dataset and 24% and 8.27% on the TUM dataset. Compared to traditional methods, this dual-mask collaborative filtering strategy improves the accuracy of dynamic feature removal and enhances the reliability of dynamic object detection, validating its robustness and applicability in complex dynamic environments.

1. Introduction

Simultaneous Localization and Mapping (SLAM) is a critical technology for autonomous robots, self-driving vehicles, unmanned aerial vehicles, and related fields. The aim of SLAM is to allow devices to construct maps of unknown environments while simultaneously performing autonomous localization and navigation [1,2,3]. With the rapid advancement of computer vision and deep learning in combination with continuous improvement of hardware computational capabilities, vision-based SLAM systems now demonstrate three core advantages over other sensors such as LiDAR, millimeter-wave radar, and satellite navigation systems. First, in terms of cost-effectiveness, SLAM can achieve centimeter-level positioning accuracy using only consumer grade cameras [4,5]. Second, in terms of environmental adaptability, independence from external signal sources ensures continuous localization capability in environments where global navigation satellite systems are unavailable (e.g., indoor spaces, tunnels). Most significantly, SLAM systems natively support synergistic processing with computer vision tasks, enabling direct environmental understanding through real-time semantic segmentation algorithms, a unique capability unattainable by other sensors. These competitive advantages establish visual SLAM as a predominant research direction in cost-sensitive applications requiring environmental intelligence, including autonomous mobile robots and augmented-reality navigation systems. Modern visual SLAM systems exhibit highly mature architectures typically comprising several core components, including a feature extraction frontend, a state estimation backend, and loop closure detection modules. Furthermore, advanced SLAM algorithms such as ORB-SLAM3 [6], DSO [7], VSO [8], and LSD-SLAM [9] have achieved highly satisfactory performance levels.
However, in practical application scenarios, vision-based SLAM systems typically assume a static environment in which the relative positions among scene points remain constant and all visual changes are attributed to camera motion [10]. This assumption often does not hold in real-world environments, especially in scenarios with dynamic objects, where the positional accuracy and map construction quality of traditional visual SLAM methods may be seriously compromised. Moreover, existing methods generally adopt a strategy of directly discarding all potential dynamic feature points. Although this approach can effectively avoid dynamic interference, it may also mistakenly eliminate some static feature points, thereby affecting the stability and map quality of the SLAM system [11].
In response to these issues, the objective of this study is to propose a dynamic feature point filtering method to enhance the feature point selection capability in dynamic environments. By combining semantic segmentation masks with geometric consistency constraints, we aim to more accurately identify and preserve static feature points, thereby improving the localization accuracy and map construction quality of SLAM systems. Concurrently, our method aims to reduce the interference of dynamic objects on the SLAM system, ensuring stable operation in complex and dynamic environments. Geometric consistency constraints are applied to the filtered feature points for validation (via reprojection error checks) and optimization (via bundle adjustment). Notably, the information of outlier feature points is preserved in order to assist in the ongoing judgment of dynamic regions under ever-changing environments [12,13,14]. The main contributions of this paper are as follows:
  • A dual-mask dynamic feature point filtering method is proposed that combines semantic segmentation masks with outlier information to precisely filter dynamic feature points during the feature matching stage. Instead of simply eliminating the entire dynamic area, this method identifies the local motion areas of dynamic objects. This minimizes false eliminations while still preserving as many static feature points as possible. The low standard deviation achieved in our experimental evaluation serves as evidence demonstrating the effectiveness of dynamic masking in reducing false eliminations and preserving static feature points.
  • A dynamic feature point update mechanism is designed. On the basis of dual-mask filtering, outliers are effectively identified and eliminated through PnP pose optimization to dynamically update the set of geometric outliers. This provides a reliable reference for subsequent frames, further improving the detection and elimination of dynamic areas.
  • Empirical evaluations on public dynamic datasets, including KITTI and TUM, demonstrate that the proposed method significantly outperforms ORB-SLAM3 by reducing both the Absolute Trajectory Error (ATE) and Relative Pose Error (RPE) in dynamic environments, validating its superiority in terms of both accuracy and robustness. Specifically, the proposed method achieves average reductions in error of 3.43% (ATE) and 11.42% (RPE) on the KITTI dataset and 24% (ATE) and 8.27% (RPE) on the TUM dataset.
The remainder of this paper is structured as follows. Section 2 reviews the research progress in visual SLAM and focuses on discussing existing dynamic object handling techniques while analyzing the strengths and weaknesses of these methods and the challenges that they face. Section 3 introduces the design and implementation of the proposed system, including core ideas, system architecture, and key technologies. Section 4 evaluates the performance of the proposed method through experiments on the KITTI and TUM datasets and compares it with existing methods to demonstrate the system’s performance improvements. Finally, Section 5 concludes the main work of the paper, discusses the limitations of the research, and looks forward to potential future research directions and application prospects.

2. Related Work

2.1. Visual SLAM Literature

The core task of visual SLAM is to estimate camera poses and construct environmental maps through pixel matching in image sequences. Early MonoSLAM [15] utilized extended Kalman filtering as the backend, supporting single-threaded processing of a small number of sparse feature points. Subsequently, PTAM [16] introduced keyframes and a dual-thread architecture to achieve parallel processing of camera tracking and mapping, which significantly improved system performance. Building upon this foundation, ORB-SLAM2 [17] added loop closure detection and local bundle adjustment optimization, further enhancing robustness and accuracy. The latest ORB-SLAM3 [6] integrates visual–inertial information and introduces a multi-map system to improve adaptability in multi-sensor configurations. However, these methods generally assume a static environment; in dynamic scenes, the feature points on moving objects violate this assumption, easily introducing errors that reduce localization accuracy and map quality. Thus, optimizing SLAM for dynamic scenarios has become a current hot topic of research.

2.2. Geometry-Constrained Dynamic SLAM

Methods based on geometric constraints typically employ robust weighting functions or motion consistency constraints to eliminate dynamic targets, thereby improving localization accuracy. These methods usually rely on the motion trajectory or geometric relationships of feature points to detect dynamic objects [18,19,20]. Kim et al. [21] eliminated outliers by estimating a non-parametric background model of the depth scene, achieving camera motion estimation on the basis of this model. Li et al. [22] proposed an RGB-D SLAM system based on keyframes, incorporating static weights of foreground edge points into the Intensity-Assisted Iterative Closest Point method to improve pose estimation accuracy. Sun et al. [23] used optical flow to track dynamic targets, then constructed and updated foreground models based on RGB-D data to remove dynamic feature points. Additionally, StaticFusion [24] simultaneously estimates camera motion and probabilistic static segmentation for static background reconstruction. Overall, geometry-based approaches typically require the integration of additional geometric computation modules, which increases the system’s implementation complexity. In contrast, the method proposed in this paper fully leverages the existing PnP optimization residuals within ORB-SLAM3. By matching outlier points between consecutive frames and analyzing their geometric consistency, it can effectively identify dynamic features without introducing additional geometric models or structural assumptions, thereby achieving lower computational cost and greater practical efficiency.

2.3. Semantic Segmentation-Based Dynamic SLAM

Dynamic SLAM methods based on semantic segmentation utilize deep learning networks to semantically parse images, thereby eliminating potential dynamic targets to enhance localization accuracy and map quality [12,14,25,26,27]. DS-SLAM [28] combines ORB-SLAM2 with a segmentation network for instance segmentation and verifies the motion consistency of feature points using epipolar constraints; however, it is limited to predefined dynamic categories (e.g., pedestrians). DynaSLAM [29] employs Mask R-CNN for instance segmentation, which it uses along with geometric constraints to identify dynamic regions in RGB-D images. This approach significantly improves robustness in dynamic scenes, albeit with a high computational cost that limits real-time performance due to inference times exceeding 200 ms. VDO-SLAM [30] uses dense optical flows to track object motion and enhances mapping capabilities for multi-target dynamic scenes through joint optimization of camera and object poses. However, it is primarily suitable for offline processing. For greater efficiency and performance balance, methods such as Detect-SLAM [31] and RDS-SLAM [32] focus on semantic processing of keyframes. This approach performs instance segmentation on keyframes to reduce computation, but relies heavily on keyframe quality and lacks dynamic frame selection strategies, limiting system adaptability in complex scenes. Overall, while semantic-based dynamic SLAM has made progress in dynamic target perception, it still faces challenges such as computational complexity, erroneous elimination of static feature points, and over-reliance on keyframes, making it difficult to balance real-time performance with accuracy [33]. To address these issues, our paper proposes a dual-mask strategy that integrates instance segmentation with outlier residual information. The proposed approach first employs YOLOv11 to perform coarse segmentation of potential dynamic objects, then refines the detection by leveraging geometric residuals to further eliminate anomalous regions. This method effectively reduces the system’s dependence on keyframes and enables precise removal of dynamic features within local areas. By improving the system’s tolerance to erroneous rejection of static points, the proposed strategy enhances the overall robustness and stability of the system while maintaining real-time performance.

3. Methodology

This paper presents a visual SLAM system capable of handling dynamic environments. The proposed system is based on the monocular ORB-SLAM3 framework, with a focus on elimination of dynamic features to improve localization and mapping accuracy in dynamic scenes. As depicted in Figure 1, the overall framework of the system includes modules for image input, feature point extraction, semantic segmentation, dynamic feature point filtering, and pose optimization. The core contribution of this paper is centered on dynamic feature point filtering and dynamic region updating. Starting with image frame input, which serves as the carrier of the original visual information, the system utilizes the ORB-SLAM3 framework algorithm to extract feature points, providing a foundation for subsequent operations. Concurrently, the YOLOv11 neural network model is employed for semantic segmentation to generate segmentation masks that contain prior information about the location of dynamic content such as pedestrians and vehicles. Based on these segmentation masks derived from semantic segmentation along with feature points extracted using geometric constraints, a further in-depth analysis is conducted to precisely differentiate outliers into static and dynamic points. The overall process of the proposed system includes the following main steps:
  • Image Input: Receives original video frames as the starting point for SLAM processing.
  • Feature Point Extraction: Uses ORB-SLAM3 to extract feature points from each frame.
  • Semantic Segmentation: Utilizes YOLOv11 to perform semantic segmentation, creating masks for potential dynamic objects.
  • Mask Processing: Performs morphological dilation on the generated masks to enhance the capture precision of dynamic object boundaries.
  • Dynamic Feature Point Filtering: Combines a dual-mask strategy with geometric information to effectively filter out dynamic feature points.
  • Pose Optimization: Improves camera localization accuracy through PnP optimization calculations and dynamically updates the feature point set to support subsequent frame processing.

3.1. Detecting Dynamic Objects and Generating Masks with YOLOv11

To effectively detect dynamic objects and achieve high-precision pose estimation in dynamic environments, this paper employs the advanced YOLOv11 [34] deep learning model for semantic segmentation to identify potential dynamic objects. As a cutting-edge technology in the field of instance segmentation, YOLOv11 provides pixel-level instance segmentation and instance labeling of images. These capabilities are fundamental for enhancing our system’s ability to handle dynamic contexts.
Throughout the processing workflow, the semantic segmentation functionality of YOLOv11 is of primary importance within this system. It achieves pixel-level segmentation of each frame, allowing for accurate identification of different object regions. YOLOv11 interprets the input RGB raw images to recognize possible dynamic entities (e.g., people, bicycles, cars, motorcycles, buses) and items that might be moved (books, chairs, etc.). Dynamic objects such as these encompass the most common elements found in most environments. By combining all objects, a mask reflecting all dynamic regions in a single image is obtained, providing precise spatial localization for the subsequent feature point filtering.
To further enhance the semantic mask’s coverage accuracy over object boundaries, in this paper we apply a range of morphological dilation (e.g., 5 pixels) to the masks generated by YOLOv11. This effectively expands the boundaries of dynamic objects and compensates for local discontinuities and gaps caused by instance segmentation errors. This process strengthens the continuity and integrity of the mask around the target edge areas, addressing the common issue of "truncation" or "under-labeling" in instance segmentation at object edges. This treatment ensures that dynamic targets are not incorrectly classified as static regions due to incomplete boundaries during the geometric processing stage. The edge regions of dynamic objects often contain many inconsistent motion feature points or feature discrepancies caused by occlusion, especially in monocular SLAM systems. If these points are not fully captured, they may lead to mismatches and greatly impact the accuracy of pose estimation. Therefore, the morphological dilation step aids in improving the reliability of feature point matching as well as the overall robustness of the system in complex dynamic scenes.

3.2. Dual-Mask Method for Dynamic Feature Point Filtering

Under the assumption of a constant velocity model in monocular visual SLAM, the proposed system typically uses the pose from the previous frame as the initial value for the current frame’s initial pose, matching it with the 3D map points corresponding to feature points from the previous frame. To achieve feature matching, feature points associated with known map points from the previous frame are reprojected into the current frame and matches are sought based on their descriptor information. However, points determined as geometric outliers in the previous frame often have large reprojection errors, possibly due to inaccurate localization, descriptor anomalies, or being located on the surface of dynamic objects. Consequently, traditional ORB-SLAM systems usually adopt the strategy of excluding feature points marked as outliers from the previous frame, preventing them from participating in the current frame’s feature matching process. This avoids wasting computational resources and reduces potential mismatches.
To enhance the system’s ability to discriminate dynamic areas, this paper proposes a dual-mask dynamic feature filtering method that integrates semantic priors with geometric consistency verification. This approach re-utilizes feature points identified as outliers in the previous frame as clues for potential dynamic regions, combining them with instance masks provided by semantic segmentation to spatially constrain dynamic objects. When the number of outliers detected in the current frame falling within the semantic dynamic mask region exceeds a set threshold, these points are considered to be dynamic features. In this way, the system realizes dynamic detection driven by both semantic and geometric cues. This method not only combines information about dynamic physical objects from a semantic perspective, it also relies on geometric verification to identify unstable regions, effectively avoiding the influence of both pure semantic misjudgments (such as stationary pedestrians or parked vehicles) and pure geometric misjudgments (such as outlier points caused by sparse features).
For local dynamic targets, for instance a person moving only their head or hand, it would be inappropriate to label the entire body region as dynamic; to address this, the system initially identifies feature points that have been verified as outliers in the current frame and marks those that fall within the semantic dynamic mask as dynamic features. Subsequently, these outlier points undergo morphological dilation (approximately 60 pixels) to expand their influence area, creating a local dynamic point mask. Finally, this dilated mask is intersected with the semantic mask. This limits the dynamic region to only cover portions of potential dynamic objects, thereby obtaining a more precise final dynamic mask.
The processing flow is depicted in Figure 2. The input consists of two consecutive frames, from which the system first obtains semantic masks through semantic segmentation; then, combining geometric consistency checks, it identifies dynamic outliers to generate an outlier mask. Finally, the semantic mask and the outlier mask are combined to produce an accurate dynamic object mask, achieving more precise dynamic feature point removal.
The final mask generation process can be expressed by the following mathematical equation:
M final = M obj E t ε
where M obj represents the dynamic object mask generated by semantic segmentation, E t denotes the set of points identified as outliers through geometric verification and feature matching in the current frame, and ε represents the morphological dilation operation on the outlier set with a kernel size of ε . The parameter ε is empirically set to 50 pixels, a value that was determined through extensive comparative experiments on the KITTI and TUM RGB-D datasets. This setting achieves a balance between dynamic and static regions. A larger value of ε tends to result in the direct removal of entire dynamic objects, while a smaller value leads to the elimination of only the dynamic points.
Furthermore, E ( t 1 ) is the set of outliers from the previous frame, which can be utilized to assist in the prediction of dynamic areas in the next frame. Introducing temporal consistency further enhances the system’s dynamic perception stability in consecutive frames. This dual-mask strategy effectively removes dynamic interference features while preserving stable feature points in static regions, improving overall matching accuracy and the robustness of pose estimation.

3.3. Geometric Consistency Verification

During the pose estimation phase of ORB-SLAM3, the system minimizes the reprojection error of matched feature points through a PnP nonlinear optimization module to optimize the camera pose. To ensure the geometric validity of the matched points, ORB-SLAM3 incorporates a geometric consistency verification mechanism that verifies matched points through a reprojection model to identify and exclude mismatches. The core of geometric consistency verification lies in calculating the reprojection error. Let a map point in the world coordinate system be P w and the camera pose be ( R c w , t c w ) , which in the camera coordinate system is
P c = R cw · P w + t cw .
Projecting this onto the image plane using the intrinsic camera parameters yields the predicted pixel coordinates u ^ , which when compared with the actual observed point u results in the reprojection error:
e = u ^ u .
ORB-SLAM3 minimizes the reprojection error of all matched points to solve for the camera pose during pose optimization, as follows:
min R , t i ρ ( u ^ i u i 2 ) .
In the above equation, the matrix R represents the camera’s rotation matrix (belonging to the special orthogonal group S O ( 3 ) ) with constraints R T R = I and det ( R ) = 1 . These conditions ensure that the rotation matrix is orthogonal with a determinant of 1, conforming to the mathematical properties of rotational transformations in three-dimensional space, and are used to describe the rotational behavior of the camera’s pose in space. The translation vector t R 3 can theoretically take any value in 3D Euclidean space without scene-specific constraints. Here, i represents the calculation of the i-th feature point, while ρ is a robust kernel function (such as the Huber kernel) used to mitigate the influence of outliers on optimization. When the reprojection error e of a feature point exceeds a set threshold, that point is determined to be a geometric outlier and is no longer involved in the pose optimization calculation for the current frame. The proposed approach utilizes the Levenberg–Marquardt nonlinear least-squares optimization algorithm to address the robust kernel-function formulated optimization problem. This represents a geometrically consistent validation process that operates synchronously during optimization iterations. As such, it forms a reliable outlier filtering mechanism within the system, particularly with innate recognition capabilities for feature points on dynamic objects.
During the feature matching process in the current frame, ORB-SLAM3 directly excludes those feature points that were marked as outliers in the previous frame, meaning that these outliers participate in neither feature matching nor pose estimation. To more effectively utilize this information, in this paper we propose an improved strategy. First, outliers are further classified in the feature matching stage, and those falling within the potential dynamic object mask are marked as dynamic points. During the pose estimation process, while outliers are excluded, both dynamic points and regular feature points participate in the computation, with the dynamic points being assigned a weight of zero. This approach aims to identify current-frame outliers. This approach provides support for the judgment of dynamic objects and areas in the next frame, thereby forming a closed-loop system process.
The strategy proposed in this paper is more efficient compared to methods that introduce additional binocular geometric constraints or lightweight PnP pose estimations for dynamic area filtering prior to the feature matching stage. We fully leverage the complete optimization process already inherent in ORB-SLAM3, allowing for dynamic point identification without additional computations. This reduces system overhead while also improving both the accuracy of dynamic object recognition and the overall robustness of the system.

4. Experimental Results and Analysis

We conducted experimental validation of the proposed system using the public datasets KITTI and TUM. In our experiments, the proposed method was compared and evaluated against the mainstream ORB-SLAM3 system, with the Absolute Trajectory Error (ATE) and Relative Pose Error (RPE) selected as performance metrics. The ATE reflects the overall deviation between the system’s estimated trajectory and the actual trajectory, intuitively demonstrating the cumulative error caused by mismatches in dynamic features. The RPE measures the accuracy of relative pose estimation between adjacent frames but is more susceptible to instantaneous dynamic disturbances; as a result, it is primarily used to assess a system’s local stability. For error quantification, the ATE is assessed using the Root Mean Square Error (RMSE) to reflect the degree of overall trajectory deviation, while the RPE is measured using the mean in order to reflect the system’s stability performance in inter-frame pose estimation. The above evaluation methods enable a comprehensive verification of the system’s accuracy and robustness in different dynamic environments.
To reduce interference from external factors, each test sequence was subjected to ten repeated experiments, with the median taken as the final experimental outcome. All experiments were conducted on a desktop computer equipped with an Intel i5-14600KF processor, NVIDIA GeForce RTX 4070S graphics card, and 32GB of memory.

4.1. Evaluation on the KITTI Dataset

The KITTI dataset provides real vehicular scenes containing both dynamic and static objects, and is widely used to evaluate the effectiveness of visual SLAM systems in separating moving and stationary objects in dynamic environments. In this experiment, the performance of the SLAM system proposed in this paper was compared and analyzed against ORB-SLAM3 under a monocular setting across eleven image sequences, with a focus on the impact of the dynamic object elimination strategy on the system’s localization accuracy and robustness. The performance of different systems in each sequence is summarized in Table 1 and Table 2, where bold font indicates the best result for that metric.
In Table 1 and Table 2, both System(D) and System(D+O) represent the dynamic SLAM systems proposed in this paper; System(D) refers to the system that uses a dynamic mask to directly eliminate the entire dynamic object, while System(D+O) refers to the system that uses a combination of dynamic mask and outlier mask in a dual-mask method to precisely exclude dynamic points. In the absolute trajectory evaluation on the KITTI dataset, System(D+O) is the most accurate in most sequences; in relative pose estimation, System(D) and System(D+O) are similar, with both performing better than the ORB-SLAM3 system.
Meanwhile, in the evaluation of ATE, the average standard deviations of ORB-SLAM3, System(D), and System(D+O) are 19.01, 16.90, and 17.64, respectively. Compared to ORB-SLAM3, System(D) achieves a reduction of approximately 11.1%, while System(D+O) achieves a reduction of about 7.2%. In the evaluation of RPE, the average standard deviations are 0.373, 0.282, and 0.325 for ORB-SLAM3, System(D), and System(D+O), respectively, with System(D) and System(D+O) reducing the standard deviation by 24.4% and 12.9%, respectively. These results indicate that System(D), which directly removes dynamic objects using a dynamic mask strategy, achieves the lowest average standard deviations in both ATE and RPE evaluations. This suggests significantly reduced error fluctuations compared to ORB-SLAM3 and verifies the effectiveness of the proposed strategy in suppressing error dispersion and enhancing the stability of the SLAM system, with particularly notable advantages in relative pose estimation.
To further investigate the specific impact of potential dynamic objects on SLAM system performance, Table 3 lists the distribution of potential dynamic objects in each sequence along with an analysis of their impact on localization accuracy. For brevity, in this paper we denote System(D) as “D”, System(D+O) as “DO”, and ORB-SLAM3 as “3”.
The analysis shows that the distribution of dynamic and static objects significantly affects system performance. In the KITTI dataset sequences, ORB-SLAM3 only outperforms the dynamic SLAM system in sequences with fewer dynamic objects, such as sequences 03 and 06. This suggests that the dynamic SLAM system may erroneously eliminate static objects, resulting in reduced trajectory estimation accuracy.
Conversely, in scenarios with many dynamic objects, such as sequences 01, 04, and 07, the dynamic SLAM system significantly outperforms ORB-SLAM3 in terms of both ATE and RPE, further validating the effectiveness of the dynamic object elimination mechanism in complex dynamic environments. To quantify the degree of system performance improvement, this paper uses the following improvement rate (enhancement rate) calculation equation:
η = o r o × 100 %
where o represents the error value of ORB-SLAM3, r represents the error value of the proposed method in this paper, and η is the percentage of improvement.
As shown in Table 4, based on this evaluation, System(D) achieves an average improvement of 6.69% in absolute trajectory accuracy and 8.69% in relative pose estimation accuracy compared to ORB-SLAM3, while System(D+O) achieves an average improvement of 3.43% and 11.42%, respectively. These results demonstrate the comprehensive advantage of the proposed systems in dynamic environments.
Further comparison between System(D) and System(D+O) reveals that because the KITTI dataset primarily consists of vehicular scenes, with potential dynamic objects mostly being vehicles, System(D) employs a strategy of globally removing dynamic objects, which requires high accuracy in dynamic object recognition. Any misclassification can lead to a significant performance degradation in the system. In scenes with a higher presence of static vehicles and pedestrians and with fewer dynamic vehicles (such as sequences 00, 08, 09, and 10), System(D+O) outperforms System(D) in both trajectory and relative pose accuracy. This indicates that its fine-grained dual-mask method effectively prevents performance fluctuations caused by the misremoval of static objects, demonstrating stronger stability.
In scenarios with a high density of dynamic objects but a lack of static ones (such as sequences 01 and 04), System(D) exhibits superior performance due to the absence of misjudged static object interference, as its coarse-grained elimination strategy is more advantageous in such contexts.
In summary, the experimental results demonstrate that the dual-mask dynamic object elimination method proposed in this paper can effectively reduce pose estimation errors in dynamic scenes. The proposed system achieves accuracy surpassing that of ORB-SLAM3 in most KITTI sequences, exhibiting robustness and versatility. Even in complex boundary scenarios with many static objects and fewer dynamic ones, the performance of our system remains stable, further validating the key role of the proposed dual-mask strategy in enhancing accuracy and stability in dynamic SLAM systems.

4.2. TUM RGB-D Dataset Evaluation

The TUM RGB-D dataset includes real indoor scenes with human activity and static objects, and is designed to assess system accuracy and stability in high-dynamic indoor environments. In this experiment, we conducted a comparative analysis of the impact of dynamic contexts on the performance of the proposed SLAM system versus ORB-SLAM3 in a monocular setting across six image sequences from the TUM RGB-D dataset. The performance of the different systems across the sequences is summarized in Table 5, Table 6 and Table 7, where bold font indicates the best results in each metric.
As presented in Table 7, compared to ORB-SLAM3, the proposed System(D) achieves an average reduction of 28% in ATE and 4.27% in RPE. The enhanced System(D+O) achieves an average reduction of 24% in ATE and 8.27% in RPE. These quantitative results verify that our methods effectively mitigate cumulative errors induced by dynamic feature mismatches. Furthermore, the experimental data in Table 5 and Table 6 reveal that both System(D) and System(D+O) exhibit lower standard deviation (Std) values, demonstrating the high adaptability and stability of our proposed approaches in dynamic environments. Similarly, Figure 3 and Figure 5 visually demonstrate that both System(D) and System(D+O) exhibit narrower standard deviation ranges compared to ORB-SLAM3, indicating more stable performance.
From a specific scenario analysis, walking sequences such as w_halfsphere, w_rpy, and w_static contain significant overall motion by individuals, causing notable dynamic interference. Traditional methods are prone to feature matching failures and drift estimation due to the presence of dynamic points. In contrast, the proposed approaches exhibit greater accuracy improvements in these highly dynamic sequences, indicating that our dynamic point elimination mechanism has good suppression capabilities for widespread disturbances.
In the experimental results for the fre3_walking_rpy(w_rpy) sequence, Figure 3 shows that the trajectory error of ORB-SLAM is in the range of 0.019–0.035, the trajectory error of System(D) is in the error range is 0.018–0.03, and the trajectory error of System(D+O) is in the range of 0.017–0.043. The accuracy of System(D+O) is almost the same as that of System(D). Figure 4 presents the trajectory deviation map in comparison to ground truth, showing that ORB-SLAM3 experiences significant drift, while System(D) offers the most accurate trajectory estimation. In these results, System(D+O) exhibits marginally lower performance compared to System(D).
Conversely, sitting sequences such as s_halfsphere, s_rpy, and s_static are less dynamic, with only minor movements of the hands and head. In such scenarios, systems are more likely to misjudge local dynamics as global dynamics, leading to loss of static information. Therefore, the fine-grained dynamic discrimination strategy introduced in this paper is particularly crucial.
In the experimental results for the fre3_sitting_xyz(s_xyz) sequence, the constrained local movements in this sitting scenario are primarily limited to head and hand motions, resulting in negligible dynamic interference. Figure 5 shows that the trajectory error of ORB-SLAM is in the range of 0.019–0.035 and that the trajectory error is relatively small, while the trajectory error of System(D) is in the error range is 0.018–0.03 and the trajectory error of System(D+O) is in the range of 0.017–0.043. As evidenced in Figure 6, the trajectory deviation map demonstrates no visually discernible differences between the compared systems. Thus, both System(D) and System(D+O) continue to exhibit enhancements in the ATE and RPE metrics compared with the ORB-SLAM3 system.
When comparing the performance of System(D+O) and System(D), the former demonstrates superior ATE performance in the sitting sequences, indicating that its fine-grained dynamic point removal strategy can more effectively distinguish between local microdynamics and globally static regions, resulting in improved overall localization accuracy. System(D+O) also demonstrates measurable improvements over System(D), with enhancements of 3.18% in ATE and 8.15% in RPE. In contrast, the coarse-grained removal mechanism employed by System(D) is more prone to incorrectly eliminating static objects in such fine-grained scenarios, resulting in decreased accuracy. Conversely, in the highly dynamic walking sequences, the global removal strategy of System(D) exhibits greater advantages, as it can rapidly eliminate interfering features within large-scale moving regions. As a result, System(D) shows slightly better ATE performance in certain sequences compared to System(D+O).
In conclusion, the dual-mask dynamic point filtering method proposed in this paper demonstrates exceptional performance in complex indoor dynamic scenarios. In particular, it shows a distinct advantage in the ATE metric. System(D+O) is suitable for refined scenarios with a mix of dynamic areas and static backgrounds, while System(D) is more appropriate for handling environments with extensive highly dynamic disturbances. Regardless of the level of dynamism, the method presented in this paper significantly enhances the precision of pose estimation and the overall robustness of the system, demonstrating good versatility and potential for practical application.

5. Time Evaluation

In practical applications, real-time performance is a key metric for evaluating SLAM systems. We tested the average computation time of the main modules under different datasets. The experiments used the pretrained YOLOv11-seg.pt model for semantic segmentation. The results shown in Table 8 indicate that the semantic segmentation processing speed with GPU hardware acceleration meets real-time requirements (KITTI: 15.79 ms/frame, TUM: 13.36 ms/frame). For the geometric consistency check phase, we utilized the existing PnP nonlinear optimization module in ORB-SLAM3, where the process involves matching outliers from the previous frame to the current frame to achieve geometric consistency detection. This is a simple modification, but proves to be both practical and cost-effective. Therefore, the additional time at this stage comes only from the calculation of the retained outliers, with an increase in processing time of less than 1 ms.
Next, we conducted a comparison with DS-SLAM. In the geometric consistency detection phase of DS-SLAM, optical flow is used to quickly match feature points, then the fundamental matrix is computed to complete the epipolar geometry constraint process for detecting feature point anomalies. The DS-SLAM experiments were conducted on a computer equipped with an Intel i7 CPU, P4000 GPU, and 32 GB of memory, with the process taking 29.51 ms. The semantic segmentation stage of DS-SLAM [28] uses a neural network called SegNet, which takes 37.57 ms. These data are taken from the DS-SLAM team’s experimental report.
Next, we conducted a comparison with DynaSLAM. The geometric consistency detection phase in DynaSLAM uses a multi-view geometry method to detect dynamic objects by analyzing the depth and angle between feature points in RGB-D images. DynaSLAM took 235.98 ms on the fre3_walking_rpy sequence from the TUM dataset. These data were taken from the DynaSLAM team’s experimental results [29]. The semantic segmentation stage of DynaSLAM uses a Mask R-CNN neural network. He et al. [35] reported that the running time of Mask R-CNN with an Nvidia Tesla M40 GPU was 195 ms per image.
Although both DS-SLAM and DynaSLAM were evaluated based on ORB-SLAM2 and the method proposed in this paper was implemented on ORB-SLAM3, the processing time of dynamic detection stages such as geometric consistency checks and semantic segmentation remains comparable across systems. Notably, the method proposed in this paper accomplishes these tasks with significantly lower computational cost and reduced processing time.

6. Conclusions

This paper introduces a monocular ORB-SLAM3 system optimized for dynamic environments. The proposed system is aimed at improving the localization accuracy and mapping robustness of visual SLAM in dynamic scenarios. By integrating semantic segmentation with YOLOv11, morphological dilation operations, and a geometric reprojection consistency discrimination mechanism, the proposed method achieves a dual-mask approach for precise identification and effective filtering of dynamic features, helping to mitigate the interference caused by dynamic objects on traditional SLAM systems.
Experimental results conducted on the KITTI and TUM RGB-D datasets demonstrate that the proposed method improves global trajectory accuracy across various dynamic scenes. Employing a dynamic object elimination mechanism reduces the influence of dynamic targets on trajectory estimation. In particular, this approach exhibits superiority over traditional methods in highly dynamic scenarios, where it helps to improve robustness. However, a notable limitation of this method is that the improvement in efficiency is relatively marginal, meaning that it may lead to a slight decrease in accuracy in scenarios with a large number of potentially dynamic objects but few truly dynamic ones. These issues primarily stem from misclassifications of dynamic objects, where some static feature points are incorrectly identified as dynamic during the geometric consistency verification phase. In contrast, the dual-mask approach proposed in this study demonstrates enhanced robustness; even when dynamic points are mistakenly detected, removal is confined to local regions rather than entire objects, helping to preserve the overall stability and robustness of the system. Overall, the proposed system shows improvements in both ATE and RPE, effectively overcoming the cumulative errors associated with mismatches in dynamic environments. This provides strong support for high-precision localization and mapping in practical applications such as autonomous driving and robotic navigation.
In terms of time and computational cost, classical dynamic SLAM methods typically introduce additional geometric constraints prior to the feature matching stage to determine whether candidate regions produced by instance segmentation truly correspond to dynamic objects. For example, DS-SLAM employs optical flow to implement a simplified geometric consistency check based on ORB-SLAM2, offering a lightweight mechanism to distinguish dynamic objects. On the other hand, DynaSLAM utilizes depth information to perform a more complex multi-view geometric analysis for dynamic object detection. In contrast, the method proposed in this paper leverages the existing optimization process within the SLAM pipeline to identify dynamic points, thereby avoiding additional computational overhead while maintaining reliable outlier detection capability. This leads to improved overall computational efficiency as well as better accuracy.
Future research should consider further optimizing the system’s real-time performance and local motion models to enhance its stability and accuracy in highly dynamic environments. The method proposed in this paper not only offers a superior solution for SLAM systems operating in dynamic environments but also provides important guidance and insights for future research and applications of SLAM systems.

Author Contributions

Conceptualization, Z.L., C.J. and Y.S.; methodology, Z.L. and C.J.; software, C.J. and L.H.; validation, C.J.; data curation, C.J.; writing—original draft preparation, C.J. and H.S.; writing—review and editing, Z.L., C.J. and Y.S.; supervision, Z.L. and Y.S.; project administration, Z.L. and Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guangxi Science and Technology Major Project for Laboratory Innovation Capacity Construction, grant number AD25069071. The APC was funded by the same project.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The data can be found here: TUM dataset: https://cvg.cit.tum.de/data/datasets/rgbd-dataset (accessed on 18 June 2025); KITTI dataset: https://www.cvlibs.net/datasets/kitti/eval_odometry.php (accessed on 18 June 2025).

Conflicts of Interest

Author Zhanrong Li was employed by the company Nanning Huishi Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Thrun, S. Probabilistic robotics. Commun. ACM 2002, 45, 52–57. [Google Scholar] [CrossRef]
  2. Huang, S.; Dissanayake, G. A critique of current developments in simultaneous localization and mapping. Int. J. Adv. Robot. Syst. 2016, 13, 1729881416669482. [Google Scholar] [CrossRef]
  3. Yang, D.; Bi, S.; Wang, W.; Yuan, C.; Wang, W.; Qi, X.; Cai, Y. DRE-SLAM: Dynamic RGB-D Encoder SLAM for a Differential-Drive Robot. Remote Sens. 2019, 11, 380. [Google Scholar] [CrossRef]
  4. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  5. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  6. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.M.; D. Tardós, J. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  7. Engel, J.; Koltun, V.; Cremers, D. Direct Sparse Odometry. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 611–625. [Google Scholar] [CrossRef]
  8. Lianos, K.N.; Schonberger, J.L.; Pollefeys, M.; Sattler, T. Vso: Visual semantic odometry. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 234–250. [Google Scholar]
  9. Engel, J.; Schöps, T.; Cremers, D. LSD-SLAM: Large-Scale Direct Monocular SLAM. In Computer Vision–ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Cham, Switzerland, 2014; pp. 834–849. [Google Scholar]
  10. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. Trans. Rob. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  11. Zhong, M.; Hong, C.; Jia, Z.; Wang, C.; Wang, Z. DynaTM-SLAM: Fast filtering of dynamic feature points and object-based localization in dynamic indoor environments. Robot. Auton. Syst. 2024, 174, 104634. [Google Scholar] [CrossRef]
  12. Chen, Y.; Lu, T. SGD-SLAM: A visual SLAM system with a dynamic feature rejection strategy combining semantic and geometric information for dynamic environments. Meas. Sci. Technol. 2025, 36, 026305. [Google Scholar] [CrossRef]
  13. Liang, R.; Yuan, J.; Kuang, B.; Liu, Q.; Guo, Z. DIG-SLAM: An accurate RGB-D SLAM based on instance segmentation and geometric clustering for dynamic indoor scenes. Meas. Sci. Technol. 2023, 35, 015401. [Google Scholar] [CrossRef]
  14. Yan, L.; Hu, X.; Zhao, L.; Chen, Y.; Wei, P.; Xie, H. DGS-SLAM: A Fast and Robust RGBD SLAM in Dynamic Environments Combined by Geometric and Semantic Information. Remote Sens. 2022, 14, 795. [Google Scholar] [CrossRef]
  15. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-Time Single Camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef]
  16. Klein, G.; Murray, D. Parallel Tracking and Mapping for Small AR Workspaces. In Proceedings of the 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, Nara, Japan, 13–16 November 2007; pp. 225–234. [Google Scholar] [CrossRef]
  17. Mur-Artal, R.; Tardós, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. Trans. Rob. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  18. Fu, D.; Xia, H.; Qiao, Y. Monocular Visual-Inertial Navigation for Dynamic Environment. Remote Sens. 2021, 13, 1610. [Google Scholar] [CrossRef]
  19. Dai, W.; Zhang, Y.; Li, P.; Fang, Z.; Scherer, S. RGB-D SLAM in Dynamic Environments Using Point Correlations. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 373–389. [Google Scholar] [CrossRef]
  20. Wang, R.; Wan, W.; Wang, Y.; Di, K. A New RGB-D SLAM Method with Moving Object Detection for Dynamic Indoor Scenes. Remote Sens. 2019, 11, 1143. [Google Scholar] [CrossRef]
  21. Kim, D.H.; Kim, J.H. Effective Background Model-Based RGB-D Dense Visual Odometry in a Dynamic Environment. Trans. Rob. 2016, 32, 1565–1573. [Google Scholar] [CrossRef]
  22. Li, S.; Lee, D. RGB-D SLAM in Dynamic Environments Using Static Point Weighting. IEEE Robot. Autom. Lett. 2017, 2, 2263–2270. [Google Scholar] [CrossRef]
  23. Sun, Y.; Liu, M.; Meng, M.Q.H. Motion removal for reliable RGB-D SLAM in dynamic environments. Robot. Auton. Syst. 2018, 108, 115–128. [Google Scholar] [CrossRef]
  24. Scona, R.; Jaimez, M.; Petillot, Y.R.; Fallon, M.; Cremers, D. StaticFusion: Background Reconstruction for Dense RGB-D SLAM in Dynamic Environments. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 3849–3856. [Google Scholar] [CrossRef]
  25. He, L.; Li, S.; Qiu, J.; Zhang, C. DIO-SLAM: A Dynamic RGB-D SLAM Method Combining Instance Segmentation and Optical Flow. Sensors 2024, 24. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, Z.; Zhang, Q.; Li, J.; Zhang, S.; Liu, J. A Computationally Efficient Semantic SLAM Solution for Dynamic Scenes. Remote Sens. 2019, 11, 1363. [Google Scholar] [CrossRef]
  27. Tateno, K.; Tombari, F.; Laina, I.; Navab, N. Cnn-slam: Real-time dense monocular slam with learned depth prediction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 6243–6252. [Google Scholar]
  28. Yu, C.; Liu, Z.; Liu, X.J.; Xie, F.; Yang, Y.; Wei, Q.; Fei, Q. DS-SLAM: A semantic visual SLAM towards dynamic environments. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1168–1174. [Google Scholar]
  29. Bescos, B.; Fácil, J.M.; Civera, J.; Neira, J. DynaSLAM: Tracking, Mapping, and Inpainting in Dynamic Scenes. IEEE Robot. Autom. Lett. 2018, 3, 4076–4083. [Google Scholar] [CrossRef]
  30. Zhang, J.; Henein, M.; Mahony, R.; Ila, V. VDO-SLAM: A Visual Dynamic Object-aware SLAM System. arXiv 2020, arXiv:2005.11052. [Google Scholar] [CrossRef]
  31. Zhong, F.; Wang, S.; Zhang, Z.; Chen, C.; Wang, Y. Detect-SLAM: Making Object Detection and SLAM Mutually Beneficial. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1001–1010. [Google Scholar] [CrossRef]
  32. Liu, Y.; Miura, J. RDS-SLAM: Real-Time Dynamic SLAM Using Semantic Segmentation Methods. IEEE Access 2021, 9, 23772–23785. [Google Scholar] [CrossRef]
  33. Peng, J.; Qian, W.; Zhang, H. PPS-SLAM: Dynamic Visual SLAM with a Precise Pruning Strategy. Comput. Mater. Contin. 2025, 82, 2849–2868. [Google Scholar] [CrossRef]
  34. Ultralytics. Ultralytics YOLO: State-of-the-art Deep Learning for Object Detection. 2025. Available online: https://github.com/ultralytics/ultralytics (accessed on 8 May 2025).
  35. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
Figure 1. System Overview. This figure illustrates the main components and workflow of the proposed system. The core modules of the SLAM process are indicated by solid lines, while the dashed lines highlight the key aspects of the proposed research involving dynamic feature point processing and pose optimization. This design ensures that the proposed system can effectively discern and eliminate unstable feature points when facing dynamic environments, enhancing both overall localization and mapping quality.
Figure 1. System Overview. This figure illustrates the main components and workflow of the proposed system. The core modules of the SLAM process are indicated by solid lines, while the dashed lines highlight the key aspects of the proposed research involving dynamic feature point processing and pose optimization. This design ensures that the proposed system can effectively discern and eliminate unstable feature points when facing dynamic environments, enhancing both overall localization and mapping quality.
Applsci 15 07095 g001
Figure 2. Dynamic area extraction process flowchart: the blue region indicates the final mask; in the scene, the person on the left exhibits head shaking and arm movement to operate the mouse, while the person on the right steps forward with the left foot, leading to motion on the left side of the body.
Figure 2. Dynamic area extraction process flowchart: the blue region indicates the final mask; in the scene, the person on the left exhibits head shaking and arm movement to operate the mouse, while the person on the right steps forward with the left foot, leading to motion on the left side of the body.
Applsci 15 07095 g002
Figure 3. ATE and RPE results for SLAM systems on the fre3_walking_rpy dataset.
Figure 3. ATE and RPE results for SLAM systems on the fre3_walking_rpy dataset.
Applsci 15 07095 g003
Figure 4. Map of ATE and RPE for SLAM systems on the fre3_walking_rpy dataset.
Figure 4. Map of ATE and RPE for SLAM systems on the fre3_walking_rpy dataset.
Applsci 15 07095 g004
Figure 5. ATE and RPE results for SLAM systems on the fre3_sitting_xyz dataset.
Figure 5. ATE and RPE results for SLAM systems on the fre3_sitting_xyz dataset.
Applsci 15 07095 g005
Figure 6. Map of ATE and RPE for SLAM systems on the fre3_sitting_xyz dataset.
Figure 6. Map of ATE and RPE for SLAM systems on the fre3_sitting_xyz dataset.
Applsci 15 07095 g006
Table 1. Comparative evaluation of Absolute Trajectory Error (ATE) for monocular SLAM systems.
Table 1. Comparative evaluation of Absolute Trajectory Error (ATE) for monocular SLAM systems.
SequenceORB-SLAM3System(D)System(D+O)
RMSE Mean Std RMSE Mean Std RMSE Mean Std
KITTI 006.245.433.056.585.593.446.145.353.09
KITTI 01295.24264.75129.86232.57205.63108.08269.23241.76118.03
KITTI 0224.4719.1715.2120.0016.0313.9121.6616.1614.01
KITTI 030.790.550.561.140.930.650.860.690.52
KITTI 041.561.360.761.010.850.551.191.000.64
KITTI 056.325.882.316.465.972.465.645.192.22
KITTI 0614.8612.917.3516.0414.007.8515.7713.817.62
KITTI 072.942.161.772.472.250.992.441.901.41
KITTI 0864.3850.0140.5461.3947.0939.3759.9646.1838.25
KITTI 099.038.094.008.877.933.968.867.904.00
KITTI 108.156.634.728.036.544.657.576.294.24
Table 2. Comparative evaluation of Relative Pose Error (RPE) for monocular SLAM systems.
Table 2. Comparative evaluation of Relative Pose Error (RPE) for monocular SLAM systems.
SequenceORB-SLAM3System(D)System(D+O)
RMSE Mean Std RMSE Mean Std RMSE Mean Std
KITTI 000.1860.0960.1540.3390.1000.3060.1440.1090.109
KITTI 012.0261.9020.9482.0791.7511.3021.9621.7230.869
KITTI 020.1630.1620.1030.1510.1170.0950.1470.1270.097
KITTI 030.0240.0190.0140.0300.0230.0200.0250.0200.016
KITTI 040.0580.0480.0330.0470.0390.0260.0540.0440.032
KITTI 050.3210.2240.2300.3760.2430.2860.3460.2420.247
KITTI 060.4090.2530.3340.4030.2410.3210.4990.2580.327
KITTI 071.8580.2641.8140.3450.1990.2790.9890.2451.439
KITTI 080.3550.2840.2130.3430.2710.2100.3390.2680.208
KITTI 090.2480.2480.1990.2520.1500.2030.2420.1440.195
KITTI 100.0670.0670.0470.0670.0480.0480.0620.0460.041
Table 3. Impact analysis of potential dynamic objects on SLAM system performance.
Table 3. Impact analysis of potential dynamic objects on SLAM system performance.
SequenceStatic ObjectsDynamic ObjectsATE (RMSE)RPE (Mean)
KITTI 00ManyFewDO< 3 < D3 < D < DO
KITTI 01NoneManyD < DO < 3DO < D < 3
KITTI 02ManyMediumD < DO < 3D < DO < 3
KITTI 03FewFew3 < DO < D3 < DO < D
KITTI 04NoneManyD < DO < 3D < DO < 3
KITTI 05ManyMediumDO < D < 33 < D = DO
KITTI 06ManyFew3 < DO < DD < 3 < DO
KITTI 07ManyManyDO < D < 3D < DO < 3
KITTI 08ManyMediumDO < D < 3DO < D < 3
KITTI 09ManyMediumDO < D < 3DO < D < 3
KITTI 10MediumFewDO < D < 3DO < D < 3
Table 4. Comparative performance improvement rates of ORB-SLAM3.
Table 4. Comparative performance improvement rates of ORB-SLAM3.
SequenceSystem(D)System(D+O)
ATE (RMSE) RPE (Mean) ATE (RMSE) RPE (Mean)
KITTI 001.63%−13.54%−5.47%−4.30%
KITTI 018.81%9.43%21.23%7.94%
KITTI 0211.48%21.35%18.27%28.04%
KITTI 03−9.12%−4.61%−44.76%−16.70%
KITTI 0423.50%7.63%34.80%17.44%
KITTI 0510.74%−8.08%−2.16%−8.37%
KITTI 06−6.14%−2.09%−7.93%4.55%
KITTI 0716.86%7.26%15.84%24.50%
KITTI 086.86%5.58%4.64%4.67%
KITTI 091.92%41.98%1.82%39.60%
KITTI 107.03%30.67%1.46%28.28%
Table 5. Comparative evaluation of Absolute Trajectory Error (ATE) for monocular SLAM systems.
Table 5. Comparative evaluation of Absolute Trajectory Error (ATE) for monocular SLAM systems.
SequenceORB-SLAM3System(D)System(D+O)
RMSE Mean Std RMSE Mean Std RMSE Mean Std
s_halfsphere0.0260.0230.0110.0250.0230.0100.0270.0260.008
s_static0.3230.3230.0060.3120.3120.0030.2850.2850.003
s_xyz0.0310.0260.0170.0270.0210.0170.0240.0190.015
w_halfsphere0.0440.0380.0230.0320.0270.0170.0300.0260.014
w_rpy0.4550.4540.0330.1370.1280.0480.1380.1380.011
w_static0.7300.7300.0030.3680.3680.0030.6400.6400.003
Table 6. Comparative evaluation of Relative Pose Error (RPE) for monocular SLAM systems.
Table 6. Comparative evaluation of Relative Pose Error (RPE) for monocular SLAM systems.
SequenceORB-SLAM3System(D)System(D+O)
RMSE Mean Std RMSE Mean Std RMSE Mean Std
s_halfsphere0.0140.0120.0080.0150.0120.0100.0130.0110.007
s_static0.0100.0060.0080.0100.0050.0080.0080.0040.006
s_xyz0.0180.0130.0130.0170.0120.0120.0160.0120.011
w_halfsphere0.0310.0180.0260.0200.0150.0120.0200.0150.012
w_rpy0.0190.0140.0120.0190.0110.0160.0200.0140.015
w_static0.0080.0050.0060.0090.0060.0070.0100.0050.006
Table 7. Comparative performance improvement rates of ORB-SLAM3.
Table 7. Comparative performance improvement rates of ORB-SLAM3.
SequenceSystem(D)System(D+O)
ATE (RMSE) RPE (Mean) ATE (RMSE) RPE (Mean)
s_halfsphere2.86%−4.42%−6.05%0.56%
s_static3.39%2.05%11.65%21.65%
s_xyz14.11%6.47%24.29%6.35%
w_halfsphere28.50%14.60%32.28%14.60%
w_rpy69.87%20.73%69.54%2.88%
w_static49.56%−14.00%12.27%3.57%
Table 8. Time evaluation.
Table 8. Time evaluation.
DatasetsSemantic SegmentationGeometric
Consistency Check
CPUGPU
KITTI360.78 ms15.79 ms0.78 ms
TUM317.87 ms13.36 ms0.83 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Jiang, C.; Sun, Y.; Su, H.; He, L. Precise Feature Removal Method Based on Semantic and Geometric Dual Masks in Dynamic SLAM. Appl. Sci. 2025, 15, 7095. https://doi.org/10.3390/app15137095

AMA Style

Li Z, Jiang C, Sun Y, Su H, He L. Precise Feature Removal Method Based on Semantic and Geometric Dual Masks in Dynamic SLAM. Applied Sciences. 2025; 15(13):7095. https://doi.org/10.3390/app15137095

Chicago/Turabian Style

Li, Zhanrong, Chao Jiang, Yu Sun, Haosheng Su, and Longning He. 2025. "Precise Feature Removal Method Based on Semantic and Geometric Dual Masks in Dynamic SLAM" Applied Sciences 15, no. 13: 7095. https://doi.org/10.3390/app15137095

APA Style

Li, Z., Jiang, C., Sun, Y., Su, H., & He, L. (2025). Precise Feature Removal Method Based on Semantic and Geometric Dual Masks in Dynamic SLAM. Applied Sciences, 15(13), 7095. https://doi.org/10.3390/app15137095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop