Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,525)

Search Parameters:
Keywords = slamming

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 894 KB  
Review
Indoor Mapping as a Spatiotemporal Framework for Mitigating Greenhouse Gas Emissions in Buildings: A Review
by Vinuri Nilanika Goonetilleke, Muditha K. Heenkenda and Kamil Zaniewski
Geomatics 2026, 6(2), 27; https://doi.org/10.3390/geomatics6020027 - 19 Mar 2026
Abstract
Climate change is a critical global challenge, and the building sector accounts for nearly 30% of global greenhouse gas (GHG) emissions, remaining a key target for mitigation. Indoor environments contribute significantly to GHG emissions, primarily through heating, cooling, lighting, and occupant-driven energy use. [...] Read more.
Climate change is a critical global challenge, and the building sector accounts for nearly 30% of global greenhouse gas (GHG) emissions, remaining a key target for mitigation. Indoor environments contribute significantly to GHG emissions, primarily through heating, cooling, lighting, and occupant-driven energy use. Indoor mapping, serving as the foundation for Digital Twins (DTs), provides a spatiotemporal framework that integrates sensor data with Building Information Modelling (BIM), Geographic Information Systems (GIS), and Internet of Things (IoT) to support energy-efficient, low-carbon building operations. This review examined the role of indoor mapping in understanding, modelling, and reducing GHG emissions in buildings. It synthesized current advancements in indoor spatial data acquisition, ranging from Light Detection And Ranging (LiDAR) and Simultaneous Localization and Mapping (SLAM) to deep learning-based floor plan extraction, and evaluated their contribution to improved indoor environmental analysis. The review highlighted emerging techniques, challenges, and gaps, particularly the limited integration of physical indoor spaces with virtual layers representing assets, occupants, and equipment. Addressing this gap requires embedding spatial modelling as an intermediate analytical layer that structures and contextualizes sensor data to support spatiotemporal decision-making. Overall, this review demonstrated that indoor mapping plays a critical role in transforming spatial information into actionable insights, enabling more accurate energy modelling, enhanced real-time building management, and stronger data-driven strategies for GHG mitigation in the built environment. Full article
Show Figures

Graphical abstract

23 pages, 6668 KB  
Article
Development of a Visual SLAM-Based Autonomous UAV System for Greenhouse Plant Monitoring
by Jing-Heng Lin and Ta-Te Lin
Drones 2026, 10(3), 205; https://doi.org/10.3390/drones10030205 - 15 Mar 2026
Abstract
Autonomous monitoring is essential for precision agriculture in greenhouses, yet deploying unmanned aerial vehicles (UAVs) in confined, GPS-denied environments remains limited by payload, power, and cost constraints. This study developed and validated an autonomous UAV system for reliable, low-cost operation in such conditions. [...] Read more.
Autonomous monitoring is essential for precision agriculture in greenhouses, yet deploying unmanned aerial vehicles (UAVs) in confined, GPS-denied environments remains limited by payload, power, and cost constraints. This study developed and validated an autonomous UAV system for reliable, low-cost operation in such conditions. The proposed system employs a dual-link edge-computing architecture: a lightweight onboard controller handles flight control and sensor acquisition, while visual simultaneous localization and mapping (V-SLAM) is offloaded to an edge computer via the FPV video link. Phenotyping (flower detection and tracking/counting) is performed offline from the side-view RGB stream and does not participate in the flight control loop. Using muskmelon (Cucumis melo L.) flower development as a case study, the UAV autonomously executed daily missions for 27 days in a commercial greenhouse, performing flower detection and tracking to monitor phenological dynamics. Localization and control accuracy were evaluated against a validated UWB reference system, achieving 5.4~8.0 cm 2D RMSE for trajectory tracking and 12.7 cm translation RMSE for greenhouse mapping. This work demonstrates a practical architecture for autonomous monitoring in GPS-denied agricultural environments, with operational boundaries characterized through the sustained field deployment. The system’s design principles may extend to other indoor or communication-limited scenarios requiring lightweight, intelligent robotic operation. Full article
(This article belongs to the Section Drones in Agriculture and Forestry)
Show Figures

Figure 1

17 pages, 1708 KB  
Article
Robust Visual–Inertial SLAM and Biomass Assessment for AUVs in Marine Ranching
by Yangyang Wang, Ziyu Liu, Tianzhu Gao and Xijun Du
Symmetry 2026, 18(3), 495; https://doi.org/10.3390/sym18030495 - 13 Mar 2026
Viewed by 58
Abstract
Environmental perception is a cornerstone for autonomous underwater vehicles (AUVs) to achieve robust self-localization and scene understanding, which are pivotal for the intelligent management of marine ranching. However, underwater image degradation and weak-textured scenes significantly hinder reliable self-localization and fine-grained environmental perception. To [...] Read more.
Environmental perception is a cornerstone for autonomous underwater vehicles (AUVs) to achieve robust self-localization and scene understanding, which are pivotal for the intelligent management of marine ranching. However, underwater image degradation and weak-textured scenes significantly hinder reliable self-localization and fine-grained environmental perception. To address the perceptual asymmetry arising from these challenges, this paper proposes a robust visual–inertial simultaneous localization and mapping (SLAM) and biomass assessment scheme for marine ranching. Specifically, we first propose a robust tightly coupled underwater visual–inertial localization scheme, which leverages a multi-sensor fusion strategy to solve the image degradation problem of localization in complex underwater environments. Furthermore, we propose a novel underwater scene perception method, which enables the simultaneous visual reconstruction of aquaculture species and the quantitative mapping of their spatial distribution in marine ranching. Finally, we develop a low-cost, agile, and portable multisensor-integrated system that consolidates autonomous localization and aquaculture biomass assessment modules, with its performance validated through extensive real-world underwater experiments. The experimental results demonstrate that the proposed methods can effectively overcome the interference of complex underwater environments and provide high-precision perception support for both AUV state estimation and aquaculture asset management. Full article
(This article belongs to the Special Issue Symmetry in Next-Generation Intelligent Information Technologies)
Show Figures

Figure 1

20 pages, 24767 KB  
Article
VINA-SLAM: A Voxel-Based Inertial and Normal-Aligned LiDAR–IMU SLAM
by Ruyang Zhang and Bingyu Sun
Sensors 2026, 26(6), 1810; https://doi.org/10.3390/s26061810 - 13 Mar 2026
Viewed by 160
Abstract
Environments with sparse or repetitive geometric structures, such as long corridors and narrow stairwells, remain challenging for LiDAR–inertial simultaneous localization and mapping (LiDAR–IMU SLAM) due to insufficient geometric observability and unreliable data associations. To address these issues, we propose VINA-SLAM, a novel LiDAR–IMU [...] Read more.
Environments with sparse or repetitive geometric structures, such as long corridors and narrow stairwells, remain challenging for LiDAR–inertial simultaneous localization and mapping (LiDAR–IMU SLAM) due to insufficient geometric observability and unreliable data associations. To address these issues, we propose VINA-SLAM, a novel LiDAR–IMU SLAM framework that constructs a unified global voxel map to explicitly exploit structural consistency. VINA-SLAM continuously tracks surface normals stored in the global voxel map using a normal-guided correspondence strategy, enabling stable scan-to-map alignment in degenerate scenes. Furthermore, a tangent-space metric is introduced to supplement missing rotational constraints around planar regions, providing reliable initial pose estimates for local optimization. A tightly coupled sliding-window bundle adjustment is then formulated by jointly incorporating IMU factors, voxel normal consistency factors, and planar regularization terms. In particular, the minimum eigenvalue of each voxel’s covariance is used as a statistically principled planar constraint, improving the Hessian conditioning and cross-view geometric consistency. The proposed system directly aligns raw LiDAR scans to the voxelized map without explicit feature extraction or loop closure. Experiments on 25 sequences from the HILTI and MARS-LVIG datasets show that VINA-SLAM reduces ATE by 25–40% on average while maintaining real-time performance at 10 Hz in the evaluated geometrically degenerate environments. Full article
Show Figures

Figure 1

30 pages, 3812 KB  
Review
Video-Based 3D Reconstruction: A Review of Photogrammetry and Visual SLAM Approaches
by Ali Javadi Moghadam, Abbas Kiani, Reza Naeimaei, Shirin Malihi and Ioannis Brilakis
J. Imaging 2026, 12(3), 128; https://doi.org/10.3390/jimaging12030128 - 13 Mar 2026
Viewed by 121
Abstract
Three-dimensional (3D) reconstruction using images is one of the most significant topics in computer vision and photogrammetry, with wide-ranging applications in robotics, augmented reality, and mapping. This study investigates methods of 3D reconstruction using video (especially monocular video) data and focuses on techniques [...] Read more.
Three-dimensional (3D) reconstruction using images is one of the most significant topics in computer vision and photogrammetry, with wide-ranging applications in robotics, augmented reality, and mapping. This study investigates methods of 3D reconstruction using video (especially monocular video) data and focuses on techniques such as Structure from Motion (SfM), Multi-View Stereo (MVS), Visual Simultaneous Localization and Mapping (V-SLAM), and videogrammetry. Based on a statistical analysis of SCOPUS records, these methods collectively account for approximately 6863 journal publications up to the end of 2024. Among these, about 80 studies are analyzed in greater detail to identify trends and advancements in the field. The study also shows that the use of video data for real-time 3D reconstruction is commonly addressed through two main approaches: photogrammetry-based methods, which rely on precise geometric principles and offer high accuracy at the cost of greater computational demand; and V-SLAM methods, which emphasize real-time processing and provide higher speed. Furthermore, the application of IMU data and other indicators, such as color quality and keypoint detection, for selecting suitable frames for 3D reconstruction is investigated. Overall, this study compiles and categorizes video-based reconstruction methods, emphasizing the critical step of keyframe extraction. By summarizing and illustrating the general approaches, the study aims to clarify and facilitate the entry path for researchers interested in this area. Finally, the paper offers targeted recommendations for improving keyframe extraction methods to enhance the accuracy and efficiency of real-time video-based 3D reconstruction, while also outlining future research directions in addressing challenges like dynamic scenes, reducing computational costs, and integrating advanced learning-based techniques. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

26 pages, 4902 KB  
Article
Multi-Sensor-Assisted Navigation for UAVs in Power Inspection: A Fusion Approach Using LiDAR, IMU and GPS
by Anjun Wang, Wenbin Yu, Xuexing Dong, Yang Yang, Shizeng Liu, Jiahao Liu and Hongwei Mei
Appl. Sci. 2026, 16(6), 2632; https://doi.org/10.3390/app16062632 - 10 Mar 2026
Viewed by 131
Abstract
High-precision localization is essential for autonomous navigation and environment perception of unmanned aerial vehicles (UAVs) in complex power inspection scenarios. To overcome the limited accuracy and accumulated drift of conventional GPS-based single-sensor localization, this paper proposes a LiDAR–IMU–GPS-aided navigation method that combines a [...] Read more.
High-precision localization is essential for autonomous navigation and environment perception of unmanned aerial vehicles (UAVs) in complex power inspection scenarios. To overcome the limited accuracy and accumulated drift of conventional GPS-based single-sensor localization, this paper proposes a LiDAR–IMU–GPS-aided navigation method that combines a tightly coupled front-end and a loosely coupled back-end. The front-end employs an improved Lie-group-based UKF-SLAM framework to explicitly handle the nonlinearities of rotational motion, thereby improving the stability of local pose estimation. The back-end integrates GPS absolute constraints, loop closure detection, and point cloud registration via pose graph optimization, which effectively suppresses long-term accumulated drift. The framework achieves accurate and robust localization for UAV power inspection. Experiments on public benchmark datasets and real-world power inspection scenarios demonstrate the effectiveness of the proposed method. On the MH_02_easy sequence, the absolute trajectory error is reduced from 0.521 m to 0.170 m compared with ROVIO, while in a real inspection sequence the cumulative error is reduced by more than 99% after back-end optimization. Moreover, the system maintains stable navigation under GPS-degraded conditions, indicating strong robustness and practical applicability. Full article
Show Figures

Figure 1

25 pages, 6915 KB  
Article
EXAONE-VLA: A Unified Vision–Language Framework for Mobile Manipulation via Semantic Topology and Hierarchical LLM Reasoning
by Jeong-Seop Park, Yong-Jun Lee, Jong-Chan Park, Sung-Gil Park, Jong-Jin Woo and Myo-Taeg Lim
Appl. Sci. 2026, 16(5), 2600; https://doi.org/10.3390/app16052600 - 9 Mar 2026
Viewed by 219
Abstract
This paper proposes a unified vision–language framework that translates user instructions into navigation for the mobile base and actions for the manipulator in indoor environments. In general, occupancy grid maps constructed via SLAM capture solely the geometric layout of the environment. This renders [...] Read more.
This paper proposes a unified vision–language framework that translates user instructions into navigation for the mobile base and actions for the manipulator in indoor environments. In general, occupancy grid maps constructed via SLAM capture solely the geometric layout of the environment. This renders the robot incapable of leveraging the semantic information required for object distinction. The proposed method encodes semantic information from vision–language models and the robot’s pose in a textual format, referred to as a semantic topological graph. Specifically, the models including GroundingDINO, LG EXAONE, and SAM2 extract object-level semantic information, which is subsequently used to identify room characteristics. A large language model then interprets user instructions to identify the final destination for navigation within the semantic topological graph, followed by reasoning to determine the suitable action network. Notably, the proposed text-based representation facilitates a substantial reduction in inference time, and its effectiveness is validated through real-world experiments. Full article
(This article belongs to the Special Issue Deep Reinforcement Learning for Multiagent Systems)
Show Figures

Figure 1

28 pages, 6157 KB  
Article
RI-DVP: A Physics–Geometry Dual-Driven Framework for Static Map Construction in Sparse LiDAR Scenarios
by Xiaokai Li, Li Wang, Haolong Luo and Guangyun Li
Remote Sens. 2026, 18(5), 821; https://doi.org/10.3390/rs18050821 - 6 Mar 2026
Viewed by 223
Abstract
High-fidelity static map construction is essential for reliable autonomous navigation, yet dynamic environments introduce severe artifacts caused by moving objects (also referred to as dynamic artifacts) in accumulated maps. While geometry-based methods perform well on dense point clouds, their performance notably degrades on [...] Read more.
High-fidelity static map construction is essential for reliable autonomous navigation, yet dynamic environments introduce severe artifacts caused by moving objects (also referred to as dynamic artifacts) in accumulated maps. While geometry-based methods perform well on dense point clouds, their performance notably degrades on sparse 16-beam LiDAR due to the “Sparsity Trap”: dynamic objects are frequently missed by ray-based geometry, and purely geometric cues fail in radiometrically ambiguous scenarios. To address this, we propose RI-DVP, a physics–geometry dual-driven framework. Unlike conventional approaches, RI-DVP first performs a physics-inspired radiometric normalization that compensates for range attenuation and incidence-angle effects to establish a consistent signal baseline. Subsequently, a Dual-Residual Aggressive Removal (DRAR) module jointly exploits geometric residuals—bounded by a range-dependent spatial uncertainty envelope—and calibrated intensity residuals to detect geometrically indistinguishable objects. To balance recall and precision, a Hierarchical Static Reversion strategy (HSR) employs two-stage recovery to retrieve large-scale structures and correct fine-grained artifacts via topology-based adhesion reasoning. Experiments on SemanticKITTI and custom sparse datasets demonstrate that RI-DVP outperforms state-of-the-art geometric baselines, improving Dynamic Accuracy by over 36 percentage points in sparse scanning scenarios using a VLP-16 LiDAR sensor (Velodyne Acoustics, Inc., Morgan Hill, CA, USA) compared to baselines that fail under the sparsity trap while achieving real-time performance at approximately 15.3 Hz. Full article
(This article belongs to the Special Issue LiDAR Technology for Autonomous Navigation and Mapping)
Show Figures

Figure 1

33 pages, 12968 KB  
Article
Tunnel-SLAM: Low-Cost LiDAR/Vision/RTK/Inertial Integration on Vehicles for Roadway Tunnels
by Zeyu Li, Xian Wu, Jianhui Cui, Ying Xu, Rufei Liu, Rui Tu and Wei Jiang
Electronics 2026, 15(5), 1101; https://doi.org/10.3390/electronics15051101 - 6 Mar 2026
Viewed by 242
Abstract
Reliable positioning and mapping in roadway tunnels are crucial for vehicle-based monitoring and inspection, especially considering the challenging environmental conditions such as rapidly changing illumination, low-texture environments, and repetitive structural elements. While general LiDAR-inertial odometry (LIO) frameworks and loop-closure detection methods are effective [...] Read more.
Reliable positioning and mapping in roadway tunnels are crucial for vehicle-based monitoring and inspection, especially considering the challenging environmental conditions such as rapidly changing illumination, low-texture environments, and repetitive structural elements. While general LiDAR-inertial odometry (LIO) frameworks and loop-closure detection methods are effective in general scenarios, they often suffer from severe drift or incorrect loop constraints under these specific conditions. These challenges are further exacerbated by the inherent uncertainties associated with low-cost sensors. This paper introduces a narrow field-of-view LiDAR-centric RTK-visual-inertial SLAM system enhanced by three key modules: semantic-assisted loop detection and matching, two-stage RTK quality control, and adaptive factor graph optimization (FGO). In the first module, the proposed semantic loop descriptor (SLD) matching is used to determine the potential loop closure locations and then integrates the corresponding constraint as graph nodes. The quality control module addresses RTK outlier rejection during tunnel entry and exit, employing an event-driven stochastic model to characterize the uncertainty between RTK and the other sensors, effectively suppressing RTK-induced errors. FGO module performs optimization by incorporating LIO, RTK, and loop closure factors, employing a keyframe-based strategy to produce globally optimized poses while continuously updating the map. The proposed Tunnel-SLAM was evaluated against state-of-the-art SLAM algorithms in four extended roadway tunnels, ranging in traveling distance approximately from 5 to 10 km. Experimental results demonstrate that the proposed SLAM achieved a final drift of less than 2 m with loop closure, demonstrating significantly reducing the drift, while other existing SLAM frameworks fail catastrophically or have large drift. Full article
(This article belongs to the Special Issue Simultaneous Localization and Mapping (SLAM) of Mobile Robots)
Show Figures

Figure 1

23 pages, 9498 KB  
Article
Interdisciplinary Analysis of Water UBH: The Palombaro Purgatorio Vecchio Infrastructure in Matera
by Daniele Altamura, Giandamiano Fiore, Angelarosa Manicone, Enrico Lamacchia, Arcangelo Priore, Nicola Masini, Ruggero Ermini, Antonella Guida and Graziella Bernardo
Heritage 2026, 9(3), 102; https://doi.org/10.3390/heritage9030102 - 4 Mar 2026
Viewed by 233
Abstract
Historical water management infrastructures, often comprising underground environments, represent a significant example of the interplay between built heritage and the natural substrate. This study proposes an interdisciplinary, integrated and multi-scalar investigative methodology for such structures. Through the analysis of the case study of [...] Read more.
Historical water management infrastructures, often comprising underground environments, represent a significant example of the interplay between built heritage and the natural substrate. This study proposes an interdisciplinary, integrated and multi-scalar investigative methodology for such structures. Through the analysis of the case study of Palombaro Purgatoro Vecchio, a large historical public water cistern located in Matera in Italy, this paper presents a rigorous methodology replicable in different contexts. Bibliographic and archival research establish the knowledge base regarding the structure’s historical evolution; territorial and hydromorphic analyses, supported by GIS, highlight the dynamics of the surrounding watersheds. Meanwhile, a digital survey integrating SLAM and photogrammetry provides geometric-dimensional data, serving as the foundation for analysing construction techniques and materials. The selection of accessible and manageable technologies promotes a practical, replicable investigative methodology aimed at the protection, comprehension, enhancement and dissemination of water UBH. Full article
(This article belongs to the Special Issue Exploring Underground Built Heritage)
Show Figures

Figure 1

20 pages, 17849 KB  
Article
UAV–UGV Collaborative Localization in GNSS-Denied Large-Scale Environments: An Anchor-Free VIO–UWB Fusion with Adaptive Weighting and Outlier Suppression
by Haoyuan Xu, Gaopeng Zhao and Yuming Bo
Drones 2026, 10(3), 175; https://doi.org/10.3390/drones10030175 - 4 Mar 2026
Viewed by 307
Abstract
In GNSS-denied large-scale outdoor environments, UAVs and UGVs that rely solely on visual–inertial odometry (VIO) suffer from accumulated global drift as the trajectory grows. Meanwhile, inter-platform ultra-wideband (UWB) ranging exhibits unknown, time-varying noise under NLOS/multipath, rendering naïve weighting unreliable. This paper presents an [...] Read more.
In GNSS-denied large-scale outdoor environments, UAVs and UGVs that rely solely on visual–inertial odometry (VIO) suffer from accumulated global drift as the trajectory grows. Meanwhile, inter-platform ultra-wideband (UWB) ranging exhibits unknown, time-varying noise under NLOS/multipath, rendering naïve weighting unreliable. This paper presents an anchor-free collaborative localization framework for UAV–UGV teams that fuses pairwise UWB ranges (including UAV–UAV, UAV–UGV, and UGV–UGV) with onboard VIO in a factor-graph backend via a two-stage robust scheme. First, we bound VIO drift using per-agent state covariance and reject UWB outliers with a Mahalanobis gate, preventing early-stage bias when VIO is still accurate. Then, during global optimization, we adaptively estimate the Fisher information of UWB factors from measurement–state residuals, enabling online self-tuning of measurement confidence under time-varying SNR. Real-world experiments with three UAVs and two UGVs over multi-level rooftops and forest–open areas (~1.6 km2) show that, compared to an outlier-only variant, the proposed method further reduces localization RMSE by about 24.6% and maximum error by about 31.2% for both UAVs and UGVs, maintaining strong performance during long trajectories dominated by VIO drift and NLOS ranges. The approach requires no fixed anchors or GNSS and is applicable to UAV–UGV teams for disaster response, cooperative mapping/inspection, and bandwidth-limited operations. Full article
(This article belongs to the Section Artificial Intelligence in Drones (AID))
Show Figures

Figure 1

36 pages, 15804 KB  
Article
An RGB-D SLAM Algorithm Based on a Multi-Layer Refraction Model for Underwater Scenarios
by Xianshuai Sun, Yabiao Wang, Yuming Zhao, Zhigang Li, Zhen He and Xiaohui Wang
J. Mar. Sci. Eng. 2026, 14(5), 485; https://doi.org/10.3390/jmse14050485 - 3 Mar 2026
Viewed by 188
Abstract
The use of depth cameras in low-texture environments is crucial for ensuring the feasibility of visual simultaneous localization and mapping (SLAM) algorithms. Nevertheless, in underwater scenarios, light propagation through multi-layered media gives rise to refractive distortion. Directly utilizing distorted images acquired by depth [...] Read more.
The use of depth cameras in low-texture environments is crucial for ensuring the feasibility of visual simultaneous localization and mapping (SLAM) algorithms. Nevertheless, in underwater scenarios, light propagation through multi-layered media gives rise to refractive distortion. Directly utilizing distorted images acquired by depth cameras for visual SLAM computations inevitably introduces substantial errors in localization and mapping. Additionally, the waterproof glass mounted in front of the depth camera renders traditional air-based camera calibration ineffective, thereby introducing calibration inaccuracies. To mitigate these challenges, we propose a comprehensive SLAM algorithm framework for underwater multi-layered media refraction correction based on RGB-D cameras. Firstly, a multi-layer refraction calibration module is developed to calibrate the depth camera in air. Subsequently, the calibrated parameters are leveraged to construct an underwater multi-layer refraction correction module, which retrieves undistorted color images and aligned depth images. Finally, the corrected color images and depth images are fed into the front-end of the visual SLAM algorithm to generate dense point cloud maps. Both simulation and real-world experiments are conducted to validate the accuracy of the multi-layer refraction calibration results and the precision of the dense point clouds obtained via multi-layer refraction correction. Furthermore, the superiority of the proposed method is demonstrated through both qualitative and quantitative evaluations. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

30 pages, 8087 KB  
Article
A Novel SLAM Approach for Trajectory Generation of a Dual-Arm Mobile Robot (DAMR) Using Sensor Fusion
by Narendra Kumar Kolla and Pandu Ranga Vundavilli
Automation 2026, 7(2), 42; https://doi.org/10.3390/automation7020042 - 3 Mar 2026
Viewed by 202
Abstract
Simultaneous Localization and Mapping (SLAM) is essential for autonomous movement in intelligent robotic systems. Traditional SLAM using a single sensor, such as an Inertial Measurement Unit (IMU), faces challenges including noise and drift. This paper introduces a novel Cartographer-based SLAM approach for DAMR [...] Read more.
Simultaneous Localization and Mapping (SLAM) is essential for autonomous movement in intelligent robotic systems. Traditional SLAM using a single sensor, such as an Inertial Measurement Unit (IMU), faces challenges including noise and drift. This paper introduces a novel Cartographer-based SLAM approach for DAMR trajectory generation in indoor environments to reduce drift errors and improve localization accuracy. This SLAM approach integrates multi-sensor data with extended Kalman filter (EKF) fusion from wheel odometry, an RGB-D camera (RTAB-Map), and an IMU for precise mapping with DAMR trajectory generation and is compared with the heading reference trajectory generated by robot pose estimation and frame transformation. This system is implemented in the Robot Operating System (ROS 2) for coordinated data acquisition, processing, and visualization. After experimental verification, the DAMR trajectories generated are closer to the reference trajectory and drift errors are tuned. The experimental results revealed that the DAMR trajectory with multi-sensor data integration using the EKF effectively improved the positioning accuracy and robustness of the system. The proposed approach shows improved alignment with the reference trajectory, yielding a mean displacement error of 0.352% and an absolute trajectory error of 0.007 m, highlighting the effectiveness of the fusion approach for accurate indoor robot navigation. Full article
(This article belongs to the Section Robotics and Autonomous Systems)
Show Figures

Figure 1

16 pages, 2080 KB  
Article
Lidar–Vision Depth Fusion for Robust Loop Closure Detection in SLAM Systems
by Bingzhuo Liu, Panlong Wu, Rongting Chen, Yidan Zheng and Mengyu Li
Machines 2026, 14(3), 282; https://doi.org/10.3390/machines14030282 - 3 Mar 2026
Viewed by 238
Abstract
Loop Closure Detection (LCD) is a key component of Simultaneous Localization and Mapping (SLAM) systems, responsible for correcting odometric drift and maintaining global consistency in localization and mapping. However, single-modality LCD methods suffer from inherent limitations: LiDAR-based approaches are affected by point cloud [...] Read more.
Loop Closure Detection (LCD) is a key component of Simultaneous Localization and Mapping (SLAM) systems, responsible for correcting odometric drift and maintaining global consistency in localization and mapping. However, single-modality LCD methods suffer from inherent limitations: LiDAR-based approaches are affected by point cloud sparsity, limiting feature representation in unstructured environments, while vision-based methods are sensitive to illumination and weather variations, reducing robustness. To address these issues, this paper presents a LiDAR–vision multimodal fusion LCD algorithm. Spatiotemporal alignment between LiDAR point clouds and images is achieved through extrinsic calibration and timestamp interpolation to ensure cross-modal consistency. Harris corner detection and BRIEF descriptors are employed to extract visual features, and a LiDAR-projected sparse depth map is used to complete depth information, mapping 2D features into 3D space. A hybrid feature representation is then constructed by fusing LiDAR geometric triangle descriptors with visual BRIEF descriptors, enabling efficient loop candidate retrieval via hash indexing. Finally, an improved RANSAC algorithm performs geometric verification to enhance the robustness of relative pose estimation. Experiments on the KITTI and NCLT datasets show that the proposed method achieves average F1 scores of 85.28% and 77.63%, respectively, outperforming both unimodal and existing multimodal approaches. When integrated into a SLAM framework, it reduces the Absolute Error (ATE) RMSE by 11.2–16.4% compared with LiDAR-only methods, demonstrating improved loop detection accuracy and overall system robustness in complex environments. Full article
Show Figures

Figure 1

22 pages, 6376 KB  
Article
Simulator-Based Digital Twin of a Robotics Laboratory
by Lluís Ribas-Xirgo
Machines 2026, 14(3), 273; https://doi.org/10.3390/machines14030273 - 1 Mar 2026
Viewed by 305
Abstract
Simulator-based digital twins are widely used in robotics education and industrial development to accelerate prototyping and enable safe experimentation. However, they often hide implementation details that are essential for understanding, diagnosing, and correcting system failures. This paper introduces a technology-independent model-based design framework [...] Read more.
Simulator-based digital twins are widely used in robotics education and industrial development to accelerate prototyping and enable safe experimentation. However, they often hide implementation details that are essential for understanding, diagnosing, and correcting system failures. This paper introduces a technology-independent model-based design framework that provides students with full visibility of the computational mechanisms underlying robotic controllers while remaining feasible within a 150-h undergraduate course. The approach relies on representing controller behavior using networks of Extended Finite State Machines (EFSMs) and their stacked extension (EFS2M), which unify all abstraction levels of the control architecture—from low-level reactive behaviors to high-level deliberation—under a single formal model. A structured programming template ensures traceable, optimization-free software synthesis, facilitating debugging and enabling self-diagnosis of design flaws. The framework includes real-time synchronized simulation, transparent switching between virtual and physical robots, and a smart data logger that captures meaningful events for model updating and error detection. Integrated into the Intelligent Robots course, the system supports topics such as kinematics, control, perception, and simultaneous localization and mapping (SLAM) while avoiding dependency on specific middleware such as Robot Operating System (ROS) 2. Over three academic years, students reported positive hands-on experiences, strong adaptability to diverse modeling approaches, and consistently high survey ratings reflecting the course’s overall quality. The proposed environment thus offers an effective methodology for teaching end-to-end robot controller design through transparent, simulation-driven digital twins. Full article
(This article belongs to the Section Automation and Control Systems)
Show Figures

Figure 1

Back to TopTop