Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,516)

Search Parameters:
Keywords = SLAM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 15267 KB  
Article
3D Semantic Map Reconstruction for Orchard Environments Using Multi-Sensor Fusion
by Quanchao Wang, Yiheng Chen, Jiaxiang Li, Yongxing Chen and Hongjun Wang
Agriculture 2026, 16(4), 455; https://doi.org/10.3390/agriculture16040455 (registering DOI) - 15 Feb 2026
Abstract
Semantic point cloud maps play a pivotal role in smart agriculture. They provide not only core three-dimensional data for orchard management but also empower robots with environmental perception, enabling safer and more efficient navigation and planning. However, traditional point cloud maps primarily model [...] Read more.
Semantic point cloud maps play a pivotal role in smart agriculture. They provide not only core three-dimensional data for orchard management but also empower robots with environmental perception, enabling safer and more efficient navigation and planning. However, traditional point cloud maps primarily model surrounding obstacles from a geometric perspective, failing to capture distinctions and characteristics between individual obstacles. In contrast, semantic maps encompass semantic information and even topological relationships among objects in the environment. Furthermore, existing semantic map construction methods are predominantly vision-based, making them ill-suited to handle rapid lighting changes in agricultural settings that can cause positioning failures. Therefore, this paper proposes a positioning and semantic map reconstruction method tailored for orchards. It integrates visual, LiDAR, and inertial sensors to obtain high-precision pose and point cloud maps. By combining open-vocabulary detection and semantic segmentation models, it projects two-dimensional detected semantic information onto the three-dimensional point cloud, ultimately generating a point cloud map enriched with semantic information. The resulting 2D occupancy grid map is utilized for robotic motion planning. Experimental results demonstrate that on a custom dataset, the proposed method achieves 74.33% mIoU for semantic segmentation accuracy, 12.4% relative error for fruit recall rate, and 0.038803 m mean translation error for localization. The deployed semantic segmentation network Fast-SAM achieves a processing speed of 13.36 ms per frame. These results demonstrate that the proposed method combines high accuracy with real-time performance in semantic map reconstruction. This exploratory work provides theoretical and technical references for future research on more precise localization and more complete semantic mapping, offering broad application prospects and providing key technological support for intelligent agriculture. Full article
(This article belongs to the Special Issue Advances in Robotic Systems for Precision Orchard Operations)
Show Figures

Figure 1

29 pages, 33196 KB  
Article
Robust Autonomous Perception for Indoor Service Machines via Geometry-Aware RGB-D SLAM and Probabilistic Dynamic Modeling
by Zhiyu Wang, Weili Ding and Wenna Wang
Machines 2026, 14(2), 222; https://doi.org/10.3390/machines14020222 (registering DOI) - 12 Feb 2026
Viewed by 73
Abstract
Reliable autonomous perception is essential for indoor service machines operating in human-centered environments, where weak textures, repetitive structures, and frequent dynamic interference often degrade localization stability. Conventional RGB-D SLAM systems typically rely on static-scene assumptions or binary semantic masking, which are insufficient for [...] Read more.
Reliable autonomous perception is essential for indoor service machines operating in human-centered environments, where weak textures, repetitive structures, and frequent dynamic interference often degrade localization stability. Conventional RGB-D SLAM systems typically rely on static-scene assumptions or binary semantic masking, which are insufficient for handling persistent and fine-grained environmental dynamics. This paper presents a robust autonomous perception framework based on geometry-aware RGB-D SLAM, with a particular emphasis on probabilistic dynamic modeling at the feature level. The proposed system integrates multi-granularity geometric representations, including point features, parallel-line structures, and planar regions, to enhance geometric observability in low-texture indoor environments. On this basis, a probabilistic dynamic model is introduced to explicitly characterize feature reliability under motion, where dynamic probabilities are initialized by object detection and continuously updated through temporal consistency, spatial propagation, and multi-view geometric verification. Large-scale planar structures further serve as stable anchors to support robust pose estimation. Experimental results on the TUM RGB-D dynamic benchmark demonstrate that the proposed method significantly improves localization robustness, reducing the average ATE RMSE by approximately 66% compared with representative dynamic SLAM baselines. Additional evaluations on a real-world indoor dataset further validate its effectiveness for long-term autonomous perception under dense motion and frequent occlusions. Full article
(This article belongs to the Section Robotics, Mechatronics and Intelligent Machines)
Show Figures

Figure 1

22 pages, 4393 KB  
Article
Visual–Inertial Fusion-Based Restoration of Image Degradation in High-Dynamic Scenes with Rolling Shutter Cameras
by Jianbin Ye, Cengfeng Luo, Qiuxuan Wu, Yuejun Ye, Shenao Li, Yiyang Chen and Aocheng Li
Sensors 2026, 26(4), 1189; https://doi.org/10.3390/s26041189 - 12 Feb 2026
Viewed by 92
Abstract
Rolling shutter CMOS cameras are widely used in mobile and embedded vision, but rapid motion and vibration often cause coupled degradations, including motion blur and rolling shutter (RS) geometric distortion. This paper presents a visual–inertial fusion framework that estimates unified motion-related degradation parameters [...] Read more.
Rolling shutter CMOS cameras are widely used in mobile and embedded vision, but rapid motion and vibration often cause coupled degradations, including motion blur and rolling shutter (RS) geometric distortion. This paper presents a visual–inertial fusion framework that estimates unified motion-related degradation parameters from IMU and image measurements and uses them to restore both photometric and geometric image quality in high-dynamic scenes. We further introduce an exposure-aware deblurring pipeline that accounts for the nonlinear photoelectric conversion characteristics of CMOS sensors, as well as a perspective-consistent RS compensation method to improve geometric consistency under depth–motion coupling. Experiments on real mobile data and public RS-visual–inertial sequences demonstrate improved image quality and downstream SLAM pose accuracy compared with representative baselines. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

25 pages, 7517 KB  
Article
VCC: Vertical Feature and Circle Combined Descriptor for 3D Place Recognition
by Wenguang Li, Yongxin Ma, Jiying Ren, Jinshun Ou, Jun Zhou and Panling Huang
Sensors 2026, 26(4), 1185; https://doi.org/10.3390/s26041185 - 11 Feb 2026
Viewed by 132
Abstract
Loop closure detection remains a critical challenge in LiDAR-based SLAM, particularly for achieving robust place recognition in environments with rotational and translational variations. To extract more concise environmental representations from point clouds and improve extraction efficiency, this paper proposes a novel composite descriptor—the [...] Read more.
Loop closure detection remains a critical challenge in LiDAR-based SLAM, particularly for achieving robust place recognition in environments with rotational and translational variations. To extract more concise environmental representations from point clouds and improve extraction efficiency, this paper proposes a novel composite descriptor—the vertical feature and circle combined (VCC) descriptor, a novel 3D local descriptor designed for efficient and rotation-invariant place recognition. The VCC descriptor captures environmental structure by extracting vertical features from voxelized point clouds and encoding them into circular arc-based histograms, ensuring robustness to viewpoint changes. Under the same hardware, experiments conducted on different datasets demonstrate that the proposed algorithm significantly improves both feature representation efficiency and loop closure recognition performance when compared with the other descriptors, completing loop closure retrieval within 30 ms, which satisfies real-time operation requirements. The results confirm that VCC provides a compact, efficient, and rotation-invariant representation suitable for LiDAR-based SLAM systems. Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

27 pages, 6570 KB  
Article
LiDAR–Inertial–Visual Odometry Based on Elastic Registration and Dynamic Feature Removal
by Qiang Ma, Fuhong Qin, Peng Xiao, Meng Wei, Sihong Chen, Wenbo Xu, Xingrui Yue, Ruicheng Xu and Zheng He
Electronics 2026, 15(4), 741; https://doi.org/10.3390/electronics15040741 - 9 Feb 2026
Viewed by 152
Abstract
Simultaneous Localization and Mapping (SLAM) is a fundamental capability for autonomous robots. However, in highly dynamic scenes, conventional SLAM systems often suffer from degraded accuracy due to LiDAR motion distortion and interference from moving objects. To address these challenges, this paper proposes a [...] Read more.
Simultaneous Localization and Mapping (SLAM) is a fundamental capability for autonomous robots. However, in highly dynamic scenes, conventional SLAM systems often suffer from degraded accuracy due to LiDAR motion distortion and interference from moving objects. To address these challenges, this paper proposes a LiDAR–Inertial–Visual odometry framework based on elastic registration and dynamic feature removal, with the aim of enhancing system robustness through detailed algorithmic supplements. In the LiDAR odometry module, an elastic registration-based de-skewing method is introduced by modeling second-order motion, enabling accurate point cloud correction under non-uniform motion. In the visual odometry module, a multi-strategy dynamic feature suppression mechanism is developed, combining IMU-assisted motion consistency verification with a lightweight YOLOv5-based detection network to effectively filter out dynamic interference with low computational overhead. Furthermore, depth information for visual key points is recovered using LiDAR assistance to enable tightly coupled pose estimation. Extensive experiments on the TUM and M2DGR datasets demonstrate that the proposed method achieves a 96.3% reduction in absolute trajectory error (ATE) compared with ORB-SLAM2 in highly dynamic scenarios. Real-world deployment on an embedded computing device further confirms the framework’s real-time performance and practical applicability in complex environments. Full article
Show Figures

Figure 1

27 pages, 9251 KB  
Article
Novel Approach to Simultaneous Subsampling and Noise Filtering of Real-World SLAM-Acquired Point Clouds
by Martin Boušek, Martin Štroner, Hana Váchová and Jakub Kučera
Appl. Sci. 2026, 16(4), 1696; https://doi.org/10.3390/app16041696 - 8 Feb 2026
Viewed by 172
Abstract
Laser scanners based on the Simultaneous Localization and Mapping (SLAM) principle generate extremely dense point clouds burdened with a high level of surface noise arising from random measurement errors and repeated scanning of identical regions. This increases data volume and complicates subsequent processing. [...] Read more.
Laser scanners based on the Simultaneous Localization and Mapping (SLAM) principle generate extremely dense point clouds burdened with a high level of surface noise arising from random measurement errors and repeated scanning of identical regions. This increases data volume and complicates subsequent processing. The present study introduces four novel noise filtering and subsampling algorithms that selectively preserve the points closest to the true surface. Each algorithm assigns a filtering characteristic to individual points based either on their distance from a locally estimated (planar or quadratic) surface or on the degree of local eccentricity in the spherical neighborhood of the point. The proposed methods were tested on point clouds acquired using three SLAM scanners (Emesent Hovermap ST-X, FARO Orbis, and ZEB Horizon) in three different scenes with reference data acquired by a static terrestrial scanner Leica P40. All four proposed methods effectively reduced surface noise and data volume (improving the RMSDs by 45.4–75.8% compared to the original cloud after thinning to 10% of cloud size). This clearly outperformed the standard subsampling tools, namely random subsampling (RMSD remained constant after subsampling), octree, or spatial subsampling (worsening of RMSDs with increasing subsampling). The most reliable surface noise removal in point clouds dominated by planar surfaces (building interior with planar walls) was achieved using the method based on local plane fitting. In contrast, the use of a quadratic surface proved more effective for uneven or rugged surfaces. Full article
(This article belongs to the Section Civil Engineering)
Show Figures

Figure 1

29 pages, 5833 KB  
Article
Spacio-Linear Screening for Ligand-Docking Cavities in Protein Structures: SLAM Algorithm
by Julia Panov, Alexander Elbert, Dean S. Rosenthal, Moshe Levi, Konstantin Chumakov, Raul Andino, Leonid Brodsky and Hanoch Kaphzan
Life 2026, 16(2), 285; https://doi.org/10.3390/life16020285 - 7 Feb 2026
Viewed by 144
Abstract
Identifying structurally similar ligand-binding sites in unrelated proteins can facilitate drug repurposing, reveal off-target effects, and deepen our understanding of protein function. A number of tools were developed for structural screening, but many of them suffer from limited sensitivity and scalability. Using a [...] Read more.
Identifying structurally similar ligand-binding sites in unrelated proteins can facilitate drug repurposing, reveal off-target effects, and deepen our understanding of protein function. A number of tools were developed for structural screening, but many of them suffer from limited sensitivity and scalability. Using a data bank of crystallized protein structures, we aimed to discover novel protein targets for a ligand by leveraging a known ligand-binding query protein with a resolved structure. Here, we present SLAM (Spacio-Linear Alignment of Macromolecules), a novel alignment-based algorithm that detects local 3D similarities between ligand-binding cavities or protein-exposed surfaces of query and target proteins. SLAM encodes spatial substructure neighborhoods into short linear sequences of physicochemically annotated atoms, then applies pairwise sequence alignment combined with distance-correlation scoring to identify high-fidelity structural matches. Benchmarking using the Kahraman-36 dataset demonstrated that SLAM outperforms the state-of-the-art ProBiS algorithm in true-positive rate for predicting ligand-docking compatibility. Furthermore, SLAM identifies candidate ligands that may inhibit functionally critical domains of CRISPR-Cas proteins and predicts novel binding partners of toxic per- and polyfluoroalkyl Substance (PFAS) compounds (PFOA, PFOS) with plausible mechanistic links to toxicity. In conclusion, SLAM is a robust computationally efficient and flexible structural screening tool capable of detecting subtle physicochemical compatibilities between protein surfaces, promising to accelerate target discovery in pharmacology and elucidate protein–ligand interactions in environmental toxicology. Full article
(This article belongs to the Section Biochemistry, Biophysics and Computational Biology)
Show Figures

Figure 1

23 pages, 5683 KB  
Article
Optimizing RTAB-Map Viewability to Reduce Cognitive Workload in VR Teleoperation: A User-Centric Approach
by Hojin Yoon, Haegyeom Choi, Jaehoon Jeong and Donghun Lee
Mathematics 2026, 14(3), 579; https://doi.org/10.3390/math14030579 - 6 Feb 2026
Viewed by 215
Abstract
In industrial environments, providing intuitive spatial information via 3D maps is essential for maximizing the efficiency of teleoperation. However, existing SLAM algorithms generating 3D maps predominantly focus on improving robot localization accuracy, often neglecting the optimization of viewability required for human operators to [...] Read more.
In industrial environments, providing intuitive spatial information via 3D maps is essential for maximizing the efficiency of teleoperation. However, existing SLAM algorithms generating 3D maps predominantly focus on improving robot localization accuracy, often neglecting the optimization of viewability required for human operators to clearly perceive object depth and structure in virtual environments. To address this, this study proposes a methodology to optimize the viewability of RTAB-Map-based 3D maps using the Taguchi method, aiming to enhance VR teleoperation efficiency and reduce cognitive workload. We identified eight key parameters that critically affect visual quality and utilized an L18 orthogonal array to derive an optimal combination that controls point cloud density and noise levels. Experimental results from a target object picking task demonstrated that the optimized 3D map reduced task completion time by approximately 9 s compared to the RGB image condition, achieving efficiency levels approaching those of the physical-world baseline. Furthermore, evaluations using NASA-TLX confirmed that intuitive visual feedback minimized situational awareness errors and substantially alleviated cognitive workload. This study suggests a new direction for constructing high-efficiency teleoperation interfaces from a Human–Robot Interaction perspective by expanding SLAM optimization criteria from geometric precision to user-centric visual quality. Full article
(This article belongs to the Special Issue Advances in Machine Learning and Intelligent Systems)
Show Figures

Figure 1

27 pages, 12640 KB  
Article
A Suitable Scan-to-BIM Process Using OS Software and Low-Cost Sensors: Trend, Solutions and Experimental Validation
by Massimiliano Pepe, Przemysław Klapa, Andrei Crisan, Ahmed Kamal Hamed Dewedar and Donato Palumbo
Architecture 2026, 6(1), 24; https://doi.org/10.3390/architecture6010024 - 5 Feb 2026
Viewed by 252
Abstract
Open-source software is transforming visualization-oriented digital documentation and conceptual BIM by lowering financial and technical barriers, enabling broader participation in the digitalization of the AEC sector. This study develops and validates a cost-effective Scan-to-BIM workflow that combines low-cost hardware with freely available software [...] Read more.
Open-source software is transforming visualization-oriented digital documentation and conceptual BIM by lowering financial and technical barriers, enabling broader participation in the digitalization of the AEC sector. This study develops and validates a cost-effective Scan-to-BIM workflow that combines low-cost hardware with freely available software for 3D data acquisition, processing, and modeling. Photogrammetry and SLAM-based techniques generate accurate point clouds, which, once verified against terrestrial laser scanning data, can be integrated into open-source BIM environments. The workflow leverages COLMAP for 3D reconstruction and BlenderBIM for parametric modeling, combining geometric and semantic information to produce fully interoperable models. While open-source tools offer accessibility and transparency, they require supplementary validation in precision-critical applications and may involve trade-offs in accuracy, stability, and automation compared to commercial solutions. Application to a case study shows how efficient and rapid the process is, representing the trend for the scientific community. Full article
Show Figures

Figure 1

32 pages, 53691 KB  
Article
Underwater SLAM and Calibration with a 3D Profiling Sonar
by António Ferreira, José Almeida, Aníbal Matos and Eduardo Silva
Remote Sens. 2026, 18(3), 524; https://doi.org/10.3390/rs18030524 - 5 Feb 2026
Viewed by 163
Abstract
High resolution underwater mapping is fundamental to the sustainable development of the blue economy, supporting offshore energy expansion, marine habitat protection, and the monitoring of both living and non-living resources. This work presents a pose-graph SLAM and calibration framework specifically designed for 3D [...] Read more.
High resolution underwater mapping is fundamental to the sustainable development of the blue economy, supporting offshore energy expansion, marine habitat protection, and the monitoring of both living and non-living resources. This work presents a pose-graph SLAM and calibration framework specifically designed for 3D profiling sonars, such as the Coda Octopus Echoscope 3D. The system integrates a probabilistic scan matching method (3DupIC) for direct registration of 3D sonar scans, enabling accurate trajectory and map estimation even under degraded dead reckoning conditions. Unlike other bathymetric SLAM methods that rely on submaps and assume short-term localization accuracy, the proposed approach performs direct scan-to-scan registration, removing this dependency. The factor graph is extended to represent the sonar extrinsic parameters, allowing the sonar-to-body transformation to be refined jointly with trajectory optimization. Experimental validation on a challenging real world dataset demonstrates outstanding localization and mapping performance. The use of refined extrinsic parameters further improves both accuracy and map consistency, confirming the effectiveness of the proposed joint SLAM and calibration approach for robust and consistent underwater mapping. Full article
(This article belongs to the Special Issue Underwater Remote Sensing: Status, New Challenges and Opportunities)
Show Figures

Figure 1

14 pages, 2391 KB  
Article
The Anti-SLAMF7 Antibody, Elotuzumab, Induces Antibody-Dependent Cellular Cytotoxicity Against CLL Cell Lines
by Dominik Kľoc, Bianca Dubiková, Simona Žiláková, Ján Sykora, Michaela Šuliková, Slavomír Kurhajec, Ján Sabo, Tomáš Guman and Marek Šarišský
Molecules 2026, 31(3), 531; https://doi.org/10.3390/molecules31030531 - 3 Feb 2026
Viewed by 222
Abstract
SLAMF7, also known as CD319, a SLAM (signaling lymphocytic activation molecule) family receptor, is relatively weakly expressed on chronic lymphocytic leukemia (CLL) B cells. This study evaluated the ability of elotuzumab (E), an anti-SLAMF7/CD319 antibody, to induce antibody-dependent cellular cytotoxicity (ADCC) against CLL [...] Read more.
SLAMF7, also known as CD319, a SLAM (signaling lymphocytic activation molecule) family receptor, is relatively weakly expressed on chronic lymphocytic leukemia (CLL) B cells. This study evaluated the ability of elotuzumab (E), an anti-SLAMF7/CD319 antibody, to induce antibody-dependent cellular cytotoxicity (ADCC) against CLL cell lines (MEC-1, MEC-2, CI, HG-3, PGA-1, WA-OSEL). ADCC was assessed by flow cytometry using E (100 μg/mL), rituximab (R, 100 μg/mL), and their combination (E + R). CLL lines served as targets (T), while peripheral blood mononuclear cells (PBMCs) or NK cells from healthy donors served as effectors (E) at an 8:1 E:T ratio for 4 h. With PBMCs, E-induced ADCC ranged from 1.3 ± 1.2% (PGA-1) to 14.6 ± 8.1% (MEC-1); R-induced ADCC ranged from 9.2 ± 4.6% (PGA-1) to 16.6 ± 9.4% (WA-OSEL). With NK cells, E-induced ADCC ranged from 1.8 ± 3.7% (PGA-1) to 27.3 ± 4.7% (MEC-1); R-induced ADCC ranged from 5.1 ± 4.3% (PGA-1) to 27.5 ± 13.6% (CI). E outperformed R in MEC-1, while R was superior elsewhere. Cell lines with higher SLAMF7/CD319 expression displayed increased sensitivity to E. Cell lines with del17p showed higher SLAMF7/CD319 expression. The combination of E + R showed no significant synergy over monotherapies. In conclusion, elotuzumab induced significant ADCC in CLL cells, warranting further therapeutic evaluation. Full article
Show Figures

Figure 1

23 pages, 6708 KB  
Article
Feasibility Domain Construction and Characterization Method for Intelligent Underground Mining Equipment Integrating ORB-SLAM3 and Depth Vision
by Siya Sun, Xiaotong Han, Hongwei Ma, Haining Yuan, Sirui Mao, Chuanwei Wang, Kexiang Ma, Yifeng Guo and Hao Su
Sensors 2026, 26(3), 966; https://doi.org/10.3390/s26030966 - 2 Feb 2026
Viewed by 204
Abstract
To address the limited environmental perception capability and the difficulty of achieving consistent and efficient representation of the workspace feasible domain caused by high dust concentration, uneven illumination, and enclosed spaces in underground coal mines, this paper proposes a digital spatial construction and [...] Read more.
To address the limited environmental perception capability and the difficulty of achieving consistent and efficient representation of the workspace feasible domain caused by high dust concentration, uneven illumination, and enclosed spaces in underground coal mines, this paper proposes a digital spatial construction and representation method for underground environments by integrating RGB-D depth vision with ORB-SLAM3. First, a ChArUco calibration board with embedded ArUco markers is adopted to perform high-precision calibration of the RGB-D camera, improving the reliability of geometric parameters under weak-texture and non-uniform lighting conditions. On this basis, a “dense–sparse cooperative” OAK-DenseMapper Pro module is further developed; the module improves point-cloud generation using a mathematical projection model, and combines enhanced stereo matching with multi-stage depth filtering to achieve high-quality dense point-cloud reconstruction from RGB-D observations. The dense point cloud is then converted into a probabilistic octree occupancy map, where voxel-wise incremental updates are performed for observed space while unknown regions are retained, enabling a memory-efficient and scalable 3D feasible-space representation. Experiments are conducted in multiple representative coal-mine tunnel scenarios; compared with the original ORB-SLAM3, the number of points in dense mapping increases by approximately 38% on average; in trajectory evaluation on the TUM dataset, the root mean square error, mean error, and median error of the absolute pose error are reduced by 7.7%, 7.1%, and 10%, respectively; after converting the dense point cloud to an octree, the map memory footprint is only about 0.5% of the original point cloud, with a single conversion time of approximately 0.75 s. The experimental results demonstrate that, while ensuring accuracy, the proposed method achieves real-time, efficient, and consistent representation of the 3D feasible domain in complex underground environments, providing a reliable digital spatial foundation for path planning, safe obstacle avoidance, and autonomous operation. Full article
Show Figures

Figure 1

45 pages, 5418 KB  
Review
Visual and Visual–Inertial SLAM for UGV Navigation in Unstructured Natural Environments: A Survey of Challenges and Deep Learning Advances
by Tiago Pereira, Carlos Viegas, Salviano Soares and Nuno Ferreira
Robotics 2026, 15(2), 35; https://doi.org/10.3390/robotics15020035 - 2 Feb 2026
Viewed by 596
Abstract
Localization and mapping remain critical challenges for Unmanned Ground Vehicles (UGVs) operating in unstructured natural environments, such as forests and agricultural fields. While Visual SLAM (VSLAM) and Visual–Inertial SLAM (VI-SLAM) have matured significantly in structured and urban scenarios, their extension to outdoor natural [...] Read more.
Localization and mapping remain critical challenges for Unmanned Ground Vehicles (UGVs) operating in unstructured natural environments, such as forests and agricultural fields. While Visual SLAM (VSLAM) and Visual–Inertial SLAM (VI-SLAM) have matured significantly in structured and urban scenarios, their extension to outdoor natural domains introduces severe challenges, including dynamic vegetation, illumination variations, a lack of distinctive features, and degraded GNSS availability. Recent advances in Deep Learning have brought promising developments to VSLAM- and VI-SLAM-based pipelines, ranging from learned feature extraction and matching to self-supervised monocular depth prediction and differentiable end-to-end SLAM frameworks. Furthermore, emerging methods for adaptive sensor fusion, leveraging attention mechanisms and reinforcement learning, open new opportunities to improve robustness by dynamically weighting the contributions of camera and IMU measurements. This review provides a comprehensive overview of Visual and Visual–Inertial SLAM for UGVs in unstructured environments, highlighting the challenges posed by natural contexts and the limitations of current pipelines. Classic VI-SLAM frameworks and recent Deep-Learning-based approaches were systematically reviewed. Special attention is given to field robotics applications in agriculture and forestry, where low-cost sensors and robustness against environmental variability are essential. Finally, open research directions are discussed, including self-supervised representation learning, adaptive sensor confidence models, and scalable low-cost alternatives. By identifying key gaps and opportunities, this work aims to guide future research toward resilient, adaptive, and economically viable VSLAM and VI-SLAM pipelines, tailored for UGV navigation in unstructured natural environments. Full article
(This article belongs to the Special Issue Localization and 3D Mapping of Intelligent Robotics)
Show Figures

Figure 1

23 pages, 6426 KB  
Article
An Improved Map Information Collection Tool Using 360° Panoramic Images for Indoor Navigation Systems
by Kadek Suarjuna Batubulan, Nobuo Funabiki, I Nyoman Darma Kotama, Komang Candra Brata and Anak Agung Surya Pradhana
Appl. Sci. 2026, 16(3), 1499; https://doi.org/10.3390/app16031499 - 2 Feb 2026
Viewed by 262
Abstract
At present, pedestrian navigation systems using smartphones have become common in daily activities. For their ubiquitous, accurate, and reliable services, map information collection is essential for constructing comprehensive spatial databases. Previously, we have developed a map information collection tool to extract building information [...] Read more.
At present, pedestrian navigation systems using smartphones have become common in daily activities. For their ubiquitous, accurate, and reliable services, map information collection is essential for constructing comprehensive spatial databases. Previously, we have developed a map information collection tool to extract building information using Google Maps, optical character recognition (OCR), geolocation, and web scraping with smartphones. However, indoor navigation often suffers from inaccurate localization due to degraded GPS signals inside buildings and Simultaneous Localization and Mapping (SLAM) estimation errors, causing position errors and confusing augmented reality (AR) guidance. In this paper, we present an improved map information collection tool to address this problem. It captures 360° panoramic images to build 3D models, apply photogrammetry-based mesh reconstruction to correct geometry, and georeference point clouds to refine latitude–longitude coordinates. For evaluations, experiments in various indoor scenarios were conducted. The results demonstrate that the proposed method effectively mitigates positional errors with an average drift correction of 3.15 m, calculated via the Haversine formula. Geometric validation using point cloud analysis showed high registration accuracy, which translated to a 100% task completion rate and an average navigation time of 124.5 s among participants. Furthermore, usability testing using the System Usability Scale (SUS) yielded an average score of 96.5, categorizing the user interface as ’Best Imaginable’. These quantitative findings substantiate that the integration of 360° imaging and photogrammetric correction significantly enhances navigation reliability and user satisfaction compared with previous sensor fusion approaches. Full article
(This article belongs to the Section Electrical, Electronics and Communications Engineering)
Show Figures

Figure 1

20 pages, 1202 KB  
Article
Adaptive ORB Accelerator on FPGA: High Throughput, Power Consumption, and More Efficient Vision for UAVs
by Hussam Rostum and József Vásárhelyi
Signals 2026, 7(1), 13; https://doi.org/10.3390/signals7010013 - 2 Feb 2026
Viewed by 240
Abstract
Feature extraction and description are fundamental components of visual perception systems used in applications such as visual odometry, Simultaneous Localization and Mapping (SLAM), and autonomous navigation. In resource-constrained platforms, such as Unmanned Aerial Vehicles (UAVs), achieving real-time hardware acceleration on Field-Programmable Gate Arrays [...] Read more.
Feature extraction and description are fundamental components of visual perception systems used in applications such as visual odometry, Simultaneous Localization and Mapping (SLAM), and autonomous navigation. In resource-constrained platforms, such as Unmanned Aerial Vehicles (UAVs), achieving real-time hardware acceleration on Field-Programmable Gate Arrays (FPGAs) is challenging. This work demonstrates an FPGA-based implementation of an adaptive ORB (Oriented FAST and Rotated BRIEF) feature extraction pipeline designed for high-throughput and energy-efficient embedded vision. The proposed architecture is a completely new design for the main algorithmic blocks of ORB, including the FAST (Features from Accelerated Segment Test) feature detector, Gaussian image filtering, moment computation, and descriptor generation. Adaptive mechanisms are introduced to dynamically adjust thresholds and filtering behavior, improving robustness under varying illumination conditions. The design is developed using a High-Level Synthesis (HLS) approach, where all processing modules are implemented as reusable hardware IP cores and integrated at the system level. The architecture is deployed and evaluated on two FPGA platforms, PYNQ-Z2 and KRIA KR260, and its performance is compared against CPU and GPU implementations using a dedicated C++ testbench based on OpenCV. Experimental results demonstrate significant improvements in throughput and energy efficiency while maintaining stable and scalable performance, making the proposed solution suitable for real-time embedded vision applications on UAVs and similar platforms. Notably, the FPGA implementation increases DSP utilization from 11% to 29% compared to the previous designs implemented by other researchers, effectively offloading computational tasks from general purpose logic (LUTs and FFs), reducing LUT usage by 6% and FF usage by 13%, while maintaining overall design stability, scalability, and acceptable thermal margins at 2.387 W. This work establishes a robust foundation for integrating the optimized ORB pipeline into larger drone systems and opens the door for future system-level enhancements. Full article
Show Figures

Figure 1

Back to TopTop