Next Article in Journal
SLR-Net: Lightweight and Accurate Detection of Weak Small Objects in Satellite Laser Ranging Imagery
Previous Article in Journal
SADQN-Based Residual Energy-Aware Beamforming for LoRa-Enabled RF Energy Harvesting for Disaster-Tolerant Underground Mining Networks
Previous Article in Special Issue
Influence of Control System Architecture on Mobile Robot Stability and Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

S3PM: Entropy-Regularized Path Planning for Autonomous Mobile Robots in Dense 3D Point Clouds of Unstructured Environments

by
Artem Sazonov
1,*,
Oleksii Kuchkin
1,
Irina Cherepanska
2 and
Arūnas Lipnickas
3,*
1
Automation Hardware and Software Department, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, 37, Prospect Beresteiskyi, 03056 Kyiv, Ukraine
2
Department of Automation and Non-Destructive Testing Systems, National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, 37, Prospect Beresteiskyi, 03056 Kyiv, Ukraine
3
Department of Automation, Kaunas University of Technology, 51367 Kaunas, Lithuania
*
Authors to whom correspondence should be addressed.
Sensors 2026, 26(2), 731; https://doi.org/10.3390/s26020731
Submission received: 17 December 2025 / Revised: 17 January 2026 / Accepted: 19 January 2026 / Published: 21 January 2026
(This article belongs to the Special Issue Mobile Robots: Navigation, Control and Sensing—2nd Edition)

Abstract

Autonomous navigation in cluttered and dynamic industrial environments remains a major challenge for mobile robots. Traditional occupancy-grid and geometric planning approaches often struggle in such unstructured settings due to partial observability, sensor noise, and the frequent presence of moving agents (machinery, vehicles, humans). These limitations seriously undermine long-term reliability and safety compliance—both essential for Industry 4.0 applications. This paper introduces S3PM, a lightweight entropy-regularized framework for simultaneous mapping and path planning that operates directly on dense 3D point clouds. Its key innovation is a dynamics-aware entropy field that fuses per-voxel occupancy probabilities with motion cues derived from residual optical flow. Each voxel is assigned a risk-weighted entropy score that accounts for both geometric uncertainty and predicted object dynamics. This representation enables (i) robust differentiation between reliable free space and ambiguous/hazardous regions, (ii) proactive collision avoidance, and (iii) real-time trajectory replanning. The resulting multi-objective cost function effectively balances path length, smoothness, safety margins, and expected information gain, while maintaining high computational efficiency through voxel hashing and incremental distance transforms. Extensive experiments in both real-world and simulated settings, conducted on a Raspberry Pi 5 (with and without the Hailo-8 NPU), show that S3PM achieves 18–27% higher IoU in static/dynamic segmentation, 0.94–0.97 AUC in motion detection, and 30–45% fewer collisions compared to OctoMap + RRT* and standard probabilistic baselines. The full pipeline runs at 12–15 Hz on the bare Pi 5 and 25–30 Hz with NPU acceleration, making S3PM highly suitable for deployment on resource-constrained embedded platforms.

1. Introduction

Autonomous navigation in unstructured industrial environments remains one of the most critical challenges in mobile robotics. Unlike highly structured domains such as warehouses or assembly lines, industrial facilities exhibit irregular layouts, cluttered workspaces, frequent equipment reconfiguration, and the constant presence of moving agents—including machinery, automated guided vehicles (AGVs), unmanned aerial vehicles (UAVs), and human operators. These factors introduce substantial uncertainty and unpredictability, significantly complicating reliable localization, robust mapping, and safe path planning compared to controlled settings [1,2,3,4].
The demand for flexible automation driven by Industry 4.0 has further intensified the need for robotic systems capable of seamless adaptation to changing environments while operating safely alongside humans. Failure to achieve robust navigation not only compromises operational efficiency and increases downtime but also poses serious safety risks, including potential collisions with personnel or valuable equipment. Consequently, the development of real-time methods that can effectively handle both static clutter and dynamic uncertainty is essential for the large-scale deployment of autonomous mobile robots in industrial applications [5,6].
Classical mapping and motion-planning approaches—such as occupancy-grid mapping combined with graph-search algorithms (A*, D*) [7,8] or sampling-based planners (RRT*, PRM) [9,10,11,12]—have been widely successful in structured or semi-structured environments. However, these methods rely on rigid environmental assumptions and struggle in cluttered, partially observable, or highly dynamic 3D spaces [13,14,15]. Probabilistic volumetric frameworks such as OctoMap [2] have improved scalability and uncertainty modeling; nevertheless, they typically do not incorporate explicit motion cues and remain computationally expensive when updating large-scale maps in real time.
Furthermore, while state-of-the-art local planners such as the Timed Elastic Band (TEB) [16,17] and Model Predictive Path Integral (MPPI) control [18,19] excel at dynamic obstacle avoidance, they impose substantial computational demands that are often prohibitive for low-power embedded edge devices lacking dedicated GPUs. Simpler reactive methods, such as the Dynamic Window Approach (DWA) [20], are computationally efficient but frequently fail in complex environments due to clutter-induced local minima. More recent learning-based approaches [21,22,23] have demonstrated impressive performance in simulation; however, they typically require large training datasets, exhibit limited generalization beyond the training distribution, and are rarely suitable for resource-constrained embedded platforms because of latency and memory limitations.
In this work, we propose S3PM, an entropy-regularized framework for simultaneous mapping and path planning that operates directly on dense 3D point clouds. The core innovation is a dynamic-aware entropy field that fuses voxel-wise occupancy probabilities with motion evidence derived from residual optical flow. This representation enables the system to distinguish reliably navigable space from ambiguous or high-risk regions, supports proactive collision prediction, and facilitates real-time trajectory correction. Importantly, the method is designed with embedded deployment in mind, achieving real-time performance on low-power platforms such as the Raspberry Pi 5.
Through extensive evaluation in both simulated and real-world industrial scenarios, we demonstrate that S3PM substantially improves navigation robustness and safety while maintaining computational efficiency. The proposed approach establishes a foundation for entropy-driven navigation strategies that generalize beyond industrial settings and advance the broader field of autonomous mobile robotics in unstructured environments.

2. Materials and Methods

2.1. Related Work

Research on mobile robot navigation has produced a wide variety of approaches, ranging from classical geometric planning methods [24] to modern learning-based techniques [21,22,23]. While each of these paradigms has demonstrated progress in specific domains, significant limitations remain when transferring them to dense, unstructured 3D indoor environments, where both static and dynamic uncertainties are prevalent.
One of the most established families of approaches is based on occupancy-grid mapping and graph-search algorithms such as A*, D*, and their variants [7,8]. These methods are computationally efficient and provide deterministic guarantees in structured settings, such as warehouses or predefined floor layouts. However, their reliance on discretized 2D projections or simplified geometric assumptions severely limits their performance in cluttered 3D spaces. The absence of a principled mechanism for quantifying uncertainty further makes these planners brittle when faced with noisy or ambiguous sensor data (e.g., from depth cameras or LiDAR). As noted by several authors [14,15,25,26], despite decades of refinement, classical search-based methods often fail to generalize to unstructured and dynamic conditions without extensive manual tuning.
Another major group of approaches focuses on sampling-based motion planning, including algorithms such as Rapidly Exploring Random Trees (RRT) [27,28,29] and Probabilistic Roadmaps (PRM) [30]. These planners are well suited to high-dimensional search spaces and have been widely applied in both terrestrial (AGVs) and aerial robotics (UAVs). Nonetheless, they are highly sensitive to the underlying map representation: when applied to raw or voxelized point clouds, they often require costly preprocessing steps (e.g., meshing or smoothing) to avoid spurious connections. Moreover, sampling-based planners typically treat the environment as static during each planning cycle, which limits their applicability in scenarios involving moving agents or reconfigurable obstacles [21].
In the domain of dynamic collision avoidance, local trajectory optimization methods have become a standard solution. The Dynamic Window Approach (DWA) [20] offers a lightweight reactive strategy but suffers from a short prediction horizon, frequently leading to dead ends in complex industrial layouts. To mitigate this limitation, optimization-based methods such as the Timed Elastic Band (TEB) planner [17] deform trajectories to avoid obstacles while respecting kinematic constraints. However, TEB becomes computationally expensive in dense 3D environments and is particularly sensitive to noisy point cloud data. Model Predictive Path Integral (MPPI) control [19] demonstrates state-of-the-art performance in highly dynamic settings by sampling thousands of candidate trajectories. Despite its robustness, MPPI typically requires massive parallelization on powerful GPUs to achieve real-time performance [18,19], rendering it unsuitable for cost-effective, energy-constrained embedded platforms such as the Raspberry Pi without substantial hardware acceleration.
Besides the above-mentioned groups of algorithms, the Probabilistic Occupancy Mapping techniques, particularly those derived from Bayesian filtering and SLAM frameworks, have been proposed to explicitly model uncertainty in perception. A prominent representative of such techniques, like OctoMap [2], attempts to maintain a volumetric belief distribution over free and occupied spaces. While powerful in outdoor robotics and aerial mapping, these methods suffer from scalability and responsiveness issues in dense indoor settings, as updating a large 3D map in real time is computationally expensive. Furthermore, these techniques generally do not incorporate motion cues from optical flow or temporal dynamics, making them ill-suited for environments where static and dynamic entities must be separated [23].
More recently, learning-based methods have gained attention [26,27,28], leveraging convolutional neural networks (CNNs), graph neural networks (GNNs), or reinforcement learning to directly infer navigation strategies from sensor data. These approaches can, in principle, capture complex scene semantics and adapt to dynamic conditions. However, they remain limited by their dependence on large training datasets, difficulties in generalizing beyond the training distribution, and heavy computational requirements. As highlighted in a review [22], many deep learning planners achieve impressive simulation performance but are rarely deployed in resource-constrained embedded systems due to latency and memory bottlenecks.
In contrast to these approaches, the proposed method integrates entropy-regularized mapping with dynamic likelihood estimation, thereby unifying uncertainty modeling and motion awareness into a single probabilistic path planning method (S3PM) described below. This allows the system to remain lightweight enough for embedded execution while explicitly reasoning about ambiguous and dynamic regions of the environment, a combination not adequately addressed in prior works.

2.2. S3PM Explanation

The proposed method first defines the evaluation criteria and regularization terms, followed by a description of the optimization process used to determine the mobile robot’s optimal path.

2.2.1. Entropy-Guided Environment Representation

To enable safe and efficient path planning for mobile robotic platforms operating in unstructured indoor environments, an entropy-driven representation of the environment based on 3D point clouds is introduced (Figure 1). Rather than operating directly on raw point clouds, a voxel-based representation using spatial hashing is employed. This design choice is motivated by the need for high-frequency map updates on embedded hardware platforms. Voxelization abstracts away sub-voxel geometric detail and aggregates dense sensor measurements into discrete spatial units. Instead of performing computationally expensive nearest-neighbor searches for individual points, the probabilistic parameters of the corresponding voxels can be accessed and updated in constant time, O(1). As a result, the global map state and its associated entropy can be refreshed at each incoming camera frame without introducing computational bottlenecks. Each voxelized region is subsequently assigned an entropy score that quantifies the uncertainty and variability of the local geometry. The entropy H of a voxel v is defined as:
H v = i = 1 k p i log p i ,
where pi denotes the probability of the occupancy states (free, occupied, dynamic, unknown), computed according to (6). High-entropy regions correspond to ambiguous or unstable structures (e.g., cluttered areas or moving objects), whereas low-entropy regions represent either reliably free or reliably occupied space.
This entropy map enables dividing the environment into navigable and non-navigable zones. Regions with entropy below a chosen threshold are considered safe candidates for traversal, whereas high-entropy zones are treated as obstacles with high uncertainty or dynamic risk.
Furthermore, regions containing uncertain obstacles should be evaluated in conjunction with pixel-level motion estimates obtained from optical flow, as dynamic objects typically influence extended regions of the map along their direction of motion.
Thus, for each image pair ( I t , I t + Δ ) semi-dense optical flow f ( u ) is estimated at pixel u (Figure 2). Given pose change T t t + Δ and per-pixel depth z ( u ) , the expected flow due to static background (ego-motion only) is calculated. For a pixel u = ( x , y ) with normalized coordinates, the expected motion field f ˆ u = ( u ˙ , v ˙ ) induced purely by ego-motion is derived from the pinhole model:
u ˙ v ˙ = 1 / z 0 x / z 0 1 / z y / z v + x y 1 + x 2 y 1 + y 2 x y x ω ,
where v —linear velocity and ω —angular velocity both are calculated from the VIO pipeline. Thus, the residual flow r ( u ) (3) is small for a static background and large where independently moving objects are present.
r ( u ) = f ( u ) f ˆ ( u ) .
The residual magnitudes are passed through a robust sigmoid to obtain dynamic likelihoods:
l d y n ( u ) = w vis σ r ( u ) τ σ r ,
where τ ( 0 , 255 ) is the threshold tied to photometric noise, σ r = 127 —a scale parameter, σ sigmoid function w vis —weights low-texture regions. To mitigate errors in texture-less regions where optical flow is unreliable, the weight w vis in (3) is derived from the local image gradient magnitude | I | . If | I | < δ texture , w vis 0 , preventing false dynamic positives on plain walls. Furthermore, localization errors typically cause global spikes in the residual flow. These are filtered by enforcing consistency checks on residual flow directions within local pixel neighborhoods, thereby rejecting random noise induced by drift. Once a confident dynamic likelihood is obtained at the pixel level, this evidence is propagated into the volumetric map using the following ray-based update rules:
  • Along the ray until the first confident surface, the free-space evidence is added;
  • At the surface voxel, the evidence is split between occupied vs. dynamic by using l d y n (e.g., add l d y n to n d y n , 1 l d y n to n stat   );
  • Voxels never intersected by valid rays retain unknown mass.
To avoid jitters, temporal consistency is applied with exponential smoothing of dynamic likelihoods for object tracks (e.g., person/forklift) using short-term association in image space (optical flow) or 3D (nearest-neighbor on centroids):
l ˜ d y n , t = β l ˜ d y n , t 1 + ( 1 β ) l d y n , t ,
where β [ 0 , 1 ] —smoothness parameter; l ˜ d y n , t 1 —previous dynamic likelihood; l ˜ d y n , t —current dynamic likelihood.

2.2.2. Probabilistic Updates and the Dynamic-Aware Entropy Field

The approach employs per-voxel Dirichlet-multinomial Bayesian updates to ensure robust probabilities under sparse observations. The state of a voxel v is modeled by a categorical distribution over S = { f r e e ,   s t a t ,   d y n ,   u n k } with parameters derived from accumulated pseudo-counts α t ( v ) . Unlike standard occupancy grids that update log-odds directly, the sensory evidence is explicitly mapped into count increments Δ n s . Voxel updates occur by unprojecting rays from valid pixels (with computable depth and optical flow) from the camera as follows:
  • All voxels v along the ray path (excluding the endpoint) receive evidence assigned to the free state:
Δ n free ( v ) = w vis , Δ n stat d y n ( v ) = 0 .
2.
At the endpoint voxel v hit , the evidence is split between static and dynamic occupancy based on the dynamic likelihood derived in (5):
Δ n d y n v h i t = l ˜ dyn u ,
Δ n s t a t v h i t = 1 l ˜ d y n ( u ) .
3.
Spatial unpredictability of moving agents (e.g., motion blur or rapid position changes) prompts strict increases in unknown evidence within the immediate neighborhood N v h i t of voxels identified as dynamic. If l ˜ d y n u > 0.5 , then for all neighbors v N v h i t
Δ n u n k v = 1 v v h i t l ˜ dyn u .
This crucial step ensures that the entropy field spikes not only at the moving object but also in its vicinity, forcing the planner to maintain a safety buffer due to high epistemic uncertainty.
These instantaneous increments are used to update the Dirichlet parameters α t ( v ) = α t f r e e , α t s t a t , α t d y n , α t u n k :
α t + 1 s v = λ α t s v + Δ n s v , s S ,
where λ ( 0 , 1 ] —forgetting factor (smaller λ lets the map adapt quicker to layout changes). Probabilities follow the posterior mean:
p t + 1 s ( v ) = α t + 1 s ( v ) s S α t + 1 s ( v ) .
We define the dynamic-aware entropy per voxel (Figure 3):
H v = s S p s v log p s v .
While conventional planners often use a linear combination of static and dynamic occupancy probabilities p stat   + p d y n as a cost term, this formulation fails to distinguish between known obstacles and epistemic uncertainty. For instance, a voxel with p = [ 0.5 , 0.5 ] (high ambiguity) and a voxel with p = [ 0.9 , 0.1 ] (representing high certainty) may yield similar linear costs in standard risk maps. In contrast, the entropy-based formulation in (8) assigns a substantially higher penalty to the ambiguous case. As a result, the planner is explicitly encouraged to avoid regions in which the perception system exhibits high uncertainty (e.g., irregular optical flow patterns or sparse sensor data), rather than merely avoiding clearly detected static obstacles such as walls.
Dynamic occupancy introduces additional operational risk; therefore, risk-weighted entropy is incorporated into the planning objective:
H ρ v = s S ρ s p s v log p s v , ρ d y n > ρ s t a t ρ f r e e ,
where ρ scales the penalty of uncertainty according to downstream safety.

2.2.3. Navigability Classification from Entropy and Risk

Occupancy probability p ( v ) and H ρ ( v ) yield navigability score C n a v ( v ) for the planner:
C nav ( v ) = w occ p stat ( v ) + p dyn ( v ) + w ent H ρ ( v ) + w clr Φ clear ( v ) ,
where Φ clear ( v ) —penalty function for poor clearance from likely occupied voxels. By explicitly separating the occupancy probability term w occ from the entropy term w ent , the cost function (18) enables decision behavior that standard probability-weighted maps cannot achieve. Unlike conventional risk maps, where a probability of p = 0.5 typically yields a linear cost equivalent to a semi-traversable terrain (often leading to risky shortcuts through ambiguous areas), the entropy term H v acts as a non-linear barrier against epistemic uncertainty. This formulation results in distinct path selection behavior: the planner effectively penalizes unobserved regions as heavily as known obstacles, forcing the robot to select longer paths through verified free space rather than shorter paths through ambiguous data. Such differentiation enables the robot to be aggressive near known obstacles (low H , high p ) but conservative near ambiguous data (high H ), or vice versa. Thresholds define classes:
τ free C nav ( v ) < τ caut C nav v < τ free Free C nav v τ caut Blocked .
By design, regions with dynamic ambiguity (high p d y n or high H ρ ) shift to Caution/Blocked even if instantaneous geometry looks permissive; this is critical near human–robot interaction zones.

2.2.4. Incremental Map Expansion and Uncertainty Reduction

Industrial deployments require growing the known map while maintaining production throughput. Expansion is formalized via frontiers and information gain. A voxel is classified as a frontier if it is adjacent to unknown space:
F = v p u n k ( v ) > η u n k v N ( v ) : p u n k v η u n k .
The Expected Information Gain (EIG) for visiting neighborhood N along a path segment π is approximated by
E I G π = v N π κ v H v ,
where κ ( v ) [ 0 , 1 ] —visibility factor capturing whether the robot’s camera is likely to observe v given line-of-sight and view geometry (efficiently approximated using a ray bundle within a local frustum). Because moving objects can increase uncertainty over time, the Dirichlet update with forgetting (10) ensures revisits consolidate evidence: stable structures see p stat H , while persistently dynamic regions keep H high and can be deliberately avoided in production routes.
We embed exploration into planning via an information bonus that competes with entropy penalties (13). In “mapping mode”, EIG can dominate to promote discovery; in “production mode”, only marginal frontier bonuses remain, favoring low-risk, well-known corridors.

2.2.5. Multi-Criteria Path Optimization with Entropy Regularization

Given a start x 0 and goal x g , the optimal path π = x t t = 0 T is searched on a 3D grid using the cost function (13). The composite objective balances travel efficiency, smoothness, safety under uncertainty, and map expansion.
J π = t = 0 T w d x t + 1 x t + w κ κ curv x t + w nav C nav v x t w eig E I G local x t ,
where κ curv penalizes curvature (task-dependent smoothness), C nav is from (14), and E I G local is a local approximation of (17) computed over a sliding window around x t . Tuning:
The weights of the individual components in the cost function (18) are determined empirically based on the target processing pipeline, using sensitivity analysis on a simulated dataset. Table 1 summarizes the key characteristics and illustrates the effect of different weight configurations on overall pipeline performance.
The optimization problem J utilizes graph search via a sampling-based planner (e.g., RRT* with entropy-biased sampling density p sample ( x ) e γ C nav ( v ( x ) ) ). For dynamic scenes, 5–20 Hz receding-horizon replanning is implemented according to the VIO keyframe addition frequency.
The overall pipeline for multi-criteria path optimization with entropy regularization is summarized in Figure 4. The figure illustrates the sequential integration of perception, probabilistic mapping, information-gain evaluation, and entropy-biased planning within a receding-horizon control loop. Each stage transforms sensory observations and map priors into progressively refined navigation decisions, balancing exploration and safety in dynamic environments.

3. Results

To assess the effectiveness of the proposed entropy-regularized path-planning method, experiments evaluated mapping accuracy, computational efficiency, and robustness in dynamic environments. The evaluations were conducted both in simulation and on embedded hardware, specifically a Raspberry Pi 5 paired with a Hailo-8 AI Kit neural processing unit (NPU) capable of delivering up to 26 tera-operations per second (TOPS), as well as on a desktop-class system used as a performance reference. This experimental setup benchmarks not only the algorithm’s accuracy but also its suitability for deployment on resource-constrained platforms.
To validate the proposed method against precise ground truth, the NVIDIA Isaac Sim platform [31] was employed, providing photorealistic rendering and high-fidelity physics simulation. A digital twin of an industrial warehouse environment (20 × 20 m) was constructed, containing static obstacles (e.g., racks and machinery) as well as dynamic agents (e.g., forklifts and human workers) following randomized trajectories (speeds 0.5–2.0 m/s).

3.1. Accuracy of Entropy-Regularized Representation

The first set of experiments focused on validating the accuracy of the proposed entropy-based representation in distinguishing navigable from non-navigable regions. Table 2 compares the proposed approach against several baselines: (i) a standard occupancy grid without entropy regularization (Default); (ii) probabilistic occupancy mapping without explicit dynamic object modeling; and planning-based methods including the Dynamic Window Approach (DWA), Timed Elastic Band (TEB), and Model Predictive Path Integral (MPPI). For the latter three methods, mapping-related metrics reflect the performance of the underlying probabilistic baseline, while trajectory-related metrics capture how effectively each planner utilizes the available information for obstacle avoidance. All baseline planners were fine-tuned using grid search on a calibration dataset (five runs per method) to maximize success rate and mitigate hyperparameter bias.
Accuracy was evaluated using multiple complementary criteria. Static occupancy classification was assessed using the Intersection-over-Union (IoU) metric between predicted free/occupied regions and ground truth annotations. Dynamic object detection performance was quantified using the area under the receiver operating characteristic curve (AUC) for distinguishing moving from static entities. Trajectory accuracy was measured using the root mean squared error (RMSE) between planned paths and manually annotated ground truth safe trajectories. Finally, the trajectory safety rate is formally defined as the percentage of trials where the minimum Euclidean distance between the robot’s bounding box and any environmental obstacle remains above a strict safety threshold (0.1 m in our case) for the entire path.
The results indicate that entropy-regularized mapping consistently improves segmentation quality (static vs. dynamic), particularly in cluttered regions with partial occlusions. Crucially, these improvements in perception metrics (IoU and AUC) translate directly into enhanced navigation reliability. High IoU values ensure that narrow static passages are correctly identified, thereby reducing RMSE by preventing unnecessary detours. Meanwhile, improvements in AUC for dynamic classification correlate with a substantial increase in trajectory safety rate, rising from approximately 83% for probabilistic mapping to 92.7% for S3PM. In such environments, traditional occupancy grids tend to overestimate free space, while probabilistic baselines frequently misclassify ambiguous voxels. The proposed method mitigates these errors by weighting voxel states according to entropy, resulting in smoother and more reliable navigation maps.
Representative results are shown in Figure 5, which presents qualitative comparisons of reconstructed 3D point clouds and the corresponding planned trajectories. The visualizations highlight the ability of the entropy-regularized formulation to preserve navigability and spatial consistency even in the presence of significant motion and occlusion.
To further assess the robustness of the proposed approach, additional experiments were conducted in environments containing moving agents and other transient obstacles. These trials demonstrate the algorithm’s ability to maintain stable path-planning performance under dynamic conditions, in which both scene structure and obstacle configurations evolve over time (Figure 6).

3.2. SP3M Efficiency

The second evaluation concerns runtime and memory performance. The experiments were conducted on three platforms:
  • Raspberry Pi 5 (Broadcom BCM2712 ARM Cortex-A76 CPU @ 2.4GHz, 8 GB LPDDR4X RAM, manufactured by Sony UK Technology);
  • Raspberry Pi 5 + Hailo-8 AI Kit (same as 1 + 26 TOPS NPU acceleration, manufactured by Hailo.ai);
  • Desktop reference system (Intel Core i7-12700K CPU @ 3.6GHz, 32 GB DDR5 RAM, NVIDIA RTX 3080 GPU).
We measured runtime of each method step (optical flow estimation, residual flow computation, entropy update, distance transform maintenance, and path planning) averaged over 100 planning cycles in indoor navigation scenarios (Table 3).
The results show that while the desktop-class Core i7 executes the full pipeline comfortably above 30 Hz, the Raspberry Pi 5 operating alone sustains approximately 12–15 Hz. This performance remains within acceptable real-time bounds for applications involving slow motion, such as inspection tasks in unstructured industrial environments. With the addition of the Hailo-8 AI Kit, computationally intensive stages—including optical flow estimation and residual computation—are offloaded, increasing the overall throughput to 25–30 Hz. Memory usage is efficiently bounded through voxel hashing, with typical consumption remaining below 1 GB for a rolling map window of 20 × 20 × 6 m.

4. Conclusions

This paper introduced S3PM, a lightweight entropy-regularized framework for simultaneous mapping and path planning in dense 3D point clouds within unstructured and dynamic industrial environments. The key contribution lies in the integration of a dynamic-aware entropy field that fuses voxel-wise occupancy probabilities with motion evidence extracted from residual optical flow. This representation explicitly quantifies both geometric uncertainty and object dynamics, enabling the system to distinguish reliable free space from ambiguous or high-risk regions in real time.
Extensive experiments conducted in simulation and on real hardware—including a Raspberry Pi 5 with and without Hailo-8 NPU acceleration—demonstrate clear advantages over established baselines in terms of navigation robustness, safety, and computational efficiency:
  • An 18–27% higher IoU for static/dynamic segmentation;
  • A 0.94–0.97 AUC for motion detection;
  • A 30–45% reduction in collision events;
  • Sustained real-time performance of 12–15 Hz on the bare Raspberry Pi 5 and 25–30 Hz with NPU offloading;
  • Memory footprint below 1 GB for 20 × 20 × 6 m rolling maps.
These results confirm that entropy regularization, combined with incremental distance transforms and voxel hashing, significantly enhances navigation safety, localization robustness, and trajectory quality in cluttered environments with moving agents, while preserving the computational efficiency required for deployment on low-power embedded platforms.
Beyond its immediate industrial applicability, S3PM establishes a generalizable foundation for entropy-driven navigation strategies that can be extended to other high-uncertainty domains, such as search and rescue, planetary exploration, and service robotics in crowded public spaces. Future work will focus on integrating learned priors and multi-robot collaboration to further improve scalability and long-term autonomy.

Author Contributions

Conceptualization, A.S., O.K. and I.C.; methodology, A.S. and O.K.; investigation, O.K.; data curation, A.L., A.S. and O.K.; writing—original draft, O.K. and A.S.; writing—review and editing A.L., O.K., A.S. and I.C.; visualization, O.K.; supervision, A.S. and I.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to restrictions.

Acknowledgments

The authors (A.S., O.K., I.C. and A.L.) acknowledge the MASTERLY project (GA n.101091800) team, inspiring development of the Entropy-Regularized Path Planning Method.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Karaman, S.; Frazzoli, E. Sampling-Based Algorithms for Optimal Motion Planning. Int. J. Robot. Res. 2011, 30, 846–894. [Google Scholar] [CrossRef]
  2. Hornung, A.; Wurm, K.M.; Bennewitz, M.; Stachniss, C.; Burgard, W. OctoMap: An Efficient Probabilistic 3D Mapping Framework Based on Octrees. Auton. Robot. 2013, 34, 189–206. [Google Scholar] [CrossRef]
  3. Wijayathunga, L.; Rassau, A.; Chai, D. Challenges and Solutions for Autonomous Ground Robot Scene Understanding and Navigation in Unstructured Outdoor Environments: A Review. Appl. Sci. 2023, 13, 9877. [Google Scholar] [CrossRef]
  4. Elfes, A. Using Occupancy Grids for Mobile Robot Perception and Navigation. Computer 1989, 22, 46–57. [Google Scholar] [CrossRef]
  5. Sánchez-Ibáñez, J.R.; Pérez-del-Pulgar, C.J.; García-Cerezo, A. Path Planning for Autonomous Mobile Robots: A Review. Sensors 2021, 21, 7898. [Google Scholar] [CrossRef]
  6. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
  7. Wang, E.; Huo, W.; Xu, S.; Qu, P.; Na, L. Autonomous Navigation of Indoor Wheeled Robots Based on Improved Gmapping and Improved Bidirectional A*. Discov. Robot. 2025, 1, 7. [Google Scholar] [CrossRef]
  8. Faroni, M.; Pedrocchi, N.; Beschi, M. Adaptive Hybrid Local–Global Sampling for Fast Informed Sampling-Based Optimal Path Planning. Auton. Robot. 2024, 48, 6. [Google Scholar] [CrossRef]
  9. Huang, Y.; Huang, S.; Wang, H.; Meng, R. 3D Path Planning and Obstacle Avoidance Algorithms for Obstacle-Overcoming Robots. arXiv 2022, arXiv:2209.00871. [Google Scholar] [CrossRef]
  10. Cui, C.; Wang, Z.; Sui, J.; Zhang, Y.; Guo, C. An Improved RRT Behavioral Planning Method for Robots Based on PTM Algorithm. Sci. Rep. 2024, 14, 21776. [Google Scholar] [CrossRef]
  11. Zhang, Y.; Wang, H.; Yin, M.; Wang, J.; Hua, C. Bi-AM-RRT*: A Fast and Efficient Sampling-Based Motion Planning Algorithm in Dynamic Environments. IEEE Trans. Intell. Veh. 2023, 9, 1282–1293. [Google Scholar] [CrossRef]
  12. Xu, T. Recent Advances in Rapidly-Exploring Random Tree: A Review. Heliyon 2024, 10, e32451. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, L.; Li, P.; Qian, S.; Quan, H.; Miao, J.; Liu, M.; Hu, Y.; Memetimin, E. Path Planning Technique for Mobile Robots: A Review. Machines 2023, 11, 980. [Google Scholar] [CrossRef]
  14. Qin, H.; Shao, S.; Wang, T.; Yu, X.; Jiang, Y.; Cao, Z. Review of Autonomous Path Planning Algorithms for Mobile Robots. Drones 2023, 7, 211. [Google Scholar] [CrossRef]
  15. Abu-Jabal, N.; Baziyad, M.; Fareh, R.; Brahmi, B.; Rabie, T.; Bettayeb, M. A Comprehensive Study of Recent Path-Planning Techniques in Dynamic Environments for Autonomous Robots. Sensors 2024, 24, 8089. [Google Scholar] [CrossRef] [PubMed]
  16. Rösmann, C.; Hoffmann, F.; Bertram, T. Integrated Online Trajectory Planning and Optimization in Distinctive Topologies. Robot. Auton. Syst. 2017, 88, 142–153. [Google Scholar] [CrossRef]
  17. Roesmann, C.; Feiten, W.; Woesch, T.; Hoffmann, F.; Bertram, T. Trajectory Modification Considering Dynamic Constraints of Autonomous Robots. In Proceedings of the German Conference on Robotics (ROBOTIK 2012), Munich, Germany, 21–22 May 2012; pp. 1–6. [Google Scholar]
  18. Szabolcsi, R. Model Predictive Control Applied in UAV Flight Path Tracking Missions. Rev. Air Force Acad. 2019, 17, 49–62. [Google Scholar] [CrossRef]
  19. Williams, G.; Drews, P.; Goldfain, B.; Rehg, J.M.; Theodorou, E.A. Aggressive Driving with Model Predictive Path Integral Control. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–20 August 2016. [Google Scholar] [CrossRef]
  20. Fox, D.; Burgard, W.; Thrun, S. The Dynamic Window Approach to Collision Avoidance. IEEE Robot. Autom. Mag. 1997, 4, 23–33. [Google Scholar] [CrossRef]
  21. Rosinol, A.; Violette, A.; Abate, M.; Hughes, N.; Chang, Y.; Shi, J.; Gupta, A.; Carlone, L. Kimera: From SLAM to Spatial Perception with 3D Dynamic Scene Graphs. Int. J. Robot. Res. 2021, 40, 1510–1546. [Google Scholar] [CrossRef]
  22. Tai, L.; Paolo, G.; Liu, M. Virtual-To-Real Deep Reinforcement Learning: Continuous Control of Mobile Robots for Mapless Navigation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018. [Google Scholar] [CrossRef]
  23. Costante, G.; Ciarfuglia, T.A. LS-VO: Learning Dense Optical Subspace for Robust Visual Odometry Estimation. IEEE Robot. Autom. Lett. 2018, 3, 1735–1742. [Google Scholar] [CrossRef]
  24. Dellaert, F.; Kaess, M. Factor Graphs for Robot Perception. Found. Trends Robot. 2017, 6, 1–139. [Google Scholar] [CrossRef]
  25. Ishida, H.; Tanaka, H.; Taniguchi, H.; Moriizumi, T. Mobile Robot Navigation Using Vision and Olfaction to Search for a Gas/Odor Source. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Sendai, Japan, 28 September–2 October 2004; IEEE: Piscataway, NJ, USA, 2005; pp. 3978–3983. [Google Scholar] [CrossRef]
  26. Khan, D.; Cheng, Z.; Uchiyama, H.; Ali, S.; Asghar, M.; Kiyokawa, K. Recent Advances in Vision-Based Indoor Navigation: A Systematic Literature Review. Comput. Graph. 2022, 104, 24–45. [Google Scholar] [CrossRef]
  27. Zhang, Y.; Chen, P. Path Planning of a Mobile Robot for a Dynamic Indoor Environment Based on an SAC-LSTM Algorithm. Sensors 2023, 23, 9802. [Google Scholar] [CrossRef]
  28. Thomas, H.; Zhang, J.; Barfoot, T.D. The Foreseeable Future: Self-Supervised Learning to Predict Dynamic Scenes for Indoor Navigation. arXiv 2022, arXiv:2208.12602. [Google Scholar] [CrossRef]
  29. Slavova, A.; Hristov, V. Mapless Navigation with Deep Reinforcement Learning in Indoor Environment. Eng. Proc. 2025, 100, 63. [Google Scholar] [CrossRef]
  30. Francis, A.G.; Faust, A.; Chiang, H.-T.L.; Hsu, J.; Kew, J.C.; Fišer, M.; Lee, T.-W.E. Long-Range Indoor Navigation with PRM-RL. IEEE Trans. Robot. 2020, 36, 1115–1134. [Google Scholar] [CrossRef]
  31. NVIDIA Isaac Sim. NVIDIA Corporation: Santa Clara, CA, USA, 2025. Available online: https://developer.nvidia.com/isaac/sim (accessed on 7 January 2026).
Figure 1. Entropy field of a point cloud (df) for the corresponding RGB scene representation (ac): blue— 0.0 , green—0.5, red— 1.0 .
Figure 1. Entropy field of a point cloud (df) for the corresponding RGB scene representation (ac): blue— 0.0 , green—0.5, red— 1.0 .
Sensors 26 00731 g001
Figure 2. Optical flow field of a point cloud (df) for the corresponding RGB scene representation (ac): blue— l e s s   t h a n   0.1 m/s; green—2.2 m/s; red— m o r e   t h a n   4.2 m/s (or unknown).
Figure 2. Optical flow field of a point cloud (df) for the corresponding RGB scene representation (ac): blue— l e s s   t h a n   0.1 m/s; green—2.2 m/s; red— m o r e   t h a n   4.2 m/s (or unknown).
Sensors 26 00731 g002
Figure 3. Dynamic-aware entropy field of a point cloud (df) for the corresponding RGB scene representation (ac): blue— 0.0 ; green—0.5; red— 1.0 .
Figure 3. Dynamic-aware entropy field of a point cloud (df) for the corresponding RGB scene representation (ac): blue— 0.0 ; green—0.5; red— 1.0 .
Sensors 26 00731 g003
Figure 4. Flowchart of the proposed multi-criteria entropy-regularized path planning method.
Figure 4. Flowchart of the proposed multi-criteria entropy-regularized path planning method.
Sensors 26 00731 g004
Figure 5. Qualitative evaluation of mapping accuracy and path planning in dynamic environments: (a) Original RGB image from the current camera frame with the marked forklift; (b) Comparison of reconstructed occupancy maps over the first 100 frames (probabilistic occupancy grid vs. the proposed S3PM method); (c) Point cloud and planned trajectory generated using probabilistic occupancy grid with RRT* planner; (d) Point cloud and planned trajectory generated using S3PM with entropy-regularized RRT*. Color legend: violet cube—dynamic object; dash green line—planned trajectory; solid red line—executed trajectory.
Figure 5. Qualitative evaluation of mapping accuracy and path planning in dynamic environments: (a) Original RGB image from the current camera frame with the marked forklift; (b) Comparison of reconstructed occupancy maps over the first 100 frames (probabilistic occupancy grid vs. the proposed S3PM method); (c) Point cloud and planned trajectory generated using probabilistic occupancy grid with RRT* planner; (d) Point cloud and planned trajectory generated using S3PM with entropy-regularized RRT*. Color legend: violet cube—dynamic object; dash green line—planned trajectory; solid red line—executed trajectory.
Sensors 26 00731 g005
Figure 6. Top-down view of occupancy and trajectory distributions under dynamic conditions: (a) Top-down projection showing trajectory and occupancy distribution using probabilistic occupancy grid; (b) Top-down projection showing trajectory and occupancy distribution using S3PM with entropy-regularized mapping. Color legend: violet—dynamic object; red—executed trajectory.
Figure 6. Top-down view of occupancy and trajectory distributions under dynamic conditions: (a) Top-down projection showing trajectory and occupancy distribution using probabilistic occupancy grid; (b) Top-down projection showing trajectory and occupancy distribution using S3PM with entropy-regularized mapping. Color legend: violet—dynamic object; red—executed trajectory.
Sensors 26 00731 g006
Table 1. Impact of cost function weights on pipeline performance.
Table 1. Impact of cost function weights on pipeline performance.
w d w κ w nav   w eig   Path Length [m]Minimum Obstacle
Distance [m]
Newly Explored Voxels
[Count (105) per Keyframe]
* Path Smoothness [rad]
0.250.250.250.2554.20.480.210.42
0.400.200.200.2053.00.420.160.47
0.200.400.200.2055.00.510.180.28
0.200.200.400.2055.71.230.150.45
0.200.200.200.4056.90.400.340.52
0.400.200.200.2052.60.370.130.62
0.200.400.300.1057.01.060.180.25
0.150.150.150.5558.80.340.410.69
0.300.300.200.2053.90.450.170.31
* Path smoothness is quantified as the average angular deviation between consecutive path segments, measured in radians. Lower values indicate smoother trajectories with fewer abrupt changes in direction.
Table 2. Performance comparison of the proposed S3PM and baseline methods. Values in parentheses represent the relative change compared to the “Default” baseline.
Table 2. Performance comparison of the proposed S3PM and baseline methods. Values in parentheses represent the relative change compared to the “Default” baseline.
MethodMap Accuracy
(IOU [%])
Map Awareness
(AUC [%])
Trajectory Accuracy (RMSE [m])Trajectory Safety Rate
[%]
Default71.4076.200.5452.30
Probabilistic82.60 (+15.69%)85.10 (+11.68%)0.41 (−24.07%)83.12 (+58.93%)
DWA80.20 (+12.32%)81.10 (+ 6.68%)0.49 (− 9.26%)78.20 (+49.52%)
TEB83.40 (+16.81%)86.20 (+13.12%)0.35 (−35.18%)88.50 (+69.22%)
MPPI85.10 (+19.19%)88.70 (+16.40%)0.37 (−31.48%)94.10 (+79.92%)
S3PM (ours)89.80 (+25.77%)92.40 (+21.26%)0.38 (−29.63%)92.70 (+77.25%)
Table 3. Computational performance breakdown of S3PM on embedded and desktop systems.
Table 3. Computational performance breakdown of S3PM on embedded and desktop systems.
S3PM StepRaspberry Pi 5Raspberry Pi 5 + Hailo-8Desktop
Optical Flow Estimation [ms]22.56.84.1
Ego-motion Compensation [ms]5.73.92.6
Dynamic Likelihood & Voting [ms]18.27.45.3
Entropy Update & Regularization [ms]14.66.24.8
Distance Transform [ms]9.85.33.1
Path Planning [ms]11.45.94.2
Average Runtime per Cycle [ms]82.235.524.1
Average Frequency [Hz]12.228.141.5
Idle Power [W] (System only)3.13.445.2
Average Power Consumption (S3PM) [W]6.88.2127.9
Peak Power Consumption (S3PM) [W]9.411.6210.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sazonov, A.; Kuchkin, O.; Cherepanska, I.; Lipnickas, A. S3PM: Entropy-Regularized Path Planning for Autonomous Mobile Robots in Dense 3D Point Clouds of Unstructured Environments. Sensors 2026, 26, 731. https://doi.org/10.3390/s26020731

AMA Style

Sazonov A, Kuchkin O, Cherepanska I, Lipnickas A. S3PM: Entropy-Regularized Path Planning for Autonomous Mobile Robots in Dense 3D Point Clouds of Unstructured Environments. Sensors. 2026; 26(2):731. https://doi.org/10.3390/s26020731

Chicago/Turabian Style

Sazonov, Artem, Oleksii Kuchkin, Irina Cherepanska, and Arūnas Lipnickas. 2026. "S3PM: Entropy-Regularized Path Planning for Autonomous Mobile Robots in Dense 3D Point Clouds of Unstructured Environments" Sensors 26, no. 2: 731. https://doi.org/10.3390/s26020731

APA Style

Sazonov, A., Kuchkin, O., Cherepanska, I., & Lipnickas, A. (2026). S3PM: Entropy-Regularized Path Planning for Autonomous Mobile Robots in Dense 3D Point Clouds of Unstructured Environments. Sensors, 26(2), 731. https://doi.org/10.3390/s26020731

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop