Next Article in Journal
Quantitative Assessment of Aerosol Leakage in Protective Clothing During Nursing Tasks: The Impact of Body Morphology and Pumping Effects
Previous Article in Journal
Am I Top of the Pops? Does Feedback of Live GPS Between Sets of Hurling-Specific Small-Sided Games Improve Subsequent Running and Physiological Performance?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PFMS-RRT*: A Progress-Aware Fused-Sampling RRT* with Multi-Level Strategy Extension for Path Planning

1
College of Oceanography and Space Informatics, China University of Petroleum (East China), Qingdao 266580, China
2
Qingdao Institute of Software, College of Computer Science and Technology, China University of Petroleum (East China), Qingdao 266580, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(6), 3107; https://doi.org/10.3390/app16063107
Submission received: 11 February 2026 / Revised: 12 March 2026 / Accepted: 20 March 2026 / Published: 23 March 2026

Abstract

Sampling-based planners such as RRT* are attractive for robot navigation in complex spaces, but they often suffer from high randomness, low efficiency, slow convergence, and suboptimal path quality in cluttered environments. To address these limitations, this paper proposes PFMS-RRT*, a progress-aware fused-sampling RRT* with a multi-level strategy extension. The method builds on a bidirectional RRT* framework and introduces three main components: (i) a progress-aware fused sampling scheme that adapts an oriented elliptical sampling region based on inter-tree progress and stagnation, mixes locally guided elliptical samples with globally explorative Halton-sequence samples, and dynamically balances exploration and exploitation; (ii) a three-level goal-guided extension mechanism that escalates from direct steering to local probing and then multi-direction detours to maintain forward progress when obstacles block expansion; and (iii) a smooth tangential artificial potential field (APF) extension used as a fallback, with a failure-driven probabilistic switching rule that increases APF usage after repeated extension failures. Simulations in four representative 2D environments (sparse, corridor-like dense, random dense, and narrow passage) show that PFMS-RRT* consistently yields shorter paths, lower and more stable runtime, and fewer nodes than several RRT* variants while maintaining competitive or improved obstacle clearance.

1. Introduction

With the advancement of unmanned technology, mobile robots have attracted widespread attention in the fields of field exploration [1], disaster relief [2], and agricultural production [3]. Path planning aims to generate a collision-free path from the start to the goal, and it is one of the core problems in mobile robotics [4].
Based on their characteristics, path planning algorithms can be categorized into three major types. The first type includes traditional planning algorithms, such as graph-search algorithms [5] and sampling-based algorithms [6]. The second type consists of swarm intelligence optimization algorithms, such as particle swarm optimization (PSO) [7] and genetic algorithms (GA) [8]. The third type comprises learning-based methods, such as reinforcement learning [9].
Due to their outstanding performance in high-dimensional and complex environments, sampling-based algorithms have attracted considerable attention from researchers. Rapidly-exploring Random Tree (RRT) [10] is a classical representative of sampling-based methods. It features a simple and efficient structure as well as probabilistic completeness. However, because it grows through random iterations in the search space and does not explicitly consider path cost or quality, the resulting paths are often complex and of poor quality. RRT* [11] is an extension of RRT that incorporates parent selection and node rewiring strategies, significantly improving path quality. Nevertheless, RRT* still suffers from slow planning speed and long convergence time. To further accelerate the convergence of RRT*, many improved variants have been proposed.
Some improvement approaches focus on optimizing the sampling stage of RRT*, thereby improving algorithmic efficiency. Informed RRT* adopts a non-uniform sampling strategy [12] that reduces search blindness by restricting the sampling region and decreasing redundant samples. SOF-RRT* introduces a spatial probability-weighted sampling strategy [13], assigning higher sampling probability to regions with larger feasible areas. This reduces redundant sampling and improves the effective sampling rate. F-RRT* algorithm optimizes the path cost by creating an exclusive parent node for the random sampling point, rather than selecting a parent node from the existing set of vertices [14]. NRRT* [15] and NIRRT* [16] use neural networks to predict sampling points and perform biased sampling concentrated around these points. Although these algorithms enhance the sampling stage, they lack improvements to the extension stage, which limits their performance in complex environments.
To further improve performance, many algorithms refine RRT* at the extension stage. BI-RRT* adopts a bidirectional exploration mechanism [17], constructing search trees from the start and the goal to improve planning efficiency. E-RRT* replaces line segments with ellipsoids when connecting adjacent nodes, improving performance in narrow environments [18]. FC-RRT* performs bidirectional extension using prior task information [19]. FN-RRT* constrains the tree size using a fixed-node-number control mechanism [20], thereby accelerating convergence. Although these RRT* variants significantly improve extension strategies, they pay insufficient attention to sampling and do not introduce targeted sampling improvements, leading to suboptimal planning performance in scenarios with complex obstacle distributions.
There are also approaches that integrate other techniques and algorithms. IA-RRT* modifies the cost evaluation function of the A* algorithm [21], providing a relatively weak guidance direction for the search. This narrows the search region and prevents the algorithm from converging too quickly and becoming trapped in a locally optimal path. GB-RRT* combines a grid-based search strategy and proposes a minimum-dose path planning scheme suitable for complex environments with obstacles and narrow passages [22]. VPS-APF-RRT* incorporates an improved artificial potential field (APF), further enhancing the algorithm’s obstacle-avoidance capability [23].
To address the issues of high randomness, low efficiency, long search time, and poor path quality of RRT* in complex environments, a Progress-aware Fused Sampling and multi-level strategy Extension RRT* algorithm, namely PFMS-RRT*, is proposed. The main improvements of the proposed algorithm are summarized as follows:
(1)
A progress-aware fused sampling strategy that adaptively reshapes the sampling distribution according to planning progress, reducing redundant samples and improving the probability of generating informative samples.
(2)
A three-level goal-guided extension mechanism with progressively stronger fallback behaviors to improve extension success rate and search stability in cluttered environments.
(3)
A smooth tangential APF-based extension method as a fallback when goal-guided extension is ineffective, which alleviates stagnation near obstacles, reduces invalid expansions, and enhances extension robustness in cluttered environments.

2. Related Works

This section provides a comprehensive overview of the RRT* algorithm, which serves as the foundation for all algorithms proposed in this paper. Subsequently, an in-depth exploration of the Bi-RRT*, APF-RRT*, and Informed RRT* algorithms will be conducted to lay the groundwork for introducing the PFMS-RRT* algorithm.

2.1. RRT*

RRT* is an improved sampling-based motion planning algorithm derived from the classic Rapidly-exploring Random Tree (RRT), designed for path planning in high-dimensional environments with complex obstacles. Unlike RRT, which aims to quickly find a feasible path, RRT* preserves the global exploration capability brought by random sampling while introducing rewiring and choose parent mechanisms. As a result, the generated path is continuously optimized over iterations and enjoys an important property: asymptotic optimality—that is, as the number of samples approaches infinity, RRT* converges to the optimal path with probability 1.
The overall procedure of the RRT* algorithm is shown in Algorithm 1. To facilitate understanding, an explanation of the pseudocode is provided below:
  • Sample: In each iteration, a random state x rand is generated in the state space.
  • Nearest: Given the sampled state x rand , the algorithm searches the current vertex set V ( T ) and finds the closest vertex x nearest according to a predefined distance metric.
  • Steer: Starting from x nearest , the algorithm steers toward x rand to generate a new candidate state x new . In practice, the extension is usually limited by a maximum step size η to ensure incremental growth and stable collision checking.
  • Free: The function Free checks whether the local path connecting x nearest to x new lies entirely in the obstacle-free region. If a collision is detected, the candidate is rejected and the algorithm proceeds to the next iteration.
  • ChooseParent: After obtaining x new , the algorithm identifies a set of nearby vertices X near = Near ( x new , V ( T ) ) within a neighborhood radius. Among these candidates, ChooseParentselects the parent vertex x parent X near that minimizes the total cost from the start to x new . If no better candidate exists, x nearest is used as the parent.
  • Rewire: Once x new is added to the tree, Rewire attempts to improve the existing structure by reconnecting vertices in X near . For each x X near , if routing through x new yields a lower cost and the edge ( x new , x ) is collision-free, then the parent of x is updated to x new .
Algorithm 1: RRT*
Applsci 16 03107 i001

2.2. Bi-RRT*

Bi-RRT* generates two random trees from the start state and the goal state, respectively, and searches by expanding them toward each other. Owing to its bidirectional search mechanism, Bi-RRT* expands more efficiently than single-tree methods, thereby producing a feasible path more quickly. The pseudocode of this algorithm is shown in Algorithm 2.
Algorithm 2: Bi-RRT*
Applsci 16 03107 i002

2.3. Informed-RRT*

Informed-RRT* improves upon RRT* by refining the sampling strategy: instead of sampling over the entire map, it focuses on an elliptical region defined by the start and goal states. This informed, targeted sampling substantially reduces the search space, thereby accelerating convergence toward the optimal solution and improving overall sampling efficiency. The detailed procedure is presented in Algorithm 3.
The specific implementation of the SampleInformed function in Algorithm 3 is given in Algorithm 4.
Algorithm 3: Informed-RRT*
Applsci 16 03107 i003
Algorithm 4: Sample Informed
Applsci 16 03107 i004

2.4. APF-RRT*

APF-RRT* is an improved variant of RRT* that incorporates the guiding idea of the Artificial Potential Field method. During the sampling or expansion stage, it introduces a goal-attractive force and an obstacle-repulsive force, causing the random tree to grow preferentially toward the goal while avoiding obstacles, thereby reducing blind exploration and ineffective sampling.

3. PFMS-RRT* Algorithm

As shown in Figure 1, PFMS-RRT* inherits the dual-tree expansion mechanism of Bi-RRT* and introduces three core improvements: (1) in the sampling phase, progress-aware fused sampling replaces random sampling to generate informative samples; (2) in the expansion phase, a three-level goal-guided extension mechanism is proposed to ensure robust tree growth of the algorithm in cluttered environments with obstacles; and (3) an improved smooth tangential APF is integrated as a fallback extension strategy when the goal-guided extension fails. In addition, a failure-driven probabilistic switching scheme adaptively increases the chance of invoking the APF extender when repeated extension failures are observed. The pseudocode of PFMS-RRT* is presented in Algorithm 5. Firstly, the start tree T _ s and goal tree T _ g are initialized. the progress metric d prev is initialized to the Euclidean distance between the start state x start and the goal state x goal , and the stagnation countersis reset to 0. Secondly, an informative sample x is generated via the progress-aware fused sampling strategy. Thirdly, the extension strategy switching probability p sw ( X ) driven by the failure metric is updated. According to p sw ( X ) , the algorithm determines whether to skip the three-level goal-guided extension. If it is skipped, the smooth tangential APF is directly employed for extension. Otherwise, the three-level goal-guided extension is executed first, and the smooth tangential APF is used as a fallback strategy if the extension fails. The following sections provide a detailed introduction to each stage.

3.1. Progress-Aware Fused Sampling

Progress-aware fused sampling dynamically concentrates samples in regions that are more likely to reduce the separation between the two trees while preserving global exploration. At each iteration, the nearest inter-tree vertex pair ( x s * , x g * ) is computed, and the current progress metric is defined as
d k = x s * x g * 2 .
where T s and T g denote the start and goal trees, respectively; ( x s * , x g * ) is the nearest inter-tree vertex pair (one vertex from each tree); x s * V ( T s ) and x g * V ( T g ) ; · 2 denotes the Euclidean norm; d k is the inter-tree separation (progress metric) at iteration k. If d k decreases compared with the previous value d prev , the planner is considered to make progress and the stagnation counter s is reset; otherwise, s is increased by one.
Figure 1. Overall diagram of PFMS-RRT*.
Figure 1. Overall diagram of PFMS-RRT*.
Applsci 16 03107 g001
Algorithm 5: PFMS-RRT* x start , x goal , O , K max
Applsci 16 03107 i005
To adapt the sampling domain to the planning progress, as shown in Figure 2, an oriented elliptical region is constructed whose semi-axes a k and b k are scaled by d k and an additional stagnation amplification factor γ k (line 11–12 in Algorithm 6). When the stagnation counter exceeds a threshold s th , a stagnation level k is increased (line 8–10 in Algorithm 6), which enlarges γ k and thus expands the ellipse. This expansion encourages broader exploration and helps the planner escape local trapping near cluttered obstacles.
To mitigate the bias induced by a fixed ellipse center, as shown in Figure 3, the center c k is randomly switched (with probability 0.5) between (i) the midpoint of the nearest inter-tree pair and (ii) the midpoint of the start–goal pair. The ellipse orientation ϕ k aligns with the corresponding pair direction. A locally guided sample q ell is then drawn uniformly inside the oriented ellipse (line 13–20 in Algorithm 6).
Meanwhile, a globally explorative sample q g is generated using a Halton low-discrepancy sequence, which improves space-filling properties over purely random sampling. The sequence index is denoted by n N + and is incremented by one each time a global sample is generated. The n-th Halton point is defined as
( u n , v n ) = H 2 ( n ) , H 3 ( n ) ( 0 , 1 ) 2 ,
where H b ( n ) denotes the Halton radical-inverse function in base b. The point ( u n , v n ) is mapped to the workspace bounding box [ x min , x max ] × [ y min , y max ] as
x n = x min + u n ( x max x min ) , y n = y min + v n ( y max y min ) .
No scrambling is applied. Skipping is used only through different initial indices for the two trees: the start tree uses n 0 s = 1 and the goal tree uses n 0 g = 5000 .
The final sample q is selected from { q ell , q g } by a mixing probability p g , k (line 27–31 in Algorithm 6). To strengthen global exploration during stagnation, p g , k increases with the stagnation level k , thereby achieving an adaptive balance between fast local convergence and escaping local minima.
Algorithm 6: Progress Aware Fusion Sample ( T s , T g , d prev , s )
Applsci 16 03107 i006
   Notation for Algorithm 6.  V ( T ) denotes the vertex set of a tree T. d prev is the progress metric value from the previous iteration. s is the stagnation counter. s th is the stagnation threshold used to trigger stagnation-level increase. k is the stagnation level at iteration k. γ k is the stagnation amplification factor, bounded by γ max . (If the implementation uses a slope coefficient, denote it explicitly as γ k and use γ k = min ( γ max , 1 + γ k k ) .) a k and b k are the semi-major and semi-minor axes of the oriented elliptical sampling region. a min , a max and b min , b max are the lower/upper bounds of a k and b k , respectively. c k ( p ) = 1 2 ( x s * + x g * ) is the midpoint of the nearest inter-tree pair. c k ( s g ) = 1 2 ( x start + x goal ) is the midpoint of the start–goal pair. c k is the ellipse center, randomly chosen between c k ( p ) and c k ( s g ) with probability 0.5 . φ k is the ellipse orientation angle computed by atan 2 ( · , · ) . R ( φ k ) is the 2 × 2 planar rotation matrix with angle φ k . u 1 , u 2 U ( 0 , 1 ) are i.i.d. uniform random variables; r = u 1 and θ = 2 π u 2 are polar coordinates for uniform sampling in an ellipse. q is the locally guided sample drawn uniformly inside the oriented ellipse. n is the index of the Halton sequence. H 2 ( n ) and H 3 ( n ) are the 2nd and 3rd dimensions of the Halton sequence, producing u n and v n . x min , x max , y min , y max define the workspace bounding box. q g ( n ) is the global sample generated by the Halton sequence mapping to the workspace. p g , k is the mixing probability for choosing the global sample, with increment coefficient Δ p . p g , 0 > 0 is a small global-sampling probability. q is the final sample selected from { q , q g } .

3.2. Three-Level Goal-Guided Extension Mechanism

After obtaining the sample in Section 3.1, the tree is expanded using the three-level goal-guided extension mechanism. The three-level goal-guided extension is the core strategy of PFMS-RRT* in the expansion phase, which aims to efficiently generate collision-free nodes toward the goal direction through hierarchical exploration. Its pseudocode is presented in Algorithm 7, and the core workflow is outlined below.
The tree is expanded using a three-level goal-guided extension mechanism, which follows a coarse-to-fine strategy to balance fast convergence and robust obstacle avoidance. Given the current node x k , the guiding target g, and obstacles O, as shown in Figure 4, the first level performs the fastest direct goal steering by advancing one step along the unit direction h ^ k = ( g x k ) / g x k 2 , i.e.,  x k + 1 ( 1 ) = x k + η h ^ k . If the segment ( x k , x k + 1 ( 1 ) ) is collision-free, the new node is accepted immediately.
As shown in Figure 5, when direct steering is blocked, the second level conducts a local small-range probing search. First, the lateral shift direction s { 1 , + 1 } is selected, where s = 1 corresponds to a leftward shift and s = 1 corresponds to a rightward shift. Given the lateral displacement set Δ = { 0.3 η , 0.8 η } and the forward advance set Λ = { 0.6 η , η } , the lateral shift base point is first constructed:
x base ( s , δ ) = x k + s δ v ^ k
then the candidate point is obtained:
x try ( s , δ , λ ) = x base ( s , δ ) + λ h ^ k
where x k is the current node (state) to be extended; g is the guiding target (typically the sampled state or the opposite-tree target); h ^ k = ( g x k ) / g x k 2 is the unit direction from x k to g; v ^ k = [ h ^ k , y , h ^ k , x ] is the lateral unit vector orthogonal to h ^ k ; s { 1 , + 1 } indicates left ( 1 ) or right ( + 1 ) lateral shift; δ Δ is the lateral displacement magnitude, and  λ Λ is the forward step size; η is the nominal extension step length.
A pure lateral shift may cause the expansion tree to drift back and forth locally. Therefore, the forward progress along the goal direction is defined as
prog ( x try ) = x try x k , h ^ k
with the constraint
prog ( x try ) τ η
This constraint ensures that the extension bypasses obstacles while continuing to approach the goal, rather than performing a purely lateral movement. A one-step look-ahead point is added in front of the candidate point:
x look = x try + η h ^ k
and define
L = Free ( x try , x look ) { 0 , 1 }
where · , · denotes the Euclidean inner product; prog ( x try ) measures the forward progress of candidate x try along h ^ k ; τ ( 0 , 1 ] is a progress threshold coefficient enforcing minimum forward progress τ η ; x look is a one-step look-ahead point used for feasibility checking; L is a binary feasibility indicator.
By comprehensively considering the goal distance, progress effect, look-ahead feasibility, and lateral shift cost, an evaluation function J ( x try ) is constructed. The candidate that minimizes this function among all feasible candidates is selected as x k + 1 ( 2 ) :
J ( x try ) = g x try 2 w p · prog ( x try ) w L · L η + w δ · δ
x k + 1 ( 2 ) = arg min s { ± 1 } , δ Δ , λ Λ F ( x try ( s , δ , λ ) ) J ( x try ( s , δ , λ ) )
where J ( x try ) is the candidate evaluation function; w p , w L , and  w δ are nonnegative weights for progress reward, look-ahead feasibility reward, and lateral-shift penalty, respectively; g x try 2 is the Euclidean distance to the guiding target; Δ and Λ are discrete sets of lateral and forward step candidates. In (9), x k + 1 ( 2 ) is the best second-level candidate selected by minimizing J ( · ) over all feasible ( s , δ , λ ) combinations. The overall procedure is detailed in Lines 7–24 of Algorithm 7.
As shown in Figure 6, if no feasible node is found in the second level, the third level activates a stronger multi-direction detour fallback by sampling multiple rotated directions d ^ ( θ , σ ) = R ( σ θ ) h ^ k with discrete angles Θ and step sizes S , Θ denotes a discrete set of steering angles around the goal direction, and  S denotes a discrete set of step sizes used to generate candidate detour motions. Among all collision-free candidates, the one minimizing J det ( x t r y ) = g x t r y 2 + λ θ | θ | is chosen. This three-level mechanism improves the extension success rate and maintains continuous forward progress in cluttered environments.
Algorithm 7: Three Level Goal Guided Extend ( x k , g , O )
Applsci 16 03107 i007

3.3. Smooth Tangential APF Extension

To improve extension stability and obstacle-avoidance efficiency near obstacles, a smooth tangential artificial potential field is introduced on top of the conventional APF. The conventional APF consists of an attractive potential U att , a repulsive potential U rep , and the total potential U total , where the acting forces are given by the negative gradients of the corresponding potentials. However, the repulsive field may vary sharply around the boundary of the obstacle influence region, which can cause abrupt changes in the resultant force direction, leading to jittery extensions and ineffective samples. In narrow passages or near concave obstacles, the planner may also suffer from local minima or inefficient detours, thus degrading the extension success rate. The pseudocode of the smooth tangential APF extension is shown in Algorithm 8.
For each obstacle o, a smoothing factor is defined as
s o ( x ) = min max ρ 0 ρ o ( x ) ρ 0 , 0 , 1 ,
w o ( x ) = s o 2 ( x ) ,
where ρ o ( x ) denotes the distance from point x to obstacle o. In this work, this distance is computed based on the exact geometric boundary of each obstacle. ρ 0 > 0 is the obstacle influence radius (threshold distance), s o ( x ) [ 0 , 1 ] is a smoothing factor, and  w o ( x ) = s o 2 ( x ) is its squared weight, so that the tangential effect changes continuously at the boundary of the influence region. Let n ^ o ( x ) denote the outward unit normal at the expansion point with respect to obstacle o, and define the tangential unit vector as
t ˜ o ( x ) = n ^ o , y ( x ) n ^ o , x ( x ) .
where n ^ o ( x ) is the outward unit normal direction at x with respect to obstacle o; t ˜ o ( x ) is the tangential unit vector obtained by rotating n ^ o ( x ) by 90 ; t ^ o ( x ) is the final tangential direction chosen to be consistent with the goal-guiding direction h ^ k . To avoid moving along an undesired tangential direction, the sign is selected according to the goal-guiding direction h ^ k , yielding the final tangential direction t ^ o ( x ) . The tangential force is then defined by
F tan , o ( x ) = k tan w o ( x ) 1 ρ o ( x ) 1 ρ 0 t ^ o ( x ) ,
F tan ( x ) = o O F tan , o ( x ) ,
The resultant force is
F ( x ) = F att ( x ) + F rep ( x ) + F tan ( x ) ,
u ^ ( x ) = F ( x ) F ( x ) 2 ,
where k tan is the tangential force gain, k r is the repulsive force gain, and k a is the attractive force gain. F tan , o ( x ) is the tangential force induced by obstacle o and F tan ( x ) = o O F tan , o ( x ) . F att ( x ) and F rep ( x ) are the attractive and repulsive forces, respectively; F ( x ) is the resultant force; and u ^ ( x ) = F ( x ) / F ( x ) 2 is its unit direction. η is the step length used to generate the next node x k + 1 = x k + η u ^ ( x ) .
Algorithm 8: Smooth Tangential APF Extend ( x k , g , O )
Applsci 16 03107 i008

3.4. Failure-Driven Switching Probability Update

PFMS-RRT* integrates two extension strategies: three-level goal-guided extension mechanism and smooth tangential APF extension. The three-level goal-guided extension mechanism mainly advances toward the guiding target and converges faster in open areas, but it may suffer frequent collisions and stagnation in cluttered regions or around narrow-passage entrances. In contrast, the smooth tangential APF extension introduces a smooth tangential component within the obstacle influence region, enabling boundary-following motion instead of being pushed away purely by normal repulsion, and is thus more robust in locally obstacle-constrained regions. Based on this complementarity, a failure-driven adaptive switching scheme is adopted: when repeated failures indicate that goal-guided probing is ineffective locally, the probability of activating the smooth tangential APF extension is increased to enhance obstacle avoidance and restore forward progress. The failure-driven adaptive switching scheme is shown in Algorithm 9.
A binary success indicator S ( X ) { 0 , 1 } is defined for each extension attempt, where S ( X ) = 1 denotes success and S ( X ) = 0 denotes failure (i.e., returning ⌀). A failure counter f ( X ) is maintained as
f ( X ) 0 , S ( X ) = 1 , f ( X ) + 1 , S ( X ) = 0 .
The counter is then mapped to a probability term via an exponential decay model
p g ( X ) = exp β f ( X ) , β > 0 ,
and the switching probability is defined as
p s w ( X ) = 1 p g ( X ) .
where f ( X ) is the accumulated failure counter; β > 0 is the exponential decay rate controlling the sensitivity to failures; p g ( X ) is the probability term mapped from the failure counter; p s w ( X ) is the switching probability of invoking the APF-based extension.
Consequently, accumulated failures monotonically increase p s w ( X ) , making the planner more likely to invoke smooth tangential APF extension. Once a feasible extension is found, f ( X ) is reset and p s w ( X ) drops quickly, allowing for the planner to return to the more efficient three-level goal-guided extension mechanism.
Algorithm 9: Update Switch Prob ( f ( X ) , S ( X ) )
Applsci 16 03107 i009

3.5. Probabilistic Completeness and Asymptotic Optimality Discussion

Let the state of PFMS-RRT* at iteration k be
X k = V k s , V k g ,
where V k s and V k g denote the vertex sets of the two search trees grown from the start and goal, respectively. Let X be the state space and X free X be the free space. At each iteration, PFMS-RRT* generates a sample q k from a mixture distribution
q k μ k = ( 1 p g , k ) μ k ell + p g , k μ global ,
where μ k ell is the uniform distribution over the oriented ellipse constructed from the current progress information, μ global is a global distribution covering the workspace (e.g., Halton low-discrepancy samples mapped to X), and p g , k ( 0 , 1 ] is the global-sampling probability that can be adaptively adjusted according to stagnation.
(Probabilistic completeness). Assume that for any measurable set A X free with Vol ( A ) > 0 , the global distribution satisfies μ global ( A ) > 0 , and there exists a constant p min > 0 such that p g , k p min for all k (i.e., global sampling is triggered infinitely often with a non-zero lower-bounded probability). Then, for any ε > 0 and any q free X free , with probability one there exist infinitely many iterations whose samples fall into the ε -neighborhood of q free , i.e.,
P k : q k q free < ε = 1 .
Consequently, together with bidirectional tree growth and collision checking/connection, if a feasible path exists, PFMS-RRT* is probabilistically complete in the sense that
lim k P X k contains a feasible start goal connection = 1 .
Note that this conclusion relies on the condition that the global sampling component continuously provides non-zero coverage of X free ; if p g , k is set to zero or the sampling support is restricted to a strict subset of X free for all sufficiently large k, completeness may be weakened.
(Asymptotic optimality). Let c k denote the best path cost found by PFMS-RRT* up to iteration k (once a connection exists), and let c * be the optimal cost. Since PFMS-RRT* retains the RRT* parent selection and rewiring operations, the best-so-far cost is non-increasing, i.e., c k + 1 c k when rewiring improves the solution. Moreover, under the same assumptions above (global sampling has full support over X free and is invoked infinitely often with p g , k p min > 0 ), and assuming that a standard goal-steering/steer attempt is executed with non-zero probability (so that the algorithm does not exclude the cost-improving events used in the classical RRT* convergence argument), PFMS-RRT* preserves asymptotic optimality in the sense that
P lim k c k = c * = 1 .

4. Experiment

In this section, the proposed PFMS-RRT* algorithm will be compared to GB-RRT* [22], F-RRT* [14], BI-RRT* [17], BI-APF-RRT* [24] and IBI-P-RRT* [25] algorithms. The simulation platform is MATLAB R2024b. All experiments were conducted on a desktop PC with CPU: 13th Gen Intel(R) Core(TM) i5-13600KF, RAM: 32 GB, and OS: Windows 10.

4.1. Experiment Configuration

Four representative 2D obstacle environments are used in the experiments, as shown in Figure 7a–d (Environment A–D).
Environment A (sparse mixed-obstacle scenario): the number of obstacles is relatively small, with rectangles and ellipses mixed in the workspace. The free space is largely open with a few local occlusions, and this scenario is mainly used to evaluate convergence speed and path quality under general obstacle distributions.
Environment B (structured corridor with dense small obstacles): The layout is highly structured, where long rectangular obstacles form boundary and corridor constraints along the top/bottom and sides, while many small elliptical obstacles are placed inside. This setting reflects dense obstacle interference and frequent collision checking, and is used to assess extension success rate and stability in cluttered environments.
Environment C (random dense mixed-obstacle scenario): Obstacles are more numerous and distributed more randomly, with rectangles and ellipses interleaved. The connectivity of the free space becomes more complex, often requiring detours through locally constrained regions, and this scenario is used to evaluate exploration capability and robust obstacle avoidance in unstructured complex environments.
Environment D (narrow-passage scenario): Multiple large rectangular obstacles form a typical narrow-passage structure, leaving only a slender traversable corridor with local bottlenecks. This is a classic challenging scenario for testing whether the sampling and extension strategies can pass through narrow passages and avoid stagnation.
To ensure a fair comparison, all environments share the same map scale and coordinate setting: the planar workspace is a 100 × 100 square with ( x , y ) [ 0 , 100 ] × [ 0 , 100 ] m . Obstacles are represented by gray occupied regions, and the remaining area is free space for feasible sampling and tree expansion. The start and goal configurations are fixed for all environments as x start = ( 0 , 0 ) , x goal = ( 100 , 100 ) . The initial step size is set to 3. All algorithms were executed independently for 30 runs per environment, with a maximum iteration limit of 3000 for each algorithm.

4.2. Parameter Settings

As summarized in Table 1, the key parameters of PFMS-RRT* were fixed across all experiments to ensure reproducibility and fair comparison. The step size η was set to balance exploration efficiency and local collision avoidance, while the stagnation threshold s th and the amplification upper bound γ max were used to detect persistent lack of progress and to enlarge the elliptical sampling corridor in a controlled manner. The global-sampling increment Δ p regulates how quickly the planner increases global exploration when stagnation occurs, preventing premature over-exploration while still enabling escape from difficult regions. The ellipse bounds ( a min , a max ; b min , b max ) constrain the locally guided sampling region to avoid over-concentration or excessive expansion, thus maintaining both efficiency and robustness. For the weights for progress reward, look-ahead feasibility reward, lateral-shift penalty w p , w L , and w δ were set to fixed values ( w p = 0.8 , w L = 0.8 , w δ = 0.05 ), and then kept unchanged in all reported experiments. These values were chosen based on term scaling and intended behavior: the forward progress term prog ( x try ) is typically on the order of the step size η , and the look-ahead feasibility reward is bounded by one step through L η with L 0 , 1 ; therefore, setting w p = w L = 0.8 assigns comparable importance to promoting progress and preferring candidates that admit a feasible one-step continuation, while the distance-to-goal term | g x try | 2 still provides global guidance. In contrast, a smaller lateral-shift penalty w δ = 0.05 discourages unnecessary detours without suppressing lateral motion, which is essential for escaping local dead-ends when direct steering is blocked. For the failure-driven switching mechanism, the decay rate β controls the sensitivity of the switching probability to consecutive failures, enabling an adaptive balance between goal-guided extension and APF-based extension. In the APF module, the gains ( k a , k r , k tan ) and influence radius ρ 0 were chosen to provide stable obstacle avoidance and smooth boundary-following behavior without introducing excessive detours.
As summarized in Table 2, for the compared baseline planners, we followed the parameter settings reported in their original papers to maintain consistency with prior work. In addition, common experimental settings such as the workspace bounds, termination conditions, and collision-checking procedures were kept the same for all methods, so that performance differences primarily reflect the algorithmic designs rather than parameter discrepancies.

4.3. Metric Definitions

In this paper, four metrics are reported to evaluate planning performance: Path length, Runtime, Number of nodes, and Avgdistance.

4.3.1. Path Length

Let the planned path be an ordered waypoint sequence P = { p i } i = 0 N . The path length is
L ( P ) = i = 0 N 1 p i + 1 p i 2 .

4.3.2. Runtime

Runtime is the wall-clock planning time measured from the start of the main planning loop until a feasible start–goal connection is found (or the termination budget is reached). Unless otherwise stated, it includes sampling, collision checking, and RRT* rewiring, and excludes visualization.

4.3.3. Number of Nodes

The number of nodes is the total number of vertices in the final tree(s) when the algorithm terminates. For bidirectional planners, it is reported as the sum of nodes in the start tree and the goal tree.

4.3.4. Avgdistance

Let the path P = { p i } i = 0 N form N line segments e i = [ p i , p i + 1 ] . We evaluate clearance by sampling the path on segments (not vertices only). For a segment e and obstacle set O, define
d min ( e ) = min o O d ( e , o ) ,
where d ( e , o ) = 0 if e intersects obstacle o, otherwise d ( e , o ) is the Euclidean distance between segment e and the obstacle boundary o (segment-to-circle for circular obstacles and segment-to-polygon-edge for rectangular obstacles). The average clearance is
Avgdistance = 1 N i = 0 N 1 d min ( e i ) .
All methods (PFMS-RRT* and baselines) are evaluated using the same definitions and distance computation for fair comparison.

4.4. Algorithm Performance Analysis

As shown in Figure 8, in Environment A, all six path-planning algorithms are able to successfully find a collision-free feasible path from the start to the goal. The red curve denotes the final planned path, while the blue and pink curves indicate the bidirectional search trees expanded from the start and the goal, respectively. However, the differences mainly lie in sampling efficiency and path quality. For GB-RRT* and F-RRT*, the search trees are more dispersed and exhibit more ineffective expansions in the vicinity of obstacles, manifested by repeated growth of the blue and pink branches near obstacles with insufficient utilization of the truly traversable regions. This behavior leads to unnecessary detours and polyline-like corrections in the resulting path. In contrast, PFMS-RRT* concentrates its tree expansion in effective feasible regions, producing more coherent obstacle-avoidance expansions with fewer rollbacks, and thus generating a more compact and smoother path. This phenomenon is directly related to the proposed progress-aware fusion sampling strategy. As a result, ineffective sampling around obstacle neighborhoods and redundant nodes are significantly reduced, improving both convergence efficiency and path quality.
The quantitative results in Table 3 further support the above observations. In Environment A, PFMS-RRT* achieves an average path length of 149.61   m , an average runtime of 0.09   s , an average node count of 53.60 , and an average minimum distance from path points to obstacles of 1.26   m . Compared with BI-APF-RRT*, BI-RRT*, and IBI-P-RRT*, PFMS-RRT* reduces the path length by 4.40 % , 4.35 % , and 4.07 % , respectively; compared with F-RRT* and GB-RRT*, the path length is reduced by 0.94 % and 1.25 % , respectively. In terms of efficiency, PFMS-RRT* decreases runtime by 68.97 % , 57.14 % , 43.75 % , 80.00 % , and 62.50 % relative to BI-APF-RRT*, BI-RRT*, F-RRT*, GB-RRT*, and IBI-P-RRT*, respectively; meanwhile, the average node count is reduced by 60.87 % , 51.52 % , 42.28 % , 78.87 % , and 55.32 % , respectively. Regarding safety, PFMS-RRT* improves the average minimum obstacle clearance by 6.78 % , 11.50 % , 28.57 % , 36.96 % , and 5.00 % over BI-APF-RRT*, BI-RRT*, F-RRT*, GB-RRT*, and IBI-P-RRT*, respectively. Overall, progress-aware fusion sampling significantly reduces runtime and tree size by eliminating ineffective samples and redundant expansions, while the more focused exploration also contributes to shorter, smoother paths with stronger clearance performance.
As shown in Figure 9, Environment B contains multiple narrow passages and corner-like structures, imposing higher requirements on reachability and extension stability. In this scenario, the baseline planners are more likely to exhibit path jitter at passage entrances and corners, which is reflected by frequent changes in expansion direction and unstable node distributions inside the corridor. As a consequence, the resulting paths often include additional polyline turns and detours before and after entering the narrow passages. In contrast, PFMS-RRT* can enter and traverse the narrow passages more stably, showing better convergence during search and significantly improved path continuity compared with the baselines.
These improvements are directly attributed to the proposed extension strategy. During expansion, PFMS-RRT* employs a three-level fallback goal-guided mechanism: when the standard goal-directed extension is blocked, the planner can still maintain forward progress via progressively strengthened detour behaviors, thereby reducing the probability of local failures and mitigating stagnation around corridor entrances. Moreover, when goal-guided probing becomes ineffective (or when the switching condition is triggered), the planner switches to the smooth tangential APF extension, which encourages boundary-following growth and smoother steering near obstacle boundaries. This makes the node expansion better aligned with corridor boundaries and yields more stable turning behaviors, ultimately improving passage-traversal stability and safety margin in Environment B.
The quantitative results in Table 4 further validate the above observations. PFMS-RRT* achieves an average path length of 150.90   m , an average runtime of 0.10   s , an average node count of 54.13 , and an average minimum obstacle clearance (Avgdistance) of 1.02   m . Compared with BI-APF-RRT*, BI-RRT*, and IBI-P-RRT*, PFMS-RRT* reduces the path length by 4.31 % , 5.16 % , and 4.78 % , respectively; compared with F-RRT* and GB-RRT*, the path length is reduced by 2.80 % and 2.25 % , respectively. In terms of efficiency, PFMS-RRT* decreases runtime by 81.48%, 56.52%, 41.18%, 75.00%, and 76.74% relative to BI-APF-RRT*, BI-RRT*, F-RRT*, GB-RRT*, and IBI-P-RRT*, respectively; meanwhile, the average node count is reduced by 67.26%, 52.85%, 42.25%, 75.32%, and 67.17%, respectively. Regarding safety, PFMS-RRT* improves Avgdistance by 3.03%, 13.33%, 52.24%, 45.71%, and 18.60% over the corresponding baselines. Overall, the strong reductions in runtime and node count indicate that progress-aware fused sampling effectively suppresses redundant exploration in dense obstacle fields, while the improved clearance reflects more stable obstacle-neighborhood expansions.
As shown in Figure 10, Environment C features highly fragmented free space, which can easily induce redundant sampling and detour-prone paths. The baseline planners generally suffer from increased ineffective exploration in this scenario. In particular, IBI-P-RRT* exhibits rollback behaviors, indicating that it is more prone to repeated probing in regions with poor local connectivity, which leads to less natural path shapes and degraded search efficiency. In contrast, PFMS-RRT* concentrates its sampling distribution more clearly on effective regions and produces more natural obstacle-avoidance trajectories, resulting in improvements in both overall search efficiency and path quality.
These results highlight the synergy between fused sampling and multi-level strategy extension. On the one hand, the progress-aware elliptical sampling suppresses repeated sampling in low-value regions by progressively focusing on informative areas. On the other hand, the three-level extension together with the smooth tangential APF provides more stable local guidance in fragmented spaces, thereby reducing the likelihood of rollbacks and unnecessary detours.
As reported in Table 5, PFMS-RRT* attains an average path length of 148.97   m , an average runtime of 0.10   s , an average node count of 52.53 , and an Avgdistance of 1.44   m . Compared with BI-APF-RRT*, BI-RRT*, and IBI-P-RRT*, PFMS-RRT* reduces the path length by 7.94%, 7.96%, and 6.75%, respectively; compared with F-RRT* and GB-RRT*, the path length is reduced by 1.99% and 2.81%, respectively. For efficiency, PFMS-RRT* decreases runtime by 68.75%, 58.33%, 47.37%, 87.80%, and 62.96% relative to BI-APF-RRT*, BI-RRT*, F-RRT*, GB-RRT*, and IBI-P-RRT*, respectively; the average node count is reduced by 63.60%, 56.74%, 52.75%, 86.86%, and 57.78%, respectively. In terms of safety, PFMS-RRT* improves Avgdistance by 2.86%, 9.09%, 17.07%, 18.03%, and 5.88% over the corresponding baselines. These gains indicate that the proposed sampling improvement drives substantial reductions in tree size and runtime in fragmented clutter, while the robust extension behaviors help maintain strong clearance and avoid inefficient stagnation.
As shown in Figure 11, Environment D is a typical strongly constrained narrow-passage scenario, where the feasible corridor is extremely limited. In this setting, most baseline planners exhibit unstable extensions or noticeable path oscillations inside the passage. Their search trees repeatedly grow within the narrow region but often fail to form a continuous and effective through-corridor connection, so the resulting paths tend to include multiple corrections or boundary-hugging segments. In contrast, PFMS-RRT* can expand stably along the passage and generates a smoother path that better conforms to the corridor structure, demonstrating stronger robustness and adaptability to constrained environments.
This advantage is closely related to the introduction of the smooth tangential APF extension. Under tight boundary constraints, the tangential potential field is more effective at guiding the expansion to follow obstacle boundaries smoothly, thereby reducing collision-prone probing and abrupt direction changes within the passage. Consequently, PFMS-RRT* improves both reachability and safety in Environment D.
As summarized in Table 6, PFMS-RRT* achieves an average path length of 177.39   m , an average runtime of 0.11   s , an average node count of 62.33 , and an Avgdistance of 0.86   m . Compared with BI-APF-RRT*, BI-RRT*, and IBI-P-RRT*, PFMS-RRT* reduces the path length by 14.82%, 10.60%, and 9.43%, respectively; compared with F-RRT* and GB-RRT*, the path length is reduced by 12.39% and 9.31%, respectively. In terms of efficiency, PFMS-RRT* decreases runtime by 92.95%, 90.35%, 75.00%, 80.36%, and 88.78% relative to BI-APF-RRT*, BI-RRT*, F-RRT*, GB-RRT*, and IBI-P-RRT*, respectively; the average node count is reduced by 72.09%, 66.07%, 71.75%, 78.17%, and 65.72%, respectively. Regarding safety, PFMS-RRT* improves Avgdistance by 3.61%, 6.17%, 13.16%, 17.81%, and 4.88% over BI-APF-RRT*, BI-RRT*, F-RRT*, GB-RRT*, and IBI-P-RRT*, respectively. Overall, the pronounced reductions in runtime and node count strongly suggest that the proposed sampling and multi-strategy extension effectively mitigate stagnation in narrow passages, while the tangential APF fallback helps maintain stable boundary-following behavior and competitive clearance under severe local constraints.
Figure 12 compares the core performance metrics of six RRT* variant algorithms across the four environments using bar charts with overlaid scatter points and 95% confidence-interval error bars. PFMS-RRT* achieves significantly shorter path lengths in all scenarios, with a particularly pronounced advantage in the challenging Environment D. F-RRT exhibits the largest variation in path length, and its mean value in Environment D is much higher than that of the other algorithms. PFMS-RRT* maintains relatively low runtime in most scenarios, with only a slight increase in Environment D. In contrast, GB-RRT* shows notably higher runtime in Environment B, and its scatter points are highly concentrated, indicating unstable efficiency in that scenario. GB-RRT* also produces consistently larger node counts across all environments, suggesting substantial redundancy in its sampling and exploration. By comparison, PFMS-RRT* maintains a low node count throughout, reflecting a more efficient sampling strategy. Moreover, the average distance between nodes and obstacles for PFMS-RRT* varies little across environments, indicating stronger planning stability, whereas F-RRT shows the most dispersed scatter distribution and the largest performance fluctuation. Overall, PFMS-RRT* delivers the best comprehensive performance over the four environments, producing shorter paths with fewer nodes, lower and more stable runtime, and improved robustness. GB-RRT* suffers from high exploration redundancy, and F-RRT exhibits large performance variability. The bidirectional variants perform between PFMS-RRT* and the single-tree methods, achieving a relatively good balance between efficiency and stability.

4.5. Ablation Study

PFMS-RRT* integrates multiple mechanisms, including progress-aware fused sampling, multi-level strategy extension, and a smooth tangential APF extension. To quantify the contribution of each component, ablation variants are evaluated by disabling one component at a time while keeping all other settings unchanged.

Ablation Variants

  • PFMS-RRT*: The complete method.
  • No-FS: Without fused sampling improvement.
  • No-ML: Without multi-level strategy extension.
  • No-APF: Without smooth tangential APF extension.
In this ablation study, Environments A and D are selected as representative cases. Environment A is a relatively simple scenario that reflects typical open-space navigation, while Environment D is the most challenging environment with denser obstacles and tighter passages. Evaluating both allows for the contribution of each module to be observed in both easy and difficult settings under the same experimental budget.
Table 7 and Table 8 report ablation results in the representative Environments A and D. Overall, disabling the multi-level extension (No-ML) causes the largest performance degradation across both environments, indicating that the multi-level strategy extension is the primary contributor to efficiency and robustness in cluttered scenes. In Environment A, No-ML increases the average node count from 53.60 to 145.20 (+170.9%) and runtime from 0.09 s to 0.23 s (+155.6%), showing substantially more ineffective expansions when the hierarchical obstacle-bypassing mechanism is removed. In the more challenging Environment D, No-ML similarly increases the node count from 62.33 to 126.23 (+102.5%) and runtime from 0.11 s to 0.24 s (+118.2%), and also leads to a larger path-length increase (177.39 m → 183.29 m, +3.33%) and reduced average clearance (0.86 m → 0.74 m, 14.0 %). These results confirm that the multi-level extension effectively improves obstacle bypassing and reduces repeated failed attempts in constrained regions.
Removing fused sampling (No-FS) and removing the APF (No-APF) lead to moderate but consistent degradations, suggesting that these components provide complementary benefits. In particular, No-FS noticeably reduces the average clearance Avgdistance (1.26 m → 1.07 m, 15.1 % in Environment A; 0.86 m → 0.81 m, 5.8 % in Environment D), indicating that progress-aware fused sampling improves the quality of sampled guidance and tends to yield safer paths with larger obstacle clearance. Importantly, disabling the APF (No-APF) also degrades both safety and efficiency: Avgdistance decreases from 1.26 m to 1.17 m ( 7.1 %) in Environment A and from 0.86 m to 0.82 m ( 4.7 %) in Environment D, accompanied by longer paths (e.g., 149.61 m → 150.52 m in Environment A and 177.39 m → 178.36 m in Environment D). This indicates that the smooth tangential APF provides a useful robustness-enhancing backup. Overall, the ablation study demonstrates that PFMS-RRT* benefits from complementary contributions of the three components.

4.6. Sensitivity Experiment

To evaluate the influence of the weight parameters in Equation (10), a sensitivity experiment was conducted for w p , w L , and w δ . These three parameters respectively control the reward of forward progress, the reward of one-step look-ahead feasibility, and the penalty on lateral offset during the secondary local probing stage. Their values jointly determine the balance among target-oriented expansion, local feasibility maintenance, and obstacle-avoidance flexibility.
The sensitivity results of w p are reported in Table 9. It can be observed that when w p varies from 0.4 to 0.8, the overall performance remains highly stable. The path length is maintained around 149.29–149.61 m, the runtime remains approximately 0.09 s, the number of nodes is almost unchanged at 53.60, and the average obstacle distance stays near 1.26–1.27 m. This indicates that within a moderate range, adjusting the forward-progress reward does not significantly alter the global planning behavior, suggesting that the proposed evaluation mechanism is relatively robust to small perturbations of w p .
However, when w p further increases to 1.0, 1.2, and 1.4, the path length gradually rises from 150.06 m to 153.21 m and finally 155.05 m, while the average distance to obstacles decreases from 1.24 m to 1.18 m and then 1.13 m. This trend indicates that an excessively large w p causes the local probing strategy to overemphasize direct forward movement, thereby weakening its ability to maintain a proper balance between aggressive advancement and safe detouring. In other words, an overly large forward-progress reward makes the candidate selection more radical, which may reduce local adaptability and eventually deteriorate path quality.
Overall, the results suggest that w p has a relatively wide stable interval, and values in the range of 0.4–0.8 can provide both short paths and stable safety margins. The setting w p = 0.8 achieves a good compromise between planning efficiency and local adaptability.
The results for w L are listed in Table 10. Compared with w p , the influence of w L shows a more distinct threshold behavior. When w L is set to 0.4 or 0.6, the path length remains relatively large at 153.20 m, and the average obstacle distance reaches 1.47 m, which is noticeably larger than that obtained with higher w L values. This suggests that when the look-ahead feasibility reward is insufficient, the probing process tends to generate more conservative local expansions, resulting in longer paths and larger clearance from obstacles.
Once w L increases to 0.8, the path length drops sharply to 149.61 m, while the average obstacle distance decreases to about 1.26 m. Further increasing w L to 1.0, 1.2, and 1.4 produces almost identical results, with the path length stabilized around 149.60 m and the node number unchanged. This implies that after the one-step look-ahead feasibility term reaches a sufficient weight, it can reliably guide the local probing toward candidates with better continuation potential, thereby significantly improving path compactness. At the same time, the absence of noticeable performance degradation at larger w L values indicates that the proposed strategy is not overly sensitive to this parameter once it exceeds a certain effective threshold.
Therefore, the experimental results demonstrate that w L mainly acts as an enabling factor for local feasibility awareness. A value that is too small weakens the look-ahead guidance effect, whereas values from 0.8 upward yield stable and favorable performance. Considering both effectiveness and consistency, w L = 0.8 is a reasonable and economical choice.
The sensitivity analysis of w δ is summarized in Table 11. Unlike w p and w L , the role of w δ is to penalize lateral deviation during candidate evaluation. The results show that the influence of w δ is comparatively mild within a small range. Specifically, when w δ increases from 0.01 to 0.05, the path length only changes slightly from 149.59 m to 149.61 m, and the average obstacle distance remains nearly constant at 1.29 m. This indicates that small variations of the lateral-offset penalty do not substantially affect the local decision process.
When w δ is further increased to 0.10 and 0.15, the path length rises to 149.98 m, and the average distance to obstacles drops to 1.26 m. At w δ = 0.20 , the path length further increases to 151.04 m, while the average obstacle distance decreases to 1.21 m. These results suggest that an excessively large lateral penalty suppresses necessary sideways adjustment during local obstacle avoidance, making the probing process less flexible and slightly reducing path quality and safety margin.
In general, w δ behaves as a weak regularization term rather than a dominant decision factor. The algorithm is relatively insensitive to small perturbations of this parameter, but too large a value may gradually limit local detouring capability. Therefore, a small-value interval such as 0.01–0.05 is sufficient, and w δ = 0.05 provides a suitable balance between suppressing redundant lateral oscillations and preserving necessary obstacle-avoidance flexibility.
Combining the above observations, the three parameters exhibit different sensitivity patterns. The parameter w p mainly affects the aggressiveness of forward expansion, and overly large values may lead to longer paths and reduced safety margins. The parameter w L plays a more decisive role in enabling one-step feasibility awareness, and its performance improves significantly once it reaches a sufficient level. In contrast, w δ acts as a mild regularizer, and its influence is relatively limited unless it becomes too large.
More importantly, the results in Table 9, Table 10 and Table 11 show that the proposed method maintains stable runtime and node number under different parameter settings, indicating that the overall framework is robust and does not rely on delicate parameter tuning. Based on the sensitivity experiment, the final parameter configuration is set to
w p = 0.8 , w L = 0.8 , w δ = 0.05 ,
which achieves a favorable trade-off among path quality, runtime, node scale, and obstacle clearance.

5. Conclusions

This paper proposes PFMS-RRT*, a progress-aware fused-sampling RRT* with multi-level strategy extension, to address the high randomness, low efficiency, long search time, and poor path quality of RRT* in complex obstacle environments. The method introduces three key components: (1) progress-aware fused sampling that adaptively reshapes the sampling distribution according to planning progress to reduce redundant samples and generate more informative ones; (2) a three-level goal-guided extension with progressively stronger fallback behaviors to improve extension success rate and stability in clutter; and (3) a smooth tangential APF-based fallback extension (with failure-driven probabilistic switching) to alleviate stagnation near obstacles and enhance robustness under local constraints.
Extensive simulations in four representative 2D environments, including structured corridors, dense random clutter, and narrow passages, demonstrate that PFMS-RRT* consistently delivers superior overall performance compared with several RRT* variants. In particular, PFMS-RRT* achieves shorter paths, lower and more stable runtime, and substantially fewer nodes, while maintaining competitive or improved obstacle clearance (Avgdistance), indicating stronger planning stability across scenarios.
Future work includes incorporating robot footprint and kinodynamic constraints (e.g., minimum turning radius and velocity/acceleration limits), extending the framework to higher-dimensional planning (the sampling region generalizes from an ellipse to an ellipsoid), and validating the method in more realistic settings with uncertainty and dynamic obstacles.

Author Contributions

Conceptualization, Z.L.; methodology, Z.L.; validation, J.L.; formal analysis, Z.L.; investigation, C.L.; data curation, J.L.; writing—original draft preparation, Z.L. and J.L.; writing—review and editing, Z.L. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APFArtificial Potential Field
APF-RRT*Artificial Potential Field based RRT*
Bi-RRT*Bidirectional RRT*
GAGenetic Algorithm
Informed RRT*Informed Rapidly-exploring Random Tree Star
NIRRT*Neural Informed RRT*
NRRT*Neural RRT*
PFMS-RRT*Progress-aware Fused Sampling RRT* with Multi-level Strategy extension
PSOParticle Swarm Optimization
RRTRapidly-exploring Random Tree
RRT*Rapidly-exploring Random Tree Star
SOF-RRT*Spatial Offset Sampling RRT*

References

  1. Kotsinis, D.; Karras, G.C.; Bechlioulis, C.P. Autonomous Exploration in Unknown Indoor 2D Environments Using Harmonic Fields and Monte Carlo Integration. Sensors 2025, 25, 4894. [Google Scholar] [CrossRef] [PubMed]
  2. Lin, Z.; Wu, K.; Shen, R.; Yu, X.; Huang, S. An Efficient and Accurate A-Star Algorithm for Autonomous Vehicle Path Planning. IEEE Trans. Veh. Technol. 2024, 73, 9003–9008. [Google Scholar] [CrossRef]
  3. Cho, Y.; Choi, D.; Ko, K.; Park, J.H. Position Selection and Collision-Free Path Planning for Fruit Picking Robots. Appl. Sci. 2025, 15, 4419. [Google Scholar] [CrossRef]
  4. Gong, J.; Chen, J.; Cai, D.; Wei, W.; Long, Y. Disturbance Observer-Based Passivity and Impedance Control for Trajectory Tracking in Autonomous Hydraulic Excavators. Autom. Constr. 2025, 170, 105898. [Google Scholar] [CrossRef]
  5. Cobano, J.A.; Merino, L.; Caballero, F. Exploiting Euclidean Distance Field Properties for Fast and Safe 3D Planning with a Modified Lazy Theta*. Robot. Auton. Syst. 2026, 198, 105317. [Google Scholar] [CrossRef]
  6. Yang, J.; He, K.; Wu, J.; Ruan, H.; Cai, S.; Zhang, Y. Bidirectional Region Expansion Guided Sampling-Based RRT*: Achieving High Stability and Ultra-High Planning Efficiency. Expert Syst. Appl. 2026, 310, 131326. [Google Scholar] [CrossRef]
  7. Ge, J.; Wang, T.; Hu, K.; Wang, J.; Wu, J.; Wang, J.; Wu, J. Optimization of Power Grid Material Warehousing and Supply Chain Distribution Path Planning Based on Improved PSO Algorithm. Sci. Rep. 2025, 15, 45132. [Google Scholar] [CrossRef]
  8. Hong, L.; Chu, Y.; Wang, B.; Liang, Y.; Huo, P. Research on the Shortest Path Planning for Non-Standard Workpiece Arc Welding Based on Genetic Algorithm. Int. J. Adv. Manuf. Technol. 2026, 142, 4453–4466. [Google Scholar] [CrossRef]
  9. Liu, G.; Gao, S.; Jiang, J.; Luo, Z.; Jiang, G. Mountain UGV Path Planning via Optimized Dueling Double DQN (D3QN): Structural Optimization, Path-Guided Rewards, and Phased Action Policy. Control Eng. Pract. 2026, 169, 106792. [Google Scholar] [CrossRef]
  10. Wang, Y.; Jiang, W.; Luo, Z.; Yang, L.; Wang, Y. Path Planning of a 6-DOF Measuring Robot with a Direction Guidance RRT Method. Expert Syst. Appl. 2024, 238, 122057. [Google Scholar] [CrossRef]
  11. Wang, X.; Ma, X.; Li, X.; Ma, X.; Li, C. Target-Biased Informed Trees: Sampling-Based Method for Optimal Motion Planning in Complex Environments. J. Comput. Des. Eng. 2022, 9, 755–771. [Google Scholar] [CrossRef]
  12. Gammell, J.D.; Srinivasa, S.S.; Barfoot, T.D. Informed RRT*: Optimal Sampling-Based Path Planning Focused via Direct Sampling of an Admissible Ellipsoidal Heuristic. In 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems; IEEE: Piscataway, NJ, USA, 2014; pp. 2997–3004. [Google Scholar] [CrossRef]
  13. Yu, S.; Chen, J.; Liu, G.; Tong, X.; Sun, Y. SOF-RRT*: An Improved Path Planning Algorithm Using Spatial Offset Sampling. Eng. Appl. Artif. Intell. 2023, 126, 106875. [Google Scholar] [CrossRef]
  14. Liao, B.; Wan, F.; Hua, Y.; Ma, R.; Zhu, S.; Qing, X. F-RRT*: An Improved Path Planning Algorithm with Improved Initial Solution and Convergence Rate. Expert Syst. Appl. 2021, 184, 115457. [Google Scholar] [CrossRef]
  15. Wang, J.; Chi, W.; Li, C.; Wang, C.; Meng, M.Q.-H. Neural RRT*: Learning-Based Optimal Path Planning. IEEE Trans. Autom. Sci. Eng. 2020, 17, 1748–1758. [Google Scholar] [CrossRef]
  16. Huang, Z.; Chen, H.; Pohovey, J.; Driggs-Campbell, K. Neural Informed RRT*: Learning-Based Path Planning with Point Cloud State Representations under Admissible Ellipsoidal Constraints. In 2024 IEEE International Conference on Robotics and Automation (ICRA); IEEE: Piscataway, NJ, USA, 2024; pp. 8742–8748. [Google Scholar] [CrossRef]
  17. Wang, B.; Ju, D.; Xu, F.; Feng, C. Bi-RRT*: An Improved Bidirectional RRT* Path Planner for Robot in Two-Dimensional Space. IEEJ Trans. Electr. Electron. Eng. 2023, 18, 1639–1652. [Google Scholar] [CrossRef]
  18. Ji, H.; Xie, H.; Wang, C. E-RRT*: Path Planning for Hyper-Redundant Manipulators. IEEE Robot. Autom. Lett. 2023, 8, 8128–8135. [Google Scholar] [CrossRef]
  19. Wang, J.; Li, J.; Song, Y.; Tuo, Y.; Liu, C. FC-RRT*: A Modified RRT* with Rapid Convergence in Complex Environments. J. Comput. Sci. 2024, 77, 102239. [Google Scholar] [CrossRef]
  20. Qi, J.; Yuan, Q.; Wang, C.; Du, X.; Du, F.; Ren, A. Path planning and col-lision avoidance based on the RRT*FN framework for a ro-botic manipulator in various scenarios. Complex Intell. Syst 2023, 9, 7475–7494. [Google Scholar] [CrossRef]
  21. Liu, N.; Hu, zihang; Wei, M.; Guo, P.; Zhang, S.; Zhang, A. Improved A* Algorithm Incorporating RRT* Thought: A Path Planning Algorithm for AGV in Digitalised Workshops. Comput. Oper. Res. 2025, 177, 106993. [Google Scholar] [CrossRef]
  22. Chao, N.; Liu, Y.; Xia, H.; Ayodeji, A.; Bai, L. Grid-Based RRT* for Minimum Dose Walking Path-Planning in Complex Radioactive Environments. Ann. Nucl. Energy 2018, 115, 73–82. [Google Scholar] [CrossRef]
  23. Tao, F.; Ding, Z.; Wang, Z.; Ji, B. VPS-APF-RRT*: A Variable Probability Sampling RRT* Intelligent Algorithm Combined with APF for Autonomous Vehicles Path Planning. Expert Syst. Appl. 2026, 299, 129965. [Google Scholar] [CrossRef]
  24. Huang, Y.; Li, H.; Dai, Y.; Lu, G.; Duan, M. A 3D Path Planning Algorithm for UAVs Based on an Improved Artificial Potential Field and Bidirectional RRT*. Drones 2024, 8, 760. [Google Scholar] [CrossRef]
  25. Yuan, L.; Zhao, J.; Li, W.; Hou, J. Improved Informed-RRT* Based Path Planning and Trajectory Optimization for Mobile Robots. Int. J. Precis. Eng. Manuf. 2023, 24, 435–446. [Google Scholar] [CrossRef]
Figure 2. Schematic diagram of elliptical sampling region variation.
Figure 2. Schematic diagram of elliptical sampling region variation.
Applsci 16 03107 g002
Figure 3. Schematic diagram of central jumping.
Figure 3. Schematic diagram of central jumping.
Applsci 16 03107 g003
Figure 4. Direct goal steering.
Figure 4. Direct goal steering.
Applsci 16 03107 g004
Figure 5. Local small-range probing.
Figure 5. Local small-range probing.
Applsci 16 03107 g005
Figure 6. Multi-direction detour fallback.
Figure 6. Multi-direction detour fallback.
Applsci 16 03107 g006
Figure 7. Experimental environments.
Figure 7. Experimental environments.
Applsci 16 03107 g007
Figure 8. Examples of path planning results in environment A.
Figure 8. Examples of path planning results in environment A.
Applsci 16 03107 g008
Figure 9. Examples of path planning results in environment B.
Figure 9. Examples of path planning results in environment B.
Applsci 16 03107 g009
Figure 10. Examples of path planning results in environment C.
Figure 10. Examples of path planning results in environment C.
Applsci 16 03107 g010
Figure 11. Examples of path planning results in environment D.
Figure 11. Examples of path planning results in environment D.
Applsci 16 03107 g011
Figure 12. Comparison diagrams of core performance metrics for comparative algorithms across four environments.
Figure 12. Comparison diagrams of core performance metrics for comparative algorithms across four environments.
Applsci 16 03107 g012
Table 1. Key parameter settings used in the experiments.
Table 1. Key parameter settings used in the experiments.
Parameter (Symbol)Value
Step size ( η )3
Stagnation threshold ( s th )4
Amplification upper bound ( γ max )3.0
Global-sampling increment ( Δ p )0.15
Ellipse bounds ( a min , a max ; b min , b max ) ( 12 ,   90 ;   8 ,   90 )
Progress reward weight ( w p )0.8
Look-ahead feasibility reward weight ( w L )0.8
Lateral-shift penalty weight ( w δ )0.05
Switching decay rate ( β )0.8
Repulsive force gain ( k r )1.5
Tangential force gain ( k tan )1.2
Attractive force gain ( k a )0.5
Influence radius ( ρ 0 )10
Table 2. Method-specific baseline parameters used in the experiments.
Table 2. Method-specific baseline parameters used in the experiments.
BaselineMethod-Specific ParametersValue
GB-RRT*Goal bias probability ( p goal )0.2
F-RRT*Goal bias in random branch ( P goal )0.2
Bisection tolerance ( ε bisect )0.25
Max bisection steps ( N bisect )30
Constraint radius factor ( α )1.0
Max constraint tries ( N tries )200
BI-RRT*None
BI-APF-RRT*Attractive force gain ( k a )0.5
Repulsive force gain ( k r )30
Influence radius ( ρ 0 )10
Random attraction gain ( k rand )10
IBI-P-RRT*Attractive force gain ( k a )0.2
Repulsive force gain ( k r )0.5
Influence radius ( ρ 0 )5
Repulsive exponent (n)2
Random attraction gain ( k rand )15
Table 3. Performance comparison of different algorithms in Environment A. Values are reported as mean (standard deviation).
Table 3. Performance comparison of different algorithms in Environment A. Values are reported as mean (standard deviation).
AlgorithmPath Length (m)Runtime (s)Number of NodesAvgdistance (m)
PFMS-RRT*149.61 (0.69)0.09 (0.01)53.60 (1.22)1.26 (0.04)
BI-APF-RRT*156.49 (5.19)0.29 (0.15)136.97 (45.93)1.18 (0.27)
BI-RRT*156.41 (4.49)0.21 (0.06)110.57 (21.74)1.13 (0.35)
F-RRT*151.03 (3.00)0.16 (0.03)92.87 (15.12)0.98 (0.27)
GB-RRT*151.50 (3.38)0.45 (0.19)253.70 (86.36)0.92 (0.33)
IBI-P-RRT*155.95 (5.18)0.24 (0.10)119.97 (32.31)1.20 (0.29)
Table 4. Performance comparison of different algorithms in Environment B. Values are reported as mean (standard deviation).
Table 4. Performance comparison of different algorithms in Environment B. Values are reported as mean (standard deviation).
AlgorithmPath Length (m)Runtime (s)Number of NodesAvgdistance (m)
PFMS-RRT*150.90 (0.93)0.10 (0.01)54.13 (1.11)1.02 (0.09)
BI-APF-RRT*157.70 (6.41)0.54 (1.09)165.33 (109.11)0.99 (0.28)
BI-RRT*159.11 (7.61)0.23 (0.08)114.80 (23.63)0.90 (0.18)
F-RRT*155.25 (5.89)0.17 (0.04)93.73 (19.24)0.67 (0.20)
GB-RRT*154.37 (2.93)0.40 (0.26)219.37 (112.16)0.70 (0.27)
IBI-P-RRT*158.48 (3.65)0.43 (0.40)164.87 (93.95)0.86 (0.26)
Table 5. Performance comparison of different algorithms in Environment C. Values are reported as mean (standard deviation).
Table 5. Performance comparison of different algorithms in Environment C. Values are reported as mean (standard deviation).
AlgorithmPath Length (m)Runtime (s)Number of NodesAvgdistance (m)
PFMS-RRT*148.97 (0.89)0.10 (0.01)52.53 (1.11)1.44 (0.03)
BI-APF-RRT*161.82 (8.41)0.32 (0.14)144.33 (37.78)1.40 (0.27)
BI-RRT*161.86 (9.58)0.24 (0.08)121.43 (28.91)1.32 (0.25)
F-RRT*152.00 (3.71)0.19 (0.06)111.17 (32.17)1.23 (0.19)
GB-RRT*153.28 (3.96)0.82 (0.44)399.70 (163.71)1.22 (0.24)
IBI-P-RRT*159.76 (8.01)0.27 (0.10)124.43 (29.36)1.36 (0.23)
Table 6. Performance comparison of different algorithms in Environment D. Values are reported as mean (standard deviation).
Table 6. Performance comparison of different algorithms in Environment D. Values are reported as mean (standard deviation).
AlgorithmPath Length (m)Runtime (s)Number of NodesAvgdistance (m)
PFMS-RRT*177.39 (0.64)0.11 (0.01)62.33 (1.60)0.86 (0.07)
BI-APF-RRT*208.25 (34.67)1.56 (1.42)223.30 (76.35)0.83 (0.18)
BI-RRT*198.42 (28.78)1.14 (1.50)183.70 (79.55)0.81 (0.14)
F-RRT*202.48 (31.90)0.44 (0.17)220.67 (72.49)0.76 (0.20)
GB-RRT*195.61 (25.24)0.56 (0.14)285.57 (52.72)0.73 (0.13)
IBI-P-RRT*195.87 (26.94)0.98 (0.74)181.83 (54.77)0.82 (0.14)
Table 7. Ablation results in Environment A.
Table 7. Ablation results in Environment A.
VariantAvg Path Length (m)Avg Node CountAvg Time (s)Avg Dist (m)
PFMS-RRT*149.6153.600.091.26
No-FS150.9556.400.111.07
No-ML151.29145.200.231.13
No-APF150.5256.100.111.17
Table 8. Ablation results in Environment D.
Table 8. Ablation results in Environment D.
VariantAvg Path Length (m)Avg Node CountAvg Time (s)Avg Dist (m)
PFMS-RRT*177.3962.330.110.86
No-FS178.5666.230.130.81
No-ML183.29126.230.240.74
No-APF178.3665.130.130.82
Table 9. Sensitivity analysis results of w p .
Table 9. Sensitivity analysis results of w p .
w p Path Length (m)Runtime (s)Number of NodesAvgdistance (m)
0.40149.290.0953.601.27
0.60149.410.0953.601.27
0.80149.610.0953.601.26
1.00150.060.0953.601.24
1.20153.210.0953.561.18
1.40155.050.0953.601.13
Table 10. Sensitivity analysis results of w L .
Table 10. Sensitivity analysis results of w L .
w L Path Length (m)Runtime (s)Number of NodesAvgdistance (m)
0.40153.200.0953.601.47
0.60153.200.0953.601.47
0.80149.610.0953.601.26
1.00149.600.0953.601.26
1.20149.600.0953.601.26
1.40149.600.0953.601.26
Table 11. Sensitivity analysis results of w δ .
Table 11. Sensitivity analysis results of w δ .
w δ Path Length (m)Runtime (s)Number of NodesAvgdistance (m)
0.01149.590.0953.601.29
0.05149.610.0953.601.29
0.10149.980.0953.601.26
0.15150.280.0953.601.26
0.20151.040.0953.601.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Li, J.; Luo, C. PFMS-RRT*: A Progress-Aware Fused-Sampling RRT* with Multi-Level Strategy Extension for Path Planning. Appl. Sci. 2026, 16, 3107. https://doi.org/10.3390/app16063107

AMA Style

Li Z, Li J, Luo C. PFMS-RRT*: A Progress-Aware Fused-Sampling RRT* with Multi-Level Strategy Extension for Path Planning. Applied Sciences. 2026; 16(6):3107. https://doi.org/10.3390/app16063107

Chicago/Turabian Style

Li, Zhongwei, Jiaming Li, and Cai Luo. 2026. "PFMS-RRT*: A Progress-Aware Fused-Sampling RRT* with Multi-Level Strategy Extension for Path Planning" Applied Sciences 16, no. 6: 3107. https://doi.org/10.3390/app16063107

APA Style

Li, Z., Li, J., & Luo, C. (2026). PFMS-RRT*: A Progress-Aware Fused-Sampling RRT* with Multi-Level Strategy Extension for Path Planning. Applied Sciences, 16(6), 3107. https://doi.org/10.3390/app16063107

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop