Next Article in Journal
Ground Risk Buffer Estimation for Unmanned Aerial Vehicle Test Flights Based on Dynamics Analysis
Previous Article in Journal
Electric Propulsion and Hybrid Energy Systems for Solar-Powered UAVs: Recent Advances and Challenges
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Hierarchical Framework for Adaptive UAV Trajectory Planning in Dynamic Target Tracking

School of Aeronautics and Astronautics, Sun Yat-sen University, Shenzhen 518107, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(12), 847; https://doi.org/10.3390/drones9120847
Submission received: 27 October 2025 / Revised: 26 November 2025 / Accepted: 6 December 2025 / Published: 10 December 2025
(This article belongs to the Section Artificial Intelligence in Drones (AID))

Highlights

What are the main findings?
  • A novel kinodynamic planning framework is proposed, which efficiently integrates ESDF-based obstacle awareness directly into the search process for tracking a dynamic target.
  • The introduced environment-adaptive risk-weight mechanism dynamically adjusts the search step size, enhancing planning efficiency and range without additional computational cost.
What are the implications of the main findings?
  • The proposed intent-free motion prediction method, relying solely on spatiotemporal dynamics, offers a lightweight and versatile solution for tracking targets with unknown intentions.
  • The method’s validation on a fully autonomous quadcopter demonstrates its practical applicability for real-world dynamic scenarios requiring robust motion planning and obstacle avoidance.

Abstract

Planning trajectories for unmanned aerial vehicles (UAVs) actively tracking a moving target poses a significant challenge. This paper introduces a motion planning system utilizing UAVs to track a dynamic target actively. The system can handle complex tracking tasks, ensuring safety and dynamic feasibility. Initially, a front-end path search involves employing adaptive Kinodynamic path searching to identify the route. A heuristic approach is applied to determine an initial trajectory, considering the minimum time and control cost. Subsequently, we introduce the cost of combining optimization and use Euclidean Signed Distance Field (ESDF) gradient information to improve trajectory smoothness and dynamic feasibility. Simulation results confirm the consistent superiority of our proposed method over the Fast-Tracker algorithm, resulting in shorter tracking distance and flight time. Our method achieves smoother trajectories while preserving dynamic feasibility. A real-world single-UAV experiment further validates the effectiveness of our approach.

1. Introduction

Unmanned aerial vehicles (UAVs), known for their ease of deployment and versatility, have become indispensable tools in various military and civilian applications, including search and rescue missions, reconnaissance, and surveillance [1,2,3]. In drone technology, a crucial application involves monitoring complex environments and dynamic targets often characterized by uncertain intentions. Active tracking of moving targets presents a significant challenge in this domain. This challenge is further intensified when tracking occurs in cluttered and a priori unknown environments. Here, the UAV must simultaneously map its surroundings, predict the target’s motion, and plan a safe, kinodynamically feasible trajectory, all under severe computational constraints. The failure of any component in this chain can lead to mission failure, highlighting the need for a robust and integrated approach.
Tracking dynamic targets presents a complex challenge, constrained by limited onboard computation, UAV dynamics, and stringent safety requirements for obstacle avoidance [4,5,6]. This challenge is traditionally addressed within the framework of maneuvering target tracking. The prevailing approach in this field relies on multiple-model methods, which simultaneously employ a set of models (and corresponding filters) to cover the possible dynamic behaviors of a target, as comprehensively surveyed by Li and Jilkov [7].
Such filtering-based approaches form the core of many practical Electro-Optical Systems (EOSs) for UAVs, as demonstrated by Kim et al. [8], who implemented a Kalman filter for image tracking and real-time 3D target localization on a small UAV. Their work exemplifies the successful application of classic estimation theory and also highlights the inherent system complexity and computational burden associated with achieving high-quality imaging, tracking, and measurement under SWaP constraints. To circumvent the dependency on explicit dynamic models, an alternative paradigm formulates the problem as a reactive control task.
Vision-based systems, for instance, achieve autonomous tracking by using visual trackers like LCT (Long-Term Correlation Tracking) with pixel-level feedback [9,10,11], effectively bypassing the need for explicit state estimation via filtering. For instance, Kumar et al. [12] demonstrated the detection, localization, and tracking of spherical targets using only onboard sensors. The primary focus of these approaches is the visual tracking scheme and the design of UAV control laws. By leveraging specific tracking strategies, they can achieve real-time performance. A fundamental limitation of these methods, however, is their general lack of environmental perception, which prevents them from incorporating critical safety and dynamic feasibility constraints. Regarding target prediction methodologies, some approaches employ Kalman filters integrated with coarse-grained motion models to forecast future trajectories [13,14,15]; however, their reliability is often limited by model inaccuracies. To address this limitation, hybrid estimation methods, such as the Interacting Multiple Model (IMM) algorithm, have been developed to better handle target maneuverability and nonlinearities [16,17]. Alternatively, intent-free planning methods have been proposed that predict target motion through polynomial or Bézier curve regression [18,19]. A significant drawback of such methods is their susceptibility to the Runge phenomenon and catastrophic failure during boundary extrapolation.
To address these limitations, researchers use the hierarchical framework method to address these challenges [20,21]. Hierarchical motion planning involves an advanced geometric path planner for initial path generation and a low-level time parameterization scheme for trajectory optimization. The advanced geometric path planner identifies an obstacle-free path to satisfy primary geometric constraints. Subsequently, the low-level time parameterization method accounts for drone dynamics and generates a time-parameterized trajectory. Thus, the advanced geometric path planner, often called front-end path planning, complements the low-level time parameterization, known as back-end trajectory optimization [22,23]. For rotor-based unmanned aerial vehicles with non-trivial dynamics, generating trajectories directly in high-dimensional state space is time-consuming, whereas developing geometric paths and applying smoothing calculations is more efficient.
Although effective in meeting real-time demands, this decoupling often sacrifices global optimality. The separation of geometric and dynamic constraints can result in trajectories trapped in local minima, especially during local replanning with a non-zero initial state. This frequently manifests as suboptimal smoothness and inefficient maneuvering when facing sudden obstacles or evasive targets. For instance, when facing sudden obstacles or target maneuvers during the tracking process, the planned trajectory is constrained within a topologically equivalent class [18,24,25]. This constraint can result in entrapment within local minima, leading to suboptimal flight smoothness and safety. Chen et al. [18] address the challenging problem of tracking a moving target in cluttered environments using a quadrotor. Their initial proposal involves an online trajectory planning method that generates smooth, dynamically feasible, and collision-free polynomial trajectories that track a visually tracked moving target. Ding et al. [26] propose a real-time B-spline based kinodynamic (RBK) search algorithm, transforming a position-only shortest path search into an efficient kinodynamic search by leveraging B-spline parameterization properties. Meanwhile, Dmitri et al. [27] describe a practical path-planning algorithm for autonomous vehicles operating in unknown environments with online obstacle detection. However, both [25,26] suffer from significant time complexity due to the need for fine-grained search. This may lead to low computational efficiency, which is not conducive to rapid route planning for complex environments and drones.
Beyond these, a distinct line of research employs hierarchical learning and control frameworks, such as those using Safe Reinforcement Learning (SRL) from human demonstration [28] and Hierarchical SRL with prescribed performance bounds [29]. These methods excel in learning complex skills from demonstration and providing strong theoretical safety and stability guarantees, often via Lyapunov analysis. However, their reliance on extensive pre-training data, complex neural network computations, and precise environment models presents a fundamental challenge for resource-constrained UAVs tasked with real-time, onboard dynamic target tracking. The computational overhead and data dependency of these SRL-based approaches stand in contrast to the need for a lightweight, self-contained algorithm that requires no prior demonstrations or complex online policy learning.
Therefore, a comprehensive solution is required that not only maintains the efficiency of hierarchical planning but also bridges the gap between geometry and dynamics to achieve higher performance. Such a solution should deliver real-time computational efficiency, robustness to prediction uncertainties, strict dynamic feasibility, and tight perception-planning integration.
To this end, we propose a solution framework to adaptively navigate and plan paths for unmanned aerial vehicles tracking dynamic targets in unknown obstacle scenarios. The contributions of this article are summarized as follows, with comparisons to state-of-the-art methods such as Fast Tracker [19]:
(1)
A lightweight, intent-free method for motion prediction driven purely by spatiotemporal dynamics, departing from any form of fitting-based prediction (e.g., trajectory fitting in Fast Tracker) to avoid extrapolation inaccuracies.
(2)
An environment-adaptive risk-weight mechanism for direct step-size determination in ESDF fine-grained search, enabling more efficient and reactive planning compared to fixed-step approaches.
(3)
The proposed approach is validated through extensive simulations and subsequently implemented on a fully autonomous quadcopter system. Real-world experiments are conducted to evaluate its performance in dynamic scenarios requiring motion planning and obstacle avoidance.
Overall, trajectory planning efficiency is enhanced through a kinodynamic framework that integrates ESDF calculation directly into the search process. This method accounts for dynamic constraints and incorporates an adaptive risk-weight strategy, which reduces time complexity without extra computational cost, thereby extending the planner’s range and precision.

2. Framework Overview

In UAV perception, both LiDAR and cameras have proven effective for environmental sensing. However, LiDAR’s significant weight and power consumption present substantial drawbacks for small UAVs, where payload and energy are critically constrained. For instance, typical high-resolution LiDARs used in robotics (e.g., the Livox Mid-360 laser scanner from Livox Technology Co., Ltd., Shenzhen, China) can weigh over 250 g and consume 6 W or more [30], whereas a lightweight camera system (e.g., the Intel RealSense D435i from Intel Corporation, Santa Clara, CA, USA) often weighs less than 100 g and consumes under 3 W [22]. This stark contrast in Size, Weight, and Power (SWaP) underscores the distinct advantages of cameras, making them particularly suitable for vision-based UAV applications. Consequently, our work employs a depth camera as the primary sensor to address the challenging problem of autonomous mobile target tracking. This task finds critical applications in security surveillance, filmmaking, and wildlife monitoring [13,20], where the depth camera provides essential geometric information for robust 3D tracking and obstacle avoidance.
As illustrated in Figure 1, our system is specifically designed for tracking a dynamic target in complex, unknown environments. The core problem we address is formally defined as: given the UAV’s current state x 0 ( t ) and the estimated state of the target x target ( t ) , compute a trajectory ς ( t ) for the UAV that minimizes a cost function J ( ς ) while satisfying (i) the UAV’s kinematic and dynamic constraints and collision-free conditions ς ( t ) 𝒪 = , where the obstacle set 𝒪 is constructed from the depth camera data; (ii) real-time replanning requirements to adapt to the target’s unpredictable motion. This formulation of the tracking problem, which fuses target estimates with depth-based obstacle mapping, aligns with and extends previous research on drone-based pursuit [5,20] and specifically builds upon studies published in the Drones journal concerning moving target tracking [10,11].
Figure 1 illustrates the software architecture of our vision-based quadrotor automatic navigation system. The suggested trajectory planning framework (highlighted in yellow) is a planning module that encompasses the Adaptive Kinodynamic path search and B-spline-based optimization method (AKBS). The hardware components (highlighted in orange) comprise the sensing hardware and the UAV with its actuator. The sensing hardware includes a depth camera and an IMU. The Euclidean Signed Distance Field (ESDF) [31] and the state estimation module (VINS Fusion) [2] offer low-frequency local maps and high-frequency pose estimation. Following the map update, the planning module will initiate the replanning strategy and a fixed-frequency trajectory prediction.
Our trajectory planning framework utilizes an adaptive kinodynamic path search to generate initial, collision-free trajectories efficiently. This approach is detailed in Section 4. Subsequently, a computationally efficient optimization step refines these trajectories by adjusting their control points to enhance smoothness and guarantee dynamic feasibility, as elaborated in Section 5. A geometric controller tracks the desired trajectories generated by the real-time replanning module during flight [32]. The resulting attitude and throttle commands are then transmitted to the PX4 flight controller to guide the aerial vehicle toward the dynamic target. The experimental setup and results are detailed in Section 6.

3. Intent-Free Target Trajectory Prediction

Trajectory prediction is essential for aerial drones to track moving targets, as conceptually illustrated in Figure 2. The figure contrasts two approaches: methods without prediction (red trajectories) can only react to the target’s current estimated position, resulting in inefficient, oscillatory paths that lag behind the target’s motion and greatly increase the risk of loss—a catastrophic failure in applications like surveillance. In contrast, our prediction-informed method (green trajectory) anticipates the target’s future motion. This enables the UAV to plan a smoother, more efficient path that proactively intercepts the predicted path, thereby minimizing tracking error. The surrounding obstacles, perceived by the depth camera, further define the feasible space for these trajectories, underscoring the necessity of prediction in complex environments. However, generating an optimal trajectory is challenging, especially for local replanning from a non-zero initial state, due to the target’s unknown intent and dynamic characteristics. Consequently, the planning module must perform high-frequency replanning. Longer prediction horizons generally lead to more globally optimal planning results, provided the predictions remain accurate.
Assuming the target’s current position T 0 R 2 and velocity T ˙ 0 R 2 can be acquired via onboard sensors or communication, the objective is to predict its future position T ( t p ) at a fixed, known time horizon t p . Algorithm 1 outlines the proposed intent-free trajectory prediction method for moving target. The proposed intent-free trajectory prediction algorithm is designed to address a critical challenge in vision-based UAV tracking: maintaining pursuit of a dynamic target in the absence of high-level intent information. In practical scenarios such as tracking a non-cooperative vehicle or a wild animal, the target’s ultimate goals and planned route are unknown to the UAV. Our algorithm operates under this key assumption, relying solely on real-time kinematic observations (position and velocity) from the onboard vision system. This makes it particularly suitable for applications like surveillance and filmmaking, where the target’s behavior is unpredictable, and the planner must react swiftly based on limited information. The core objective is to generate a short-term, kinematically feasible prediction to enable the UAV’s planner to proactively intercept the target, as opposed to lagging behind it.
Algorithm 1: Intent-free trajectory prediction algorithm for moving target
Input:  T 0 , T ˙ 0 , t p
Output: Predicted trajectory T *
1 i n i ( )
2 T ( t p ) C o m E n d ( )
3While t < t p do
4    for a a l i s t do
5                     x ( t + Δ t ) S t a t e ( a , x ( t ) )
6      end for
7             T * ( t ) S o r t ( )
8Return T *
Prior to the presentation of Algorithm 1, this section details the kinematic modeling of dynamic target. The model employs a second-order chain integral formulation, utilizing acceleration as the control input, as defined below (Note that the following formulation is independent of any knowledge regarding the target’s motion):
x ˙ ( t ) = A x ( t ) + B u ( t )   A = 0 2 × 2 I 2 0 2 × 2 0 2 × 2 , B = 0 2 × 2 I 2
The state vector x ( t ) = T ( t ) T ˙ ( t ) T encompasses the target’s kinematic properties, where T ( t ) = T x ( t ) T y ( t ) T denotes its 2D position, and T ˙ ( t ) represents its velocity. The control input u ( t ) = a =   T ¨ ( t ) is defined as the target’s acceleration, which drives the state evolution. The system matrix A and the input matrix B govern the internal dynamics and how the input affects the state, respectively. I and 0 represent the identity matrix and the zero matrix, respectively.
In Algorithm 1, C o m E n d ( ) calculates a preliminary endpoint position T ( t p ) based on the target’s motion model:
T ˜ ( t p ) = T 0 + T ˙ 0 t p
The position T ˜ ( t p ) serves as an estimate of the final state, providing the heuristic cost required to reach it. In line 4, a l i s t denotes the discrete control input, defined as follows:
a l i s t = a max , N 1 N a max , , N 1 N a max , a max ,
In Line 6, the process S t a t e ( ) computes the target state x ( t + Δ t ) using Equation (1). This subsequent state is then used to select the optimal control input by applying a time-energy optimization principle:
arg min a a l i s t   f = g ( a ) + h ( x ( t + Δ t ) , T ˜ ( t p ) )
where g denotes the actual cost, and h denotes the heuristic cost.
g ( a ) = λ 1 a   h ( x ( t + Δ t ) , T ˜ ( t p ) = T ( t + Δ t ) T ˜ ( t p )
The coefficient λ 1 balances the weight of these two cost components. Notably, Equation (5) employs the Euclidean distance from the target to the estimated position as a proxy for time. Once Equation (5) is computed, the optimal control input corresponding to the target position T ( t + Δ t ) is selected and appended to the predicted trajectory sequence T * . This operation is executed in Line 7.
Compared to learning-based predictors [10], our method is data-efficient and does not require extensive offline training on specific motion datasets. It is robust to novel, unseen target behaviors since it is based on fundamental kinematics rather than pattern matching. However, a key limitation is its inability to leverage known contextual information or historical patterns for long-horizon prediction, which learning-based methods can potentially exploit if such data is available. Compared to other model-based predictors, such as the Constant Velocity (CV) or Constant Acceleration (CA) Kalman filters, our algorithm offers a significant advantage by explicitly generating and evaluating a set of kinematically feasible motion primitives. While a CV/CA model assumes a single, fixed motion model, our approach naturally encompasses a spectrum of maneuvers (e.g., sharp turns, deceleration) through the discrete control set a l i s t making it more adaptable to agile targets.
The primary trade-off lies in computational cost. Simple filters like CV are computationally lighter, whereas our method involves online optimization over a control space, demanding greater processing power. This design choice is justified for our target tracking application, where prediction accuracy over a short horizon is paramount for collision-free interception. In summary, our predictor fills a niche between overly simplistic filters and data-hungry learning models, offering a principled, kinematics-driven approach for short-term prediction in unstructured environments.

4. Front-End: Adaptive Kinodynamic Path Searching

In dynamic target tracking, the trajectory planning module requires high-frequency replanning. Traditional methods, which generate trajectories based on the shortest path [33] (right curve, Figure 3), often result in dynamically infeasible and unsafe trajectories. This issue is particularly pronounced during local replanning with a non-zero initial state. In contrast, the proposed method produces a dynamically feasible and safe trajectory (left curve, Figure 3) by fully accounting for the aerial vehicle’s initial state and performing path searches informed by obstacle proximity. This work operates under the assumption that the target’s position is either perfectly perceived by the sensing algorithms or communicated to the UAV without obstruction.
Our kinodynamic search method builds upon a hybrid A* algorithm [18], expanding nodes (motion primitives) generated by discretizing the control input. This process identifies a safe and dynamically feasible trajectory within a voxel grid map. A key contribution is the thorough utilization of the ESDF map’s prior information to design adaptive search steps and pruning schemes, enabling an efficient, collision-free trajectory search. The complete front-end search process is outlined in Algorithm 2, where O and C denote the open and closed sets, respectively. The open set is a priority queue that stores candidate nodes to be expanded, typically sorted by the sum of the cost from the start node and a heuristic estimate to the goal. The findMinFCost ( ) operation selects the node with the lowest cost for expansion. The closed set records nodes that have already been visited and expanded to avoid redundant searches and prevent infinite loops. Expand ( ) implements the core node expansion process. It generates new candidate nodes by applying kinematically feasible motion primitives from the current node. The contain ( ) method (used as C . contain ( ) and O . contain ( ) ) checks whether a specific node is currently present in the open set or closed set, respectively. This is fundamental for the graph search logic to determine if a node is a new discovery, needs an update, or has already been processed. The Occupy ( ) function checks whether the node lies in an occupied space (collision check) using the ESDF. The algorithm maintains a parent-pointer for each node during the expansion phase (line 19: n i . p a r e n t n c ), which records the optimal predecessor of the node. When the NearEnd ( ) condition is met at node n c , the Return Path() operation is invoked. This operation does not rely on a separately stored path variable. Instead, it dynamically reconstructs the optimal path by starting from the terminal node and sequentially tracing backwards through the chain of parent pointers until the start node is reached.
Algorithm 2: Adaptive Kinodynamic Path Searching
1:Input: initial state x 0 , goal state x T
2:Initialize()
3:while O is not empty do
4: n c O . findMinFCost ( ) , O . pop ( )
5:if NearEnd ( n c ) then
6:              Return Path()
7:end if
8: n o d e s Expand ( n c )
9:for   n i  in  n o d e s  do
10:    if C . contain ( n i )   Occupy ( n i ) then
11:        continue;
12:     end
13:             g t n i . g + EdgeCost ( n i , n c )
14:     if ¬ O . contain ( n i ) then
15:                     O . insert ( n i )
16:    else if g t > n i . g then
17:             continue;
18:     else if
19:                                     n i . g g t , n i . p a r e n t n c
20:                                     n i . f c n i . g + Heu ( n i , x T )
21:                end if
22:             end for
23:   end while
Instead of straight lines, we use motion primitives respecting the aerial drone dynamic as graph edges. Motion primitives are represented by the struct node.

4.1. Motion Primitive

While the original Hybrid A* algorithm in [27] was designed for car-like robots in a 2D plane, our work constitutes a non-trivial extension to 3D space for quadrotor navigation, with key adaptations in state representation and motion primitives. The search state is defined as x = [ p x , v x , a x , p y , v y , a y , p z , v z , a z ] T , and node expansion (Expand()) is governed by the UAV’s kinodynamic model presented in Equation (6) of this paper, which captures the full 3D translational dynamics—a fundamental departure from the 2D bicycle model in [27]. A critical enhancement for aerial navigation is the treatment of orientation: although yaw is not explicitly included in the kinodynamic search, the desired yaw angle φ for target tracking is geometrically derived from the generated path by aligning the UAV’s heading with the horizontal velocity vector ( v x , v y ) . Furthermore, leveraging the differential flatness property of UAV systems [34], the entire state and control inputs can be algebraically determined from the flat outputs ( x , y , z , φ ) . Thus, the trajectory and yaw profile generated by our planner fully define the UAV’s motion, enabling accurate tracking by the downstream geometric controller [25]. We first discuss how to generate motion primitives in Expand(). Similarly to [34], using the differential flatness theory, the state of the aerial drone in path searching is represented by x = [ p x , v x , a x , p y , v y , a y , p z , v z , a z ] T . With the third derivative of displacement j = p as the control input u , the state transition equation is as follows:
x k + 1 = A x k + B u k ,
A = i = 1 3 1 τ 1 2 τ 2 0 1 τ 0 0 1   B = i = 1 3 0 0 1 ,
Here, the symbol denotes the direct sum, which constructs the matrices A and B by stacking the enclosed matrix block diagonally three times, with all off-diagonal blocks being zero. τ is the expansion step, u = [ u max , u max ] R 3 . In three-dimensional space, uniform discrete control inputs u max , N 1 N u max , , N 1 N u max , u max . To fully leverage the prior information from the ESDF—a product of the mapping module that is reused by the planner without incurring additional computational cost—we introduce a risk weight parameter κ , defined as:
κ = < d , v > d v 90 2 π ,
where d is the gradient of the ESDF. As shown in Figure 4, when the velocity direction of the aerial drone is toward the obstacle, the gradient direction given by the ESDF is opposite to the velocity direction, which aggravates the maneuvering danger. Conversely, obstacle information can be disregarded when the gradient direction aligns with the velocity direction. The expansion step τ of the motion primitive is given by the following formula:
τ ( κ ) = min ( τ max 1 + e α κ , τ min ) ,
where α R + is the rate-of-change coefficient. Evidently, the expansion step size is maximized when the directional angle is minimal. Conversely, the step size is minimized when the direction is opposed, as illustrated in Figure 4b.
Unlike traditional search-based path planning algorithms such as A*, which typically consider a fixed number of spatially adjacent nodes (e.g., 8 in 2D or 26 in 3D), the proposed algorithm incorporates a dynamically adjustable point expansion step size mechanism, as shown in Figure 5.
The non-uniform step expansion strategy differentiates our search algorithm and enhances aerial pathfinding in several key aspects. First, the well-designed mapping function in Equation (9) dynamically adjusts the expansion step size based on the risk weight κ , which is derived from obstacle proximity and velocity alignment. This mechanism ensures conservative steps near obstacles for safety while allowing larger steps in open spaces to enhance search efficiency, thereby preventing node expansion in hazardous proximity. Second, since computational time correlates with the number of expanded nodes, larger steps reduce the total node count, simultaneously promoting trajectories away from obstacles for improved safety. Third, compared to traditional A* [33], our method sparsely populates the search space, lowering computational load and enhancing real-time performance. Finally, the strategy fully incorporates the initial state during expansion, enabling dynamically feasible solutions for high-frequency replanning from non-zero initial states.

4.2. Actual Cost and Heuristic Cost

Following the terminology of A*, the cost of each node can be expressed as f c = g c + h c . In this formula, g c denotes the actual cost from the initial state x 0 to the current state x c , while h c represents the heuristic cost to search faster.
To achieve a trade-off between control-effort and trajectory duration, we minimize the cost function defined as:
J ( T ) = 0 T u 2 d t + ρ T ,
In Algorithm 2, EdgeCost ( ) calculates the cost of a motion primitive for each node extension, which is determined by the discrete control inputs and their duration. Under this formulation, the optimal search path is composed of a sequence of these motion primitives. Therefore, the cost function g c is defined as follows:
g c = i = 1 m e i = i = 1 m ( u i 2 + ρ ) τ i ,
An admissible and informative heuristic is essential for accelerating the search process. To this end, the problem is formulated as a two-point boundary value problem [18,32]. Furthermore, the acceleration constraint is relaxed to reduce computational load.
α i β i γ i = 1 T 5 320 120 T 200 T 72 T 2 40 T 2 12 T 3 Δ p i Δ v i ,
Here, Δ p denotes the positional difference between the start and end points, and Δ v represents the velocity difference. i represents three directions. Equation (12) incorporates the UAV’s initial velocity into the path search.
Heu ( ) calculates this heuristic cost J ( T ) by applying Pontryagin’s minimum principle to minimize the cost-to-go from the current state to the target state.
J ( T ) = γ 2 + β γ T + 1 3 β 2 T 2 + 1 3 α γ T 2 + 1 4 α β T 3 + 1 20 α 2 T 4 ,
So, f c = g c + h c = i = 1 m ( u i 2 + ρ ) τ i + J ( T ) . Adjusting the expansion step size of motion primitives offers a tunable trade-off between path quality and computational completeness. Empirical results from both simulations and physical experiments validate the algorithm’s robust performance.

5. Back-End: Trajectory Optimization

While the proposed search algorithm efficiently generates an initial kinodynamically feasible path, the resulting path, like many sampling-based or graph-search methods, may still be in close proximity to obstacles due to the discrete nature of the search and the myopic nature of the cost function. This is a known limitation of such front-end path finders [35]. To address this issue and fully leverage the distance information in free space, we employ a B-spline-based trajectory optimization as the back-end. This two-stage strategy combines the robustness of the search algorithm in complex environments with the ability of gradient-based optimization to refine the trajectory for safety and smoothness. The key is to transform the sparse, discrete path waypoints from the front-end into a continuous, smooth, and collision-free trajectory that explicitly maintains a safe distance from obstacles.

5.1. Boundary Constraint

Boundary constraints define the fixed conditions for the initial state x ( t 0 ) and the final state x ( t f ) This work employs a B-spline for representing the UAV’s three-dimensional trajectory [36,37]. The core of our optimization is the formulation of a collision cost term that actively pushes the trajectory away from obstacles using the ESDF. As defined in Equation (20), this term penalizes control points where the distance to the nearest obstacle is below a safety threshold, ensuring the final optimized trajectory maintains a safe clearance.
The initial state of k degree B-spline is determined only by the initial k control points and the time interval. Given position constraint p = [ p 0 , p 1 , , p n 1 ] , the relationship between position and control points is defined as follows:
C i = 1 6 [ p i + 4 p i + 1 + p i + 2 ] ,
Owing to the recursive property of B-splines, their derivatives are also B-splines. Consequently, the velocity and acceleration control points, denoted as V i and A i , respectively, can be derived from Equation (13):
V i = C i + 1 C i Δ t   ,   A i = V i + 1 V i Δ t ,
where Δ t represents the knot span. For a uniform B-spline, the knot vector is defined such that the time difference between consecutive knots is constant. This time difference, denoted as Δ t , is termed the knot span. Therefore, the boundary constraint of the aerial drone can be expressed by the following formula:
A C = B ,   A = 1 6 1 4 1 3 Δ t 0 3 Δ t 6 Δ t 2 12 Δ t 2 6 Δ t 2 ,   B = p 0 v 0 a 0   o r   p e n d v e n d a e n d ,
where C represent control points. The vectors p 0 , p e n d , v 0 , v e n d , a 0 , a e n d represent the position, velocity, and acceleration at the trajectory’s start and end points, respectively. To accommodate high-frequency replanning, Equation (15) is applied iteratively throughout the process.

5.2. Trajectory Optimization

For a k-degree B-spline defined by n + 1 control points C 0 , C 1 , , C n , boundary constraints fix the positions of the first and last k control points. Therefore, we optimize the remaining n + 1 − 2k control points and define the total cost function as follows:
J a l l = ω s J s ( ) + ω c J c ( ) + ω d J d ( ) ,
The total cost function includes smoothness cost ( J s ), collision cost ( J c ), and dynamic cost ( J d ). ω s , ω c , and ω d represent a trade-off between smoothness, safety, and dynamic feasibility.
In the context of 3D motion planning for UAVs, the dynamic constraints and operational requirements for vertical motion often differ from those in the horizontal plane. For instance, the acceleration and jerk limits along the z-axis are typically more conservative due to the need to directly counteract gravity and ensure stable ascent/descent. To accommodate these distinct characteristics and to enable independent control over the agility in the horizontal plane and the stability in the vertical direction, we formulate separate smoothness cost terms for the xy-plane and the z-axis.
The smoothness cost J s is defined using the squared norm of higher-order derivatives of the trajectory segments. To independently regulate the smoothness in the horizontal plane and the vertical axis, as motivated above, we decompose the cost into two orthogonal components using projection matrices. Let T x y = d i a g ( 1 , 1 , 0 ) and T x y = d i a g ( 1 , 1 , 0 ) be the projection matrices that extract the horizontal (x, y) and vertical (z) components from a 3D vector, respectively. The smoothness cost is then formulated as:
J s = i = 1 n k + 2 ω x y T x y A i 2 2 + i = 1 n k + 2 ω z T z A i 2 2 ,
where ω x y and ω z are their respective weighting coefficients. This formulation allows for fine-grained tuning; for example, a larger ω z can be set to enforce a smoother, more conservative vertical profile, while a smaller ω x y permits more aggressive maneuvers in the horizontal plane.
Noting the particular property of the B-spline, we convert the higher-order derivative information into the geometry information of the trajectory. So we rewrite J s :
J s = i = 1 n k + 2 ω x y T x y ( C i + 2 x y 2 C i + 1 x y + C i x y ) 2 2 + i = 1 n k + 2 ω z T z ( C i + 2 z 2 C i + 1 z + C i z ) 2 2 ,
While the two formulations are fundamentally equivalent, Equation (18) offers distinct advantages: it relies solely on the trajectory’s geometric information, independent of time allocation, and it directly relates the cost to the optimization variables.
For the collision cost J c , it is expressed as the repulsive force of the obstacle acting on each control point:
J c = i = k n k f c ( d ( C i ) ) f c ( C i ) = 0 d ( C i ) > d t h r ( d ( C i ) d t h r ) 2 d ( C i ) d t h r ,
Here, d ( C i ) represents the distance from the i-th control point to the nearest obstacle, obtained from the ESDF gradient map, and d t h r is the safety distance threshold. A penalty is applied when d ( C i ) falls below d t h r .
Dynamic feasibility is enforced by constraining all velocity and acceleration control points [38]. Consequently, penalties are imposed on the maximum velocity and acceleration of the aerial vehicle, formulated as:
J d = i = 1 n k ω v f d ( V i ) + i = 1 n k 1 ω a f d ( A i ) f d ( c r | r = x , y , z ) = a 1 c r 2 + b 1 c r + c 1   ( c r c j ) ( λ c m c r ) 3   ( c j < c r λ c m ) 0   ( λ c m c r λ c m ) ( c r λ c m ) 3   ( λ c m < c r < c j ) a 2 c r 2 + b 2 c r + c 2   ( c r c j ) ,
The function f d ( c ) is an even function that applies no penalty when the velocity or acceleration magnitude remains below a specified threshold. When the threshold is exceeded, a two-stage penalty is activated: a cubic term imposes a progressively increasing cost for exceeding the maximum allowable rate, while a quadratic term is added to ensure numerical stability and prevent unbounded growth.

5.3. Optimization Strategy

This local optimization strategy, also referred to as the Re-Planning Strategy, is illustrated in Figure 6. The process begins by establishing a global target using the prediction data from Algorithm 1. If this global target lies outside the horizon space (denoted by the dotted circle), a line is projected from the UAV’s current position to the global target. The local target is then defined as the point where this line intersects the horizon space boundary. Since planning outside this space is inefficient, all path search and trajectory optimization are confined within it.
Re-planning is triggered under two conditions, as shown in Figure 7. (1) Safety Check after Optimization: Our planning framework consists of a front-end path search that uses a discrete obstacle inflation for geometric feasibility, followed by a back-end trajectory optimization that employs soft constraints for smoothness and more precise obstacle avoidance. After the optimized trajectory is generated, a final safety validation is performed by discretizing the trajectory and rigorously checking for collisions against the ESDF. A re-planning cycle is immediately triggered if any discretized point on the optimized trajectory penetrates an obstacle (i.e., the signed distance ≤ 0). This ensures that the final output is strictly safe, even if the soft-constrained optimizer occasionally produces a trajectory that slightly cuts corners due to cost balancing. (2) Periodic Re-planning: The motion planning module is also invoked periodically at fixed time intervals of T = 0.5 s. This ensures system reactivity to new obstacles and dynamic changes that may not have been present during the previous planning cycle.

6. Result and Analysis

6.1. Implementation Details

The parameters of the proposed framework are summarized in Table 1. The grid resolution and obstacle inflation size were determined by considering the UAV’s physical dimensions and the available onboard computational resources. Specifically, the obstacle inflation size is set to 0.099 m to create a safety buffer around obstacles, effectively compensating for localization and control errors, which is critical for robust collision avoidance.
In the path searching module, the number of discrete steps N significantly influences computational complexity, which grows exponentially with it. The time steps are not fixed but are allowed to adaptively vary between τ min = 0.2 s and τ max = 0.8 s, a range configured to thoroughly evaluate the UAV’s maximum flight speed and the trajectory tracking controller’s performance under dynamic constraints. The change-rate coefficient α was found to have negligible impact on performance within the range of 3 to 5, indicating a degree of algorithmic robustness to this parameter.
For trajectory optimization, the maximum velocity Vmax and acceleration Amax constraints are dictated not only by the UAV’s physical limits but also by the experimental field size and safety considerations. The key weight coefficients— ω s (smoothness), ω c (safety), and ω d (dynamic)—were initialized based on values commonly adopted in related literature [38] and then fine-tuned through extensive simulation to balance flight agility against safety and smoothness. A sensitivity analysis revealed that the trajectory quality is most susceptible to the safety threshold d t h r and the time-penalty coefficient ρ . An excessively small safety threshold increases collision risk, while an overly large one may render the optimization problem infeasible in cluttered spaces. Similarly, ρ directly governs the trade-off between trajectory duration and smoothness.
Finally, the prediction horizon t p for target prediction is set to 2 s. This value represents a standard compromise; longer prediction periods lead to rapidly decaying accuracy due to increased uncertainty, while shorter horizons limit the system’s ability to react to target maneuvers.
The proposed motion planning method is implemented in C++14 and utilizes the lightweight L-BFGS nonlinear solver for optimization [39]. The target is modeled as a freely moving vehicle with a basic planning module. The mapping component of our framework integrates occupancy grid maps with ESDF. The occupancy map is maintained and updated following the methodology of FASTER [21], where the local map’s origin and data are refreshed based on odometry and depth images.

6.2. Simulation Result

We evaluated our AKBS-Tracker (Adaptive Kinodynamic path search and B-Spline-based optimization method) against the well-known Fast-Tracker system, a hierarchical framework for drone tracking, comparing trajectory smoothness, time, and tracking distance [15].
Trajectory Smoothness. A comparative analysis of planning results between AKBS-Tracker and Fast-Tracker under identical environmental conditions is presented in Figure 8. AKBS-Tracker generates shorter paths owing to its adaptive expansion step and demonstrates superior trajectory smoothness. This enhancement stems from the use of B-spline curves and a reduced number of path points in the back-end optimization, which inherently promotes smoothness. Additionally, the convex hull property of the B-spline contributes to safety. This property guarantees that the entire curve segment is strictly contained within the convex hull of its surrounding control points. Consequently, by ensuring that these convex hulls remain entirely within obstacle-free space, the trajectory itself is guaranteed to be collision-free. The resulting trajectory is not only visually smoother and more continuous but also improves the overall stability and reliability of the motion.
Time. To highlight the time advantage of AKBS-Tracker, the maximum speed of the Fast-tracker is set to 3 m/s. As shown in Figure 9, the performance of AKBS-Tracker and Fast-Tracker is compared in terms of speed and task completion time. Even when operating at a higher speed limit, Fast-Tracker required 13.49 s to complete the tracking task, which is 5.2% slower than the 12.78 s achieved by AKBS-Tracker, as shown in Table 2. To ensure statistical significance, the experiment was repeated 10 times with different endpoints. AKBS-Tracker demonstrated consistent performance, with an average duration of 12.84 s and a maximum of 13.07 s.
Tracking distance. Figure 10 presents a comparison of tracking distance, where the proposed method consistently outperforms the Fast-Tracker planner. The red curve, representing the optimal case, achieves a maximum tracking distance of 3.03 m—merely 56.56% of the distance required by Fast-Tracker. Even the worst-performing case reaches only 85.82% of the competitor’s distance. This superior performance is attributed to the high-frequency replanning and the non-uniform expansion strategy of the front-end search algorithm. By incorporating the UAV’s initial velocity state, the algorithm facilitates faster convergence to the target point, while the non-uniform expansion significantly accelerates the search process.
To quantitatively assess the reliability and safety of the proposed planner, we conducted 30 simulation trials across a diverse set of scenarios with varying obstacle densities (ranging from 60 to 80 static obstacles) and target locations. The planner successfully generated and executed collision-free trajectories in 29 cases, resulting in a success rate of 96.7%. The relevant test results, including dataset, are available as per the Data Availability Statement.
A single planning failure occurred in the most cluttered scenario (80 obstacles). We analyzed this case to understand the system’s limitations. The front-end search successfully provided an optimal, collision-free initial path. However, the back-end optimizer failed to refine this path into a smooth, collision-free trajectory within the allocated time budget. The failure was attributed to the highly non-convex nature of the collision cost function in such a dense environment, which caused the gradient-based solver to become trapped in a poor local minimum. This highlights a known challenge for local optimization methods and represents a boundary of the current system’s performance.

6.3. Onboard Autonomous Flight

The experiment was conducted in an unknown indoor environment. The UAV platform, depicted in Figure 11, primarily comprises flight control hardware, an IMU, and a support frame. The onboard computing unit is an Intel NUC11-TNKi5, and the depth camera is a RealSense D435i.
The indoor motion planning task was performed in a cluttered environment, as illustrated in Figure 12. The UAV’s maximum velocity was constrained to 0.5 m/s to comply with the safety regulations of the testing facility.
The UAV’s flight is documented in Figure 13 and Figure 14. Figure 13 shows a grayscale image from the onboard camera, and Figure 14 provides a third-person view of the same flight. The experimental results confirm that the proposed algorithm achieves excellent real-time performance, running efficiently on the onboard microcomputer.
Figure 15 shows the UAV’s flight trajectory, demonstrating successful obstacle avoidance in an unknown environment and real-time tracking of local targets, which are the points where the line-of-sight to the global goal intersects the planning horizon (as defined in Section 5.3 and Figure 6). An initial discrepancy between the planned and actual trajectory is observed, which converges over time. This is attributed to the path search algorithm’s initial neglect of the yaw angle, causing a transient tracking error. Regarding velocity, Figure 16 confirms that the UAV’s speed remained within the predefined maximum limit, thus adhering to the specified constraint.

7. Conclusions

This paper introduced an integrated onboard motion planning framework for autonomous drones, with contributions in two key areas. First, we propose a lightweight, intent-free method for motion prediction driven purely by spatiotemporal dynamics, which provides a critical foundation for tracking unpredictable targets. Building upon this, a hierarchical planner combining front-end kinodynamic path searching and back-end B-spline trajectory optimization was developed to address the challenge of high-speed onboard replanning. The front-end search efficiently generates a safe, dynamically feasible initial path by incorporating a novel risk-weight parameter to reduce computational complexity. Subsequently, the back-end optimization refines this path for smoothness and safety using boundary constraints and gradient-based methods, leveraging the convex hull property of B-splines to enhance optimization efficiency. The proposed framework was rigorously validated through extensive simulations and real-world experiments, demonstrating robust performance. Our conclusions can be summarized as follows:
(1)
For trajectory planning for single drone target tracking, the proposed AKBS-Tracker generates shorter trajectories than the traditional Fast-Tracker, with notable advantages in trajectory smoothness, time and tracking distance.
(2)
The AKBS-Tracker decreased flight time effectively while maintaining the planned speed within the defined maximum speed limit, demonstrating its advantage in sustaining dynamic feasibility.
(3)
In a real-world experiment conducted in an unfamiliar indoor environment, the AKBS-Tracker successfully executed motion planning for a single UAV. It adeptly maintained a smooth and dynamic trajectory while skillfully avoiding collisions with obstacles. This experiment validates the system’s capability to operate safely and effectively in practical scenarios.
This research significantly improves autonomous UAV efficiency, safety, and practical applicability for target tracking. The potential impact extends to various applications, such as surveillance, delivery, and exploration. Nevertheless, acknowledging the limitations of our study is crucial. Given the limits of our study, we did not take into account the communications of tracking and visual occlusion. Subsequent research should take into account the robustness to sensor noise and communication delay. This can be achieved by scaling the approach, integrating advanced sensors, creating adaptive algorithms, conducting real-world experiments, and facilitating practical implementation in various applications.

Author Contributions

Conceptualization, S.H.; methodology, S.H.; software, Z.T.; validation, S.H.; formal analysis, S.H.; investigation, W.Z.; resources, S.H.; data curation, Z.T.; writing—original draft preparation, W.Z.; writing—review and editing, B.Z.; visualization, S.H.; supervision, B.Z.; project administration, B.Z.; funding acquisition, B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 52202513, 52302511, and 52202454) and the Guangdong Basic and Applied Basic Research Foundation (No. 2023A1515010023).

Data Availability Statement

The experimental data generated and analyzed during this study are available in the GitHub repository: https://github.com/520hs/Adaptive-Trajectory-Planning-for-UAV-Target-Tracking-Based-on-Improved-Hierarchical-Framework.git (accessed on 4 December 2025). The version of the data used in this study corresponds to Release v1.0 of the GitHub repository. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Qin, T.; Cao, S.; Pan, J.; Shen, S. A General Optimization-based Framework for Global Pose Estimation with Multiple Sensors. IET Cyber-Syst. Robot. 2025, 7, e70023. [Google Scholar] [CrossRef]
  2. Bai, Z.; Zhang, B.; Song, M.; Tian, Z. Rapid Integrated Design Verification of Vertical Take-Off and Landing UAVs Based on Modified Model-Based Systems Engineering. Drones 2024, 8, 755. [Google Scholar] [CrossRef]
  3. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  4. Ji, J.; Pan, N.; Xu, C.; Gao, F. Elastic Tracker: A Spatio-temporal Trajectory Planner for Flexible Aerial Tracking. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 47–53. [Google Scholar]
  5. Tang, L.; Wang, H.; Li, P.; Wang, Y. Real-time Trajectory Generation for Quadrotors using B-spline based Non-uniform Kinodynamic Search. In Proceedings of the 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), Dali, China, 6–8 December 2019; pp. 1133–1138. [Google Scholar] [CrossRef]
  6. Gu, Y.; Guo, K.; Guo, L.; Qiao, J.; Jia, J.; Yu, X.; Xie, L. An enhanced UAV safety control scheme against attacks on desired trajectory. Aerosp. Sci. Technol. 2021, 119, 107212. [Google Scholar] [CrossRef]
  7. Li, X.R.; Jilkov, V.P. Survey of Maneuvering Target Tracking. Part V: Multiple-Model Methods. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 1255–1321. [Google Scholar] [CrossRef]
  8. Kim, J.-H.; Lee, D.W.; Cho, K.-R.; Jo, S.-Y.; Kim, J.-H.; Min, C.-O.; Han, D.-I.; Cho, S.-J. Development of an Electro-Optical System for Small UAV. Aerosp. Sci. Technol. 2010, 14, 505–511. [Google Scholar] [CrossRef]
  9. Kim, J.; Shim, D.H. A vision-based target tracking control system of a quadrotor by using a tablet computer. In Proceedings of the 2013 International Conference on Unmanned Aircraft Systems (ICUAS), Atlanta, GA, USA, 28–31 May 2013; pp. 1165–1172. [Google Scholar] [CrossRef]
  10. Zhao, X.; Huang, X.; Cheng, J.; Xia, Z.; Tu, Z. A Vision-Based End-to-End Reinforcement Learning Framework for Drone Target Tracking. Drones 2024, 8, 628. [Google Scholar] [CrossRef]
  11. Ferreira, D.; Basiri, M. Dynamic Target Tracking and Following with UAVs Using Multi-Target Information: Leveraging YOLOv8 and MOT Algorithms. Drones 2024, 8, 488. [Google Scholar] [CrossRef]
  12. Thomas, J.; Welde, J.; Loianno, G.; Daniilidis, K.; Kumar, V. Autonomous Flight for Detection, Localization, and Tracking of Moving Targets with a Small Quadrotor. IEEE Robot. Autom. Lett. 2017, 2, 1762–1769. [Google Scholar] [CrossRef]
  13. Kendall, A.G.; Salvapantula, N.N.; Stol, K.A. On-board object tracking control of a quadcopter with monocular vision. In Proceedings of the 2014 International Conference on Unmanned Aircraft Systems (ICUAS), Orlando, FL, USA, 27–30 May 2014; pp. 404–411. [Google Scholar] [CrossRef]
  14. Cheng, H.; Lin, L.; Zheng, Z.; Guan, Y.; Liu, Z. An autonomous vision-based target tracking system for rotorcraft unmanned aerial vehicles. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1732–1738. [Google Scholar] [CrossRef]
  15. Nägeli, T.; Alonso-Mora, J.; Domahidi, A.; Rus, D.; Hilliges, O. Real-Time Motion Planning for Aerial Videography with Dynamic Obstacle Avoidance and Viewpoint Optimization. IEEE Robot. Autom. Lett. 2017, 2, 1696–1703. [Google Scholar] [CrossRef]
  16. Li, X.R.; Bar-Shalom, Y. Performance Prediction of the Interacting Multiple Model Algorithm. IEEE Trans. Aerosp. Electron. Syst. 1993, 29, 755–771. [Google Scholar] [CrossRef]
  17. Canolla, A.; Jamoom, M.B.; Pervan, B. Interactive Multiple Model Sensor Analysis for Unmanned Aircraft Systems (UAS) Detect and Avoid (DAA). In Proceedings of the 2018 IEEE/ION Position, Location and Navigation Symposium (PLANS), Monterey, CA, USA, 23–26 April 2018. [Google Scholar]
  18. Chen, J.; Liu, T.; Shen, S. Tracking a moving target in cluttered environments using a quadrotor. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 446–453. [Google Scholar] [CrossRef]
  19. Han, Z.; Zhang, R.; Pan, N.; Xu, C.; Gao, F. Fast-Tracker: A Robust Aerial System for Tracking Agile Target in Cluttered Environments. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021. [Google Scholar]
  20. Ding, W.; Gao, W.; Wang, K.; Shen, S. An Efficient B-Spline-Based Kinodynamic Replanning Framework for Quadrotors. IEEE Trans. Robot. 2019, 35, 1287–1306. [Google Scholar] [CrossRef]
  21. Zhou, B.; Pan, J.; Gao, F.; Shen, S. RAPTOR: Robust and Perception-Aware Trajectory Replanning for Quadrotor Fast Flight. IEEE Trans. Robot. 2021, 37, 1992–2009. [Google Scholar] [CrossRef]
  22. Zhou, B.; Gao, F.; Wang, L.; Liu, C.; Shen, S. Robust and Efficient Quadrotor Trajectory Generation for Fast Autonomous Flight. IEEE Robot. Autom. Lett. 2019, 4, 3529–3536. [Google Scholar] [CrossRef]
  23. Zhou, B.; Gao, F.; Pan, J.; Shen, S. Robust Real-time UAV Replanning Using Guided Gradient-based Optimization and Topological Paths. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020. [Google Scholar]
  24. Pan, N.; Zhang, R.; Yang, T.; Cui, C.; Xu, C.; Gao, F. Fast-Tracker 2.0: Improving autonomy of aerial tracking with active vision and human location regression. IET Cyber-Syst. Robot. 2021, 3, 292–301. [Google Scholar] [CrossRef]
  25. Liu, Y.; Wang, H.; Fan, J.; Wu, J.; Wu, T. Control-oriented UAV highly feasible trajectory planning: A deep learning method. Aerosp. Sci. Technol. 2021, 110, 106435. [Google Scholar] [CrossRef]
  26. Ding, W.; Gao, W.; Wang, K.; Shen, S. Trajectory Replanning for Quadrotors Using Kinodynamic Search and Elastic Optimization. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 7595–7602. [Google Scholar]
  27. Dolgov, D.; Thrun, S.; Montemerlo, M.; Diebel, J. Practical Search Techniques in Path Planning for Autonomous Driving. Ann Arbor 2008, 1001, 48105. [Google Scholar]
  28. Tan, J.; Xue, S.; Guo, Z.; Li, H.; Zheng, X.; Cao, H. Adaptive hierarchical control of quadcopters via safe reinforcement learning from human demonstration. Eng. Appl. Artif. Intell. 2026, 163, 113013. [Google Scholar] [CrossRef]
  29. Tan, J.; Xue, S.; Li, H.; Guo, Z.; Cao, H.; Chen, B. Hierarchical Safe Reinforcement Learning Control for Leader-Follower Systems With Prescribed Performance. IEEE Trans. Autom. Sci. Eng. 2025, 22, 19568–19581. [Google Scholar] [CrossRef]
  30. Yin, H.; Xu, X.; Lu, S.; Chen, X.; Xiong, R.; Shen, S.; Stachniss, C.; Wang, Y. A Survey on Global LiDAR Localization: Challenges, Advances and Open Problems. Int. J. Comput. Vis. 2024, 132, 3139–3171. [Google Scholar] [CrossRef]
  31. Han, L.; Gao, F.; Zhou, B.; Shen, S. FIESTA: Fast Incremental Euclidean Distance Fields for Online Motion Planning of Aerial Robots. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019. [Google Scholar]
  32. Mueller, M.W.; Hehn, M.; D’Andrea, R. A Computationally Efficient Motion Primitive for Quadrocopter Trajectory Generation. IEEE Trans. Robot. 2015, 31, 1294–1310. [Google Scholar] [CrossRef]
  33. Tang, L.; Wang, H.; Liu, Z.; Wang, Y. A real-time quadrotor trajectory planning framework based on B-spline and nonuniform kinodynamic search. J. Field Robot. 2020, 38, 452–475. [Google Scholar] [CrossRef]
  34. Faessler, M.; Franchi, A.; Scaramuzza, D. Differential Flatness of Quadrotor Dynamics Subject to Rotor Drag for Accurate Tracking of High-Speed Trajectories. IEEE Robot. Autom. Lett. 2018, 3, 620–626. [Google Scholar] [CrossRef]
  35. Quan, L.; Han, L.; Zhou, B.; Shen, S.; Gao, F. A survey of uav motion planning. IET Cyber-Syst. Robot. 2020, 2, 8. [Google Scholar] [CrossRef]
  36. Qin, K. General matrix representations for b-splines. Vis. Comput. 2000, 16, 177–186. [Google Scholar] [CrossRef]
  37. Wang, L.; Guo, Y. Speed Adaptive Robot Trajectory Generation Based on Derivative Property of B-Spline Curve. IEEE Robot. Autom. Lett. 2023, 8, 1905–1911. [Google Scholar] [CrossRef]
  38. Zhou, X.; Wang, Z.; Ye, H.; Xu, C.; Gao, F. EGO-Planner: An ESDF-Free Gradient-Based Local Planner for Quadrotors. IEEE Robot. Autom. Lett. 2021, 6, 478–485. [Google Scholar] [CrossRef]
  39. Liu, D.C.; Nocedal, J. On the limited memory bfgs method for large scale optimization. Math. Program. 1989, 45, 503–528. [Google Scholar] [CrossRef]
Figure 1. Framework of aerial drone autonomous navigation system.
Figure 1. Framework of aerial drone autonomous navigation system.
Drones 09 00847 g001
Figure 2. Motion planning methods that lack trajectory prediction produce inefficient trajectories, as illustrated by the red curves originating from the current goal. In contrast, a method incorporating trajectory prediction generates a dynamically feasible and efficient trajectory, shown by the orange-red curve.
Figure 2. Motion planning methods that lack trajectory prediction produce inefficient trajectories, as illustrated by the red curves originating from the current goal. In contrast, a method incorporating trajectory prediction generates a dynamically feasible and efficient trajectory, shown by the orange-red curve.
Drones 09 00847 g002
Figure 3. Traditional methods generate a trajectory (right curve) from the shortest path but face the danger of dynamical infeasibility and insecurity. In contrast, our trajectory obtains a dynamically feasible and safe trajectory.
Figure 3. Traditional methods generate a trajectory (right curve) from the shortest path but face the danger of dynamical infeasibility and insecurity. In contrast, our trajectory obtains a dynamically feasible and safe trajectory.
Drones 09 00847 g003
Figure 4. (a) Illustration of d , ϑ changing in the planning process. ϑ is the angle between gradient and velocity. (b) The mapping angle to the step size, the more dangerous the scene, the smaller the step size.
Figure 4. (a) Illustration of d , ϑ changing in the planning process. ϑ is the angle between gradient and velocity. (b) The mapping angle to the step size, the more dangerous the scene, the smaller the step size.
Drones 09 00847 g004
Figure 5. Comparison of A* algorithm and our search algorithm node expansion strategy.
Figure 5. Comparison of A* algorithm and our search algorithm node expansion strategy.
Drones 09 00847 g005
Figure 6. Re-planning Strategy. The local target is the point where this line intersects the boundary of the horizon range (the circle). Outside the circle lies an obstacle beyond perception.
Figure 6. Re-planning Strategy. The local target is the point where this line intersects the boundary of the horizon range (the circle). Outside the circle lies an obstacle beyond perception.
Drones 09 00847 g006
Figure 7. Two scenarios of re-planning.
Figure 7. Two scenarios of re-planning.
Drones 09 00847 g007
Figure 8. (ac) illustrate the AKBS-Tracker’s process for tracking a moving target, while (d) presents the corresponding result from the Fast-Tracker for comparison. The blue box contains a magnified view of the area.
Figure 8. (ac) illustrate the AKBS-Tracker’s process for tracking a moving target, while (d) presents the corresponding result from the Fast-Tracker for comparison. The blue box contains a magnified view of the area.
Drones 09 00847 g008
Figure 9. The comparison of the two planners in terms of speed and time.
Figure 9. The comparison of the two planners in terms of speed and time.
Drones 09 00847 g009
Figure 10. Comparison of two planners in tracking distance.
Figure 10. Comparison of two planners in tracking distance.
Drones 09 00847 g010
Figure 11. Overview of our quadrotor system (a) A rack containing flight control hardware and IMU (b) Onboard computing unit (c) The depth camera d435i.
Figure 11. Overview of our quadrotor system (a) A rack containing flight control hardware and IMU (b) Onboard computing unit (c) The depth camera d435i.
Drones 09 00847 g011
Figure 12. Schematic diagram of UAV onboard flight experiment.
Figure 12. Schematic diagram of UAV onboard flight experiment.
Drones 09 00847 g012
Figure 13. Drone real flight image–Drone perspective. (a) represent the shot I. (b) represent the shot II. (c) represent the shot III. (d) represent the shot IV.
Figure 13. Drone real flight image–Drone perspective. (a) represent the shot I. (b) represent the shot II. (c) represent the shot III. (d) represent the shot IV.
Drones 09 00847 g013
Figure 14. Drone real flight image–third person view. (a) represent the shot I. (b) represent the shot II. (c) represent the shot III. (d) represent the shot IV. The red circle mark the UAVs.
Figure 14. Drone real flight image–third person view. (a) represent the shot I. (b) represent the shot II. (c) represent the shot III. (d) represent the shot IV. The red circle mark the UAVs.
Drones 09 00847 g014
Figure 15. Trajectory curve from real-world UAV flight test.
Figure 15. Trajectory curve from real-world UAV flight test.
Drones 09 00847 g015
Figure 16. Speed curve from real-world UAV flight test.
Figure 16. Speed curve from real-world UAV flight test.
Drones 09 00847 g016
Table 1. Parametric setting for framework.
Table 1. Parametric setting for framework.
ModuleParameterDescription and UnitValue
Environmental Perception and Map RepresentationResolutionMap resolution (m)0.1
ObsInflateObstacle inflate size (m)0.099
Target Prediction t p Predicted duration (s)2
λ 1 Predicted coefficient2
Path SearchingNNumber of Discrete Steps3
τ min Minimum step constraint (s)0.2
τ max Maximum step constraint (s)0.8
α change rate coefficient3
ρ time penalty coefficient10
Trajectory OptimizationVmaxVelocity constraint (m/s)2.5
AmaxAcceleration constraint (m/s2)3.5
ω s Smoothness weight1
ω c Safety weight0.5
ω d dynamic weight0.1
ω x y horizontal (x, y) weight1
ω z Vertical plane weight3
d t h r Safety threshold0.4
ω v Velocity weight1
ω a Acceleration weight1
UAV parametersmMass of UAV (kg)2.5
LWheelbase (mm)410
Table 2. Flight times planned by different planners.
Table 2. Flight times planned by different planners.
TrackerFastAKBSAKBSAKBSAKBS
TimeTimeAverage TimeMax TimeMin Time
flight time (s)13.4912.7812.838713.0712.74
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, S.; Tian, Z.; Zhu, W.; Zhang, B. An Improved Hierarchical Framework for Adaptive UAV Trajectory Planning in Dynamic Target Tracking. Drones 2025, 9, 847. https://doi.org/10.3390/drones9120847

AMA Style

Hu S, Tian Z, Zhu W, Zhang B. An Improved Hierarchical Framework for Adaptive UAV Trajectory Planning in Dynamic Target Tracking. Drones. 2025; 9(12):847. https://doi.org/10.3390/drones9120847

Chicago/Turabian Style

Hu, Sen, Zhong Tian, Weiyu Zhu, and Bangchu Zhang. 2025. "An Improved Hierarchical Framework for Adaptive UAV Trajectory Planning in Dynamic Target Tracking" Drones 9, no. 12: 847. https://doi.org/10.3390/drones9120847

APA Style

Hu, S., Tian, Z., Zhu, W., & Zhang, B. (2025). An Improved Hierarchical Framework for Adaptive UAV Trajectory Planning in Dynamic Target Tracking. Drones, 9(12), 847. https://doi.org/10.3390/drones9120847

Article Metrics

Back to TopTop