Next Article in Journal
Study on Corrugated Tube Structure Within the Tempcore Process Based on Large Steel Bar Cooling Efficiency
Previous Article in Journal
A Multi-Layer Triboelectric Material Deep Groove Ball Bearing Triboelectric Nanogenerator: Speed and Skidding Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Human–Machine Collaboration: A Trust-Aware Trajectory Planning Framework for Assistive Aerial Teleoperation

1
National Key Laboratory of Avionics Integration and Aviation System-of-Systems Synthesis, China Aeronautical Radio Electronics Research Institute, Shanghai 201100, China
2
Department of Automation, University of Science and Technology of China, Hefei 230026, China
3
Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei 230088, China
*
Author to whom correspondence should be addressed.
Machines 2025, 13(9), 876; https://doi.org/10.3390/machines13090876
Submission received: 14 August 2025 / Revised: 14 September 2025 / Accepted: 17 September 2025 / Published: 20 September 2025
(This article belongs to the Special Issue Advances in AI-Powered Human-Machine-Augmented Intelligence)

Abstract

Human–machine collaboration in assistive aerial teleoperation is frequently compromised by trust imbalances, which arise from the vehicle’s complex dynamics and the operator’s constrained perceptual feedback. We introduce a novel framework that enhances collaboration by dynamically integrating a model of human trust into the unmanned aerial vehicle’s trajectory planning. We first propose a Machine-Performance-Dependent trust model, specifically tailored for aerial teleoperation, that quantifies trust based on real-time safety and visibility metrics. This model then informs a trust-aware trajectory planning algorithm, which generates smooth and adaptive trajectories that continuously align with the operator’s trust level and intent inferred from control inputs. Extensive simulations conducted in diverse forest environments validate our approach. The results demonstrate that our method achieves task efficiency comparable to that of a trust-unaware baseline while significantly reducing operator workload and improving trajectory smoothness, achieving reductions of up to 23.2% and 43.2%, respectively, in challenging dense environments. By embedding trust dynamics directly into the trajectory optimization loop, this work pioneers a more intuitive, efficient, and resilient paradigm for assistive aerial teleoperation.

1. Introduction

Unmanned aerial vehicle (UAV) teleoperation technology enables UAVs to access hazardous or human-inaccessible environments, performing tasks such as search and rescue [1] and surveillance [2]. It not only ensures the safety of operators but also gives full play to the advantages of human judgment and skills. However, in the traditional direct teleoperation mode, operators are responsible for all control tasks of the UAV, including flight path planning and obstacle avoidance [3], while the UAV itself lacks any high-level decision-making autonomy. This mode not only requires operators to possess extensive professional knowledge and experience but also leads to cognitive fatigue due to frequent manual operations [4]. Therefore, the question of how to reduce operational difficulty and alleviate the operator’s burden, especially when performing tasks in complex and hazardous environments, remains a key challenge in UAV teleoperation technology [3,5].
With the advancement of UAV autonomous technology, the assistive aerial teleoperation mode has been widely recognized as an effective solution to address the aforementioned challenges, for example, by reducing high cognitive load and mitigating the need for extensive operator expertise [3,6]. In this mode, operators provide commands through remote control, while the UAV offers assistance by inferring the operators’ intent [5,6,7] and autonomously plans appropriate action strategies based on mission requirements. By combining human decision-making with machine-assisted execution, this collaborative approach can significantly reduce the operator’s workload while improving task efficiency and safety.
However, while assistive aerial teleoperation technology has made significant progress in reducing operators’ workload and other aspects, it has also triggered the issue of trust imbalance in human–machine collaboration [8,9]. Due to the complex high-dimensional nonlinear dynamic characteristics of a UAV and the limited perceptual capabilities of operators through remote control terminals [10], it is difficult for humans to intuitively understand the UAV’s status and performance. Such information asymmetry [11] can lead to human trust imbalance, resulting in very serious consequences: over-trust may cause operators to fail to intervene in a timely manner in the UAV’s erroneous behaviors [12], thereby leading to mission failure. while a lack of trust may cause operators to frequently conduct unnecessary interventions [13], increasing the operational burden and reducing task execution efficiency.
Addressing this critical issue of trust imbalance reveals a twofold research gap, spanning from theory to application, that has hindered progress in the field. First, in the area of trust modeling, research specific to assistive aerial teleoperation remains limited. Most existing trust models are developed for other scenarios, such as ground mobile robots [14,15] and autonomous vehicles [16,17]. However, as these models do not account for the higher environmental complexity and the operator’s limited perceptual capabilities inherent to UAV teleoperation, they cannot be directly applied. Second, there is a practical gap in trust application. Even with a suitable model, the question of how to operationally use this dynamic trust information to guide the UAV’s behavior remains largely unanswered. Most current assistive planners focus on aligning with operator intent [3,5,7], but they lack a mechanism to modulate their behavior based on the operator’s dynamic trust level. A planning framework that can integrate a real-time trust metric into its decision-making process is a crucial missing piece.
To address this twofold research gap, this paper introduces a novel framework for trust-aware trajectory planning in assistive aerial teleoperation. Our approach is founded upon the Machine-Performance-Dependent (MPD) human trust model framework, which provides a quantitative basis for linking objective system performance to an operator’s trust. We adapt and specialize this general framework for the UAV teleoperation scenario and, crucially, integrate it into a real-time trajectory planner. By tackling both the modeling and application gaps, the key contributions of this work are summarized as follows:
  • We propose an MPD trust model tailored for assistive aerial teleoperation. The model quantifies the UAV’s performance capacity using safety and visibility metrics, which, in contrast to methods relying on abstract or binary scores, enables a more nuanced and dynamic representation of human trust that can inform trajectory planning.
  • Building upon this model, we develop a trust-aware trajectory planner that, for the first time, integrates a dynamic trust level directly into the optimization loop. This enables the planner to continuously adapt its assistance strategy, moving beyond simple intent-matching to achieve true human–machine collaboration.
  • We validate our approach through extensive simulations in challenging, randomly generated forest environments. The results confirm that our trust-aware method significantly reduces operator workload and enhances trajectory smoothness, achieving superior collaborative performance without compromising task efficiency compared to a trust-unaware baseline.
Our approach assumes a stable communication link between the operator and the UAV. The trust model parameters are calibrated using experimental data to estimate human factors. However, the model does not account for real-time adaptation to individual differences. The proposed trust model is validated in simulated forest environments, and its applicability in highly dynamic urban settings remains a subject for future work.
The remainder of this paper is structured as follows. Section 2 surveys the literature on assistive aerial teleoperation, along with related work on human-factor-aware trajectory planning and trust modeling. Section 3 outlines the problem formulation and introduces the foundational MPD trust model. Building on this, Section 4 details the proposed trust model tailored for UAVs, while Section 5 presents our core trust-aware trajectory planning framework. Section 6 describes the experiments and analyzes the results. Section 7 interprets the findings, illustrates the framework’s practical significance through a scenario-based evaluation, and outlines open issues for future work. Finally, Section 8 concludes this paper with a summary of our contributions.

2. Related Work

2.1. Assistive Aerial Teleoperation

Existing research in assistive aerial teleoperation primarily focuses on two levels: operational control and trajectory planning [3]. At the control level, some works model the task as a local safety problem by generating constraints to mitigate collision risks [18,19]. Other works, such as Backman et al. [20], propose a shared autonomy approach that learns a policy module to provide assistive control inputs for safe landing via the reinforcement learning algorithm TD3. At the planning level, other works concentrate on longer-horizon human–machine collaborative planning. For instance, Yang and Michael [7] propose a motion-primitive-based sampling method to generate long-horizon trajectories which align with the human intention while circumventing obstacles. Similarly, Wang et al. [6] present a gaze-enhanced teleoperation framework that generates a safe and perception-aware trajectory by capturing the operator’s exploratory intent from gaze information. However, while these approaches show increasing technical sophistication, their interaction models often lack a deep consideration of human factors, such as operator preferences and trust, which are critical for effective collaboration.

2.2. Trajectory Planning Considering Human Factors

In the field of autonomous driving, significant research has focused on trajectory planning that incorporates human factors to create more intelligent and acceptable vehicle behaviors. Some works focus on understanding the human-like behaviors of surrounding vehicles for better prediction. For instance, Zhang et al. [21] propose an interaction-aware and driving-style-aware model that first uses an unsupervised clustering algorithm to classify the driving styles of other drivers and then employs an encoder–decoder architecture for long-term trajectory prediction. Other works aim to make the autonomous vehicle’s own path feel more natural to the user. Igneczi et al. [22] develop a curve trajectory model based on a linear driver model, which calculates human-preferred node points from road curvature data to design a path that emulates average human driving patterns. Furthermore, to cater to individual user needs, some methods allow for explicit personalization. Huang et al. [23] present a personalized trajectory planning method for lane-change maneuvers, which uses an improved RRT algorithm to generate a cluster of feasible trajectories and then utilizes fuzzy linguistic preference relations and TOPSIS to select the optimal one that matches the user’s subjective preferences for safety and comfort. While these methods effectively integrate human factors like driving style and preference into ground-based trajectory planning, their underlying models and assumptions are tied to the structured nature of road networks and vehicle dynamics, which do not directly apply to the challenges of 3D aerial teleoperation. Among these human factors, the operator’s dynamic trust in the system’s capability is arguably the most critical yet least understood, especially in high-stakes aerial tasks.

2.3. Trust Models in Human–Machine Interaction

In fields beyond UAV teleoperation, particularly in autonomous driving and ground mobile robotics, trust modeling has been explored to facilitate smoother and safer human–machine collaboration. For instance, Li et al. [14] proposed a bidirectional trust-based variable autonomy (BTVA) approach for the field of semi-autonomous mobile robots to balance control between a human and a robot. Similarly, Saeidi and Wang [15] proposed a trust framework for the same field to resolve human–machine intent conflicts by continuously modulating the degree of automation. In the autonomous driving domain, Hu et al. [16] proposed a quantitative trust model to dynamically allocate control authority by measuring the deviation between vehicle performance and driver expectations. Furthermore, Fang et al. [17] proposed a hierarchical framework based on mutual trust evaluation for shared steering control to optimize authority allocation by evaluating driver skill and human–machine conflicts. However, these frameworks’ underlying assumptions do not hold in UAV teleoperation. The key distinction is the shift from navigating on a plane to maneuvering in complex 3D space, a challenge compounded by the operator’s limited situational awareness [10]. A critical consequence is that the machine’s objective capability becomes highly volatile and context-dependent. Existing trust models for more predictable ground systems do not explicitly parameterize this performance volatility. Our work, therefore, builds upon the MPD human trust model, which addresses this gap by modeling the machine’s context-dependent capability to achieve a more accurate and adaptive trust assessment.

3. Problem Formulation and Preliminaries

We formalize assistive aerial teleoperation as a human–UAV collaborative trajectory planning problem and employ the MPD trust model to characterize human trust dynamics. This section provides an overview of human–UAV collaborative trajectory planning and the MPD trust model.

3.1. Human–UAV Collaborative Trajectory Planning

Let the state vector of the UAV be denoted as x ( t ) X R n , where X represents the state space and n is the dimension of the state vector. Typically, x ( t ) comprises the UAV’s position, velocity, and other variables relevant to its dynamics. The control input is defined as u ( t ) U R m , where U denotes the admissible set of control inputs and m is the dimension of the control input. The control input u ( t ) governs the state transitions of the UAV.
To ensure both safety and operational feasibility during flight, the UAV’s trajectory must satisfy multiple constraints. These constraints arise from the system’s physical limitations and environmental factors. Specifically, we define ξ ( t ) as the position component of the state vector x ( t ) , representing the trajectory over a time horizon t [ 0 , T ] . The dynamic feasibility constraints are expressed as shown in (1):
ξ ( i ) ( t ) x max ( i ) , t [ 0 , T ]
where the i-th-order derivative of the trajectory ξ ( t ) is upper-bounded by a known limit x max ( i ) of the UAV. Additionally, the UAV must avoid obstacles in the environment during flight. To this end, the trajectory should always lie within the safe region, as specified by the constraint in (2):
ξ ( t ) X safe , t [ 0 , T ]
where X safe = X O , and O denotes the set of obstacles.
In human–UAV collaborative scenarios without a predefined target, it is more appropriate to model the operator’s intent as directional preferences rather than specific state-space targets [24]. The operator conveys their intent by providing control signals u h through input devices, such as joysticks. To integrate this intent into trajectory planning, we introduce an intent cost function C u h ( ξ ) , which quantifies the alignment between the trajectory ξ ( t ) and the operator’s directives. Concurrently, a flight quality cost function C q ( ξ ) is defined to assess metrics such as trajectory smoothness and energy efficiency.
Thus, the human–UAV collaborative trajectory planning problem can be formulated as the following optimization problem: given the operator’s input u h , compute a trajectory ξ within the feasible space to minimize the combined cost, as shown in (3):
min ξ ( i ) x max ( i ) , ξ X safe C u h ( ξ ) + C q ( ξ )

3.2. The MPD Human Trust and Its Dynamic Model

The MPD trust (denoted by T m h ) is defined as the ratio of subjective machine capability (denoted by C m h , referring to the machine capability perceived by the human) to objective machine capability (denoted by C m m , referring to the machine’s actual capability at a given time). This relationship is formally expressed in (4):
T m h ( k ) = C m h ( k ) C m m ( k ) , k 0 T m h ( t ) = T m h ( k ) , t [ t k , t k + 1 ) , k 0
where t k is the time when the human updates their subjective machine capability for the k-th time ( t 0 : = 0 is the starting time of the considered human–machine system).
A dynamic model for MPD human trust was built, as detailed in (5):
T m h ( k ) = ( 1 α h * ( k ) ) T m h ( k 1 ) + α h * ( k ) α h P P k , k > 0 T m h ( t ) = T m h ( k ) , t [ t k , t k + 1 ) , k > 0 T m h ( 0 ) = α h C
where P k is defined in (6):
P k : = P m m ( Δ k ) C m m ( k 1 )
and α h * ( k ) is given in (7):
α h * ( k ) = α h + , Δ * P m h ( k ) > 0 ; 0 , Δ * P m h ( k ) = 0 ; α h , Δ * P m h ( k ) < 0 .
with Δ * P m h ( k ) specified in (8):
Δ * P m h ( k ) = α h P P m m ( Δ k ) T m h ( k 1 ) C m m ( k 1 )
The four human factors—the static human factor via informed capability α h C , the static human factor via observed performance α h P , conservativeness and risk-awareness human factors α h + and α h , and the actual machine performance P m m ( Δ k ) during time interval [ t k 1 , t k ) —can be estimated using experimental data.

4. MPD Trust Model for Assistive Aerial Teleoperation

This section specifies the general MPD trust model, presented in (5), for the context of assistive aerial teleoperation. The initial step is to determine the machine’s performance P m m ( Δ k ) and its objective capability C m m ( k ) . While C m m ( k ) is often assumed to be a constant value C m m for a given task, as expressed in (9),
C m m ( k ) = C m m , k 0
this assumption may not hold in practice. Especially for UAVs operating in unknown environments, the machine’s capability can vary due to dynamic environmental factors. Therefore, this work focuses on quantifying both P m m ( Δ k ) and the context-dependent C m m ( k ) to formulate a specialized trust model for assistive aerial teleoperation.

4.1. Quantifying Machine Performance P m m ( Δ k ) and Objective Machine Capability C m m ( k )

In practice, the objective machine capability C m m ( k ) is an idealized benchmark derived from data on actual machine performance collected over an extended period. Consequently, both the machine performance P m m ( Δ k ) and the objective machine capability C m m ( k ) originate from the same set of evaluation metrics. Their primary distinction lies in the temporal scope of the assessment: P m m ( Δ k ) is evaluated over a discrete time interval [ t k 1 , t k ) , whereas C m m ( k ) reflects the aggregated performance across a long-term horizon.
We first define the evaluation metrics for machine performance and objective machine capability in the context of assistive aerial teleoperation. We then present the specific algorithms used to compute P m m ( Δ k ) and C m m ( k ) .

4.1.1. UAV Trajectory Performance Metrics

To ensure compliance with both flight safety and operator visibility constraints, this work proposes a multi-dimensional metric that incorporates a safety factor and a visibility factor.
(a)
Safety
     Safety is defined as a quantitative index of collision risk. This index reflects the principle that the collision risk increases as the radial distance to obstacles Δ d ( k ) decreases and the velocity component towards them v d ( k ) increases. Specifically, suppose that at time k, the position vector from the UAV to the nearest obstacle is d ( k ) = [ d x ( k ) , d y ( k ) , d z ( k ) ] . The corresponding minimum radial distance Δ d ( k ) is modeled as the Euclidean norm of this position vector, as given in (10):
Δ d ( k ) = d ( k ) = d x 2 ( k ) + d y 2 ( k ) + d z 2 ( k )
     Meanwhile, the flight velocity vector is v ( k ) = [ v x ( k ) , v y ( k ) , v z ( k ) ] . The velocity component v d ( k ) in the obstacle direction is calculated by vector projection, as expressed in (11):
v d ( k ) = v ( k ) · d ( k ) Δ d ( k )
where v ( k ) · d ( k ) denotes the dot product of the velocity vector and the obstacle direction vector. To characterize the relative motion risk, we employ a kinematic risk indicator v d 2 ( k ) 2 Δ d ( k ) , which represents the constant deceleration required to avoid a collision [14]. The safety factor S ( k ) is then formulated as given in (12):
S ( k ) = e γ s v d 2 ( k ) 2 Δ d ( k )
where γ s R + is a tunable coefficient that regulates risk sensitivity.
(b)
Visibility
     Visibility is a metric designed to quantify the level of occlusion within the UAV’s field of view (FOV) [25]. When the occluded area within the FOV exceeds a threshold, the environmental information available to the operator will significantly diminish, thereby affecting decision-making efficiency. As shown in Figure 1, the UAV trajectory point p ( k ) = [ p x ( k ) , p y ( k ) , p z ( k ) ] , and the target position c ( k ) = [ c x ( k ) , c y ( k ) , c z ( k ) ] . It is assumed that the UAV always faces the target. The blue area in the figure is the UAV’s FOV, and the blue dashed area is defined as the confident FOV, i.e., the core observation area where no occlusion is required in the ideal state. Considering that the computational complexity of the analytical solution for the occluded area of obstacles in the FOV is extremely high, we approximate the visibility by constructing a series of spherical regions { B 1 ( k ) , B 2 ( k ) , , B N ( k ) } , where N is the total number of spherical regions. The center c i ( k ) and radius r i ( k ) of the spherical regions are defined as given in (13):
c i ( k ) = p ( k ) + ψ i ( c ( k ) p ( k ) ) r i ( k ) = ρ · ψ i · c ( k ) p ( k )
where ψ i = i N , i { 1 , 2 , , N } , and ρ is a parameter related to the size of the FOV. The visibility of the FOV is measured by comparing the minimum distance from the center point c i ( k ) to the nearest obstacle with the radius r i ( k ) of the corresponding sphere. The visibility of each area B i ( k ) is ensured by the following condition, as given in (14):
Ξ ( c i ( k ) ) r i ( k )
where Ξ ( c i ( k ) ) denotes the minimum distance from c i ( k ) to the nearest obstacle. Accordingly, the visibility factor F ( k ) at time k is defined as given in (15):
F ( k ) = 1 i = 1 N max ( 0 , r i 2 ( k ) Ξ 2 ( c i ( k ) ) ) i = 1 N r i 2 ( k )

4.1.2. Determining Machine Performance P m m ( Δ k )

The machine performance P m m ( Δ k ) is quantified by aggregating trajectory performance metrics within the time interval [ t k 1 , t k ) . Research in human–machine interaction suggests that trust formation is subject to two key temporal effects. First, recent performance immediately influences trust, an effect known as immediacy. Second, the cumulative effect of historical performance shapes long-term trust assessment [26]. This dual influence necessitates a modeling approach for machine performance that balances short-term responsiveness with long-term stability.
Specifically, within the interval [ t k 1 , t k ) , we define a sampling interval Δ t , where the interval length is Δ k = t k t k 1 = N Δ t , comprising N equally spaced sampling points. To incorporate observation data dynamically, we employ a sliding window spanning the most recent N sampling points, covering a duration of N Δ t . Within this window, the safety indicator P S Δ k and visibility indicator P F Δ k of the UAV trajectory are computed using an Exponentially Weighted Moving Average (EWMA) to emphasize recent performance, as follows in (16):
P S Δ k = i = 1 N α i 1 S ( t k i Δ t ) i = 1 N α i 1 P F Δ k = i = 1 N α i 1 F ( t k i Δ t ) i = 1 N α i 1
where t k i Δ t represents the i-th sampling point backward from the current time, and α ( 0 , 1 ) is a decay factor that adjusts the degree of influence of historical data. A smaller value of α places greater emphasis on more recent data. The machine performance P m m ( Δ k ) is then defined as given in (17):
P m m ( Δ k ) = ω S P S Δ k + ω F P F Δ k
where ω S and ω F [ 0 , 1 ] are weighting coefficients for safety and visibility, respectively, satisfying ω S + ω F = 1 . The value P m m ( Δ k ) [ 0 , 1 ] quantifies the overall trajectory performance within [ t k 1 , t k ) .

4.1.3. Determining Objective Machine Capacity C m m ( k )

Given the dynamic and uncertain environments that UAVs encounter during missions, their objective machine capability C m m ( k ) may vary due to environmental factors. To calculate C m m ( k ) , we integrate prior knowledge from the system’s predefined capability model with posterior knowledge derived from real-time environmental feedback. The machine capability at time step k is thus expressed as given in (18):
C m m ( k ) = ( 1 α m ) C m m ( k 1 ) + α m P m m ( Δ k ) , k > 0 C m m ( 0 ) = C m m
where P m m ( Δ k ) represents the actual machine performance over [ t k 1 , t k ) , and α m ( 0 , 1 ) is a dynamic factor reflecting the contribution of recent performance to the capability assessment.

4.2. The MPD Human Trust Model for Assistive Aerial Teleoperation

The MPD human trust model for assistive aerial teleoperation can now be formulated as follows in (19):
T U h ( k ) = ( 1 α h * ( k ) ) T U h ( k 1 ) + α h * ( k ) α h P P k , k > 0 T U h ( t ) = T U h ( k ) , t [ t k , t k + 1 ) , k > 0 T U h ( 0 ) = α h C
where P k , α h * ( k ) , and Δ * P m h ( k ) can be similarly defined as in (6), (7), and (8), respectively. In particular, P m m ( Δ k ) and C m m ( k ) can be calculated as in (17) and (18), respectively.
Regarding the human factors from (5), the initial trust value, α h C , must be determined before the system operates. In this work, we initialize this value using a pre-task trust scale. The three human-factor parameters, α h P , α h + , and α h , must be estimated from runtime data. Various methods are suitable for this purpose, such as the Least Squares Estimation approach.

5. Trust-Aware Human–UAV Collaborative Trajectory Planning

This section presents a trust-aware method for human–UAV collaborative trajectory planning. First, a trajectory generation method is adopted to generate a candidate trajectory set. Subsequently, the optimal trajectory is selected from the candidate set by minimizing a cost function that incorporates the discrete Fréchet distances to the current trajectory and the human input primitive, as well as the trust value associated with each trajectory. Finally, the selected trajectories are reparameterized and optimized in conjunction with human trust levels using MINCO’s representation. The specific process is detailed in Algorithm 1 and an overview of the entire framework is shown in Figure 2.
Algorithm 1 Trust-aware human–UAV collaborative trajectory planning.
Require: Initial state x c , User input u h , Motion primitive tree Γ , Sample set S , Human trust T U h
Ensure: Human ideal trajectory ξ *
 1:
Initialization:  S { η c } , Γ
 2:
while terminating condition not met do
 3:
    for each sampled node η n from S  do
 4:
        Add the sampled node η n to the tree Γ
 5:
        Remove the sampled node from S
 6:
        Generate the primitive γ u n and node set C n
 7:
        if  γ u n A safe  then
 8:
           Compute cost f c n for each node in C n using Equation (22)
 9:
           Add the lowest- f c n node from C n to S
10:
        end if
11:
    end for
12:
end while
13:
Convert the tree Γ into a path set { ξ i } i = 1 M
14:
Compute the Discrete Fréchet Distance and average trust for each ξ
15:
Select the optimal path ξ * from the path set { ξ i } i = 1 M
16:
Adjust the trajectory weights λ according to the human trust T U h
17:
Optimize the trajectory of ξ *
18:
return  ξ *

5.1. Trajectory Generation Guided by Human Intentions

5.1.1. Motion Primitive Tree Generation

Existing studies have demonstrated that quadrotors are differentially flat systems [27]; thus, we can compute precise inputs to make the vehicle follow a specified trajectory in terms of the flat outputs x, y, z, and yaw angle θ . We define the state as x = [ x , y , z , θ ] X , where the state space X R 4 . The operator provides inputs u h = [ v x , ω , v z ] with respect to a level frame C , which is a world-z-aligned frame rigidly attached to the vehicle’s origin.
Inspired by [5], we generate motion primitives as polynomial trajectories. This method constrains the trajectory by the vehicle’s current state and a target velocity profile guided by the operator’s input. The state of the motion primitive is set to evolve according to the unicycle model [28], as expressed in (20):
f ˙ u h ( t ) = [ v x cos ( ω t ) , v x sin ( ω t ) , v z , ω ]
The motion primitive, denoted as γ u h ( t ) , is an eighth-order polynomial trajectory in the world frame W of the form γ u h ( t ) = i = 0 8 c i t i . Its coefficients are uniquely determined. Given the current state x c and its higher-order derivatives (up to the fourth order), along with a duration T h , the operator’s input primitive γ u h ( t ) for t [ 0 , T h ] is generated by enforcing the following boundary conditions, as given in (21):
γ ( j ) ( 0 ) = x c ( j ) for j = 0 , 1 , 2 , 3 , 4 γ ˙ ( T h ) = R t C W f ˙ u h ( T h ) γ ( j ) ( T h ) = 0 for j = 2 , 3 , 4
where R t C W is the transformation matrix from the level frame C to the world frame W at time t. The primary role of this guiding primitive γ u h ( t ) is to direct the search process, ensuring the generated trajectories align with the user’s desired direction.
For generating the tree, we fix the forward velocity v x to the value provided by the human operator’s input u h . We then perform discrete sampling on the remaining action parameters: angular velocity ω , vertical velocity v z , and duration T. By iteratively sampling the action space A n , we propagate from node η n to its child node set C n . We generate a corresponding candidate primitive γ u n ( t ) by applying the exact same polynomial generation process defined by Equations (20) and (21). We then check if it lies within the safe space X safe to ensure all generated primitives are collision-free. This acts as a hard constraint, immediately discarding any inherently unsafe paths and guaranteeing that only collision-free trajectories are considered for the subsequent stages.
To evaluate the directional similarity between a candidate primitive and the human-guided primitive, we compute a cost based on their direction vectors. We define a mapping p ( · ) that extracts the spatial coordinates from a full state vector. The cost f c n is then given by (22):
f c n = 1 d n · d h d n = p ( γ u n ( T n ) ) p ( x n ) p ( γ u n ( T n ) ) p ( x n ) , d h = p ( γ u h ( T h ) ) p ( x c ) p ( γ u h ( T h ) ) p ( x c )
where T n is the duration of the motion primitive corresponding to node η n , x n is the initial state corresponding to node η n , and x c is the current state. We sort the costs f c n and select the node with the minimum cost from the candidate node set C n to add to the sampling set S , which is used for the next round of motion primitive expansion.
When x n x c exceeds the range of the local perception, or the number of nodes reaches a preset threshold, the motion primitive tree ceases to expand. By connecting nodes at different depths, a series of trajectories { ξ i } i = 1 M are obtained.

5.1.2. Trajectory Selection

Given a set of candidate trajectories { ξ i } i = 1 M , we select the human-desired trajectory ξ * from them. We need to balance three objectives: minimizing the deviation from the current trajectory to ensure a smooth transition, minimizing the deviation from the human input primitives [3], and maximizing the average trust value associated with each trajectory. To this end, we introduce a selection cost function, which evaluates the proximity of each trajectory to the current trajectory ξ L and the operator input primitive γ u h by calculating the discrete Fréchet distance δ d F (see Figure 3) [29] while also considering the trust value T ( ξ ) of each trajectory. A brief definition of the discrete Fréchet distance is given as follows.
Consider a discrete sampling of two continuous functions f and g, forming sequences P = f 1 , f 2 , , f n and Q = g 1 , g 2 , , g m , consisting of n and m discrete points, respectively. An order-preserving complete correspondence between P and Q is defined by a pair ( α , β ) of discrete monotone reparameterizations [5], where α maps 1 , , k to 1 , , n and β maps 1 , , k to 1 , , m . The discrete Fréchet distance of P and Q is given by (23):
δ d F ( f , g ) : = min ( α , β ) max i [ 1 , k ] d ( f α ( i ) , g β ( i ) )
where ( α , β ) range over all order-preserving complete correspondences between P and Q . Thus, for time-parameterized trajectories Γ and Φ , their Fréchet distance is defined as given in (24):
δ d F ( Γ , Φ ) : = min ( α , β ) max i [ 1 , k ] d ( Γ α ( i ) , Φ β ( i ) )
where Γ i = Γ ( i · Δ t ) denotes sampling Γ at a fixed time interval Δ t .
Based on the above definitions, we derive the trajectory selection mechanism as follows: Given the current trajectory ξ L , the operator input primitive γ u h , and the average trust value T ( ξ ) for each candidate trajectory, the desired trajectory is determined by (25):
ξ * = agr min ξ { ξ i } i = 1 M w L δ d F ( ξ , ξ L ) + w U δ d F ( ξ , γ u h ) w T T ( ξ )
where w L weights the smoothness with the current trajectory, w U weights the deviation from the human input primitive, and w T weights the trust value.

5.2. Optimization Problem Formulation

Selected from the motion primitive tree is the desired human trajectory ξ * . In this section, a trajectory representation method called MINCO [30] is adopted to reparameterize the trajectory. MINCO performs spatiotemporal decoupling transformation on the M-piece trajectory ξ ( t ) with flat output through a linear complexity mapping, thereby decoupling the spatial parameters from the temporal parameters. Specifically, the s-order MINCO trajectory is actually a 2 s -order polynomial spline with constant boundary conditions, which is defined as given in (26):
T MINCO = { ξ ( t ) : [ 0 , T ] R m | c = M ( q , T ) , q R m ( M 1 ) , T R > 0 M }
where ξ ( t ) is an m-dimensional M-piece polynomial trajectory. c = ( c 1 T , , c M T ) T R 2 M s × m is the polynomial coefficient, and M ( q , T ) is a linear-complexity smooth map from intermediate points q and a time allocation T for all pieces to the coefficients of splines. The i-th piece of the trajectory, ξ i ( t ) , is defined as given in (27):
ξ ( t ) = ξ i ( t t i 1 ) , t [ t i 1 , t i ] ξ i ( t ) = c i T β ( t ) , t [ 0 , T i ]
where β ( t ) : = [ 1 , t , , t N ] T is the natural basis, c i is the coefficient matrix, T i = t i t i 1 , and T = i = 1 M T i .
Define the constraint points ξ ˚ i , j . The sampling manner on each piece of the trajectory is given by ξ ˚ i , j = ξ i j κ i T i , where j 0 , 1 , 2 , , κ i and κ i denotes the number of sampling points in the i-th piece. These points serve to convert the time integrals of certain constraint functions into a weighted sum of sampling penalty functions. Considering our requirements, all constraints are converted into penalty terms, and the trajectory optimization problem is formulated as given in (28):
min q , T J t , J o , J e , J d · λ
where J t is the execution time penalty, J o is the safety penalty, J e is the control effort penalty, and J d is the dynamic feasibility penalty. λ = λ t , λ o , λ e , λ d T is the weight vector used to balance these respective penalties.

5.2.1. Execution Time Penalty J t

To enhance the flight speed of the UAV, we impose a penalty term on the total execution time J t = i = 1 M T i . Its gradients are given by J t / c = 0 and J t / T = 1 .

5.2.2. Safety Penalty J o

Inspired by [31], the safety penalty term J o for obstacle avoidance is computed using the Euclidean Signed Distance Field (ESDF). We penalize sampling points that are too close to obstacles, as expressed in (29):
ψ o ( ξ ˚ i , j ) = Ξ t h r Ξ ( ξ ˚ i , j )
where Ξ t h r is the safety distance threshold set according to practical scenarios, and Ξ ( ξ ˚ i , j ) represents the distance from ξ ˚ i , j to the nearest obstacle. The safety penalty term is then derived by computing the weighted sum of the sampled constraint functions, specifically as given in (30):
J o = T i κ i j = 0 κ i ω ¯ j max { ψ o ( ξ ˚ i , j ) , 0 } 3
where ( ω ¯ 0 , ω ¯ 1 , , ω ¯ κ i 1 , ω ¯ κ i ) = ( 1 / 2 , 1 , , 1 , 1 / 2 ) are the quadrature coefficients following the trapezoidal rule [32]. The detailed derivation of the gradient of J o is as given in (31):
J o c i = J o ψ o ψ o c i J o T i = J o T i + J o ψ o ψ o t t T i t T i = j κ i , t = j κ i T i
where t is the relative time on this piece. For the case where Ξ ( ξ ˚ i , j ) < Ξ t h r , the gradient is given as expressed in (32):
ψ o c i = β ( t ) d T , ψ o t = d T ξ ˙ ( t )
where d is the gradient of the ESDF at ξ ˚ i , j . Otherwise, the gradients become ψ o / c i = 0 and ψ o / t = 0 .

5.2.3. Control Effort Penalty J e

According to [33], the s-order control input and its gradient for the i-th piece of the trajectory can be expressed as given in (33):
J e = i = 1 M 0 T i ξ i ( s ) ( t ) 2 d t J e c i = 2 0 T i β ( s ) ( t ) β ( s ) ( t ) T d t c i J e T i = c i T β ( s ) ( T i ) β ( s ) ( T i ) T c i

5.2.4. Dynamical Feasibility Penalty J d

When the trajectory of the UAV approaches physical constraints, the dynamic feasibility penalty mechanism is triggered. To ensure that the UAV can execute the trajectory, we impose limits on the maximum values of velocity, acceleration, and jerk, as expressed in (34):
G * = ξ i ( n ) ( t ) 2 m a x * 2 , t [ 0 , T i ]
where * = { v , a , j } , n = { 1 , 2 , 3 } , and t = j T i / κ i denote velocity, acceleration, jerk, coordinate axis indices, and sampling time, respectively. V max , A max , and J max are the maximum allowable velocity, acceleration, and jerk. The dynamic feasibility penalty [6] term is obtained by calculating the weighted sum of sampled constraint functions, as given in (35):
( J d ) * = i = 1 M T i κ i j = 0 κ i ω ¯ j max G * ( ξ ˚ i , j ) , 0 3
where ( ω ¯ 0 , ω ¯ 1 , , ω ¯ κ i 1 , ω ¯ κ i ) = ( 1 / 2 , 1 , , 1 , 1 / 2 ) . The gradient can be expressed as given in (36):
( J d ) * c i = ( J d ) * G G c i ( J d ) * T i = ( J d ) * G G t t T i + j = 0 κ i ω ¯ j κ i max ( G * ( ξ ˚ i , j ) , 0 ) 3 G * c i = 2 β ( n ) ( t ) ξ i ( n ) ( t ) T , G * t = 2 β ( n + 1 ) ( t ) T c i ξ i ( n ) ( t )
In practical implementation, to reserve some power resources and ensure the feasibility of the trajectory, we take 90% of the maximum allowable velocity, acceleration, and jerk as the thresholds.

5.3. Trajectory Weight Adjustment Based on Human Trust

Considering the changes in human trust in the UAV, the optimized trajectory should balance search efficiency and safety, and λ t and λ o are further optimized as given in (37):
λ t = λ t , 0 + Δ λ t λ o = λ o , 0 + Δ λ o
Based on the historical trajectory of the UAV and the MPD human trust model, we can calculate the human trust level T U h . We categorize the human trust level T U h into three ranges from low to high: [ 0 , T L ) , [ T L , T H ) , and [ T H , ) . Here, T L and T H represent predefined trust thresholds. When the operator’s trust level in the UAV is low ( T U h [ 0 , T L ) ), the operator may compete with the UAV for control, posing a significant safety risk. In this scenario, Δ λ t is assigned a smaller value, while Δ λ o is assigned a larger value to enhance the safety of the UAV during human intervention. As the operator’s trust level increases ( T U h [ T L , T H ) ), Δ λ t is gradually increased, and Δ λ o is gradually decreased to achieve a balance between trajectory rapidity and safety. Furthermore, when the operator’s trust level is high ( T U h [ T H , ) ), it reflects strong confidence in the UAV’s capabilities. Consequently, Δ λ t is set to a larger value, and Δ λ o is set to a smaller value to optimize search efficiency and reduce the operator’s long-term workload. To accommodate the requirements of these distinct trust intervals, the specific expressions for the efficiency weight adjustment Δ λ t ( T ) and the safety weight adjustment Δ λ o ( T ) , where T denotes the human trust level T U h , are given by (38) and illustrated in Figure 4:
Δ λ t ( T ) = Λ t 1 + e k 1 ( T T H ) Δ λ o ( T ) = Λ o 1 + e k 2 ( T T L )
The key parameters are set based on a combination of pilot studies and design principles. The trust thresholds are set to T L = 0.6 and T H = 0.9 based on pilot studies. The adjustment bounds follow a safety-first principle where Λ o > Λ t , ensuring safety concerns elicit a stronger system response than efficiency gains. Finally, the transition slopes ( k 1 , k 2 ) were tuned through iterative user feedback to achieve a behavioral transition that operators perceived as smooth and predictable.

6. Experiments and Results

In this section, we design and conduct an experiment on human–UAV collaboration to compare the effectiveness of our proposed method against a baseline approach that does not account for human trust. The experiments are performed on a high-performance desktop equipped with an Intel Core i9-14900KF CPU (32 cores, base frequency 3.2 GHz, maximum turbo frequency 6.0 GHz) and an NVIDIA GeForce RTX 4090 GPU (24 GB VRAM).

6.1. Implementation Details

To evaluate the proposed method, we conducted simulation experiments in three simulated random forest environments with varying tree densities. The simulation area was a rectangular volume measuring 70 m in length, 20 m in width, and 3 m in height, containing 50, 100, or 200 trees to represent sparse, medium, and dense scenarios, respectively. For the experiments, 15 students (ages 22–28) with no prior drone piloting experience were recruited. After a 5 min familiarization session, each operator was tasked with navigating a simulated quadrotor vehicle from one end of the forest to the other end. The experiment followed a within-subjects design, where each participant tested both our proposed method and a baseline approach. The order of the two methods was counterbalanced across participants. For each method, the operator performed the task once in each of the three random forest environments. The task was deemed complete upon reaching the opposite side. Operators were provided with a first-person view of the vehicle and sent commands via a joystick to guide its flight. They were permitted to intervene in the vehicle’s flight at any time.
We evaluate two aspects of the performance of our assistive aerial teleoperation system: (1) navigation efficiency, which mainly includes the distance, duration, and average velocity during the mission; and (2) the cost of human–UAV interaction, for which the core evaluation metrics are the number of operator inputs, jerk integral, and average trust. Specifically, the number of operator inputs indicates the intervention frequency of the assistive system, while the jerk integral reflects the piloting burden and trajectory smoothness. If the former two interaction indicators (operator inputs and jerk integral) are high, it indicates that the path searched by the assistive system does not match the human’s true intention, leading to poor flight quality and requiring more operator interaction to correct, which is often accompanied by lower trust. In contrast, if these indicators are low, it indicates that the operator agrees with the system’s path planning, which implies fluid human–machine collaboration and a smooth flight trajectory, ultimately building and maintaining a higher level of trust.
The motion primitive parameters used in the experiment are shown in Table 1. The maximum linear velocity input by the operator is set to 2 m/s. The parameters of the visibility metric are N = 20 , ρ = 0.8 . The parameters of the safety metric are γ s = 2 . The specific parameters for the MPD trust model are detailed in Table 2. In our implementation, these parameters are empirically determined by referencing previous work and our own experiments, and their systematic optimization is beyond the scope of this paper.

6.2. Results Analysis

The quantitative results of our experiments are summarized in Table 3, which compares the performance of our proposed trust-aware approach against the trust-unaware baseline across four key dimensions: task completion efficiency, operator workload, trajectory quality, and human–machine trust. These experiments were conducted in three distinct environments to assess the robustness of both methods under varying levels of complexity. Trajectory comparisons are visually illustrated in Figure 5 (2D) and Figure 6 (3D), offering additional insights into the behavioral differences between the two approaches.
In terms of fundamental task completion metrics, both methods exhibit comparable performance across all environments. Specifically, flight distance and duration show no significant differences between the trust-aware and trust-unaware approaches. For instance, in the dense environment, the trust-aware method recorded a flight distance of 71.38 m and a duration of 49.86 s, compared to 71.73 m and 54.89 s for the trust-unaware method. Similar trends are observed in the sparse and medium environments, indicating that incorporating the trust model does not compromise basic task execution efficiency. Both methods guide the UAV to plan effective paths of similar lengths, where the paths for both methods closely align in terms of overall distance across sparse, medium, and dense settings. However, the trust-aware method demonstrates a slight advantage in average velocity, particularly in the dense environment. This improvement is primarily attributed to smoother flight trajectories and fewer unnecessary pauses. The trust-aware method produces paths with fewer sharp turns and abrupt changes compared to the trust-unaware baseline, especially in the dense environment, enhancing the overall efficiency of motion.
The number of operator inputs serves as a core metric for evaluating operator intervention frequency and effort, directly reflecting the workload imposed on the operator. As shown in Table 3, the trust-aware method significantly reduces the number of operator inputs compared to the trust-unaware method across all scenarios. In the sparse environment, inputs decreased from 39 to 34 (a reduction of approximately 12.8%), and this gap widened in the dense environment, where inputs dropped from 56 to 43 (a reduction of approximately 23.2%). These results strongly demonstrate that our trust-aware method effectively alleviates the operator’s workload. The trust-aware system’s ability to dynamically adjust its assistance strategy based on real-time trust assessments underlies this improvement. When trust is high, the system operates more autonomously, reducing the need for operator fine-tuning. Conversely, when trust declines, the system adopts conservative behaviors that align better with the operator’s expectations, minimizing interventions caused by mismatches between system decisions and human intent.
The jerk integral metric reflects the smoothness of the UAV’s flight, with lower values indicating smoother trajectories with fewer abrupt changes. The experimental results in Table 3 reveal that the trajectory quality of the trust-aware method consistently surpasses that of the trust-unaware baseline. In the sparse environment, the jerk integral was reduced by approximately 22.5%, and this improvement was even more pronounced in the dense environment, with a reduction of approximately 43.2%. These smoother trajectories imply safer, more energy-efficient flights and provide a more comfortable first-person view experience for the operator. The reduction in the jerk integral is closely tied to the decrease in operator inputs; fewer external interventions prevent frequent, abrupt changes in flight state. Furthermore, the trust-aware model’s ability to predict and align with the operator’s long-term intent results in more proactive and coherent paths, contrasting with the passive, jerky reactions of the trust-unaware baseline, particularly evident in the dense environment.
Finally, the average trust metric directly validates the effectiveness of our trust-aware approach. In all environments, the average trust level of operators was significantly higher with the trust-aware method compared to the trust-unaware baseline. For example, in the medium-density environment, the trust-aware method achieved an average trust of 0.750 compared to 0.657 for the trust-unaware method. Trust levels for both methods decreased as environmental complexity increased, which is expected due to the greater challenges in denser settings. However, the decline was gentler for the trust-aware method (e.g., from 0.813 in the sparse environment to 0.693 in the dense environment compared to a steeper drop from 0.744 to 0.611 for the trust-unaware method). This suggests that by actively modeling and adapting to the operator’s trust state, our system establishes and maintains a more robust and resilient human–machine trust relationship. The higher trust levels both result from and contribute to the benefits observed (fewer inputs, smoother trajectories). As the system’s behavior becomes more reliable and predictable, operators are more willing to reduce interventions, creating a virtuous cycle.

7. Discussion

7.1. Discussion of Findings

The empirical findings provide compelling evidence that our trust-aware framework substantially elevates the quality of human–UAV interactions, as evidenced by reductions in operator workload, smoother trajectories, and more resilient trust dynamics—all without compromising core task metrics such as flight efficiency. This aligns with the framework’s core innovation: integrating an MPD trust model into the real-time planner, which dynamically predicts and adapts to operator intent and trust states. Unlike traditional trust-unaware baselines that react passively to inputs, our approach proactively aligns system behavior with human expectations and fosters more intuitive collaboration. These improvements extend beyond mere performance gains, underscoring the framework’s value in making human–machine systems safer and more user-centric, particularly in variable environments where baseline methods falter.
The core of this success lies in the formation of a virtuous cycle in operator–UAV interaction, a phenomenon observed across different environments. Specifically, this cycle occurs when the system’s trust-aware behavior aligns with operator expectations, which reduces unnecessary interventions and in turn reinforces the operator’s trust, fostering smoother and more effective collaboration. This cycle not only validates the important role of trust in human–machine interaction (e.g., [13]) but also provides a crucial extension to previous studies focused on intent-aligned trajectory generation [3,5]. While these prior works demonstrated that aligning with operator intent improves collaboration, our findings reveal that intent-alignment alone is insufficient under trust-eroding conditions. Specifically, our results in dense environments show that without a trust model, operator interventions remained high despite intent alignment, whereas our trust-aware approach successfully mitigated conflicts by dynamically increasing system caution. Our work thus extends these intent-based models by incorporating a vital second feedback loop, the operator’s dynamic trust state, which proves essential for robust performance in complex scenarios.
Furthermore, our framework makes a tangible contribution to the formal practice of risk analysis in human–UAV interaction. As conceptualized in frameworks like that of Sanz et al. [34], effective risk management for UAVs involves a cycle of hazard identification, assessment, and reduction. Our work provides a concrete, real-time implementation of this cycle. Specifically, the MPD trust model functions as a dynamic risk assessment tool, quantifying the risk of human–machine goal conflicts based on system performance and the operator’s trust level. The adaptive trajectory planner then serves as the risk reduction mechanism, translating this assessed risk into tangible actions by modulating the safety and efficiency weights. By formalizing the link between trust dynamics and system safety behaviors, our approach offers a more structured method for managing risk in collaborative aerial tasks, moving beyond the static safety rules common in prior systems.
However, while these findings affirm the framework’s efficacy in simulated scenarios, they also reveal critical limitations that warrant careful consideration, primarily stemming from the inherent sim-to-real gap. It must be acknowledged that our controlled simulation cannot fully replicate the complex uncertainties of the physical world. These include unmodeled aerodynamics, complex sensor noise profiles, and unpredictable communication latencies. Crucially, such factors could directly corrupt the real-time safety and visibility metrics that are fundamental inputs to our MPD trust model. Furthermore, beyond these physical uncertainties, our study utilized a limited operator cohort, and the framework’s generalizability across operators with varying expertise levels and risk tolerances remains an underexplored but vital area. These limitations, therefore, highlight critical open research issues regarding the framework’s real-world robustness and its capacity for personalization, paving the way for future enhancements.

7.2. Practical Significance and Scenario-Based Illustration

To illuminate the practical significance of our trust-aware framework, we present an illustrative scenario in a high-stakes application: post-disaster search and rescue.
Imagine a rescue mission where an operator must remotely pilot a UAV through a collapsed building to locate survivors. When using a UAV equipped only with basic obstacle avoidance, the operator faces immense cognitive load, simultaneously interpreting video feeds, navigating complex 3D spaces, and fighting against a system that may refuse to approach cluttered areas where a survivor might be. This often leads to human–machine conflict; to complete the mission, the operator might disable the assistance system, risking a catastrophic crash due to a moment of inattention. The loss of an expensive UAV not only incurs a financial cost but, more critically, could prematurely end a life-saving search.
Now, let us envision the same mission executed with our trust-aware framework. Its practical value unfolds in several stages:
  • Initial trust calibration: As the UAV enters the building, it first navigates through less cluttered corridors. Its smooth, predictable trajectories, reflecting high safety and visibility, rapidly build the operator’s trust in the system’s competence.
  • Intent-driven investigation: The operator spots a potential sign of life and issues a command to move closer for a better view. Because the operator’s trust is high, our framework accurately infers a strong, deliberate intent. It then generates a precise trajectory that navigates assertively yet safely around debris, fulfilling the operator’s goal without resistance. This is a stark contrast to a trust-unaware system that might have simply halted.
  • Adaptive safety preservation: As the UAV ventures deeper, it enters an area with poor visibility, causing its onboard sensors to become less reliable. Our framework’s machine performance metrics objectively detect this degradation. Consequently, the trust model dynamically lowers its trust value, shifting the system into a more cautious state. If the anxious operator now commands a rapid forward movement, the system will provide stronger assistance, moderating the speed and trajectory to prioritize the UAV’s survival. It intelligently prevents a trust-induced error, safeguarding the mission’s most critical asset.
In summary, this scenario demonstrates clear practical significance. By fostering a more intuitive human–machine collaboration, our framework simultaneously enhances mission effectiveness and reduces operator cognitive load. Crucially, its ability to dynamically adapt to risk ensures the safety of critical assets, ultimately empowering human–machine teams to confidently undertake complex, high-stakes operations that were previously too hazardous or inefficient, thereby significantly expanding their operational boundaries.

7.3. Open Issues and Future Work

Future work will address several open issues by focusing on three core directions to further enhance the trust-aware framework’s generalization, accuracy, and application scope.
First, in human–machine interaction, a central open issue is how to bridge the sim-to-real gap while enabling genuine user adaptivity. Existing studies have primarily validated trust-aware frameworks in simulation [17,35], and their robustness against real-world uncertainties, such as sensor noise and communication latency, remains insufficiently tested. At the same time, current generic trust models often ignore differences in operator expertise, limiting their ability to meet personalized needs. To address this, our future research will encompass the following: (i) transition the framework from simulated to physical environments to verify robustness in practice; and (ii) develop personalized trust models that identify and adapt to the trust characteristics of operators with varying experience levels, thereby creating a system that is both robust and user-adaptive.
Second, in trust modeling, another open issue is how to reduce the over-reliance on empirical, manually tuned parameters. Existing approaches largely depend on human expertise [14,17] for parameter settings, which restricts scalability and adaptability. To overcome this, we will explore data-driven paradigms such as Inverse Reinforcement Learning (IRL) [13], enabling the model to autonomously learn and optimize intrinsic parameters from interaction data, replacing manual tuning with a more robust self-learning mechanism.
Finally, at the application level, an important open issue is how to scale trust-aware frameworks beyond constrained laboratory tasks into complex, high-stakes real-world scenarios. Prior work has mostly focused on basic navigation [3,5,7] in controlled settings, whereas missions such as large-scale inspection and search and rescue require balancing multiple performance metrics—including speed, safety, and coverage—while maintaining stability in highly dynamic environments that include challenges such as dynamic obstacles. To meet this challenge, we will extend the framework toward these high-stakes missions, using them as testbeds to rigorously evaluate and strengthen the adaptive capabilities of the trust-aware planner under realistic conditions.

8. Conclusions

This work proposes and validates a trust-aware trajectory planning framework to address the critical issue of human–machine trust imbalance in assistive aerial teleoperation. By embedding an MPD trust model into a real-time planner, the approach effectively reduces operator workload, improves flight safety, and establishes a more stable human–machine trust relationship, all while maintaining task efficiency. The results demonstrate that introducing dynamic trust quantification into the control loop of autonomous systems holds significant potential, enabling the creation of a virtuous cycle of reduced human intervention, greater system predictability, and enhanced collaborative performance. Taken together, this work establishes a foundation for developing more intuitive and efficient human–machine collaboration models, making an important contribution to the advancement of intelligent aerial systems. In the future, by addressing the open issues outlined in the Discussion section, this framework holds promise for extension to a broader range of autonomous system applications.

Author Contributions

Conceptualization, Q.Z.; methodology, K.H.; software, K.H.; validation, K.H. and X.J.; formal analysis, X.J.; investigation, Q.Z.; resources, Q.Z.; writing—original draft preparation, K.H.; writing—review and editing, P.L. and Y.Z.; visualization, K.H.; supervision, Y.K.; project administration, P.L. and Y.Z.; funding acquisition, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the State Key Laboratory of Avionics Integration and Aviation System-of-Systems Synthesis (grant number 2024AIASS0701) and supported by the Dreams Foundation of Jianghuai Advance Technology Center (grant number 2023-ZM01G003).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MPDMachine-Performance-Dependent
UAVUnmanned Aerial Vehicle
MPTMotion Primitive Tree
DFDDiscrete Fréchet Distance
IRLInverse Reinforcement Learning

References

  1. Perez-Grau, F.; Ragel, R.; Caballero, F.; Viguria, A.; Ollero, A. Semi-autonomous teleoperation of UAVs in search and rescue scenarios. In Proceedings of the 2017 International Conference on Unmanned Aircraft Systems (ICUAS), Miami, FL, USA, 13–16 June 2017; pp. 1066–1074. [Google Scholar] [CrossRef]
  2. Kanso, A.; Elhajj, I.H.; Shammas, E.; Asmar, D. Enhanced teleoperation of UAVs with haptic feedback. In Proceedings of the 2015 IEEE International Conference on Advanced Intelligent Mechatronics (AIM), Busan, Republic of Korea, 7–11 July 2015; pp. 305–310. [Google Scholar] [CrossRef]
  3. Shi, H.; Luo, L.; Gao, S.; Yu, Q.; Hu, S. Human-guided motion planner with perception awareness for assistive aerial teleoperation. Adv. Robot. 2024, 38, 152–167. [Google Scholar] [CrossRef]
  4. Boksem, M.A.; Meijman, T.F.; Lorist, M.M. Effects of mental fatigue on attention: An ERP study. Cogn. Brain Res. 2005, 25, 107–116. [Google Scholar] [CrossRef]
  5. Yang, X.; Cheng, J.; Michael, N. An Intention Guided Hierarchical Framework for Trajectory-based Teleoperation of Mobile Robots. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 482–488. [Google Scholar] [CrossRef]
  6. Wang, Q.; He, B.; Xun, Z.; Xu, C.; Gao, F. GPA-Teleoperation: Gaze Enhanced Perception-Aware Safe Assistive Aerial Teleoperation. IEEE Robot. Autom. Lett. 2022, 7, 5631–5638. [Google Scholar] [CrossRef]
  7. Yang, X.; Michael, N. Assisted Mobile Robot Teleoperation with Intent-aligned Trajectories via Biased Incremental Action Sampling. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 10998–11003. [Google Scholar] [CrossRef]
  8. Gebru, B.; Zeleke, L.; Blankson, D.; Nabil, M.; Nateghi, S.; Homaifar, A.; Tunstel, E. A Review on Human–Machine Trust Evaluation: Human-Centric and Machine-Centric Perspectives. IEEE Trans.-Hum.-Mach. Syst. 2022, 52, 952–962. [Google Scholar] [CrossRef]
  9. Alhaji, B.; Beecken, J.; Ehlers, R.; Gertheiss, J.; Merz, F.; Müller, J.P.; Prilla, M.; Rausch, A.; Reinhardt, A.; Reinhardt, D.; et al. Engineering Human–Machine Teams for Trusted Collaboration. Big Data Cogn. Comput. 2020, 4, 35. [Google Scholar] [CrossRef]
  10. Ibrahim, B.; Elhajj, I.H.; Asmar, D. 3D Autocomplete: Enhancing UAV Teleoperation with AI in the Loop. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 17829–17835. [Google Scholar] [CrossRef]
  11. Steyvers, M.; Tejeda, H.; Kerrigan, G.; Smyth, P. Bayesian modeling of human–AI complementarity. Proc. Natl. Acad. Sci. USA 2022, 119, e2111547119. [Google Scholar] [CrossRef]
  12. Lyons, J.B.; Stokes, C.K. Human–human reliance in the context of automation. Hum. Factors 2012, 54, 112–121. [Google Scholar] [CrossRef]
  13. Nam, C.; Walker, P.; Li, H.; Lewis, M.; Sycara, K. Models of Trust in Human Control of Swarms With Varied Levels of Autonomy. IEEE Trans.-Hum.-Mach. Syst. 2020, 50, 194–204. [Google Scholar] [CrossRef]
  14. Li, Y.; Cui, R.; Yan, W.; Zhang, S.; Yang, C. Reconciling Conflicting Intents: Bidirectional Trust-Based Variable Autonomy for Mobile Robots. IEEE Robot. Autom. Lett. 2024, 9, 5615–5622. [Google Scholar] [CrossRef]
  15. Saeidi, H.; Wang, Y. Incorporating Trust and Self-Confidence Analysis in the Guidance and Control of (Semi)Autonomous Mobile Robotic Systems. IEEE Robot. Autom. Lett. 2019, 4, 239–246. [Google Scholar] [CrossRef]
  16. Hu, C.; Shi, Y.; Ge, S.; Hu, H.; Zhao, J.; Zhang, X. Trust-Based Shared Control of Human-Vehicle System Using Model Free Adaptive Dynamic Programming. IEEE Trans. Intell. Veh. 2024, 1–13. [Google Scholar] [CrossRef]
  17. Fang, Z.; Wang, J.; Liang, J.; Yan, Y.; Pi, D.; Zhang, H.; Yin, G. Authority Allocation Strategy for Shared Steering Control Considering Human-Machine Mutual Trust Level. IEEE Trans. Intell. Veh. 2024, 9, 2002–2015. [Google Scholar] [CrossRef]
  18. Nieuwenhuisen, M.; Droeschel, D.; Schneider, J.; Holz, D.; Läbe, T.; Behnke, S. Multimodal obstacle detection and collision avoidance for micro aerial vehicles. In Proceedings of the 2013 European Conference on Mobile Robots, Barcelona, Spain, 25–27 September 2013; pp. 7–12. [Google Scholar] [CrossRef]
  19. Odelga, M.; Stegagno, P.; Bülthoff, H.H. Obstacle detection, tracking and avoidance for a teleoperated UAV. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 2984–2990. [Google Scholar] [CrossRef]
  20. Backman, K.; Kulić, D.; Chung, H. Learning to Assist Drone Landings. IEEE Robot. Autom. Lett. 2021, 6, 3192–3199. [Google Scholar] [CrossRef]
  21. Zhang, Q.; Xing, Y.; Wang, J.; Fang, Z.; Liu, Y.; Yin, G. Interaction-Aware and Driving Style-Aware Trajectory Prediction for Heterogeneous Vehicles in Mixed Traffic Environment. IEEE Trans. Intell. Transp. Syst. 2025, 26, 10710–10724. [Google Scholar] [CrossRef]
  22. Igneczi, G.F.; Horvath, E.; Toth, R.; Nyilas, K. Curve Trajectory Model for Human Preferred Path Planning of Automated Vehicles. Automot. Innov. 2024, 7, 59–70. [Google Scholar] [CrossRef]
  23. Huang, C.; Huang, H.; Hang, P.; Gao, H.; Wu, J.; Huang, Z.; Lv, C. Personalized Trajectory Planning and Control of Lane-Change Maneuvers for Autonomous Driving. IEEE Trans. Veh. Technol. 2021, 70, 5511–5523. [Google Scholar] [CrossRef]
  24. Yang, X.; Sreenath, K.; Michael, N. A framework for efficient teleoperation via online adaptation. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 5948–5953. [Google Scholar] [CrossRef]
  25. Wang, Q.; Gao, Y.; Ji, J.; Xu, C.; Gao, F. Visibility-aware Trajectory Optimization with Application to Aerial Tracking. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 5249–5256. [Google Scholar] [CrossRef]
  26. Lee, J.D.; See, K.A. Trust in Automation: Designing for Appropriate Reliance. Hum. Factors 2004, 46, 50–80. [Google Scholar] [CrossRef]
  27. Mellinger, D.; Kumar, V. Minimum snap trajectory generation and control for quadrotors. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 2520–2525. [Google Scholar] [CrossRef]
  28. Falanga, D.; Foehn, P.; Lu, P.; Scaramuzza, D. PAMPC: Perception-Aware Model Predictive Control for Quadrotors. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar] [CrossRef]
  29. Eiter, T.; Mannila, H. Computing Discrete Fréchet Distance; Technical Report CD-TR 94/64; Christian Doppler Laboratory: Vienna, Austria, 1994. [Google Scholar]
  30. Wang, Z.; Zhou, X.; Xu, C.; Gao, F. Geometrically Constrained Trajectory Optimization for Multicopters. IEEE Trans. Robot. 2022, 38, 3259–3278. [Google Scholar] [CrossRef]
  31. Zhou, B.; Gao, F.; Wang, L.; Liu, C.; Shen, S. Robust and Efficient Quadrotor Trajectory Generation for Fast Autonomous Flight. IEEE Robot. Autom. Lett. 2019, 4, 3529–3536. [Google Scholar] [CrossRef]
  32. Press, W.H.; William, H.; Teukolsky, S.A.; Saul, A.; Vetterling, W.T.; Flannery, B.P. Numerical Recipes 3rd Edition: The Art of Scientific Computing; Cambridge University Press: New York, NY, USA, 2007. [Google Scholar]
  33. Quan, L.; Yin, L.; Xu, C.; Gao, F. Distributed Swarm Trajectory Optimization for Formation Flight in Dense Environments. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 4979–4985. [Google Scholar] [CrossRef]
  34. Sanz, D.; Valente, J.; del Cerro, J.; Colorado, J.; Barrientos, A. Safe operation of mini UAVs: A review of regulation and best practices. Adv. Robot. 2015, 29, 1221–1233. [Google Scholar] [CrossRef]
  35. Broad, A.; Schultz, J.; Derry, M.; Murphey, T.; Argall, B. Trust Adaptation Leads to Lower Control Effort in Shared Control of Crane Automation. IEEE Robot. Autom. Lett. 2017, 2, 239–246. [Google Scholar] [CrossRef]
Figure 1. Illustration of the visibility metric definition.
Figure 1. Illustration of the visibility metric definition.
Machines 13 00876 g001
Figure 2. The framework diagram presents a human–UAV system for assistive aerial teleoperation that incorporates human trust into trajectory planning. The system uses human input to guide trajectory planning and generate an initial trajectory while simultaneously modeling human trust via the MPD trust model. Subsequently, the system optimizes the initial trajectory to match the dynamic trust level, ultimately forming an optimized path for UAV trajectory tracking.
Figure 2. The framework diagram presents a human–UAV system for assistive aerial teleoperation that incorporates human trust into trajectory planning. The system uses human input to guide trajectory planning and generate an initial trajectory while simultaneously modeling human trust via the MPD trust model. Subsequently, the system optimizes the initial trajectory to match the dynamic trust level, ultimately forming an optimized path for UAV trajectory tracking.
Machines 13 00876 g002
Figure 3. Illustration of the calculation of discrete Fréchet distance (DFD) in trajectory selection. The DFD of each candidate trajectory with respect to the operator input primitives and the current trajectory is calculated.
Figure 3. Illustration of the calculation of discrete Fréchet distance (DFD) in trajectory selection. The DFD of each candidate trajectory with respect to the operator input primitives and the current trajectory is calculated.
Machines 13 00876 g003
Figure 4. Trust-aware trajectory weight adjustment function.
Figure 4. Trust-aware trajectory weight adjustment function.
Machines 13 00876 g004
Figure 5. Two-dimensional trajectory comparison between trust-aware and trust-unaware methods under different obstacle densities. The three panels show trajectories in sparse (top, 50 trees), medium (middle, 100 trees), and dense (bottom, 200 trees) environments. The colorbar on the right indicates obstacle heights in meters. Green and red circles represent start and end points, respectively.
Figure 5. Two-dimensional trajectory comparison between trust-aware and trust-unaware methods under different obstacle densities. The three panels show trajectories in sparse (top, 50 trees), medium (middle, 100 trees), and dense (bottom, 200 trees) environments. The colorbar on the right indicates obstacle heights in meters. Green and red circles represent start and end points, respectively.
Machines 13 00876 g005
Figure 6. Three-dimensional visualization of trajectory comparison across sparse, medium, and dense obstacle environments.
Figure 6. Three-dimensional visualization of trajectory comparison across sparse, medium, and dense obstacle environments.
Machines 13 00876 g006
Table 1. Motion primitive parameters.
Table 1. Motion primitive parameters.
ParameterMinMaxNum. Discretizations
Duration T 0.3 s 1.5 s 5
Angular vel. ω 0.75 rad / s 0.75 rad / s 15
Z vel. v z 0.75 m / s 0.75 m / s 3
Table 2. MPD trust model parameters.
Table 2. MPD trust model parameters.
Parameter α h C α h P α h + α h C m m
Value0.50.920.090.420.85
Table 3. Results for the random forest environment.
Table 3. Results for the random forest environment.
ApproachDistance (m)Duration (s)Avg. Vel. (m/s)Jerk Integral (m2/s3)Num. InputsAvg. Trust
Sparse (50 trees)
Trust-unaware71.15 ± 0.0646.62 ± 1.491.53 ± 0.0527.49 ± 1.6339 ± 40.744 ± 0.013
Trust-aware70.04 ± 0.6345.53 ± 0.471.54 ± 0.0221.30 ± 3.3934 ± 30.813 ± 0.016
Medium (100 trees)
Trust-unaware71.33 ± 1.1648.88 ± 1.711.46 ± 0.0644.88 ± 10.3845 ± 70.657 ± 0.015
Trust-aware70.29 ± 2.4445.84 ± 2.391.52 ± 0.0331.48 ± 10.6034 ± 50.750 ± 0.013
Dense (200 trees)
Trust-unaware71.73 ± 1.7154.89 ± 5.291.31 ± 0.1484.69 ± 9.6856 ± 80.611 ± 0.017
Trust-aware71.38 ± 2.0049.86 ± 1.051.43 ± 0.0848.11 ± 9.4443 ± 50.693 ± 0.023
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhuang, Q.; Huang, K.; Jin, X.; Li, P.; Zhao, Y.; Kang, Y. Enhancing Human–Machine Collaboration: A Trust-Aware Trajectory Planning Framework for Assistive Aerial Teleoperation. Machines 2025, 13, 876. https://doi.org/10.3390/machines13090876

AMA Style

Zhuang Q, Huang K, Jin X, Li P, Zhao Y, Kang Y. Enhancing Human–Machine Collaboration: A Trust-Aware Trajectory Planning Framework for Assistive Aerial Teleoperation. Machines. 2025; 13(9):876. https://doi.org/10.3390/machines13090876

Chicago/Turabian Style

Zhuang, Qianzheng, Kangjie Huang, Xiaoran Jin, Pengfei Li, Yunbo Zhao, and Yu Kang. 2025. "Enhancing Human–Machine Collaboration: A Trust-Aware Trajectory Planning Framework for Assistive Aerial Teleoperation" Machines 13, no. 9: 876. https://doi.org/10.3390/machines13090876

APA Style

Zhuang, Q., Huang, K., Jin, X., Li, P., Zhao, Y., & Kang, Y. (2025). Enhancing Human–Machine Collaboration: A Trust-Aware Trajectory Planning Framework for Assistive Aerial Teleoperation. Machines, 13(9), 876. https://doi.org/10.3390/machines13090876

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop