Next Article in Journal
Effectiveness of Powered Hand Exoskeleton on Upper Extremity Function in People with Chronic Stroke
Previous Article in Journal
Prescribed Performance Global Non-Singular Fast Terminal Sliding Mode Control of PMSM Based on Linear Extended State Observer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Encouraging Guidance: Floating Target Tracking Technology for Airborne Robotic Arm Based on Reinforcement Learning

1
College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
2
College of Electronic Engineering, Nanjing Xiaozhuang University, Nanjing 211171, China
*
Author to whom correspondence should be addressed.
Actuators 2025, 14(2), 66; https://doi.org/10.3390/act14020066
Submission received: 17 December 2024 / Revised: 22 January 2025 / Accepted: 29 January 2025 / Published: 31 January 2025
(This article belongs to the Section Actuators for Robotics)

Abstract

:
Aerial robots equipped with operational robotic arms are a powerful means of achieving aerial contact operations, and their core competitiveness lies in target tracking control at the end of the airborne robotic arm (ARA). In order to improve the learning efficiency and flexibility of the ARA control algorithm, this paper proposes the encouraging guidance of an actor–critic (Eg-ac) algorithm based on the actor–critic (AC) algorithm and applies it to the floating target tracking control of ARA. It can quickly lock in the exploration direction and achieve stable tracking without increasing the learning cost. Firstly, this paper establishes approximate functions, policy functions, and encouragement functions for the state value of ARA. Secondly, an adoption rate controller (ARC) module was designed based on the concept of heavy rewards and light punishments (HRLP). Then, the kinematic and dynamic models of ARA were established. Finally, simulation was conducted using stable baselines3 (SB3). The experimental results show that, under the same computational cost, the convergence speed of the Eg-ac is improved by 21.4% compared to deep deterministic policy gradient (DDPG). Compared with soft actor–critic (SAC) and DDPG, Eg-ac has improved learning efficiency by at least 20% and has a more agile and stable floating target tracking effect.

1. Introduction

With the continuous development of aerial robot technology, flying robots carrying different onboard equipment to reach designated positions and complete high-altitude tasks have gradually become a situation of interest in aerial operations [1,2,3,4]. Fast and safe working methods are gradually replacing traditional manual operations. At present, there is still a certain gap between the various applications of aerial robots and the widespread applications in the concept [5]. There are still many technical issues that need to be further addressed, such as designing specific robot configurations, operating tools, and operating modes for different application goals [6,7,8,9]. Moreover, in-depth research and the optimization of control strategies for aerial work robots have become urgent technical challenges that need to be addressed.
This article takes obstacle removal, line repair, bolt tightening, and other aerial contact operations of high-altitude transmission lines as the research background. At present, these tasks still mainly rely on manual operations with significant safety hazards [10,11], and the intelligent operation of airborne robotic arms is a powerful means to solve the above problems. As shown in Figure 1, the aerial robot system studied in this article can perform line hanging operations on high-altitude routes. This robot integrates a multirotor unmanned aerial vehicle, a hanging line walking system, and an airborne multidegree of freedom robotic arm that can accurately reach the target position in complex aerial environments and complete hanging line, walking, and line clearing tasks. Due to the complex aerial operation environment and the constantly changing external disturbances, the core competitiveness of this system lies in the precise control of the aerial operation robotic arm. Only by improving positioning accuracy and tracking performance can the accuracy and effectiveness of aerial operations be ensured. Based on this, this work aimed to study the control problem of an airborne operation robotic arm in an aerial robot system and conduct research on the target tracking task of the robotic arm tool end.
Prior to this, scholars both domestically and internationally have conducted extensive research on the control issues of manipulator arms [12,13,14,15,16,17]. In recent years, control algorithms based on machine learning [18,19] have gradually emerged, and reinforcement learning frameworks have become a promising algorithmic tool for improving the control of aerial robotic arms [20,21,22]. The resulting deep reinforcement learning [23,24,25,26,27,28] combines the perceptual ability of deep learning with the decision-making ability of reinforcement learning, achieving end-to-end control from input to output, attracting the attention of a large number of researchers. And research on robotic arm control based on deep reinforcement learning is also becoming increasingly popular. After the successful application of deep learning in object detection, H. Sekkat et al. [29] proposed a neural inverse kinematic solution based on deep reinforcement learning for object detection using deep learning models, which evolved grasping tasks by achieving expected goals. This method calculates the joint angle of the detected position through inverse kinematics, causing the robot arm to move toward the position of the target object. The simulation results showed that the accuracy of the end effector grip joint angle and posture of the robot is satisfactory. T. Lindner [30] studied six combinations of four reinforcement learning algorithms for robot localization tasks. These algorithms were used for the positioning control of robot arm models, taking into account the evaluation of positioning accuracy, motion trajectory, and the number of steps required to achieve the target. The simulation and experimental results indicated that the RL algorithm can be successfully applied to the learning of robot arm positioning control. K. M. Oikonomou et al. [31] argued that although deep neural networks (DNNs) have achieved significant results in many robot applications, energy consumption remains a major limitation. They proposed a hybrid variant based on the deep deterministic policy gradient (DDPG) learning method for training six-degree-of-freedom robotic arms for target arrival tasks. Among them, a peak neural network was introduced into the actor model, and a DNN was introduced into the critic model. Finally, the hybrid DDPG model was compared with the classical DDPG model, demonstrating the superiority of the hybrid method. Coincidentally, B. Y. Song [32] proposed a dual-delay space robotic arm trajectory planning technique based on deep reinforcement learning to solve complex dynamics and control problems in the process of space debris removal. This technique can achieve end-to-end control effects comparable to the human grasping of objects. This study utilized joint and end effector control strategies developed using trajectory planners, trajectory trackers, and seven different weighted reward functions to implement a trajectory planning method for a floating space robotic arm to capture space debris. The experimental results indicated that this capture policy can maintain a high capture success rate. P. Wu [33] believed that traditional robotic arm control algorithms often struggle to adapt to the challenges posed by dynamic obstacles. Therefore, a reinforcement learning-based dynamic obstacle avoidance method was proposed to solve the real-time processing problem of dynamic obstacles. This method introduces a feature extraction network with an integrated gating mechanism based on traditional reinforcement learning algorithms. In addition, an adaptive dynamic reward mechanism was designed to optimize obstacle avoidance strategies. Verification showed that this method can effectively avoid randomly moving obstacles and significantly improved in convergence speed compared to traditional algorithms.
The summary above indicates that deep reinforcement learning has been widely applied in the research of space robotic arm control algorithms for tasks such as arrival, grasping, and trajectory tracking. Most of the robotic arms studied are series-connected with fixed bases, and the algorithms currently being researched can also achieve the basic goal of completing tasks. However, most algorithms do not consider the training efficiency and external disturbances of the robotic arm while pursuing positioning accuracy. When it comes to the end positioning of unstable bases and tracking of dynamic target objects, some algorithms seem inadequate and often yield unsatisfactory results.
The aerial operation-type robotic arm studied in this paper is mounted on the fuselage of a rotary wing of an unmanned aerial vehicle. Due to the special working environment, its end tool will sway with the shaking of the base and fuselage. In the task of clearing obstacles on the route studied in this article, most of the objects were floating objects with irregular floating phenomena. Therefore, the positioning accuracy and flexibility requirements of the tool end were high. Conventional control struggles to quickly locate the target object due to lag, while current reinforcement learning algorithms for robotic arms can achieve ideal results through learning. Generally speaking, value-based methods output the values of actions and are typically used in environments with discrete action spaces [23]. The policy-based approach, which outputs the probability of a direct action or action, is usually more suitable for environments with high-dimensional or continuous action spaces [26]. The behavior space of the airborne robotic arm (ARA) end positioning control is continuous and large-scale, and basic algorithms such as Monte Carlo reinforcement learning based on complete sampling and temporal differential reinforcement learning based on incomplete sampling may have low efficiency and even fail to achieve good solutions [34]. In view of this, this study adopted a policy-based learning method that regards the policy as a parameterized policy function of the ARA state and joint power output. Through establishing an objective function and using the rewards generated by the interaction between the ARA and the environment, the parameters of the policy function are learned. Currently, the most commonly used policy learning methods have a high degree of randomness in the early stages of learning, which may result in useless exploration or even completely opposite exploration paths to the target task. This situation can lead to a slower iteration speed in the early stages of learning, thereby reducing the convergence speed of the objective function. Although it ensures the exploration function of the algorithm, it increases the computational cost as a result.
In summary, this article proposes an encouragement-guided policy learning algorithm, Eg-ac, that adds an encourager to the actor–critic algorithm. The main function of the encourager is to generate strategies consistent with the final target task direction based on the current and target states of the ARA. In addition, the algorithm also incorporates a heavy rewards and light punishments (HRLP) reward mechanism and an adoption rate controller (ARC) module. The basic idea of HRLP is to increase rewards for good behavior and reduce punishments for bad behavior, thereby accelerating learning and encouraging exploration. The ARC is used to randomly generate adoption rates for the encourager under certain constraints, and the final policy is for actors and the encourager to output the ARA’s behavior strategies under the regulation of the ARC. The advantage of doing so is that the ARA executes the policy and obtains the corrected state before the next training begins. As the saying goes, ’a good start is half the battle’, and correcting the ARA status at any time is beneficial for approaching the final target position more quickly in subsequent training, thereby shortening the number of training steps included in an experience and improving training efficiency. Figure 2 illustrates the control policy for ARA’s floating target tracking control learning using the Eg-ac algorithm proposed in this paper. Firstly, this paper establishes approximate functions, policy functions, and encouragement functions for the state value of the ARA. The value function can evaluate and optimize strategies, and the optimized policy function, through the ARC joint encouragement function, will output more reasonable behavioral strategies for the ARA, which will in turn make the value function more accurate in reflecting the value of the state. The three function types mutually promote each other and ultimately obtain the optimal tracking policy for the ARA. Then, this paper establishes the kinematic and dynamic models of the ARA, which obtain the joint positions through inverse kinematic calculation and input them into the dynamic system to obtain the current state. Finally, this study conducted simulations using the open source reinforcement learning library stable baselines3 (SB3) built on the pytorch framework.
The following sections of this paper are arranged as follows. The second part elaborates on the principle and process of the Eg-ac algorithm. The third part describes the establishment of the ARA model and the setting of simulation parameters. The fourth part presents the simulation details. The fifth part summarizes the results and the method.

2. Algorithm Principles and Processes

2.1. ARA Status Description

The ARA system is a continuous power system with state characteristic s t s and system input a t a , denoted as Equation (1).
s t + 1 = f s t , a t + g ( t )
Among them, s t + 1 is the prediction of the next state based on the current state feature s t taking action a t , f is an unknown function, and g t is a random noise function for exploring a small range around the control variable, where t represents all the time steps recorded in the previous step during the training process.
Assuming that each temporal process of ARA has Markovian properties, establish a Markov decision process tuple S , A , P , R , γ for the ARA:
s t = [ p a x , p a y , p a z , p ˙ a x , p ˙ a y , p ˙ a z , p t x , p t y , p t z ] T , s t S a = p ¨ x , p ¨ y , p ¨ z , a A P s t s t + 1 a = P S t + 1 | S t = s t , A t = a R s t a = Ε R t + 1 | S t = s t , A t = a γ 0 ,   1
Among them, s t is a set of state representations in state set S . p a x , p a y , and p a z are the position coordinates of the ARA in the three axis directions of the end position in the body coordinate system at the current time. p ˙ a x , p ˙ a y , and p ˙ a z are the velocities of the ARA in the three coordinate axes of the end position in the body coordinate system. p t x , p t y , and p t z are the target position coordinates of the ARA in the three coordinate axis directions of the end position in the body coordinate system. s t = [ p a x , p a y , p a z , p ˙ a x , p ˙ a y , p ˙ a z , p t x , p t y , p t z ] T describes the state parameters of the ARA in the body coordinate system. Since the ARA operation object studied in this article is a floating object ahead, the pitch angle of the ARA operation end position in the body coordinate system was set to remain unchanged, and the state parameters do not need to consider the end effector attitude. a is a set of control variables in control set A , where p ¨ x , p ¨ y , and p ¨ z represent the acceleration of the ARA in the three coordinate axis directions of the end position in the body coordinate system. P s t s t + 1 a represents a set of transition probabilities in the state transition probability matrix P , which represents the transition probability from the current state s t to the next state s t + 1 through control variable a . The reward obtained during this process is R t + 1 . Subscripts t and t + 1 in the above variables represent the current time step and the next time step, respectively. R s t a represents a set of reward values in the reward function R , which are the rewards obtained from the current state s t after passing through control variable a . The γ in tuple S , A , P , R , γ is the decay factor, and S t is the finite set of states for the current time step t . A t is a finite set of control variables for the current time step t .
Establish the motion state update equation for the ARA as follows:
s t + 1 = f S t = s t , A t = a | π θ s t , a + g ( t ) π θ s t , a = P A t = a | S t = s t , θ
Among them, π θ s t , a is the ARA control strategy. θ is the strategy parameter. P A t = a | S t = s t , θ is the probability distribution matrix of executing control variable a under the given state s t and strategy parameter θ . f S t = s t , A t = a | π θ s t , a is the dynamic equation obtained by continuously updating s t , a , s t + 1 of the strategy, and g ( t ) is the noise function.

2.2. Eg-ac Control Policy

The Eg-ac algorithm proposed in this article is based on the AC algorithm, and the specific control policy is as follows:
π Eg - ac s t , a = P A t = a μ a E | S t = s t , θ μ ,   θ E
Among them, θ μ is the actor policy network parameter, and θ E is the encourager policy parameter. The Eg-ac policy advocates that the predicted behavior given is the dynamic weighted behavior value of the actor’s behavior and the encourager’s output behavior, i.e., A t = a μ a E , where the dynamic weight comes from the ARC module. The symbol represents weighted addition. The critic network estimates the behavior value based on the ARC output behavior and calculates the gradient of the objective function and the loss of the critic, thereby updating the parameters of each network. It is worth noting that the design of the Eg-ac control strategy references the HRLP reward mechanism and incorporates the ARC model. The specific design details are as follows.
(a)
ARC Module
In order to improve learning efficiency while also considering algorithm exploration function, this paper proposes introducing an ARC before behavior output to randomly generate adoption rates η A R C for the encourager under certain constraints (e.g., η A R C < 30 % ). N contains a set of data corresponding to the joint positions of the ARA that will be described later. And the actor will also generate a probability of adoption with a value of 1 η A R C . And after training to a certain extent, η A R C 0 is set. It is known that the encourager produces controllable and directionally correct behaviors. In the initial training stage, the actor network is not yet mature, and the output values generated are irregular, even going against the expected values. In this situation, appropriately reducing the adoption rate of the actor and partially accepting encourager’s suggestion output value will correct the final output action value appropriately. In the actual execution process, the output action value develops in a positive direction, that is, moving along the expected trajectory. Firstly, the punishment level will be reduced, and the reward level will be increased. Secondly, the corrected state will be obtained before the next training, which is beneficial for approaching the final target position more quickly in subsequent training, thereby shortening the number of training steps included in an experience and improving training efficiency. The reason for setting η A R C 0 after training to a certain extent is that the encourager only plays a corrective guidance role in the initial stage, and it adopts an error control method. When it comes to target tracking tasks, there will be lag like in traditional control methods. Therefore, later participation will affect the exploratory ability of the actor itself, not only hindering the improvement of training effectiveness but also limiting the flexibility of the algorithm. Therefore, with the real-time correction of the ARA status, the next training step can approach the final target position more quickly, thus completing an experience as soon as possible, shortening the training steps, and improving training efficiency.
(b)
HRLP Reward Mechanism
The presentation of the HRLP reward mechanism is based on the output behavior of the ARC module mentioned above. Due to the regulation of the ARC, the positive behavior output is amplified, which is then used as an execution strategy to obtain a positive ARA state and generate greater reward values. Bad behavior can be corrected, and executing the behavior will result in a less inappropriate state than before, thereby reducing punishment. In summary, HRLP is responsible for determining rewards based on the output values of policy functions. Its basic idea is to increase rewards for good behavior and reduce punishments for bad behavior, thereby accelerating learning and encouraging exploration.
Let the actor network function be μ θ μ , the critic network function be Q θ Q , the encourager function be E θ E , the target actor network function be μ θ μ , and the target critic network function be Q θ Q . The specific control process of the Eg-ac policy is shown in Figure 3, and g t is the random noise function. In the randomly selected state–action pairs s i , a i , r i + 1 , s i + 1 from the experience pool, state s i generates a specific behavior a μ through the actor network as one of the action input values for the ARC module, and the next state s i + 1 generates a predicted behavior a through the target actor network as the action input value for the target critic network to calculate the estimated value. The encourager network inputs the pre and post state values and error values from historical data into the PD controller, generating specific behavior a E as another action input value for the ARC module. The ARC module uses a randomly generated set of adoption rates to integrate two sets of actions and outputs them to the critic network to calculate the behavior value Q s i , a i corresponding to the state s i and the integrated behavior a i = a μ a E . The target critic network generates the behavioral value Q s i + 1 , a for calculating the target value Q Target based on the subsequent states s i + 1 and a .
In the above process, in assuming that the reward obtained from interacting with the environment under strategy π θ is R s , a , the objective function is designed as follows:
J θ = Ε π θ R s , a
The ultimate goal of the algorithm is to find the optimal behavior a i = ( 1 η A R C ) a μ + η A R C a E output by the ARC to maximize the value of the output behavior, where the reward obtained by executing a i is r i + 1 , and this transformation process is stored as s i , a i , r i + 1 , s i + 1 . In assuming that each batch has Γ execution steps, the system carries out a certain amount of the conversion process and randomly selects X samples to calculate the target value Q Target and the loss L o s s c of the critic network:
Q Target = r i + 1 + γ Q s i , μ s i | θ μ | θ Q
L o s s c = 1 X i Q Target Q s i , a i | θ Q 2
The update method of the actor network adopts the gradient ascent method. Considering that most optimizers are designed for gradient descent, this study achieves the effect of gradient ascent by minimizing the negative Q value and uses the average negative Q value of all samples in the batch as the loss. The gradient θ A J of the objective function and the loss L o s s a of the actor network are as follows:
θ A J 1 X a Q s , a | θ Q | s = s i , a = μ s i E s i θ μ μ s | θ μ | s i
L o s s a = 1 X i Q s i , a i | θ Q
Target network update:
ς θ Q + 1 ς θ Q θ Q
ς θ μ + 1 ς θ μ θ μ
Among them, g t is the noise function, γ is the update parameter, and ς is the learning rate.
The proportional proprietary controller corresponding to the encourager is defined as
E ( s i ) = θ E e k   e ˙ k T e k = s i s d
Among them, θ E is the controller adjustment parameter, e k is the error of state e ˙ k , which is the derivative of the state error, and s i and s d are the sampled state and its reference value, respectively.
Algorithm 1 provides an overview of the proposed algorithm.
Algorithm 1: Eg-ac Algorithm
Input   : γ , ς , θ Q , θ μ , θ E
Output   :   optimized   θ Q , θ μ
initialize   E s | θ E   with   weights   θ E by Equation (12)
randomly   initialize   ARC   with   weights   η A R C
randomly   initialize   Q s , a | θ Q ,   μ s | θ μ ,   with   weights   θ Q ,   θ μ
initialize   target   network   Q s , a | θ Q ,   μ s | θ μ   with   weights   θ Q θ Q ,   θ A θ A
initialize the experience cache space T
for episode from 1 to Limit do
        Initialize a noise function
        receive the initial state
        for t = 1 to  Γ  do
                 Perform   action   a t   via   ARC ,   get   reward   r t + 1 ,   next   state   s t + 1 , store
                 transition   s t , a t , r t + 1 , s t + 1 in T 
                sample a batch of random X   transitions   s i , a i , r i + 1 , s i + 1 in T, where
                   i = 0 , 1 , 2 , X
                set target value function via Equation (6)
                update critic by minimizing the loss via Equation (7)
                update actor policy by policy gradient via Equations (8) and (9)
                update networks via Equations (10) and (11)
        end for
end for
This paper compares the performance of the SAC algorithm, DDPG algorithm, and Eg-ac algorithm in the following sections.

3. Preparation Work for Simulation

This section will elaborate on the simulation preparation work for the algorithm proposed in the paper. Section 3.1 first describes conducting a kinematic analysis of the ARA based on the Denavi–Hartenberg (D-H) method [35] to obtain the inverse kinematic solution of the ARA. Next, a dynamic analysis of the ARA was conducted to prepare for subsequent simulations. Section 3.2 introduces the simulation network structure and parameter settings.

3.1. Establishment of ARA Mathematical Model

In the Eg-ac algorithm, both the input and output of the ARC module are composed of the acceleration at the end of the ARA operation, while in the simulation process, the ARA power execution system requires inputs for each joint position. This requires the use of an inverse kinematic solution, and the ARA power system is required during the algorithm training process, so this study conducted kinematic and dynamic analyses of the ARA.

3.1.1. ARA Kinematic Analysis

The ARA operating mechanism of this paper was designed as a five-degree-of-freedom robotic arm, and Figure 4 is a three-dimensional schematic of the robotic arm. In order to enhance the structural stiffness of the main connecting rod and improve positioning accuracy, the motion chain of the robotic arm is not a simple open-loop system. Instead, a four-bar linkage mechanism is used to connect driving motor 3 to connecting rod 3, and two four-bar linkage mechanisms are used to connect driving motor 4 to connecting rod 4. The movements of the other three driving motors are directly connected to the joints. The layout of the driver and connecting rod is shown in the figure. Drive motor 2 controls the position of joint 2, while the posture of connecting rod 3 remains unchanged relative to the base; drive motor 3 can adjust the posture of connecting rod 3 relative to the base, while the posture of connecting rod 4 relative to the base remains unchanged. Similarly, driving motor 4 specifically adjusts the posture of connecting rod 4 relative to the base, regardless of the position of joints 2 and 3. The control of each joint is relatively independent, unlike serial joints where the movement of the latter joint is relative to the movement of the previous link.
The coordinate systems of each connecting rod are marked in Figure 4. Among them, the relative coordinate of the workbench coordinate system O ^ t X ^ t Y ^ t Z ^ t with respect to the body coordinate system O ^ 0 X ^ 0 Y ^ 0 Z ^ 0 is l 3 , 0 , 0 . The plane where the origin of each coordinate system is located is designated as the operating arm plane. The position of the ARA shown in the figure corresponds to joint vector Θ = 0 , 90 , 90 , 90 , 0 . Table 1 describes symbols used in the text.
Calculate the linkage transformation matrix based on the D-H parameters in Table 2.
T 1 0 = c θ 1 s θ 1 0 0 s θ 1 c θ 1 0 0 0 0 1 0 0 0 0 1 ,   T 2 1 = c θ 2 s θ 2 0 0 0 0 1 0 s θ 2 c θ 2 0 0 0 0 0 1 ,   T 3 2 = c θ 3 s θ 3 0 l 2 s θ 3 c θ 3 0 0 0 0 1 0 0 0 0 1 , T 4 3 = c θ 4 s θ 4 0 l 3 s θ 4 c θ 4 0 0 0 0 1 0 0 0 0 1 ,   T 5 4 = c θ 5 s θ 5 0 0 0 0 1 0 s θ 5 c θ 5 0 0 0 0 0 1
The linkage transformation matrix T 5 0 is multiplied by the matrices of Equation (13):
T 5 0 = T 1 0 T 2 1 T 3 2 T 4 3 T 5 4
T 5 0 = = r 11 r 12 r 13 p x r 21 r 22 r 23 p y r 31 r 32 r 33 p z 0 0 0 1
r 11 = c 1 c 234 c 5 s 1 s 5 r 21 = s 1 c 234 c 5 + c 1 s 5 r 31 = s 234 c 5 r 12 = c 1 c 234 s 5 s 1 c 5 r 22 = s 1 c 234 s 5 + c 1 c 5 r 32 = s 234 s 5 r 13 = c 1 s 234 r 23 = s 1 s 234 r 33 = c 234 p x = c 1 l 2 c 2 + l 3 c 23 p y = s 1 l 2 c 2 + l 3 c 23 p z = l 2 s 2 l 3 s 23
By the kinematic equations of the ARA, the position and orientation of the end effector coordinate system can be calculated from the joint vectors. In order to simplify the solution and avoid the occurrence of multiple solutions, this study chose to use partial algebraic and partial geometric solutions for the inverse kinematic solution of the ARA.
θ 1 = A tan 2 p y , p x θ 2 = A tan 2 p z , p x 2 + p y 2 A tan 2 l 3 sin θ 3 , l 2 + l 3 cos θ 3 θ 3 = A tan 2 1 cos 2 θ 3 , cos θ 3 θ 4 = θ 234 θ 2 θ 3 θ 5 = A tan 2 r 21 c 1 r 11 s 1 , r 22 c 1 r 12 s 1
Among them, function A tan 2 ( y , x ) is a bivariate arctangent function, and when calculating tan 1 y x , the quadrant where the angle is located can be determined based on the signs of x and y . θ 234 and cos θ 3 are, respectively,
θ 234 = A tan 2 r 13 c 1 + r 23 s 1 , r 33
cos θ 3 = p x 2 + p y 2 + p z 2 l 2 2 l 3 2 2 l 2 l 3

3.1.2. ARA Dynamic Analysis

Use Lagrange method to dynamically model the designed ARA. The joint position variable set q i ( i = 1 , 2 , , 5 ) of ARA is defined as a generalized coordinate system, and its Lagrange function is defined as
L = T U T = i = 1 5 T i U = i = 1 5 U i
Among them, L is the Lagrange function; T and U are the total kinetic energy and total potential energy of the obstacle-clearing operation mechanism, respectively; and T i and U i are the kinetic energy and potential energy of the first link of the robotic arm, respectively. Due to the fact that link 4 and link 5 are located at the end of the robotic arm, only the kinetic and potential energies of the first three links are considered. The total kinetic energy is
T = 1 2 I 1 q ˙ 1 2 + 1 2 m 2 r 2 2 q ˙ 1 2 cos 2 q 2 + 1 2 m 2 r 2 2 q ˙ 2 2 + 1 2 m 3 q ˙ 1 2 a 2 sin q 2 + r 3 sin q 2 + q 3 2 + 1 2 m 3 a 2 2 q ˙ 2 2 + 1 2 m 3 r 3 2 q ˙ 2 + q ˙ 3 2 + m 3 a 2 r 3 q ˙ 2 2 + q ˙ 2 q ˙ 3
The total potential energy is
U = m 2 g r 2 m 2 g r 2 cos q 2 + m 3 g a 2 m 3 g a 2 cos q 2 + m 3 g r 3 m 3 g r 3 cos q 2 + q 3
Among them, I i is the inertia tensor of joint i ; q ˙ i is the generalized velocity of joint i ; m i is the mass of joint i ; a i is the length of connecting rod i ; r i is the distance from the centroid of joint i to the joint axis; and g is the acceleration due to gravity.
The Lagrange equation can be expressed as Equations (23) and (24):
d d t ( L q ˙ i ) L q i = τ
d d t ( T q ˙ ) T q + U q = τ
Among them, τ is the joint torque, and τ i is the joint torque at joint i .
Calculate the relevant partial derivatives as follows:
T q ˙ = I 1 + m 2 r 2 2 cos 2 q 2 + m 3 a 2 sin q 2 + r 3 sin q 2 + q 3 2 q ˙ 1 m 2 r 2 2 + m 3 a 2 2 + m 3 r 3 2 + 2 m 3 a 2 r 3 q ˙ 2 + m 3 r 3 2 + m 3 a 2 r 3 q ˙ 3 m 3 r 3 2 + m 3 a 2 r 3 q ˙ 2 + m 3 r 3 2 q ˙ 3
T q = 0 1 2 m 3 q ˙ 1 2 a 2 2 m 2 r 2 2 q ˙ 1 2 sin 2 q 2 + m 3 q ˙ 1 2 a 2 r 3 sin 2 q 2 + q 3 + 1 2 m 3 q ˙ 1 2 r 3 2 sin 2 q 2 + 2 q 3 m 3 q ˙ 1 2 a 2 r 3 sin q 2 cos q 2 + q 3 + m 3 q ˙ 1 2 r 3 2 sin q 2 + q 3 cos q 2 + q 3
U q = 0 m 2 g r 2 sin q 2 + m 3 g a 2 sin q 2 + m 3 g r 3 sin q 2 + q 3 m 3 g r 3 sin q 2 + q 3
Obtain the dynamic equation of the ARA:
τ 1 = I 1 + m 2 r 2 2 cos 2 q 2 + m 3 a 2 sin q 2 + r 3 sin q 2 + q 3 2 q ¨ 1 + 2 m 3 a 2 sin q 2 + 2 m 3 r 3 sin q 2 + q 3 a 2 cos q 2 + r 3 cos q 2 + q 3 m 2 r 2 2 sin ( 2 q 2 ) q ˙ 1 q ˙ 2 + 2 m 3 a 2 r 3 cos q 2 + q 3 sin q 2 + 2 m 3 r 3 2 cos q 2 + q 3 sin q 2 + q 3 q ˙ 1 q ˙ 3
τ 2 = m 2 r 2 2 + m 3 a 2 2 + m 3 r 3 2 + 2 m 3 a 2 r 3 q ¨ 2 + m 3 r 3 2 + m 3 a 2 r 3 q ¨ 3 1 2 m 3 q ˙ 1 2 a 2 2 m 2 r 2 2 q ˙ 1 2 sin 2 q 2 m 3 q ˙ 1 2 a 2 r 3 sin 2 q 2 + q 3 1 2 m 3 q ˙ 1 2 r 3 2 sin 2 q 2 + 2 q 3 + m 2 r 2 + m 3 a 2 g sin q 2 + m 3 r 3 g sin q 2 + q 3
τ 3 = m 3 r 3 2 + m 3 a 2 r 3 q ¨ 2 + m 3 r 3 2 q ¨ 3 m 3 q ˙ 1 2 a 2 r 3 sin q 2 cos q 2 + q 3 1 2 m 3 q ˙ 1 2 r 3 2 sin 2 q 2 + 2 q 3 + m 3 r 3 g sin q 2 + q 3
Simplify the dynamic equation and obtain the matrix form of the dynamic equation for the ARA:
τ = D ( q ) q ¨ + B ( q , q ˙ ) + G ( q )
Among them,
D ( q ) = I 1 + m 2 r 2 2 cos 2 q 2 + m 3 a 2 sin q 2 + r 3 sin q 2 + q 3 2 0 0 0 m 2 r 2 2 + m 3 a 2 2 + m 3 r 3 2 + 2 m 3 a 2 r 3 m 3 r 3 2 + m 3 a 2 r 3 0 m 3 r 3 2 + m 3 a 2 r 3 m 3 r 3 2
B ( q , q ˙ ) = 2 m 3 a 2 sin q 2 + 2 m 3 r 3 sin q 2 + q 3 a 2 cos q 2 + r 3 cos q 2 + q 3 m 2 r 2 2 sin ( 2 q 2 ) q ˙ 1 q ˙ 2 + 2 m 3 a 2 r 3 cos q 2 + q 3 sin q 2 + 2 m 3 r 3 2 cos q 2 + q 3 sin q 2 + q 3 q ˙ 1 q ˙ 3 1 2 m 3 q ˙ 1 2 a 2 2 m 2 r 2 2 q ˙ 1 2 sin 2 q 2 m 3 q ˙ 1 2 a 2 r 3 sin 2 q 2 + q 3 1 2 m 3 q ˙ 1 2 r 3 2 sin 2 q 2 + 2 q 3 m 3 q ˙ 1 2 a 2 r 3 sin q 2 cos q 2 + q 3 1 2 m 3 q ˙ 1 2 r 3 2 sin 2 q 2 + 2 q 3
G ( q ) = 0 m 2 r 2 + m 3 a 2 g sin q 2 + m 3 r 3 g sin q 2 + q 3 m 3 r 3 g sin q 2 + q 3

3.2. Network Architecture Design

Based on the preparation work analyzed by the above institutions, this study used the open source reinforcement learning library SB3 built on the pytorch framework for simulation. The real-time position p a x , p a y , p a z and velocity p ˙ a x , p ˙ a y , p ˙ a z of the ARA were obtained in the workbench coordinate system through the ARA power system. Then, the ARA state is fed into the input of the Eg-ac algorithm. The output behavior a = p ¨ a x , p ¨ a y , p ¨ a z describes the position control quantity under system control, and its learning process is the iterative strategy function solving process. The ARA power system will provide current status data and determine reward values at each time step.
The Eg-ac algorithm is based on the actor–critic algorithm, so this article refers to the processing method in DDPG. Target actor and target critic networks were added with the same structure as the actor and critic networks in the algorithm. The architectures of the actor and critic networks are shown in Figure 5. This study designed a critic with three hidden layers. The hidden layers for the processing states and behaviors are first operated separately, and the value of the state behavior is output by fully connecting them together through the last hidden layer. The input received by the actor network is the number of ARA features, and the specific value of each behavioral feature is output to the ARC module. The actor network is designed with three hidden layers, fully connected between layers. In order to achieve exploration, the algorithm adds a random noise with a mean of 0 to the generated behavior, allowing it to explore a certain range around the exact behavior. To prevent overfitting during the optimization process, the dropout algorithm is introduced to regularize the network. The dropout algorithm was proposed by Hinton [36] in 2012. In each training batch, through ignoring half of the feature detectors (setting half of the hidden layer nodes to 0), it can effectively alleviate the occurrence of overfitting and achieve regularization to a certain extent.
Before simulation, the following parameters were set as follows:
(1)
The definition of the reward value is the negative value of the distance between the end of the ARA and the target position, which means that the larger the error, the lower the reward value.
r i = p t x p a x 2 + p t y p a y 2 + p t z p a z 2
(2)
Each training batch is executed Γ = 256 times. When there is an error e 0.05 between the end of the ARA task and the target position during an experience, the task is considered completed, and the current experience is stopped and reset to enter the next experience.
(3)
Let the simulation time step be t and the total training steps be Ν = 20 , 000 .
(4)
Set learning rates ς = 0.001 , γ = 0.99 , and batch sampling quantity X = 100 .
(5)
Controller parameters in the encourager: θ E = 0.5   0.05 .
(6)
Output range setting: Q s , a | θ Q 10 , 10 , μ s | θ μ 0.05 , 0.05 , E s | θ E 0.05 , 0.05 .
(7)
Set the random noise range to g ( t ) 0.002 , 0.002 .
(8)
The size of the ARA connecting rod is l 2 = 0.2 ( m ) , l 3 = 0.15 ( m ) .
(9)
To ensure that the pitch and roll angles of the ARA end remain unchanged, and to make the end face of joint 5 perpendicular to the plane of the operating arm, the following settings are used: θ 4 = θ 2 + θ 3 , θ 5 = 0 .
(10)
The adoption rate parameter of the ARC is composed of a set of random numbers η 1 , η 2 η 5 η i 30 % , i = 1 , 2 , 5 with five elements generated by the computer. The process diagram is shown in Figure 6. In this process, the input variables are first subjected to inverse kinematics (IK) to obtain the joint positions θ i i = 1 , 2 , 5 of the ARA and then integrated into θ i E μ i = 1 , 2 , 5 through adoption rate calculation. Finally, the final behavior is output through kinematic calculations (KCs). Among them, θ i E μ = η i θ i E + 1 η i θ i μ i = 1 , 2 , 5 . Set η A R C 0 when the training steps are t 3000 .
(11)
The simulated computer configuration used an Intel(R) Core (TM) i5-7300HQ (ASUS, Taipei, China).

4. Simulation Results and Analysis

The ARA was trained using the SAC, DDPG, and Eg-ac algorithms. During the process, every four episodes were grouped together, and the reward values obtained for each episode were recorded and averaged as an average reward value, as shown in Figure 7. For clarity and intuitiveness, the green boxes in the images represent enlarged areas. The trend of the reward episode lines for the three algorithms in the figure is basically the same, but compared with the three phases, Eg-ac has a faster learning efficiency and the fastest and most stable reward rise slope, followed by DDPG, and SAC converges the slowest and fluctuates slightly near convergence. When the reward values of the three algorithms reach their highest, the horizontal axis position shows that the learning speed of Eg-ac is at least 20% faster than that of the other two algorithms. The figure shows that for the same number of training steps, the numbers of episodes for the three algorithms are approximately as follows: SAC (4100), DDPG (4800), and Eg-ac (5800). The reason why Eg-ac has the most episodes is because it can quickly find the direction, reach the target position, and move on to the next experience with fewer steps in each episode. The same conclusion can also be drawn from the more intuitive table in Figure 8. In Figure 8, as mentioned above, every four episodes form a group, and the average number of time steps contained in each episode is taken as the average step count. It can be seen that Eg-ac has shown significant advantages in ARA training, as it can shorten learning time and converge quickly.
To evaluate the training cost of the proposed algorithm, the training time and execution frequency of the three algorithms were recorded. Figure 9 shows the comparison of the training time. The training time consumed by the Eg-ac algorithm is between that of the other two algorithms, and the training time of the three algorithms is almost the same. The average number of steps taken per unit time was calculated, i.e., the execution frequency, in groups of four episodes. As shown in Figure 10, the large initial values are due to factors such as network and parameter initialization time. In terms of execution frequency, the DDPG algorithm has the highest number of training steps per unit time. SAC is the lowest, while Eg-ac values are between the two. The three sets of values are quite close, so it can be concluded that Eg-ac does not have significant differences in computational cost compared to the other algorithms, and the training time difference between the three is not significant, which is within an acceptable range.
Due to the fact that both the Eg-ac and DDPG algorithms have two target networks in their network structures, while the critic loss function uses mean squared error (MSE) loss, the MSE loss will cause the prediction of the critic network to approach the target value, so its convergence speed is related to the training efficiency and learning effect of the algorithm. Figure 11 shows the loss comparison curves between the Eg-ac algorithm and the DDPG algorithm. The loss trend of both algorithms first increases and then decreases, with MSE loss decreasing and gradually lowering the loss value. The difference between the two curves is that the peak of Eg-ac is relatively earlier, indicating a faster reaction and a consistently lower loss than DDPG, ultimately remaining at around 0.1, while DDPG is around 0.2. It can be seen that the designs of the ARC module and HRLP reward mechanism in the Eg-ac algorithm effectively increase the training efficiency of the algorithm and improve the convergence speed by about 21.4%.
Based on the above analysis, this study selected two sets of floating target trajectories for algorithm model validation experiments. Experiment 1 used a spatial spiral curve as the floating trajectory to simulate ARA tracking when the aircraft body shakes due to wind disturbance. Experiment 2 used a planar curve as the floating trajectory to simulate ARA tracking when the aircraft is stationary.
In Experiment 1, it is assumed that the floating trajectory of the obstacle to be cleared by the ARA under the dual effects of the base and wind disturbance is a spatial arc. The models were trained using three different algorithms, and the motion parameters of the ARA were recorded. Figure 12 shows the tracking trajectories after the execution of the three algorithms. It can be intuitively analyzed that the trajectory of SAC is the most unstable, with position jitter occurring during the tracking process, while the stability of the Eg-ac and DDPG algorithms was better. Figure 13 shows the tracking errors of three algorithms during execution. Analysis shows that Eg-ac has a smaller tracking error compared to DDPG, which is basically stable at around 0.01 m, while DDP is about 0.03 m. Figure 14 and Figure 15, respectively, show the displacement and velocity curves of the ARA in the three coordinate axes of the workbench during Experiment 1. From the analysis of the simulation curve above, it can be seen that under the same number of training steps and training time, the algorithm proposed in this paper exhibits more accurate and stable tracking performance in the process of floating target tracking.
In Experiment 2, it is assumed that the obstacle to be cleared by the ARA will generate a floating trajectory as a planar arc without external disturbance. The models were trained using three different algorithms, and the motion parameters of the ARA were recorded. Figure 16 shows the tracking trajectories after the execution of the three algorithms. Figure 17 shows the tracking errors of three algorithms during execution. As shown in the figure, during the planar tracking process, the tracking error of Eg-ac is basically stable at around 0.01 m, while DDPG is about 0.02 m. Figure 18 and Figure 19, respectively, show the displacement and velocity curves of the ARA in the three coordinate axes of the workbench during Experiment 2. Based on the analysis of the simulation curve above, it can be concluded that, similar to the conclusion of Experiment 1, the algorithm proposed in this paper exhibits more accurate and stable tracking performance in floating target tracking under the same number of training steps and training time.
In order to more intuitively demonstrate the computational costs during the execution of the three algorithms, this study recorded the computation time during the execution of each algorithm. Figure 20 shows the partial single-step computation time of three algorithms. The average computation time of the three algorithms is 997 (μs), which further proves that the algorithm proposed in this paper does not waste too much computational cost while improving learning efficiency and tracking performance. All the simulation data above can be found in the Supplementary Materials section.

5. Conclusions

This article proposes an Eg-ac algorithm based on the AC algorithm and applies it to the floating target tracking control of the ARA. The research objectives of the algorithm proposed in this paper were to quickly lock the exploration direction during the process of the ARA reaching the floating target position, improve learning efficiency, and obtain stable tracking results without increasing learning costs. Based on the above objectives, this study established approximate functions, strategy functions, and incentive functions for ARA state values in algorithm construction and designed an ARC module. Among them, the ARC generates the adoption rate for the encourager and outputs the ARA behavior strategy under the regulation of the ARC. Given that the inverse kinematic settlement and dynamic system execution of the ARA are required during the algorithm model training process, this paper establishes the kinematic and dynamic models of the ARA based on the D-H method. The target positions of each joint are obtained through inverse kinematic calculation, and the current state is obtained through the dynamic system. Finally, simulation was conducted using the open source reinforcement learning library SB3 built on the pytorch framework. The experimental results show that under the same computational cost, the loss function convergence speed of the Eg-ac algorithm designed in this study was improved by 21.4% compared to that of DDPG. Compared with SAC and DDPG, Eg-ac improved learning efficiency by at least 20% and has a more agile and stable floating target tracking performance.
While proposing a better algorithm in this article, there are some inevitable aspects that need improvement or can make the proposed algorithm better, such as the following: (1) Significant oscillation phenomena when approaching the target need to be improved upon. (2) The simulation did not consider the end effector attitude of the ARA. If it is necessary to consider the end effector attitude, research on attitude angles needs to be conducted. (3) There are various types of airborne disturbances, and in future work, it is necessary to further refine the disturbance effects and improve system stability. The authors will focus on addressing these areas in future research.

Supplementary Materials

The following supporting information can be downloaded at: https://github.com/keepfoolisher/Floating-point-tracking-technology/releases/tag/v1, accessed on 9 December 2024.

Author Contributions

Conceptualization, J.W. and Z.Y.; methodology, J.W.; software, J.W. and H.Z.; validation, L.L. and D.C.; formal analysis, J.W.; investigation, H.Z., C.X. and J.W.; resources, Z.Y.; data curation, J.W.; writing—original draft preparation, J.W.; writing—review and editing, J.W., C.X. and Z.Y.; visualization, J.W., D.C. and Z.W.; supervision, Z.Y.; project administration, Z.Y.; funding acquisition, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the Guangxi Power Grid Company’s 2023 Science and Technology Innovation Project under Grant GXKJXM20230169 and in part by the Guizhou Provincial Science and Technology Projects, grant number Guizhou-Sci-Co-Supp [2020]2Y044.

Data Availability Statement

The data supporting the findings of this study are available within the Supplementary Materials.

Acknowledgments

The authors would like to thank the reviewers for their constructive comments and suggestions that helped improve this paper.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Liao, L.; Yang, Z.; Wang, C.; Xu, C.; Xu, H.; Wang, Z.; Zhang, Q. Flight Control Method of Aerial Robot for Tree Obstacle Clearing with Hanging Telescopic Cutter. Control Theory Appl. 2023, 40, 343–352. [Google Scholar]
  2. Wang, M.; Chen, Z.; Guo, K.; Yu, X.; Zhang, Y.; Guo, L.; Wang, W. Millimeter-Level Pick and Peg-in-Hole Task Achieved by Aerial Manipulator. IEEE Trans. Robot. 2024, 40, 1242–1260. [Google Scholar] [CrossRef]
  3. Kang, K.; Prasad, J.V.R.; Johnson, E. Active Control of a UAV Helicopter with a Slung Load for Precision Airborne Cargo Delivery. Unmanned Syst. 2016, 4, 213–226. [Google Scholar] [CrossRef]
  4. Amri bin Suhaimi, M.S.; Matsushita, K.; Kitamura, T.; Laksono, P.W.; Sasaki, M. Object Grasp Control of a 3D Robot Arm by Combining EOG Gaze Estimation and Camera-Based Object Recognition. Biomimetic 2023, 8, 208. [Google Scholar] [CrossRef]
  5. Villa, D.K.D.; Brandão, A.S.; Sarcinelli-Filho, M. A Survey on Load Transportation Using Multirotor UAVs. J. Intell. Robot. Syst. 2020, 98, 267–296. [Google Scholar] [CrossRef]
  6. Tagliabue, A.; Kamel, M.; Verling, S.; Siegwart, R.; Nieto, J. Collaborative transportation using MAVs via passive force control. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017. [Google Scholar]
  7. Darivianakis, G.; Alexis, K.; Burri, M.; Siegwart, R. Hybrid Predictive Control for Aerial Robotic Physical Interaction towards Inspection Operations. In Proceedings of the IEEE International Conference on Robotics & Automation, Hong Kong, China, 31 May–7 June 2014. [Google Scholar]
  8. Molina, J.; Hirai, S. Aerial pruning mechanism, initial real environment test. Robot. Biomim. 2018, 7, 127–132. [Google Scholar]
  9. Roderick, W.R.T.; Cutkosky, M.R.; Lentink, D. Bird-inspired dynamic grasping and perching in arboreal environments. Sci. Robot. 2021, 6, eabj7562. [Google Scholar] [CrossRef] [PubMed]
  10. Sun, X. Application of Intelligent Operation and Maintenance Technology in Power System. Integr. Circuit Appl. 2023, 40, 398–399. [Google Scholar]
  11. Zhang, Q.; Liao, L.; Xiao, S.; Yang, Z.; Chen, K.; Wang, Z.; Xu, H. Research on the aerial robot flight control technology for transmission line obstacle clearance. Appl. Sci. Technol. 2023, 50, 57–63. [Google Scholar]
  12. Suarez, A.; Heredia, G.; Ollero, A. Physical-Virtual impedance control in ultralightweight and compliant Dual-Arm aerial manipulators. IEEE Robot. Autom. Lett. 2018, 3, 2553–2560. [Google Scholar] [CrossRef]
  13. Zhang, G.; He, Y.; Dai, B.; Gu, F.; Yang, L.; Han, J.; Liu, G. Aerial Grasping of an Object in the Strong Wind: Robust Control of an Aerial Manipulator. Appl. Sci. 2019, 9, 2230. [Google Scholar] [CrossRef]
  14. Nguyen, H.; Lee, D. Hybrid force/motion control and internal dynamics of quadrotors for tool operation. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013. [Google Scholar]
  15. Zhong, H.; Miao, Z.; Wang, Y.; Mao, J.; Li, L.; Zhang, H.; Chen, Y.; Fierro, R. A Practical Visual Servo Control for Aerial Manipulation Using a Spherical Projection Model. IEEE Trans. Ind. Electron. 2020, 67, 10564–10574. [Google Scholar] [CrossRef]
  16. Alexis, K.; Huerzeler, C.; Siegwart, R. Hybrid predictive control of a coaxial aerial robot for physical interaction through contact. Control Eng. Pract. 2014, 32, 96–112. [Google Scholar] [CrossRef]
  17. Zhuo, H.; Yang, Z.; You, Y.; Xu, N.; Liao, L.; Wu, J.; He, J. A Hierarchical Control Method for Trajectory Tracking of Aerial Manipulators Arms. Actuators 2024, 13, 333. [Google Scholar] [CrossRef]
  18. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529. [Google Scholar] [CrossRef] [PubMed]
  19. Hasselt, H.V.; Guez, A.; Silver, D. Deep reinforcement learning with double q-learning. arXiv 2015, arXiv:1509.06461. [Google Scholar] [CrossRef]
  20. Qiu, Z.; Liu, Y.; Zhang, X. Reinforcement Learning Vibration Control and Trajectory Planning Optimization of Translational Flexible Hinged Plate System. Eng. Appl. Artif. Intell. 2024, 133, 108630. [Google Scholar] [CrossRef]
  21. Yang, A.; Chen, Y.; Naeem, W.; Fei, M.; Chen, L. Humanoid motion planning of robotic arm based on human arm action feature and reinforcement learning. Mechatronics 2024, 7, 102630. [Google Scholar] [CrossRef]
  22. Zhang, S.; Xia, Q.; Chen, M.; Cheng, S. Multi-Objective Optimal Trajectory Planning for Robotic Arms Using Deep Reinforcement Learning. Sensors 2023, 23, 5974. [Google Scholar] [CrossRef]
  23. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing Atari with Deep Reinforcement Learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  24. Peters, J.; Schaal, S. Policy Gradient Methods for Robotics. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 2219–2225. [Google Scholar]
  25. Silver, D.; Lever, G.; Heess, N.; Degris, T.; Wierstra, D.; Riedmiller, M. Deterministic Policy Gradient Algorithms. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), Beijing, China, 21–26 June 2014; pp. 387–395. [Google Scholar]
  26. Haarnoja, T.; Zhou, A.; Hartikainen, K.; Tucker, G.; Ha, S.; Tan, J.; Kumar, V.; Zhu, H.; Gupta, A.; Abbeel, P.; et al. Soft Actor-Critic Algorithms and Applications. arXiv 2019, arXiv:1812.05905. [Google Scholar]
  27. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  28. Fujimoto, S.; van Hoof, H.; Meger, D. Addressing Function Approximation Error in Actor-Critic Methods. arXiv 2018, arXiv:1802.09477. [Google Scholar]
  29. Sekkat, H.; Tigani, S.; Saadane, R.; Chehri, A. Vision-based robotic arm control algorithm using deep reinforcement learning for autonomous objects grasping. Appl. Sci. 2021, 11, 7917. [Google Scholar] [CrossRef]
  30. Lindner, T.; Milecki, A.; Wyrwał, D. Positioning of the Robotic Arm Using Different Reinforcement Learning Algorithms. Int. J. Control. Autom. Syst. 2021, 19, 1661–1676. [Google Scholar] [CrossRef]
  31. Oikonomou, K.M.; Kansizoglou, I.; Gasteratos, A. A Hybrid Reinforcement Learning Approach with a Spiking Actor Network for Efficient Robotic Arm Target Reaching. IEEE Robot. Autom. Lett. 2023, 8, 3007–3014. [Google Scholar] [CrossRef]
  32. Song, B.Y.; Wang, G.L. A Trajectory Planning Method for Capture Operation of Space Robotic Arm Based on Deep Reinforcement Learning. J. Comput. Inf. Sci. Eng. 2024, 24, 091003-1. [Google Scholar] [CrossRef]
  33. Wu, P.; Su, H.; Dong, H.; Liu, T.; Li, M.; Chen, Z. An obstacle avoidance method for robotic arm based on reinforcement learning. Ind. Robot 2024, 52, 9–17. [Google Scholar] [CrossRef]
  34. Wu, J.; Yang, Z.; Liao, L.; He, N.; Wang, Z.; Wang, C. A State-Compensated Deep Deterministic Policy Gradient Algorithm for UAV Trajectory Tracking. Machines 2022, 10, 496. [Google Scholar] [CrossRef]
  35. Denavit, J.; Hartenberg, R.S. A kinematic notation for lower-pair mechanisms based on matrices. Trans. ASME J. Appl. Mech. 1955, 22, 215–221. [Google Scholar] [CrossRef]
  36. Hinton, G.E.; Srivastava, N.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R.R. Improving neural networks by preventing coadaptation of feature detectors. Comput. Sci. 2012, 3, 212–223. [Google Scholar]
Figure 1. Aerial robot system.
Figure 1. Aerial robot system.
Actuators 14 00066 g001
Figure 2. Control policy of Eg-ac algorithm.
Figure 2. Control policy of Eg-ac algorithm.
Actuators 14 00066 g002
Figure 3. Eg-ac control policy.
Figure 3. Eg-ac control policy.
Actuators 14 00066 g003
Figure 4. Schematic of ARA structure design and coordinate establishment.
Figure 4. Schematic of ARA structure design and coordinate establishment.
Actuators 14 00066 g004
Figure 5. Network architecture settings in simulation.
Figure 5. Network architecture settings in simulation.
Actuators 14 00066 g005
Figure 6. Schematic of ARC processing process.
Figure 6. Schematic of ARC processing process.
Actuators 14 00066 g006
Figure 7. Comparison of rewards.
Figure 7. Comparison of rewards.
Actuators 14 00066 g007
Figure 8. Comparison of episode steps.
Figure 8. Comparison of episode steps.
Actuators 14 00066 g008
Figure 9. Comparison of training time.
Figure 9. Comparison of training time.
Actuators 14 00066 g009
Figure 10. Execution frequency.
Figure 10. Execution frequency.
Actuators 14 00066 g010
Figure 11. Loss curve of critic network.
Figure 11. Loss curve of critic network.
Actuators 14 00066 g011
Figure 12. Three-dimensional curve of ARA tracking trajectory in Experiment 1.
Figure 12. Three-dimensional curve of ARA tracking trajectory in Experiment 1.
Actuators 14 00066 g012
Figure 13. Tracking error curve in Experiment 1.
Figure 13. Tracking error curve in Experiment 1.
Actuators 14 00066 g013
Figure 14. Displacement curves in each coordinate axis direction in Experiment 1: (a) X-axis displacement; (b) Y-axis displacement; (c) Z-axis displacement.
Figure 14. Displacement curves in each coordinate axis direction in Experiment 1: (a) X-axis displacement; (b) Y-axis displacement; (c) Z-axis displacement.
Actuators 14 00066 g014
Figure 15. Tracking speed curve in Experiment 1: (a) X-axis velocity; (b) Y-axis velocity; (c) Z-axis velocity; (d) tracking linear velocity.
Figure 15. Tracking speed curve in Experiment 1: (a) X-axis velocity; (b) Y-axis velocity; (c) Z-axis velocity; (d) tracking linear velocity.
Actuators 14 00066 g015aActuators 14 00066 g015b
Figure 16. ARA tracking trajectory in Experiment 2.
Figure 16. ARA tracking trajectory in Experiment 2.
Actuators 14 00066 g016
Figure 17. Tracking error curve in Experiment 2.
Figure 17. Tracking error curve in Experiment 2.
Actuators 14 00066 g017
Figure 18. Displacement curves in each coordinate axis direction in Experiment 2: (a) X-axis displacement; (b) Y-axis displacement; (c) Z-axis displacement.
Figure 18. Displacement curves in each coordinate axis direction in Experiment 2: (a) X-axis displacement; (b) Y-axis displacement; (c) Z-axis displacement.
Actuators 14 00066 g018
Figure 19. Tracking speed curve in Experiment 2: (a) X-axis velocity; (b) Y-axis velocity; (c) Z-axis velocity; (d) tracking linear velocity.
Figure 19. Tracking speed curve in Experiment 2: (a) X-axis velocity; (b) Y-axis velocity; (c) Z-axis velocity; (d) tracking linear velocity.
Actuators 14 00066 g019aActuators 14 00066 g019b
Figure 20. Comparison of running time.
Figure 20. Comparison of running time.
Actuators 14 00066 g020
Table 1. Symbol explanation.
Table 1. Symbol explanation.
SymbolExplanation
T i i 1 Transformation   matrix   of   coordinate   system   i   relative   to   coordinate   system   i 1
c θ i   or   c i Abbreviation   of   cos θ i
s θ i   or   s i Abbreviation   of   sin θ i
c 12 c 12 = cos θ 1 + θ 2
s 12 s 12 = sin θ 1 + θ 2
α i 1 Angle   of   rotation   from   Z ^ i 1   to   Z ^ i   around   axis   X ^ i 1
a i 1 Distance   from   Z ^ i 1   to   Z ^ i   along   axis   X ^ i 1
d i Distance   from   X ^ i 1   to   X ^ i   along   axis   Z ^ i
θ i Angle   of   rotation   from   X ^ i 1   to   X ^ i   around   axis   Z ^ i
Table 2. Connecting rod parameters of ARA.
Table 2. Connecting rod parameters of ARA.
i α i 1 a i 1 d i θ i
1000 θ 1
2−90°00 θ 2
30 l 2 0 θ 3
40 l 3 0 θ 4
590°00 θ 5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, J.; Yang, Z.; Zhuo, H.; Xu, C.; Liao, L.; Cheng, D.; Wang, Z. Encouraging Guidance: Floating Target Tracking Technology for Airborne Robotic Arm Based on Reinforcement Learning. Actuators 2025, 14, 66. https://doi.org/10.3390/act14020066

AMA Style

Wu J, Yang Z, Zhuo H, Xu C, Liao L, Cheng D, Wang Z. Encouraging Guidance: Floating Target Tracking Technology for Airborne Robotic Arm Based on Reinforcement Learning. Actuators. 2025; 14(2):66. https://doi.org/10.3390/act14020066

Chicago/Turabian Style

Wu, Jiying, Zhong Yang, Haoze Zhuo, Changliang Xu, Luwei Liao, Danguo Cheng, and Zhiyong Wang. 2025. "Encouraging Guidance: Floating Target Tracking Technology for Airborne Robotic Arm Based on Reinforcement Learning" Actuators 14, no. 2: 66. https://doi.org/10.3390/act14020066

APA Style

Wu, J., Yang, Z., Zhuo, H., Xu, C., Liao, L., Cheng, D., & Wang, Z. (2025). Encouraging Guidance: Floating Target Tracking Technology for Airborne Robotic Arm Based on Reinforcement Learning. Actuators, 14(2), 66. https://doi.org/10.3390/act14020066

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop