Next Article in Journal
Mortal vs. Machine: A Compact Two-Factor Model for Comparing Trust in Humans and Robots
Previous Article in Journal
Event-Based Closed-Loop Control for Path Following of a Purcell’s Three-Link Swimmer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Physical Reinforcement Learning with Integral Temporal Difference Error for Constrained Robots

by
Luis Pantoja-Garcia
1,
Vicente Parra-Vega
2,* and
Rodolfo Garcia-Rodriguez
3
1
Tecnologico de Monterrey, Institute of Advanced Materials for Sustainable Manufacturing, Av. General Ramon Corona 2514, Zapopan 45210, Jalisco, Mexico
2
Robotics and Advanced Manufacturing Department, Research Center for Advanced Studies (Cinvestav-Ipn), Mexico City 07360, Mexico
3
Facultad de Ciencias de la Administración, Universidad Autónoma de Coahuila, Saltillo 25225, Mexico
*
Author to whom correspondence should be addressed.
Robotics 2025, 14(8), 111; https://doi.org/10.3390/robotics14080111
Submission received: 5 June 2025 / Revised: 29 July 2025 / Accepted: 12 August 2025 / Published: 14 August 2025
(This article belongs to the Section AI in Robotics)

Abstract

The paradigm of reinforcement learning (RL) refers to agents that learn iteratively through continuous interactions with their environment. However, when the value function is unknown, a neural network is used, which is typically encoded into an unknown temporal difference equation. When RL is implemented in physical systems, explicit convergence and stability analyses are required to guarantee the worst-case operations for any trial, even when the initial conditions are set to zero. In this paper, physical RL (p-RL) refers to the application of RL in dynamical systems that interact with their environments, such as robot manipulators in contact tasks and humanoid robots in cooperation or interaction tasks. Unfortunately, most p-RL schemes lack stability properties, which can even be dangerous for specific robot applications, such as those involving contact (constrained) tasks or interaction tasks. Considering an unknown and disturbed DAE2 robot, in this paper a p-RL approach is developed to guaranteeing robust stability throughout a continuous-time-adaptive actor–critic, with local exponential convergence of force–position tracking error. The novel adaptive mechanisms lead to robustness, while an integral sliding mode enforces tracking. Simulations are presented and discussed to show our proposal’s effectiveness, and some final remarks are addressed concerning the structural aspects.

1. Introduction

Reinforcement learning (RL) has emerged as an alternative decision-making learning technique that issues a reinforcement to compute the action, or control, based on recollection of the discount reward encoded throughout the performance evaluation; a distinct feature of RL versus control is precisely that it evaluates the performance each time and iteratively corrects the action to improve it, leading to a task being learned. Unfortunately, for us roboticists, it was developed and substantiated for the tuple of state–action, with the state s being the output and the action π the input of a system Σ —however, without an explicit structure of Σ . Thus, to deploy RL in the physical layer, e.g., a robot, a lot of effort is required to guarantee that the physical structural properties of all of the components are met, with concerns arising about the safety of its deployment in a real-world application [1]. In addition, due to the iterative realm of these schemes, the safety requirements are exacerbated, leading to sim-to-real schemes to transfer RL schemes to the physical realm. In addition, when a rigid-body robot engages in contact with a rigid environment, high-frequency contact is of concern due to the high energy transfer index. Fortunately, the robotics community has proposed RL schemes based on a Lagrangian structure for Σ and derived RL motor learning, ready to meet such structures based on passivity-based and model-free adaptive schemes [2]. In the latter type of scheme, ordinary differential equations (ODEs) model the Lagrangian robot for free-motion tasks, and safety issues are handled since the passivity is preserved; uncertainties and unmodeled dynamics are considered; smooth action is synthesized; and no trial-and-error is involved, but explicit energy-based stability conditions are established. In this context, the energy-based stability and adaptive mechanism that substitutes pre-training sets apart this approach of RL for robots as a machine learning scheme since the learning paradigm of a black-box interactive process of exploration and exploitation is circumvented using a structure energy-based adaptation that guides reward recollection without pre-training, quite convenient for physical world applications. In this paper, we extended such an approach to constrained motion modeled using a differential algebraic system of equations (DAE2) to study position–force learning tasks.
Most robotic contact tasks require precise control of the normal force exerted on a surface while following a desired end-effector trajectory on the tangent plane. Active force–position control considers the DAE2’s constrained dynamics, wherein the derivative of the momentum represents the Lagrangian multiplier (contact force) to be controlled. Most schemes circumvent the DAE2 formulation, using ODE (unconstrained motion) dynamical models, where no state is related to the contact force [3]. Popular schemes that use the latter have proliferated, among them impedance control [4,5] and stiffness control, which cannot guarantee position–force convergence. On the other hand, the DAE2 formulation leverages orthogonalized subspaces to model orthogonal operators from the holonomic constraint, thereby projecting force and velocity errors into orthogonal complements. Then, control is designed to enforce force and position convergence simultaneously and independently [6], followed by model-based schemes [7]. Additionally, to compensate for unknown robot dynamics, even Lipschitz disturbances, a method called neurocontrol is proposed where neural networks, NNs, are used as approximator functions or to compensate for unknown dynamics. Some approaches to contact tasks based on NNs are presented in [8,9,10]. In recent years, the application of reinforcement learning (RL) to dynamical systems has attracted attention due to its ability to learn through interaction by encoding different behaviors [11]. Crucial elements of RL are the instrumentation of a “reward” and its “effort” (the value function), which qualify a robot’s task performance accordingly. Depending on such an evaluation, it issues a reinforcement to the “action”, which leads to an improved performance while minimizing the value function.
Actor–critic is a technique used to instrument RL based on a critic NN used to approximate the value function of the temporal difference error equation. In contrast, the actor NN approximates inverse dynamics to yield the action [12]. In control schemes, the actor NN is coupled with a damping term to stabilize the robot. This scheme has gained popularity due to its ability to handle uncertainties while gradually improving the performance to facilitate the successful completion of robotic tasks for free motion, such as in [13,14], to name a few. For contact/interaction tasks, see [15,16,17,18]; notably, in [19], an admittance controller computes the desired force based on RL, leading to a stable force error. Similarly, in [15], adaptive dynamic programming based on an LQR is explored, with bounded errors. In [16], an impedance-based advantage actor–critic (A2C) deals with constraint uncertainties, while [17] uses an impedance-based natural gradient actor–critic (NAC), with experiments. Additionally, [18] presents a model-based AC for optimal impedance tuning, and it is used for cooperation with multi-robots in [20,21], even for humanoid robots [22] without a stability analysis. Although variants of the AC approach have shown effectiveness for contact/interaction tasks, the literature has overlooked the force–position convergence, which is critical due to interactions that occur at a high frequency (due to rigid-body contact), in particular when the robot is constrained by a rigid environment, i.e., a holonomic constraint that yields a DAE2 model.
An early review [23] and recent work [24] on deep RL for manipulation have pointed out the need to explicitly address certain challenges in constrained robot dynamics resulting from the physical interaction, which we refer to as physical RL (p-RL). This acronym aims to distinguish it from RL schemes for unconstrained robots; the underlying foundation of p-RL is based on sound physical interaction principles or contact tasks rather than simply software apps (s-RL). That is, in p-RL, the physical layer of a robotic system cannot withstand hundreds, or thousands, or millions, of training or pre-training tests, as are typically required nowadays in RL, due to the risk of compromising the robot’s structural integrity. Thus, p-RL schemes need to be substantiated through a model-based stability analysis to produce an actionable (physically speaking) action command. Actionable p-RL implies control that is smooth, causal, model-free, and bounded, without pre-training requirements.
In this manuscript, we propose actionable p-RL for a DAE2 robot, showing position–force convergence with a stability analysis. A continuous-time AC scheme deploys RL, extending the methodology [25] from unconstrained to constrained robot dynamics. Hereby, fast and robust adaptation for the critic is obtained by handling robust convergence of the so-called integral temporal difference (ITD) error. At the same time, a passivity-based actor is proposed over orthogonalized error manifolds, where robust convergence is guaranteed based on chatterless integral sliding modes.

2. Preliminaries

2.1. Robot Dynamics and the Problem Statement

Let us take a robot manipulator with n degrees of freedom whose end-effector is constrained to evolve onto a rigid surface, modeled by the following differential algebraic equation (DAE index 2, or DAE-2)
H ( q ) q ¨ + C ( q , q ˙ ) q ˙ + τ f r + g ( q ) = τ + J φ T ( q ) λ + τ d ,
s . t . φ ( q ) = 0 ,
where q R n stands for the generalized position and q ˙ R n the generalized velocity, H ( q ) R n × n is the inertia matrix, C ( q , q ˙ ) R n × n is the Coriolis matrix, g ( q ) R n is the gravity vector, J φ R l × n is the constrained Jacobian, λ R 1 stands for the contact force in operational space, and τ f r represents affine friction and τ d a Lipschitz disturbance. Notice that the rigid surface is modeled by the implicit equation φ ( q ) = 0 R l , for l = 1 independent contact point. Additionally, the DAE system is of index 2 since the input appears in the second derivative of the constraint, and its state is ξ = ( q , q ˙ , p = λ ) , where q R n r , q ˙ R n r , and the momentum p R 1 .

2.2. The Actor–Critic Learning Control Problem

Leveraging the functional Y r = H ( q ) q ¨ r + C ( q , q ˙ ) q ˙ r + g ( q ) + τ f r , where q ˙ r is a nominal velocity reference to the shape orthogonalized error coordinate system S r = q ˙ q ˙ r , [26], then the open-loop error equation is defined as
H ( q ) S ˙ r + C ( q , q ˙ ) S r + B S r = τ + τ d + J φ T λ Y r .
Thus, the problem statement is to design a model-free p-RL control mechanism for a DAE2 robot (1) and (2) in order to generate a smooth action policy τ such that the state ξ = ( q , q ˙ , p ) ξ d = ( q d , q ˙ d , p d = λ d ) without any previous training or learning and any knowledge of the robot dynamics, Y r . The use of the term “model-free” in this work refers to a lack of knowledge of the robot’s dynamic model, but the kinematics are assumed to be known.

3. Learning Actor–Critic Design

In the AC algorithm proposed here, the critic learning process is based on a stability-guaranteed integral temporal difference error. On the other hand, the actor synthesizes a partial action, playing the role as an approximator of the inverse dynamics. In this way, the proposed controller is composed of the action plus a stabilizer term, typically a PID-like term.
In the following sections, it is shown how stability is ensured in the NN-critic using Lyapunov stability theory. After that, the adaptation laws for the NN-actor are designed to guarantee closed-loop stability.

3.1. The Value Function and the Temporal Difference Error

Let the value function in continuous time, representing the cumulative rewards, given as [27],
R = t 0 e m t ψ r ( m ) d m ,
where r represents the instant reward, and ψ corresponds to a discount factor with t 0 m . In continuous time, the value function is driven by the temporal difference (TD) error [11], such that differentiating (4), we have
R ˙ = 1 ψ R r ( t ) .
In cases where the value function R is not exactly known, the learning goal is to guarantee that
δ = R ˙ 1 ψ R + r = 0 .
Expressing (6) as a lower-order equation γ = δ = t 0 t f δ ( τ ) d τ , the integral temporal difference (ITD) error arises as
γ = R 1 ψ R + r .
Notice that if γ 0 , so does (6), and henceforth (5) applies. The foundations of this chain of implications lead to the learning rationale encoded by the unknown value function (4).
Remark 1.
The motivation behind the introduction of the ITD error is to drive the learning and stability properties of the critic by γ, specifically the quasi-sliding mode condition. Additionally, the inclusion of the ITD eliminates the need for R ¨ in the stability analysis. In the next section, the use of the TD error and the ITD as tools to design critic adaptive weights is explained.

3.2. The Critic Neural Network

Assuming that the smooth value function R of (4) is unknown, yet it is Lipschitz, then it can be approximated by a neural network. To this end, a large vector of the ideal weights W c and a basis Z c ( X c ) exist for a set of ideal inputs X c such that R = W c T Z c . Now, considering a smaller set of weights W c W c and basis Z c Z c ( ) , a small reconstruction error ϵ R appears that approximates the value function, i.e.,
R = W c T Z c + ϵ R .
The linear parametrization (8) suggests that a subset of adaptive weights W ^ c W c and a basis exist such that the value function R can be approximated as
R ^ = W ^ c T Z c ,
where W ^ c R p × 1 are adaptive weights; Z c = σ ( χ c ) is the bipolar sigmoid function, with the scaled input χ c = V c T X c , with V c a constant matrix; and X c is the input vector to the neural network. Using (9) and its derivative, the neural approximation of δ becomes
δ ^ = R ^ ˙ 1 ψ R ^ + r = W ^ ˙ c T Z c + W ^ c T Z ˙ c 1 ψ W ^ c T Z c + r ,
where δ ^ ( t 0 ) 0 in general. Note that using the NN parametrization (9) in Equation (7), we obtain γ ^ ˙ = δ ^ . The challenge now becomes proposing an adaptive law W ^ ˙ c such that δ ^ converges, meaning that the learning process is achieved within a small approximation error.
Proposition 1
(Adaptive Critic Weight Design). Let the adaptation law for the critic’s weights be defined as follows:
W ^ ˙ c = K ω sgn ( W ^ c ) K sgn ( γ ^ ) Z c Z c T Z c ,
where K w , K are positive gains. Then, γ ^ ϵ γ for ϵ γ a constant, which leads to γ ^ ˙ = 0   δ ^ = 0 . Finally, due to the reconstruction error, we obtain δ ϵ δ .
Proof. 
See Appendix A.1. □

3.3. The Actor Neural Network

The actor network has the role of approximating the inverse dynamics parametrized by the functional Y r . Then, analogously constant weights W a and the basis Z a exist such that
Y r = W a T Z a + ϵ a ,
where ϵ a is the neural reconstruction error. Henceforth, an actor neural network exists that approximates (12) as follows
Y ^ r = W ^ a T Z a ,
where W ^ a R h × n are adaptive weights, and Z a = σ ( χ a ) is the bipolar sigmoid function, with the input χ a = V a T X a , where V a are constant weights, and X a its input vector. The adaptation law W ^ ˙ a still needs to be designed.
Proposition 2
(Adaptive Actor Weight Design). Let the following adaptive law update the actor weights:
W ^ ˙ a = Γ a Z a S r T Γ a W ^ a ( γ ^ r ) 2 ,
where Γ a R + h × h is the feedback gain, and ( γ ^ r ) 2 constitutes the reinforcement. Then, weights W ^ a are bounded.
Proof. 
See Appendix A.2. □

3.4. Model-Free Actor–Critic Control Design

Assuming that Y r is unknown and approximated by the NN-actor, let the controller be given as
τ = K d S r + W ^ a T Z a + J φ T λ d S ˙ d F + K F tanh ( μ S q F ) + η S v F ,
where S r = q ˙ q ˙ r is an orthogonalized extended velocity error arising when the nominal reference q ˙ r is
q ˙ r = Q ( q ) q ˙ α Δ q + S d q K q sgn ( S q q ) + J φ + ( q ) β S v F ,
where Q ( q ) = I J φ + ( q ) J φ ( q ) is the orthogonal projection of J φ , J φ + is the right pseudoinverse, Δ q = q q d is the position error, q d C 2 is the desired position, and β is a positive definite gain. Thus, the extended velocity error S r can be rewritten as follows:
S r = Q ( q ) S v q J φ + ( q ) β S v F ,
where
S v q = S q q + K q sgn ( S q q ) position   error   manifold ,
S v F = S q F + K F sgn ( S q F ) force   error   manifold ,
for S q q = S q S d q , S q = Δ q ˙ + α Δ q , S d q = S q ( t 0 ) e k 1 t , S q F = S F S d F , S F = Δ λ , S d F = S F ( t 0 ) e k 2 t , and Δ λ = λ λ d , where λ d is the desired contact force, and α , k 1 , k 2 , K q , K F , β are positive definite feedback gains. Using (3) and (15), we obtain the following closed-loop error equation:
H ( q ) S ˙ r = C ( q , q ˙ ) S r B S r + τ d K d S r + W ^ a T Z a Y r + J φ T ( S ˙ v F + η S v F ) + K F J φ T ϵ z ,
where ϵ z = ( tanh ( μ S q F ) sgn ( S q F ) ) . Notice that the state of the closed-loop error equation is complemented with (10), (11) and (14). We are now ready to announce the main result.
Theorem 1.
Consider the constrained robot dynamics (1) in the closed loop with the proposed actor–critic learning controller (15), which leads to the system of Equations (10), (11), (14) and (20). Then, from Propositions 1 and 2 and selecting K d , K q , K F large enough, the integral sliding mode at S q q = 0 and S q F = 0 is enforced. Thus, the local exponential convergence of the tracking errors Δ q , Δ q ˙ , and Δ λ is guaranteed, and δ ^ 0 without any knowledge of the DAE2 dynamics.
Proof. 
See Appendix A.2. □

4. Simulations

Simulations are carried out in Matlab Simulink using the integrator ode4, with a fixed step of 1 × 10 4 . The simulation consists of a 3-dof robot in contact with a rigid plane parallel to the x y plane, see Figure 1, at a height of z d = 0.15 m. The desired trajectory onto the plane x y is a circle with the radius a = 0.15 centered at ( x c , y c ) = ( 0.5 , 0 ) ; then, the desired Cartesian path is x d = x c + a cos ( w t ) and y d = y c + a sin ( w t ) , with the frequency w = 1 ; the desired force is λ d = 25 + 2 sin ( t ) N.
The robot parameters are the following: link lengths l 1 = 0.5 , l 2 = 0.5 , and l 3 = 0.35 m; link masses m 1 = 7.2 , m 2 = 5 , and m 3 = 1.9 kg; distances to the center of mass l c 1 = 0.25 , l c 2 = 0.19 , and l c 3 = 0.12 ; and inertias I 1 = 0.15 , I 2 = 0.02 , and I 3 = 0.016 . The initial conditions are q ( t 0 ) = ( 0 , 0 , π / 2 ) T   r a d , q ˙ ( t 0 ) = ( 0 , 0 , 0 ) T   r a d / s , corresponding to being in contact with the plane.
The reward is designed to continuously supervise the tracking performance. Thus, a weighted summation of the position and velocity errors is considered, without weighting the integral of the tracking errors, regardless of its orthogonal projection. Then, r = Δ q T P Δ q + Δ q ˙ T A Δ q ˙ + ( Δ λ ) 2 , with P = d i a g ( 0.9 ) R n × n , A = d i a g ( 0.1 ) R n × n so that the position is greatly weighted given that the robot is constrained in its position. The adaptation gains are selected as K w = 7 , K = 50 , Γ a = I × 1500 and the feedback gains are selected as K d = d i a g ( 50 100 50 ) , β = 1 , α = d i a g ( 3 3 3 ) , K q = d i a g ( 1 1 1 ) , K F = 1.15 , k 1 = 50 , k 2 = 10 , η = 2000 .

Results

In Figure 2, we can see that the integral temporal difference error converges close to zero and remains constant; consequently, the temporal difference error, a derivative of the ITD error, converges to zero. This means that the adaptive laws for the NN-critic guarantee the convergence of the TD error. The integral temporal difference error dynamics facilitate motor learning through reward recollection; the larger the ITD error or the reward, the greater the reinforcement to improve the task performance. This is independent of how the reward policy is designed, including Pavlov’s negative rewards [28].
The joint position and velocity errors are shown in Figure 3, which converge approximately after 1 s, while Figure 4a shows the convergence of the force error in approximately 0.02 s, considerably faster than the position and velocity errors, shown in Figure 4b. Extended velocity errors are shown in Figure 5a, which remain bounded. Additionally, Figure 5b and Figure 6a show how the invariant manifolds for position-velocity and force, respectively, remain near zero. In Figure 6b, the performance of the smooth control signals is shown. Finally, Figure 7 shows how the actor–critic weights remain bounded, as is expected according to the theory. The convergence of the tracking errors stems from the underlying integral sliding mode, notably without any chattering in the controller, which contrasts with the existing literature, which lacks a similar result. Chattering refers to a very high-frequency control component that puts the structural integrity of the robot, let alone its constrained motion, in danger. It is unclear how to achieve smooth and robust control and convergence using another technique.
To compare the performance of the proposed approach, a comparative study is presented under the same assumptions as those for the model-free DAE robot, including the parameters and simulation conditions. Specifically, it is used as a force–position neurocontroller for a DAE-2 robot. For the comparative study, the following performance metrics are used: I T A E p o s = t 0 t t Δ q 2 represents the accumulation of the position error over time; I T A E v e l = t 0 t t Δ q ˙ 2 represents the accumulation of the velocity error over time; I A C = t 0 t τ and I A C V = t 0 t d d t τ represent the control effort index and the smoothness of the control signal index, respectively.
Table 1 presents the numerical values obtained for each case, indicating that our proposal yields smaller values, representing a better performance than that of the neurocontroller.

5. Conclusions

A novel model-free AC controller for contact tasks, which yields motor learning in DAE2 robots, is presented. This scheme is based on a method called p-RL, which refers to the application of RL to dynamical systems that interact with their environments. The proposed p-RL is instrumented through an AC architecture for robotic contact tasks, instrumenting novel adaptive laws. It is worth mentioning that the stability involves the nominal DAE2 dynamics; however, the AC-based controller scheme is free of model dynamics.
Update laws are designed to satisfy Lyapunov stability rather than being concerned with parametric variations, as in gradient descent. In this way, local minima are avoided, resulting in stronger stability properties in the time domain of the complete state of the closed-loop error equation.
A simulation for a 3D robot constrained by a rigid plane is presented, demonstrating tracking with smooth control actions, without requiring knowledge of the disturbed dynamics. An experimental evaluation is underway, including a comparison with various reward policies, as well as asymptotic model-free controllers that guarantee tracking, such as neurocontrol and integral sliding modes. Finally, we have employed a general formulation of DAE2 robots, which enables a straightforward extension to multirobot tasks, such as cooperative manipulators and robotic hand manipulation, particularly in critical contact tasks where continuous oversight of the performance is required.
Safety concerns arise when attempting to deploy an RL scheme (obtained under the premise of an abstract system Σ ) in a physical system. Lyapunov-based schemes are a straightforward methodology to deal with, as devised more than two decades ago in the sound paper [29], or, for instance, in synthesizing control Lyapunov functions from barrier Lyapunov functions or Lyapunov stability, as in our case.
Unlike the traditional RL schemes reported in the literature, the proposed p-RL does not require any training or learning in the actor or the critic. Although sim-to-real transfer may be recommended to reduce high real-world trial costs, it is not a fundamental step for the proposed approach. That is, the process of transferring knowledge or policies learned in a simulated environment to the real world is not necessary. Only the process of tuning the gains is required to satisfy the stability conditions, where physics-based simulators can be used to improve the transferability.
Last but not least, model-free force–position control for DAE2 robots was solved over three decades ago, without chattering [7]. Therefore, the question arises as to what contribution is obtained by introducing motor learning. When motor learning is deployed, tracking is commensurate with the performance evaluation, not only asymptotic tracking. This latter notion enforces tracking as t , whereas motor learning supervises the performance at each instant, not in the limit, at infinity.

Author Contributions

Conceptualization and investigation, L.P.-G., V.P.-V. and R.G.-R.; methodology and formal analysis, L.P.-G., V.P.-V. and R.G.-R.; software and resources, L.P.-G.; writing—original draft preparation, L.P.-G., V.P.-V. and R.G.-R.; writing—review, editing and supervision, V.P.-V. and R.G.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created, or used from any publicly archived datasets.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proofs of the Propositions and Theorem

Appendix A.1. Proof of Proposition 1

Consider the following Lyapunov candidate function (LCF):
V γ = 1 2 γ ^ 2 + 1 2 W ^ c T W ^ c
which qualifies as the LCF for system (10) and (11) for the origin ( γ ^ , W ^ c ) = ( 0 , 0 ) . Taking the time derivative to (A1) and substituting Equation (10) leads to
V ˙ γ = γ ^ γ ^ ˙ + W ^ c T W ^ ˙ c = γ ^ R ^ ˙ 1 ψ R ^ + r + W ^ c T W ^ ˙ c = γ ^ W ^ ˙ c T Z c + W ^ c T Z ˙ c 1 ψ W ^ c T Z c + r + W ^ c T W ^ ˙ c
Using the weight adaptation law defined in (11), (A2) can be written as follows:
V ˙ γ = γ ^ K sgn ( γ ^ ) Z c Z c T Z c K ω sgn ( W ^ c ) T Z c + W ^ c T Z ˙ c 1 ψ W ^ c T Z c + r + W ^ c T K sgn ( γ ^ ) Z c Z c T Z c K ω sgn ( W ^ c )
Considering ψ 1 and using Z c Z c T Z c 1 , Equation (A3) becomes
V ˙ γ = K γ ^ sgn ( γ ^ ) K ω γ ^ sgn ( W ^ c ) T Z c ϵ 1 + γ ^ W ^ c T Z ˙ c W ^ c T Z c γ ^ ψ ϵ 2 + γ ^ r K W ^ c T sgn ( γ ^ ) Z c Z c T Z c ϵ 3 K ω W ^ c T sgn ( W ^ c )
where ϵ 1 , ϵ 2 , ϵ 3 are bounded terms. Then, (A4) becomes
V ˙ γ K γ ^ K ω W ^ c + γ ^ Z ˙ c W ^ c + K ω γ ^ ϵ 1 ϵ 2 W ^ c + γ ^ r K ϵ 3 W ^ c ( K K ω ϵ 1 ) γ ^ ( K ω γ ^ Z ˙ c ϵ 2 K ϵ 3 ) W ^ c + γ ^ r η 1 γ ^ η 2 W ^ c + ϵ γ
If K > K ω ϵ 1 and K ω > ( | Z ˙ c | γ ^ + ϵ 2 ) / ( 1 + ϵ 3 ) , then η 1 > 0 , η 2 > 0 . In this condition, (A5) guarantees that ( γ ^ , W ^ c ) ϵ = ϵ γ / inf ( η 1 , η 2 ) , a very small vicinity. This implies that the L 1 norm induces a quasi-sliding mode at ( γ ^ , W ^ c ) = ( ϵ 5 , ϵ 6 ) in finite time but not at ( γ ^ , W ^ c ) = ( 0 , 0 ) (QED).

Appendix A.2. Proof of the Theorem

Consider the derivative of the Lyapunov candidate function V T = 1 2 S r T H ( q ) S r + 1 2 S v F β S v F + 1 2 t r ( W ˜ a T Γ a 1 W ˜ a ) + V γ , along with its solutions (14) and (20); for γ ˜ = Δ γ = γ γ ^ , we obtain
V ˙ T = S r T H ( q ) S ˙ r + S r T H ˙ 2 S r + S v F T β S ˙ v F + t r W ˜ a T Γ a 1 W ^ ˙ a + V ˙ γ = S r T ( K d + B ) S r + W ^ a T Z a Y r + τ d + K F J φ T ϵ z S v F T β η S v F t r W ˜ a T ( Z a S r T W ^ a ( γ ^ r ) 2 ) + V ˙ γ = S r T ( K d + B ) S r W ˜ a T Z a ϵ a + τ d + K F J φ T ϵ z S v F T β η S v F + S r T W ˜ a T Z a t r W ˜ a W ^ a ( γ ^ r ) 2 + V ˙ γ S r T ( K d + B ) S r + S r T ( τ d ϵ a ) S v F T β η S v F + S r T K F J φ T ϵ z ( γ ^ r ) 2 t r W ˜ a T ( W ˜ a W a ) η 1 γ ^ η 2 W ^ c + γ ^ r
where we have used the trace property, and V ˙ γ η 1 γ ^ η 2 W ^ c + γ ^ r . Thus, (A6) becomes
V ˙ T S r T ( K d + B ) S r S v F T β η S v F + S r ϵ a + ϵ τ d + E z η 1 γ ^ η 2 W ^ c ( γ ^ r ) 2 W ˜ a W a m a x 2 2 + ( γ ^ r ) 2 W a m a x 4 + γ ^ r
where ϵ τ d | | τ d | | and E z K F J φ T ϵ z . Then, for large enough gains K d , β , η , η 1 and η 2 , the closed-loop signals S r , W ˜ c , W ˜ a , and γ ^ remain bounded. Now, we continue to prove the existence of the sliding mode at S q q and S q F . From (18) and (19), we obtain
S q q T S ˙ q q = K q S q q + S ˙ v q S q q μ q S q q
S q F T S ˙ q F = K F S q F + S ˙ v F S q F μ F S q F
where μ q = K q ϵ q and μ F = K F ϵ F , with ϵ q > S ˙ v q and ϵ q > S ˙ v F ; then, tuning K q > ϵ q and K F > ϵ F yields μ q , μ F > 0 , guaranteeing the sliding mode condition at S q q = 0 and S q F = 0 , with ( Δ q , Δ q ˙ ) , Δ λ as the unique solution to these manifolds, respectively; thus, local exponential convergence for the position, velocity, and force tracking errors is proven (QED).

References

  1. Gu, S.; Yang, L.; Du, Y.; Chen, G.; Walter, F.; Wang, J.; Knoll, A. A Review of Safe Reinforcement Learning: Methods, Theories, and Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 11216–11235. [Google Scholar] [CrossRef]
  2. Pantoja-Garcia, L.; Parra-Vega, V.; Garcia-Rodriguez, R. Automatic reinforcement for robust model-free neurocontrol of robots without persistent excitation. Int. J. Adapt. Control Signal Process. 2024, 38, 221–236. [Google Scholar] [CrossRef]
  3. Chiaverini, S.; Siciliano, B.; Villani, L. A survey of robot interaction control schemes with experimental comparison. IEEE/ASME Trans. Mechatronics 1999, 4, 273–285. [Google Scholar] [CrossRef]
  4. Hogan, N. Impedance Control: An Approach to Manipulation. In Proceedings of the 1984 American Control Conference, San Diego, CA, USA, 6–8 June 1984; pp. 304–313. [Google Scholar] [CrossRef]
  5. Chien, M.C.; Huang, A.C. Adaptive Impedance Control of Robot Manipulators based on Function Approximation Technique. Robotica 2004, 22, 395–403. [Google Scholar] [CrossRef]
  6. McClamroch, N.; Wang, D. Feedback stabilization and tracking of constrained robots. IEEE Trans. Autom. Control 1988, 33, 419–426. [Google Scholar] [CrossRef]
  7. Parra-Vega, V.; Arimoto, S.; Liu, Y.; Naniwa, T. Model-based adaptive hybrid control for robot manipulators under holonomic constraints. IFAC Proc. Vol. 1994, 27, 475–480. [Google Scholar] [CrossRef]
  8. Jung, S.; Hsia, T. Neural network impedance force control of robot manipulator. IEEE Trans. Ind. Electron. 1998, 45, 451–461. [Google Scholar] [CrossRef]
  9. Bechlioulis, C.P.; Doulgeri, Z.; Rovithakis, G.A. Neuro-Adaptive Force/Position Control With Prescribed Performance and Guaranteed Contact Maintenance. IEEE Trans. Neural Netw. 2010, 21, 1857–1868. [Google Scholar] [CrossRef] [PubMed]
  10. Peng, G.; Yang, C.; He, W.; Chen, C.L.P. Force Sensorless Admittance Control With Neural Learning for Robots With Actuator Saturation. IEEE Trans. Ind. Electron. 2020, 67, 3138–3148. [Google Scholar] [CrossRef]
  11. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction, 2nd ed.; The MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  12. Barto, A.G.; Sutton, R.S.; Anderson, C.W. Looking Back on the Actor–Critic Architecture. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 40–50. [Google Scholar] [CrossRef]
  13. He, W.; Gao, H.; Zhou, C.; Yang, C.; Li, Z. Reinforcement Learning Control of a Flexible Two-Link Manipulator: An Experimental Investigation. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 7326–7336. [Google Scholar] [CrossRef]
  14. Vu, V.T.; Dao, P.N.; Loc, P.T.; Huy, T.Q. Sliding Variable-based Online Adaptive Reinforcement Learning of Uncertain/Disturbed Nonlinear Mechanical Systems. J. Control Autom. Electr. Syst. 2021, 32, 281–290. [Google Scholar] [CrossRef]
  15. Zhan, H.; Huang, D.; Chen, Z.; Wang, M.; Yang, C. Adaptive dynamic programming-based controller with admittance adaptation for robot–environment interaction. Int. J. Adv. Robot. Syst. 2020, 17, 1729881420924610. [Google Scholar] [CrossRef]
  16. Zhang, T.; Xiao, M.; Zou, Y.b.; Xiao, J.d.; Chen, S.y. Robotic Curved Surface Tracking with a Neural Network for Angle Identification and Constant Force Control based on Reinforcement Learning. Int. J. Precis. Eng. Manuf. 2020, 21, 869–882. [Google Scholar] [CrossRef]
  17. Liang, L.; Chen, Y.; Liao, L.; Sun, H.; Liu, Y. A novel impedance control method of rubber unstacking robot dealing with unpredictable and time-variable adhesion force. Robot. Comput.-Integr. Manuf. 2021, 67, 102038. [Google Scholar] [CrossRef]
  18. Zhao, X.; Han, S.; Tao, B.; Yin, Z.; Ding, H. Model-Based Actor−Critic Learning of Robotic Impedance Control in Complex Interactive Environment. IEEE Trans. Ind. Electron. 2022, 69, 13225–13235. [Google Scholar] [CrossRef]
  19. Perrusquía, A.; Yu, W.; Soria, A. Position/force control of robot manipulators using reinforcement learning. Ind. Robot. Int. J. Robot. Res. Appl. 2019, 46, 267–280. [Google Scholar] [CrossRef]
  20. Dao, P.N.; Do, D.K.; Nguyen, D.K. Adaptive Reinforcement Learning-Enhanced Motion/Force Control Strategy for Multirobot Systems. Math. Probl. Eng. 2021, 2021, 5560277. [Google Scholar] [CrossRef]
  21. Dao, P.N.; Liu, Y.C. Adaptive reinforcement learning in control design for cooperating manipulator systems. Asian J. Control 2022, 24, 1088–1103. [Google Scholar] [CrossRef]
  22. Katić, D.M.; Rodić, A.D.; Vukobratović, M.K. Hybrid dynamic control algorithm for humanoid robots based on reinforcement learning. J. Intell. Robot. Syst. Theory Appl. 2008, 51, 3–30. [Google Scholar] [CrossRef]
  23. Kormushev, P.; Calinon, S.; Caldwell, D.G. Reinforcement Learning in Robotics: Applications and Real-World Challenges. Robotics 2013, 2, 122–148. [Google Scholar] [CrossRef]
  24. Liu, R.; Nageotte, F.; Zanne, P.; de Mathelin, M.; Dresp-Langley, B. Deep Reinforcement Learning for the Control of Robotic Manipulation: A Focussed Mini-Review. Robotics 2021, 10, 22. [Google Scholar] [CrossRef]
  25. Pantoja-Garcia, L.; Parra-Vega, V.; Garcia-Rodriguez, R.; Vázquez-García, C.E. A Novel Actor—Critic Motor Reinforcement Learning for Continuum Soft Robots. Robotics 2023, 12, 141. [Google Scholar] [CrossRef]
  26. Parra-Vega, V.; Arimoto, S. A passivity-based adaptive sliding mode position-force control for robot manipulators. Int. J. Adapt. Control Signal Process. 1996, 10, 365–377. [Google Scholar] [CrossRef]
  27. Doya, K. Temporal Difference Learning in Continuous Time and Space. In Advances in Neural Information Processing Systems, Proceedings of the 1995 Conference, Denver, CO, USA, 27–30 November 1995; Touretzky, D., Mozer, M., Hasselmo, M., Eds.; MIT Press: Cambridge, MA, USA, 1995; Volume 8, pp. 1073–1079. [Google Scholar]
  28. Madden, G.J.; Mahmoudi, S.; Brown, K. Pavlovian learning and conditioned reinforcement. J. Appl. Behav. Anal. 2023, 56, 498–519. [Google Scholar] [CrossRef]
  29. Perkins, T.; Barto, A. Lyapunov design for safe reinforcement learning control. In Safe Learning Agents: Papers from the 2002 AAAI Symposium, Palo Alto, CA, USA, 25–27 March 2002; AAAI Press: Menlo Park, CA, USA, 2002; pp. 23–30. [Google Scholar]
Figure 1. A robot manipulator in contact with a rigid plane follows the desired trajectory while exerting a desired force on it.
Figure 1. A robot manipulator in contact with a rigid plane follows the desired trajectory while exerting a desired force on it.
Robotics 14 00111 g001
Figure 2. TD error performance: (a) the ITD error converges faster to the origin, (b) the TD error converges in approximately 0.02 s.
Figure 2. TD error performance: (a) the ITD error converges faster to the origin, (b) the TD error converges in approximately 0.02 s.
Robotics 14 00111 g002
Figure 3. Tracking errors: (a) position tracking error convergence for each joint, Δ q i ; (b) velocity tracking error convergence for each joint, Δ q ˙ i , for i = 1 , 2 , 3 .
Figure 3. Tracking errors: (a) position tracking error convergence for each joint, Δ q i ; (b) velocity tracking error convergence for each joint, Δ q ˙ i , for i = 1 , 2 , 3 .
Robotics 14 00111 g003
Figure 4. Tracking errors: (a) force tracking error convergence; (b) real vs. desired Cartesian trajectory.
Figure 4. Tracking errors: (a) force tracking error convergence; (b) real vs. desired Cartesian trajectory.
Robotics 14 00111 g004
Figure 5. Error manifolds: (a) orthogonalized extended velocity errors, S r i , (b) sliding position manifold, S q q i for i = 1 , 2 , 3 .
Figure 5. Error manifolds: (a) orthogonalized extended velocity errors, S r i , (b) sliding position manifold, S q q i for i = 1 , 2 , 3 .
Robotics 14 00111 g005
Figure 6. (a) Sliding force manifold, S q F , (b) control signal performance, τ i , for i = 1 , 2 , 3 .
Figure 6. (a) Sliding force manifold, S q F , (b) control signal performance, τ i , for i = 1 , 2 , 3 .
Robotics 14 00111 g006
Figure 7. Adaptation of neural weights for the Critic NN (a), and the Actor NN (b).
Figure 7. Adaptation of neural weights for the Critic NN (a), and the Actor NN (b).
Robotics 14 00111 g007
Table 1. Metrics obtained from our proposal and the neurocontroller for contact tasks.
Table 1. Metrics obtained from our proposal and the neurocontroller for contact tasks.
Control LawPerformance Metrics
ITAE pos ITAE vel IAC IACV
Our Proposal4.796 × 10−49.194 × 10−2258.8142.9
Neurocontroller7.142 × 10−49.246 × 10−2259151.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pantoja-Garcia, L.; Parra-Vega, V.; Garcia-Rodriguez, R. Physical Reinforcement Learning with Integral Temporal Difference Error for Constrained Robots. Robotics 2025, 14, 111. https://doi.org/10.3390/robotics14080111

AMA Style

Pantoja-Garcia L, Parra-Vega V, Garcia-Rodriguez R. Physical Reinforcement Learning with Integral Temporal Difference Error for Constrained Robots. Robotics. 2025; 14(8):111. https://doi.org/10.3390/robotics14080111

Chicago/Turabian Style

Pantoja-Garcia, Luis, Vicente Parra-Vega, and Rodolfo Garcia-Rodriguez. 2025. "Physical Reinforcement Learning with Integral Temporal Difference Error for Constrained Robots" Robotics 14, no. 8: 111. https://doi.org/10.3390/robotics14080111

APA Style

Pantoja-Garcia, L., Parra-Vega, V., & Garcia-Rodriguez, R. (2025). Physical Reinforcement Learning with Integral Temporal Difference Error for Constrained Robots. Robotics, 14(8), 111. https://doi.org/10.3390/robotics14080111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop