Next Article in Journal
The Chemistry of Polydopamine Film Formation: The Amine-Quinone Interplay
Previous Article in Journal
Fiber Embroidery of Self-Sensing Soft Actuators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Control Synergies for Rapid Stabilization and Enlarged Region of Attraction for a Model of Hopping

Robotics and Motion Laboratory, Department of Mechanical Engineering, The University of Texas at San Antonio, One UTSA Circle, San Antonio, TX 78249, USA
*
Authors to whom correspondence should be addressed.
Biomimetics 2018, 3(3), 25; https://doi.org/10.3390/biomimetics3030025
Submission received: 29 July 2018 / Revised: 30 August 2018 / Accepted: 1 September 2018 / Published: 6 September 2018

Abstract

:
Inspired by biological control synergies, wherein fixed groups of muscles are activated in a coordinated fashion to perform tasks in a stable way, we present an analogous control approach for the stabilization of legged robots and apply it to a model of running. Our approach is based on the step-to-step notion of stability, also known as orbital stability, using an orbital control Lyapunov function. We map both the robot state at a suitably chosen Poincaré section (an instant in the locomotion cycle such as the mid-flight phase) and control actions (e.g., foot placement angle, thrust force, braking force) at the current step, to the robot state at the Poincaré section at the next step. This map is used to find the control action that leads to a steady state (nominal) gait. Next, we define a quadratic Lyapunov function at the Poincaré section. For a range of initial conditions, we find control actions that would minimize an energy metric while ensuring that the Lyapunov function decays exponentially fast between successive steps. For the model of running, we find that the optimization reveals three distinct control synergies depending on the initial conditions: (1) foot placement angle is used when total energy is the same as that of the steady state (nominal) gait; (2) foot placement angle and thrust force are used when total energy is less than the nominal; and (3) foot placement angle and braking force are used when total energy is more than the nominal.

Graphical Abstract

1. Introduction

Stability is of paramount importance for successful deployment of legged robots. However, balance and control of legged robots is a difficult problem resulting from multiple complexities: system nonlinearity, naturally unstable dynamics, limited foot–ground interaction, and discretely changing dynamics due to support transfer [1,2,3,4,5,6,7,8,9]. There are generally two definitions of stability in legged robots. One is the robot’s ability to not fall down and the other is the robot’s ability to follow a given reference trajectory. In this paper, we consider the second definition. Metrics such as viability and gait sensitivity norm (GSN) evaluate the ability to not fall down. The set of all states that a legged robot can experience and avoid falling down is referred to as viability kernel [10], but finding the viability kernel is intractable for high dimensional systems. Using subsets of the viability kernel such as N-step capture regions based on a longer preview of steps [11] and based on optimal step location and timing adjustments for the subsequent step [12] has been shown to create robust walking gaits for simple models (e.g., linear inverted pendulum model). The GSN is defined as the two norm of sensitivity of a suitable gait indicator (e.g., step time, step width, and step velocity) to a representative disturbance (e.g., terrain variation, a push/pull to the robot, model parameter mismatch, and sensors) [13]. This metric is straightforward to compute but is sensitive to the choice of a good gait indicator that correlates with falling. Metrics such as region of attraction, the maximum eigenvalue of the limit cycle, and Lyapunov function evaluate the ability to follow a given reference trajectory.
The majority of past works (e.g., [13,14,15]) have considered control stability but only under a narrow range of perturbations. There is a need to consider control techniques that enlarge the region of attraction (range of initial conditions that can be stabilized to a given steady state gait). Enlarging the region of attraction will improve the robustness of the system by increasing the size of the disturbances that the robot can withstand. In addition, improving the rate of stabilization is highly desirable when the system is subject to repeated disturbances (e.g., uneven terrain). In this paper, we use a model of running to investigate how control actions or control inputs can be combined (control synergies) to provide a wide region of attraction while simultaneously achieving exponentially fast stability.
The most well-known and impactful work in the area of running robots was by Raibert [16] who built a series of hydraulic powered monopedal, bipedal, and quadrupedal robots in the early 1980s. All of his robots were controlled using three decoupled control laws: (1) an axial thrust along the stance leg was used for apex height control; (2) a hip torque during stance phase was used for torso stabilization; and (3) foot placement angle during flight phase was used for velocity control. His robots were able to run on flat ground as well as uneven terrain, negotiate stairs, increase and decrease their speeds as desired, and even perform somersaults.
A widely used model of running is the spring loaded inverted pendulum (SLIP) that consists of a point mass body and a springy leg. The SLIP model has been shown to be a good descriptive model for the center of mass trajectories for runners as diverse as humans, horses, cockroaches and crabs [17]. In a pure SLIP model, the only control variable is the foot placement angle. Foot placement angle may be controlled to achieve a wide range of stable solutions [18]. Furthermore, a slightly backward motion of the swing leg (also known as swing leg retraction (SLR)) just before touchdown has been shown to improve robustness to changes in terrain height [19]. Moreover, recent results have shown that a slightly forward motion of the swing leg (also known as swing leg protraction (SLP)) just before touchdown can impart gait stability under certain conditions [20]. However, since the SLIP model is conservative (total energy is conserved), any perturbation that causes the robot state to deviate from the nominal total energy, cannot be rejected. Therefore, foot placement angle is only able to reject a very narrow set of perturbations, those that lie on the total energy curve of the nominal gait.
To increase the range of initial conditions that can be stabilized in the SLIP model, there needs to be mechanisms to add or remove energy from the system. Shemer and Degani [21] used SLR for foot placement angle and change in leg length for adding/removing energy for rough terrain hopping. Ernst et al. [22] used foot placement angle and spring stiffness, the latter of which is analogous to change in leg length, to improve robustness. Andrews et al. [23] used a fixed impulse in stance phase to add energy, change in the nominal length of the spring during stance to remove energy, and swing leg retraction for rough terrain running. These results suggest that it is important to have means of energy removal and addition besides control of foot placement angle for robust running.
Periodic or steady state motion of legged systems, also referred to as a limit cycle, is evaluated by finding the fixed point using a Poincaré section. A Poincaré section is an instant in the motion (e.g., support transfer, mid-stance, and apex) and the fixed point is the initial condition at the Poincaré section that maps onto itself after a single step. The stability of the limit cycle is given by the largest eigenvalue of the Jacobian of the fixed point [24]. An eigenvalue less than 1 indicates a stable limit cycle and an eigenvalue equal to or greater than 1 indicates an unstable limit cycle. That is, in the former, small perturbations will die out in a few steps, while, in the latter, small perturbations will grow leading to system failure. The eigenvalue-based approach has been extensively used to analyze the stability of passive as well as actively powered models of walking and running [14,25].
One could also design a controller such that the eigenvalue is less than 1 using open-loop control [26] and/or feedback control [27,28]. The issue with this approach is that the eigenvalues are based on linearization of the system dynamics and works well only for small perturbations. To increase the range of initial conditions that may be stabilized a Lyapunov function based method is more effective [29]. Tedrake [30] used convex optimization to find multiple Lyapunov functions, each of which funnels the robot state from the goal state to nominal limit cycle. This method is known as sum of squares optimization [31] and involves finding coefficients of a suitable quadratic function such that the Lyapunov function decays with time.
The rate of convergence to the fixed point is of practical importance because a faster convergence leads to a more stable system. Fast convergence is particularly important when the system is subjected to persistent disturbances (e.g., rough terrain locomotion) where slow convergence may lead to destabilization of the system. Eigenvalue-based approaches produce asymptotic stability [32], which may be too slow. The fastest possible convergence is a one-step dead-beat stabilization in which perturbations are nullified in a single step [33]. Carver et al. [34] demonstrated that the number of steps needed for dead-beat stabilization depends on the number of goals (e.g., forward velocity and motion direction) and number of control actions. If there are n goals and m control actions such that m n , then it is possible to cancel the effect of perturbations in a single step. Zamani and Bhounsule [35] used a control Lyapunov function to ensure exponential rate of convergence between steps for a model of walking. Bhounsule and Zamani [36] found that one-step dead-beat stabilization is more sensitive to modeling errors than exponential stabilization because, in the former, modeling errors lead to overcorrection thereby leading to instability.
The outline of this paper is as follows. The main motivation and novelty of the research is described in detail in Section 2. Next, the model of running is presented in Section 3. In Section 4, we describe necessary tools for analysis of periodic running gaits and use multiple control actions to enlarge the range of stabilizable initial conditions while achieving exponential convergence. Section 5 presents simulation results. A discussion of results is in Section 6 followed by the conclusions in Section 7.

2. Biological Relevance and Novelty

As biological organisms have more muscles (actuators) than mechanical degrees of freedom (joints), there are infinitely many ways of performing a given task. This is famously known as the degrees of freedom problem (see Bernstein [37]). There is evidence that the central nervous system simplifies control by constraining muscles to be activated in fixed groups [38]. These groups are called synergies. Each synergy is defined as a set of muscles recruited by a single neural command signal. It has been found that four synergies can explain postural control in cats [39], five synergies are used by spinal cord injury patients to reproduce foot kinematics of able-bodied people [40], human walking and running have the same five synergy patterns and a shift in the temporal activation of one of the synergies distinguishes walking from running [41].
The present work on robotic running is inspired from synergies in biology. Our work combines biologically inspired synergies with optimal control and feedback control theory. It has been shown that a sufficient condition for balance and stability of legged robots is to adequately bound the position and velocity of the center of mass between steps [42]. We encapsulate this notion of stability by bounding a Lyapunov function (a scalar) that is a weighted quadratic sum of position and velocity of the center of mass at a suitable instant of time in the locomotion cycle (e.g., mid-stance, apex and support transfer). Next, we find control actions (e.g., foot placement angle and axial forces in the leg) that can influence the Lyapunov function. We exploit the observation that there are multiple control actions that can modulate the single output, the Lyapunov function. This leads to redundancy. That is, there are infinitely many control actions that can achieve a given change in Lyapunov function between steps. We impose additional constraints (e.g., rate of stabilization) and minimize an objective function (e.g., minimize energy) to find the optimal combination of control actions, which are then used to identify control synergies.
This paper provides a new perspective on the role of control inputs in enhancement of performance (stability and energy optimality). We demonstrate how to find simple control synergies that enable energy efficient but rapid stabilization of perturbations. These synergies enlarge the region of attraction, which is an important consideration for hardware deployment.

3. Model

3.1. Running Model

Figure 1 shows a model of running. The model consists of a point mass body with mass m = 80 kg and maximum leg length 0 = 1 m. Gravity points downwards and is denoted by g = 9 . 81 m/s 2 . There is a prismatic actuator that can generate an axial force F along the leg and a hip actuator that can place the swing leg at an angle θ instantaneously fast. Although the hip actuator does not affect the dynamics of the swing leg it does affect the dynamics of the center of mass by controlling the foot placement angle during support transfer.

3.2. Equations of Motion

The states of the model are given by { x , x ˙ , y , y ˙ } where x , y are the x- and y-position of the center of mass and x ˙ , y ˙ are the respective velocities. A single step of the runner is given by the following equation:
F l i g h t a p e x F l i g h t t o u c h d o w n S t a n c e c o m p r e s s i o n m i d s t a n c e S t a n c e r e s t i t u t i o n t a k e o f f F l i g h t o n e s t e p / p e r i o d o n e l i m i t c y c l e a p e x F l i g h t
We describe Equation (1) in detail next. The model starts at the apex where the state vector is { x i , x ˙ i , y i , 0 } . The model then falls under gravity,
x ¨ = 0 , y ¨ = g
until contact with the ground is detected by the condition y 0 cos ( θ ) = 0 , where θ is the foot placement angle measured relative to the vertical. Thereafter, the ground contact interaction is given by
m x ¨ = F x , m y ¨ = F y m g ,
where x and y are taken relative to the contact point and F > 0 is the linear actuator force along the leg. For the first half of the stance phase from touchdown to mid-stance (defined by y ˙ = 0 ; note that the event y ˙ = 0 is different from the event corresponding to full leg compression, which is given by ˙ = 0 ), referred to as the compression phase, the actuator force is a braking force F = F b = P b + k ( 0 ) . For the second half of the stance phase from mid-stance to take-off, referred to as the restitution phase, the actuator force is a thrust force F = F t = P t + k ( 0 ) . P b and P t are constant control forces during compression and restitution, respectively. In Equation (3), = x 2 + y 2 is the instantaneous leg length measured relative to the contact point and k is the constant (fixed) gain analogous to the spring constant. In all simulations we take k = 32 , 000 N/m.
The take-off phase is when the leg is fully extended, that is, 0 = 0 . Thereafter, the mass has a flight phase and ends up in the next apex state, { x i + 1 , x ˙ i + 1 , y i + 1 , 0 } .

4. Methods

4.1. Step-to-Step Analysis Using Poincaré Map

We use the step-to-step map referred to as the Poincaré map to build the stabilizing controller. The Poincaré map F is a function that maps the state from an instant in the locomotion cycle to itself after one step. The Poincaré map in this paper is defined at the apex and is given by the condition y ˙ = 0 . Since x i does not affect the Poincaré map (x is monotonically increasing), the state at the apex is given by x i = { x ˙ i , y i } . Given the state at the apex, x i , and the control, u i = { θ , P b , P t } , we compute the state at the next step
x i + 1 = F ( x i , u i ) .
The nominal limit cycle is found by fixing x i + 1 = x i = x 0 and searching for u i = u 0 = { θ , 0 , 0 } (we assume braking and thrust forces are zero) such that
x 0 = F ( x 0 , u 0 ) .

4.2. Control Synergies for Enlarging the Region of Attraction

Control of foot placement angle has been widely used as a strategy for control of running gaits under perturbations [17,43]. However, using only foot placement control will limit the range of perturbation that can be stabilized (see Section 5). To overcome this limitation, we use control synergies.

4.2.1. Key Ideas Behind Control Synergies

A sufficient condition for balance and stability of legged robots is to adequately bound the position and velocity of the center of mass between steps [42]. To bound the position and velocity of the center of mass, we define a Lyapunov function, V ( x ) . The Lyapunov function is defined only at an instant of the locomotion cycle (e.g., mid-stance, just before support transfer) and can be controlled using control actions during the step (e.g., foot placement angle, push-off amplitude). In essence, we have defined a control Lyapunov function V ( x , u ) . Using parameter optimization, we find control actions u that can achieve a desired behavior of the Lyapunov function (e.g., boundedness, asymptotic stability, and exponential stability) over a wide range of initial conditions. Finally, we group the control actions based on the initial conditions, to find the control synergies.
In this paper, the Lyapunov function is defined at the apex and is the weighted sum of squares of the horizontal velocity and the height. The control actions are the two constant forces P c and P r and foot placement angle θ . Our objective is to find the control actions that lead to exponential convergence to the fixed point. This has two benefits: (1) rapid convergence to the limit cycle which enables quick transitions between limit cycles to create aperiodic gaits [44,45]; and (2) high robustness to persistent external disturbances (e.g., running over rough terrain). By visualizing the control actions as a function of initial conditions, we find three distinct control synergies over the range of initial conditions.

4.2.2. Exponential Convergence Using Orbital Control Lyapunov Function

We define a Lyapunov function for the ith limit cycle as follows:
V ( Δ x i ) = ( Δ x i ) T S 0 Δ x i = ( x i x 0 ) T S 0 ( x i x 0 ) ,
where the positive definite matrix S 0 = d i a g { s 11 , s 22 } and the superscript T denotes transpose. The condition for exponential stabilization is
V ( Δ x i + 1 ) V ( Δ x i ) α V ( Δ x i ) ,
where 0 < α < 1 is the rate of decay of the Lyapunov function between steps. Thus, the condition for exponential stability can be rewritten in terms of control as follows:
V ( Δ x i + 1 ) ( 1 α ) V ( Δ x i ) 0 , ( x i + 1 x 0 ) T S 0 ( x i + 1 x 0 ) ( 1 α ) ( x i x 0 ) T S 0 ( x i x 0 ) 0 , F ( x i , u i ) x 0 T S 0 F ( x i , u i ) x 0 ( 1 α ) ( x i x 0 ) T S 0 ( x i x 0 ) 0 .
Equation (8) is the condition on the control Lyapunov function for exponential orbital stabilization (step-to-step stabilization) which is different from exponential local stabilization [46,47]. Specifically, we select u i such that the above condition is met. The variable α is set to 0.9 in all simulations.

4.2.3. Region of Attraction

The region of attraction (ROA) of the controller is the set of all initial conditions x i that would converge to the nominal limit cycle, x 0 . In our case, we are interested in all x i for which we can find u i such that Equation (8) is satisfied. To find the ROA for a given limit cycle, we need to find the level set, ( x i x 0 ) T S 0 ( x i x 0 ) = c such that Equation (8) is held. A small value of constant c leads to a small region of attraction but a large value of c leads to a higher chance of reaching the actuator and/or kinematic limits. As a compromise, we restrict to c 1 . The algorithm is given in Algorithm 1.
Algorithm 1: ROA( x 0 )
  Input: fixed point x 0
  Output: Initial conditions x i ’s
 1 FIND( S 0 ) such that ( x i x 0 ) T S 0 ( x i x 0 ) = 1 intersects y = 0 .
 2 foreach c ( ϵ , 1 ) do // ϵ is a small positive number
 3  COMPUTE( x i ’s) on the level set ( x i x 0 ) T S 0 ( x i x 0 ) = c .
 4  FIND( u i ) for each x i by solving optimization problem described by Equation (10).
 5  if u i then
 6    continue
 7  else
 8    break
 9  end
 10 end
The state constraint y = 0 in Algorithm 1 is a conservative estimate that defines the feasibility of running. That is, if apex height y 0 then there is no flight phase assuming that the leg is vertical at the apex.

4.2.4. Using Optimization for Exponential Stabilization

We formulate an optimization problem with the objective of minimizing an energy metric and the exponential convergence given by Equation (8) as an inequality constraint. The energy metric is referred to as the mechanical cost of transport (MCOT), defined as the energy used per unit weight per unit distance traveled [48,49]
MCOT = MCOT θ + MCOT b + MCOT t = E θ m g D + E b m g D + E t m g D = | k ( 0 ) ˙ | d t m g D + | P b ˙ | d t m g D + | P t ˙ | d t m g D ,
where | x | is the absolute value of x, D is the step length, and ˙ = x x ˙ + y y ˙ . The absolute value is a non-smooth function of its argument, so we smooth it using square root smoothing [50]. That is, | x | = x 2 + ϵ 2 where ϵ is a small number (we set ϵ = 0.01 ).
The optimization problem is defined as follows:
minimize u i MCOT subject   to : x i + 1 = F ( x i , u i ) F ( x i , u i ) x 0 T S 0 F ( x i , u i ) x 0 ( 1 α ) ( x i x 0 ) T S 0 ( x i x 0 ) 0 .

5. Results

First, we show that using only foot placement control will limit the range of perturbation that can be stabilized. To this end, we set P b = P t = 0 in the model shown in Figure 1. Thus, the axial force in the leg is only the actuated spring force F = k ( 0 ) and the only control variable is u i = θ i . We set the fixed point to x 0 = { x ˙ 0 , y 0 } = { 5 , 1 . 3 } and find a control u 0 = θ 0 such that Equation (5) is satisfied. We use numerical integration ode113 in MATLAB (MathWorks, Natick, MA, USA) to find the step-to-step map F . Then, Equation (5) is solved numerically using fsolve in MATLAB, which is a function that finds the zeros of the given function. We obtained a foot placement angle of θ 0 = 0 . 3465 rad.
Figure 2 shows a plot of the vertical height at the apex ( y i ) versus the horizontal velocity at apex ( x ˙ i ). The fixed points x 0 is shown as point (a). Since the system is conservative (no dissipation), each fixed point lies on a constant total energy (TE) line that is found using the sum of potential and kinetic energy at the apex, T E i = 0 . 5 m x ˙ i 2 + m g y i . The constant energy line corresponding to the fixed point is shown as TE 0 . Because the system is conservative, only initial states on the constant total energy line converge back to the fixed point. The initial states on the constant TE 0 line that start at a higher height ( y 1 > y 0 ) and a lower speed ( x ˙ 1 < x ˙ 0 ) compared to nominal (such as point (b)), can converge to the fixed point by decreasing the foot placement angle ( θ 1 < θ 0 ). The initial states on the constant TE 0 line that start at a lower height ( y 2 < y 0 ) and a higher speed ( x ˙ 2 > x ˙ 0 ) compared to nominal (such as point (c)), can converge to the fixed point by increasing the foot placement angle ( θ 2 > θ 0 ). The initial states not on the constant TE 0 line (such as point (d)) cannot converge to the fixed point because there is no means of changing the total energy of the system. From this analysis, we observe that the foot placement angle is able to convert kinetic energy to potential energy or vice versa while maintaining the total energy of the system. This limits the range of initial conditions that can be stabilized to be on the total energy line corresponding to the fixed point. Thus, the foot placement angle has a very narrow region of attraction.
Next, we show that the control synergies developed in Section 4.2 can enlarge the region of attraction. Figure 3 shows the three control actions as a function of apex state, the height y and the horizontal velocity x ˙ . Each ellipse represents the region of attraction. The total energy line corresponding to the nominal solution is T E 0 = 0 . 5 m x ˙ 0 2 + m g y 0 (black dashed line). This line divides the ellipse into two halves: top right half has higher total energy than the nominal, T E i > T E 0 (where i is any initial condition in the top right half); and bottom left half has lower total energy than the nominal, T E i < T E 0 . Thus, for the top right half, the braking force P b is non-zero and serves to brake or extract energy from the system. Similarly, for the bottom left half, the control strategy is to apply a constant thrust force to add energy to the system. The three control actions perform three distinct roles: foot placement angle cannot change the total energy but can convert potential energy to kinetic energy and vice versa, braking force can only decrease the total energy, and thrust force can only increase the total energy. While each of the individual control actions has a small region of attraction, their combination can substantially increase the region of attraction. This indicates control synergies, that is, control actions can be grouped into distinct groups. Furthermore, the control synergies serve to enlarge the range of initial conditions that can be stabilized.
Figure 4 shows the associated MCOT (see Equation (9)) for individual control actions and the total MCOT. Figure 4a shows the total MCOT. The total MCOT for the fixed point (shown by + sign) is 0 . 39 . The total MCOT is mostly flat along x-axis or the horizontal velocity axis but increases monotonically along y-axis or the vertical height axis. Figure 4b–d plots the contribution of individual control actions while Figure 4a is the sum of Figure 4b–d.

6. Discussion

In this paper, we consider a running model with three control actions: (1) hip actuator for foot placement angle; (2) axial actuator for a constant braking force in the compression phase; and (3) the same axial actuator for a constant thrust force in the restitution phase. We use Poincaré map at the apex of flight phase to create and control periodic running motions. The fixed point at the Poincaré map is characterized by the horizontal speed ( x ˙ ) and vertical height (y). We first show that using only foot placement angle for control leads to a very narrow region of attraction, a curve in the x ˙ y plane. We can increase the region of attraction substantially to an ellipse by using all three control actions. Furthermore, we can achieve rapid stabilization to the fixed point using exponential stabilization by using a suitable control Lyapunov function.
The fundamental problem of stable control of running gaits can be achieved by bounding the position and velocity of the center of mass [42]. We have captured the essence of balance in a scalar, a Lyapunov function based on apex height and horizontal velocity, which are sufficient number of states that characterize a steady running gait. We have shown that three control actions can affect the value of the Lyapunov function. Thus, there are infinitely many combinations of control actions that can achieve a given value for the Lyapunov function. We choose control actions that not only achieve a given rate of decay of initial condition but also achieve low energetic cost. The optimization reveals three distinct control synergies depending on the initial conditions: (1) foot placement angle is used when the total energy is the same as that of the steady state (nominal) gait; (2) foot placement angle and thrust force are used when the total energy is less than the nominal; and (3) foot placement angle and braking force are used when the total energy is more than the nominal. These synergies are analogous to biological synergies wherein groups of muscles are activated in synchrony to achieve coordination and balance.
A special case of the running model presented here is the SLIP model, which has a linear spring along the leg and with foot placement angle as the only control variable. The SLIP model is often used as a template to create running controllers for robots (e.g., ATRIAS [51]). However, since the model conserves energy, it has limited ability to reject exogenous disturbances and perturbations (see Figure 2). By adding means to remove and add energy to the system using constant forces in the compression and restitution phases, we are able to substantially increase the range of initial conditions that may be stabilized. The three control actions—foot placement angle, braking force, and thrust force—serve different functions. The foot placement angle θ converts potential energy into kinetic energy or vice versa without changing the total energy, the braking force P b decreases the total energy of the system, and the thrust force P t increases the total energy of the system. When these control actions are taken individually, they can control a limited set of initial conditions. For example, the foot placement angle cannot change the total energy or braking force cannot increase the total energy. However, when these control actions are combined, they can substantially increase the range of initial conditions that can be stabilized, as demonstrated here.
Raibert gave simple decoupled control laws for the control of running gaits [16]. It involved using foot placement angle to control horizontal speed, axial force for height control, and hip torque during stance for body attitude control. The foot placement and force control actions in our study are analogous to those in Raibert’s control laws. However, instead of intuitive tuning, we use energy optimization to discover and tune the control actions. Our control approach is generic and other optimization metrics such as speed, effort, jerk, etc., or their combination can be seamlessly incorporated.
We have considered exponential orbital stabilization with a convergence rate such that perturbations in initial condition decay by a factor of 10 ( α = 0 . 9 in Equation (8)). However, the fastest convergence is achieved by correcting perturbations in a single step, also known as one-step dead-beat stabilization [33]. However, we have found that a one-step dead-beat stabilization is sensitive to modeling errors and hence we do not advocate using it [36]. This happens because one-step dead-beat control can overcorrect in the presence of modeling errors, leading to instability.
Our work has limitations which we highlight next. Although we have used the three control actions—foot placement angle, braking force, and thrust force—there are other control actions that may lead to different synergies and results. For example, some other control actions are: rate of swing leg retraction, free length of the simulated spring, and spring constant. Using visualization of the control actions for a given convergence rate, we can understand the role of different control actions in stabilizing running gait. However, for a different set of control actions, the relations might be more complicated. In this case, it should be possible to use machine learning and/or neural networks to find hidden relations and structure of the control synergies. Each trajectory optimization for the simple model takes about 15 s to complete, which is too slow for online implementation. One simple strategy would be to save all initial conditions and corresponding control actions as a look up table and use it for online implementation. A better approach, particularly useful for a large number of fixed points, would be to fit a control law (e.g., quadratic function, neural networks, etc.) for each control action to simplify online implementation [52,53,54]. We have ignored actuator limits and kinematic limits, which will restrict the region of attraction to smaller ellipses, particularly at faster speeds.

7. Conclusions

Here, we demonstrate how control synergies can be developed in a model of running: (1) to aid energy-efficient and exponential stabilization; and (2) to enlarge the range of initial conditions that can be stabilized. This is significant because past approaches have considered asymptotic stabilization, which is considerably slower than exponential stabilization, and for a narrow range of initial conditions. Our conclusion is that control synergies provide a simple, effective, and convenient means of representing control strategies that would improve stability and increase the agility of legged robots.

Author Contributions

Conceptualization, A.Z. and P.A.B.; Funding acquisition, P.A.B.; Methodology, A.Z. and P.A.B.; Software, A.Z. and P.A.B.; Validation, A.Z. and P.A.B.; Formal analysis, A.Z. and P.A.B.; Writing—original draft preparation, A.Z. and P.A.B.; Writing—review and editing, A.Z. and P.A.B.

Funding

This research was funded by the National Science Foundation (NSF) grant IIS 1566463 to P.A.B.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pratt, J. Exploiting Inherent Robustness and Natural Dynamics in the Control of Bipedal Walking Robots. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2000. [Google Scholar]
  2. Zamani, A.; Khorram, M.; Moosavian, S.A.A. Dynamics and Stable Gait Planning of a Quadruped Robot. In Proceedings of the 11th International Conference on Control, Automation and Systems (ICCAS), Gyeonggi-do, Korea, 26–29 October 2011; pp. 25–30. [Google Scholar]
  3. Khorram, M.; Moosavian, S.A.A. Push recovery of a quadruped robot on challenging terrains. Robotica 2017, 35, 1670–1689. [Google Scholar] [CrossRef]
  4. Kasaei, M.; Lau, N.; Pereira, A. An Optimal Closed-Loop Framework to Develop Stable Walking for Humanoid Robot. In Proceedings of the 18th IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC), Torres Vedras, Portugal, 25–27 April 2018; pp. 30–35. [Google Scholar]
  5. Khadiv, M.; Moosavian, S.A.A.; Yousefi-Koma, A.; Sadedel, M.; Mansouri, S. Optimal gait planning for humanoids with 3D structure walking on slippery surfaces. Robotica 2017, 35, 569–587. [Google Scholar] [CrossRef]
  6. Massah, A.; Zamani, A.; Salehinia, Y.; Sh, M.A.; Teshnehlab, M. A hybrid controller based on CPG and ZMP for biped locomotion. J. Mech. Sci. Technol. 2013, 27, 3473–3486. [Google Scholar] [CrossRef]
  7. Moosavian, S.A.A.; Khorram, M.; Zamani, A.; Abedini, H. PD Regulated Sliding Mode Control of a Quadruped Robot. In Proceedings of the 2011 IEEE International Conference on Mechatronics and Automation (ICMA), Beijing, China, 7–10 August 2011; pp. 2061–2066. [Google Scholar]
  8. Faraji, H.; Tachella, R.; Hatton, R.L. Aiming and Vaulting: Spider Inspired Leaping for Jumping Robots. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 2082–2087. [Google Scholar]
  9. Rohani, F.; Richter, H.; Van Den Bogert, A.J. Optimal design and control of an electromechanical transfemoral prosthesis with energy regeneration. PLoS ONE 2017, 12, e0188266. [Google Scholar] [CrossRef] [PubMed]
  10. Wieber, P.B. On the Stability of Walking Systems. In Proceedings of the 3rd IARP International Workshop on Humanoid and Human Friendly Robotics, Tsukuba, Japan, 11–12 December 2002. [Google Scholar]
  11. Koolen, T.; Boer, T.D.; Rebula, J.; Goswami, A.; Pratt, J.E. Capturability-based analysis and control of legged locomotion. Part 1: Theory and application to three simple gait models. Int. J. Robot. Res. 2012, 31, 1094–1113. [Google Scholar] [CrossRef]
  12. Khadiv, M.; Herzog, A.; Moosavian, S.A.A.; Righetti, L. A robust walking controller based on online step location and duration optimization for bipedal locomotion. arXiv, 2017; arXiv:1704.01271. [Google Scholar]
  13. Hobbelen, D.; Wisse, M. A disturbance rejection measure for limit cycle walkers: The gait sensitivity norm. IEEE Trans. Robot. 2007, 23, 1213–1224. [Google Scholar] [CrossRef]
  14. McGeer, T. Passive dynamic walking. Int. J. Robot. Res. 1990, 9, 62–82. [Google Scholar] [CrossRef]
  15. Hobbelen, D.G.; Wisse, M. Swing-leg retraction for limit cycle walkers improves disturbance rejection. IEEE Trans. Robot. 2008, 24, 377–389. [Google Scholar] [CrossRef]
  16. Raibert, M. Legged Robots That Balance; MIT Press: Cambridge, MA, USA, 1986. [Google Scholar]
  17. Schwind, W.J. Spring Loaded Inverted Pendulum Running: A Plant Model. Ph.D. Thesis, University of Michigan, Ann Arbor, MI, USA, 1998. [Google Scholar]
  18. Seyfarth, A.; Geyer, H.; Günther, M.; Blickhan, R. A movement criterion for running. J. Biomech. 2002, 35, 649–655. [Google Scholar] [CrossRef] [Green Version]
  19. Seyfarth, A.; Geyer, H.; Herr, H. Swing-leg retraction: A simple control model for stable running. J. Exp. Biol. 2003, 206, 2547–2555. [Google Scholar] [CrossRef] [PubMed]
  20. Bhounsule, P.A.; Zamani, A. Stable bipedal walking with a swing-leg protraction strategy. J. Biomech. 2017, 51, 123–127. [Google Scholar] [CrossRef] [PubMed]
  21. Shemer, N.; Degani, A. A flight-phase terrain following control strategy for stable and robust hopping of a one-legged robot under large terrain variations. Bioinspir. Biomim. 2017, 12, 046011. [Google Scholar] [CrossRef] [PubMed]
  22. Ernst, M.; Geyer, H.; Blickhan, R. Extension and customization of self-stability control in compliant legged systems. Bioinspir. Biomim. 2012, 7, 046002. [Google Scholar] [CrossRef] [PubMed]
  23. Andrews, B.; Miller, B.; Schmitt, J.; Clark, J.E. Running over unknown rough terrain with a one-legged planar robot. Bioinspir. Biomim. 2011, 6, 026009. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Strogatz, S. Nonlinear Dynamics and Chaos; Addison-Wesley: Reading, MA, USA, 1994. [Google Scholar]
  25. Coleman, M.J.; Chatterjee, A.; Ruina, A. Motions of a rimless spoked wheel: A simple three-dimensional system with impacts. Dyn. Stab. Syst. 1997, 12, 139–159. [Google Scholar] [CrossRef]
  26. Mombaur, K.; Georg Bock, H.; Schlöder, J.; Longman, R. Stable walking and running robots without feedback. In Climbing and Walking Robots; Springer: Berlin/Heidelberg, Germany, 2005; pp. 725–735. [Google Scholar]
  27. Bhounsule, P.A.; Ameperosa, E.; Miller, S.; Seay, K.; Ulep, R. Dead-Beat Control of Walking for a Torso-Actuated Rimless Wheel Using an Event-Based, Discrete, Linear Controller. In Proceedings of the ASME 2016 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference (IDETC/CIE), Charlotte, NC, USA, 21–24 August 2016; pp. 1–9. [Google Scholar]
  28. Kuo, A.D. Stabilization of lateral motion in passive dynamic walking. Int. J. Robot. Res. 1999, 18, 917–930. [Google Scholar] [CrossRef]
  29. Navabi, M.; Mirzaei, H. Robust optimal adaptive trajectory tracking control of quadrotor helicopter. Latin Am. J. Solids Struct. 2017, 14, 1040–1063. [Google Scholar] [CrossRef]
  30. Tedrake, R. LQR-Trees: Feedback Motion Planning on Sparse Randomized Trees. In Proceedings of the Robotics Science and Systems (RSS), Zaragoza, Spain, 28 June–1 July 2009. [Google Scholar]
  31. Prajna, S.; Papachristodoulou, A.; Parrilo, P.A. Introducing SOSTOOLS: A General Purpose Sum of Squares Programming Solver. In Proceedings of the 41st IEEE Conference on Decision and Control, Las Vegas, NV, USA, 10–13 December 2002; Volume 1, pp. 741–746. [Google Scholar]
  32. Grizzle, J.; Abba, G.; Plestan, F. Asymptotically stable walking for biped robots: Analysis via systems with impulse effects. IEEE Trans. Autom. Control 2001, 46, 51–64. [Google Scholar] [CrossRef]
  33. Antsaklis, P.; Michel, A. Linear Systems; Birkhauser: Basel, Switzerland, 2006. [Google Scholar]
  34. Carver, S.; Cowan, N.; Guckenheimer, J. Lateral stability of the spring-mass hopper suggests a two-step control strategy for running. Chaos 2009, 19, 026106. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Zamani, A.; Bhounsule, P.A. Foot Placement and Ankle Push-Off Control for the Orbital Stabilization of Bipedal Robots. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 4883–4888. [Google Scholar]
  36. Bhounsule, P.A.; Zamani, A. A discrete control lyapunov function for exponential orbital stabilization of the simplest walker. J. Mech. Robot. 2017, 9, 051011. [Google Scholar] [CrossRef]
  37. Bernstein, N. The Co-Ordination and Regulation of Movements; Pergamon Press: London, UK, 1967. [Google Scholar]
  38. Torres-Oviedo, G.; Macpherson, J.M.; Ting, L.H. Muscle synergy organization is robust across a variety of postural perturbations. J. Neurophysiol. 2006, 96, 1530–1546. [Google Scholar] [CrossRef] [PubMed]
  39. Ting, L.H.; Macpherson, J.M. A limited set of muscle synergies for force control during a postural task. J. Neurophysiol. 2005, 93, 609–613. [Google Scholar] [CrossRef] [PubMed]
  40. Ivanenko, Y.P.; Grasso, R.; Zago, M.; Molinari, M.; Scivoletto, G.; Castellano, V.; Macellari, V.; Lacquaniti, F. Temporal components of the motor patterns expressed by the human spinal cord reflect foot kinematics. J. Neurophysiol. 2003, 90, 3555–3565. [Google Scholar] [CrossRef] [PubMed]
  41. Cappellini, G.; Ivanenko, Y.P.; Poppele, R.E.; Lacquaniti, F. Motor patterns in human walking and running. J. Neurophysiol. 2006, 95, 3426–3437. [Google Scholar] [CrossRef] [PubMed]
  42. Pratt, J.E.; Tedrake, R. Velocity-based stability margins for fast bipedal walking. In Fast Motions in Biomechanics and Robotics; Springer: Berlin, Germany, 2006; pp. 299–324. [Google Scholar]
  43. Wu, A.; Geyer, H. The 3-D spring–mass model reveals a time-based deadbeat control for highly robust running and steering in uncertain environments. IEEE Trans. Robot. 2013, 29, 1114–1124. [Google Scholar] [CrossRef]
  44. Bhounsule, P.A.; Zamani, A.; Pusey, J. Switching between Limit Cycles in a Model of Running Using Exponentially Stabilizing Discrete Control Lyapunov Function. In Proceedings of the 2018 American Control Conference (ACC), Milwaukee, WI, USA, 27–29 June 2018; pp. 3714–3719. [Google Scholar]
  45. Motahar, M.S.; Veer, S.; Poulakakis, I. Composing Limit Cycles for Motion Planning of 3D Bipedal Walkers. In Proceedings of the 55th IEEE Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016; pp. 6368–6374. [Google Scholar]
  46. Ames, A.D.; Galloway, K.; Sreenath, K.; Grizzle, J.W. Rapidly exponentially stabilizing control Lyapunov functions and hybrid zero dynamics. IEEE Trans. Autom. Control 2014, 59, 876–891. [Google Scholar] [CrossRef]
  47. Veer, S.; Motahar, M.S.; Poulakakis, I. Generation of and Switching among Limit-Cycle Bipedal Walking Gaits. In Proceedings of the 56th IEEE Conference on Decision and Control (CDC), Melbourne, Australia, 12–15 December 2017; pp. 5827–5832. [Google Scholar]
  48. Zamani, A.; Bhounsule, P.A.; Taha, A. Planning Energy-Efficient Bipedal Locomotion on Patterned Terrain. In Proceedings of the Unmanned Systems Technology XVIII, International Society for Optics and Photonics, Baltimore, MD, USA, 17–21 April 2016; p. 98370A. [Google Scholar]
  49. Zamani, A.; Bhounsule, P.; Hurst, J. Energy-efficient planning for dynamic legged robots on patterned terrain. In Proceedings of the Dynamic Walking Conference, Holly, MI, USA, 4–7 June 2016; p. 1. [Google Scholar]
  50. Srinivasan, M. Why Walk and Run: Energetic Costs and Energetic Optimality in Simple Mechanics-Based Models of a Bipedal Animal. Ph.D. Thesis, Cornell University, Ithaca, NY, USA, 2006. [Google Scholar]
  51. Grimes, J.A.; Hurst, J.W. The design of ATRIAS 1.0 a unique monopod, hopping robot. In Proceedings of the 14th International Conference on Climbing and Walking Robots and the Support Technologies for Mobile Machines (CLAWAR), Baltimore, MD, USA, 23–26 July 2012; pp. 548–554. [Google Scholar]
  52. Bhounsule, P.; Zamani, A.; Krause, J.; Farra, S.; Pusey, J. Control policies for large region of attraction for dynamically balancing legged robots: A sampling-based approach. Robotica 2018. sumbitted. [Google Scholar]
  53. Alambeigi, F.; Khadem, S.M.; Khorsand, H.; Hasan, E.M.S. A comparison of performance of artificial intelligence methods in prediction of dry sliding wear behavior. Int. J. Adv. Manuf. Technol. 2016, 84, 1981–1994. [Google Scholar] [CrossRef]
  54. Alambeigi, F.; Wang, Z.; Hegeman, R.; Liu, Y.H.; Armand, M. A robust data-driven approach for online learning and manipulation of unmodeled 3-D heterogeneous compliant objects. IEEE Robot. Autom. Lett. 2018, 3, 4140–4147. [Google Scholar] [CrossRef]
Figure 1. A complete step for the running model. The model starts in (a) the flight phase at the apex position (vertical velocity is zero), followed by (b,c) the stance phase, and finally ending in (d) the flight phase at the apex position of the next step. The running model has a prismatic actuator that is used to provide an axial braking force F b in the compression phase and axial thrust force F t in the restitution phase, and a hip actuator (not shown) that can place the swinging leg at an angle θ with respect to the vertical as the leg lands on the ground. P b , P t , k, and 0 are constant control forces during compression and restitution, the constant gain, and the maximum leg length, respectively.
Figure 1. A complete step for the running model. The model starts in (a) the flight phase at the apex position (vertical velocity is zero), followed by (b,c) the stance phase, and finally ending in (d) the flight phase at the apex position of the next step. The running model has a prismatic actuator that is used to provide an axial braking force F b in the compression phase and axial thrust force F t in the restitution phase, and a hip actuator (not shown) that can place the swinging leg at an angle θ with respect to the vertical as the leg lands on the ground. P b , P t , k, and 0 are constant control forces during compression and restitution, the constant gain, and the maximum leg length, respectively.
Biomimetics 03 00025 g001
Figure 2. Control of foot placement angle. The plot shows the velocity in the x-direction ( x ˙ i ) versus the vertical height ( y i ) at the Poincaré section, which is at the apex in the flight phase. (a) The fixed point is { x ˙ 0 , y 0 } = { 5 , 1.3 } , which corresponds to the foot placement angle θ 0 = 0.3465 rad. (b) Initial states such as { x ˙ 1 , y 1 } need to decrease the foot placement angle θ 1 to converge to the limit cycle. (c) Initial states such as { x ˙ 2 , y 2 } need to increase the foot placement angle θ 2 to converge to the limit cycle. (d) Initial conditions such as { x ˙ 3 , y 3 } which are not on the TE 0 line cannot converge to the fixed point { x ˙ 0 , y 0 } because there is no means to change the total energy of the system (drawing not shown).
Figure 2. Control of foot placement angle. The plot shows the velocity in the x-direction ( x ˙ i ) versus the vertical height ( y i ) at the Poincaré section, which is at the apex in the flight phase. (a) The fixed point is { x ˙ 0 , y 0 } = { 5 , 1.3 } , which corresponds to the foot placement angle θ 0 = 0.3465 rad. (b) Initial states such as { x ˙ 1 , y 1 } need to decrease the foot placement angle θ 1 to converge to the limit cycle. (c) Initial states such as { x ˙ 2 , y 2 } need to increase the foot placement angle θ 2 to converge to the limit cycle. (d) Initial conditions such as { x ˙ 3 , y 3 } which are not on the TE 0 line cannot converge to the fixed point { x ˙ 0 , y 0 } because there is no means to change the total energy of the system (drawing not shown).
Biomimetics 03 00025 g002
Figure 3. Contour plots for control actions as a function of horizontal velocity and apex height. (a) Foot placement angle; (b) constant braking force in stance phase; and (c) constant thrust force in the restitution phase.
Figure 3. Contour plots for control actions as a function of horizontal velocity and apex height. (a) Foot placement angle; (b) constant braking force in stance phase; and (c) constant thrust force in the restitution phase.
Biomimetics 03 00025 g003
Figure 4. Contour plots for mechanical cost of transport (MCOT) function of horizontal velocity and apex height. (a) Total MCOT; (b) MCOT due to foot placement angle θ ; (c) MCOT due to braking force P b ; and (d) MCOT due to thrust force P t .
Figure 4. Contour plots for mechanical cost of transport (MCOT) function of horizontal velocity and apex height. (a) Total MCOT; (b) MCOT due to foot placement angle θ ; (c) MCOT due to braking force P b ; and (d) MCOT due to thrust force P t .
Biomimetics 03 00025 g004

Share and Cite

MDPI and ACS Style

Zamani, A.; Bhounsule, P.A. Control Synergies for Rapid Stabilization and Enlarged Region of Attraction for a Model of Hopping. Biomimetics 2018, 3, 25. https://doi.org/10.3390/biomimetics3030025

AMA Style

Zamani A, Bhounsule PA. Control Synergies for Rapid Stabilization and Enlarged Region of Attraction for a Model of Hopping. Biomimetics. 2018; 3(3):25. https://doi.org/10.3390/biomimetics3030025

Chicago/Turabian Style

Zamani, Ali, and Pranav A. Bhounsule. 2018. "Control Synergies for Rapid Stabilization and Enlarged Region of Attraction for a Model of Hopping" Biomimetics 3, no. 3: 25. https://doi.org/10.3390/biomimetics3030025

Article Metrics

Back to TopTop