1. Introduction
According to the World Health Organization in 2019 [
1], 2.4 billion people experienced conditions that could benefit from rehabilitation services (1 in 3 impaired people globally). Between 1990 and 2019, there was a 69.4% rise in the population living with disabilities. Musculoskeletal disorders and strokes are among the leading causes of disability worldwide [
2]. These conditions often result in limited functional capacity, particularly in terms of mobility.
One of the most common ways to support mobility recovery in individuals is through rehabilitation services, where physical rehabilitation plays a fundamental role. These services are patient-centered and aim to promote independence. Within these treatments, robotic-assisted rehabilitation has proven to be efficient because this method allows for personalized, precise, and repetitive movements, increasing the intensity of the treatment [
3,
4]. Such devices include exoskeletons, which are robotic systems designed based on bio-inspired models (similar in shape and functions to the human body) that are externally attached to the user [
5,
6].
The main challenges in robotic rehabilitation with exoskeletons are (a) those that come from their physical characteristics and (b) those that come from their application. The former addresses engineering design aspects for the physical components, workspace, and robot kinematic/kinetic models. The latter contemplates the nature of rehabilitation treatments, such as precision (personalized treatments), physical assistance by the physiotherapist (resistive or assisted therapy), and coupling between the robot and the patient (generates highly nonlinear systems with changing dynamic models), among others [
7,
8].
These latter aspects present a significant challenge for developers as they require the implementation of automated control strategies that ensure compliance while also providing security to users. A key aspect in the design of control strategies is the tuning of controller gains. Typically, this tuning is performed heuristically or based on the biomechanical characteristics of a particular patient population. However, this approach presents the following drawbacks: (a) the process is slow and tedious [
8], and (b) its application is limited to a small group of users. Therefore, it is essential to develop approaches for automatic controller gain tuning that account for system complexities, model uncertainties, and external disturbances (such as human–exoskeleton interaction). The following section provides a brief literature review on these approaches.
2. Literary Review: Controllers and Gain Tuning Methods
Due to the nature of exoskeleton applications, it is crucial to implement controllers that ensure optimal performance, as these systems interact directly with the user and are exposed to external disturbances. This review focuses on analyzing the most commonly used controllers in exoskeletons and the techniques employed for gain tuning, covering both classical and more recent approaches.
Table 1 provides a brief overview of several controllers main characteristics, along with examples of studies where they have been applied.
As shown in
Table 1, most controllers combine different approaches to achieve good performance in the application. However, the Proportional–Integral–Derivative (PID) controller stands out as the most widely used due to its simplicity in implementation and tuning, as well as its effectiveness in linear systems. It is followed by sliding mode control (SMC), neural networks (NNs), fuzzy logic (FL), and supertwisting control (ST).
In terms of robustness against uncertainties and disturbances, SMC and ST controllers are particularly noteworthy, with the latter also being more accurate in complex, nonlinear systems. On the other hand, NN- and FL-based approaches are also capable of effectively handling uncertainty and achieve high accuracy, particularly when trained and properly designed. The choice of controller depends on system complexity, the nature of disturbances, and the required level of precision or robustness.
In addition to selecting the type of controller, another crucial factor for its efficiency is the choice of the gain tuning method, as this directly influences the control system stability and performance. For this purpose, various approaches are employed, which differ in complexity and effectiveness.
Table 2 presents some of the most commonly used gain tuning methods, ranging from traditional techniques, such as heuristic tuning, to advanced approaches, including optimization algorithms and machine learning.
It should be noted that in most of the reviewed works, the controller gains are presented as constant values over time, without specifying how these values were assigned. In these studies, only the effectiveness of the proposed controller is demonstrated.
According to
Table 2, for linear systems, methods such as Ziegler–Nichols (the most common for parameter initialization) and tuning by trial and error (either manually or pre-tuned) tend to be effective and easy to implement. However, in nonlinear systems, which are inherently susceptible to uncertainties and disturbances, the most effective methods include heuristic optimization, robust control, and adaptive control. This is because neural networks require considerable time to process each input and target data sample due to the size of each training set.
Over the years, nature has inspired the development of algorithms to solve complex real-world problems. However, there is no universal applicable heuristic method for optimizing the gains of a nonlinear controller in the presence of disturbances, as the effectiveness of each algorithm depends on the type of nonlinear system, the nature of the disturbances, and the specific requirements of the control problem. Some of the most common heuristic methods used for these types of problems include genetic algorithms (GAs), particle swarm optimization (PSO) [
8,
9,
14,
15,
16,
19,
21,
35], simulated annealing (SA), ant colony optimization (ACO), cuckoo search (CS) algorithm, and differential evolution (DE) algorithms.
For optimizing gains in a nonlinear controller in the presence of disturbances, genetic algorithms and PSO are the most widely used and effective due to their flexibility and ability to identify suitable solutions in nonlinear problems with multiple local optima. However, when rapid convergence is required or there are computational limitations, PSO is an excellent choice, as it is an algorithm that is part of swarm intelligence, a significant branch of artificial intelligence [
44,
45]. Conversely, if robustness against the possibility of getting trapped in local optima is desired and more computational time is available, genetic algorithms or differential evolution are more appropriate options.
In the context of exoskeletons for rehabilitation, where there is physical interaction with users, methods that require precise knowledge of the mathematical model are generally not used, especially if the system is of high order and exhibits nonlinearities or external disturbances [
29,
46,
47,
48]. Therefore, in these applications, it is essential to have a method that adjusts the controller parameters in real time, adapting to the evolution of the system and changes in user characteristics. Designing such solutions are complex and may require specialized hardware and software.
As a result, it is necessary to implement control approaches based on metaheuristic optimization that adapt in real time to different users. These algorithms identify efficient configurations for the controller compared to traditional methods. By adopting these robust approaches, the efficiency of rehabilitation tasks is improved.
Optimization-based approaches overcome the drawbacks of those that do not employ it, so the key steps of this approach are outlined below:
Establish an objective function: The aim is to meet specific criteria while considering the associated cost, whether minimum or maximum. This can be achieved through one or more objective functions, or by using a weighted objective function.
Optimizer selection: Direct methods, such as the response surface method and gradient descent, as well as swarm-based methods like evolutionary algorithms (EAs), are employed. Direct methods are sensitive to noise, which can reduce their performance. Additionally, they typically provide solutions based on a single criterion, tend to converge to local optima, and may encounter issues related to an unidentified search space. In contrast, swarm-based methods have demonstrated superior performance compared to these traditional approaches [
8,
44].
Lately, the particle swarm (PS) algorithm has emerged as one of the EAs utilized for optimizing controller parameters in exoskeletons, due to the following charact- eristics [
19,
21,
49]: (1) its simplicity and high convergence speed; (2) it does not need any assumptions about the problem to be optimized; (3) it explores extensive solution spaces; (4) it is capable of handling intricate complex nonlinear systems with a high number of dimensions and challenges in dynamic optimization across diverse domains; (5) it does not demand a precise dynamic model or accurate system parameters beforehand; (6) a limited number of parameters require tuning; and (7) it exhibits a high likelihood and efficiency in discovering global optima, among others.
Despite the merits of the PS algorithm, it also has some disadvantages [
15,
16,
19,
21]: (1) the time required for convergence might be extended, potentially leading to undesirable chattering phenomena in a control context; (2) the potential problem of early convergence could hinder the ability to adapt to sudden changes in the system and the environment; and (3) there is a risk of early convergence, leading to entrapment in local optima, among others.
In this case, due to the application objectives, the PS algorithm was chosen to optimize the gains of the sliding mode controller. This technique is a robust and well-established control method, designed to handle uncertain dynamics, provide a rapid response to changes, offer robustness against parameter variations, and facilitate implementation [
23,
27,
28,
30]. However, its main disadvantage is the occurrence of oscillations. To mitigate this problem and reduce convergence time, an alternative approach is proposed that incorporates an exponential function as the reaching law, which preserves the robustness of SMC [
28,
31,
50,
51]. This enables the controller to adapt in real time to changes in the morphology and motor abilities of exoskeleton users.
To the best of our knowledge, the main contributions of this study involve the following:
Online gain tuning using the exponential reaching law (ERL) and the PSO algorithm on a highly nonlinear system is proposed for the first time.
This is the first real-time implementation of the robust PSO-ERL algorithm on a 7 DOF exoskeleton for online gain tuning for trajectory tracking.
It presents the findings from the experimental assessment of gain tuning among healthy subjects exhibiting diverse physiological characteristics.
The work is structured as follows:
Section 3 covers the PSO algorithm’s mathematical aspects, the exoskeleton’s dynamical model (nonlinear considering external disturbances), and the controller (features, stability analysis, etc.).
Section 4 details the implementation conditions, experimental results, and analyses performed. The last section concludes and suggests several future study directions.
4. Results: Real-Time Implementation
This study performs gain tuning ( and ) of the ERL controller on a highly nonlinear exoskeleton robot, subject to external disturbances, such as initial offsets and user variations. The problem is formulated as an optimization task and addressed using the PSO algorithm.
4.1. Exoskeleton Description
The controller is applied to the ETS-MARSE (École de Technologie Supérieure—Motion Assistive Robotic-exoskeleton for Superior Extremity) with 7 DOF (see
Figure 2). This exoskeleton facilitates rehabilitation by assisting movements for individuals with impaired upper extremities. Its design takes into account the upper extremity alignment, featuring 3 degrees of freedom (DOF) for the shoulder (S), 1 DOF for the elbow (E), and 3 DOF for the wrist (W). The modified Denavit–Hartenberg parameters (D-H) for the exoskeleton are detailed in
Table 3.
The exoskeleton’s real-time system consists of three processing units, detailed as follows:
Host PC (Intel Core i7-4770 CPU @3.4 GHz, and 16 GB RAM): facilitates the transmission of higher-level controls to the exoskeleton via the human–machine interface (developed in LabView 2017).
A real-time PC (NI PXI-8108, Intel dual-core @2.53 GHz processor and 8 GB RAM): operates the top-level control and manages the exoskeleton dynamics with a sampling time of 1 ms.
FPGA (NI PXI-7813R): utilized for handling analog and digital inputs and outputs to the actuators and sensors, executing the low-level control (PI) current control loop with a sampling time of 50 µs.
The robot utilizes brushless DC motors, specifically Maxon EC-45 and Maxon EC-90, in combination with harmonic drives. Motors 1 and 2 have a gear ratio of 120:1, while motors 3–7 have a gear ratio of 100:1.
4.2. Tests Description
In this section, the tests to be conducted are described, with the following considerations:
To demonstrate the efficiency of the PSO algorithm when tuning the ERL controller gains, the following tests were performed:
Assess the ERL performance controller under pre-tuned gains, denoted as PT conditions.
Assess the ERL controller’s real-time performance through online tuning using PSO, denoted as PSO conditions.
Assess the ERL controller’s performance using the globally tuned gains obtained after applying the PSO algorithm, denoted as PSOG conditions.
4.2.1. First Test
The gains were chosen heuristically and theirs values are shown in
Table 5.
4.2.2. Second Test
The ERL controller parameters vary over time and are adaptively tuned online. This flexibility allows for efficient tracking of trajectories, even when confronted with external disturbances like patient forces or physiotherapist assistance. The control scheme is depicted in
Figure 4.
Commonly, to address the optimization problem, a single objective function is used that encompasses all DOF. In this work, a weighted objective function is employed: it includes position and velocity errors for each degree of freedom (DOF), where the position error is prioritized. So, the function targeted as the objective in the PSO algorithm is presented in Equation (
33).
Here, , with and constrained to the range 0 to 1.
In this scenario, each particle comprises the following two gains,
and
. The goal is to dynamically identify the optimal particle in real time, minimizing the objective function
. Therefore Algorithm 1 introduces the proposed robust PSO-ERL algorithm.
Algorithm 1 Proposed robust PSO-ERL algorithm. |
- 1:
Define the number of particles in the swarm, n. - 2:
Initialize the swarm with random values. - 3:
i ← 1 - 4:
repeat - 5:
Insert a particle i (, ) - 6:
Compute using the ERL approach Equation ( 12) - 7:
Calculate using Equation ( 7). - 8:
Utilize as the input for the system. - 9:
Compute the objective function f using Equation ( 33). - 10:
Store the value of f for particle i. - 11:
i ← i + 1 - 12:
until i m - 13:
Determine and . - 14:
- 15:
GOTO 4
|
In this case, Equation (
20a) represents the gain value (
or
), and Equation (
20b) represents the corresponding gain change rate. The particle’s initial values (
and
) are randomly generated within their respective search space, as defined in
Table 6. Subsequently, the objective function is calculated in each iteration to assess each particle, iteratively guiding the PSO-ERL to find the optimal solution. The particle exhibiting the objective function minimum value corresponds to the best global gain.
The remaining parameters for the PSO algorithm and the controller’s experimental conditions are detailed in
Table 7.
To select the parameters in
Table 7, we considered the following:
: Promotes global exploration.
: Global optimum is prioritized.
: Prioritizing position error minimization.
Once the second test was completed (applying the PSO algorithm to determine the gains
and
of the controller), the values of the global gains were obtained for each user, which are shown in
Table 8.
4.2.3. Third Test
It consists in executing the ERL controller with the gain values shown in
Table 8.
So, in the next section, we perform a comparative analysis of the results obtained in the three experiments.
4.3. Experimental Results and Comparative Analysis
Figure 5 shows the results of the trajectory tracking, tracking error, control signals, and cartesian tracking and error graphs of user 1, respectively. Furthermore,
Figure 6 corresponds to user 2. In each figure, there are three signals, each one corresponding to the experiments carried out (PT, PSO, and PSOG).
To evaluate the PSO algorithm’s behavior,
Figure 7 presents the evolution of the values of
K and
of a particle, showing an oscillatory and convergent behavior. Choosing the parameters
,
,
guarantees the PSO algorithm’s deterministic convergence, whose eigenvalues of Equation (
26) are equal to
.
According to the results obtained, the average RMSE (root mean square error) for both users in the first test was and the average RMSE for the same users in the third test was . This shows that the proposed approach reduces the average RMSE of user 1 by 28.3% and by 23.5% for user 2.
It should be noted that chattering is not eliminated entirely, however, the average RMST (root mean square torque) for the same users in the first test was and for the third test was . It is concluded that the RMST in the third experiment increased by 0.4% and 1.6% compared to the first test, which means a small increase in the control activity while greatly improving the tracking error.
Table 9 and
Table 10 show the users’ results for the RMSE and RMST of each joint, as well as the average for the first and third tests.
Additionally, it is worth noting that the elapsed time for the first and third experiments was around 238 µs, while for the second experiment it was around 263 µs, representing an 11% increase.
5. Discussion
Rehabilitation exoskeletons are high-order systems with nonlinearities and are subject to external disturbances [
29,
46,
47,
48]. To optimize their performance, it is essential to develop controllers that account for user interaction and allow real-time dynamic adjustment of their parameters.
In most studies, controllers use constant gains that ensure adequate performance in experimental tests, enabling trajectory tracking adapted to a specific group of users [
23,
54,
55,
56]. However, online adaptive control is a promising alternative for improving system response to disturbances [
12,
15,
19,
21,
29,
55,
57].
This study implements a robust control approach with dynamic gain adjustment, based on mathematical formulations that consider modeling uncertainties and disturbances [
23,
27,
28,
30]. A sliding mode controller with an exponential reaching law was used [
31,
50], which simultaneously reduced chattering and convergence time without compromising system robustness. Additionally, it provides a quick response to system changes, resistance to parameter variations, and ease of implementation.
To optimize the gains of the nonlinear controller in the presence of disturbances, the PSO algorithm was employed due to its ability to find optimal solutions in problems with multiple local minima, its low computational cost, and its adaptability to various applications [
15,
44,
45].
To the best of our knowledge, this is the first study to experimentally apply this method to a 7 DOF exoskeleton for online gain adjustment in a sliding mode controller. The results obtained with healthy subjects of different morphologies showed an improvement of over 20% in trajectory tracking, along with a significant reduction in chattering.
This work presents a promising solution for tuning robust controllers in dynamically coupled systems. However, controller performance depends on the correct selection of gains within the search space, defined according to the system’s physical characteristics. If the exoskeleton lacks dynamic stability and energy efficiency, the algorithm alone will not guarantee significant improvements.