Next Article in Journal
Dynamic Study on a Passive Damping Scheme for Permanent Magnet Electrodynamic Suspension Vehicle Utilizing Onboard Magnets End Effects
Next Article in Special Issue
Weight-Incorporating A* Algorithm with Multi-Factor Cost Function for Enhanced Mobile Robot Path Planning
Previous Article in Journal
Comprehensive Comparative Analysis of Lower Limb Exoskeleton Research: Control, Design, and Application
Previous Article in Special Issue
Robotic Wound Closure: Detection and Control of Patient Motion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model Predictive Control for Pneumatic Manipulator via Receding-Horizon-Based Extended State Observers

1
School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
2
School of Artificial Intelligence, Tiangong University, Tianjin 300387, China
3
Xingyu Electronics (Ningbo) Co., Ltd., Ningbo 315500, China
4
Tianjin Key Laboratory of Intelligent Unmanned Swarm Technology and System, Tianjin 300072, China
*
Author to whom correspondence should be addressed.
Actuators 2025, 14(7), 343; https://doi.org/10.3390/act14070343
Submission received: 13 May 2025 / Revised: 4 July 2025 / Accepted: 8 July 2025 / Published: 10 July 2025
(This article belongs to the Special Issue Actuators in Robotic Control—3rd Edition)

Abstract

This paper presents a model predictive control (MPC)-enabled disturbance-rejection controller approach for a pneumatic manipulator system subjected to complex nonlinear terms within the system. To facilitate the handling of the complex nonlinear terms, they are modeled as disturbances. To address these disturbances, a receding-horizon-based extended state observer (RH-ESO) incorporating a decision variable is developed. The optimal disturbance estimation error is determined through a receding-horizon optimization procedure, which provides the best estimate of the disturbance. Using this optimal estimate, the MPC-enabled disturbance-rejection controller is proposed for the pneumatic manipulator system to achieve angle tracking control. Moreover, the proposed approach ensures both the recursive feasibility of the optimization problem and the uniform boundedness of the closed-loop system. The simulation results further demonstrate the effectiveness and validity of the proposed methodology.

1. Introduction

The research on pneumatic soft robotic arms has accelerated with the recent progress in flexible actuation [1]. Owing to their intrinsic compliance, these arms are now prevalent in soft manipulation [2], bio-inspired robotics [3], and industrial automation [4]. Their primary actuators—pneumatic artificial muscles (PAMs)—provide high power-to-weight ratios [5], large deformations [6], low cost [7], and straightforward fabrication [8]. Nevertheless, severe nonlinearities, rate-dependent hysteresis, input coupling, and sensitivity to disturbances complicate precise control [9]. To mitigate these difficulties, several robust strategies have been investigated, including adaptive control [10], sliding mode control (SMC) [11], and active disturbance-rejection control (ADRC) [12]. An adaptive hysteresis–compensation scheme in [10] achieves high-precision trajectory tracking for PAM joints, whereas a constraint-aware adaptive fuzzy controller in [13] maintains prescribed tracking accuracy under time-varying motion constraints. These advances highlight the growing maturity of PAM control; however, further work is needed to balance tracking accuracy, robustness, and computational efficiency for real-world deployment.
SMC has become the preferred robust strategy for PAM-actuated systems owing to its inherent resilience to parameter variations and external disturbances [11]. Literature [11] combines adaptive SMC with fuzzy-logic approximation on a PAM-driven humanoid arm, achieving asymptotic trajectory tracking despite unknown nonlinearities. More recently, the study in [14] incorporates a disturbance observer into the SMC framework and, with rigorous stability analysis, guarantees uniformly ultimately bounded tracking errors under discontinuous friction.
As a robust anti-disturbance methodology, ADRC has been widely deployed in power systems [15], industrial robots [12], and multi-agent networks [16]. By dynamically estimating and compensating both endogenous uncertainties and exogenous disturbances, ADRC largely decouples closed-loop performance from precise modeling requirements [17]. Its core component—the extended state observer (ESO)—treats disturbances as augmented state variables, permitting simultaneous estimation of the original and disturbance states and thereby yielding a full description of the extended system dynamics [18]. Numerous ESO variants have been proposed. A reduced-order ESO for mobile robots is introduced in [19], while an event-triggered Takagi–Sugeno fuzzy ESO with adaptive parameters for bandwidth-constrained nonlinear systems is reported in [20]. To attenuate measurement noise, a composite observer that couples an ESO with a Kalman filter is presented in [21]. ESO-based solutions have also been tailored to pneumatic artificial muscles: a super-twisting ESO for trajectory regulation is presented in [22], and an ESO-assisted sliding mode controller is detailed in [23]. Despite these advances, designing an ESO that maximizes closed-loop performance remains unresolved. The composite observer in [21], for example, achieves minimum-variance state estimation rather than disturbance estimation explicitly optimized for control objectives. Motivated by this gap, the present work seeks to develop a new ESO that delivers optimal control performance.
In addition to exogenous disturbances, practical systems must respect stringent operational limits—especially state bounds and actuator saturation—which directly influence performance and safety [24,25,26]. Model predictive control (MPC) mitigates these limitations by solving a receding-horizon optimization that enforces state and input constraints, delivering superior closed-loop behavior with explicit safety guarantees [27,28]. Literature [27] introduces a hybrid disturbance-predictive, adaptive, event-triggered MPC that improves prediction accuracy while reducing communication demand, whereas [29] develops a periodic event-triggered MPC for nonlinear systems under bounded disturbances. Despite such progress, applications of MPC to constraint-ridden pneumatic manipulators remain scarce—an observation that motivates the present study.
This work addresses the critical challenge of constrained control synthesis for pneumatic manipulator systems operating under concurrent state and input constraints as well as external disturbances. The principal theoretical and methodological advancements are systematically articulated as follows: (1) a novel ESO with an additive term, whose estimation error dynamics are explicitly incorporated into the MPC cost function formulation, is proposed that dynamically adjusts estimation gain through real-time error feedback; (2) an MPC-enabled disturbance-rejection controller is designed for the pneumatic manipulator system to address both constraints and exogenous perturbations in soft actuator applications; and (3) sufficient stability conditions are established in terms of linear matrix inequalities (LMIs), theoretically ensuring the uniform ultimate boundedness of all closed-loop signals under the proposed MPC-based disturbance-rejection controller.

2. Problem Formulation and Preliminaries

2.1. The Model of the Pneumatic Manipulator System

In this paper, the model of the pneumatic manipulator system considered is based on [30]. In [30], the authors have established the dynamic model of the pneumatic manipulator for the corresponding experimental platform. In the experimental setup of [30], compressed air is provided by an air compressor and supplied to pressure-proportional valves as the air source. Voltage signals from the industrial computer are transmitted to twelve pressure-proportional valves, which regulate the internal pressures and pulling forces of the twelve PAMs. During the control process, the system operates in three scenes: an initial scene, a preloading scene, and a movement scene. When the pneumatic manipulator is not pressurized, the input signals of the twelve pressure-proportional valves are zero, indicating that all PAMs are in a relaxed state without air pressure, and the deflection angle of the manipulator is zero. When the manipulator is required to track a given deflection angle, its motion is driven by antagonistic PAMs: one PAM is inflated while the other is deflated, resulting in an angular deflection of the manipulator. From this process, it can be seen that, although the input signal to the system is voltage, it is ultimately converted into internal pressure and pulling force of the PAMs. Therefore, in the dynamic model, the actual control input is the force, and the control objective is angle regulation. The objective of this study is to design an MPC scheme based on a novel ESO, enabling the pneumatic manipulator to track a given deflection angle signal with optimal control cost.
The pneumatic manipulator model is given according to [30]
θ ¨ = R ˜ k 0 2 π N 2 m l ˜ 2 ( 3 L 0 2 b 2 ) u 3 R ˜ 2 L 0 p 0 π N 2 m l ˜ 2 θ g sin θ l ˜ + 3 R ˜ 3 k 0 2 π N 2 m l ˜ 2 θ 2 u ,
where θ denotes the deflection angle of the joint mechanism; u is the control input. The system configuration parameters include R ˜ (moment arm distance between PAM actuators and the rotating linkage), m (equivalent total mass incorporating the lower chassis and adjacent joint components), and l ˜ (effective length of the driving linkage). The pneumatic actuation characteristics are defined by k 0 (pressure–voltage conversion coefficient of the proportional valves), p 0 (preloading pressure of PAMs), and L 0 (nominal rest length of the pneumatic artificial muscles). Additional manufacturing parameters specify N (number of synthetic fiber winding turns in PAM construction) and b (total length of rayon).
Then, system (1) is reformulated as the following state-space form as
s ˙ 1 = s 2 s ˙ 2 = b 1 s 1 + b 0 u + w d y = s 1
where s 1 θ , s 2 θ ˙ , and
b 0 R ˜ k 0 2 π N 2 m l ˜ 2 ( 3 L 0 2 b 2 ) , b 1 3 R ˜ 2 L 0 p 0 π N 2 m l ˜ 2 , w d 3 R ˜ 3 k 0 2 π N 2 m l ˜ 2 s 1 2 u g l ˜ sin ( s 1 ) .
Then, system (2) can be transformed as the following discrete-time state-space form
s 1 ( i + 1 ) = s 1 ( i ) + T s 2 ( i ) , s 2 ( i + 1 ) = T b 1 s 1 ( i ) + s 2 ( i ) + T b 0 u ( i ) + T w d ( i ) , y ( i ) = s 1 ( i ) ,
where T is the sampling period; i is the time step.
Considering the physical constraints on s 1 ( i ) , s 2 ( i ) , and u ( i ) of the pneumatic manipulator system, this study formulates these limitations in the following mathematical representation:
| s 1 ( i ) | s ¯ 1 ,
| s 2 ( i ) | s ¯ 2 ,
| u ( i ) | u ¯ ,
respectively, where s ¯ 1 , s ¯ 2 , and u ¯ are known positive constants.
Then, constraints (4)–(6) can be rewritten as
s ( i ) S { x : b s T s ( i ) h s } ,
u ( i ) U { u : b u T u ( i ) h u } ,
where
s ( i ) s 1 ( i ) s 2 ( i ) , h s s ¯ 1 s ¯ 1 s ¯ 2 s ¯ 2 , h u u ¯ u ¯ , b s 1 1 0 0 0 0 1 1 , b u 1 1 .
According to (7) and (8), the constraint of w d ( i ) is easily obtained as
w d ( i ) W { w d ( i ) : b w T w d ( i ) h w } ,
where
b w 1 1 , h w w ¯ d w ¯ d , w ¯ d 3 R ˜ 3 k 0 2 π N 2 m l ˜ 2 s ¯ 1 2 u ¯ + g l ˜ sin ( s ¯ 1 ) .
Remark 1.
Subject to the mechanical constraints of the pneumatic manipulator’s physical structure, its actual deflection angle is practically confined to ± 15 ; i.e., θ ( i ) [ 15 , + 15 ] . Consequently, the disturbance boundary is defined as w ¯ d .

2.2. The Receding-Horizon-Based ESO

In this paper, a receding-horizon-based ESO (RH-ESO) is proposed based on the receding-horizon mechanism. Decision variables are used in the RH-ESO to simultaneously optimize observer performance, suppress high-gain phenomena, and mitigate overshoot in disturbance estimation errors. The RH-ESO is designed as
s ^ 1 ( i + 1 ) = s ^ 1 ( i ) + T s ^ 2 ( i ) + T β 1 e 1 ( i ) + T l 1 ( i ) ,
s ^ 2 ( i + 1 ) = T b 1 s ^ 1 ( i ) + s ^ 2 ( i ) + T s ^ 3 ( i ) + T b 0 u ( i ) + T β 2 e 1 ( i ) + T l 2 ( i ) ,
s ^ 3 ( i + 1 ) = s ^ 3 ( i ) + T β 3 e 1 ( i ) + T l 3 ( i ) ,
where s ^ 1 ( i ) is the estimation of s 1 ( i ) ; s ^ 2 ( i ) is the estimation of s 2 ( i ) ; s ^ 3 ( i ) is the estimation of w d ( i ) ; l m ( m = 1 , 2 , 3 ) is the decision variable that is obtained by solving an optimization problem subsequently; β m ( m = 1 , 2 , 3 ) are the adjustable gains; ε m ( i ) s m ( i ) s ^ m ( i ) ( m = 1 , 2 , 3 ) are the estimation errors that can be obtained as
ε 1 ( i + 1 ) = ( 1 T β 1 ) ε 1 ( i ) + T ε 2 ( i ) T l 1 ( i ) ,
ε 2 ( i + 1 ) = ( T b 1 T β 2 ) ε 1 ( i ) + ε 2 ( i ) + T ε 3 ( i ) T l 2 ( i ) ,
ε 3 ( i + 1 ) = T β 3 ε 1 ( i ) + ε 3 ( i ) T l 3 ( i ) + T Δ d ( i ) ,
where Δ d ( i ) w d ( i + 1 ) w d ( i ) T .
From (13), we have
ε 3 ( i ) = 1 T [ ε 2 ( i + 1 ) ( T b 1 T β 2 ) ε 1 ( i ) ε 2 ( i ) + T l 2 ( i ) ] ,
Substituting (15) into (14), we have
ε 3 ( i + 1 ) = T β 3 ε 1 ( i ) + 1 T [ ε 2 ( i + 1 ) ( T b 1 T β 2 ) ε 1 ( i ) ε 2 ( i ) + T l 2 ( i ) ] T l 3 ( i ) + T Δ d ( i )
From (12), we have
ε 2 ( i ) = 1 T ε 1 ( i + 1 ) ( 1 T β 1 ) ε 1 ( i ) + T l 1 ( i )
Substituting (17) into (16), we have
ε 3 ( i + 1 ) = T β 3 ε 1 ( i ) + 1 T [ ε 2 ( i + 1 ) ( T b 1 T β 2 ) ε 1 ( i ) 1 T ε 1 ( i + 1 ) ( 1 T β 1 ) ε 1 ( i ) + T l 1 ( i ) + T l 2 ( i ) ] T l 3 ( i ) + T Δ d ( i ) = 1 T 2 ε 1 ( i + 1 ) + 1 T 2 b 1 β 1 T + β 2 T β 3 ε 1 ( i ) + 1 T ε 2 ( i + 1 ) + 1 T l 1 ( i ) + l 2 ( i ) T l 3 ( i ) + T Δ d ( i )
From (18), we have
ε 3 ( i ) = 1 T 2 ε 1 ( i ) + 1 T 2 b 1 β 1 T + β 2 T β 3 ε 1 ( i 1 ) + 1 T ε 2 ( i ) + 1 T l 1 ( i 1 ) + l 2 ( i 1 ) T l 3 ( i 1 ) + T Δ d ( i 1 ) ,
Then, substituting (17) and (19) into (13), we have
ε 2 ( i + 1 ) = 2 T ε 1 ( i + 1 ) + T b 1 3 T + 2 β 1 T β 2 ε 1 ( i ) + 1 T T b 1 β 1 + T β 2 T 2 β 3 ε 1 ( i 1 ) + 2 l 1 ( i ) l 1 ( i 1 ) ,
From (20), we have
ε 2 ( i ) = 2 T ε 1 ( i ) + T b 1 3 T + 2 β 1 T β 2 ε 1 ( i 1 ) + 1 T T b 1 β 1 + T β 2 T 2 β 3 ε 1 ( i 2 ) + 2 l 1 ( i 1 ) l 1 ( i 2 ) ,
Then, substituting (21) into (12), we have
ε 1 ( i + 1 ) = 3 T β 1 ε 1 ( i ) + T 2 b 1 T 2 β 2 + 2 T β 1 3 ε 1 ( i 1 ) + 1 T 2 b 1 T β 1 + T 2 β 2 T 3 β 3 ε 1 ( i 2 ) T l 1 ( i ) + 2 T l 1 ( i 1 ) T l 1 ( i 2 ) T 2 l 2 ( i 1 ) + T 2 l 2 ( i 2 ) T 3 l 3 ( i 2 ) + T 3 Δ d ( i 2 )
Let
m 11 = 3 T β 1 , m 12 = T 2 b 1 T 2 β 2 + 2 T β 1 3 , m 13 = 1 T 2 b 1 T β 1 + T 2 β 2 T 3 β 3 , m 14 = T , m 15 = 2 T , m 16 = T , m 17 = T 2 , m 18 = T 2 , m 19 = T 3 , m 21 = 2 T , m 22 = T b 1 3 T + 2 β 1 T β 2 , m 23 = 1 T T b 1 β 1 + T β 2 T 2 β 3 , m 24 = 2 , m 25 = 1 , m 26 = T , m 27 = T , m 28 = T 2 , m 31 = 1 T 2 , m 32 = 1 T 2 b 1 β 1 T + β 2 T β 3 , m 33 = 1 T , m 34 = 1 T , m 35 = 1 , m 36 = T .
Then, (18), (21), and (22) can be converted as
ε 1 ( i + 1 ) = m 11 ε 1 ( i ) + m 12 ε 1 ( i 1 ) + m 13 ε 1 ( i 2 ) + m 14 l 1 ( i ) + m 15 l 1 ( i 1 ) + m 16 l 1 ( i 2 ) + m 17 l 2 ( i 1 ) + m 18 l 2 ( i 2 ) + m 19 l 3 ( i 2 ) + T 3 Δ d ( i 2 ) , ε 2 ( i + 1 ) = m 21 ε 1 ( i + 1 ) + m 22 ε 1 ( i ) + m 23 ε 1 ( i 1 ) + m 24 l 1 ( i ) + m 25 l 1 ( i 1 ) + m 26 l 2 ( i ) + m 27 l 2 ( i 1 ) + m 28 l 3 ( i 1 ) + T 2 Δ d ( i 1 ) , ε 3 ( i + 1 ) = m 31 ε 1 ( i + 1 ) + m 32 ε 1 ( i ) + m 33 ε 2 ( i + 1 ) + m 34 l 1 ( i ) + m 35 l 2 ( i ) + m 36 l 3 ( i ) + T Δ d ( i ) ,
respectively.
Denote
ε s ( i ) [ ε 1 ( i ) ε 2 ( i ) ε 3 ( i ) ε 1 ( i 1 ) ε 1 ( i 2 ) ε 1 ( i 3 ) ] T , l ( i ) [ l 1 ( i ) l 2 ( i ) l 3 ( i ) ] T , l ( i ) = [ l T ( i 1 ) l T ( i 2 ) ] T , J ( i ) [ T 3 Δ d ( i 2 ) T 2 Δ d ( i 1 ) T Δ d ( i ) ] , l ( 1 ) 0 6 × 1 , l ( 2 ) = [ l T ( 1 ) 0 1 × 3 ] T .
The dynamics of ε s ( i ) can be obtained as
ε s ( i + 1 ) = G ε ε s ( i + 1 ) + A ˜ ε ε s ( i ) + B ˜ ε l ( i ) + D ˜ ε l ( i ) + J ( i ) ,
where
G ε 0 0 0 0 0 0 m 21 0 0 0 0 0 m 31 m 33 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 , A ˜ ε m 11 0 0 m 12 m 13 0 m 22 0 0 m 23 0 0 m 32 0 0 0 0 0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 , B ˜ ε m 14 0 0 m 24 0 0 m 34 m 35 m 36 0 0 0 0 0 0 0 0 0 , D ˜ ε m 15 m 17 0 m 16 m 18 m 19 m 25 m 27 m 28 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 .
Then, it yields
ε s ( i + 1 ) = ( I G ε ) 1 A ˜ ε ε s ( i ) + ( I G ε ) 1 B ˜ ε l ( i ) + ( I G ε ) 1 D ˜ ε l ( i ) + ( I G ε ) 1 J ( i ) ,
Denote
A ε ( I G ε ) 1 A ˜ ε , B ε ( I G ε ) 1 B ˜ ε , D ε ( I G ε ) 1 D ˜ ε , F e ( I G ε ) 1 .
Thus, there exists
ε s ( i + 1 ) = A ε ε s ( i ) + B ε l ( i ) + D ε l ( i ) + F e J ( i ) .
To mitigate abrupt changes in observed values induced by overshoot during observation error convergence, this study imposes constraints on the observation error to enforce smooth convergence characteristics. This restriction ensures gradual variations in both state and disturbance estimates, thereby enhancing the closed-loop system performance through stabilized estimation dynamics. The constraints on ε m ( i ) ( m = 1 , 2 , 3 ) are given as
| ε m ( i ) | h ε m ,
where h ε m are given positive constants. Then, it is easy to obtain
b ε s T ε s ( i ) h ε s ,
where b ε s and h ε s are obtained from constraint (23).
Remark 2.
Constraint (24) is designed to bound the disturbance estimation error, which effectively mitigates the occurrence of high-gain phenomena in the RH-ESO and reduces the overshoot in disturbance estimation. By incorporating this constraint, the disturbance estimation error can achieve smooth convergence characteristics, thereby facilitating the subsequent derivation of control inputs with enhanced stability and reduced transient fluctuations. This systematic constraint implementation ensures coordinated performance improvements throughout the closed-loop control system.

2.3. The Tracking Differentiator

In this subsection, the tracking differentiator (TD) is used to handle the transient process of the system state. Denote v 0 as the given input, v 1 ( i ) as the output tracking of v 0 , and v 2 ( i ) as the differential signal of v 1 ( i ) . In this paper, v 0 = θ 0 is the deflection angle to be tracked. The TD is designed as
v 1 ( i + 1 ) = v 1 ( i ) + T v 2 ( i ) , v 2 ( i + 1 ) = v 2 ( i ) + T fhan ( i ) ,
where
fhan ( i ) = r s sign ( a ( i ) ) , | a ( i ) | > d r s a ( i ) d , | a ( i ) | d a ( i ) = v 2 ( i ) + a 0 ( i ) d 2 sign ( d s ( i ) ) , | d s ( i ) | > d 0 v 2 ( i ) + d s ( i ) T , | d s ( i ) | d 0 d s ( i ) = v 1 ( i ) v 0 + T v 2 ( i ) , d = r s T , d 0 = r s T 2 a 0 ( i ) = d 2 + 8 r s | d s ( i ) |
with r s > 0 and d > 0 as given constants.

2.4. The MPC-Enabled Disturbance-Rejection Controller

To design the controller, the state tracking errors are denoted as r 1 ( i ) v 1 ( i ) s ^ 1 ( i ) and r 2 ( i ) v 2 ( i ) s ^ 2 ( i ) . Then, the control input is designed as
u ( i ) = k 1 r 1 ( i ) + k 2 r 2 ( i ) + c ( i ) 1 b 0 s ^ 3 ( i ) ,
where k 1 and k 2 are feedback gains to be designed; c ( i ) is the decision variable of the MPC scheme to be obtained subsequently.
According to r 1 ( i ) , r 2 ( i ) , and (25), there exists
r 1 ( i + 1 ) = r 1 ( i ) + T r 2 ( i ) T β 1 ε 1 ( i ) T l 1 ( i ) , r 2 ( i + 1 ) = r 2 ( i ) T b 0 u ( i ) + T fhan ( i ) T b 1 s ^ 1 ( i ) T s ^ 3 ( i ) T β 2 ε 1 ( i ) T l 2 ( i ) .
Then, it is obtained that
r 1 ( i + 1 ) = r 1 ( i ) + T r 2 ( i ) T β 1 ε 1 ( i ) T l 1 ( i ) , r 2 ( i + 1 ) = T b 0 k 1 r 1 ( i ) + ( 1 T b 0 k 2 ) r 2 ( i ) T b 0 c ( i ) T β 2 ε 1 ( i ) T l 2 ( i ) + q ( i ) ,
where q ( i ) = T fhan ( i ) T b 1 s ^ 1 ( i ) .
Denote r ( i ) [ r 1 ( i ) r 2 ( i ) ] T . Then, it follows that
r ( i + 1 ) = A r r ( i ) + B r ε s ( i ) + C r c ( i ) + D r l ( i ) + E r q ( i ) .
where
A r 1 T T b 0 k 1 1 T b 0 k 2 , B r T β 1 0 1 × 5 T β 2 0 1 × 5 , C r 0 T b 0 , D r T 0 0 0 T 0 , E r 0 1 .
Then, r ( i ) is constrained by r ( i ) R { r ( i ) : b r T r ( i ) h r } , where
b r 1 1 0 0 0 0 1 1 , h r r ¯ 1 r ¯ 1 r ¯ 2 r ¯ 2 T ,
with r ¯ 1 and r ¯ 2 being given positive constants. Then, the constraint on q ( i ) is obtained as
q ( i ) Q { q ( i ) : b q T q ( i ) q ¯ } ,
with q ¯ T r s + T b 1 ( s ¯ 1 + r ¯ 1 ) . Denote ζ ( i ) [ r T ( i ) ε s T ( i ) ] T and h ( i ) [ c T ( i ) l T ( i ) ] T . Then, we have
ζ ( i + 1 ) = A ζ ζ ( i ) + B ζ h ( i ) + C ζ l ( i ) + D ζ q ( i ) ,
where
A ζ A r B r 0 6 × 2 A ε , B ζ C r D r 0 6 × 1 B ε , C ζ 0 2 × 6 D ε , D ζ E r 0 6 × 1 .
The nominal system of ζ ( i ) is obtained as
ζ ¯ ( i + 1 ) = A ζ ζ ¯ ( i ) + B ζ h ( i ) + C ζ l ( i ) ,
where ζ ¯ ( i ) [ r ¯ T ( i ) ε ¯ s T ( i ) ] T with
r ¯ ( i + 1 ) = A r r ¯ ( i ) + B r ε ¯ s ( i ) + C r c ( i ) , ε ¯ s ( i + 1 ) = A ε ε ¯ s ( i ) + B ε l ( i ) + D ε l ( i ) .
Accordingly, the constraints on ζ ¯ ( i ) and h ( i ) are obtained as
ζ ¯ ( i ) M { ζ ¯ ( i ) : b ζ T ζ ¯ ( i ) h ζ } , h ( i ) N { h ( i ) : b ˜ u T ζ ¯ ( i ) + b h T h ( i ) h h + h q ( i ) } ,
respectively, where
b ζ b r 0 0 b ε s , h ζ h r h ε s , h q ( i ) b u T s ^ 3 ( i ) 0 6 × 1 , b ˜ u [ k 1 k 2 ] T b u 0 2 × 6 0 6 × 2 0 6 × 6 , b h b u 0 1 × 3 0 1 × 1 0 1 × 3 , h h h u 0 6 × 1 .

2.5. The MPC Scheme

In this paper, the MPC scheme is proposed to deal with the state and input constraints of the pneumatic manipulator as well as generate a part of the control input signal. To establish the MPC scheme, a cost function is designed as
J ( l ^ ( i ) , h ^ ( i ) ) s = 0 T s 1 ζ ^ ( s , i ) Q 2 + h ^ ( s , i ) R 2 + ζ ^ ( T s , i ) P 2 ,
where ζ ^ ( s , i ) is the predictive state with ζ ^ ( 0 , i ) = ζ ¯ ( i ) = ζ ( i ) ; h ^ ( s , i ) is the predictive decision variable; ζ ^ ( i ) [ ζ ^ T ( 0 , i ) ζ ^ T ( 1 , i ) ζ ^ T ( T s , i ) ] T is the predictive state sequence; h ^ ( i ) [ h ^ T ( 0 , i ) h ^ T ( 1 , i ) h ^ T ( T s 1 , i ) ] T is the predictive decision variable sequence with h ^ ( 0 , i ) = h ( i ) ; Q and R denote the positive definite weighting matrices; P > 0 represents the terminal penalty matrix requiring design. The predictive horizon is denoted by T s .
Next, the optimization problem of the MPC scheme can be formulated as
Prob 1:
( l ^ * ( i ) , h ^ * ( i ) ) arg min h ^ ( i ) J ( l ^ ( i ) , h ^ ( i ) )
subject to
ζ ^ ( s + 1 , i ) = A ζ ζ ^ ( s , i ) + B ζ h ^ ( s , i ) + C ζ l ^ ( s , i )
b ζ T ζ ^ ( s , i ) h ζ ϕ ( s , i ) , s Z [ 0 , T s 1 ]
b ˜ u T ζ ^ ( s , i ) + b h T h ^ ( s , i ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( s , i ) , s Z [ 0 , T s 1 ] ,
ζ ^ ( T s , i ) ζ T ( T s , i ) ,
where
ϕ ( s , i ) j = 0 s 1 max q ( s 1 j , i ) Q b ζ T A ζ j D ζ e ζ ( s 1 j , i ) , χ ( s , i ) j = 0 s 1 max q ( s 1 j , i ) Q b ˜ u T A ζ j D ζ e ζ ( s 1 j , i ) , e ζ ( s + 1 , i ) = A ζ e ζ ( s , i ) + D ζ q ( s , i ) , ϕ ( 0 , i ) 0 10 × 1 , χ ( 0 , i ) 0 8 × 1 , e ζ ( 0 , i ) = 0 4 × 1 , ζ T ( T s , i ) ζ ^ ( T s , i ) H ^ ( T s , i ) , ζ ^ ( T s , i ) { ζ ^ ( T s , i ) : b ζ T ζ ^ ( T s , i ) h ζ ϕ ( T s , i ) } , H ^ ( T s , i ) { ζ ^ ( T s , i ) : b ˜ u T ζ ^ ( T s , i ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( T s , i ) } , ζ ^ * ( i ) [ ζ ^ * T ( 0 , i ) ζ ^ * T ( 1 , i ) ζ ^ * T ( T s , i ) ] T , h ^ * ( i ) [ h ^ * T ( 0 , i ) h ^ * T ( 1 , i ) h ^ * T ( T s 1 , i ) ] T , l ^ ( s , i ) = l ( i ) , s = 0 0 3 × 1 I 3 × 3 h ^ ( 0 , i ) I 3 × 3 0 3 × 3 l ( i ) , s = 1 0 3 × 1 I 3 × 3 h ^ ( s 1 , i ) 0 3 × 1 I 3 × 3 h ^ ( s 2 , i ) , s Z [ 2 , T s 1 ]
Remark 3.
The optimality of the RH-ESO proposed in this paper is fundamentally different from that in [21]. In [21], the system under consideration includes a slowly time-varying disturbance in the state equations and Gaussian white noise in the output equations. It should be noted that the optimality in [21] is defined in terms of a minimum-variance performance index for estimating the states s 1 ( i ) and s 2 ( i ) and does not address an optimal estimate of the disturbance w d ( i ) . By contrast, the focus of this paper is on the rolling optimization of the disturbance estimation error ε s ( i ) to obtain the optimal estimate of the slowly varying disturbance w d ( i ) at the current instant. This “optimal” estimate is understood as the one that minimizes the system control performance index—that is, the quadratic tracking error of the system state.
The structure diagram of the pneumatic manipulator system is presented in Figure 1.
Definition 1
([31]). System (3) is said to be uniformly bounded if, for all Ξ 1 > 0 , there exists a constant Ξ 2 ( Ξ 1 ) such that
s ( 0 ) Ξ 1 s ( i ) Ξ 2 ( Ξ 1 ) ,
holds for all s ( 0 ) R n and i Z 0 .
Following the above discussion, this paper aims to achieve the following two objectives: (1) establish sufficient conditions to ensure that the proposed MPC scheme is recursively feasible, and (2) establish easily verifiable sufficient conditions to ensure that system (3) under control input (25) is uniformly bounded.

3. Results

This section establishes the recursive feasibility of Prob 1 and provides rigorous Lyapunov-based stability guarantees for the closed-loop system.
Firstly, the following lemma is given to prove the feasibility of Prob 1.
Lemma 1.
If Prob 1 is feasible at time instant t, then there exists at least one feasible solution for Prob 1 at time instant i + 1 .
Proof. 
Denote candidate decision variable sequences h ˜ ( i + 1 ) and l ˜ ( i + 1 ) as
h ˜ ( i + 1 ) h ˜ T ( 0 , i + 1 ) h ˜ T ( T s , i + 1 ) T = h ^ T * ( 1 , i ) h ^ T * ( T s 1 , i ) 0 4 × 1 T , l ˜ ( i + 1 ) l ˜ T ( 0 , i + 1 ) l ˜ T ( T s , i + 1 ) T = l ^ T ( 1 , i ) l ^ T ( T s 1 , i ) 0 3 × 1 T ,
where h ˜ ( s , i + 1 ) is a candidate decision variable for s Z [ 0 , T s ] . l ^ ( s , i + 1 ) can be obtained from h ˜ ( s , i + 1 ) for s Z [ 0 , T s ] . Then, we have
ζ ^ ( 0 , i + 1 ) = A ζ ζ ^ ( 0 , i ) + B ζ h ^ * ( 0 , i ) + C ζ l ^ * ( 0 , i ) + D ζ q ( i ) .
By considering the predictive dynamics of ζ ( i + 1 ) based on h ˜ ( i + 1 ) , it follows that the corresponding state predictions can be derived from the system’s predictive dynamics:
ζ ˜ ( s + 1 , i + 1 ) = A ζ η ˜ ( s , i + 1 ) + B ζ h ˜ ( s , i + 1 ) + C ζ l ˜ ( s , i + 1 ) + D ζ q ( s , i + 1 ) ,
where ζ ˜ ( s + 1 , i + 1 ) is the feasible state for s Z [ 0 , T s ] .
Combining (26), (30), and (31), it is obtained that
ζ ˜ ( s , i + 1 ) = ζ ^ * ( s , i ) + A ζ s 1 D ζ q ( i ) .
Denote l ˜ ( i + 1 ) ζ ˜ T ( 0 , i + 1 ) ζ ˜ T ( T s , i + 1 ) T as the feasible predictive state sequence.
In order to ensure the existence of at least one solution for Prob 1 at each subsequent time step i + 1 , it is necessary that the following conditions are satisfied:
b ζ T ζ ˜ ( s , i + 1 ) h ζ ϕ ( s , i + 1 ) , s Z [ 0 , T s ] ,
b ˜ u T ζ ˜ ( s , i + 1 ) + b h T h ˜ ( s , i + 1 ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( s , i + 1 ) , s Z [ 0 , T s 1 ] ,
ζ ˜ ( T s , i + 1 ) ζ T ( T s , i + 1 ) .
In the following analysis, conditions (33)–(35) will be demonstrated one by one. To begin with, we first provide the proof for condition (33). Assuming that Prob 1 has a feasible solution at time instant t, it directly follows that constraint (27) will be satisfied, which in turn implies
b ζ T ζ ^ * ( s , i ) h ζ ϕ ( s , i ) ,
for s Z [ 0 , T s 2 ] . Based on (32) and (36), we have b ζ T ( ζ ˜ ( s , i + 1 ) A ζ s D ζ q ( i ) ) h ζ ϕ ( s , i ) , which yields
b ζ T ζ ˜ ( s , i + 1 ) h ζ ϕ ( s , i ) + b ζ T A ζ s D ζ q ( i ) = h ζ j = 0 s 1 max q ( s 1 j , i ) Q b ζ T A ζ j D ζ q ( s 1 j , i ) + b ζ T A ζ s D ζ q ( i ) max q ( s j , i ) Q q ( s j , i ) h ζ j = 0 s 1 max q ( s 1 j , i ) Q b ζ T A ζ j D ζ q ( s 1 j , i ) = h ζ ϕ ( s , i + 1 ) ,
for s Z [ 0 , T s ] . Thus, condition (33) holds true for all s within the interval Z [ 0 , T s ] .
Next, the condition (34) is verified for the time instant i + 1 . By utilizing the sequences h ˜ ( i + 1 ) and l ˜ ( i + 1 ) , we have h ˜ ( s , i + 1 ) = h ^ ( s , i ) and l ˜ ( s , i + 1 ) = l ^ ( s , i ) . Based on (28) and (32), it follows that
b ˜ u T ( ζ ˜ ( s , i + 1 ) A ζ s D ζ q ( i ) ) + b h T h ^ ( s , i ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( s , i ) ,
which indicates
b ˜ u T ζ ˜ ( s , i + 1 ) + b h T h ˜ ( s , i + 1 ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( s , i ) + b ˜ u T A ζ s D ζ q ( i ) ) = h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T j = 0 s 1 max q ( s 1 j , i ) Q b ˜ u T A ζ j D ζ q ( s 1 j , i ) + b ˜ u T A ζ s D ζ q ( i ) max q ( s j , i ) Q q ( s j , i ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T j = 0 s 1 max q ( s 1 j , i ) Q b ˜ u T A ζ j D ζ q ( s 1 j , i ) = h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( s , i + 1 )
for s Z [ 0 , T s 2 ] . For s = T s 1 , there exist h ˜ ( T s , i + 1 ) = 0 4 × 1 and l ˜ ( T s , i + 1 ) = 0 6 × 1 . Based on (29) and (32), h ˜ ( T s , i + 1 ) and l ˜ ( T s , i + 1 ) , it is easy to obtain
b ˜ u T ( ζ ˜ ( T s , i + 1 ) A ζ T s 1 D ζ q ( i ) ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( T s , i ) ,
which means
b ˜ u T ζ ˜ ( T s , i + 1 ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( T s , i ) + b ˜ u T A ζ T s 1 D ζ q ( i ) = h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T j = 0 T s 2 max q ( T s 2 j , i ) Q b ˜ u T A ζ j D ζ q ( T s 2 j , i ) b ˜ u T A ζ T s 1 D ζ max q ( T s 1 j , i ) Q q ( T s 1 j , i ) + b ˜ u T A ζ T s 1 D ζ q ( i ) h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T j = 0 T s 2 max q ( T s 2 j , i ) Q b ˜ u T A ζ j D ζ q ( T s 2 j , i ) = h h + ( w ¯ d + h ε 3 ) [ 1 1 0 1 × 6 ] T χ ( T s , i + 1 ) .
Thus, condition (34) holds true for all s within the interval Z [ 0 , T s 1 ] .
Finally, the satisfaction of condition (35) is established. Thus, ζ T ( T s , i + 1 ) ζ ^ ( T s , i ) H ^ ( T s , i + 1 ) . From (33), it follows that ζ ˜ ( T s , i + 1 ) belongs to the set ζ ^ ( T s , i ) . Additionally, based on (29), it can be readily shown that ζ ˜ ( T s , i + 1 ) also lies within H ^ ( T s , i ) .
Therefore, since both membership conditions are satisfied, we conclude that ζ ˜ ( T s , i + 1 ) ζ T ( T s , i + 1 ) holds true, thereby ensuring the validity of condition (35). □
Theorem 1.
If there exist appropriate parameters k 1 , k 2 , β 1 , β 2 , β 3 , and a positive definite matrix P, make the linear matrix inequality (LMI)
A ζ T P A ζ P + Q < 0 ,
hold; then, ε s ( i ) [ e s 1 ( i ) e s 2 ( i ) ] T is uniformly bounded under the control input (25) with e s 1 ( i ) v 1 ( i ) s 1 ( i ) and e s 2 ( i ) v 2 ( i ) s 2 ( i ) . Moreover, the state ε s ( i ) converges to the set E { ε s ( i ) : ε s ( i ) M λ ̲ 1 2 ( Q ) σ 1 2 } with
σ 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s ) + s = 1 T s 1 2 λ ¯ 2 ( Q ) Θ ( s ) Ω ( s ) + λ ¯ 2 ( Q ) Ω 2 ( s ) Θ ( s ) max b ζ T ζ ^ * ( s , i ) h ζ ϕ ( s , i ) ζ ^ * ( s , i ) , Ω ( s ) max b q T q ( i ) q ¯ A ζ s 1 D ζ q ( i ) , M 1 0 1 0 0 0 1 0 1 0 .
Proof. 
Considering e s 1 ( i ) r 1 ( i ) ε 1 ( i ) and e s 2 ( i ) r 2 ( i ) ε 2 ( i ) , it can be obtained that e s 1 ( i ) r 1 ( i ) + ε 1 ( i ) and e s 2 ( i ) r 2 ( i ) + ε 2 ( i ) . Subsequently, the proof of uniform boundedness of e s 1 ( i ) and e s 2 ( i ) is given first.
Denote the optimal cost of Prob 1 at time instant i as
J ( l ^ * ( i ) , h ^ * ( i ) ) s = 0 T s 1 ζ ^ * ( s , i ) Q 2 + h ^ * ( s , i ) R 2 + ζ ^ * ( T s , i ) P 2
Then, a Lyapunov candidate function is denoted as V ( i ) J ( l ^ * ( i ) , h ^ * ( i ) ) . Consequently, there exists
Δ V ( i ) J ( l ^ * ( i + 1 ) , h ^ * ( i + 1 ) ) J ( l ^ * ( i ) , h ^ * ( i ) )
Given the optimal solution that Prob 1 is obtained at time instant i, then Prob 1 has at least one feasible solution at i + 1 by Lemma 1. Therefore, the feasible cost of Prob 1 at time instant i + 1 can be denoted as
J ( l ˜ ( i + 1 ) , h ˜ ( i + 1 ) ) s = 0 T s 1 ζ ˜ ( s , i + 1 ) Q 2 + h ˜ ( s , i ) R 2 + ζ ˜ ( T s , i + 1 ) P 2
Denote
Δ V ˜ ( i ) J ( l ˜ ( i + 1 ) , h ˜ ( i + 1 ) ) J ( l ^ * ( i ) , h ^ * ( i ) )
Split Δ V ˜ ( i ) Δ 1 ( i ) + Δ 2 ( i ) + Δ 3 ( i ) , where
Δ 1 ( i ) s = 1 T s 1 ζ ˜ ( s , i + 1 ) Q 2 ζ ^ * ( s , i ) Q 2 + s = 1 T s 1 h ˜ ( s , i + 1 ) R 2 h ^ * ( s , i ) R 2 Δ 2 ( i ) ζ ˜ ( T s , i + 1 ) Q 2 + h ˜ ( T s , i + 1 ) R 2 + ζ ˜ ( T s , i + 1 ) P 2 ζ ^ * ( T s , i ) P 2 Δ 3 ( i ) ζ ^ * ( 0 , i ) Q 2 h ^ * ( 0 , i ) R 2
For Δ 1 ( i ) , there exists h ˜ ( s , i + 1 ) = h ^ * ( s , i ) for s Z [ 1 , T s 1 ] . Therefore, it is obtained that
Δ 1 ( i ) = s = 1 T s 1 ζ ^ * ( s , i ) + A ζ s 1 D ζ q ( i ) Q 2 s = 1 T s 1 ζ ^ * ( s , i ) Q 2 = s = 1 T s 1 ζ ^ * ( s , i ) + A ζ s 1 D ζ q ( i ) Q + ζ ^ * ( s , i ) Q × s = 1 T s 1 ζ ^ * ( s , i ) + A ζ s 1 D ζ q ( i ) Q ζ ^ * ( s , i ) Q s = 1 T s 1 2 ζ ^ * ( s , i ) Q + A ζ s 1 D ζ q ( i ) Q A ζ s 1 D ζ q ( i ) Q s = 1 T s 1 2 λ ¯ 2 ( Q ) Θ ( s ) Ω ( s ) + λ ¯ 2 ( Q ) Ω 2 ( s )
where
Θ ( s ) max b ζ T ζ ^ * ( s , i ) h ζ ϕ ( s , i ) ζ ^ * ( s , i ) , Ω ( s ) max b q T q ( i ) q ¯ A ζ s 1 D ζ q ( i )
For Δ 2 ( i ) , there exists h ˜ ( T s , i + 1 ) = 0 4 × 1 . Then, it follows that
Δ 2 ( i ) ζ ˜ ( T s , i + 1 ) P 2 ζ ^ * ( T s , i ) P 2 + ζ ˜ ( T s , i + 1 ) P 2 ζ ˜ ( T s , i + 1 ) P 2 + ζ ˜ ( T s , i + 1 ) Q 2 = ζ ˜ ( T s , i + 1 ) P 2 ζ ^ * ( T s , i ) P 2 ζ ˜ ( T s , i + 1 ) Q ¯ 2 ζ ^ * ( T s , i ) + A ζ T s 1 D ζ q ( i ) P + ζ ^ * ( T s , i ) P × A ζ T s 1 D ζ q ( i ) P ζ ˜ ( T s , i + 1 ) Q ¯ 2 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s ) ζ ˜ ( T s , i + 1 ) Q ¯ 2
where Q ¯ = P Q A ζ T P A ζ . If there exist appropriate parameters k 1 , k 2 , β 1 , β 2 , β 3 , and a positive definite matrix P, make the LMI
A ζ T P A ζ P + Q < 0
hold, and then ζ ˜ ( T s , i + 1 ) Q ¯ 2 0 . In this case, it follows that
Δ 2 ( i ) 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s )
For Δ 3 ( i ) , it yields Δ 3 ( i ) ζ ^ * ( 0 , i ) Q 2 .
Combining Δ 1 ( i ) , Δ 2 ( i ) , and Δ 3 ( i ) , we have
Δ ˜ V ( i ) = Δ 1 ( i ) + Δ 2 ( i ) + Δ 3 ( i ) ζ ^ * ( 0 , i ) Q 2 + 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s ) + s = 1 T s 1 2 λ ¯ 2 ( Q ) Θ ( s ) Ω ( s ) + λ ¯ 2 ( Q ) Ω 2 ( s )
According to the principle of optimality, the feasible cost is greater than the optimal cost, which indicates Δ V ( i ) Δ ˜ V ( i ) . Then, there exists
Δ V ( i ) ζ ^ * ( 0 , i ) Q 2 + 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s ) + s = 1 T s 1 2 λ ¯ 2 ( Q ) Θ ( s ) Ω ( s ) + λ ¯ 2 ( Q ) Ω 2 ( s )
Letting V ( i ) < 0 , it follows that
ζ ^ * ( 0 , i ) Q 2 > 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s ) + s = 1 T s 1 2 λ ¯ 2 ( Q ) Θ ( s ) Ω ( s ) + λ ¯ 2 ( Q ) Ω 2 ( s )
It indicates that ζ ( i ) = ζ ^ * ( 0 , i ) will converge to the set E ˜ { ζ ( i ) : ζ ( i ) Q 2 σ } under h * ( 0 , i ) as i + with
σ 2 λ ¯ 2 ( P ) Θ ( T s ) Ω ( T s ) + λ ¯ 2 ( P ) Ω 2 ( T s ) + s = 1 T s 1 2 λ ¯ 2 ( Q ) Θ ( s ) Ω ( s ) + λ ¯ 2 ( Q ) Ω 2 ( s )
Next, the proof that ε s ( i ) is uniformly bounded is given. It is easy to get ε s ( i ) = M ζ ( i ) , where
M = 1 0 1 0 0 0 1 0 1 0
Then, as i + , there exists
e x ( + ) = M ζ ( + ) M ζ ( + ) σ ¯
with σ ¯ M λ ̲ 1 2 ( Q ) σ 1 2 , which suggests that ε s ( i ) will converge to the set
E { ε s ( i ) : ε s ( i ) M λ ̲ 1 2 ( Q ) σ 1 2 }
as i + . That is, the tracking error ε s ( i ) is uniformly bounded under the control input (25). This completes the proof. □
Remark 4.
The main advantages of the proposed methodology are twofold: (1) a novel RH-ESO that significantly enhances disturbance estimation accuracy is proposed. By reformulating the disturbance estimation error as prediction states and recursively solving a constrained optimization problem at each sampling instant, this innovative structure guarantees optimal observer gain selection with proven convergence properties. (2) The MPC-enabled disturbance-rejection controller is proposed with disturbance compensation, which enables the closed-loop system to simultaneously achieve prescribed optimal control performance and maintain active disturbance-rejection capabilities.

4. Numerical Example

4.1. Time-Invariant Reference Input Signal Tracking

In this subsection, a numerical example of a pneumatic manipulator system is employed to validate the method and demonstrate its effectiveness. The simulation specifies an initial deflection angle of 0 and a target deflection angle of 10 ; i.e., v 0 ( t ) = 10 . In this part, system (1) is used to demonstrate the effectiveness of the proposed method.
Parameters of the pneumatic manipulator system used in this subsection are presented in Table 1.
In the following, the simulation results of the pneumatic manipulator system using the proposed MPC-enabled disturbance-rejection controller and the ADRC in Figure 2 are used to show the effectiveness of the proposed method in this paper.
Figure 2a plots the deflection angle s 1 ( i ) of the pneumatic manipulator system under the proposed MPC-enabled disturbance-rejection controller and the ADRC, respectively. From Figure 2a, it is evident that the proposed method can make the deflection angle closer to the set 10 . Figure 2b plots the angular velocity s 2 ( i ) of the pneumatic manipulator system.
Figure 2c–e plot the estimation errors ε m ( i ) ( m = 1 , 2 , 3 ), respectively. It is evident that the proposed method achieves smaller estimation error magnitudes in steady-state conditions compared to the traditional ADRC approach. From Figure 2c–e, it is easy to find that the estimation errors under the proposed method demonstrate a smaller magnitude in steady-state conditions compared to that of the ADRC approach. Figure 2f plots the control input of the pneumatic manipulator system. In Figure 2f, it can be found that the proposed method effectively mitigates excessive control signal magnitudes, preventing potential actuator saturation and enhancing operational stability. As a result, from Figure 2a–f, the effectiveness of the proposed method is verified.

4.2. Time-Varying Reference Input Signal Tracking

In this subsection, the simulation specifies an initial deflection angle of 0 and a time-varying target deflection angle of v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) . Furthermore, two simulation comparison groups are considered: (i) the back-stepping control (BSC) method described in [30], and (ii) a classical proportional–derivative (PD) controller. Note that system (37) is used in this subsection to demonstrate the effectiveness of the proposed method.
θ ¨ = R ˜ k 0 2 π N 2 m l ˜ 2 ( 3 L 0 2 b 2 ) u 3 R ˜ 2 L 0 p 0 π N 2 m l ˜ 2 θ g sin θ l ˜ + 3 R ˜ 3 k 0 2 π N 2 m l ˜ 2 θ 2 u + d + f ,
where d = 2 sin t 20 is an external disturbance term and f = sin θ 2 is a model uncertainty term.
The parameters of the pneumatic manipulator system and the MPC scheme are the same as in Table 1. For given reference input signal v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) , the parameters of RH-ESO, ADRC, PD, and BSC are presented in Table 2, and the results are given in Figure 3.
Figure 3a plots the deflection angle of the pneumatic manipulator system (37). Figure 3b plots the deflection angle tracking error of the pneumatic manipulator system (37). Figure 3c plots the decision variables l m ( t ) ( m = 1 , 2 , 3 ) of the RH-ESO (9)–(11). Figure 3d plots the total costs throughout the entire simulation time of the pneumatic manipulator system (37) under the controller of this paper, ADRC, PD, and BSC, respectively. Note that the cost throughout the entire simulation time is calculated by the following formula:
J ( t ) = 0 t ( s 1 ( τ ) v 0 ( τ ) Q 2 + u ( τ ) R 2 ) d τ
where t = 50 , Q = I , and R = 0.01 I .
Under the same input signal, i.e., v 0 ( t ) = sin ( 0.06 × 2 π t + π / 2 ) , the controller gain k 1 of the RH-ESO is increased from 70 to 130. The corresponding simulation results are shown in Figure 4.
Figure 4. Simulation results for reference signal sin ( 0.06 × 2 π t + π / 2 ) of the pneumatic manipulator system using the parameters given in Table 3. (a) Deflection angle of the pneumatic manipulator system. (b) Deflection angle tracking error of the pneumatic manipulator system. (c) Decision variables of the RH-ESO. (d) Costs of the pneumatic manipulator system under these methods.
Figure 4. Simulation results for reference signal sin ( 0.06 × 2 π t + π / 2 ) of the pneumatic manipulator system using the parameters given in Table 3. (a) Deflection angle of the pneumatic manipulator system. (b) Deflection angle tracking error of the pneumatic manipulator system. (c) Decision variables of the RH-ESO. (d) Costs of the pneumatic manipulator system under these methods.
Actuators 14 00343 g004
From Figure 3a and Figure 4a, it can be clearly seen that the method mentioned in [30] exhibits overshoot and vibration phenomena in the early stages of simulation. The reason is that the control quantity fluctuates greatly. Nevertheless, during the early stage of state convergence, the system still experiences considerable overshoot accompanied by minor residual oscillations. This behavior is primarily attributable to the controller structure adopted in [30], which is given as
u ( t ) = k p b 1 ( t ) s 1 ( t ) k i e ˙ ( t ) + α 2 ( t ) 0 t ( ξ 3 ς 3 ( τ ) + ς 2 ( τ ) ) d τ + k p υ ¨ 1 ( t ) k p b 0 s ^ 3 ( t ) b 0
In this controller, an acceleration-level variable v ¨ 1 ( t ) and a velocity-level variable e ˙ ( t ) are employed. On the one hand, v ¨ 1 ( t ) is driven only by the reference signal, yet it exhibits large variations during the initial stage of the tracking differentiator, which seeks to follow the reference rapidly. On the other hand, the rate of change of e ˙ ( t ) in the control law u ( t ) is much higher than that of e ( t ) . When e ( t ) is large, e ˙ ( t ) oscillates frequently with considerable amplitude. These rapid oscillations further induce high-frequency large-magnitude oscillations in u ( t ) , leading to overshoot and oscillations in the state s 1 ( t ) at the outset. As e ( t ) gradually converges, its amplitude diminishes, the amplitude of e ˙ ( t ) likewise decreases, and the oscillations in u ( t ) become weaker and smaller. Consequently, during the subsequent tracking process, s 1 ( t ) no longer exhibits frequent oscillations or large excursions.
From Figure 3a and Figure 4a, it can be clearly seen that the estimated value s ^ 1 ( t ) is oscillating during the simulation. The oscillations in the angle estimate s ^ 1 ( i ) are directly caused by the decision vector l * T ( i ) = l 1 * T i l 2 * T i l 3 * T i T . In the receding-horizon optimization problem proposed here, the variable to be optimized is defined as h ( i ) [ c T ( i ) l T ( i ) ] T . Because the receding-horizon strategy minimizes a cost that explicitly anticipates the evolution of states and decision variables over a future horizon, solving the optimization at time step i immediately yields the optimal decision vector l * ( i ) . At the next instant, however, the optimal vector l * ( i + 1 ) is obtained by re-solving the problem with the updated initial state ζ ¯ ( i + 1 ) [ r ¯ T ( i + 1 ) ε ¯ s T ( i + 1 ) ] T and therefore bears no explicit relation to l * ( i ) . Consequently, the value of l * ( i ) is determined solely by the optimization outcome. The oscillatory behavior of l * ( i ) thus corroborates the role of the receding-horizon mechanism: at each sampling instant, it predicts the state trajectory over the horizon from the current initial state and returns the decision vector that is optimal for that instant.
From Figure 3 and Figure 4, it can be seen that the proposed method achieves smoother bias angle tracking and the lowest overall control cost. Moreover, by comparing the simulation results, it is evident that the proposed method is insensitive to the controller gain k 1 , whereas the traditional ADRC method is more sensitive to these gains. This further demonstrates the effectiveness of the proposed method.

4.3. Analysis of Estimation Error Under Different Observer Initial Values

In this subsection, we investigate the evolution of the estimation error when the RH-ESO initial states differ from that of the system. In Table 4, three initial states of RH-ESO and the same initial value of the system are given.
Firstly, the case in which the reference signal is v 0 ( t ) = 10 is shown. The parameters of the proposed RH-ESO used in this case are those in Table 1. The corresponding results are given in Figure 5, Figure 6 and Figure 7.
From Figure 6 and Figure 7, it is easy to observe that there exist large values of the estimation errors ε 2 ( t ) and ε 3 ( t ) . According to (12)–(14), when the same error signal ε 1 ( i ) is used, the magnitudes of the terms β 2 ε 1 ( i ) and β 3 ε 1 ( i ) greatly exceed that of β 1 ε 1 ( i ) . This results in relatively large values of ε 2 ( i ) and ε 3 ( i ) during the initial phase of the simulation. In addition, another simulation result can be obtained. From Figure 5, Figure 6 and Figure 7, it is easy to obtain that, the closer the initial value of the observer is to the reference signal, the higher the final observation accuracy when the reference is a constant signal.
Secondly, the case in which the reference signal is v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) is shown. The parameters of the proposed RH-ESO in this case are those in Table 2.
The corresponding results are given in Figure 8, Figure 9 and Figure 10.
From Figure 9 and Figure 10, we can find that ε 2 ( t ) and ε 3 ( t ) remain relatively large at the beginning of the simulation. However, they are nevertheless much smaller than those obtained in Figure 6 and Figure 7. This reduction is due to a significant reduction in the RH-ESO gain used in sinusoidal tracking. Therefore, when the observation error shows a large value in the initial stage, the observer gain can be appropriately reduced to reduce the large value in the initial stage.
Furthermore, under these settings, the estimation accuracy is not as good as that achieved by a constant reference, and changing the initial state of the observer will not result in significant differences in the final observation error. The occurrence of this behavior is due to the fact that interference depends on time-varying states and control inputs, and the tracking of disturbance has a lag, resulting in a lag in the estimation of the system state. Therefore, the tracking error cannot converge precisely to zero but instead appears as a sine curve-like image. In general, a high observer gain effectively reduces the steady-state error; however, it also causes excessive overshoot at the beginning of the simulation. Therefore, the observer gain—high or low—should be selected flexibly according to the specific control requirements. In this simulation, the angle tracking error ultimately stays within ±0.05°, and the observation error during the initial phase is small and thus acceptable. Accordingly, the selected RH-ESO parameters are deemed appropriate.

5. Conclusions

This paper has proposed an MPC-enabled disturbance-rejection controller approach for a pneumatic manipulator system subjected to external disturbances. An RH-ESO has been designed that incorporates a decision variable to deal with the disturbances modeled from the complex nonlinear terms within the system. The optimal disturbance estimation error has been determined through a receding-horizon optimization procedure to obtain the best estimate of the disturbance. Using this optimal estimate, the MPC-enabled disturbance-rejection controller has been developed to enable precise angular trajectory tracking in the pneumatic manipulator system. Moreover, the recursive feasibility of Prob 1 has been guaranteed through rigorous proofs. Additionally, the uniform boundedness of the closed-loop system is mathematically guaranteed. Finally, the simulation results demonstrate the superior performance and effectiveness of the proposed approach.

Author Contributions

Conceptualization, Y.X.; methodology, X.H.; software, Y.X.; validation, X.H., D.Z. and L.W.; formal analysis, P.L.; investigation, Y.X.; resources, X.H.; data curation, D.Z.; writing—original draft preparation, Y.X.; writing—review and editing, D.Z.; supervision, D.Z.; project administration, L.W.; funding acquisition, Y.X. and X.H. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank the anonymous reviewers for their detailed comments, which helped to improve the quality of the paper. This work was supported in part by National Natural Science Foundation of China (Grant Nos. 62303061 and 62403353) and National Key Laboratory of Science and Technology on Space-Born Intelligent Information Processing under (Grant No. TJ-02-22-03).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data used in this paper are provided in the Section 4.

Conflicts of Interest

The authors Dongjie Zhu and Liangchao Wu were employed by the company Xingyu Electronics (Ningbo) Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MPCModel predictive control
RH-ESOReceding-horizon-based extended state observer
SMCSliding mode control
ADRCActive disturbance-rejection control
PAMsPneumatic artificial muscles
LMIsLinear matrix inequalities
TDTracking differentiator
Nomenclature
R The real number set
Z 0 The non-negative integer set
Z [ a , b ] The set of positive integers { a , a + 1 , , b }
A R m × n The dimension of A is m × n
A > 0 ( A < 0 ) Matrix A is positive definite (or negative definite)
λ ¯ ( A ) The maximum eigenvalue of A
IThe unit matrix with appropriate dimensions
ζ ( s , i ) The s step ahead prediction of state ζ conditioned on measurements
available at time instant i

References

  1. Xie, Z.; Mohanakrishnan, M.; Wang, P.; Liu, J.; Xin, W.; Tang, Z.; Wen, L.; Laschi, C. Soft robotic arm with extensible stiffening layer. IEEE Robot. Autom. Lett. 2023, 8, 3597–3604. [Google Scholar] [CrossRef]
  2. Low, J.H.; Lee, W.W.; Khin, P.M.; Thakor, N.V.; Kukreja, S.L.; Ren, H.L.; Yeow, C.H. Hybrid tele-manipulation system using a sensorized 3-D-printed soft robotic gripper and a soft fabric-based haptic glove. IEEE Robot. Autom. Lett. 2017, 2, 880–887. [Google Scholar] [CrossRef]
  3. Yu, N.; Zhai, Y.; Yuan, Y.; Wang, Z. A bionic robot navigation algorithm based on cognitive mechanism of hippocampus. IEEE Trans. Autom. Sci. Eng. 2019, 16, 1640–1652. [Google Scholar] [CrossRef]
  4. Chi, H.R.; Radwan, A.; Huang, N.F.; Tsang, K.F. Guest editorial: Next-generation network automation for industrial internet-of-things in Industry 5.0. IEEE Trans. Ind. Inform. 2022, 19, 2062–2064. [Google Scholar] [CrossRef]
  5. Miron, G.; Plante, J.S. Design principles for improved fatigue life of high-strain pneumatic artificial muscles. Soft Robot. 2016, 3, 177–185. [Google Scholar] [CrossRef]
  6. Kim, W.; Park, H.; Kim, J. Compact flat fabric pneumatic artificial muscle (ffpam) for soft wearable robotic devices. IEEE Robot. Autom. Lett. 2021, 6, 2603–2610. [Google Scholar] [CrossRef]
  7. Tsai, T.C.; Chiang, M.H. A lower limb rehabilitation assistance training robot system driven by an innovative pneumatic artificial muscle system. Soft Robot. 2023, 10, 1–16. [Google Scholar] [CrossRef] [PubMed]
  8. Cho, Y.; Kim, W.; Park, H.; Kim, J.; Na, Y. Bidirectional double-spring pneumatic artificial muscle with inductive self-sensing. IEEE Robot. Autom. Lett. 2023, 8, 8160–8167. [Google Scholar] [CrossRef]
  9. Liu, G.; Diao, S.; Liu, Z.; Zhang, X.; Xiao, X.; Men, S.; Sun, N. Practical Finite-Time Compliant Control for Horizontal Pneumatic Artificial Muscle Systems Under Force-Sensorless Reflecting. IEEE Trans. Autom. Sci. Eng. 2024, 22, 9515–9527. [Google Scholar] [CrossRef]
  10. Wang, Q.; Yang, T.; Liu, G.; Qin, Y.; Fang, Y.; Sun, N. Adaptive compensation tracking control for parallel robots actuated by pneumatic artificial muscles with error constraints. IEEE Trans. Ind. Inform. 2023, 20, 1585–1595. [Google Scholar] [CrossRef]
  11. Liang, D.; Sun, N.; Wu, Y.; Liu, G.; Fang, Y. Fuzzy-sliding mode control for humanoid arm robots actuated by pneumatic artificial muscles with unidirectional inputs, saturations, and dead zones. IEEE Trans. Ind. Inform. 2021, 18, 3011–3021. [Google Scholar] [CrossRef]
  12. Khaled, T.A.; Akhrif, O.; Bonev, I.A. Dynamic path correction of an industrial robot using a distance sensor and an ADRC controller. Ieee/Asme Trans. Mechatronics 2020, 26, 1646–1656. [Google Scholar] [CrossRef]
  13. Diao, S.; Liu, G.; Liu, Z.; Zhou, L.; Sun, W.; Wang, Y.; Sun, N. Prescribed-Time Adaptive Fuzzy Control for Pneumatic Artificial Muscle-Actuated Parallel Robots With Input Constraints. IEEE Trans. Fuzzy Syst. 2023, 32, 2039–2051. [Google Scholar] [CrossRef]
  14. Qin, Y.; Zhang, H.; Wang, X.; Sun, N.; Han, J. Adaptive set-membership filter based discrete sliding mode control for pneumatic artificial muscle systems with hardware experiments. IEEE Trans. Autom. Sci. Eng. 2023, 21, 1682–1694. [Google Scholar] [CrossRef]
  15. Hosseini, S.A.; Toulabi, M.; Dobakhshari, A.S.; Ashouri-Zadeh, A.; Ranjbar, A.M. Delay compensation of demand response and adaptive disturbance rejection applied to power system frequency control. IEEE Trans. Power Syst. 2019, 35, 2037–2046. [Google Scholar] [CrossRef]
  16. Zeng, Y.; Liang, G.; Liu, Q.; Rodriguez, E.; Pou, J.; Jie, H.; Liu, X.; Zhang, X.; Kotturu, J.; Gupta, A. Multiagent soft actor-critic aided active disturbance rejection control of DC solid-state transformer. IEEE Trans. Ind. Electron. 2024, 72, 492–503. [Google Scholar] [CrossRef]
  17. Fu, C.; Tan, W. Tuning of linear ADRC with known plant information. ISA Trans. 2016, 65, 384–393. [Google Scholar] [CrossRef]
  18. Guo, B.Z.; Zhao, Z.l. On the convergence of an extended state observer for nonlinear systems with uncertainty. Syst. Control Lett. 2011, 60, 420–430. [Google Scholar] [CrossRef]
  19. Qin, B.; Yan, H.; Zhang, H.; Wang, Y.; Yang, S.X. Enhanced reduced-order extended state observer for motion control of differential driven mobile robot. IEEE Trans. Cybern. 2021, 53, 1299–1310. [Google Scholar] [CrossRef]
  20. Li, Z.; Yan, H.; Zhang, H.; Yang, S.X.; Chen, M. Novel extended state observer design for uncertain nonlinear systems via refined dynamic event-triggered communication protocol. IEEE Trans. Cybern. 2022, 53, 1856–1867. [Google Scholar] [CrossRef]
  21. Sun, H.; Madonski, R.; Li, S.; Zhang, Y.; Xue, W. Composite control design for systems with uncertainties and noise using combined extended state observer and Kalman filter. IEEE Trans. Ind. Electron. 2021, 69, 4119–4128. [Google Scholar] [CrossRef]
  22. Zhao, L.; Li, Q.; Liu, B.; Cheng, H. Trajectory tracking control of a one degree of freedom manipulator based on a switched sliding mode controller with a novel extended state observer framework. IEEE Trans. Syst. Man Cybern. Syst. 2017, 49, 1110–1118. [Google Scholar] [CrossRef]
  23. Zhao, L.; Cheng, H.; Wang, T. Sliding mode control for a two-joint coupling nonlinear system based on extended state observer. ISA Trans. 2018, 73, 130–140. [Google Scholar] [CrossRef]
  24. Mayne, D.Q.; Rawlings, J.B.; Rao, C.V.; Scokaert, P.O. Survey constrained model predictive control: Stability and optimality. Automatica 2000, 36, 789–814. [Google Scholar] [CrossRef]
  25. Zeilinger, M.N.; Morari, M.; Jones, C.N. Soft constrained model predictive control with robust stability guarantees. IEEE Trans. Autom. Control 2014, 59, 1190–1202. [Google Scholar] [CrossRef]
  26. Li, T.; Sun, X.; Lei, G.; Guo, Y.; Yang, Z.; Zhu, J. Finite-control-set model predictive control of permanent magnet synchronous motor drive systems—An overview. IEEE/CAA J. Autom. Sin. 2022, 9, 2087–2105. [Google Scholar] [CrossRef]
  27. Li, P.; Kang, Y.; Wang, T.; Zhao, Y.B. Disturbance prediction-based adaptive event-triggered model predictive control for perturbed nonlinear systems. IEEE Trans. Autom. Control 2022, 68, 2422–2429. [Google Scholar] [CrossRef]
  28. Li, N.; Zhang, K.; Li, Z.; Srivastava, V.; Yin, X. Cloud-assisted nonlinear model predictive control for finite-duration tasks. IEEE Trans. Autom. Control 2022, 68, 5287–5300. [Google Scholar] [CrossRef]
  29. Wang, M.; Cheng, P.; Zhang, Z.; Wang, M.; Chen, J. Periodic event-triggered MPC for continuous-time nonlinear systems with bounded disturbances. IEEE Trans. Autom. Control 2023, 68, 8036–8043. [Google Scholar] [CrossRef]
  30. Zhao, L.; Li, Z.; Li, H.; Liu, B. Backstepping integral sliding mode control for pneumatic manipulators via adaptive extended state observers. ISA Trans. 2024, 144, 374–384. [Google Scholar] [CrossRef]
  31. Peuteman, J.; Aeyels, D.; Sepulchre, R. Boundedness properties for time-varying nonlinear systems. SIAM J. Control Optim. 2001, 39, 1408–1422. [Google Scholar] [CrossRef]
Figure 1. Structure diagram of the pneumatic manipulator system.
Figure 1. Structure diagram of the pneumatic manipulator system.
Actuators 14 00343 g001
Figure 2. Simulation results for reference signal v 0 ( t ) = 10 of the pneumatic manipulator system. (a) The state s 1 ( i ) under the proposed method and the ADRC. (b) The state s 2 ( i ) under the proposed method and the ADRC. (c) The estimation error ε 1 ( i ) under the method in this paper and the ADRC. (d) The estimation error ε 2 ( i ) under the method in this paper and the ADRC. (e) The estimation error ε 3 ( i ) under the method in this paper and the ADRC. (f) The control input u ( i ) under the method in this paper and the ADRC.
Figure 2. Simulation results for reference signal v 0 ( t ) = 10 of the pneumatic manipulator system. (a) The state s 1 ( i ) under the proposed method and the ADRC. (b) The state s 2 ( i ) under the proposed method and the ADRC. (c) The estimation error ε 1 ( i ) under the method in this paper and the ADRC. (d) The estimation error ε 2 ( i ) under the method in this paper and the ADRC. (e) The estimation error ε 3 ( i ) under the method in this paper and the ADRC. (f) The control input u ( i ) under the method in this paper and the ADRC.
Actuators 14 00343 g002
Figure 3. Simulation results for reference signal v 0 ( t ) = sin ( 0.06 × 2 π t + π / 2 ) of the pneumatic manipulator system using the parameters given in Table 2. (a) Deflection angle of the pneumatic manipulator system. (b) Deflection angle tracking error of the pneumatic manipulator system. (c) Decision variables of the RH-ESO. (d) Costs of the pneumatic manipulator system under these methods.
Figure 3. Simulation results for reference signal v 0 ( t ) = sin ( 0.06 × 2 π t + π / 2 ) of the pneumatic manipulator system using the parameters given in Table 2. (a) Deflection angle of the pneumatic manipulator system. (b) Deflection angle tracking error of the pneumatic manipulator system. (c) Decision variables of the RH-ESO. (d) Costs of the pneumatic manipulator system under these methods.
Actuators 14 00343 g003
Figure 5. The estimation error ε 1 ( t ) under different initial values of the proposed RH-ESO.
Figure 5. The estimation error ε 1 ( t ) under different initial values of the proposed RH-ESO.
Actuators 14 00343 g005
Figure 6. The estimation error ε 2 ( t ) under different initial values of the proposed RH-ESO.
Figure 6. The estimation error ε 2 ( t ) under different initial values of the proposed RH-ESO.
Actuators 14 00343 g006
Figure 7. The estimation error ε 3 ( t ) under different initial values of the proposed RH-ESO.
Figure 7. The estimation error ε 3 ( t ) under different initial values of the proposed RH-ESO.
Actuators 14 00343 g007
Figure 8. The estimation error ε 1 ( t ) under different initial values of the proposed RH-ESO.
Figure 8. The estimation error ε 1 ( t ) under different initial values of the proposed RH-ESO.
Actuators 14 00343 g008
Figure 9. The estimation error ε 2 ( t ) under different initial values of the proposed RH-ESO.
Figure 9. The estimation error ε 2 ( t ) under different initial values of the proposed RH-ESO.
Actuators 14 00343 g009
Figure 10. The estimation error ε 3 ( t ) under different initial values of the proposed RH-ESO.
Figure 10. The estimation error ε 3 ( t ) under different initial values of the proposed RH-ESO.
Actuators 14 00343 g010
Table 1. Parameters used in the simulation.
Table 1. Parameters used in the simulation.
Parameter R ˜ l ˜ L 0 b
Value 0.1 (m) 0.3 (m) 0.3 (m) 0.3 (m)
Parameter p 0 mg T
Value1 (kpa)1 (kg) 9.8 ( m / s 2 ) 0.01 (s)
Parameter k 0 N s ¯ 1 s ¯ 2
Value10081530
Parameter u ¯ ω ¯ d h ε 1 h ε 2
Value100010110
Parameter h ε 3 β 1 β 2 β 3
Value10030070001750
Parameter k 1 k 2 r 1 r 2
Value100302050
Parameterd T s QR
Value15 I 8 × 8 0.5 I 4 × 4
Table 2. Reference signal of v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) .
Table 2. Reference signal of v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) .
Parameters of RH-ESO and ADRC β 1 β 2 β 3 k 1 k 2
Value 1505501007010
Parameters of PD k p k d
Value 1002
Parameters of BSC ξ 1 ξ 2 ξ 3 β 1 β 2 β 3 k p k i
Value222150100010050 21.7
Table 3. Reference signals of v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) .
Table 3. Reference signals of v 0 ( t ) = 10 sin ( 0.06 × 2 π t + π / 2 ) .
Parameters of RH-ESO and ADRC β 1 β 2 β 3 k 1 k 2
Value 15055010013010
Parameters of PD k p k d
Value 1002
Parameters of BSC ξ 1 ξ 2 ξ 3 β 1 β 2 β 3 k p k i
Value222150100010050 21.7
Table 4. Initial states of the system and the RH-ESO.
Table 4. Initial states of the system and the RH-ESO.
Initial states of the systemVariable s 1 ( 0 ) s 2 ( 0 )
Value00
Case 1: Initial states of the RH-ESOVariable s ^ 1 ( 0 ) s ^ 2 ( 0 )
Value00
Case 2: Initial states of the RH-ESOVariable s ^ 1 ( 0 ) s ^ 2 ( 0 )
Value50
Case 3: Initial states of the RH-ESOVariable s ^ 1 ( 0 ) s ^ 2 ( 0 )
Value80
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, Y.; Hao, X.; Zhu, D.; Wu, L.; Li, P. Model Predictive Control for Pneumatic Manipulator via Receding-Horizon-Based Extended State Observers. Actuators 2025, 14, 343. https://doi.org/10.3390/act14070343

AMA Style

Xu Y, Hao X, Zhu D, Wu L, Li P. Model Predictive Control for Pneumatic Manipulator via Receding-Horizon-Based Extended State Observers. Actuators. 2025; 14(7):343. https://doi.org/10.3390/act14070343

Chicago/Turabian Style

Xu, Yang, Xiaohui Hao, Dongjie Zhu, Liangchao Wu, and Peng Li. 2025. "Model Predictive Control for Pneumatic Manipulator via Receding-Horizon-Based Extended State Observers" Actuators 14, no. 7: 343. https://doi.org/10.3390/act14070343

APA Style

Xu, Y., Hao, X., Zhu, D., Wu, L., & Li, P. (2025). Model Predictive Control for Pneumatic Manipulator via Receding-Horizon-Based Extended State Observers. Actuators, 14(7), 343. https://doi.org/10.3390/act14070343

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop