You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

1 December 2025

Hybrid Partial-Data-Driven H Robust Tracking Control for Linear Stochastic Systems with Discrete-Time Observation of Reference Trajectory

,
and
1
College of Mathematics and Systems Science, Shandong University of Science and Technology, Qingdao 266590, China
2
College of Electronic and Information Engineering, Shandong University of Science and Technology, Qingdao 266590, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Stochastic System Analysis and Control

Abstract

A hybrid robust H tracking-control design method is studied for linear stochastic systems in which the parameters of the reference system are unknown but inferred from discrete-time observations. First, the reference system parameters are estimated by the least-squares method, and a corresponding data-dependent augmented system is constructed. Second, a Riccati matrix inequality is established for these systems, and a state-feedback H controller is designed to improve tracking performance. Third, to mitigate large tracking errors, an error-feedback control scheme is introduced to compensate for dynamic tracking deviations. These results yield a hybrid control framework that integrates data observation, state-feedback H control, and error-feedback H control to address the tracking problem more effectively. Two numerical examples and one practical example demonstrate the effectiveness of the proposed method.

1. Introduction

With advancements in artificial intelligence and data-acquisition technology, data-driven control has become a prominent paradigm in control engineering [1]. Numerous important results have emerged. For instance, Shen et al. studied iterative learning control (ILC) for discrete-time linear systems without prior probability information on randomly varying iteration length [2]. The authors of Ref. [3] proposed a proportional–integral–derivative (PID) control scheme based on adaptive updating rules and data-driven techniques. In [4], bounded-input bounded-output stability, monotonic convergence of tracking-error dynamics, and internal stability of the full-form dynamic-linearization-based model-free adaptive control (MFAC) scheme were analyzed using the contraction mapping principle. The authors of [5] proposed a data-driven adaptive control method based on the incremental triangular dynamic-linearization data model. Recently, data-driven modeling methods have advanced rapidly, yielding notable results in several areas, including predictive control for switched linear systems [6], predictive control for a modular multilevel converter [7], optimal output tracking control [8], self-triggered control [9], and distributed predictive control [10].
The H control technique is an effective method for enhancing robustness against exogenous disturbances [11] and has been widely applied in aerospace, robotics, and wireless communications [12,13,14,15]. Robust H control theory has evolved over more than 40 years. The foundational works include [16,17], which introduce the frequency-domain and algebraic Riccati equation (ARE) methods, respectively, for linear deterministic systems. The linear matrix inequality (LMI) approach was later developed in [18]. For systems subject to random noise [19], stochastic differential equations have been adopted [20], and corresponding H results have been extended to linear Itô systems [21]. H control theories for nonlinear systems have also been established, such as those in [22] for deterministic settings and those in [23]. State-feedback H control for affine stochastic Itô systems was studied in [24] using the completing-squares method. More recently, we proposed an H control design method for general nonlinear discrete-time stochastic systems using the disintegration property of conditional expectation and convex functions [25]. Additional results on H control can be found in [26,27,28] and the references therein.
Tracking-control design techniques are widely used in autonomous underwater vehicles (AUVs), unmanned surface vehicles, and unmanned aerial vehicles (UAVs). For example, Ref. [29] investigated trajectory-tracking control for underactuated autonomous underwater vehicles subject to input constraints and arbitrary attitude maneuvers. A delay-compensated control framework with prescribed performance guarantees was proposed in [30] for trajectory-tracking control of unmanned underwater vehicles (UUVs) subject to uncertain time-varying input delays. The distributed model-predictive control (MPC) framework in [31] was designed for tracking-control problems involving multiple unmanned aerial vehicles (UAVs) interconnected through a directed communication graph. To address size-related limitations in computing invariant sets and to simplify the offline model-predictive control (MPC) design, the authors of Ref. [32] developed an MPC technique based on implicit terminal components. Optimal H adaptive fuzzy observer-based indirect reference tracking-control designs were investigated in [33] for uncertain SISO and MIMO nonlinear stochastic systems with nonlinear uncertain measurement functions and measurement noise.
This paper investigates a hybrid partial-data-driven H robust tracking-control method for the following linear stochastic system to be controlled:
d x ( t ) = [ A x ( t ) + B u ( t ) + B 1 v 1 ( t ) ] d t + [ A 1 x ( t ) + B 2 v 2 ( t ) ] d W ( t ) , t [ t 0 , T f ] ,
together with the unknown reference model:
x ˙ r ( t ) = A r x r ( t ) + v 3 ( t ) ,
where x ( t ) is the system state, x r ( t ) is the reference trajectory, u ( t ) R n u is the control input, and v 1 ( t ) , v 2 ( t ) and v 3 ( t ) are the exogenous disturbances. The matrices A , B , B 1 , A 1 , and B 2 of the system to be controlled are known, whereas the matrix A r of the reference system is unknown. Only some discrete-time observations of x r ( t ) are available in t = t 0 , t 1 , , t N . To reduce the tracking error, i.e., the distance between x ( t ) and x r ( t ) , the control input u ( t ) must be properly designed.
A data-dependent augmented system is constructed, and state-feedback and error-feedback H controllers are developed accordingly. The resulting hybrid partial-data-driven H control design scheme comprises three stages: data observation, state-feedback H control, and error-feedback H control. In the first stage, information on x r ( t ) is obtained, the reference system parameters are estimated, and the corresponding data-dependent augmented system is constructed, from which a state-feedback H controller is derived. In the second stage, under the action of the state-feedback controller, the tracking error is reduced, although its magnitude remains relatively large because the control depends only on the system state and does not incorporate tracking-error values. In the third stage, an error-feedback H controller is designed to further reduce the tracking error.
Compared with the method of [33], the proposed approach has two main innovations:
  • Tracking errors are incorporated not only in the performance index but also directly in the control input through an error-feedback term.
  • The control input u ( t ) is a hybrid partial-data-driven controller with a piecewise structure.
This paper is organized as follows: In Section 2, some results on H control for linear stochastic systems are reviewed. In Section 3, the data-dependent augmented system is constructed, and the corresponding state-feedback H control and error-feedback H control design methods are studied by solving algebraic Riccati inequalities. In Section 4, the hybrid control scheme is proposed, which includes three stages, and the corresponding programming evolution is also provided. In Section 5, two numerical examples and one practical example are discussed to illustrate the effectiveness of the proposed method.
Notations: A T : the transpose of the matrix A; | x | : the Euclidean norm of the vector x R n ; tr ( A ) : the trace of square matrix A = ( a i j ) R n × n with tr ( A ) = i = 1 n a i i ; M : the norm of matrix M R m × n defined by M = tr ( M T M ) ; n y : the dimension of vector y R n y ; S n ( R ) : the set of the n-order real symmetric matrix; S + n ( R ) : the set of the n-order positive definite matrix; M > 0 ( M 0 ) : the matrix M is a positive definite (semi-definite) matrix; M 1 2 : the square root of a positive definite (semi-definite) matrix M; i.e., M 1 2 is a positive (semi-definite) matrix satisfying M = M 1 2 M 1 2 . v M : the norm of vector v R n v with weighted matrix M defined by v M = v T M v , where M > 0 .

2. Preliminaries

Let { W ( t ) , t 0 } be a 1-dimensional standard Brownian motion in a completed probability space ( Ω , F , F , P ) with E [ W ( t ) ] = 0 and E [ W ( t ) 2 ] = t . The filtration F = { F t , t 0 } satisfies the usual conditions, and F t is generated by Brownian motion, i.e.,
F t = σ { W s : 0 s t } N
where N is the totality of P -null sets. Denote L 2 ( Ω , F t , P ; R n ) as the set of random variables ξ , where ξ is F t -measurable and
E [ | ξ | 2 ] < .
Denote M F 2 ( [ t 0 , T f ] ; R n y ) as the set of F -adapted stochastic processes y = { y t , t 0 t T f } with norm
y M F 2 ( [ t 0 , T f ] ; R n y ) = E t 0 T f | y t | 2 d t < .
The stochastic linear system with exogenous disturbance v ( t ) is given as
d x ( t ) = ( A x ( t ) + B 1 v ( t ) ) d t + ( A 1 x ( t ) + B 2 v ( t ) ) d W ( t ) x ( t 0 ) = x 0 R n , t [ t 0 , T f ] .
where A , A 1 R n × n , B 1 , B 2 R n × n v are coefficient matrices, v ( t ) R n v is the exogenous disturbance, and T f is the terminal time with 0 < T f < . We first review some results of H problem for system (1) that will be used later.
Denote x ( t ; v , t 0 , x 0 ) the solutions of (1) beginning at t 0 with initial state x 0 under exogenous disturbance v = { v ( t ) , t 0 t T f } . For every v ( t ) M F 2 ( [ t N , T f ] ; R n v ) , let
z ( t ; v , t 0 , x 0 ) = C x ( t ; v , t 0 , x 0 ) + D v ( t )
be the output of system (1). Define operator L : M F 2 ( [ t 0 , T f ] ; R n v ) M F 2 ( [ t 0 , T f ] ; R n z ) with
L ( v ) = z ,
where z ( t ) = z ( t ; v , t 0 , 0 ) . The H norm of L is defined as
L = sup v 0 z M F 2 ( [ t 0 , T f ] ; R n z ) v M F 2 ( [ t 0 , T f ] ; R n v ) .
If there exists a constant ρ > 0 such that, for every v ( t ) M F 2 ( [ t 0 , T f ] ; R n v ) ,
z M F 2 ( [ t 0 , T f ] ; R n z ) ρ v M F 2 ( [ t 0 , T f ] ; R n v ) ,
i.e.,
E t 0 T f | z ( t ) | 2 d t ρ 2 E t 0 T f | v ( t ) | 2 d t ,
then (2) and (3) are equivalent to
L ρ .
The following lemma is the bounded real lemma from [34], which is also suitable for the case of H problem for system (1).
Lemma 1. 
For some given positive scalar ρ > 0 , suppose there exists a positive definite matrix P S + n ( R ) such that
P A + A T P + A 1 T P A 1 P B 1 + A 1 T P B 2 C T ( ρ 2 I B 2 P B 2 ) D T I 0 ,
wheredenotes the symmetrical part, and then L ρ . Moreover, there also exists
E t 0 T f | z ( t ; t 0 , v , ξ ) | 2 d t E [ x 0 T P x 0 ] + ρ 2 E t 0 T f | v ( t ) | 2 d t ,
for all x 0 R n and v ( t ) M F 2 ( [ t 0 , T f ] ; R n v ) .

3. State-Feedback and Error-Feedback H Tracking Control for Linear Systems Based on Partially Observable Data

The following linear control systems are considered:
d x ( t ) = [ A x ( t ) + B u ( t ) + B 1 v 1 ( t ) ] d t + [ A 1 x ( t ) + B 2 v 2 ( t ) ] d W ( t ) , x ( 0 ) = x 0 R n , t [ 0 , T f ] ,
where A , A 1 R n × n , B R n × n u , B 1 R n × n v 1 , B 2 R n × n v 2 , C R n z × n , and D R n z × n are coefficient matrices, x ( t ) R n is the system state, u ( t ) R n u is the controller, and v 1 ( t ) R n v 1 and v 2 ( t ) R n v 2 are the exogenous disturbances.
Suppose the tracking target of system (5) can be described by the following reference model:
x ˙ r ( t ) = A r x r ( t ) + v 3 ( t ) ,
where x r ( t ) R n is the desired reference state tracked by x ( t ) in (5), A r R n × n is the coefficient matrix, and v 2 ( t ) R n v 2 is the exogenous disturbance of reference system (6). In system (6), the system coefficient A r is unknown, but x r ( t ) can be observed at discrete-time t 0 , t 1 , , t N with t 0 < t 1 < < t N T f ; i.e., the observation values of x r ( t 0 ) , x r ( t 1 ) , x r ( t N ) can be obtained.
Denote e r ( t ) as the tracking errors between x ( t ) and x r ( t ) ; i.e.,
e r ( t ) = x ( t ) x r ( t ) .
Our target is, with the background of reference system coefficients unknown and only based on the obtained discrete-time observation x r ( t 0 ) , x r ( t 1 ) , x r ( t N ) , to find an H controller u ( t ) such that, for the given ρ > 0 , the tracking performance always satisfies the following inequality in the next coming time [ t N , T f ] given as
t N T f E [ e r T ( t ) Q e r ( t ) + u T ( t ) R u ( t ) ] d t ρ 2 t N T f E [ v T ( t ) v ( t ) ] d t ,
for all exogenous disturbances v ( t ) = [ v 1 ( t ) T , v 2 ( t ) T , v 3 ( t ) T ] T M F 2 ( [ t N , T f ] ; R n v ) , where Q 0 is the weighted matrix of tracking errors e ( t ) , R > 0 is the weighted matrix of controller u ( t ) , T f > 0 is the terminal time of control, and ρ is the attenuation level of external interference on the system.
Denote
ξ r ( t k ) = x r ( t k + 1 ) x r ( t k ) t k + 1 t k , k = 0 , 1 , , N 1 .
In order to obtain the H control u ( t ) , the unknown coefficient matrix A r in system (6) should be estimated first. Because the trajectory x r ( t ) of reference system (6) can be observed at discrete-time t 0 , t 1 , , t N , in order to overcome the system’s uncertainty regarding (6), the observation values of x r ( t 0 ) , x r ( t 1 ) , x r ( t N ) are used to estimate the unknown matrix A r . This requires the discrete form of (6) as follows:
x r ( t k + 1 ) x r ( t k ) t k + 1 t k = A r x r ( t k ) + ϵ k , k = 0 , 1 , 2 , , N 1 ,
i.e.,
ξ r ( t k ) = A r x r ( t k ) + ϵ k , k = 0 , 1 , 2 , , N 1 ,
where { ϵ k , k = 0 , 1 , , N 1 } are the fitting errors or residuals of this model. The least-squares method is used to estimate matrix A r , which satisfies
A ^ r = arg min A r TSSE ( A r ) ,
where TSSE is the total sum of squared errors with
TSSE ( A r ) = k = 0 N 1 ϵ k 2 .
Theorem 1. 
Let X r = [ x r ( t 0 ) , x r ( t 1 ) , x r ( t N 1 ) ] and Ξ r = [ ξ r ( t 0 ) , ξ r ( t 1 ) , ξ r ( t N 1 ) ] ; then, the least-squares estimator of matrix A r can be presented by
A ^ r = Ξ r X r T ( X r X r T ) .
which satisfies (9), where ( X r X r T ) is the Moore–Penrose inverse matrix of X r X r T .
Proof. 
Applying the discrete-time model (8), we can obtain
ϵ k = ξ r ( t k ) A r x ( t k ) , k = 0 , 1 , 2 , , N 1 .
By the definition of TSSE ( A r ) , there exists
TSSE ( A r ) = k = 0 N 1 [ ξ r ( t k ) A r x ( t k ) ] T [ ξ r ( t k ) A r x ( t k ) ] = t r ( ( Ξ r T X r T A r T ) ( Ξ r A r X r ) ) = Ξ r A r X r 2 .
So, the least-squares estimator of A r is given by (11). □
Remark 1. 
In Theorem 1, the Moore–Penrose inverse matrix is used. As proposed by the anonymous reviewer, the Moore–Penrose procedure is insufficient. But, if the sample size N + 1 is large enough, the matrix X r X r T can be an invertible matrix, which is shown in Examples 1 and 2. If X r X r T is invertible, then it is equivalent to R ( X r X r T ) = n , where R ( X r X r T ) denotes the rank of matrix X r X r T . Because R ( X r X r T ) R ( X r ) and X r is n × N matrix, the necessary condition for X r X r T to be invertible is N n . So, if X r X r T is invertible, the result of (11) can be replaced by
A r ^ = Ξ r X r T ( X r X r T ) 1 .
So, based on the results of Theorem 1, the parameter-unknown reference system (6) can be estimated with the observed data form
x ˙ r ( t ) = Ξ r X r T ( X r X r T ) x r ( t ) + v 3 ( t ) .
The augment system of (5) and (12) is written by
d x ˜ ( t ) = [ A ˜ x ˜ ( t ) + B ˜ u ( t ) + B ˜ 1 v ( t ) ] d t + [ A 1 ˜ x ˜ ( t ) + B ˜ 2 v ( t ) ] d W ( t ) x ˜ ( t N ) = x ( t N ) x r ( t N ) , t [ t N , T f ]
where
x ˜ ( t ) = x ( t ) x r ( t ) ,   v ( t ) = v 1 ( t ) v 2 ( t ) v 3 ( t ) ,   A ˜ = A 0 0 Ξ r X r T ( X r X r T ) ,   B ˜ = B 0 ,  
A 1 ˜ = A 1 0 0 0 ,   B ˜ 1 = B 1 0 0 0 0 I n v 3 ,   B ˜ 2 = 0 B 2 0 0 0 0 .
So, the corresponding performance of (13) is rewritten as
t N T f E [ x ˜ T ( t ) Q ˜ x ˜ ( t ) + u T ( t ) R u ( t ) ] d t ρ 2 t N T f E [ v T ( t ) v ( t ) ] d t
where
Q ˜ = Q Q Q Q .
Remark 2. 
Denote the observations of the reference system (6) at discrete-time t 0 , t 1 , , t N as
X = [ x r ( t 0 ) , x r ( t 1 ) , x r ( t N ) ] ,
and then the coefficients of Ξ r and X r in (12) depend on such observation data X r . So, some coefficients of the augment system (13) such as A ˜ are also data-dependent where the data X r is only the observations of x r ( t ) at discrete-time t 0 , t 1 , , t N . Because such data X r is only the observation of reference system (6) but not including the to-be-controlled system (5), and X r is only the data at discrete time at t 0 , t 1 , , t N but not the continuous interval in [ t 0 , t N ] , so the to-be-designed control is called partial-data-driven, which depends on such data X r . Now, we outline the definition of partial-data-driven H tracking control of systems (5) and (6) in which the performance is more general than (14).
Definition 1. 
For the given scalar ρ > 0 and the observations X r of reference system (6), if there exists a positive definite matrix P and a control u ( t ) , t [ t N , T f ] such that the solutions of x ˜ ( t ) for augmented system (13) satisfy the following inequality
t N T f E [ x ˜ T ( t ) Q ˜ x ˜ ( t ) + u T ( t ) R u ( t ) ] d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + ρ 2 t N T f E [ v T ( t ) v ( t ) ] d t ,
for all x ˜ ( t N ) L 2 ( Ω , F t N , P ; R 2 n ) , v ( t ) M F 2 ( [ t N , T f ] ; R n v ) , then u ( t ) is called the partial-data-driven H tracking control of systems (5) and (6).
Theorem 2. 
For the given positive number ρ > 0 , suppose the positive definite matrix P > 0 satisfies the following Riccati inequality:
H 1 ( P ) : = A ˜ T P + P A ˜ + A 1 ˜ T P A 1 ˜ + Q ˜ P B ˜ R 1 B ˜ T P + ( A ˜ 1 T P B ˜ 2 + P B ˜ 1 ) ( ρ 2 I B ˜ 2 T P B ˜ 2 ) 1 ( B ˜ 2 T P A ˜ 1 + B ˜ 1 T P ) 0 , ρ 2 I B ˜ 2 T P B ˜ 2 > 0 .
Then, the state-feedback control u ( t ) = R 1 B ˜ T P x ˜ ( t ) is the partial-data-driven H tracking control of systems (5) and (6).
Proof. 
Let V ( x ˜ ) = x ˜ T P x ˜ . Applying Itô’s formula to V ( x ˜ ( t ) ) , we have
V ( x ˜ ( T f ) ) V ( x ˜ ( t N ) ) = t N T f [ x ˜ T ( t ) ( A ˜ T P + P A ˜ + A 1 ˜ T P A 1 ˜ ) x ˜ ( t ) + 2 x ˜ T ( t ) P B ˜ u ( t )                     + 2 x T ( t ) ( P B ˜ 1 v ( t ) + A ˜ 1 T P B ˜ 2 ) v ( t ) + v T ( t ) B ˜ 2 T P B ˜ 2 v ( t ) ] d t   + t N T f 2 x ˜ T ( t ) P [ A 1 ˜ x ˜ ( t ) + B ˜ 2 v ( t ) ] d W ( t ) .
Taking expectation on both sides, we get
E [ V ( x ˜ ( T f ) ) ] E [ V ( x ˜ ( t N ) ) ] = E t N T f [ x ˜ T ( t ) ( A ˜ T P + P A ˜ + A 1 ˜ T P A 1 ˜ ) x ˜ ( t ) + 2 x ˜ T ( t ) P B ˜ u ( t ) + 2 x T ( t ) ( P B ˜ 1 v ( t ) + A ˜ 1 T P B ˜ 2 ) v ( t ) + v T ( t ) B ˜ 2 T P B ˜ 2 v ( t ) ] d t .
Since P 0 , there exists V ( x ˜ ( T f ) ) 0 . So, we have
t N T f E [ x ˜ T ( t ) Q ˜ x ˜ ( t ) + u ( t ) T R u ( t ) ] d t E [ V ( x ˜ ( t N ) ) ] + ρ 2 E t N T f | v ( t ) | 2 d t   + E t N T f [ x ˜ T ( t ) ( A ˜ T P + P A ˜ + A 1 ˜ T P A 1 ˜ + Q ) x ˜ ( t ) + 2 x ˜ T ( t ) P B ˜ u ( t )   + u ( t ) T R u ( t ) + 2 x T ( t ) ( P B ˜ 1 + A ˜ 1 T P B ˜ 2 ) v ( t ) v T ( t ) ( ρ 2 I B ˜ 2 T P B ˜ 2 ) v ( t ) ] d t .
By completing-squares method, we have
t N T f E [ x ˜ T ( t ) Q ˜ x ˜ ( t ) + u ( t ) T R u ( t ) ] d t E [ V ( x ˜ ( t N ) ) ] + ρ 2 E t N T f | v ( t ) | 2 d t + E t N T f [ x ˜ T ( t ) H 1 ( P ) x ˜ T ( t ) + u ( t ) + R 1 B T P x R 2 v ( t ) ( ρ 2 I B ˜ 2 T P B ˜ 2 ) 1 ( B ˜ 1 T P + B ˜ 2 T P A ˜ 1 ) x ( t ) M 2 ] d t .
where M = ρ 2 I B ˜ 2 T P B ˜ 2 . Taking u = R 1 B T P x and combining with inequality (17), the following inequality is obtained:
t N T f E [ x ˜ T ( t ) Q ˜ x ˜ ( t ) + u ( t ) T R u ( t ) ] d t E [ V ( x ˜ ( t N ) ) ] + ρ 2 E t N T f | v ( t ) | 2 d t ,
i.e.,
t N T f E [ x ˜ T ( t ) Q ˜ x ˜ ( t ) + u ( t ) T R u ( t ) ] d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + ρ 2 E t N T f | v ( t ) | 2 d t ,
for all x ˜ ( t N ) L 2 ( Ω , F t N , P ; R 2 n ) , v ( t ) M F 2 ( [ t N , T f ] ; R n v ) . This ends the proof. □
Remark 3. 
Let
z ˜ ( t ) = Q ˜ 1 2 x ˜ ( t ) + R 1 2 u ( t ) ,
and then the inequality (16) can be rewritten as
E t N T f | z ˜ ( t ) | 2 d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + ρ 2 t N T f E [ v T ( t ) v ( t ) ] d t .
Particularly, if x ( t N ) = 0 , we have
E t N T f | z ˜ ( t ) | 2 d t ρ 2 t N T f E [ v T ( t ) v ( t ) ] d t .
This is just the performance of (14). Under such situation, we can define the operator L as follows:
L : M F 2 ( [ t N , T f ] ; R n v ) M F 2 ( [ t N , T f ] ; R n z ˜ )
with L ( v ) ( t ) = z ˜ ( t ) , and the H norm of L satisfies L ρ . So, the inequality (16) in Definition 1 is the generalization of the performance given by (16).
Now, we consider the error-feedback control case and suppose the corresponding control u ( t ) will be designed with the form of
u ( t ) = K e ( t ) ,
where K is a to-be-designed matrix taking values in R n u × n . Denote
A ˜ K = A + B K B K 0 Ξ r X r T ( X r X r T ) .
Then, the augmented system of (5) and (12) under control u ( t ) can be described as
d x ˜ ( t ) = [ A ˜ K x ˜ ( t ) + B ˜ u ( t ) + B ˜ 1 v ( t ) ] d t + [ A 1 ˜ x ˜ ( t ) + B ˜ 2 v ( t ) ] d W ( t ) x ˜ ( t N ) = x ( t N ) x r ( t N ) , t [ t N , T f ]
So, our target is to find proper matrix K and P to satisfy the following inequality for some given ρ > 0
E t N T f e r ( t ) T Q e r ( t ) d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + ρ 2 E t N T f | v t | 2 d t ,   x ˜ ( t N ) L 2 ( Ω , F t N , P ; R 2 n ) , v ( t ) M F 2 ( [ t N , T f ] ; R n v ) .
Theorem 3. 
Suppose there exists matrix K R n v 1 × n , and positive matrix P > 0 satisfies the following Riccati inequality
H 2 ( P ) : = A ˜ K T P + P A ˜ K + A ˜ 1 T P A ˜ 1 + Q ˜                                                             + ( A ˜ 1 T P B ˜ 2 + P B ˜ 1 ) ( ρ 2 I B ˜ 2 T P B ˜ 2 ) 1 ( B ˜ 1 T P + B ˜ 2 T P A ˜ 1 ) 0 ρ 2 I B ˜ 2 T P B ˜ 2 > 0
Then, the error-feedback control u ( t ) = K e r ( t ) satisfies (19).
Proof. 
Similar to the proof of Theorem 2, applying Itô’s formula to V ( x ˜ ( t ) ) = x ˜ T ( t ) P x ˜ ( t ) , we can obtain the following inequality:
E t N T f e r ( t ) T Q e r ( t ) d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + ρ 2 E t N T f | v t | 2 d t + E t N T f x ˜ T ( t ) H 1 ( P ) x ˜ ( t ) d t   E t N T f v ( t ) ( ρ 2 I B ˜ 2 T P B ˜ 2 ) 1 ( B ˜ 1 T P + B ˜ 2 T P A ˜ 1 ) x ( t ) M 2 d t ,
where M = ρ 2 I B ˜ 2 T P B ˜ 2 . So, for every v ( t ) M F 2 ( [ t N , T f ] ; R n v ) , there exists
E t N T f e r ( t ) T Q e r ( t ) d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + ρ 2 E t N T f | v t | 2 d t ,
for all v ( t ) M F 2 ( [ t N , T f ] ; R n v ) . This ends the proof. □

4. Hybrid Partial-Data-Driven H Robust Tracking Control Scheme for Linear Stochastic Systems

Based on the results of Theorems 1–3 in Section 3, the hybrid partial-data-driven H robust tracking control scheme for (5) and (6) is proposed in this section. The hybrid control scheme includes three stages, and the interval [ t 0 , T f ] is divided into three segments with the subintervals of [ t 0 , t N ] , [ t N , T ( 1 ) ] , and [ T ( 1 ) , T f ] , where t 0 < t N < T ( 1 ) < T f . The control input of system (5) is designed with piecewise form for each stage, which is organized in detail as follows:
Stage 1: Observing state of x r ( t ) in [ t 0 , t N ] at t 0 , t 1 , t 2 , , t N .
In this stage, the coefficient A r of reference system (6) is unknown, but the observation values of x r ( t ) can be obtained at t 0 , t 1 , t 2 , , t N . The observations of x r ( t 0 ) , x r ( t 1 ) , x r ( t 2 ) , , x r ( t N ) can rewritten as a matrix
X r = [ x r ( t 0 ) , x r ( t 1 ) , x r ( t 2 ) , , x r ( t N ) ] .
Denote
Ξ r = x r ( t 1 ) x r ( t 0 ) t 1 t 0 , x r ( t 2 ) x r ( t 1 ) t 2 t 1 , , x r ( t N ) x r ( t N 1 ) t N t N 1
By results of Theorem 1, the estimator of A r can be obtained:
A ^ r = Ξ r X r T ( X r X r T ) .
Because the main objective is to observe the state of system (6), there are no control inputs in this stage; i.e.,
u * ( t ) 0 ,
when t [ t 0 , t N ] .
Stage 2: Designing state-feedback H control in [ t N , T ( 1 ) ] .
Based on the estimator of reference system (12), the augmented control system of (5) and (12) can be obtained. For the given positive scalar ρ > 0 , by solving the Riccati inequality (17), the positive definite matrix P is obtained corresponding to the augmented system (13). By the results of Theorem 2, the state-feedback H control is designed:
u * ( t ) = K 2 x ˜ ( t ) ,
when t [ t N , T ( 1 ) ] , where K 2 = R 1 B ˜ T P and x ˜ ( t ) is the state of the augmented system (13), and such control satisfies the following performance:
E t N T ( 1 ) [ e r T ( t ) Q e r ( t ) + u T ( t ) R u ( t ) ] d t E [ x ˜ T ( t N ) P x ˜ ( t N ) ] + E t N T ( 1 ) | v ( t ) | 2 d t ,   x ˜ ( t N ) L 2 ( Ω , F t N , P ; R 2 n ) , v ( t ) M F 2 ( [ t N , T ( 1 ) ] ; R n v ) .
Stage 3: Designing error-feedback H control in [ T ( 1 ) , T f ] .
In order to further decrease the errors between x ( t ) and x r ( t ) , the error-feedback control is designed. By solving the Riccati inequality (20), the positive matrix P is obtained, and the corresponding error-feedback is designed:
u * ( t ) = K 3 ( x ( t ) x r ( t ) ) ,
when t [ T ( 1 ) , T f ] , where K 3 is a part of the solutions of Riccati inequality (20).
Finally, the three-stage H control u * ( t ) of (5) and (6) is obtained with piecewise form regarding t [ t 0 , T f ] , which can be written as follows:
u * ( t ) = 0 , t [ t 0 , t N ] , K 2 x ˜ ( t ) , t [ t N , T ( 1 ) ] , K 3 ( x ( t ) x r ( t ) ) , t [ T ( 1 ) , T f ] .
This piecewise control u * ( t ) satisfies the following performance:
E T ( 1 ) T f e r ( t ) T Q e r ( t ) d t E [ x ˜ T ( T ( 1 ) ) P x ˜ ( T ( 1 ) ] + ρ 2 E T ( 1 ) T f | v t | 2 d t ,   x ˜ ( T ( 1 ) ) L 2 ( Ω , F T ( 1 ) , P ; R 2 n ) , v ( t ) M F 2 ( [ T ( 1 ) , T f ] ; R n v ) .
For the profiles of the hybrid control u * ( t ) and the corresponding performances are shown in the following Example 1–3.
In summary, the programming evolution of such suggested hybrid 3-stage H control is organized in the following Table 1.
Table 1. The programming evolution of hybrid partial-data-driven H control.

5. Examples and Simulation

Example 1. 
Consider the following 1-dimensional controlled system:
d x ( t ) = ( 2 x ( t ) + u ( t ) + v 1 ( t ) ) d t + ( 0.16 x ( t ) + v 2 ( t ) ) d W ( t ) , x ( 0 ) = 3 , t [ 0 , T f ] ,
and the reference system is also a 1-dimensional system:
x ˙ r ( t ) = a r x r ( t ) + v 3 ,
where a r is an unknown number, but the values of x r ( t ) are observed at t = t 0 , t 1 , , t N with observation vector X r as follows:
X r = [ 1.0000 , 0.9306 , 0.8793 , 0.8420 , 0.7828 , 0.7024 , 0.6394 , 0.5906 , 0.4946 , 0.4727 ,   0.3722 , 0.2397 , 0.2241 , 0.2115 , 0.2014 , 0.2699 , 0.3046 , 0.3673 , 0.3416 , 0.3529 , 0.2880 ] ,
where t 0 = 0 , t k = k Δ t , k = 1 , 2 , , 20 , and Δ t = 0.05 . By Theorem 1, we get the estimator a ^ r = 1.5182 . So, the augmented system is obtained:
d x ( t ) = [ 2 x ( t ) + u ( t ) + v 1 ( t ) ] d t + [ 0.16 x ( t ) + v 2 ( t ) ] d W ( t ) , d x r ( t ) = [ 1 , 5182 x r ( t ) + v 3 ( t ) ] d t ,   t [ t N , T f ] .
The inverse matrix of X r X r T is ( X r X r T ) 1 = 0.1507 , and the corresponding augmented matrices are
A ˜ = 2 0 0 1.5182 , B ˜ = 1 0 , B ˜ 1 = 1 0 , B ˜ 2 = 1 0 , B ˜ 3 = 0 1 , Q = R = 1 .
For ρ = 0.9 , it is easy to check that the following matrix
P 2 = 0.5066 0.2990 0.2990 0.7361
satisfies the Riccati inequality (17). Let T ( 1 ) ( t N , T f ) . By Theorem 2, the corresponding state-feedback H control
u * ( t ) = 0.5066 x ( t ) + 0.2990 x r ( t ) , t [ t N , T ( 1 ) ]
is designed. Similarly, it is easy to check that the matrix P 3 = P 2 and K 3 = 0.4267 also satisfy the Riccati inequality (20). By Theorem 3, we can obtain the corresponding error-feedback H control
u * ( t ) = 0.4267 [ x ( t ) x r ( t ) ] , t [ T ( 1 ) , T f ] .
In summary, the control u ( t ) of system (21) is designed in three stages with t in different intervals: [ t 0 , t N ] , [ t N , T ( 1 ) ] , [ T ( 1 ) , T f ] . So, the control u * ( t ) has a piecewise form:
u * ( t ) = 0 , t [ t 0 , t N ] , 0.5066 x ( t ) + 0.2990 x r ( t ) , t [ t N , T ( 1 ) ] , 0.4267 [ x ( t ) x r ( t ) ] , t [ T ( 1 ) , T f ] .
This control is the suggested hybrid H control of (21) and (22). The trajectory of u * ( t ) with piecewise form is shown in Figure 1.
Figure 1. Trajectories of u * ( t ) in Example 1.
Figure 2 shows the profiles of exogenous disturbances and Brownian motion. The trajectories of x ( t ) and x r ( t ) are illustrated in Figure 3, which are the solutions of (21) and (22) under the control u ( t ) given by (24).
Figure 2. Profiles of Brownian motion and exogenous disturbance in systems (25) and (26) in Example 1.
Figure 3. Trajectories of x ( t ) and x r ( t ) under the control u * ( t ) in Example 1.
By Figure 3 and Figure 4, we see that, under the control of u * ( t ) , the distance between x ( t ) and x r ( t ) becomes less and the errors e r ( t ) = x ( t ) x r ( t ) also become smaller when t changes in [ t 0 , T f ] . Comparing the errors in different stages, Figure 4 illustrates that the values of errors e r ( t ) at the third stage ( t [ T ( 1 ) , T f ] ) are smaller than in the other two stages, where t [ t 0 , t N ] and [ t N , T ( 1 ) ] .
Figure 4. The changes in errors | e r ( t ) | under the effect of control u * ( t ) in Example 1.
Example 2. 
Consider the following 3-dimensional controlled system:
d x ( t ) = ( A x ( t ) + B x ( t ) + B 1 v 1 ( t ) ) d t + ( A 1 x ( t ) + B 2 v 2 ( t ) ) d W ( t ) x ( 0 ) = 3 1 1 , t [ 0 , T f ]
with the matrix coefficients as follows:
A = 1.6000 0.8000 0 0.8000 2.0000 1.6000 0.8000 0 1.6000 ,   B = 1 0 0 0 1 0 0 0 1 ,   B 1 = 0.1300 0.1000 0.1400 , A 1 = 0.1280 0.0640 0 0.0640 0.1600 0.1280 0.0640 0 0.1280 ,   B 2 = 0.0980 0.1000 0.1200 ,   B 3 = 0.1500 0.1700 0.1200 , Q = R = 1 0 0 0 1 0 0 0 1 ,
and the reference system is also a 3-dimensional system:
x ˙ ( t ) = A r x r ( t ) + B 3 v 3 ,
where A r is an unknown matrix. Suppose the interval [ t 0 , T f ] is divided into three segments, i.e., [ t 0 , t N ] , [ t N , T ( 1 ) ] , and [ T ( 1 ) , T f ] . Now, we apply the suggested hybrid 3-stage-control method to design the H control of (25) and (26), which includes 3 steps:
Stage 1: Observing state of x r ( t ) in [ t 0 , t N ] at t 0 , t 1 , t 2 , , t N .
In this stage, the states of reference system (26) can be observed at t 0 , t 1 , t 2 , , t N . Suppose the observations of reference system (26) at t 0 = 0 , t 1 = Δ t , t 2 = 2 Δ t , , t N = N Δ t with N = 20 are arranged as a 3 × ( N + 1 ) matrix denoted by X r = [ x t 0 , x t 1 , , x t N ] , whose values are observed as follows:
X r = 2.0099 1.9772 1.9242 1.8567 1.7744 1.6807 1.6077 3.3307 2.8971 2.5053 2.1834 1.8992 1.6630 1.4555 1.5543 1.5005 1.4424 1.4031 1.3478 1.2990 1.2610 1.5152 1.4263 1.3366 1.2687 1.1921 1.1177 1.0487 1.2870 1.1427 1.0174 0.9122 0.8072 0.7265 0.6451 1.2263 1.1559 1.1126 1.0681 1.0206 0.9567 0.9056 0.9799 0.8970 0.8369 0.7756 0.7252 0.6752 0.6380 0.5843 0.5281 0.4838 0.4713 0.4487 0.4153 0.3883 0.8605 0.8039 0.7567 0.7249 0.6946 0.6643 0.6421 ,
where Δ t = 0.05 is the sampling period. The inverse matrix of X r X r T is
( X r X r T ) 1 = 15.6800 2.3095 16.5304 2.3095 0.5003 2.2326 16.5304 2.2326 17.7210 .
Applying the results of Theorem 1 in this stage, the estimator of A r can be obtained as follows:
A ^ r = 2.6897 1.0497 0.8069 2.0047 2.8509 3.0743 0.9389 0.1091 2.1192 .
In this stage, there is no control input u ( t ) ; i.e., u 1 , * ( t ) 0 .
Stage 2: Designing state-feedback H control in [ t N , T ( 1 ) ] .
Taking ρ = 0.9 , by solving the Riccati inequality (17) corresponding to systems (25) and (26) with observation X r , we can obtain the positive definite matrix
P 2 = 0.7505 0.0796 0.1836 0.2306 0.0259 0.0978 0.0796 0.5933 0.2790 0.0234 0.1879 0.1338 0.1836 0.2790 0.9542 0.0218 0.0622 0.3467 0.2306 0.0234 0.0218 0.4721 0.0116 0.1373 0.0259 0.1879 0.0622 0.0116 0.3991 0.2849 0.0978 0.1338 0.3467 0.1373 0.2849 0.9738 .
Furthermore, by Theorem 2, we can get the state-feedback H control u 2 , * ( t ) = K 2 [ x T ( t ) , x r T ( t ) ] T with
K 2 = 0.7505 0.0796 0.1836 0.2306 0.0259 0.0978 0.0796 0.5933 0.2790 0.0234 0.1879 0.1338 0.1836 0.2790 0.9542 0.0218 0.0622 0.3467 .
Stage 3: Designing error-feedback H control in [ T ( 1 ) , T f ] .
By solving the Riccati inequality (20), the positive definite matrix P 3 is obtained as
P 3 = 1.4993 0.0444 0.0603 0.0000 0.0000 0.0000 0.0444 1.5181 0.0488 0.0000 0.0000 0.0000 0.0603 0.0488 1.4884 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 1.4926 0.0701 0.0495 0.0000 0.0000 0.0000 0.0701 1.4750 0.0560 0.0000 0.0000 0.0000 0.0495 0.0560 1.5149 ,
and, by results of Theorem 3, the corresponding error-feedback H control u 3 , * ( t ) = K 3 ( x ( t ) x r ( t ) ) can be obtained, where K 3 is solved as
K 3 = 0.0368 0.2511 0.0297 0.2503 0.0603 0.2802 0.0346 0.1502 0.0227 .
Combining the results of Stage 1, Stage 2, and Stage 3, the input control of system (25) can be rewritten as segmentation form:
u * ( t ) = 0 , t [ t 0 , t N ] , K 2 [ x T ( t ) , x r T ( t ) ] T , t [ t N , T ( 1 ) ] , K 3 ( x ( t ) x r ( t ) ) , t [ T ( 1 ) , T f ] .
Figure 5 shows the profiles of exogenous disturbances v 1 ( t ) , v 2 ( t ) , v 3 ( t ) , and Brownian motion W ( t ) to which systems (25) and (26) are subjected. The trajectories of each component of the suggested H control u * ( t ) of (25) are illustrated in Figure 6, which is divided into 3 stages. The contrast between x ( t ) and x r ( t ) shown in Figure 7 and Figure 8 illustrate the change in errors between x ( t ) and x r ( t ) with | e r ( t ) | = | x ( t ) x r ( t ) | | . In the firs stage, i.e., t [ t 0 , t N ] , system (25) is no control input, where u 1 ( t ) = u 2 ( t ) = u 3 ( t ) = 0 . Figure 7 and Figure 8 illustrate that the distance between x ( t ) and x r ( t ) is the farthest; i.e., the values of errors | e r ( t ) | are worst in the first stage. In the second stage, i.e., t [ t N , T ( 1 ) ] , the control input of system (25) is state-feedback H control. By Figure 7 and Figure 8, we see that, compared with the first stage, the values of errors | e r ( t ) | with t [ t N , T ( 1 ) ] become smaller but still not very good. In the third stage, it is easy to see that the errors between x ( t ) and x r ( t ) become the smallest when t [ T ( 1 ) , T f ] , which is shown in Figure 7 and Figure 8.
Figure 5. Profiles of Brownian motion W ( t ) and exogenous disturbance v 1 ( t ) , v 2 ( t ) , and v 3 ( t ) in Example 2.
Figure 6. Trajectories of u * ( t ) in Example 2.
Figure 7. Trajectories of x ( t ) and x r ( t ) under the control u * ( t ) in Example 2.
Figure 8. The changes in errors | e r ( t ) | under the effect of control u * ( t ) in Example 2.
Example 3. 
Consider the system of robot manipulator discussed in [35], whose dynamic equation is given by
M ( q ) q ¨ + V ( q , q ˙ ) + G ( q ) = τ + τ d
where q ( t ) R n is the vector of generalized coordinates, M ( q ) R n × n is the inertia matrix, V ( q , q ˙ ) R n is the Coriolis and centrifugal torque vector, G ( q ) R n is the gravitational torque vector, τ R n is the generalized control input vector, and τ d is the disturbance.
See Figure 9 in order to complete the task of the robot arm moving along with a given trajectory from A to B in the time interval [ t 0 , T f ] . The control τ in (28) should be designed rationally such that q ( t ) can take proper values at every moment t [ t 0 , T f ] . So, the values of q ( t ) should track a given reference q d ( t ) . The error between them is defined as
e ( t ) = q ( t ) q d ( t ) .
Figure 9. Sketch of robot manipulator moving along the given reference trajectory.
Now, suppose M ( q ) is an invertible positive definite constant matrix and V ( q , q ˙ ) and G ( q ) are linear in q and q ˙ ; i.e., there exist matrices V 1 , V 2 , G R n × n such that V ( q , q ˙ ) = V 1 q + V 2 q ˙ and G ( q ) = G q .
Let
x = q q ˙ , x r = q d q ˙ d ,
Then, system (28) is equivalent to
d x ( t ) = [ A x ( t ) + B τ ( t ) + B 1 τ d ( t ) ] d t .
Now, we extend it to the stochastic case. Denote the control u ( t ) = τ ( t ) and the exogenous disturbance v 1 ( t ) = τ d ( t ) . Suppose the stochastic system with control input has the following form:
d x ( t ) = [ A x ( t ) + B u ( t ) + B 1 v 1 ( t ) ] d t + [ A 1 x ( t ) + B 2 v 2 ( t ) ] d W ( t ) .
where
A = 0 I M 1 ( V 1 + G ) M 1 V 2 , B = 0 M 1 , A 1 = 0 0 M 1 ( V 1 + G ) λ M 1 V 2 λ , B 1 = 0 M 1 .
In practice, system (29) is seen as a version of (28) subjected to Brownian motion W ( t ) and exogenous disturbances v 1 ( t ) and v 2 ( t ) .
We also suppose the reference trajectory x r ( t ) given by
d x r ( t ) = [ A r x r ( t ) + B 3 v 3 ( t ) ] d t ,
where A r R 2 n × 2 n is unknown, but the values of x r ( t ) can be observed at t 0 , t 1 , , t N .
Now, let n = 3 ; i.e., q is a 3-dimensional vector with
q = q 1 q 2 q 3 , q d = q d 1 q d 2 q d 3 .
Suppose the corresponding matrices of M , V 1 , V 2 , and G are provided as follows:
M = 10.2470 1.2741 4.7213 1.2741 14.3984 0.3849 4.7213 0.3849 10.5878 ,   V 1 = 23.4859 0.5090 6.4912 6.3166 29.0514 0.3357 9.8896 6.6483 20.1972 , V 2 = 36.0462 0.1154 11.2630 15.8835 52.9585 17.5902 13.5789 8.6582 34.6113 , G = 20.6781 0.1502 5.5988 11.2692 29.6475 2.1779 9.9908 6.5910 19.7089 ,
and λ = 0.05 , B 2 = B 3 = B . Suppose the observations of x r at t = t 0 , t 1 , , t N are obtained:
X r = 0.2000 0.2250 0.2417 0.2514 0.2556 0.2552 0.2512 0.2444 0.2355 0.2250 0.2200 0.2450 0.2656 0.2825 0.2961 0.3068 0.3151 0.3212 0.3253 0.3276 0.7000 0.7050 0.6999 0.6865 0.6666 0.6415 0.6126 0.5808 0.5471 0.5122 0.5000 0.3330 0.1953 0.0829 0.0078 0.0797 0.1358 0.1784 0.2097 0.2315 0.5000 0.4126 0.3373 0.2719 0.2150 0.1651 0.1211 0.0822 0.0476 0.0167 0.1000 0.1029 0.2674 0.3985 0.5010 0.5788 0.6356 0.6745 0.6984 0.7095   0.2134 0.2011 0.1885 0.1757 0.1630 0.1506 0.1386 0.1270 0.1160 0.1056 0.3285 0.3279 0.3262 0.3233 0.3194 0.3147 0.3091 0.3029 0.2960 0.2886 0.4767 0.4412 0.4061 0.3717 0.3384 0.3063 0.2756 0.2465 0.2190 0.1932 0.2456 0.2531 0.2554 0.2535 0.2484 0.2407 0.2311 0.2201 0.2082 0.1958 0.0108 0.0355 0.0576 0.0775 0.0951 0.1109 0.1249 0.1373 0.1481 0.1575 0.7102 0.7023 0.6872 0.6666 0.6415 0.6131 0.5823 0.5497 0.5161 0.4821 0.0958 0.0867 0.0781 0.0702 0.0630 0.0563 0.0501 0.0445 0.0394 0.0348 0.0306 0.2807 0.2724 0.2638 0.2549 0.2458 0.2365 0.2271 0.2176 0.2081 0.1986 0.1892 0.1691 0.1467 0.1260 0.1069 0.0895 0.0736 0.0592 0.0462 0.0345 0.0242 0.0150 0.1831 0.1704 0.1579 0.1457 0.1339 0.1226 0.1119 0.1019 0.0925 0.0837 0.0756 0.1655 0.1723 0.1779 0.1823 0.1857 0.1881 0.1896 0.1903 0.1901 0.1892 0.1876 0.4480 0.4143 0.3812 0.3491 0.3180 0.2882 0.2599 0.2330 0.2076 0.1838 0.1615 .
where t 0 = 0 , t k = k Δ t , k = 1 , 2 , , 30 with Δ t = 0.05 . By Theorem 1, we can obtain the estimator of A r :
A ^ r = 1.8585 0.2796 0.9931 0.9700 0.4932 0.0032 1.6188 0.1667 0.8638 0.1287 0.6264 0.0664 0.8675 0.1626 0.4663 0.0249 0.2543 1.0276 2.8913 0.1483 1.6393 3.8331 0.6145 0.3722 0.4257 4.2336 0.7023 1.4485 4.0145 1.1032 0.5798 1.0479 3.9362 0.6826 0.5743 3.2811 .
Construct the augmented system of (29) and (30) with A ^ r given above. Then, for ρ = 0.8 , applying the results of Theorems 2 and 3, the state-feedback H control and error-feedback H control are designed, which are given as follows:
τ * ( t ) = u * ( t ) = 0 , t [ t 0 , t N ] , K 2 [ x T ( t ) , x r T ( t ) ] T , t [ t N , T ( 1 ) ] , K 3 ( x ( t ) x r ( t ) ) , t [ T ( 1 ) , T f ] .
where
K 2 = 0.0266 0.0039 0.0122 0.0397 0.0029 0.0141 0.0056 0.0000 0.0096 0.0141 0.0010 0.0081 0.0021 0.0192 0.0007 0.0038 0.0262 0.0099 0.0076 0.0101 0.0050 0.0015 0.0103 0.0021 0.0092 0.0121 0.0277 0.0097 0.0098 0.0456 0.0097 0.0056 0.0103 0.0053 0.0031 0.0188
and
K 3 = 1.0647 0.0374 0.3934 1.3155 0.2722 0.6980 0.1236 1.2073 0.3950 0.3407 1.6582 0.2415 0.5678 0.3900 0.9147 0.6441 0.1463 1.4583
Figure 10 illustrates the trajectories of control τ * ( t ) with three stages. Figure 11 shows the trajectories of q ( t ) and reference q d ( t ) under the control τ * ( t ) designed above. It is easy to see that each component of q ( t ) performs well regarding the reference q d ( t ) . Figure 12 illustrates the error’s changes between q ( t ) and q d ( t ) , showing that there exists the smallest error in the third stage, which verifies the effect of the proposed method.
Figure 10. Trajectories of u * ( t ) in Example 3.
Figure 11. Trajectories of x ( t ) and x r ( t ) under the control u * ( t ) in Example 3.
Figure 12. The changes in errors | e r ( t ) | under the effect of control u * ( t ) in Example 3.

6. Conclusions

The robust H tracking problem is investigated for linear stochastic systems where the parameters of a reference system are unknown but some discrete-time observations are available. A hybrid data-driven H tracking-control design scheme is proposed for such problems. By using the least-squares method, the parameters of the reference system are estimated and the corresponding data-dependent augmented systems are established. Based on the solutions of algebraic Riccati inequalities, the state-feedback and error-feedback H controls are designed for such tracking problems to enhance system performance and reduce tracking error. Moreover, the programming evolution of the hybrid control method is summarized, which includes data observation, state-feedback H control, and error-feedback H control.

Author Contributions

Conceptualization, X.L. and R.Z.; methodology, X.L. and Y.Z.; software, X.L. and Y.Z.; validation, X.L., Y.Z., and R.Z.; formal analysis, X.L.; investigation, R.Z.; resources, X.L.; writing—original draft preparation, Y.Z. and X.L.; writing—review and editing, X.L. and R.Z.; visualization, R.Z.; supervision, X.L.; project administration, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSF of China, grant number 62273212.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Persis, C.D.; Tesi, P. Formulas for data-driven control: Stabilization, optimality, and robustness. IEEE Trans. Autom. Control 2020, 65, 909–924. [Google Scholar] [CrossRef]
  2. Shen, D.; Zhang, W.; Wang, Y.; Chien, C. On almost sure and mean square convergence of P-type ILC under randomly varying iteration lengths. Automatica 2016, 63, 359–365. [Google Scholar] [CrossRef]
  3. Yu, H.; Guan, Z.; Chen, T.; Yamamoto, T. Design of data-driven PID controllers with adaptive updating rules. Automatica 2020, 121, 109185. [Google Scholar] [CrossRef]
  4. Hou, Z.; Xiong, S. On model-free adaptive control and its stability analysis. IEEE Trans. Autom. Control 2019, 64, 4555–4569. [Google Scholar] [CrossRef]
  5. Pang, Z.; Ma, B.; Liu, G.; Han, Q. Data-driven adaptive control: An incremental triangular dynamic linearization approach. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 4949–4953. [Google Scholar] [CrossRef]
  6. Wang, Z.; Liu, K.; Cheng, X.; Sun, X. Online data-driven model predictive control for switched linear systems. IEEE Trans. Autom. Control 2025, 70, 6222–6229. [Google Scholar] [CrossRef]
  7. Wu, W.; Qiu, L.; Rodriguez, J.; Liu, X.; Ma, J.; Fang, Y. Data-driven finite control-set model predictive control for modular multilevel converter. IEEE J. Emerg. Sel. Top. Power Electron. 2023, 11, 523–531. [Google Scholar] [CrossRef]
  8. Sun, T.; Sun, X.; Sun, A. Optimal output tracking of aircraft engine systems: A data-driven adaptive performance seeking control. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 1467–1471. [Google Scholar] [CrossRef]
  9. Liu, W.; Sun, J.; Wang, G.; Bullo, F.; Chen, J. Data-driven self-triggered control via trajectory prediction. IEEE Trans. Autom. Control 2023, 68, 6951–6958. [Google Scholar] [CrossRef]
  10. Zhan, J.; Ma, Z.; Zhang, L. Data-driven modeling and distributed predictive control of mixed vehicle platoons. IEEE Trans. Intell. Veh. 2023, 8, 572–582. [Google Scholar] [CrossRef]
  11. Zhang, W.; Xie, L.; Chen, B.S. Stochastic H2/H Control: A Nash Game Approach; Taylor & Francis: Boca Raton, FL, USA, 2017. [Google Scholar]
  12. Chen, B.S.; Wang, C.P.; Lee, M.Y. Stochastic robust team tracking control of multi-UAV networked system under Wiener and Poisson random fluctuations. IEEE Trans. Cybern. 2021, 51, 5786–5799. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, H.; Ma, T.; Lewis, F.L.; Wan, Y. Robust formation control for multiple quadrotors with nonlinearities and disturbances. IEEE Trans. Cybern. 2020, 50, 1362–1371. [Google Scholar] [CrossRef]
  14. Zhao, J.; Zhao, H.; Song, Y.; Sun, Z.Y.; Yu, D. Fast finite-time consensus protocol for high-order nonlinear multi-agent systems based on event-triggered communication scheme. Appl. Math. Comput. 2026, 508, 129631. [Google Scholar] [CrossRef]
  15. Boshkovska, E.; Ng, D.W.K.; Zlatanov, N.; Koelpin, A.; Schober, R. Robust resource allocation for MIMO wireless powered communication networks based on a non-linear EH model. IEEE Trans. Commun. 2017, 65, 1984–1999. [Google Scholar] [CrossRef]
  16. Zames, G. Feedback and optimal sensitivity: Model reference transformations, multiplicative seminorms, and approximate inverses. IEEE Trans. Autom. Control 1981, 26, 301–320. [Google Scholar] [CrossRef]
  17. Doyle, J.C.; Glover, K.; Khargonekar, P.; Francis, B. State-space solutions to standard H2 and H control problems. IEEE Trans. Autom. Control 1989, 34, 831–847. [Google Scholar] [CrossRef]
  18. Gahinet, P.; Apkarian, P. A linear matrix inequality approach to H control. Int. J. Robust Nonlinear Control 1994, 4, 21–448. [Google Scholar] [CrossRef]
  19. Dai, X.; Zuo, H.; Deng, F. Mean square finite-time stability and stabilization of impulsive stochastic distributed parameter systems. IEEE Trans. Syst. Man Cybern. Syst. 2025, 55, 4064–4075. [Google Scholar] [CrossRef]
  20. Zhao, J.; Yuan, Y.; Sun, Z.Y.; Xie, X. Applications to the dynamics of the suspension system of fast finite time stability in probability of p-norm stochastic nonlinear systems. Appl. Math. Comput. 2023, 457, 128221. [Google Scholar] [CrossRef]
  21. Hinrichsen, D.; Pritchard, A.J. Stochastic H. SIAM J. Control Optim. 1998, 36, 1504–1538. [Google Scholar] [CrossRef]
  22. der Schaft, A.J.V. On a state-space approach to nonlinear H control. Syst. Control Lett. 1991, 16, 1–8. [Google Scholar]
  23. Ball, J.A.; Helton, J.W.; Walker, M.L. H control for nonlinear systems with output feedback. IEEE Trans. Autom. Control 1993, 38, 546–559. [Google Scholar] [CrossRef]
  24. Zhang, W.; Chen, B.S. State feedback H control for a class of nonlinear stochastic systems. SIAM J. Control Optim. 2006, 44, 1973–1991. [Google Scholar] [CrossRef]
  25. Lin, X.; Zhang, T.; Zhang, W.; Chen, B.S. New approach to general nonlinear discrete-Time stochastic H control. IEEE Trans. Autom. Control 2019, 64, 1472–1486. [Google Scholar] [CrossRef]
  26. Chen, B.S.; Yang, C.T.; Lee, M.Y. Multiplayer noncooperative and cooperative minimax H tracking game strategies for linear mean-field stochastic systems with applications to cyber-social systems. IEEE Trans. Cybern. 2022, 52, 2968–2980. [Google Scholar]
  27. Xin, Y.; Zhang, W.; Wang, D.; Zhang, T.; Jiang, X. Stochastic H2/H control for discrete-time multi-agent systems with state and disturbance-dependent noises. Int. J. Robust Nonlinear Control 2025, preprinted. [Google Scholar] [CrossRef]
  28. Zhao, J.; Wang, Z.; Lv, Y.; Na, J.; Liu, C.; Zhao, Z. Data-driven learning for H control of adaptive cruise control systems. IEEE Trans. Veh. Technol. 2024, 73, 18348–18362. [Google Scholar] [CrossRef]
  29. Liao, Y.; Zhang, T.; Yan, X.; Jiang, D. Integrated guidance and tracking control for underactuated AUVs on SE(3) with singularity-free global prescribed performance attitude control. Ocean Eng. 2025, 342, 122770. [Google Scholar] [CrossRef]
  30. Zhang, L.; Sun, Y.; Chai, P.; Tan, J.; Zheng, H. Prescribed-performance time-delay compensation control for UUV trajectory tracking in main-branch water conveyance tunnel transitions under unknown input delays. Ocean Eng. 2025, 342, 122941. [Google Scholar]
  31. Xu, B.; Dai, Y.; Suleman, A.; Shi, Y. Distributed fault-tolerant control of multi-UAV formation for dynamic leader tracking: A Lyapunov-based MPC framework. Automatica 2025, 175, 112179. [Google Scholar] [CrossRef]
  32. Luque, I.; Chanfreut, P.; Limón, D.; Maestre, J.M. Model predictive control for tracking with implicit invariant sets. Automatica 2025, 179, 112436. [Google Scholar] [CrossRef]
  33. Chen, B.S.; Hsueh, C.H.; Wu, R.S. Optimal H adaptive fuzzy observer-based reference tracking control of nonlinear stochastic systems under uncertain measurement function and noise. Fuzzy Sets Syst. 2025, 521, 109597. [Google Scholar] [CrossRef]
  34. Lin, X.; Zhang, R. H Control for stochastic systems with Poisson jumps. J. Syst. Sci. Complex 2011, 24, 683–700. [Google Scholar] [CrossRef]
  35. Song, G.; Park, H.; Kim, J. The H Robust stability and performance conditions for uncertain robot manipulators. IEEE/CAA Hournal Autom. Sin. 2025, 12, 270–272. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.