Abstract
This paper is to study finite horizon linear quadratic (LQ) non-zero sum Nash game for discrete-time infinite Markov jump stochastic systems (IMJSSs). Based on the theory of stochastic analysis, a countably infinite set of coupled generalized algebraic Riccati equations are solved and a necessary and sufficient condition for the existence of Nash equilibrium points is obtained. From a new perspective, the finite horizon mixed robust control is investigated, and summarize the relationship between Nash game and control problem. Moreover, the feasibility and validity of the proposed method has been proved by applying it to a numerical example.
Keywords:
infinite Markov jump stochastic systems (IMJSSs); Nash equilibrium points; coupled generalized algebraic Riccati equations; Nash game; H2/H∞ control MSC:
93E03
1. Introduction
Markov jump stochastic systems (MJSSs), as a typical class of stochastic hybrid dynamical systems, are widely used in practical engineering control systems. At the same time, a great deal of related research on the stability, ergodicity, robust control and filtering, and so on [,,,,,,,,] has been undertaken. However, for the Markov chain, when the state space takes values in infinite countably sets, from the application point of view, it may be more extensive. Consequently, plentiful theory literature and research achievements have emerged for infinite Markov jump stochastic systems (IMJSSs). In recent years, various efforts have been made to cope with IMJSSs in a wide variety of systems. To be specific, via the operator spectrum method, exponential stabilities for discrete-time [] and continuous-time [] IMJSSs have been investigated respectively. Aiming at the practical time-delay factors, with the aid of Lyapunov stability theory, [] discussed the stability analysis for IMJSSs with time-delay; further, stabilities for uncertain discrete-time IMJSSs with time-delay have been developed in [], and on this basis, [] addressed the finite horizon control problems. fuzzy filtering has been solved for nonlinear IMJSSs by the T-S fuzzy model approach in [].
Dynamic game theory, which has come into wider use in many fields such as engineering, economics, and management science, has attracted great attention, and a large number of results have been obtained in the literature [,,,,,]. Furthermore, the Nash game problem has been studied for MJSSs, and further, a unified treatment approach for , and control design problems was presented in []. In [,], the authors revealed the relationship between the Nash equilibrium points and control for continuous-time MJSSs. While many theoretical results have been established to study the stochastic system governed by a finite-state Markov chains, more research efforts are required to discuss IMJSSs. However, the causal and anticausal Lyapunov operators of IMJSSs are no more adjoint, which is the essential difference between finite and infinite MJSSs [,]. In most practical applications, IMJSSs have broader application prospects than MJSSs with finite jumps. As a consequence, the related research about IMJSSs is of critical importance. In the game field, [] discussed an infinite horizon linear quadratic (LQ) Nash game for continuous-time IMJSSs. Unfortunately, to the best of our knowledge, there is almost no research about the Nash game problem for discrete-time IMJSSs.
In this article, we study a finite horizon LQ non-zero sum Nash game for discrete-time IMJSSs, which considers a more general system. From a new perspective, the finite horizon mixed control is further investigated. On the one hand, this is an extension of the previous study from MJSSs [], to IMJSSs; on the other, it is a discrete-time case of []. Concretely, the main work and contribution of this thesis is as follows. First, the existence of Nash equilibrium points, which can be boiled down to the solvability of a countably infinite set of coupled generalized algebraic Riccati equations, is shown. For infinite Markov jump systems, the causal and anticausal Lyapunov operators are no more adjoint, which leads to the inequivalence between stochastic stability, asymptotic mean square stability, and exponential mean square stability. Further, this indicates the essential difference between the finite and infinite Markov jump systems. To this end, we introduce the infinite dimension Banach spaces, where the elements are countably infinite sequences of linear and bounded operators. Thus, the source of the problem resides in the way to solve this kind of equations, which are harder to solve than those in []. Second, the finite horizon mixed control is solved from the new view of a Nash game, and further, the relationship between the Nash game and control problem is summarized with some remarks. Finally, the typical example cited in the paper demonstrates the validity of the proposed method well.
The rest of this article is arranged as below. Some useful preliminary results are introduced in Section 2. The existence of Nash equilibrium points is solved by a necessary and sufficient condition in Section 3. In Section 4, some special cases with some remarks are given. In Section 5, a numerical example is illustrated, and a summary is provided in Section 6.
For convenience, the following notations are adopted. : n-dimensional real Euclidean space; : the linear space of all m by n real matrices; the Euclidean norm of or the operator norm of ; : the identity matrix; : the transpose of a matrix (or vector) N; : the pseudo-inverse of a matrix N; : N is positive (semi-positive) definite; ; .
2. Preliminaries
On a complete probability space (,,), we consider the following discrete-time IMJSS:
where represents the system state, and and are the control processes of two different players, respectively. stands for the controlled output. , is a standard r-dimensional Brownian motion. denotes a infinite Markov jump process taking values in , and the transition probability matrix is with . In this paper, we assume is nondegenerate, for all , and the stochastic processes and are mutually independent. Let . is -measurable, and }.
is defined as the Banach space of the set with the norm In the sequel, we assume that all coefficients of considered systems have a finite norm . We make use of the notation of for . Further, denotes the subspace of formed by all and . For , is equivalent to , which means . We set , , , , , , and assume for .
The two cost functions for the Nash game problem are given by
where is a given prescribed disturbance attenuation level. For simplicity, let the Nash game problem for cost Functions (2) and (3) with Equation (1) by the Nash game problem . Hence, the Nash game problem is addressed, which is to find admissible control to minimize cost Functions (2) and (3) subject to Equation (1).
Next, we list some definitions and lemmas that are needed for the follow-up procedures.
Definition 1.
A strategy pair is a Nash equilibrium point if
for all .
Lemma 1
([]). For a symmetric matrix S, we have
- (i)
- ;
- (ii)
- ;
- (iii)
- if and only if .
Lemma 2
([]). Let matrices , H, and with appropriate dimensions consider the following quadratic form:
where y and u are random variables defined on a probability space (Ω,,). Then, the following conditions are equivalent:
- (i)
- for any random variable y.
- (ii)
- There exists a symmetric matrix such that for any random variable y.
- (iii)
- and .
- (iv)
- and .
- (v)
- There exists a symmetric matrix such that .Moreover, if any of the above conditions hold, then (ii) is satisfied by . In addition, for any T satisfying (v). Finally, for any random variable y, the random variable is optimal and the optimal value is .
The following LQ result for discrete-time IMJSSs can be directly yielded from Theorem 3.1 [], which is the continuous-time case.
Lemma 3.
For the following standard LQ optimal control problem with discrete-time IMJSSs,
subject to
then, we find that solves the following coupled generalized algebraic Riccati equations:
where
and
3. Nash Equilibrium Points
This section focuses on solving the Nash gameproblem , and we assume that the linear, memoryless feedback strategy has the following form []:
under this assumption, we obtain the feedback Nash equilibrium points by a countably infinite set of coupled generalized algebraic Riccati equations.
Theorem 1.
The Nash game problem has unique Nash equilibrium points
iff the following coupled generalized algebraic Riccati equations,
admit a group of solutions with , for , where
Proof.
Sufficiency: Since Equations (10)–(13) have a group of solutions
by constructing , then, we substitute into Equation (1) to get the equation as follows:
using similar processing, Lemma 1 in [] can be generalized to the infinite Markov jump systems. In fact, via the assumption that and are mutually independent, and is also irrelevant to , we have
it follows from taking summation from 0 to T on both sides of (15) that
Further, we can obtain the following result:
where
for A combination of the method of completing square and (10) turns Equation (16) into
where ; therefore, the Nash equilibrium inequality (4) occurs naturally, that is to say
additionally, by plugging into Equation (1), Equation (1) can be converted into
in the meantime, we also have
Obviously, to illustrate that minimizes (3), we only need to prove a kind of LQ optimal control problem of IMJSSs that causes (19) minimization under the condition of (18). Via Lemma 3, we can easily prove the Nash equilibrium inequality (5).
Necessity: Assume that is a linear feedback Nash equilibrium point for the Nash game (4) and (5). Putting into Equation (1), we gain
now discover that the above problem is transformed into the following indefinite LQ optimal control with optimal solution :
it is obvious that the above indefinite LQ problem (21) is well-posed. Next, we prove that the coupled generalized algebraic Riccati Equation (22) is solvable by mathematical induction:
to this end, the value function is introduced as follows:
for , we can obtain from that the existence of motivates
it is noteworthy that Equation (22) is available for with , and the optimal value function is
in the process,
then, for , assume that the coupled generalized algebraic Riccati Equation (22) has a solution . Meanwhile, the optimal value function is
and the optimal control is
now, our goal is to illustrate that for , the existence of solution to Equation (22). By the aid of (20), the dynamic programming optimality principle applies to
where
similar results to Lemma 2 apply to the infinite Markov jump case, it is demonstrated that satisfies the following equation:
the following conclusions can be handled in the same manner as (17), that is
and
up to the present, one can infer that there exists satisfying Equation (22). Furthermore, an optimal solution of the indefinite LQ problem (21) is with . Amalgamating the above results into Equation (22), it is stated that . It only remains to indicate . In practice,
in addition, if we plug into Equation (1), then (18) is gained. We can deduce that (5) is a standard LQ control subject to (18). Using Lemma 3, can be easily accessible. Moreover, with The proof has been completed. □
Remark 1.
Note that Theorem 1 can be considered as an extension of [] to infinite jumps and multiplicative noise and a discrete-time version of [].
Remark 2.
If the infinite horizon cost function is concerned, it is much more challenging owing to the requirement of stabilization limitation for the closed-loop system. As discussed in [], the infinite horizon LQ Nash game has been considered.
4. Application to Special Case
In the previous section, the Nash game problem for discrete-time IMJSSs is solved. It should be noted that when takes different values or is regarded as exogenous disturbance. As a special case, we discuss the finite horizon robust control problem from a new perspective and further explore the relationship between Nash equilibrium points and finite horizon control with some remarks.
4.1. Finite Horizon Control
in system (1) is seen as one of the players in the Nash game problem; as a matter of fact, it is more a case of considering as an exogenous disturbance in most control systems. Let be a prescribed disturbance attenuation. In consequence, the original Nash game problem is turned into finding a controller such that
- (i)
- for the closed system of Equation (1) with ;
- (ii)
- if exists, minimizes the output energy:where is the worst case disturbance andin other words, the above problem is called the finite horizon control problem. We need to pay attention to the definition of perturbed operator given in [,].
As an applications of Nash game problem, it can be obtained from Theorem 1 that
and
where and are defined in Theorem 1.
Remark 3.
In accordance with the definition of control problem, it is crucial to note that does not come down to . If we can confirm that can be replaced by , then the following result would be given naturally.
Theorem 2.
For system (1), assume the following coupled generalized algebraic Riccati equations:
admit solutions with , for , then the finite horizon control optimal controller is , . Conversely, if the finite horizon control problem has the solution , , and , then the coupled generalized algebraic Riccati Equations (35)–(38) admit a group of solutions with , for .
Proof.
Sufficiency: It is obvious from the sufficiency part in Theorem 1 that we need only to explain . In fact, in the light of the definition of perturbed operator , when , note that the condition of (35), Equation (17) means iff . Putting into Equation (14), the obtained closed-loop system with initial state leads to the state response . One step further, naturally occurs. Then, the inescapable conclusion is that iff , which stands for .
Necessity: Assume that the finite horizon control problem has the solution , , a combination of (31) and (32) causes for all . By now, it can be deduced from the necessity of Theorem 1 that Equations (10)–(13) have solutions. Keep in mind that , hence, Equations (35)–(38) admit a group of solutions with , for . The proof has been completed. □
Remark 4.
Although the finite horizon control problem has been solved in [], we yield a similar result from the new perspective of a Nash game.
Remark 5.
By comparing Theorem 1 with Theorem 2, surely the existence of Nash equilibrium points and the solvability of finite horizon control are not equivalent for system (1), which differs from the continuous-time case as described in []. The main cause of the inequivalence lies in that the condition of Equation (35) cannot meet the Nash equilibrium problem, namely, is not equivalent to .
4.2. Some Remarks on Nash Equilibrium Points
As a point of fact, only when is the equivalence between Nash equilibrium points and finite horizon control valid. Accordingly, the relationship between the Nash game and control problem can be discussed in the following theorem.
Theorem 3.
- (i)
- There exists a linear memoryless Nash equilibrium points with
- (ii)
- The finite horizon control is solvable with
- (iii)
Proof.
Theorem 3 can be demonstrated through Theorem 1 and Theorem 2. □
Remark 6.
Keep in mind that when , , at this point, Equations (10)–(13) are consistent with Equations (35)–(38). In other words, for system (1), the existence of Nash equilibrium problem, the solvable of finite horizon control and the solvability of Equations (35)–(38) are equivalent. In addition, under the limitation of , a unified treatment for , and control can be investigated such as [].
5. Numerical Example
In this section, to solve the coupled generalized algebraic Riccati Equations (35)–(38), we provide an iterative algorithm, which can be summarized as follows:
- (i)
- When , the terminal condition and can obtain and ;
- (ii)
- (iii)
- (iv)
- Repeating the above procedures, for we can compute that , , and .
Remark 7.
Next, a numerical example will be presented to show the validity of the proposed method.
Example 1.
Consider a three-stage one-dimensional discrete-time IMJSS with coefficients in Table 1.
Table 1.
Coefficients of considered system.
6. Conclusions
This paper mainly explores a finite horizon LQ non-zero sum Nash game for discrete-time IMJSSs, whose system is governed by a countable Markov chain. The Nash equilibrium points for the considered system are solved by a countably infinite set of coupled generalized algebraic Riccati equations. Then, some special cases are given via a new perspective of the Nash game, which is the finite horizon mixed control with some remarks, and we summarized the relationship between Nash game and control problem. The contents in this paper are an extension and improvement of the previous works [,] in the MJSSs case. In fact, to solve the difficulties caused by the countable Markov chain, we introduce the infinite dimension Banach spaces, where the elements are countably infinite sequences of linear and bounded operators. Besides, to overcome the difficulty of solving a countably infinite set of coupled generalized algebraic Riccati equations, we present an iterative algorithm. In the future, the infinite horizon Nash game for IMJSSs can be considered.
Author Contributions
Conceptualization, Y.L.; methodology, Y.L.; formal analysis, Y.L.; investigation, Z.W. and X.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L.; supervision, X.L. and Z.W.; project administration, Y.L.; funding acquisition, Y.L., Z.W. and X.L. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Natural Science Foundation of Qingdao under Grant 23-2-1-7-zyyd-jch, Social Science Planning and Research Special Project of Shandong Province under Grant 22CSDJ43, Natural Science Foundation of China under grant 62273212, the Natural Science Foundation of Shandong Province under grant ZR2020MF062 and People Benefit Project of Qingdao under Grant 22-3-7-smjk-16-nsh.
Data Availability Statement
Not applicable.
Acknowledgments
We would like to thank the anonymous reviewers for their constructive suggestions to improve the quality of this paper.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Shafieepoorfard, E.; Raginsky, M.; Meyn, S.P. Rationally inattentive control of Markov processes. SIAM J. Control Optim. 2016, 54, 987–1016. [Google Scholar] [CrossRef]
- Veretennikov, A.Y.; Veretennikova, M.A. On improved bounds and conditions for the convergence of Markov chains. Izv. Math. 2022, 86, 92–125. [Google Scholar] [CrossRef]
- Khasminskii, R.Z. Stability of regime-switching stochastic differential equations. Probl. Inform. Transm. 2012, 48, 259–270. [Google Scholar] [CrossRef]
- Li, F.; Xu, S.; Shen, H.; Zhang, Z. Extended dissipativity-based control for hidden Markov jump singularly perturbed systems subject to general probabilities. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 5752–5761. [Google Scholar] [CrossRef]
- Wang, L.; Wu, Z.G.; Shen, Y. Asynchronous mean stabilization of positive jump systems with piecewise-homogeneous Markov chain. IEEE Trans. Circuits Syst. II Exp. Briefs 2021, 68, 3266–3270. [Google Scholar] [CrossRef]
- Wang, B.; Zhu, Q. Stability analysis of discrete-time semi-Markov jump linear systems with partly unknown semi-Markov kernel. Syst. Control Lett. 2020, 140, 104688. [Google Scholar] [CrossRef]
- Zhao, X.; Deng, F.; Gao, W. Exponential stability of stochastic Markovian jump systems with time-varying and distributed delays. Sci. China Inf. Sci. 2021, 64, 209202:1–209202:3. [Google Scholar] [CrossRef]
- Han, X.; Wu, K.N.; Niu, Y. Asynchronous boundary control of Markov jump Neural networks with diffusion terms. IEEE Trans. Cybern. 2023, 53, 4962–4971. [Google Scholar] [CrossRef]
- Xue, M.; Yan, H.; Zhang, H.; Shen, H.; Peng, S. Dissipativity-based filter design for Markov jump systems with packet loss compensation. Automatica 2021, 133, 109843. [Google Scholar] [CrossRef]
- Hou, T.; Ma, H. Exponential stability for discrete-time infinite Markov jump systems. IEEE Trans. Autom. Control. 2016, 61, 4241–4246. [Google Scholar] [CrossRef]
- Ma, H.; Jia, Y. Stability analysis for stochastic differential equations with infinite Markovian switchings. J. Math. Anal. Appl. 2016, 435, 593–605. [Google Scholar] [CrossRef]
- Song, R.; Zhu, Q. Stability of linear stochastic delay differential equations with infinite Markovian switchings. Int. J. Robust Nonlinear Control 2018, 28, 825–837. [Google Scholar] [CrossRef]
- Hou, T.; Liu, Y.; Deng, F. Stability for discrete-time uncertain systems with infinite Markov jump and time-delay. Sci. China Inf. Sci. 2021, 64, 152202:1–152202:11. [Google Scholar] [CrossRef]
- Hou, T.; Liu, Y.; Deng, F. Finite horizon H2/H∞ control for SDEs with infinite Markovian jumps. Nonlinear Anal. Hybrid Syst. 2019, 34, 108–120. [Google Scholar] [CrossRef]
- Liu, Y.; Hou, T. Robust H2/H∞ fuzzy filtering for nonlinear stochastic systems with infinite Markov jump. J. Syst. Sci. Complex. 2020, 33, 1023–1039. [Google Scholar] [CrossRef]
- Dockner, E.J.; Jorgensen, S.; Long, N.V. Differential Games in Economics and Management Science; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
- Chen, B.S.; Tseng, C.S.; Uang, H.J. Fuzzy differential games for nonlinear stochastic systems: Suboptimal approach. IEEE Trans. Fuzzy Syst. 2002, 10, 222–233. [Google Scholar] [CrossRef]
- Lin, Y.; Zhang, T.; Zhang, W. Infinite horizon linear quadratic Pareto game of the stochastic singular systems. J. Frankl. Inst. 2018, 355, 4436–4452. [Google Scholar] [CrossRef]
- Moon, J. Linear-quadratic stochastic leader-follower differential games for Markov jump-diffusion models. Automatica 2023, 147, 110713. [Google Scholar] [CrossRef]
- Gao, X.; Deng, F.; Zeng, P. Zero-sum game-based security control of unknown nonlinear Markov jump systems under false data injection attacks. Int. J. Robust Nonlinear Control 2022. Early Access. [Google Scholar] [CrossRef]
- Dufour, F.; Prieto-Rumeau, T. Stationary Markov Nash equilibria for nonzero-sum constrained ARAT Markov games. SIAM J. Control Optim. 2022, 60, 945–967. [Google Scholar] [CrossRef]
- Hou, T.; Zhang, W. A game-based control design for discrete-time Markov jump systems with multiplicative noise. IET Control Theory Appl. 2013, 7, 773–783. [Google Scholar] [CrossRef]
- Sheng, L.; Zhang, W.; Gao, M. Relationship between Nash equilibrium strategies and H2/H∞ control of stochastic Markov jump systems with multiplicative noise. IEEE Trans. Autom. Control. 2014, 59, 2592–2597. [Google Scholar] [CrossRef]
- Sheng, L.; Zhang, W.; Gao, M. Some remarks on infinite horizon stochastic H2/H∞ control with (x, u, v) dependent noise and Markov jumps. J. Frankl. Inst. 2015, 352, 3929–3946. [Google Scholar] [CrossRef]
- Dragan, V.; Morozan, T.; Stoica, A.M. Mathematical Methods in Robust Control of Linear Stochastic Systems, 2nd ed.; Springer: New York, NY, USA, 2013. [Google Scholar]
- Liu, Y.; Hou, T. Infinite horizon LQ Nash Games for SDEs with infinite jumps. Asian J. Control 2021, 23, 2431–2443. [Google Scholar] [CrossRef]
- Rami, M.A.; Chen, X.; Zhou, X. Discrete-time indefinite LQ control with state and control dependent noises. J. Glob. Optim. 2002, 23, 245–265. [Google Scholar] [CrossRef]
- Basar, T.; Olsder, G.J. Dynamic Noncooperative Game Theory; SIAM: Philadelphia, PA, USA, 1999. [Google Scholar]
- Hou, T.; Zhang, W.; Ma, H. Finite horizon H2/H∞ control for discrete-time stochastic systems with Markovian jumps and multiplicative noise. IEEE Trans. Autom. Control. 2010, 55, 1185–1191. [Google Scholar]
- Wang, J.; Hou, T. Finite horizon H2/H∞ control for discrete-time time-varying stochastic systems with infinite Markov jumps. In Proceedings of the 36th Chinese Control Conference, Dalian, China, 26–28 July 2017. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).