Next Article in Journal
Building of a Mathematical Model for Solving the Elastic–Gas-Dynamic Task of the Gas Lubrication Theory for Petal Bearings
Next Article in Special Issue
Boundary Value Problem for a Loaded Pseudoparabolic Equation with a Fractional Caputo Operator
Previous Article in Journal
Finite Element Error Analysis of a Viscoelastic Timoshenko Beam with Thermodiffusion Effects
Previous Article in Special Issue
Transcendence and the Expression of the Spectral Series of a Class of Higher Order Differential Operators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Backward Stackelberg Games with Delay and Related Forward–Backward Stochastic Differential Equations

1
School of Science, China University of Mining and Technology, Beijing 100083, China
2
School of Mathematics and Statistics, Shandong University, Weihai 264209, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(13), 2898; https://doi.org/10.3390/math11132898
Submission received: 7 May 2023 / Revised: 26 June 2023 / Accepted: 27 June 2023 / Published: 28 June 2023

Abstract

:
In this paper, we study a kind of Stackelberg game where the controlled systems are described by backward stochastic differential delayed equations (BSDDEs). By introducing a new kind of adjoint equation, we establish the sufficient verification theorem for the optimal strategies of the leader and the follower in a general case. Then, we focus on the linear–quadratic (LQ) backward Stackelberg game with delay. The backward Stackelberg equilibrium is presented by the generalized fully coupled anticipated forward–backward stochastic differential delayed Equation (AFBSDDE), which is composed of anticipated stochastic differential equations (ASDEs) and BSDDEs. Moreover, we obtain the unique solvability of the AFBSDDE using the continuation method. As an application of the theoretical results, the pension fund problem with delay effect is considered.

1. Induction

Stackelberg games, also called leader–follower games, were introduced in [1] in order to investigate the optimal strategies in competitive systems in which there exist two asymmetric players. A Stackelberg game is a non-zero-sum game that assumes that one of the players (the follower) waits until the other player (the leader) announces their strategy, i.e., the follower chooses a strategy to optimize their own cost functional with respect to a given leader’s policy. Meanwhile, with the knowledge of the follower’s response, the leader will make a decision to optimize their own cost functional.
In recent years, considerable attention has been given to Stackelberg games. Under the deterministic framework, Simaan and Cruz [2] studied the Nash equilibrium and Tolwinski [3] obtained the closed-loop solution to a multistage LQ Stackelberg game. For the stochastic Stackelberg games, Øksendal et al. [4] established the maximum principle for a general stochastic Stackelberg differential game and applied it to news vendor problems. Bensoussan et al. [5] derived the maximum principle for the global solutions. They introduced several information structures, and obtained the optimal strategies of the follower and the leader under the adapted open-loop and adapted closed-loop memoryless information structures. Ref. [6] also studied the Stackelberg games under tow stochastic settings taken from [5], however, the diffusion term in their system allowed them to depend on the control variable. Shi et al. [7,8] studied the Stackelberg games with asymmetric information. Yong [9] illustrated the existence and uniqueness of the LQ Stackelberg game by the unique solvability of a stochastic Riccati equation. Wang and Zhang [10] considered a mean-field LQ Stackelberg differential game with one leader and two followers. Li et al. [11] proposed a kind of LQ Stackelberg game with multilevel hierarchy. They presented the Stackelberg equilibrium by forward–backward stochastic differential equations (FBSDEs) in a closed form.
On the other hand, controlled systems with delay have also received significant attention due to the path-dependence of the state and the time lag between the input and output (see [12,13,14,15,16]). Compared with the system driven by the Wiener process or the non-Gaussian process such as Rosenblatt process, the references [17,18,19,20,21] discussed Stackelberg games with different delays which are based on the system driven by a standard Brownian motion. In [17], the open-loop Stackelberg strategies were obtained for a deterministic LQ game with state delay. In [18,20], the authors were concerned with the games with delay appearing in the leader’s control. The open-loop strategies were explicitly given by some symmetric Riccati equations. In [19], the authors considered a linear–quadratic mean-field game between a leader and a group of followers, which may arise from an advertising model and an interbank lending and borrowing model with delay. Finally, in [21], the authors considered a general model in which the system involves both state delay and control delay. Under some assumptions, the state feedback representation of the Stackelberg strategy was derived.
The aforementioned papers are all based on the forward system. In addition to forward systems, controlled backward systems can also be encountered in optimal control problems and have a wide range of applications in finance and related fields. In particular, we will have a controlled BSDDE when considering delayed recursive utility or the pricing of derivative security for which the underlying goods are described by stochastic differential delayed equations (SDDEs). One can refer to [22,23,24,25,26]. Zheng and Shi [27] introduced a Stackelberg game described by a backward stochastic differential Equation (BSDE) and applied the theoretical results to solve a pension fund problem.
In this paper, we shall study the backward Stackelberg game with state delay. To the best of our knowledge, it is the first attempt to address this problem. The problem with delay remains challenging since the state feedback representation cannot be easily obtained in general cases. In order to obtain the optimal feedback of a LQ optimal control problem of BSDDEs, Meng and Shi [26] introduced a new class of delayed Riccati equations, which are very complicated to compute. In the Stackelberg game, the controlled system of the leader is more complex, which is described by coupled BSDDE and ASDE. Therefore, this paper focuses more on the solvability of the problem, rather than the feedback form of the control. The sufficient conditions of the backward equilibrium are given by the corresponding generalized Hamilton system composed of ASDE and BSDDE. For the LQ backward Stackelberg game with delay, we obtain that the solvability of the game is equivalent to that of a related AFBSDDE. Furthermore, we discuss the unique solvability of the AFBSDDE.
The rest of this paper is organized as follows. Section 2 is devoted to presenting the sufficient conditions of the follower and the leader in the general backward Stackelberg game with delay. In Section 3, the LQ backward Stackelberg game with delay is investigated. A new kind of AFBSDDE is used to characterize the unique equilibrium of the game. In Section 4, as an application, a pension fund problem in markets with delayed effects is discussed. Some conclusions and unsolved issues for future research are displayed in Section 5.
To conclude this section, we introduce some notations which will be used throughout this paper. Let R k be the k-dimensional Euclidean space with inner product · , · and norm | · | . Let R k × d be the Hilbert space consisting of all k × d matrices with the inner product A , B : = tr { A B } , for any A , B R k × d , where appears in the superscript and denotes the transposition of a matrix. In particular, S n denotes the set of all n × n symmetric matrices.
Let ( Ω , F , F , P ) be a complete probability space equipped with natural filtration F = { F t } t 0 generated by a d-dimensional standard Brownian motion W ( · ) . Assume that F t = σ { W ( s ) , 0 s t } N , where N denotes the totality of P-null sets. T > 0 is a finite time duration, and 0 δ < T is a time delay. We define F t F 0 for all t [ δ , 0 ) and F = F T + δ . We introduce some spaces which will be used throughout this paper as follows:
L p ( F T , R k ) = ξ : ξ is R k valued - F T - measurable random variable such that E | ξ | p < + , L F p ( a , b ; R k ) = ξ ( t ) : ξ ( t ) is R k - valued F - adapted process such that E a b | ξ ( t ) | p d t < + , S F p ( a , b ; R k ) = ξ ( t ) : ξ ( t ) is R k - valued F - adapted process such that E sup a t b | ξ ( t ) | p < + , L F ( a , b ; R k ) = ξ ( t ) : ξ ( t ) is R k - valued F - adapted bounded process ,
where p 1 is a real number.

2. Backward Stackelberg Game with Delay

Let us consider a backward Stackelberg game with a time delay involving two players. For the sake of simplicity, we only consider the following BSDDE driven by a standard Brownian motion W ( · ) :
d y ( t ) = g ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u ( t ) , w ( t ) ) d t z ( t ) d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) ,
where ( y ( · ) , z ( · ) ) R n × R n × d is the state process pair with terminal condition γ and initial path ( η ( · ) , ζ ( · ) ) . We assume that g : Ω × [ 0 , T ] × R n × R n × R n × d × R n × d × R m 1 × R m 2 R n is a given function and η ( · ) , ζ ( · ) are continuous functions on [ δ , 0 ] . u ( · ) and w ( · ) constitutes the control decisions of player 1 (the follower) and player 2 (the leader) with values in U 1 and U 2 . Additionally, U 1 and U 2 are non-empty convex sub-sets of R m 1 and R m 2 , respectively.
The admissible strategy sets for the follower and the leader are denoted by
U 1 : = { u ( · ) L F 2 ( 0 , T ; R m 1 ) | u ( · ) U 1 , a . e . a . s . } , U 2 : = { w ( · ) L F 2 ( 0 , T ; R m 2 ) | w ( · ) U 2 , a . e . a . s . } .
We suppose that the coefficients satisfy the following assumptions (H1):
(H1.1)
η ( · ) , ζ ( · ) L F 2 ( δ , 0 ; R n ) , γ L 2 ( F T ; R n ) ;
(H1.2)
For any y , y δ R n , z , z δ R n × d and u R m 1 , w R m 2 , the mapping g ( · , y , y δ , z , z δ , u , w ) is F -adapted;
(H1.3)
g ( · , 0 , 0 , 0 , 0 , 0 , 0 ) L F 2 ( 0 , T ; R n ) ;
(H1.4)
g is twice continuously differentiable with respect to ( y , y δ , z , z δ , u , w ) and the derivatives up to two orders of g are uniformly bounded.
The cost functionals of the follower and the leader to minimize, respectively, are described as follows:
J 1 ( γ ; u ( · ) , w ( · ) ) = E [ 0 T l 1 ( t , y ( t ) , z ( t ) , u ( t ) , w ( t ) ) d t + G 1 ( y ( 0 ) ) ] , J 2 ( γ ; u ( · ) , w ( · ) ) = E [ 0 T l 2 ( t , y ( t ) , z ( t ) , u ( t ) , w ( t ) ) d t + G 2 ( y ( 0 ) ) ] .
where l i : Ω × [ 0 , T ] × R n × R n × d × R m 1 × R m 2 R and G i : Ω × R n R ( i = 1 , 2 ) are also given continuous functions.
Now, we give the assumptions (H2) of the cost functionals for i = 1 , 2 .
(H2.1)
For any y R n , z R n × d , u R m 1 , w R m 2 , the mapping l i ( · , y , z , u , w ) is F -adapted and G i ( y ) is F 0 -measurable;
(H2.2)
l i ( · , 0 , 0 , 0 , 0 ) L F 1 ( 0 , T ; R ) and G i ( 0 ) L 1 ( F 0 ; R ) ;
(H2.3)
l i is twice continuously differentiable with respect to ( y , z , u , w ) and G i is twice continuously differentiable with respect to y. The derivatives up to two orders of l i and G i are uniformly bounded;
(H2.4)
l i y , l i z , l i u , l i w are bounded by C ( 1 + | y | + | z | + | u | + | w | ) and G i y is bounded by C ( 1 + | y | ) .
From Chen and Huang [24], it follows that the BSDDE (1) admits a unique pair of solutions ( y ( · ) , z ( · ) ) L F 2 ( δ , T ; R n ) × S F 2 ( δ , T ; R n × d ) for a sufficiently small time delay δ and any ( u ( · ) , w ( · ) ) U 1 × U 2 . The functional J i ( i = 1 , 2 ) in (2) is also well defined.
The main feature of the Stackelberg game is that the follower makes their decision according to the leader’s strategy. In our backward framework, for any w ( · ) U 2 of the leader and given terminal condition γ L 2 ( F T ; R n ) , the follower would like to choose u * ( · ) U 1 such that J 1 ( γ ; u * ( · ) , w ( · ) ) is the minimum of J 1 ( γ ; u ( · ) , w ( · ) ) over u ( · ) U 1 . The rational choice of the leader is determined by knowing the follower’s optimal control u * ( · ) . The leader would like to choose w * ( · ) U 2 to minimize J 2 ( γ ; u * ( · ) , w ( · ) ) over w ( · ) U 2 . Then, we propose the definition of the backward Stackelberg equilibrium similarly to [27].
Definition 1.
The pair ( u * ( · ) , w * ( · ) ) U 1 × U 2 is called the Stackelberg equilibrium of the backward Stackelberg game with delay, if it satisfies the following conditions. First, for a given γ L 2 ( F T , R n ) and any w ( · ) U 2 , there exists a map Γ : U 2 × L 2 ( F T , R n ) U 1 such that
J 1 ( γ ; Γ ( w ( · ) , γ ) , w ( · ) ) = inf u ( · ) U 1 J 1 ( γ ; u ( · ) , w ( · ) ) .
Secondly, there exists w * ( · ) U 2 , such that
J 2 ( γ ; Γ ( w * ( · ) , γ ) , w * ( · ) ) = inf w ( · ) U 2 J 2 ( γ ; Γ ( w ( · ) , γ ) , w ( · ) ) ,
and the optimal strategy of the follower is u * ( · ) = Γ ( w * ( · ) , γ ) .

2.1. Optimization for the Follower

For a given γ L 2 ( F T , R n ) and w ( · ) U 2 , the follower faces an optimal control problem as follows:
  • Problem  ( B S G ) f
inf u U 1 J 1 ( γ ; u ( · ) , w ( · ) ) ,
s . t . d y ( t ) = g ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u ( t ) , w ( t ) ) d t z ( t ) d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) .
Due to the adjoint relationship between BSDDE and ASDE, we introduce the following adjoint ASDE:
d x ( t ) = { g y ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) x ( t ) + E F t [ g y δ ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) | t + δ x ( t + δ ) ] l 1 y ( t , y ( t ) , z ( t ) , u * ( t ) , w ( t ) ) } d t + { g z ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) x ( t ) + E F t [ g z δ ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) | t + δ x ( t + δ ) ] l 1 z ( t , y ( t ) , z ( t ) , u * ( t ) , w ( t ) ) } d W ( t ) , t [ 0 , T ] , x ( 0 ) = G 1 y ( y ( 0 ) ) , x ( t ) = 0 , t ( T , T + δ ] .
Here, g y δ and g z δ denote the partial derivatives of g with respect to y ( t δ ) and z ( t δ ) , respectively. g y δ ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) | t + δ denotes the value of g y δ when time t takes value t + δ , i.e., g y δ ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) | t + δ = g y δ ( t + δ , y ( t + δ ) , y ( t ) , z ( t + δ ) , z ( t ) , u * ( t + δ ) , w ( t + δ ) ) .
Remark 1.
The assumptions (H1)–(H2) imply that BSDDE (1) admits a unique solution for ( u ( · ) , w ( · ) ) U 1 × U 2 , the cost functional (2) is well defined, and ASDE (3) admits a unique solution for ( u * ( · ) , w ( · ) ) U 1 × U 2 and the sufficiently small time delay δ by Theorem 2.2 in [24].
We define the Hamiltonian function H 1 : [ 0 , T ] × R n × R n × R n × d × R n × d × R m 1 × R m 2 × R n R by
H 1 ( t , y , y δ , z , z δ , u , w , x ) = l 1 ( t , y , z , u , w ) g ( t , y , y δ , z , z δ , u , w ) , x .
Then, the adjoint Equation (3) can be rewritten as the following Hamiltonian type:
d x ( t ) = { H 1 y ( t ) E F t [ H 1 y δ ( t ) | t + δ ] } d t + { H 1 z ( t ) E F t [ H 1 z δ ( t ) | t + δ ] } d W ( t ) , t [ 0 , T ] , x ( 0 ) = G 1 y ( y ( 0 ) ) , x ( t ) = 0 , t ( T , T + δ ] ,
where H 1 k ( t ) = H 1 k ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) , x ( t ) ) , k = y , y δ , z , z δ , u .
Now, we can give the sufficient conditions to ensure the optimality of the follower’s optimal control problem using Theorem 3.2 in [24].
Theorem 1.
Let (H1)–(H2) hold and x ( · ) be the solution of adjoint Equation (3) associated with ( u * ( · ) , w ( · ) ) U 1 × U 2 . Suppose that H 1 ( t , · , · , · , · , · , w ( t ) , x ( t ) ) is convex in ( y , y δ , z , z δ , u ) for any t [ 0 , T ] and G 1 ( · ) is convex in y. If
H 1 u ( t ) , u u * ( t ) 0 , u U 1 , a . e . , a . s .
holds, then u * ( · ) is an optimal strategy of the follower’s Problem ( B S G ) f .
Remark 2.
What we need to clarify is that ( y ( · ) , z ( · ) ) and x ( · ) in Equation (3) and inequality (6) should be the corresponding solutions with respect to ( u * ( · ) , w ( · ) ) .

2.2. Optimization for the Leader

Since the follower has adopted strategy u * ( t ) , then we can obtain the following control problem of the leader.
  • Problem  ( B S G ) l :
inf w U 2 J 2 ( γ ; u * ( · ) , w ( · ) ) ,
s . t . d y ( t ) = g ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) d t z ( t ) d W ( t ) , t [ 0 , T ] , d x ( t ) = { H 1 y ( t ) E F t [ H 1 y δ ( t ) | t + δ ] } d t + { H 1 z ( t ) E F t [ H 1 z δ ( t ) | t + δ ] } d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) , x ( 0 ) = G 1 y ( y ( 0 ) ) , x ( t ) = 0 , t ( T , T + δ ] .
Remark 3.
We should stress that, in Equation (7), the backward equation is a BSDDE and the forward equation is an ASDE. This is different from the results in the existing research literature, which are focused on forward–backward controlled systems with forward SDDE and anticipated backward stochastic differential Equation (ABSDE) (see [28,29,30,31]).
We define the Hamiltonian function of the leader by
H 2 ( t , y , y δ , z , z δ , x , x δ , u * , w , ξ , p , q ) = l 2 ( t , y , z , u * , w ) + p , b ( t , Σ , x , x δ , u * , w ) + q , σ ( t , Σ , x , x δ , u * , w ) ξ , g ( t , Σ , u * , w ) .
Here,
b ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) = g y ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) x ( t ) + E F t [ g y δ ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) | t + δ x ( t + δ ) ] l 1 y ( t , y ( t ) , z ( t ) , u * ( t ) , w ( t ) ) ,
σ ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) = g z ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) x ( t ) + E F t [ g z δ ( t , y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) , u * ( t ) , w ( t ) ) | t + δ x ( t + δ ) ] l 1 z ( t , y ( t ) , z ( t ) , u * ( t ) , w ( t ) ) ,
with Σ ( t ) = ( y ( t ) , y ( t δ ) , z ( t ) , z ( t δ ) ) .
In order to give the sufficient conditions for the optimality of the leader’s strategy, we need the following additional assumptions.
  • (H3)  G 1 ( y ) = 1 2 K y y , for K R n × n .
  • (H4)  H 2 ( t , · , · , · , · , · , · , u * ( t ) , · , ξ ( t ) , p ( t ) , q ( t ) ) is convex in ( y , y δ , z , z δ , x , x δ , w ) and G 2 ( · ) is convex in y for any t [ 0 , T ] .
Theorem 2.
Let (H1)–(H4) hold. Assume that the follower adopted their optimal strategy u * ( t ) U 1 , and w * ( · ) is an admissible control of the leader. ( y * ( · ) , z * ( · ) , x * ( · ) ) is the corresponding solution with respect to ( u * ( · ) , w * ( · ) ) with y * ( T ) = γ . We also assume that ( p ( t ) , q ( t ) , ξ ( t ) ) is the unique solution of the following adjoint equation:
d p ( t ) = { H 2 x ( t ) + H 2 x δ ( t ) | t δ } d t q ( t ) d W ( t ) , t [ 0 , T ] , d ξ ( t ) = { H 2 y ( t ) E F t [ H 2 y δ ( t ) | t + δ ] } d t + { H 2 z ( t ) E F t [ H 2 z δ ( t ) | t + δ ] } d W ( t ) , t [ 0 , T ] , p ( T ) = 0 , p ( t ) = 0 , q ( t ) = 0 , t [ δ , 0 ) , ξ ( 0 ) = G 1 y y ( y * ( 0 ) ) p ( 0 ) G 2 y ( y * ( 0 ) ) , ξ ( t ) = 0 , t ( T , T + δ ] ,
where
H 2 k ( t ) = H 2 k ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) , p ( t ) , q ( t ) , ξ ( t ) ) ,
with k = y , y δ , z , z δ , x , x δ , w . Then, w * ( · ) is an optimal strategy of the leader if it satisfies
H 2 w ( t ) , w w * ( t ) 0 , w U 2 , a . e . , a . s . .
Proof
Let w ( · ) be an arbitrary admissible control of the leader and Σ ( t ) be the corresponding trajectory.
We consider
J 2 ( γ , ; u * ( · ) , w * ( t ) ) J 2 ( γ ; u * ( · ) , w ( · ) ) = I + I I ,
where
I = E 0 T [ l 2 ( t , y * ( t ) , z * ( t ) , u * ( t ) , w * ( t ) ) l 2 ( t , y ( t ) , z ( t ) , u * ( t ) , w ( t ) ) ] d t , I I = E [ G 2 ( y * ( 0 ) G 2 ( y ( 0 ) ) ] .
By the convexity of H 2 and G 2 , we have
I = E 0 T [ H 2 ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) , ξ ( t ) , p ( t ) , q ( t ) ) H 2 ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) , ξ ( t ) , p ( t ) , q ( t ) ) ] d t E 0 T { p ( t ) , b ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) ) b ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) + tr { q T ( t ) ( σ ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) ) σ ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] u * ( t ) , w ( t ) ) ) } ξ ( t ) , g ( t , Σ * ( t ) , u * ( t ) , w * ( t ) ) g ( t , Σ ( t ) , u * ( t ) , w ( t ) ) } d t E 0 T { y * ( t ) y ( t ) , H 2 y ( t ) + y * ( t δ ) y ( t δ ) , H 2 y δ ( t ) + z * ( t ) z ( t ) , H 2 z ( t ) + z * ( t δ ) z ( t δ ) , H 2 z δ ( t ) + w * ( t ) w ( t ) , H 2 w ( t ) + x * ( t ) x ( t ) , H 2 x ( t ) + E F t [ x * ( t + δ ) x ( t + δ ) ] , H 2 x δ ( t ) } d t E 0 T { p ( t ) , b ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) ) b ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) + tr { q ( t ) ( σ ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) ) σ ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) ) } ξ ( t ) , g ( t , Σ * ( t ) , u * ( t ) , w * ( t ) ) g ( t , Σ ( t ) , u * ( t ) , w ( t ) ) } d t ,
I I E [ G 2 y ( y * ( 0 ) ) ( y * ( 0 ) y ( 0 ) ) ] = E [ ξ ( 0 ) + p ( 0 ) G 1 y y ( y * ( 0 ) ) ( y * ( 0 ) y ( 0 ) ) ] .
Applying It o ^ ’s formula to ξ ( t ) , y * ( t ) y ( t ) + p ( t ) , x * ( t ) x ( t ) , we derive that
E [ ξ ( 0 ) ( y * ( 0 ) y ( 0 ) ) + p ( 0 ) G 1 y ( y * ( 0 ) ) G 1 y ( y ( 0 ) ) ] = E 0 T { ξ ( t ) , g ( t , Σ * ( t ) , u * ( t ) , w * ( t ) ) g ( t , Σ ( t ) , u * ( t ) , w ( t ) ) + y * ( t ) y ( t ) , H 2 y ( t ) + E F t [ H 2 y δ ( t ) | t + δ ] + z * ( t ) z ( t ) , H 2 z ( t ) + E F t [ H 2 z δ ( t ) | t + δ ] } d t E 0 T { x * ( t ) x ( t ) , H 2 x ( t ) + H 2 x δ ( t ) | t δ p ( t ) , b ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) ) b ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) tr { q ( t ) ( σ ( t , Σ * ( t ) , x * ( t ) , E F t [ x * ( t + δ ) ] , u * ( t ) , w * ( t ) ) σ ( t , Σ ( t ) , x ( t ) , E F t [ x ( t + δ ) ] , u * ( t ) , w ( t ) ) ) } } d t .
Therefore, we can obtain the following result using the optimal condition (10).
J 2 ( γ , ; u * ( · ) , w * ( t ) ) J 2 ( γ ; u * ( · ) , w ( · ) ) E 0 T y * ( t δ ) y ( t δ ) , H 2 y δ ( t ) y * ( t ) y ( t ) , E F t [ H 2 y δ ( t ) | t + δ ] d t + E 0 T z * ( t δ ) z ( t δ ) , H 2 z δ ( t ) z * ( t ) z ( t ) , E F t [ H 2 z δ ( t ) | t + δ ] d t + E 0 T E F t [ x * ( t + δ ) x ( t + δ ) ] , H 2 x δ ( t ) x * ( t ) x ( t ) , H 2 x δ ( t ) | t δ d t
Now, we need to analyze the right-hand side of inequality (11). In fact,
E 0 T y * ( t δ ) y ( t δ ) , H 2 y δ ( t ) y * ( t ) y ( t ) , E F t [ H 2 y δ ( t ) | t + δ ] d t = E δ T δ y * ( t ) y ( t ) , E F t [ H 2 y δ ( t ) | t + δ ] d t E 0 T y * ( t ) y ( t ) , E F t [ H 2 y δ ( t ) | t + δ ] d t = E δ 0 y * ( t ) y ( t ) , E F t [ H 2 y δ ( t ) | t + δ ] d t E T δ T y * ( t ) y ( t ) , E F t [ H 2 y δ ( t ) | t + δ ] d t = 0 ,
since y * ( t ) = y ( t ) for any t [ δ , 0 ] and H 2 y δ ( t ) = 0 for any t [ T , T + δ ] . Similarly, we have
E 0 T z * ( t δ ) z ( t δ ) , H 2 z δ ( t ) z * ( t ) z ( t ) , E F t [ H 2 z δ ( t ) | t + δ ] d t = 0 ,
E 0 T E F t [ x * ( t + δ ) x ( t + δ ) ] , H 2 x δ ( t ) x * ( t ) x ( t ) , H 2 x δ ( t ) | t δ d t = 0 ,
Then, we deduce J 2 ( γ , ; u * ( · ) , w * ( t ) ) J 2 ( γ ; u * ( · ) , w ( · ) ) 0 and complete the proof. □
Remark 4.
The adjoint Equation (9) is a new kind of FBSDE. Due to the complex form of the Hamiltonian function H 2 , the conditions (H1)–(H2) may not guarantee the solvability of this equation. In addition, in Theorem 2, we look for the leader’s optimal strategy under the assumption that u * ( · ) is known. However, the follower’s optimal strategy u * ( t ) depends on the current value of the leader’s strategy w ( · ) . So, we can set
u * ( t ) = u * ( t , y ( t ) , z ( t ) , y ( t δ ) , z ( t δ ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) : = u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) ,
for its dependence on the leader’s strategy w ( · ) , the state ( y ( · ) , z ( · ) ) and the adjoint variable x ( t ) .
Then, the state equation of the leader can be rewritten as
d y ( t ) = g ( t , Σ ( t ) , u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) d t z ( t ) d W ( t ) , t [ 0 , T ] , d x ( t ) = { g y ( t , Σ ( t ) , u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) x ( t ) + E F t [ g y δ ( t , Σ ( t ) , u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) | t + δ x ( t + δ ) ] l 1 y ( t , y ( t ) , z ( t ) , u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) } d t + { g z ( t , Σ ( t ) , u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) x ( t ) + E F t [ g z δ ( t , Σ ( t ) ) , u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) | t + δ x ( t + δ ) ] l 1 z ( t , y ( t ) , z ( t ) , u * ( t , Σ ( t ) ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) , w ( t ) ) } d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) , x ( 0 ) = G 1 y ( y ( 0 ) ) , x ( t ) = 0 , t ( T , T + δ ] .
If we want to obtain the sufficient conditions of the optimal strategy in this case, we need the function u * ( t , Σ ( t ) , w ( t ) , x ( t ) , E F t [ x ( t + δ ) ] ) to satisfy some conditions. In Bensoussan et al. [5], they assumed that u * is uniquely defined and is uniformly Lipschitz-continuous with respect to the state variable, the leader’s control variable and the adjoint variable.
In our framework, the backward Stackelberg equilibrium relies on a fully coupled AFBSDDE consisting of the above equation and adjoint Equation (9). Since it is difficult to obtain the solvability of AFBSDDE in the general case, we will discuss the LQ backward Stackelberg game in the next section.

3. Linear–Quadratic Backward Stackelberg Game

In this section, we apply the theory studied in Section 2 to deal with an LQ backward Stackelberg game. We assume that the Brownian motion is one-dimensional for simplicity. The controlled system is given by
d y ( t ) = [ A ( t ) y ( t ) + B ( t ) y ( t δ ) + C ( t ) u ( t ) + D ( t ) w ( t ) + E ( t ) z ( t ) + F ( t ) z ( t δ ) ] d t z ( t ) d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) ,
where A ( · ) , B ( · ) , C ( · ) , E ( · ) , D ( · ) , and F ( · ) are deterministic continuous matrix-valued functions with suitable dimensions.
The corresponding cost functionals of the follower and the leader in system (12) are
J 1 ( γ ; u ( · ) , w ( · ) ) = 1 2 E { 0 T [ M 1 ( t ) y ( t ) , y ( t ) + R 1 ( t ) z ( t ) , z ( t ) + N 1 ( t ) u ( t ) , u ( t ) ] d t + G 1 y ( 0 ) , y ( 0 ) } , J 2 ( γ ; u ( · ) , w ( · ) ) = 1 2 E { 0 T [ M 2 ( t ) y ( t ) , y ( t ) + R 2 ( t ) z ( t ) , z ( t ) + N 2 ( t ) w ( t ) , w ( t ) ] d t + G 2 y ( 0 ) , y ( 0 ) } .
Then, we give the assumption (H5) which is a special case of (H1)–(H4).
  • (H5)  A ( · ) , B ( · ) , E ( · ) , F ( · ) L F ( 0 , T ; R n × n ) , C ( · ) L F ( 0 , T ; R n × m 1 ) , D ( · ) L F ( 0 , T ; R n × m 2 ) . For i = 1 , 2 , we suppose M i ( · ) , R i ( · ) L F ( 0 , T ; S n ) , N i ( · ) L F ( 0 , T ; S m i ) , G i S n , and M i ( · ) 0 , R i ( · ) 0 , N i ( · ) 0 , G i 0 .
We still consider the follower’s optimal control problem first.
  • Problem  ( L Q B S G ) f : For γ L 2 ( F T ; R n ) and any given leader’s control w ( · ) U 2 , we find u * ( · ) U 1 such that J 1 ( γ ; u * ( · ) , w ( · ) ) = inf u ( · ) U 1 J 1 ( γ ; u ( · ) , w ( · ) ) .
Applying Theorem 1 to Problem ( L Q B S G ) f , we have the following result:
Theorem 3.
Suppose that (H5) holds. For γ L 2 ( F T ; R n ) and any given leader’s control w ( · ) U 2 , the follower’s optimal control problem ( L Q B S G ) f can be uniquely solvable if and only if there exists a unique four-tuple ( y ( t ) , z ( t ) , u * ( t ) , x ( t ) ) satisfying
d y ( t ) = [ A ( t ) y ( t ) + B ( t ) y ( t δ ) + C ( t ) u * ( t ) + D ( t ) w ( t ) + E ( t ) z ( t ) + F ( t ) z ( t δ ) ] d t z ( t ) d W ( t ) , t [ 0 , T ] , d x ( t ) = { A ( t ) x ( t ) + E F t [ B ( t + δ ) x ( t + δ ) ] M 1 ( t ) y ( t ) } d t + { E ( t ) x ( t ) + E F t [ F ( t + δ ) x ( t + δ ) ] R 1 ( t ) z ( t ) } d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) , x ( 0 ) = G 1 y ( 0 ) , x ( t ) = 0 , t ( T , T + δ ] ,
such that
u * ( t ) = N 1 1 ( t ) C ( t ) x ( t ) .
The second equation in (14) is a linear ASDE and it admits a unique solution under assumption (H5) for a sufficiently small δ . Since the leader can choose their optimal strategy with the knowledge of the follower’s reaction, then the state equation of the leader’s optimal control is
d y ( t ) = [ A ( t ) y ( t ) + B ( t ) y ( t δ ) + C ( t ) N 1 1 ( t ) C ( t ) x ( t ) + D ( t ) w ( t ) + E ( t ) z ( t ) + F ( t ) z ( t δ ) ] d t z ( t ) d W ( t ) , t [ 0 , T ] , d x ( t ) = { A ( t ) x ( t ) + E F t [ B ( t + δ ) x ( t + δ ) ] M 1 ( t ) y ( t ) } d t + { E ( t ) x ( t ) R 1 ( t ) z ( t ) + E F t [ F ( t + δ ) x ( t + δ ) ] } d W ( t ) , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) , x ( 0 ) = G 1 y ( 0 ) , x ( t ) = 0 , t ( T , T + δ ] .
  • Problem  ( L Q B S G ) l :Find the optimal strategy w * ( · ) U 2 such that J 2 ( γ ; u * ( · ) , w * ( · ) ) = inf w ( · ) U 2 J 2 ( γ ; u * ( · ) , w ( · ) ) .
The adjoint equation of the leader is the following FBSDE:
d p ( t ) = { A ( t ) p ( t ) + B ( t ) p ( t δ ) C ( t ) N 1 1 ( t ) C ( t ) ξ ( t ) + E ( t ) q ( t ) + F ( t ) q ( t δ ) } d t q ( t ) d W ( t ) , t [ 0 , T ] , d ξ ( t ) = { A ( t ) ξ ( t ) + E F t [ B ( t + δ ) ξ ( t + δ ) ] + M 1 ( t ) p ( t ) M 2 ( t ) y ( t ) } d t + { E ( t ) ξ ( t ) + E F t [ F ( t + δ ) ξ ( t + δ ) ] + R 1 ( t ) q ( t ) R 2 ( t ) z ( t ) } d W ( t ) , t [ 0 , T ] , p ( T ) = 0 , p ( t ) = 0 , q ( t ) = 0 , t [ δ , 0 ) , ξ ( 0 ) = G 1 p ( 0 ) G 2 y ( 0 ) , ξ ( t ) = 0 , t ( T , T + δ ] .
Theorem 4.
Under assumption ( H 5 ) , the leader’s optimal control problem ( L Q B S G ) l can be uniquely solvable if and only if there exists a unique ( y * ( t ) , z * ( t ) , x * ( t ) , p * ( t ) , q * ( t ) , ξ * ( t ) , w * ( t ) ) satisfying Equation (16) and Equation (17) such that
w * ( t ) = N 2 1 ( t ) D ( t ) ξ * ( t ) .
We declare that, if FBSDE (16) and (17) admit a unique solution, then w * ( t ) in the form of (18) is the unique optimal control of the leader by Theorem 2. Meanwhile, we can also prove that w * ( t ) in (18) is optimal using the classical completion of the squares method. Thus, we omit the detailed proof of Theorem 4.
Theorems 3 and 4 show us an equivalence between the solvability of the LQ backward Stackelberg game and that of the coupled AFBSDDE. In order to achieve the existence of the optimal control ( u * , w * ) , we study the solvability of the following AFBSDDE.
d y ( t ) = [ A ( t ) y ( t ) + B ( t ) y ( t δ ) + I ( t ) C ( t ) x ( t ) + I ¯ ( t ) D ( t ) ξ ( t ) + E ( t ) z ( t ) + F ( t ) z ( t δ ) ] d t z ( t ) d W ( t ) , t [ 0 , T ] , d x ( t ) = { A ( t ) x ( t ) + E F t [ B ( t + δ ) x ( t + δ ) ] M 1 ( t ) y ( t ) } d t + { E ( t ) x ( t ) + E F t [ F ( t + δ ) x ( t + δ ) ] R 1 ( t ) z ( t ) } d W ( t ) , t [ 0 , T ] , d p ( t ) = { A ( t ) p ( t ) + B ( t ) p ( t δ ) I ( t ) C ( t ) ξ ( t ) + E ( t ) q ( t ) + F ( t ) q ( t δ ) } d t q ( t ) d W ( t ) , t [ 0 , T ] , d ξ ( t ) = { A ( t ) ξ ( t ) + E F t [ B ( t + δ ) ξ ( t + δ ) ] + M 1 ( t ) p ( t ) M 2 ( t ) y ( t ) } d t + { E ( t ) ξ ( t ) + E F t [ F ( t + δ ) ξ ( t + δ ) ] + R 1 ( t ) q ( t ) R 2 ( t ) z ( t ) } d W t , t [ 0 , T ] , y ( T ) = γ , y ( t ) = η ( t ) , z ( t ) = ζ ( t ) , t [ δ , 0 ) , x ( 0 ) = G 1 y ( 0 ) , x ( t ) = 0 , t ( T , T + δ ] , p ( T ) = 0 , p ( t ) = 0 , q ( t ) = 0 , t [ δ , 0 ) , ξ ( 0 ) = G 1 p ( 0 ) G 2 y ( 0 ) , ξ ( t ) = 0 , t ( T , T + δ ] ,
where I t = C ( t ) N 1 1 ( t ) , I ¯ t = D ( t ) N 2 1 ( t ) .
Set
λ ( t ) = ( y ( t ) , z ( t ) , x ( t ) ) , θ ( t ) = ( p ( t ) , q ( t ) , ξ ( t ) ) ,
and define
L 2 ( δ , T + δ ) = S F 2 ( δ , T ; R n ) × L F 2 ( δ , T ; R n × d ) × S F 2 ( 0 , T + δ ; R n ) × S F 2 ( δ , T ; R n ) × L F 2 ( δ , T ; R n × d ) × S F 2 ( 0 , T + δ ; R n )
with the norm
( λ ( · ) , θ ( · ) ) 2 = E sup t [ δ , T ] | y ( t ) | 2 + E δ T | z ( t ) | 2 d t + E sup t [ 0 , T + δ ] | x ( t ) | 2 + E sup t [ δ , T ] | p ( t ) | 2 + E δ T | q ( t ) | 2 d t + E sup t [ 0 , T + δ ] | ξ ( t ) | 2 .
Remark 5.
As mentioned before, Equation (19) is a new kind of double FBSDE composed of ASDEs and BSDDEs. This kind of FBSDE has not been studied in the previous literature. Using the method of continuation introduced in [32], we can obtain the unique solvability of AFBSDDE (19). We should note that the delay time δ must be sufficiently small for ASDEs and BSDDEs. We will give the detailed procedure below.
Theorem 5.
For sufficiently small δ, AFBSDDE (19) admits a unique solution
( λ ( · ) , θ ( · ) ) = ( y ( · ) , z ( · ) , x ( · ) , p ( · ) , q ( · ) , ξ ( · ) ) L 2 ( δ , T + δ ) .
In order to prove Theorem 5, we consider a family of AFBSDDEs with parameter ρ [ 0 , 1 ] as follows.
d y ρ ( t ) = { ρ ( A ( t ) y ρ ( t ) + B ( t ) y ρ ( t δ ) + I ( t ) C ( t ) x ρ ( t ) + I ¯ ( t ) D ( t ) ξ ρ ( t ) + E ( t ) z ρ ( t ) + F ( t ) z ρ ( t δ ) ) + ϑ 1 ( t ) + ( 1 ρ ) ( D ( t ) D ( t ) + C ( t ) C ( t ) ) ξ ρ ( t ) } d t z ρ ( t ) d W ( t ) , t [ 0 , T ] , d x ρ ( t ) = { ρ A ( t ) x ρ ( t ) + E F t [ B ( t + δ ) x ρ ( t + δ ) ] M 1 ( t ) y ρ ( t ) + κ 1 ( t ) } d t + { ρ ( E ( t ) x ρ ( t ) + E F t [ F ( t + δ ) x ρ ( t + δ ) ] R 1 ( t ) z ρ ( t ) ) + ϖ 1 ( t ) } d W ( t ) , t [ 0 , T ] , d p ρ ( t ) = { ρ ( A ( t ) p ρ ( t ) + B ( t ) p ρ ( t δ ) I ( t ) C ( t ) ξ ρ ( t ) + E ( t ) q ρ ( t ) + F ( t ) q ρ ( t δ ) ) + ϑ 2 ( t ) + ( 1 ρ ) C ( t ) C ( t ) x ρ ( t ) } d t q ρ ( t ) d W ( t ) , t [ 0 , T ] , d ξ ρ ( t ) = { ρ A ( t ) ξ ρ ( t ) + E F t [ B ( t + δ ) ξ ρ ( t + δ ) ] + M 1 ( t ) p ρ ( t ) M 2 ( t ) y ρ ( t ) + κ 2 ( t ) } d t + { ρ ( E ( t ) ξ ρ ( t ) + E F t [ F ( t + δ ) ξ ρ ( t + δ ) ] + R 1 ( t ) q ρ ( t ) R 2 ( t ) z ρ ( t ) ) + ϖ 2 ( t ) } d W ( t ) , t [ 0 , T ] , y ρ ( T ) = γ , y ρ ( t ) = η ( t ) , z ρ ( t ) = ζ ( t ) , t [ δ , 0 ) , x ρ ( 0 ) = ρ G 1 y ρ ( 0 ) + ϕ 1 , x ρ ( t ) = 0 , t ( T , T + δ ] , ξ ρ ( 0 ) = ρ G 1 p ρ ( 0 ) G 2 y ρ ( 0 ) + ϕ 2 , ξ ρ ( t ) = 0 , t ( T , T + δ ] , p ρ ( T ) = ψ , p ρ ( t ) = 0 , q ρ ( t ) = 0 , t [ δ , 0 ) ,
where ϑ i ( · ) , κ i ( · ) , ϖ i ( · ) L F 2 ( 0 , T ; R n ) , ψ L 2 ( F T ; R n ) , ϕ i L 2 ( F 0 ; R n ) ( i = 1 , 2 ) .
When ρ = 0 , FBSDE (20) is decoupled. Equation (20) has a unique solution by the properties of SDE and BSDE. When ρ = 1 , if Equation (20) has a unique solution, then the AFBSDDE (19) also has a unique solution. The following lemma gives a prior estimate for the existence of an interval of Equation (20) with respect to parameter ρ [ 0 , 1 ] .
Lemma 1.
We assume that (H5) holds. There exists a constant τ 0 > 0 such that, if AFBSDDE (20) admits a unique solution ( λ t ρ 0 , θ t ρ 0 ) for some ρ 0 [ 0 , 1 ) , then AFBSDDE (20) admits a unique solution ( λ t ρ 0 + τ , θ t ρ 0 + τ ) for sufficiently small δ and ρ 0 + τ 1 with τ [ 0 , τ 0 ] .
Proof 
(Proof.). We will use the notation
Λ = ( Y , Z , X ) , Θ = ( P , Q , Ξ ) .
Since AFBSDDE (20) admits a unique solution when ρ = ρ 0 [ 0 , 1 ) for each ϑ i ( · ) , κ i ( · ) , ϖ i ( · ) L F 2 ( 0 , T ; R n ) , ψ L 2 ( F T ; R n ) , ϕ i L 2 ( F 0 ; R n ) ( i = 1 , 2 ) , there thus exists a unique ( Λ , Θ ) L 2 ( δ , T + δ ) satisfying the following AFBSDDE for each ( λ , θ ) L 2 ( δ , T + δ ) .
d Y ( t ) = { ρ 0 ( A ( t ) Y ( t ) + B ( t ) Y ( t δ ) + I ( t ) C ( t ) X ( t ) + I ¯ ( t ) D ( t ) Ξ ( t ) + E ( t ) Z ( t ) + F ( t ) Z ( t δ ) ) + τ ( A ( t ) y ( t ) + B ( t ) y ( t δ ) + I ( t ) C ( t ) x ( t ) + I ¯ ( t ) D ( t ) ξ ( t ) + E ( t ) z ( t ) + F ( t ) z ( t δ ) ) + ( 1 ρ 0 ) D ( t ) D ( t ) + C ( t ) C ( t ) Ξ ( t ) + τ ( D ( t ) D ( t ) + C ( t ) C ( t ) ) ξ ( t ) + ϑ 1 ( t ) } d t Z ( t ) d W ( t ) , t [ 0 , T ] , d X ( t ) = { ρ 0 A ( t ) X ( t ) + E F t [ B ( t + δ ) X ( t + δ ) ] M 1 ( t ) Y ( t ) + κ 1 ( t ) + τ ( A ( t ) x ( t ) + E F t [ B ( t + δ ) x ( t + δ ) ] M 1 ( t ) y ( t ) ) } d t + { ρ 0 ( E ( t ) X ( t ) + E F t [ F ( t + δ ) X ( t + δ ) ] R 1 ( t ) Z ( t ) ) + ϖ 1 ( t ) + τ ( E ( t ) x ( t ) + E F t [ F ( t + δ ) x ( t + δ ) ] R 1 ( t ) z ( t ) ) } d W ( t ) , t [ 0 , T ] , d P ( t ) = { ρ 0 ( A ( t ) P ( t ) + B ( t ) P ( t δ ) I ( t ) C ( t ) Ξ ( t ) + E ( t ) Q ( t ) + F ( t ) Q ( t δ ) ) + ϑ 2 ( t ) + τ ( A ( t ) p ( t ) + B ( t ) p ( t δ ) I ( t ) C ( t ) ξ ( t ) + E ( t ) q ( t ) + F ( t ) q ( t δ ) ) + ( 1 ρ 0 ) C ( t ) C ( t ) X ( t ) + τ C ( t ) C ( t ) x ( t ) } d t Q ( t ) d W ( t ) , t [ 0 , T ] , d Ξ ( t ) = { ρ 0 ( A ( t ) Ξ ( t ) + E F t [ B ( t + δ ) Ξ ( t + δ ) ] + M 1 ( t ) P ( t ) M 2 ( t ) Y ( t ) ) + κ 2 ( t ) + τ ( A ( t ) ξ ( t ) + E F t [ B ( t + δ ) ξ ( t + δ ) ] + M 1 ( t ) p ( t ) M 2 ( t ) y ( t ) ) } d t + { ρ 0 ( E ( t ) Ξ ( t ) + E F t [ F ( t + δ ) Ξ ( t + δ ) ] + R 1 ( t ) Q ( t ) R 2 ( t ) Z ( t ) ) + ϖ 2 ( t ) + τ ( E ( t ) ξ ( t ) + E F t [ F ( t + δ ) ξ ( t + δ ) ] + R 1 ( t ) q ( t ) R 2 ( t ) z ( t ) ) } d W ( t ) , t [ 0 , T ] , Y ( T ) = γ , Y ( t ) = η ( t ) , Z ( t ) = ζ ( t ) , t [ δ , 0 ) , X ( 0 ) = ρ ( 0 ) G 1 Y ( 0 ) τ G 1 y ( 0 ) + ϕ 1 , X ( t ) = 0 , t ( T , T + δ ] , P ( T ) = ψ , P ( t ) = 0 , Q ( t ) = 0 , t [ δ , 0 ) , Ξ ( 0 ) = ρ 0 ( G 1 P ( 0 ) G 2 Y ( 0 ) ) + τ ( G 1 p ( 0 ) G 2 y ( 0 ) ) + ϕ 2 , Ξ ( t ) = 0 , t ( T , T + δ ] ,
where 0 ρ 0 + τ 1 .
We define the following mapping:
( Λ , Θ ) = I ρ 0 + τ ( λ , θ ) : L 2 ( δ , T + δ ) L 2 ( δ , T + δ ) .
We proceed to prove that the mapping I ρ 0 + τ is a contraction.
Suppose ( λ ¯ , θ ¯ ) L 2 ( δ , T + δ ) , ( Λ ¯ , Θ ¯ ) = I ρ 0 + τ ( λ ¯ , θ ¯ ) , and set
( λ ^ , θ ^ ) = ( λ λ ¯ , θ θ ¯ ) = ( y y ¯ , z z ¯ , x x ¯ , p p ¯ , q q ¯ , ξ ξ ¯ ) , ( Λ ^ , Θ ^ ) = ( Λ Λ ¯ , Θ Θ ¯ ) = ( Y Y ¯ , Z Z ¯ , X X ¯ , P P ¯ , Q Q ¯ , Ξ Ξ ¯ ) .
Applying I t o ^ ’s formula to X ^ ( t ) , P ^ ( t ) + Ξ ^ ( t ) , Y ^ ( t ) , we obtain
ρ 0 E G 2 Y ^ ( 0 ) , Y ^ ( 0 ) + τ E G 1 y ^ ( 0 ) , P ^ ( 0 ) + τ E G 2 y ^ ( 0 ) , Y ^ ( 0 ) τ E G 1 p ^ ( 0 ) , Y ^ ( 0 ) = E 0 T { ( 1 ρ 0 ) C ( t ) X ^ ( t ) , C ( t ) X ^ ( t ) ( 1 ρ 0 ) D ( t ) Ξ ^ ( t ) , D ( t ) Ξ ^ ( t ) ( 1 ρ 0 ) C ( t ) Ξ ^ ( t ) , C ( t ) Ξ ^ ( t ) ρ 0 M 2 ( t ) Y ^ ( t ) , Y ^ ( t ) ρ 0 I ¯ ( t ) D ( t ) Ξ ^ ( t ) , Ξ ^ ( t ) ρ 0 R 2 ( t ) Z ^ ( t ) , Z ^ ( t ) + τ A ( t ) x ^ ( t ) + E F t [ B ( t + δ ) x ^ ( t + δ ) ] M 1 ( t ) y ^ ( t ) , P ^ ( t ) τ A ( t ) p ^ ( t ) + B ( t ) p ^ ( t δ ) I ( t ) C ( t ) ξ ^ ( t ) + E ( t ) q ^ ( t ) + F ( t ) q ^ ( t δ ) , X ^ ( t ) τ C ( t ) C ( t ) x ^ ( t ) , X ^ ( t ) + τ E ( t ) x ^ ( t ) + E F t [ F ( t + δ ) x ^ ( t + δ ) ] R 1 ( t ) z ^ ( t ) , Q ^ ( t ) + τ A ( t ) ξ ^ ( t ) + E F t [ B ( t + δ ) ξ ^ ( t + δ ) ] + M 1 ( t ) p ^ ( t ) M 2 ( t ) y ^ ( t ) , Y ^ ( t ) τ A ( t ) y ^ ( t ) + B ( t ) y ^ ( t δ ) + I ( t ) C ( t ) x ^ ( t ) + I ¯ ( t ) D ( t ) ξ ^ ( t ) + E ( t ) z ^ ( t ) + F ( t ) z ^ ( t δ ) , Ξ ^ ( t ) + τ E ( t ) ξ ^ ( t ) + E F t [ F ( t + δ ) ξ ^ ( t + δ ) ] + R 1 ( t ) q ^ ( t ) R 2 ( t ) z ^ ( t ) , Z ^ ( t ) τ ( D ( t ) D ( t ) + C ( t ) C ( t ) ) ξ ^ ( t ) , Ξ ^ ( t ) } d t .
Since ρ 0 0 , 1 ρ 0 > 0 and M 2 ( · ) , I ¯ ( · ) D ( · ) , R 2 ( · ) are non-negative, we have
E 0 T [ C ( t ) X ^ ( t ) , C ( t ) X ^ ( t ) + D ( t ) Ξ ^ ( t ) , D ( t ) Ξ ^ ( t ) + C ( t ) Ξ ^ ( t ) , C ( t ) Ξ ^ ( t ) ] d t ε C 1 ( Λ ^ , Θ ^ ) 2 + τ 2 C 1 ε ( λ ^ , θ ^ ) 2 , ε > 0 .
Applying I t o ^ ’s formula to | Y ^ ( t ) | 2 , for sufficiently small δ 0 , we have the following estimate by Gronwall’s inequality:
E sup t [ δ , T ] | Y ^ ( t ) | 2 + E δ T | Z ^ ( t ) | 2 d t τ C 1 ( λ ^ , θ ^ ) 2 + C 1 E 0 T + δ [ | C ( t ) X ^ ( t ) | 2 + | D ( t ) Ξ ^ ( t ) | 2 + | C ( t ) Ξ ^ ( t ) | 2 ] d t .
Similarly, applying I t o ^ ’s formula to | X ^ ( t ) | 2 , | P ^ ( t ) | 2 and | Ξ ^ ( t ) | 2 , we can obtain
E sup t [ 0 , T + δ ] | X ^ ( t ) | 2 τ C 1 ( λ ^ , θ ^ ) 2 + C 1 E δ T [ | Y ^ ( t ) | 2 + | Z ^ ( t ) | 2 ] d t ,
E sup t [ δ , T ] | P ^ ( t ) | 2 + E δ T | Q ^ ( t ) | 2 d t τ C 1 ( λ ^ , θ ^ ) 2 + C 1 E 0 T + δ [ | C ( t ) X ^ ( t ) | 2 + | C ( t ) Ξ ^ ( t ) | 2 ] d t ,
E sup t [ 0 , T + δ ] | Ξ ^ ( t ) | 2 τ C 1 ( λ ^ , θ ^ ) 2 + C 1 E δ T [ | Y ^ ( t ) | 2 + | Z ^ ( t ) | 2 + | P ^ ( t ) | 2 + | Q ^ ( t ) | 2 ] d t ,
for sufficiently small δ 0 . Here, C 1 is a positive constant. Combing the above estimates (22)–(26), we conclude that
( Λ ^ , Θ ^ ) 2 C τ ( λ ^ , θ ^ ) 2 ,
where C is a positive constant independent of ρ 0 and τ . Taking τ = 1 2 C , it follows that
( Λ ^ , Θ ^ ) 2 1 2 ( λ ^ , θ ^ ) 2 .
Then, the contract mapping I ρ 0 + τ has a unique fixed point ( λ ρ 0 + τ ( t ) , θ ρ 0 + τ ( t ) ) , which is the unique solution of AFBSDDE (20) for ρ = ρ 0 + τ . The proof is completed. □
Lemma 1 shows that, if (20) can be uniquely solved for ρ = 0 , then there exists a unique solution of Equation (20) for ρ = 1 .
Proof of Theorem 5.
When ρ = 0 , AFBSDDE (20) is decoupled and can be uniquely solved. From Lemma 1, there exists a positive constant τ 0 such that for each τ [ 0 , τ 0 ] , AFBSDDE (20) admits a unique solution for ρ = 0 + τ and sufficiently small δ . By an inductive argument, we can increase the parameter ρ step by step from ρ = 0 to ρ = 1 . Therefore, AFBSDDE (20) has a unique solution for ρ = 1 . Especially when we let ϑ i ( · ) , κ i ( · ) , ϖ i ( · ) , ϕ i , and ψ be zero ( i = 1 , 2 ) in AFBSDDE (20), we conclude that AFBSDDE (19) admits a unique solution and completes the proof. □
The following result is a direct sequence of Theorems 3–5.
Theorem 6.
Under assumption (H5), for any given γ L 2 ( F T , R n ) and sufficiently small time delay δ,
( u * ( t ) , w * ( t ) ) = ( N 1 1 ( t ) C ( t ) x ( t ) , N 2 1 ( t ) D ( t ) ξ ( t ) )
provides a unique backward Stackelberg equilibrium for the problems ( L Q B S G ) f ( L Q B S G ) l , where ( y ( t ) , z ( t ) , x ( t ) , p ( t ) , q ( t ) , ξ ( t ) ) is the unique solution of the AFBSDDE (19).

4. Application to Pension Fund Problem

There are many areas where BSDDEs may arise, such as portfolio management, anticipating contract and variable annuities, and unit-linked products problems, as mentioned in [23]. An investor’s wealth process may depend on its own past value or the past values of the underlying investment portfolio. We can introduce different delay effects into the financial market, which will make the model more practical.
This section is devoted to the study of a pension fund problem arising from financial markets. Suppose that there are two players that continuously make contributions to a pension fund (see [27]). One of the players is usually the supervisor, government, or company (called the leader), who pays a premium proportion w ( · ) as their contribution. The other is the individual investor (called the follower) with a premium proportion u ( · ) as their contribution. Suppose the pension fund could be invested in both a bond and a stock in the market. The price of the bond (risk-free asset) satisfies
d B ( t ) = r ( t ) B ( t ) d t ,
where r ( · ) is the interest rate of the bond. In a securities market with delayed responses, the price of the stock can be described by an SDDE
d S ( t ) = μ ( t ) S ( t δ ) d t + σ S ( t ) d W ( t ) , t [ 0 , T ] , S ( t ) = ν ( t ) , t [ δ , 0 ] ,
where μ ( · ) is the appreciation rate of a return and σ is the volatility coefficient of the stock. We assume that r ( · ) , μ ( · ) , and ν ( · ) are deterministic R -valued bounded functions, and σ > 0 is a constant.
Let π ( · ) be the amount invested in the stock. Then, the dynamic of the pension fund is modeled by
d y ( t ) = [ r ( t ) y ( t ) r ( t ) π ( t ) + μ ( t ) π ( t δ ) + u ( t ) + w ( t ) ] d t + σ π ( t ) d W ( t ) , t [ 0 , T ] , y ( T ) = γ , π ( t ) = 0 , t [ δ , 0 ) ,
where γ is the terminal wealth goal. If we set z ( t ) = σ π ( t ) , the BSDDE (27) can be written as
d y ( t ) = [ r ( t ) y ( t ) + σ 1 r ( t ) z ( t ) σ 1 μ ( t ) z ( t δ ) u ( t ) w ( t ) ] d t z ( t ) d W ( t ) , t [ 0 , T ] , y ( T ) = γ , z ( t ) = 0 , t [ δ , 0 ) .
We assume that the time delay δ 0 is sufficiently small to guarantee that BSDDE (28) admits a unique solution.
Let
U 1 : = { u ( · ) L F 2 ( 0 , T ; R ) | u ( · ) 0 , a . e . a . s . } , U 2 : = { w ( · ) L F 2 ( 0 , T ; R ) | w ( · ) 0 , a . e . a . s . } .
The associated cost functionals of the follower and leader have the form
J 1 ( γ ; u ( · ) , w ( · ) ) = 1 2 E [ 0 T L 1 e β t u 2 ( t ) d t + G 1 y 2 ( 0 ) ] , J 2 ( γ ; u ( · ) , w ( · ) ) = 1 2 E [ 0 T L 2 e β t w 2 ( t ) d t + G 2 y 2 ( 0 ) ] .
where L i and G i ( i = 1 , 2 ) are positive constants, and β > 0 is a discount factor. Under the precondition that the terminal goal γ is attained, it is natural to want to minimize the cost functionals (29).
By Theorem 6, we can give an explicit characterization of the equilibrium strategy for the pension fund problem.
Proposition 1.
The pension fund problem (28) with (29) admits a unique equilibrium strategy ( u * ( · ) , w * ( · ) ) as follows:
u * ( t ) = L 1 1 e β t x ( t ) , w * ( t ) = L 2 1 e β t ξ ( t ) ,
where x ( · ) and ξ ( · ) satisfy
d y ( t ) = [ r ( t ) y ( t ) + L 1 1 e β t x ( t ) + L 2 1 e β t ξ ( t ) + σ 1 r ( t ) z ( t ) σ 1 μ ( t ) z ( t δ ) ] d t z ( t ) d W ( t ) , t [ 0 , T ] , d x ( t ) = r ( t ) x ( t ) d t + { σ 1 r ( t ) x ( t ) σ 1 E F t [ μ ( t + δ ) x ( t + δ ) ] } d W ( t ) , t [ 0 , T ] , d p ( t ) = [ r ( t ) p ( t ) L 1 1 e β t ξ ( t ) + σ 1 r ( t ) q ( t ) σ 1 μ ( t ) q ( t δ ) ] d t q ( t ) d W ( t ) , t [ 0 , T ] , d ξ ( t ) = r ( t ) ξ ( t ) d t + { σ 1 r ( t ) ξ ( t ) σ 1 E F t [ μ ( t + δ ) ξ ( t + δ ) ] } d W t , t [ 0 , T ] , y ( T ) = γ , z ( t ) = 0 , t [ δ , 0 ) , x ( 0 ) = G 1 y ( 0 ) , x ( t ) = 0 , t ( T , T + δ ] , p ( T ) = 0 , p ( t ) = 0 , q ( t ) = 0 , t [ δ , 0 ) , ξ ( 0 ) = G 1 p ( 0 ) G 2 y ( 0 ) , ξ ( t ) = 0 , t ( T , T + δ ] .

5. Concluding Remarks

In this paper, we discuss the stochastic backward Stackelberg games with delay motivated by some interesting economic and financial problems. Since the corresponding Hamilton system is particularly complicated, we focus on the linear systems. The optimal strategies of the follower and the leader are expressed by an AFBSDDE, which is a new kind of double FBSDE. Furthermore, the AFBSDDE is proved to be uniquely solvable, and we then obtain the unique equilibrium of the LQ backward Stackelberg game. Although this paper deals with systems with state delays, our approach is still valid for the systems with control delays. In addition, it is worthwhile to consider the state feedback representation of the equilibrium. Another possible research direction is to investigate the maximum principle for the leader’s optimal control problem under a structure that describes the follower’s control as a function of the leader’s control and the adjoint variable. We will study these challenging topics in the future.

Author Contributions

Methodology, H.X.; Writing—original draft, L.C. and P.Z.; Writing—review & editing, L.C.; Supervision, L.C.; Funding acquisition, H.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Yue Qi Young Scholar Project, China University of Mining and Technology, Beijing, under Grant 00/800015Z11A25; the NSF of China under Grants 61977043 and 12171053; the NSF of Shandong Province under Grants ZR2019ZD42 and ZR2020ZD24; the National Key R&D Program of China under Grant 2022YFA1006103; and the Fundamental Research Funds of the Central Universities under Grant 2023ZKPYL02.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Von Stackellberg, H. Marktform Und Gleichgewicht; Springer: New York, NY, USA, 1934; (An English translation appeared in the theory of the market ecomomy. Oxford University Press: Oxford, UK, 1952). [Google Scholar]
  2. Simaan, M.; Cruz, J.B. On the Stackelberg game strategy in non-zero games. J. Optim. Theory Appl. 1973, 11, 533–555. [Google Scholar] [CrossRef]
  3. Tolwinski, B. Closed-loop Stackelberg solution to a multistage linear-quadratic game. J. Optim. Theory Appl. 1981, 34, 485–501. [Google Scholar] [CrossRef]
  4. Øksendal, B.; Sandal, L.; Uboe, J. Stochastic Stackelberg equilibria with applications to time dependent newsvendor models. J. Econ. Dyn. Control 2013, 37, 1284–1299. [Google Scholar] [CrossRef] [Green Version]
  5. Bensoussan, A.; Chen, S.K.; Sethi, S.P. The maximum principle for global solutions of stochastic Stackelberg differential games. SIAM J. Control Optim. 2015, 5, 1956–1981. [Google Scholar] [CrossRef]
  6. Zhang, L.Q.; Zhang, W. Global solutions of stochastic Stackelberg differential games under convex control constraint. Syst. Control Lett. 2021, 156, 105020. [Google Scholar] [CrossRef]
  7. Shi, J.T.; Wang, G.C.; Xiong, J. Leader-follower stochastic differential game with asymmetric information and applications. Automatica 2016, 63, 60–73. [Google Scholar] [CrossRef] [Green Version]
  8. Shi, J.T.; Wang, G.C.; Xiong, J. Linear-quadratic stochastic Stackelberg differential game with asymmetric information. Sci. China Inform. Sci. 2017, 60, 092202. [Google Scholar] [CrossRef]
  9. Yong, J.M. A leader-follower stochastic linear quadratic differential game. SIAM J. Control Optim. 2002, 42, 1015–1041. [Google Scholar] [CrossRef]
  10. Wang, G.C.; Zhang, S.S. A mean-field linear-quadratic stochastic Stackelberg differential game with one leader and two followers. J. Syst. Sci. Complex. 2020, 33, 1383–1401. [Google Scholar] [CrossRef]
  11. Li, N.; Xiong, J.; Yu, Z.Y. Linear-quadratic generalized Stackelberg games with jump-diffusion processes and related forward-backward stochastic differential equations. Sci. China Math. 2021, 64, 2091–2116. [Google Scholar] [CrossRef]
  12. Chen, L.; Wu, Z. Maximum principle for the stochastic optimal control problem with delay and application. Automatica 2010, 48, 1074–1080. [Google Scholar] [CrossRef]
  13. Zhang, F. Sufficient maximum principle for stochastic optimal control problems with general delays. J. Optim. Theory Appl. 2022, 192, 678–701. [Google Scholar] [CrossRef]
  14. Johnson, M.; Vijayakumar, V. An investigation on the optimal control for Hilfer fractional neutral stochastic integrodifferential systems with infinite delay. Fractal Fract. 2022, 6, 583. [Google Scholar] [CrossRef]
  15. Ma, Y.K.; Dineshkumar, C.; Vijayakumar, V.; Udhayakumar, R.; Shukla, A.; Nisar, K.S. Approximate controllability of Atangana-Baleanu fractional neutral delay integrodifferential stochastic systems with nonlocal conditions. Ain Shams Eng. J. 2023, 14, 101882. [Google Scholar] [CrossRef]
  16. Johnson, M.; Vijayakumar, V.; Nisar, K.S.; Shukla, A.; Botmart, T.; Ganesh, V. Results on the approximate controllability of Atangana-Baleanu fractional stochastic delay integrodifferential systems. Alex. Eng. J. 2023, 62, 211–222. [Google Scholar] [CrossRef]
  17. Ishida, T.; Shimemura, E. Open-loop Stackelberg strategies in a linear-quadratic differential game with time delay. Int. J. Control 1987, 45, 1847–1855. [Google Scholar] [CrossRef]
  18. Xu, J.J.; Zhang, H.S. Sufficient and necessary open-loop Stackelberg strategy for two-player game with time delay. IEEE Trans. Cybern. 2016, 46, 438–449. [Google Scholar] [CrossRef] [PubMed]
  19. Bensoussan, A.; Chau, M.H.M.; Lai, Y.; Yan, S.C.P. Linear-quadratic mean field Stackelberg games with state and control delays. SIAM J. Control Optim. 2017, 4, 2748–2781. [Google Scholar] [CrossRef]
  20. Xu, J.J.; Shi, J.T.; Zhang, H.S. A leader-follower stochastic linear quadratic differential game with time delay. Sci. China Inf. Sci. 2018, 61, 86–98. [Google Scholar] [CrossRef]
  21. Meng, W.J.; Shi, J.T. A linear quadratic stochastic Stackelberg differential game with time delay. Math. Control Relat. Fields 2022, 12, 581–609. [Google Scholar] [CrossRef]
  22. Wang, G.C.; Yu, Z.Y. A Pontryagin’s maximum principle for non-zero sum differential games of BSDEs with applications. IEEE Trans. Automat. Control 2010, 55, 1742–1747. [Google Scholar] [CrossRef]
  23. Delong, L. Applications of time–delayed backward stochastic differential equations to pricing, hedging and portfolio management in insurance and finance. Appl. Math. 2012, 39, 463–488. [Google Scholar] [CrossRef]
  24. Chen, L.; Huang, J.H. Stochastic maximum principle for controlled backward delayed system via advanced stochastic differential equation. J. Optim. Theory Appl. 2015, 167, 1112–1135. [Google Scholar] [CrossRef]
  25. Shi, J.T.; Wang, G.C. A nonzero sum differential game of BSDE with time-delayed generator and application. IEEE Trans. Automat. Control 2016, 61, 1959–1964. [Google Scholar] [CrossRef]
  26. Meng, W.J.; Shi, J.T. Linear quadratic optimal control problems of delayed backward stochastic differential equations. Appl. Math. Optim. 2021, 84, 523–529. [Google Scholar] [CrossRef]
  27. Zheng, Y.Y.; Shi, J.T. A Stackelberg game of backward stochastic differential equations with applications. Dyn. Games Appl. 2020, 10, 968–992. [Google Scholar] [CrossRef] [Green Version]
  28. Chen, L.; Wu, Z. A type of general forward-backward stochastic differential equations and applications. Chin. Ann. Math. 2011, 32, 279–292. [Google Scholar] [CrossRef]
  29. Huang, J.H.; Shi, J.T. Maximum principle for optimal control of fully coupled forward-backward stochastic differential delayed equations. ESAIM Control Optim. Calc. Var. 2012, 18, 1073–1096. [Google Scholar] [CrossRef] [Green Version]
  30. Li, N.; Wang, G.C.; Wu, Z. Linear-quadratic optimal control problem of forward-backward stochastic system with delay. In Proceedings of the 36th Chinese Control Conference, Dalian, China, 26–28 July 2017; pp. 198–203. [Google Scholar]
  31. Li, N.; Wang, Y.; Wu, Z. An indefinite stochastic linear quadratic optimal control problem with delay and related forward-backward stochastic differential equations. J. Optim. Theory Appl. 2018, 179, 722–744. [Google Scholar] [CrossRef]
  32. Peng, S.; Wu, Z. Fully coupled forward-backward stochastic differential equations and applications to optimal control. SIAM J. Control Optim. 1999, 37, 825–843. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, L.; Zhou, P.; Xiao, H. Backward Stackelberg Games with Delay and Related Forward–Backward Stochastic Differential Equations. Mathematics 2023, 11, 2898. https://doi.org/10.3390/math11132898

AMA Style

Chen L, Zhou P, Xiao H. Backward Stackelberg Games with Delay and Related Forward–Backward Stochastic Differential Equations. Mathematics. 2023; 11(13):2898. https://doi.org/10.3390/math11132898

Chicago/Turabian Style

Chen, Li, Peipei Zhou, and Hua Xiao. 2023. "Backward Stackelberg Games with Delay and Related Forward–Backward Stochastic Differential Equations" Mathematics 11, no. 13: 2898. https://doi.org/10.3390/math11132898

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop