1. Introduction
A linear quadratic stochastic game model of inter-bank borrowing and lending was proposed in (
Carmona et al. 2015). In this model, each individual bank tries to minimize its costs by controlling its rate of borrowing or lending to a central bank with no obligation to pay back its loan. The finding is that, in equilibrium, the central bank acts as a clearing house providing liquidity, and hence stability is enhanced. This model was extended in (
Carmona et al. 2018), where a delay in the controls was introduced. The financial motivation is that banks are responsible for the past borrowing or lending, and need to make a repayment after a fixed time (the delay). In this model, the dynamics of the log-monetary reserves of the banks are described by stochastic delayed differential equations (SDDE). A closed-loop Nash equilibrium is identified by formulating the original SDDE in an infinite dimensional space formed by the state and the past of the control, and by solving the corresponding infinite dimensional Hamilton-Jacobi-Bellman (HJB) equation. For general stochastic equations and control theory in infinite dimension, we refer to (
Bensoussan et al. 2007;
Fabbri et al. 2017;
Da Prato and Zabczyk 2008).
In this paper, we study the mean field game (MFG) corresponding to the model proposed in (
Carmona et al. 2018) as the number of banks goes to infinity. We identify the mean field game system, which is a system of coupled partial differential equations (PDEs). The forward Kolmogorov equation describes the dynamics of the joint law of current state and past control, and the backward HJB equation describes the evolution of the value function. Recently, J.-M. Lasry and P.-L. Lions introduced the concept of “master equation” which contains all the information about the MFG. The well-posedness of this master equation in presence of a common noise and convergence of the
N-player system is analyzed in (
Cardaliaguet et al. 2015) by a PDE approach. A probabilistic approach is proposed in (
Carmona and Delarue 2014;
Chassagneux et al. 2014). See also the two-volume book (
Carmona and Delarue 2018) for a complete account of this approach.
In this paper, the master equation for our delayed mean field game is derived, a solution is given explicitly, and we show that it is the limit of the closed-loop Nash equilibrium of the N-player game system as .
The paper is organized as follows. In
Section 2, we briefly review the stochastic game model with delay presented in (
Carmona et al. 2018). Then, in
Section 3, we construct the corresponding mean field game system. In
Section 4, we define derivatives with respect to probability measures in the space
where
is the Hilbert space defined at the beginning of
Section 2.2. In addition, we derive the master equation, and exhibit an explicit solution. Furthermore, in
Section 5, we show that this solution of the master equation is an approximation of order
to the solution of the finite-player Nash system. Lastly, in
Section 6, we compare the solution of the Nash system, the solution of the mean field game system, and the solution to the master equation.
3. The Mean Field Game System
The mean field game theory describes the structure of a game with infinite many indistinguishable players. All players are rational, i.e., each player tries to minimize their cost against the mass of other players. This assumption implies that the running cost and terminal cost in (
4) only depend on
i-th player’s state
and the empirical distribution of
. Denoting this empirical distribution by
these costs, as in (
4), can be re-written as
As the number
N of players goes to
∞, the joint empirical distribution of the states and past controls
with marginals
converges to a deterministic limit denoted by
(with marginals denoted by
and
). Here, we assume that, at time 0,
satisfies the LLN (for instance with i.i.d.
), and that the propagation of chaos property holds. A full justification of this property would involve generalizing the result in
Section 2.1 of (
Carmona and Delarue 2014) to an infinite dimensional setting in order to take into account the past of the controls. This is highly technical but intuitively sound. A complete proof is beyond the scope of this paper.
In the limit, a single representative player tries to minimize his cost functional, and, dropping the index
i, his value function is defined as
subject to
The HJB equation for the value function
reads
with terminal condition
Then, we minimize in
to get
After plugging it into (
19), our backward HJB equation reads:
Next, since we “lift” the original non-Markovian optimization problem into a infinite dimensional Markovian control problem, we are able to characterize the corresponding generator for (
18), which is denoted by
,
where
is a smooth function and the time dependency comes from
given by (
20). The derivation of the adjoint
of
is given in
Appendix A. Consequently, the forward Kolmogorov equation for the distribution
reads
Combining (
21) with (
23), we obtain the mean field game system. To solve this, We make the following ansatz for the value function
where we denote the mean of state
, and the mean of past control
. Plugging (
24) into (
23), multiplying both sides of (
23) by
, and integrating over
, we have
After integration by parts, we obtain
as can be seen directly using (
24).
Similarly, plugging (
24) to (
23), multiplying both sides of (
23) by
, and integrating over
, we get
By integration by parts, we deduce
Now we are ready to verify the ansatz (
24). We first compute the derivative of the ansatz,
Then, we plug the ansatz (
24) into (
7), and by collecting
terms,
terms,
terms, and constant terms, we obtain the following system of PDEs:
with boundary conditions
As for (
13)–(
14), the system (
30)–(
31) admits a unique solution.
5. Convergence of the Nash System
From the previous section, we have seen that our master equation is well posed, and we obtained an explicit solution. Furthermore, it also describes the limit of Nash equilibria of the
N-player games as
. In this section, generalizing to the case with delay the results of (
Cardaliaguet et al. 2015) (see also
Kolokoltsov et al. 2014), we show that the solution of the Nash system (
11) converges to the solution of the master Equation (
37) as number of players
, with a
Cesaro convergence rate.
In
Section 4, we find that (
40) is a solution to the master Equation (
37). We set
, where
, denotes the joint empirical measure of
and
. The empirical measure of
is given by
, and the empirical measure of
is given by
. Note that, by direct computation, for
, and any
,
Proposition 1. For any , satisfieswhere , with terminal condition . This shows that is “almost” a solution to the Nash system (11). Proof. We compute each term in the above equation in terms of
U using the relationship (
42), and we use the fact that
U is a solution to the master equation.
From the solution (
40) of the master equation,
is Lipschitz with respect to the measures. Namely,
Thus,
Theorem 2. Let be the solution to the HJB Equation (11) of the N-player system, where fixed, and U be the solution to the master Equation (37). Fix any . Then for any , let , we have Proof. We first apply Ito’s formula to
, and use the fact that
satisfies the HJB Equation (
11) for the Nash system.
Then, we apply Ito’s formula to
, and use the fact that
u satisfies (
43)
Substracting (
46) from (
47), taking the square and applying Ito’s formula again, we obtain
Recall that
is bounded by
for
, and
is bounded by
. Let
be a family of independent random variable with common law
. By integrating (
48) from
t to
T, and taking expectation conditional on Ξ, we have
By the fact that
, and using Young’s inequality, we have
Taking average on both sides, we have
By Gronwall’s inequality and taking supremum over
, we have
which implies
Choosing Ξ uniformly distributed in
, then by continuity of
and
, and the fact that
is defined by
, we have, for any
,
□
6. Conclusions
The mean field game system acts as a characteristic of the master equation. The master equation contains all the information in the mean field game system, and it turns the forward-backward PDE into a single equation. The solution to the mean field game system is a pair , that is the value function and the joint law of current state and past law. The solution to the master equation is a function of .
Since our model is linear quadratic, we are able to solve both the mean field game system and the master equation as shown in
Section 3 and
Section 4, however, the techniques are not the same. The technique for solving the mean field game is that we first make an ansatz for the solution of the HJB equation. Then plugging this ansatz into the Fokker-Planck Equation (
23), we find that the means of state and past control are constant. Hence, the ansatz (
24) can be verified. On the other hand, a notion of derivative with respect to measure is needed in order to solve the master equation. Again, we make an ansatz (
40), which has a similar form as (
24) but is a function of
, and we verify that it satisfies the master equation.
The sets of PDEs (
30) with boundary conditions (
31) are the same for the two problems. This is due to the fact that our model is linear-quadratic and the means of states and past controls are constants.
Last but not the least, the Nash equilibrium of the corresponding N-player game is presented in
Section 2. The value function (
12) looks similar to the value function (
24) in the mean field game system and the solution (
40) to the master equation. As
, the set of PDEs (
13) becomes the same as (
30). This implies that the solution to the mean filed game appears to be the limit of the Nash system, but generally, the convergence has been known in very few specific situations. Additionally, the solution to the master equation is also a limit to the Nash system, as shown in
Section 5.
To summarize, we have extended the notion of master equation in the context of our toy model with delay, and we have shown that, as in the case without delay, this master equation provides an approximation to the corresponding finite-player game with delay. A general form of such a result, not necessarily for linear-quadratic games, is part of our ongoing research.