1. Introduction
A differential game is concerned with the problem that multiple players make decisions, according to their own advantages and 
trade-off with other peers, in the context of dynamic systems. More precisely, suppose that there are 
N players acting on a dynamic system through their decisions 
, and each of them has an individual cost functional 
, 
. The goal is to find a Nash equilibrium 
 for the 
N players, which satisfies 
, for 
, where 
 is an arbitrary decision for Player 
i. The property of 
 means that when the 
N players act with 
, it is advantageous for them to maintain their decisions throughout. If one of them acts unilaterally to change a decision, it is penalized if the others maintain their initial decisions. The problem described above is called a 
nonzero-sum differential game. In this paper, we assume 
 and focus our attention on the two-player case. In particular, if the dynamic system is given by a stochastic differential equation, then the game is called a nonzero-sum stochastic differential game, and there is much literature on this subject; see Hamadène et al. [
1], Hamadène [
2,
3], Wu [
4], and Sun and Yong [
5]. On the other hand, recently, the theory of optimal control and differential game has been found to be very useful in the field of human–machine interaction systems; see [
6,
7,
8].
Optimal switching is used to determine a sequence of stopping times at which to shift the mode of the control process to another. When there are more than two modes, it needs to decide not only 
when to switch, but also 
where to switch. As an important branch in control theory, optimal switching has been extensively investigated by means of 
variational inequalities (see Yong [
9,
10], Tang and Yong [
11], Pham [
12], and Song et al. [
13]) or 
backward stochastic differential equations (see Hamadène and Jeanblanc [
14], El Asri and Hamadène [
15], Hamadène and Zhang [
16], Hu and Tang [
17], and El Asri [
18,
19]). In addition, the 
zero-sum switching game problems have been discussed by Tang and Hou [
20], Hamadène et al. [
21], and El Asri and Mazid [
22]; it is noted that the 
nonzero-sum switching game problems have not ever been investigated. Apart from the mathematical interest in its own right, optimal switching enjoys a wide range of applications, such as resource extraction (Brekke and Øksendal [
23]), investment decisions (Duckworth and Zervos [
24]), and electricity production (Carmona and Ludkovski [
25]).
In this paper, we consider, for the first time in the literature, a two-player nonzero-sum stochastic differential game with both players using switching controls. The main contribution of this paper is to establish a verification theorem (see Theorem 1) as a sufficient criterion for Nash equilibrium between the two players, in which a set of variational inequalities is given and a regularity condition is imposed to require, taking player 1 for example, the solution 
 to be 
 on the opponent’s continuation region 
, and to be 
 on 
 except for the boundary of its own continuation region 
. It turns out that the Nash equilibrium strategies for the two players can be constructed based on these variational inequalities, and the solutions 
 and 
 coincide with the corresponding value functions (or, Nash equilibrium payoffs in the terminology of Buckdahn et al. [
26]) of the two players.
On the one hand, it is emphasized that, if we just prove the verification theorem, then the regularity condition given is more strict than what we actually need; however, the seemingly superfluous regularity condition is really necessary if we want to further apply the the so-called smooth-fit principle to solve some specific examples, otherwise we cannot find enough pasting equations for the undetermined parameters. On the other hand, we would like to mention that in this paper, the verification theorem is proved in a piecewise or stage-by-stage way along the sequence of total decision times of the two players (here, a stage means a period between two adjacent decision times), which, to the best of our best knowledge, is new in the nonzero-sum switching game literature.
The rest of this paper is organized as follows. 
Section 2 formulates the nonzero-sum game problem under consideration. 
Section 3 states and proves the verification theorem. Finally, 
Section 4 concludes the paper with some further remarks.
  2. Problem Formulation
Let R be the one-dimensional real space. For a subset , , , and , denote the spaces of all real valued continuous, continuously differentiable, and twice continuously differentiable functions on D, respectively. Let  be a probability space on which a one-dimensional standard Brownian motion , , is defined.  denotes the natural filtration of  augmented by all the null sets.
Let the two players in the game be labeled Player 1 and Player 2. In order to formulate the problem precisely, we first provide the definition of admissible switching controls for the two players.
Definition 1. Let  and  be two finite sets of possible modes for Player 1 and Player 2, respectively. An admissible switching control for Player 1 is a sequence of pairs , where  is a sequence of stopping times with  and  as , representing the decisions on “when to switch," and  is a sequence of -valued random variables with each  being -measurable, representing the decisions on “where to switch." The collection of all admissible switching controls for Player 1 is denoted by .
An admissible switching control  for Player 2 is defined similarly. The collection of all admissible switching controls for Player 2 is denoted as .
 Remark 1. Here, an example of electricity production management is provided to illustrate the meaning of “possible modes for players” in the Definition 1. Typically, a power plant has multiple modes, such as operating at full capacity, operating at partial capacity, or even shutting down all generators. Suppose that there are two managers running the power plant by switching the mode of the power plant from one to another, and the fixed costs are associated with these switchings. Both managers make decisions to maximize their payoffs and eventually reach a Nash equilibrium. In this example, the multiple modes of the power plant that can be chosen by the two managers are the so-called “possible modes for players”.
 The state process 
 is described by the following:
      where 
 are two given functions satisfying the usual Lipschitz condition, so that (
1) admits a unique strong solution. Note that the functions 
b and 
 and the functions 
f, 
, 
, 
, 
 appearing in the following payoff functionals are deterministic, as, in this paper, we adopt the verification theorem approach associated with a set of ordinary differential equations in the form of variational inequalities to deal with the game problem under consideration.
The payoff functionals for Player 1 and Player 2 to maximize are given, respectively, by
      
      and
      
      where 
 is a given function, 
 and 
 are the switching costs for Player 1 and Player 2, respectively, 
 and 
 are the corresponding gains for Player 1 and Player 2 due to the opposites’ actions, respectively, and 
 is the discount factor. It is emphasized that there are no specific conditions on the functions 
f, 
, 
, 
, 
; all of the conditions we need are listed in the verification theorem.
The objective is to find a Nash equilibrium 
, i.e.,
      
      If such a Nash equilibrium exists, then we denote
      
      as the corresponding value function. Note that 
 are 
not uniquely defined, but depend on the Nash equilibrium under consideration.
Remark 2. In this paper, we assume the Brownian motion and the state process to be one-dimensional just for simplicity of presentation. There is no essential difficulty to generalize the results to the multi-dimensional case, but with more complex notation.
   3. Verification Theorem
In this section, we establish a verification theorem as a sufficient criterion that can be used to obtain a Nash equilibrium.
Remark 3. The definitions of  and  have an immediate explanation: If Player 1 (respectively, Player 2) makes a switching from i to k (respectively, j to l), then the present Nash equilibrium payoff can be written as  (respectively, ); we have considered the payoff in the present mode of the switching control and the switching cost. The maximum point of  (respectively, ) is actually the best new mode that Player 1 (respectively, Player 2) would choose in case it wants to switch; otherwise, it would be in its interest to deviate by the definition of the Nash equilibrium.
Similarly,  (respectively, ) represents the payoff for Player 1 (respectively, Player 2) when Player 2 (respectively, Player 1) takes the best switching action and behaves optimally afterward.
 Let
      
      and
      
      In fact, 
 (respectively, 
) is the so-called continuation (or, no-switching) region for Player 1 (respectively, Player 2), in which it is better for Player 1 (respectively, Player 2) to do nothing than make a switching; see 
Figure 1 for a graphical representation of 
 and 
. The boundaries of 
 and 
 are denoted by 
 and 
, respectively.
Denote
      
      where 
 and 
 are the first-order and second-order derivatives of an arbitrary twice differentiable function 
, respectively.
Now, we state and prove the verification theorem for the nonzero-sum stochastic differential game with switching controls.
Theorem 1. Let  and  be real-valued functions, such that:
(ii) 
For  and ,and(iii) 
For  and ,(iv) 
For  and ,(v) 
For  and ,(vi) 
For  and ,Define  and  inductively asandThen, we have that  is a Nash equilibrium for the two players and  are the corresponding value functions of the game. Remark 4. It should be noticed that along the boundaries of  (respectively, ), the function  (respectively, ), , only belongs to , but does not necessarily belong to  on  (respectively, ). In this situation, we can apply the smooth approximation argument introduced by Øksendal [27] (Theorem 10.4.1 and Appendix D) to complement the smoothness needed for Itô’s formula. Here, in this proof, for convenience, we simply consider  (respectively, ), , to be  on  (respectively, ); in this connection, see also Guo and Zhang [28] (Theorem 3.1) and [29] (Theorem 2) and Aïd et al. [30] (Theorem 1).  Proof.  We only show the part for Player 1, and the counterpart for Player 2 is symmetric. We first prove that
        
        where 
 is an arbitrary switching control for Player 1. Denote 
 (with 
) as the sequence of total switching times in the game: at each 
, we have either 
 for some 
m or 
 for some 
n. Then, based on 
, the payoff functional for Player 1 can be rewritten as
        
        Applying Itô’s formula to 
 between 
 and 
, 
, we have
        
        where the inequality follows from condition (iii), as 
 when 
 and the last equality is due to the fact that no switching occurs between 
 and 
, thus, we have 
 and 
.
Summing the indices 
, we have
        
        Combining (
8) and (
9) yields
        
        In the following, the analysis is divided into three cases:
(a) If 
 for some 
m, then
        
(b) If 
 for some 
n, then
        
(c) If 
 simultaneously for some 
m and 
n, then
        
        From (
10)–(
13), we have
        
        On the other hand, the proof of
        
        is the same as above, but with all inequalities becoming equalities.    □
 Remark 5. Here, we provide some comments on the conditions (i)–(vi) given in the verification theorem. First, the condition (i) is the regularity requirement on the solutions of the variational inequalities. It is important for us, in specific cases, to solve the variational inequalities and obtain some analytical solutions by using the so-called smooth-fit principle. The condition (ii) is a typical assumption in optimal switching control theory, which comes from the dynamic programming principle. Regarding the condition (iii), if Player 2 does not make a switching (i.e., ), then the problem for Player 1 becomes a classical one-player optimal switching control problem, so we have (2). On the contrary, if Player 2 makes a switching (i.e., ), by the definition of Nash equilibrium, we expect that Player 1 does not lose anything. This is equivalent to (4) in condition (v); otherwise, it would be in its interest to deviate. Finally, the conditions (iv) and (vi) on  are imposed for the same reason.  Remark 6. 
            The proof for the case with three or more than three players can be given in a similar way to that of Theorem 1. Note that in the proof of Theorem 1, we show the property of the Nash equilibrium, that  and  separately and independently. It can be naturally generalized to the case with three or more than three players, and one needs only to modify accordingly the conditions (i)–(vi) imposed on  and  to  for  in Theorem 1.