1. Introduction
Game theory, pioneered in [
1], has many suitable applications with many different frameworks and goals. For example, [
2] developed a suitable type of strategy choice for players of non-cooperative games that came to be known as a Nash equilibrium, whereby no player can unilaterally deviate from the strategy to achieve a strictly better result for themself. Nash equilibria are known to exist under certain conditions: one such early example was from [
3], where compact Hausdorff pure strategy spaces and continuous cost functions for a game imply the existence of a Nash equilibrium by the extension of the Kakutani fixed-point theorem. A natural extension for cooperative games was made in [
4], where the strategy profile is one for which no coalition of players can collectively change their strategies to achieve strictly better results for all members of the coalition. This is known as a strong Nash equilibrium (SNE), and while there are no general results on the existence of an SNE, some work characterizing them for specific enviroments has been carried out. Recently, in [
5], some specific game properties enabled the existence of an SNE. which provided some good results. Additionally, algorithms have been created for finding an SNE, as can be found in sources like [
6,
7]. These algorithms in [
6,
7] are considerably different to the one found in this paper. A few places where Nash equilibria are explored in relation to real-world problems by utilizing variational analysis are [
8,
9]. Additionally, another work for characterizing Nash equilibria using the variational inequality was carried out in [
10].
The structure of this work proceeds as follows: First, a section on some necessary background information is given. Strong Nash equilibria are shown to be connected to efficiency in a particular case. Then, using the variational inequality, as defined in [
10], it is shown that a strategy profile is an SNE when, for every coalition, a form of the variational inequality holds. A simple characterization of an SNE is given for two-player games. Finally, an algorithm is given that produces a sequence of strategies that, when convergent, converge to a strong Nash equilibrium. A simple example of the algorithm in use is then given. Tests of the algorithm are also performed and a summary is provided. A portion of the results were part of the PhD thesis of the first author while studying at Northern Illinois University [
11].
  2. Preliminaries and Notations
This work uses only real Euclidean vector spaces. The notations and definitions that follow in this section will be used throughout the paper:
Notation 1.  Given the general vectors  and ,  refers to the vector ,  implies  for all i values,  implies  for all i values, and  implies  for all i values.
 Notation 2.  For a function  with input vector  of length n and some set , the gradient  is understood to be the gradient of f restricted to the  component of x (more than one component may be listed) and  is the gradient restricted to the component  where . This rules is used when the gradient of f is to be investigated but certain entries of x are held fixed.
 The structure of a problem and the solutions to that problem are as described by the following definition:
Definition 1.  Let  be a vector-valued function and . A problem is described by This is a multi-objective optimization problem (MOOP). A feasible solution for (P) is any point . An efficient solution (sometimes called the Pareto efficient) for (P) is a feasible solution x with the property stating that there is no other feasible solution  such that  and . Identifying all the efficient points of a problem (P) is what is meant by “solving (P)”. Lastly, a feasible solution x is said to be a weakly efficient solution to (P) if there is no other  such that .
 Efficiency first made its début in [
12] and then was modernized in [
13]. Further formalization that resulted in the form described here was provided in [
14]. A good reference on variational analysis is [
15].
With regard to game theory, a complete explanation of how a player will react or behave in a game given all the potential possible situations that the player may encounter is called a 
pure strategy. A 
mixed strategy is a vector from the simplex space of a dimension equal to the number of pure strategies available to a player, and that vector determines the probability that any of the potential pure strategies will be employed in a game for that player [
16]. Mixed strategies are continuous in nature, and so they are the type of strategies under consideration in this work.
Definition 2.  A game (G) is described by the triplet (G) with  representing the n different players, while the space  is referred to as the strategy space where each  is the closed convex simplex , which describes the  player’s collection of mixed strategies to choose from based on the  potential pure strategies available to the player. Here,  is called the  player’s strategy and  is known as a strategy profile for the game. The function  is a multi-objective function where every  player is assigned  as a cost function that they wish to minimize. These types of games are multi-objective games in the sense that all the players have their own individual objective, not in the sense that each player has multiple objectives that they are tying to minimize for themselves simultaneously.
For the index set , which will be thought of as a coalition of players, the function  is the same as  with input x but with the specific values  changed to  for , where . A coalition can comprise any combination of players (including the coalition of all players), and all the members of the coalition will be working together to reduce the cost for all coalition members. For this reason, within a coalition, the multi-objective nature of the problems becomes a group of players working together to minimize all the objectives simultaneously. For clarity, replacement cases where the vectors are denoted with a subscript may have the subscript migrated to a superscript, such as . Also, when only the  term is to be replaced, we write .
A Nash equilibrium (NE) is a point , where for all i, there is no  for which . That is to say, a unilateral strategy change by a single player will not result in a better outcome for that player. A strong Nash equilibrium (SNE) is a point  where, given any coalition , there is no other strategy profile  for which  for all . So, an SNE is a strategy profile where no coalition can make a group change to their strategies, which would result in a strictly better result for everyone in the coalition, including the coalition of all players.
 It is a well-known fact that every SNE is automatically an NE and weakly efficient (consider the coalition of all players).
  3. New Results for Strong Nash Equilibria
To begin with, a connection can be made between strong Nash equilibria and the concept of efficiency in a specific case where every player shares a function component underlying their specific composite cost function.
Proposition 1.  If  is an SNE for the game (G) with  for all i and the functions  are linear and non-constant (i.e., ) and , then it is also efficient for the MOOP of minimizing .
 Proof.  If x is not efficient, then there is a  with at least one i where  along with  for all other . Since , then  and . Since all  are linear and non-constant, for any given j, . So, for all j, . Thus, x is not weakly efficient and, thus, it cannot be an SNE. So, being an SNE for the game (G) implies efficiency for the given MOOP by way of contraposition.    □
 Some definitions for characterizing strategy profiles that are Nash equilibria are the concepts of a 
normal cone and the 
variational inequality, as described in [
10]. A vector 
v is a normal vector to 
X at 
x if and only if 
 for all 
. This is written as 
, so 
 is a set-valued mapping described by 
 and 
. In general, for a map 
f, the variational inequality for 
f and 
X is
In [
10], it was determined that a strategy profile 
x is an NE for a game with convex cost functions if and only if 
 for all 
i. This concept can be extended to an SNE.
Theorem 1.  Let (G) be a game with strategy space X and cost functions . If  is convex for all  and  has the property that for all  there is a  such thatthen  is an SNE.  Proof.  If we assume 
 is not an SNE, then there is a coalition 
 and strategy profile 
 such that 
 for all 
. But since 
 is convex for all 
, then
This means that  by definition. Therefore,  must be an SNE.    □
 Example 1.  Letbe the cost functions for a game (G) with strategy space . Now, instead of an in-depth consideration of a player’s strategy, coalitions, and oppositions, we can simply observe all the coalitions’ relevant gradients and reach a conclusion about what points must be an SNE. Observe the restricted gradients    |  | 
|  | 
|  | 
|  | 
|  | 
|  | 
|  | 
|  | 
|  | 
|  | 
|  | 
|  | 
    
From this we can work out what an SNE might be by going through all the coalitions S and making sure that  for at least some . After looking at the single-player coalitions , 2, or 3, it can be seen that  is a suitable vector for which . Indeed, when  for all ,so . When , since , then it must be true that . Lastly, similar to the  case, for , When more players are involved, we only need to find one such gradient per coalition in the normal cone. So, for , sincethen . For , notice thatand  for all , so  is in the normal cone. For , , so again it will be in the normal cone. Lastly, for , notice that , so it will also be in the normal cone.
Since, at some point, every coalition has some player’s cost function’s negative gradient in the normal cone of the coalition, Theorem 1 says that that point is an SNE. So,  is an SNE.
 The converse of Theorem 1 is not necessarily true. If  is an SNE, then  does not need to be in  for every . This can be checked in as restrictive a case as a two-player game with cost functions  and  and a strategy space that is the box .
Another proposition is provided that generalizes Theorem 1 in the context of two-player games. In fact, it gives a characterization of an SNE, and an example is also provided for describing such an SNE.
Proposition 2.  Let (G) be a game with the convex strategy space  and cost functions , where each  is convex and differentiable. Then, the following statement holds:
A pair  is an SNE if and only if  such thatwhere , and  Proof.  As each 
 is convex, 
 is an SNE if and only if it is an NE and a weakly efficient solution for 
. But 
 being an NE and a weakly efficient solution is equivalent to having (
2) and (
1) hold, respectively.    □
 Example 2.  Let , , and . For , it is easy to see that (2) holds, and  with , i.e., (1) holds. Based on Proposition 2,  is an SNE. But  does not satisfy the condition in Theorem 1.    4. Algorithm for Finding SNE
The algorithm in this work is of the same nature as other projected gradient methods like the ones found in [
17,
18] for multi-objective optimization. This algorithm differs by using the arg min function over all coalitions and those coalitions’ potential choices of strategies. So, the direction of descent that the algorithm seeks at any step is one heading towards the weakly Pareto efficient solution of a problem restricted to cost functions from a coalition (see Algorithm 1).
      
| Algorithm 1 Locating a strong Nash equilibrium for convex games | 
| Require: Strategy space  with all  convex, closed, and bounded, cost function F with convex components, initial guess  , 1:while  is not an SNE do2:(If there exists more than one argmin, choose the  that corresponds to the lexicographical minimum of the first component, the index sets .)3:(choose  to be the smallest index if  contains more than one element.)4:     such that  and  for all 5:    6:    7:end while8:if  is an SNE then9:    Set 10:else[a sequence  was created]11:    if  converges then12:        Let  be such that 13:    else[ does not converge]14:        return “Algorithm Failed”15:    end if16:end if17:return 
 | 
When F is a one-dimensional function, the algorithm works similar to a gradient descent method with a direct calculation of the step size. The additional  term is included to ensure that the minimum exists by making the argument strongly convex.
The first thing to be proven is that such step sizes exist.
Proposition 3.  The step size  calculated in Algorithm 1 exists for all .
 Proof.  Since 
 is a candidate for (
3) from Algorithm 1, the minimum will have to be less than or equal to 0. If 
, then any 
, for example, 
, would be sufficient as a value that minimizes the function.
Now, for 
, if
        then consider 
 in place of 
d,
Now, because 
 is a minimizer, it must be true that
        and so 
 for all 
. So, 
 acts as a direction of descent heading away from 
 for 
 for all 
.
Observing the constraints of the minimization problem defining , since X is convex, closed, and bounded, there is a maximum  such that . Going further, because all the functions  are convex and  is a direction of descent away from  for all  with , there is some interval  in which  for all  and .
Now, because  is convex, it will be continuous on any line segment in its domain. And because  for  defines a line segment, a minimum can be obtained for  on that segment using the extreme-value theorem. Note that  cannot be 0 because  for at least one  in .       □
 When  is an SNE, the algorithm will calculate  to be 0 over and over again, so it should be noted that the check statement at the beginning of the “while” loop may not be necessary.
In order to show that the convergence of the algorithm does actually imply that an SNE has been located, more lemmas will be required. The next lemma shows that there is an implication in convergence between the directions of descent and the indices giving the least descent from the coalitions that provide that direction.
Lemma 1.  If  converges to d and  converges to , then , as described in (4), converges.  Proof.  Since
        for all 
j, then
        where 
S is the collection of indices of non-zero components of 
d. This is because if all the limits exist, then the limit of the maximums is the maximum of the limits. Here, the minimum is only being used since in the calculation of 
, if there is more than one element given in the 
 from (
4), the smallest index of all the 
 that give the maximum is chosen.    □
 For the next lemma, it will be shown that if the directions converge to 0, then if the algorithm is convergent, it actually finds a strong Nash equilibrium.
Lemma 2.  Assuming that  and  converges from Algorithm 1, then the algorithm converges to a strong Nash equilibrium.
 Proof.  Instead consider when 
 but that 
 is not an SNE. Since every component 
 of 
F is convex from the requirements of Algorithm (1), there are an 
 and 
 such that for each 
Convexity is necessary because without that condition, it could end up in a situation where 
. But this is the same as
        for all 
 where 
. Now, take a small enough 
 so that
        for all 
 and pass 
 into 
v.
Take a 
 so that
        for all 
, and then take 
 and a large enough 
 so that 
 for all 
. In that case,
        for all 
 and 
. But since 
 is a minimizer,
But because  is fixed, . So, if  and , then it must be true that  is a strong Nash equilibrium.    □
 Moving on, we next discuss another lemma that will be helpful, as it shows that under certain conditions  cannot converge to a non-zero vector.
Lemma 3.  If all the cost functions  are continuously differentiable and  and  converges to  from Algorithm 1, then either  converges to 0 or it does not converge at all.
 Proof.  Assume 
. Note that
Indeed, otherwise, the  statement that defines  would give a positive value as the minimum because . But that is not possible because the zero vector is always a possible choice for , and that in turn would return 0 as the minimum value. Also, note that as  is defined as a minimizer for a continuously differentiable function (that is, the parameterized function  with parameter ),  for all k.
Using Lemma 1, let 
. Now, take 
b so that
This must be non-zero or else  would have to converge to 0.
Since 
 converges, it must be Cauchy, and so taking 
, a 
 can be chosen that is large enough so that
So, for all 
, this gives
        where 
 is the angle between 
 and 
. The limit is taken on both sides of the inequality, which yields
        which is a contradiction as a non-zero 
b cannot be less than or equal to half of itself. So, simply put, 
.    □
 All of these lemmas finally lead to some conditions for which the convergence of the algorithm will lead to an SNE.
Theorem 2.  With regard to Algorithm 1, when X is compact and all  are convex and continuously differentiable, then  converging to  implies that  must be an SNE.
 Proof.  The space X being compact implies it is closed and bounded. Now, because  and , either  or . However, Lemma 2 shows that  implies that  will converge to an SNE. For the case when , Lemma 3 shows that either  or  does not converge. Again, for the first case,  will imply that  converges to an SNE. So, it only remains to consider if  does not converge. But because X is compact,  has a convergent sub-sequence, known as . Since  is convergent in X, it will either converge to 0 or some other point d within the bounds of X.
Case 1: .
Assuming that 
 is not an SNE (otherwise, the theorem would be proven), there must be a coalition 
 and at least one direction 
 for which
The direction 
 can indeed be found because there is some 
d where 
 for 
 for all 
. That 
d can then be scaled down to 
 with 
 and 
, which ensures
We then take that 
 and also 
 as defined to make certain that the inequality on line (
5) holds.
Let 
 and choose a large enough 
 so that for all 
, 
 for all 
. Using this 
 inequality, along with the fact that 
 is defined as a minimizer from line (
3) in Algorithm 1, gives
Since , assuming that the limit as  goes to infinity on this whole inequality reveals the contradiction that , which is impossible. So,  would have to be an SNE.
Case 2: .
This argument’s approach is the same as that defined in Lemma 3 but on the sub-sequence  instead of . Since , it must be true that , where  from Lemma 1 using the sub-sequence  in place of .
The original assumption that 
 converges implies that 
 is Cauchy, so assuming that 
, there will be a large enough 
 so that for all 
,
It must also be noted by the definition of 
 that 
. By using the Cauchy–Schwarz inequality,
Again, a limit is taken on this inequality, which gives
        which is a contradiction as 
, so 
.    □
 So, whenever the algorithm converges, it will be to a strong Nash equilibrium. It can be seen when the algorithm is applied to an example game without an SNE that multiple cluster points will be produced [
11]. To show the algorithm in action, an example is given.
Example 3.  Consider the following problem: Note that the cost functions are convex and there is an SNE at .
Consider Algorithm 1 using  and apply it to the initial guess ; then, All of the gradients have a similarity and so, when generalizing, there are only three types of cases that need to be checked. These cases are similar to when  is , , or . The easiest case to check is when player 1 is acting on their own, which is when . After simplifying, this is carried out by finding  inwhich gives , that is to say, the direction of descent is , which gives the corresponding minimum of . When , then  and  need to be found inwhich ends up giving the direction  and a corresponding minimum of . Lastly, when , , , and  all need to be found fromwhich gives the direction of descent as  and a corresponding minimum of . From these three cases, the direction that actually minimizes the step in the algorithm is  as  is the smallest corresponding minimum, and  is found to be 1, so . The coalition of the first player is the one that the step uses but only because it is lexicographically first compared to the similar single-player coalitions  and  that we generalized.
For brevity, after that, the next direction is  and . Then, the last descent direction is , which gives the SNE .
 While this may seem like a drawn out process, each step is relatively easy to compute for many convex problems using computer programming. However easy it may be to calculate, it can be shown that Algorithm 1 may still become stuck in an endless loop even for convex functions with an SNE present [
11].
  5. Testing
Tests were programmed using Mathematica to determine some metrics on the algorithm performance. The problem types designed for testing were games with convex quadratic cost functions where the quadratic coefficient was generated by taking a random matrix with entries from  to 10 and multiplying it by its transpose to obtain a positive semi-definite matrix, and the linear coefficient was taken as a random vector with entries from  to 10. The strategy space was determined by giving each player a random interval from  to 10 with integer endpoints. The stopping criteria was to run beyond the maximum number of iterations (200 unless noted otherwise), step to a solution that has already been found that will result in a loop, or have the last two consecutive solutions within a convergence tolerance (0.00001 unless otherwise noted).
Table 1 describes the performance of the algorithm in different contexts. Column “
n” gives the number of players possible for the random games, “
” gives the number of random games the algorithm was run on, and “
’s” gives the number of random start points for every random game. The “
” column indicates if it is calculated as described in the algorithm or if it is fixed at 1. “Conv” is the count of convergent runs. “Conv It” is the average number of iterations that the algorithm took to converge. “Loops” gives the number of runs for which the algorithm ended up in a loop, with “Loop It” being the average number of iterations that it took a run to fall into a loop. “Fails” is the number of times the algorithm did not converge or ended up in a loop by the maximum number of iterations.
 As the table shows, whether  is calculated or not, the number of iterations required to find convergence is low when the algorithm converges, often taking only five to ten steps. It should also be noted that the algorithm may still fall into loops even when an SNE strategy is available, and it take more iterations to end up in those loops.
In this test,  was considered for the practical purpose that it significantly reduced the calculation time by removing a minimization algorithm that was built into the Mathematica tool set. The pairs tallied with the same problems and start points were used to further investigate the difference found between  and when  was calculated. The pairs that may cause some confusion are in the algorithm tests with increased convergence tolerance. The increased tolerance caused some issues whereby the program recorded convergences as loops due to the fact that  was calculated and coming out as practically 0 (values like ). The record keeping system that was programmed could not differentiate convergence from loop due to the small values and lumped many convergences in with the loops. This is also evident by the extremely low values in the loop iteration column. After reviewing the data, there were loops but not nearly as often as indicated on the table. So, this data needs to be carefully considered.
This data revealed some good practical techniques in execution. When using this algorithm, it should first be run using a fixed  to initially obtain a quick check for convergence. If convergence failure occurs when , the algorithm could then be run again using the line search method laid out in the original algorithm, which will have longer a computing time and require more iterations, but it can more often succeed in finding an SNE and keep the algorithm from ending up in loops. A few random start positions can be used if convergence initially fails. Care needs to be taken with near-0 calculations. And lastly, the maximum number of iterations can be lowered significantly.
  6. Conclusions
Often times, players in game theoretic problems are allowed to collaborate, and in games where there is not any particular resentment among players, there may be a desire to find outcomes that befit as many players as possible. In such games, a strong Nash equilibrium may be desirable, as this is the sort of solution where no amount of scheming can yield better results for any players. While some readers may be initially dismissive of strong Nash equilibria as being too restrictive, rare, or hard to identify, the results in this paper indicate that they can potentially be identified with relatively low effort and so seeking an SNE should not be overlooked. Considering that an SNE could be a unanimously favored solution for a problem if it exists, investing a small amount of time in seeking one would not be time wasted. What was carried out in this paper was a further exploration of the concept of strong Nash equilibria using modern methods from variational analysis. Also, a new algorithm was developed and tested for aiding in findings SNEs.
First, the concept of strong Nash equilibria was connected to the concept of efficiency under certain circumstances. Next, the concepts of the normal cone and variational inequality were used to find a sufficient condition for showing a solution to be an SNE. A characterization of two-player games using the variational inequality was also provided. The algorithm for finding an SNE was developed using a projected gradient method by calling the  function over all coalitions of players possible. It was shown that when the algorithm converges, it does so to an SNE. Some tests were performed to check the practical efficiency of the algorithm.
Further research can be carried out to explore the convergence rate of this sort of projected gradient algorithm. The relatively quick convergence of the test results, while interesting, is no proof, but it does make one hopeful that the convergence is not too slow. Additionally, the computational complexity of this algorithm and some further test results and analyses of the algorithm could be of interest. It would also be prudent to compare the algorithm results with other algorithms for calculating strong Nash equilibria like those in [
6,
7].
Another direction for advacning this research could be creating an SNE algorithm based off the weighted best reply, as mentioned in [
5]. That algorithm may be as simple as just finding the best reply to 
 and moving there. Then, the fact that a fixed best reply is in fact an SNE for convex problems may make it likely that the algorithm converging results in an SNE. Hopefully, that type of algorithm would always converge for problems that have the coalition consistency principle.