An Intelligent Algorithm for Solving the Efﬁcient Nash Equilibrium of a Single-Leader Multi-Follower Game

: This aim of this paper is to provide the immune particle swarm optimization (IPSO) algorithm for solving the single-leader–multi-follower game (SLMFG). Through cooperating with the particle swarm optimization (PSO) algorithm and an immune memory mechanism, the IPSO algorithm is designed. Furthermore, we deﬁne the efﬁcient Nash equilibrium from the perspective of mathematical economics, which maximizes social welfare and further reﬁnes the number of Nash equilibria. In the end, numerical experiments show that the IPSO algorithm has fast convergence speed


Introduction
In 1950, Nash equilibrium was formulated based on noncooperative games formed among all players, and the existence of an equilibrium point was proven [1,2]. In economics, most noncooperative game theory has focused on equilibrium in games, especially Nash equilibrium and its refinements [3]. Nash equilibrium means that every player cannot obtain additional advantage by adjusting his/her present strategy individually. The Nash equilibrium has played significant roles in many disciplines: psychology, economics, engineering management, computer sciences [4], reinsurance bargaining [5], etc. The Nash equilibrium may not be unique but multiple. Thus, the player is confused when making decisions. For the refinement of the Nash equilibrium, efficiency is introduced by using the efficient mechanism design of mathematical economics [6]. This paper proposes the efficient Nash equilibrium, which can be beneficial to all players in that social welfare is maximized and the number of Nash equilibria is greatly reduced. Hence, the study of the efficient Nash equilibrium has certain practical significance.
A single-leader-multi-follower game (SLMFG) is a special form of the leader-follower game, also called the bilevel programming problem. Yu [7] introduced the Nash equilibrium point existence theorem for the SLMFG and multi-leader-follower game. Jia et al. [8] established the existence theorem for the weakly Pareto Nash equilibrium of the SLMFG. Furthermore, SLMFGs are widely used in resource coordination [9], energy scheduling [10], cellular data traffic and 5G networks [11], hexarotors with tilted propellers [12], etc. The SLMFG contains one leader and multiple followers. The leader is capable of dominating and expecting the responses of the followers, and the leader selects the best strategy from his own feasible strategy space by knowing the responses of the followers. The followers make their optimal responses according to the leader's given strategy. Currently, many problems can be treated as leader-follower problems in reality, such as those between suppliers and retailers [13], between groups and subsidiaries, between central and local governments [14], and between defenders and multiple attackers [15].
Furthermore, a SLMFG is regarded as a bilevel programming problem with a leaderfollower hierarchical structure [16]. Currently, the study of linear bilevel programming is relatively mature, but the studies on nonlinear bilevel programming are lacking. Nonlinear bilevel programming is a NP-hard problem [17,18]. Fortunately, with the development of biological evolution and heuristic algorithms, swarm intelligent algorithms have displayed the potential for possibly solving nonlinear bilevel programming problems. Many scholars have tried to solve the Nash equilibrium of the SLMFG by using swarm intelligence algorithms, including a dynamic particle swarm optimization algorithm [19], genetic algorithms [20,21], and a nested evolutionary algorithm [22]. We consider swarm intelligence algorithms for solving the SLMFG, which has an evident theoretic foundation and applied significance.
The paper is organized as follows. In the next section, we present the model of the single-leader-multi-follower game, the efficient Nash equilibrium of the SLMFG, and some assumptions of the SLMFG. In Section 3, we consider that the SLMFG is turned into a nonlinear equation problem (NEP) by using the Karush-Kuhn-Tucker (KKT) condition and complementarity function methods. In Section 4, the IPSO algorithm consists of introducing the antibody concentration inhibition mechanism and immune memory function into the particle swarm optimization (PSO) algorithm. In Section 5, we solve some numerical experiments by utilizing the IPSO algorithm. Additionally, the IPSO algorithm has the advantages of few parameters, easy implementation, and random generation of the initial point. Furthermore, the IPSO algorithm has a fast convergence speed, as shown by observing its off-line performance. Finally, several numerical experiments showed that the IPSO algorithm is practicable: the efficient Nash equilibrium is solved and the number of Nash equilibria is greatly reduced.

Preliminaries and Prerequisites
In this section, we present the model of the SLMFG, the efficient Nash equilibrium, and some assumptions of the SLMFG. Assume that I = {1, 2, . . . , n} is a set of followers and y i (i ∈ I) is the control vector of the ith follower. The ith follower's feasible strategy set is The leader's feasible strategy set is X, and x ∈ X is the control vector of the leader. The objective function of the leader is ϕ : X × Y → R, and the followers' objective functions are f i : X × Y → R. Furthermore, the followers' best response feasible strategies regarding the leader's strategy parameter x are defined by the set-value mapping K : X → 2 Y as follows: Assume that a strategy profile for the followers is y * = (y * 1 , y * 2 , . . . , y * n ) ∈ Y; for any i ∈ I, the following equation is satisfied: (2) In that case y * is called the Nash equilibrium of the followers if the leader's strategy x * is the Nash equilibrium of the leader; hence, the strategy profile (x * , y * ) is the Nash equilibrium of the SLMFG, and this means that each follower cannot obtain additional payment by altering his/her recent strategy singly, that is, every follower makes his/her best response when the strategy of the leader is given.
The SLMFG (Figure 1) model of a leader and n followers be expressed as follows:

The Special Form of the SLMFG
From now on, we make the following some assumptions: is convex concerning y i , and the constraint function g i (x, y i , y −i ) is convex with respect to y i .
A general expression for the above model is as follows: The leader : max x∈X ϕ(x, y), The followers : ∀i ∈ I, min y∈Y f i (x, y), where x and y denote the leader's decision variable and the followers' decision variables, respectively. ϕ represents the leader's objective function and f i (i ∈ I) represent the followers' objective functions. The leader first selects his/her own strategy x, and the followers choose their own strategies y = (y 1 , . . . , y n ) through the leader's given strategy x.
We define the set for problem (4) as follows [18,23]: The feasible set of problem (4) is For fixed values of x ∈ X, the feasible set for the followers is Ω(x) = {y ∈ Y|g i (x, y i , y −i ) ≤ 0}. The projection of Ω in the leader's decision space is represented byX = {x|∃y, s.t.(x, y) ∈ Ω}. For fixed values of x ∈X, the response set for the followers is K(x) = {y|y ∈ argmin{ f i (x, y)|y ∈ Ω(x)}}, and the induced domain of problem (4) is Ξ = {(x, y)|x ∈ X, y ∈ K(x)}. Thus, the problem (4) can be translated into an optimization problem as follows: max{ϕ(x, y)|(x, y) ∈ Ξ}.
Since the Nash equilibrium may not be unique but multiple, the players are confused when making decisions. The refinement of the Nash equilibrium is an essential prerequisite. Thus, the concept of efficiency is incorporated into the Nash equilibrium, called the efficient Nash equilibrium, which maximizes social welfare and further refines the number of Nash equilibria. Efficient Nash equilibrium expresses a win-win idea under certain conditions, making it beneficial to all players and enabling it to reduce the number of Nash equilibria.

The Definition of Efficient Nash Equilibrium
We give the following some definitions: Definition 1 ((Efficiency) [6]). If a strategy maximizes social welfare in a way that the leader obtains the biggest rewards and the sum of the followers' payoff is the smallest in the entire feasible strategy space, then it is called efficient. Definition 2 (Efficient Nash equilibrium). Let S be all Nash equilibrium strategies of the SLMFG, the sum function mapping is U : X × Y → R, U k (x, y)(k = 0, 1, . . . , n) indicates the sum of the payoffs obtained from all Nash equilibrium strategies S, where k = 0 denotes the payoff sum from the leader's strategy and k = 1, . . . , n denote the payoff sums from followers' strategy. If the sum of the payoffs Z * (Z * ∈ S), then, for any other strategy, the sum of the payoffs Z(Z ∈ S) satisfies the following: and then Z * is called the efficient Nash equilibrium. This means that social welfare is maximized, and each player cannot obtain additional advantages by altering his/her present strategy individually.
The set of Nash equilibria depends on the leader's decision variable x and the followers' decision variables y = (y 1 , . . . , y n ). If x and y = (y 1 , . . . , y n ) satisfy the following equation: Then, (x * , y * ) is the efficient Nash equilibrium of the SLMFG, and this signifies that the leader obtains the largest rewards; the sum of the followers' payoffs is the smallest under the strategy profile (x * , y * ); and the leader and the followers cannot obtain additional reward by altering their current strategies unilaterally.

The Transformation of the SLMFG
When the upper-level leader's strategy is given, we need to be able to consider converting the lower-level follower problem into a nonlinear equation problem (NEP). Through the Karush-Kuhn-Tucker (KKT) condition, the SLMFG is converted into a nonlinear complementarity problem (NCP), and the nonlinear complementarity problem is transformed into a NEP through the complementarity function method. Therefore, the SLMFG problem is regarded as a nonlinear optimization problem with a bilayer structure, and the SLMFG problem is solved by using the IPSO algorithm.

The SLMFG Is Turned into a Nonlinear Equation Problem (NEP)
When the followers satisfy Assumption 1(b,c), the optimal solutions of the followers are represented by the Karush-Kuhn-Tucker (KKT) condition, and the followers can be expressed as NCP through the KKT condition [24]. Therefore, when the leader's strategy x is given, if y * contains the appropriate constraint condition, after that there is a multiplier λ * i such that (x, y * i , λ * i ) satisfies the following KKT system [25]: where g i (x, y i , y −i ) are the constraints of the followers, which depend on the control variables of the leader and the control variables of the followers. The Lagrangian function for the ith follower of system (7) is as follows: Consequently, system (7) is regarded as a first-order necessary condition for the followers. Let the SLMFG further satisfy Assumption 1(c); then, ∀y −i (i = 1, 2, . . . , n), formula (5) is turned into a convex optimization problem, and system (7) turns into the SLMFG's sufficient state. We can obtain the following result: (6) and system (7), then a strategy profile (x * , y * ) is the efficient Nash equilibrium of the SLMFG. (7) is equal to the system as follows: System (9) is NCP, thus the followers' problems of the SLMFG are converted into the NCP for the convex optimization. The Lagrangian function is uncertain differentiable, so the function needs to be smoothed further [24][25][26], but the function only needs a differentiable in this paper. However, through the complementary function, the system (9) is converted into a NEP in this paper.

The Nonlinear Complementarity Problem (NCP) Is Converted into a Nonlinear Equation Problem (NEP)
A NCP is transformed into a NEP through complementarity function methods. The function φ : R 2 → R is called a complementarity function if and only if φ(E, that is, we obtain the following complementarity function: where Θ(x, y, λ) is a FB function; then, the solution of Equation (10) is equivalent to Θ(x, y, λ) = 0. Consequently, the followers' problems are transformed into a NEP by the Fisher-Burmeister (FB) complementarity function. Thus, the NCP for the followers is converted to a NEP. For the NEP, the IPSO algorithm is designed to solve the optimal responses y * of the followers and the leader's optimal strategy x * , respectively. For followers, the fitness function is expressed as follows: Obviously, F(x, y) = 0 ⇔ Θ(x, y, λ) = 0, which means solving the value of min x∈X,y∈K(x) F(x, y) under the leader's fixed strategy. We obtain the followers' optimal strategies y * by the IPSO algorithm. For the leader, the fitness function is as follows: Furthermore, we can obtain the leader's optimal strategy x * by the IPSO algorithm. For the SLMFG, with (x * , y * ) satisfying Definition 2, the efficient Nash equilibrium (x * , y * ) is solved by using the IPSO algorithm. Finally, we obtain a reasonable, efficient Nash equilibrium solution by a refinement of the Nash equilibrium, which implies benefits to all players.

The Design of Immune Particle Swarm Optimization (IPSO) Algorithm
For a nonlinear equation equilibrium problem, the IPSO algorithm is designed by incorporating an immune memory function and an antibody concentration inhibition mechanism into the PSO algorithm. In this section, the IPSO algorithm is designed.

The Particle Swarm Optimization (PSO) Algorithm
The PSO algorithm was originally derived by Kennedy and Eberhart [27]. The PSO algorithm finds the optimal solution through collaboration and shares information among individuals. In the PSO algorithm, the solution of each optimization problem can be viewed as a "particle" in the search space. In the PSO algorithm with a population size of M and an N-dimensional space, x i = (x i1 , x i2 , . . . , x iN ) denotes the position vector of particle i and v i = (v i1 , v i2 , . . . , v iN ) represents velocity vector of particle i. According to the optimization model, each particle moves towards its own best current position p best (known as personal best) and towards the globally best particle g best (global best). At step t, the basic velocity and position of particle i are updated using the following equation: x t+1 where c 1 is the cognizant factor and c 2 is the communal factor. r 1 and r 2 are two N × N diagonal elements uniformly distributed in the section [0,1]. w is the inertia weight and w has an impact on the global and local exploration capabilities of the particle. When w is large, the global exploration ability is strong at the beginning of the process. When w is small, the local exploitation ability is stronger in the search space than when the weight is large. At present, increasing the value of the dynamic inertia weight value causes a linear decrease in the weight strategy, and the calculation formula is as follows [28]: where w max is the maximum inertia weight and w min is the minimum inertia weight. T max and T are the maximum number of iterations and the current number of iterations, respectively.

The Immune Particle Swarm Optimization (IPSO) Algorithm
The IPSO algorithm is a novel intelligent optimization algorithm [29,30] in view of the immune evolution mechanism and information sharing in biological immune systems. The optimal solution is regarded as an antibody, and the functions of object and the restricting terms are considered antigens. The IPSO algorithm is the combination of a probability concentration selection function and the PSO algorithm. During the process of particle (antibody) population updating, we always hope that highly adaptable particles (antibodies) are left behind. If a particle (antibody) is too concentrated, it is difficult to guarantee the diversity of the particle (antibody), and the algorithm may even fall into a local optimum. Therefore, those particles which have worse fitness but a better evolutionary tendency are maintained through the antibody probability concentration selection formula. Antibody M + Q consists of a nonempty set X, and the distance of the antibody F(x, y i )(i = 1, 2, . . . , M + Q) is calculated by We can define the concentration formula of particle i as follows: We can attain the probability concentration selection function for the followers as [31]: where y i (i = 1, 2, . . . , M + Q) represents the particle position for the followers and F(x, y i ) denotes the fitness value of the function for the followers. Similarly, we can also obtain the probability density selection function for the leader as follows: where x i (i = 1, 2, . . . , M + Q) represents the particle position for the leader and ψ(x i , y) is the fitness value of the function for the leader.
Increasing the new population Q primarily maintains the dynamic equilibrium of the population and takes the role of adjusting population concentration. Specifically, when the evolutionary population exhibits worse diversity and weaker global search ability, the IPSO algorithm allows the population to shift to a region with a better evolutionary tendency.

Implementation Steps of the IPSO Algorithm
The IPSO algorithm implement steps are as follows: Step 1: Initialize the parameters. The maximum number of iterations for the followers is T 1 and the maximum number of iterations for the leader is T 2 . The acceleration constants are c 1 and c 2 , the inertia weight values are w max and w min , and the precision is ε. The size of the randomly generated population is M, and the initial value x 0 is randomly generated according to the feasible domain of the leader.
Step 2: The IPSO algorithm can obtain the initial population p 0 by randomly generating the followers' initial positions y i (i = 1, 2, . . . , n) and initial velocities v yi with the followers' set-value mappings.
Step 3: The IPSO algorithm is used to calculate each particle's fitness function value for the followers and find the individual best position p ybest and population best position g ybest .
Step 4: Equation (15) is used to compute the inertia weight w.
Step 5: Equations (13) and (14) are used to update the velocities and positions of followers particles, respectively; compare the fitness functions of the particles; and retain the population best position in the memory library.
Step 6: Followers are randomly generated to obtain a new population with size Q.
Step 7: We select population M from the new population M + Q through the probability concentration selection formulation Equation (20).
Step 8: We choose particles from the memory library to replace particles with poor fitness, and then the immune system produces a next generation of particles p 1 . Then, the algorithm continues to update the particle velocities and positions by using Equations (13) and (14).
Step 9: By calculating the fitness value of particle y i 's current position, y i 's fitness value is compared with y i−1 's fitness value. If F(x 0 , y i ) < F(x 0 , y i−1 ), then y i−1 = y i ; otherwise, y i = y i−1 .
Step 10: Each particle's fitness function value for the followers is calculated, and the individual best position p ybest and population best position g ybest are found. Hence, we can compare the fitness value of the particle y i with the fitness value of the global g ybest ; if F(x 0 , g ybest ) < F(x 0 , y i ), then y i = g ybest ; otherwise, g ybest = y i .
Step 11: Stopping condition of the followers: Does the maximum number of iterations T 1 or the precision |F(x i−1 ) − F(x i )| < ε 1 satisfy the termination condition? If yes, we output the optimal particle y * (approximate solution of the followers); otherwise, we return to Step 4 .
Step 12: The followers' optimal particle y * is returned as feedback to the leader.
Step 13: The IPSO algorithm is used to calculate each particle's fitness function value for the leader and find the individual best position p xbest and population best position g xbest .
Step 14: The leader particle's initial position x i and initial velocity v xi are randomly generated. For each particle i, Equation (13) is applied for updating the velocity of the particle and Equation (14) is applied to update the position of the particle.
Step 15: A new population number of size Q is randomly generated.
Step 16: We choose population M from the new population M + Q through the probability concentration selection formula Equation (21).
Step 17: From the memory library, we choose particles to replace particles with worse fitness among the population, and the immune system generates a new generation of particles p 1 . Then, the algorithm continues to update the particle velocities and positions by using Equations (13) and (14).
Step 18: By calculating and comparing x i 's fitness value with Step 19: Each particle's fitness function value for the leader is calculated, and the individual best position p xbest and population best position g xbest are found. Hence, we can compare the particle x i 's fitness value with the global optimal particle g xbest 's fitness value; if ψ(g xbest , y * ) > ψ(x i , y * ), then x i = g xbest ; otherwise, g xbest = x i .
Step 20: Stopping condition for the leader: Is the maximum number of iterations T 2 achieved or is the precision |ψ(x i−1 ) − ψ(x i )| < ε 2 ? If yes, we output the optimal particle x * ; otherwise, we return to Step 14.
The flow chart of the IPSO algorithm is presented in Figure 2: The flow chart of the immune particle swarm optimization algorithm.

Performance Evaluation of the IPSO Algorithm
The IPSO algorithm expresses a bio-evolving swarm intelligence algorithm that is parallel to genetic algorithms (GAs), such as "natural selection" and "survival of the fittest". Thus, the measurable analysis method proposed by De Jong [32] can be used for evaluating the convergence of the IPSO algorithm according to its off-line performance. The followers : s * (x, y) = 1 T 1 The leader : u * (x, y) = 1 T 2 From above equations, we know that off-line performance represents the cumulative average of the best fitness function. When particles are closer to the fitness function value, the particles can better adapt to the SLMFG problems, thus the particles are more suitable for the objective functions under certain constraints.

Numerical Experiment
The SLMFG can be regarded as a bilevel programming problem. In the paper, the IPSO algorithm is applied for solving the leader's optimal strategy x * and the followers' optimal strategy y * , respectively. The IPSO algorithm parameters are set as follows: the population size is M = 30, the learning factors are c 1 = c 2 = 2, w max = 0.2, and w min = 0.1. The maximum numbers for the followers and leader are T 1 = 300 and T 2 = 200, respectively. The size of the new population of followers is Q = 10 and the precision of the fitness function is set to ε = 10 −3 . The Nash equilibrium of the SLMFG is solved by the IPSO algorithm, and we can calculate the efficient Nash equilibrium by a refinement of the Nash equilibrium, which implies benefits to all players. Example 1. Suppose we have a SLMFG, where the leader strategy is x and the followers' strategies are y 1 and y 2 . The leader's payoff function is ϕ(x, y), and the followers' payoff functions are f 1 (x, y 1 ) and f 2 (x, y 2 ).
The corresponding numerical results of Example 1 are given in Tables 1 and 2, and its off-line performance is shown in Figure 3a,b, respectively.     In Table 1, the average number of iterations for the follower problem is 283, and we can obtain that the approximate solution for the followers is (4, 5) T . In Table 2, the average number of iterations for the leader problem is 98, and we obtain that the leader's approximate efficient Nash equilibrium is 7.5. The efficient Nash equilibrium can minimize the income gap for the followers and maximize the rewards earned by the leader, thus strategy (7.5, 4, 5) is an efficient Nash equilibrium since Example 1 is just unique Nash equilibria. During the calculation process, the number of iterations is small and the convergence of the IPSO algorithm does not depend on the selection of the initial points. Furthermore, it greatly reduces the calculation time of the algorithm, and it does not easily fall into the local optimal solutions. As shown in Figure 3a,b, the off-line performance of the algorithm indicates that the algorithm has a fast convergence speed and is effective.

Example 2.
Suppose we have a SLMFG [33], where the leader strategy is x = (x 1 , x 2 , x 3 , x 4 ) and the strategies of the followers are y 1 = (y 11 , y 12 ) and y 2 = (y 21 , y 22 ). The leader's payoff function is ϕ(x, y 1 , y 2 ), and the followers' payoff functions are f 1 (y 1 ) and f 2 (y 2 ). max x ϕ(x, y 1 , y 2 ) = (200 − y 11 − y 21 )(y 11 + y 21 ) + (160 − y 12 − y 22 )(y 12 + y 22 ) s.t.0.4y 11 + 0.7y 12 ≤ x 1 , 0.6y 11 + 0.3y 12 ≤ x 2 , 0 ≤ y 11 , y 12 ≤ 20. min For the leader's decision vector x 0 , the followers' corresponding strategy is (y 1 , y 2 ). The optimal decision vector for the followers may not be unique when the strategy of the leader fixed. Thus, the efficient Nash equilibrium is also not unique and may even be multiple, but the number of Nash equilibria is greatly reduced, thus the Nash equilibria are refined efficiently. The followers' strategy y = (y 1 , y 2 ) is solved by the IPSO algorithm, and the leader's strategy x * is solved by the IPSO algorithm. The numerical results are shown in Table 3: · · · · · · · · · · · · · · · A run of the IPSO algorithm with 178 iterations for the followers and 105 iterations for the leader was executed. The calculation results are shown in Table 3. When the leader has the greatest benefit, there is dynamic competition among followers, that is, when one player's income grows, the other player's income is reduced. The total CPU time spent was 41 s. By Definitions 1 and 2, the leader chooses the strategy that maximizes the total rewards and minimizes followers' income gap, which means social welfare is maximized, and further each player cannot obtain additional rewards by varying his/her present strategy individually. In Table 3, the minimum total payoff min{min f 1 (y 1 ) + min f 2 (y 2 )} is equal to 54 for the followers, and the smallest income gap min{min f 1 (y 1 ) − min f 2 (y 2 )} is equal to 4. At this point, we can obtain that the efficient Nash equilibrium solution is (7, 3, 12, 18; 0, 10; 30, 0) with the leader's objective value ϕ(x * , y * ) = 6600, and the objective values of the two followers are f 1 (y * 1 ) = 25 and f 2 (y * 2 ) = 29, respectively. In [21], a run of the genetic algorithm with 600 generations shows that a Stacklberg-Nash equilibrium is (7.05, 3.13, 11.93, 17.89; 0.26, 9.92; 29.82, 0.00) with an objective value ϕ(x * , y * ) = 6599.99, and the objective values of two followers are f 1 (y * 1 ) = 23.47 and f 2 (y * 2 ) = 30.83, which means the smallest income up for the two followers' is 7.36 and the minimize total payoff is 54.30. As there is a greater income gap between followers and more total payoff than in this paper, the results in [21] are inferior to the IPSO algorithm. In [33], the value of the leader's objective function is also 6600, but this traditional mathematical analysis method has high computational complexity; the minimum total payoff is 119.42, the smallest income gap is 19.8, and the efficiency is inferior to that of the IPSO algorithm. The IPSO algorithm has a fast convergence speed, saves time, and is effective. In summary, the IPSO algorithm obtains the optimal efficient Nash equilibrium of (7, 3, 12, 18; 0, 10; 30, 0), which minimizes the income gap among all followers and maximizes the incomes of the leader.
For the optimization problem of the SLMFG, the followers' strategy is y = (y 1 , y 2 , y 3 ) and the leader's strategy is x. We use the IPSO algorithm to search the optimal solutions. A run of the IPSO algorithm with 254 iterations for the followers and 132 iterations for the leader was executed. The computation results are shown in Table 4.
The followers' strategy y * = (y * 1 , y * 2 , y * 3 ) is solved by the IPSO algorithm. y * 1 , y * 2 , and y * 3 are three identical objective function because their components are equivalent. The leader's strategy x * is solved by the IPSO algorithm. In Table 4, the efficient Nash equilibrium sets are: The efficient Nash equilibrium solution of Example 3 is multiple. By Definitions 1 and 2, to increase efficiency, each player chooses the strategy that minimizes the income gap among the followers and maximizes the total payoffs of the leader. Furthermore, the leader's objective value is 9.593, the followers' objective values are one of {1.609, 7.094, 0.000}, and max 3 ∑ i=1 f i (y i ) = 18.300. Thus, when Example 3 obtains the efficient Nash equilibrium, the leader's maximum benefit is 9.593, and the followers' total maximum benefit is 18.300. The convergence speed of the IPSO algorithm is superior to that of the algorithm in [21] with 300 generations. The calculation time of the IPSO algorithm is less than that in [34]. The IPSO algorithm has a fast convergence speed, saves more time, and is effective. In Example 3, the IPSO algorithm obtains the optimal efficient Nash equilibrium, which is also the efficient Nash equilibrium set, thereby minimizing the income gap among all followers and maximizing the reward of the leader.