You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

7 March 2025

Differential Privacy in Federated Learning: An Evolutionary Game Analysis

and
1
School of Computer Science and Technology, China University of Mining and Technology, Xuzhou 221116, China
2
School of Information and Electronic Engineering, Zhejiang Gongshang University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Multimedia Smart Security

Abstract

This paper examines federated learning, a decentralized machine learning paradigm, focusing on privacy challenges. We introduce differential privacy mechanisms to protect privacy and quantify their impact on global model performance. Using evolutionary game theory, we establish a framework to analyze strategy dynamics and define utilities for different strategies based on Gaussian noise powers and training iterations. A differential privacy federated learning model (DPFLM) is analyzed within this framework. A key contribution is the thorough existence and stability analysis, identifying evolutionarily stable strategies (ESSs) and confirming their stability through simulations. This research provides theoretical insights for enhancing privacy protection in federated learning systems.

1. Introduction

In the era of the information age, the generation and accumulation of data is accelerating at an unprecedented rate. However, this vast amount of data is often fragmented across various locations and managed by multiple entities. Traditionally, analyzing such dispersed data has relied heavily on centralized processing, which requires the aggregation of data from various locations. As concerns over data security and privacy continue to grow, conventional approaches have become increasingly inadequate, often resulting in significant privacy leakages [1]. Specifically, when private data are used for model training, the trained models may inadvertently disclose sensitive information through their behavior or structure [2,3,4].
In recent years, federated learning (FL) has emerged as a novel machine learning (ML) paradigm, gradually gaining traction. The core idea of federated learning is to enable multiple clients to collaboratively train a shared machine learning model while keeping their data decentralized and locally stored. Although federated learning can protect client privacy to some extent, significant challenges remain. For example, Shokri et al. [5] introduced membership inference attacks, highlighting a significant privacy vulnerability within federated learning. In response to such challenges, recent studies have increasingly turned to advanced cryptographic tools, such as secure multi-party computation and differential privacy, to enhance privacy protection in federated learning [6]. For instance, Wei et al. [7] investigated the trade-off between convergence performance and differential privacy protection in federated learning under a unified noise scale. Hu et al. [8] examined the relationship between model accuracy and privacy protection with a uniform noise scale. Geyer et al. [9] studied the relationship between privacy loss and model performance. However, despite significant progress in addressing privacy concerns in federated learning, many critical questions remain open for exploration. For instance, while existing studies have highlighted the importance of privacy protection through techniques such as secure multi-party computation and differential privacy, there is still a need for more comprehensive frameworks to quantify privacy leakage and to understand the impact of privacy preferences on model performance. This gap underscores the ongoing and evolving nature of privacy protection in federated learning, making it a highly relevant and active area of research.
Game theory, particularly evolutionary game theory, provides a powerful framework for analyzing the strategic behavior of participants in federated learning systems. Traditional game theory focuses on systematically analyzing competitors to choose the best strategy for maximizing personal benefits [10]. In contrast, evolutionary game theory [11,12,13] emphasizes the evolution of strategies and changes at the population level. It is widely used for analyzing interactions between groups. In the context of federated learning, evolutionary game theory can effectively model the dynamic behavior of clients as they balance the trade-off between privacy protection and model performance. Clients aim to maximize their utility by selecting strategies that protect their privacy while still contributing to the global model’s accuracy. The concept of population share in evolutionary game theory can be used to reflect the proportion of clients adopting different strategies over time, capturing the evolving dynamics of client behavior [14]. Recently, incentive allocation mechanisms in evolutionary game theory have garnered academic attention [15,16,17]. A well-designed incentive mechanism can shape the behavioral patterns of the group during evolutionary games, reduce the selfish behavior of individuals, and thereby enhance group stability. Previous studies [18,19,20] have applied game theory ideas using various models but have not considered the impact of individual strategies on overall incentives.
In reality, individual strategy choices play a crucial role in determining the performance of the global model in federated learning and have a direct impact on the actual overall incentives. Given the decentralized nature of federated learning, where each client has the autonomy to choose strategies that balance privacy protection and model performance, understanding these choices and their collective impact is essential. Traditional game theory, while powerful, often focuses on individual strategy optimization and may not fully capture the dynamic and evolving nature of client behavior in federated settings. In contrast, evolutionary game theory provides a more comprehensive framework to analyze how individual strategies evolve over time and how these changes affect the overall system dynamics. This study leverages the strengths of evolutionary game theory to address the challenges posed by individual strategy choices in federated learning and makes several key contributions:
  • Evolutionary game theory analysis: We employ evolutionary game theory to analyze the differential privacy protection problem in federated learning. Unlike traditional game theory, which focuses on individual strategy choices, our approach examines the dynamics of population shares, capturing how the collective behavior of participants evolves over time. This perspective provides a novel framework for understanding the interplay between privacy preferences and system performance.
  • Impact of differential privacy mechanisms: We quantify the impact of differential privacy mechanisms on global model performance and reflect these impacts in the incentive pool. Additionally, we characterize privacy leakage to understand its effects on user utility, providing a comprehensive analysis of the trade-offs between privacy protection and model performance.
  • Differential privacy federated learning model (DPFLM): We constructed DPFLM, a robust model that integrates differential privacy into federated learning. We also propose an algorithm to compute the utility of each strategy, enabling systematic analysis of the incentives associated with different privacy mechanisms.
  • Existence and stability analysis: We conducted a thorough analysis of the existence of solutions in evolutionary game theory. For the two-strategy case, we derived closed-form results for the evolutionarily stable strategies (ESSs). For more general cases, we validate our theoretical findings through extensive simulations, demonstrating the practical applicability of our approach.

3. System Model and Evolutionary Game Formulation

3.1. System Model

We consider a differential privacy federated learning model comprising I clients, denoted by P 1 , , P I , and a central server. When the server initiates a training task, it assigns the task to the clients and offers incentives to those who participate. Clients, in turn, decide whether to engage in the training task and, if so, employ differential privacy mechanisms to protect the privacy of their local data. Specifically, they select appropriate numbers of local iterations, E i , and add Gaussian noise with power σ i during model training and parameter uploading. The federated learning system model is illustrated in Figure 3.
Figure 3. Federated learning system model.
To analyze this model within the framework of evolutionary game theory, we define the following components:
  • Players: The players in this game are the clients, denoted by P 1 , , P I .
  • Strategies: The nth strategy available to a player can be expressed as S n = ( E n , σ n ) , where E n is the number of local iterations per round, and σ n 2 is the Gaussian noise power. We assume there are N strategies in total, denoted by S 1 , , S N . Here, ‘strategies’ represent the different combinations of local iterations and Gaussian noise power that a client can choose. These strategies are formulated based on the practical application of differential privacy in FL. Different noise powers represent varying levels of privacy protection, with higher noise offering stronger privacy but potentially impacting model accuracy and, thus, the incentive pool. The number of local iterations also influences both model contribution and potential privacy leakage. These strategies are relevant to clients because they directly control the trade-off between their privacy protection and their potential utility. Clients with different privacy preferences or data sensitivity levels might choose different strategies.
  • Population share: The population share represents the proportion of clients choosing a particular strategy within the strategy space. Let x n X denote the proportion of the population that selects strategy S n , and it satisfies n = 1 N x n = 1 .
  • Utility: The utility for each client is determined by their net income, which is the difference between the payoff obtained from the server and the training costs incurred. The utility for selecting strategy S n at time t is denoted by U n ( t ) . For simplicity, this is subsequently abbreviated as U n .
In the framework of evolutionary game theory, each player can maximize their utility by adjusting their own strategy. The selection of strategies by clients is of significant importance because it directly impacts the trade-off between utility and privacy. An optimal strategy aims to maximize client utility while adhering to privacy constraints. For instance, a client prioritizing incentives above all else might choose a strategy that introduces minimal or no noise to their data. Although this could potentially maximize their contribution to an incentive pool, it would simultaneously lead to a substantial increase in privacy leakage, potentially undermining the overall privacy guarantees of the system. Conversely, if all clients were to adopt overly conservative strategies by adding excessive noise, the aggregated model would suffer from poor training quality, resulting in a diminished incentive pool and, ultimately, reduced utility for all clients. Therefore, the strategic selection of privacy mechanisms is crucial to achieving a delicate equilibrium, enabling clients to derive meaningful utility from participation while simultaneously ensuring robust privacy protection. In essence, effective strategy selection is key to balancing client incentives with the fundamental requirement of privacy preservation within the system. As the game progresses, players’ strategies will continuously evolve, making this process a dynamic adjustment. Overall, the population shares of different strategies will also change over time. The notations are summarized in Table 1.
Table 1. List of notations.

3.2. Utility Function

The utility function in our model accounts for the impact of differential privacy on the training model. Specifically, the utility for each client is determined based on the upper bounds of convergence for both the global and local loss functions, as well as the strategies selected.

3.2.1. Incentive Pool

When clients conduct model training and employ differential privacy mechanisms, they upload their local model parameters to the server. The server then aggregates these parameters from all participants to generate a global model. The upper bound of convergence for the global loss function is used to estimate the algorithm’s convergence rate, thereby quantifying the impact of Gaussian noise and the number of iterations on model performance. The size of the incentive pool is determined based on the model’s performance. Drawing inspiration from [29], we express the global loss function L g as follows:
L g = Θ c η T G + 5 η 2 l 2 ( σ L 2 G + 6 σ G 2 G 2 ) c + d 2 c η 2 N 2 G 2 n = 1 N ( σ n 2 · I · x n )
Here, Θ represents the difference between the initial loss and the optimal loss. c is a constant. η is the model learning rate. T is the number of global updates. G is the number of global iterations. l is the Lipschitz constant. σ L 2 bounds the variance of the stochastic gradients at each client, and σ G 2 bounds the difference between the gradient of a single client and the global one. From Equation (3), we observe that the larger the Gaussian noise added by the clients, the larger the loss, implying a slower convergence rate for the global model. Consequently, the size of the incentive pool is given by
Q E = r L g ,
where r is a constant determined by the server.

3.2.2. Payoff, Cost, and Utility

The incentive pool is shared among all clients. Since the noise added by the clients affects the performance of the model, and the larger the noise, the worse the performance of the local model, the payoff should be lower for clients with larger noise power. We assume that the client with noise power σ n 2 should receive a payoff proportional to 1 σ n . Given that the size of the incentive pool is Q E , using normalization, the payoff for choosing noise power σ n 2 , denoted by p n , is given by
p n = Q E · 1 σ n I · i = 1 N x i σ i + A ,
where A represents the basic reward granted by the server to participants for their involvement in the training process. We observe that the payoffs, p 1 , , p N , are determined by the population shares x 1 , , x n .
The cost arises from the impact of Gaussian noise on privacy leakage for each client. Most current studies measure privacy leakage through experiments, which, however, requires access to user data, contradicting the original intention of privacy protection. Liu et al. [29] conducted research on the theoretical measurement of privacy leakage, making it possible to theoretically quantify privacy leakage and apply it to various scenarios and datasets. Motivated by their results, we express the cost of strategy, S n , c n , as
c n = η 2 E n 2 2 σ n 2 .
Here, η is the model learning rate. We emphasize that more local iterations make the model more suitable for the local dataset, increasing the likelihood of data leakage.
Summarizing the above, the utility for selecting strategy S n can be expressed as
U n = p n c n = Q E · 1 σ n I · i = 1 N x i σ i + A η 2 E n 2 2 σ n 2 .
Notice that U n is a function of x 1 , , x n . The following Algorithm 1 summarizes the calculation of the utility.
Algorithm 1 Utility Calculation Algorithm.
1:
Input: Selected strategy profile S n , other federated learning parameters
2:
Output: The utility for selecting strategy S n , U n
3:
Initialize the input parameters.
4:
Based on the overall noise and the number of iterations, use the global loss function, calculate the result of Q E according to Equations (3) and (4).
5:
Based on the strategies chosen, calculate the payoff and cost using Equations (5) and (6).
6:
Calculate the utility for selecting strategy S n , U n , using Equation (7).

3.3. Replicator Dynamics

Replicator dynamics (RDs) is a widely used model that describes how strategies evolve over time in evolutionary game theory. It is primarily applied to study the dynamic processes of strategy selection and evolution [37]. Therefore, this study employs replicator dynamics to analyze the dynamic process of strategy selection and evolution in our federated learning system.
The utility for selecting strategy S n is given by Equation (7), and the average utility of the clients involved in the training can be given by the following expression:
U ¯ = 1 N n = 1 N U n x n
In each federated learning task, clients can freely choose their strategies to maximize their utilities. Since the utility is related to the population share, clients cannot simply select the strategy with the minimum noise to achieve the largest size of the incentive pool. Based on the above analysis, the replicator dynamics equation for clients can be expressed as
x ˙ n ( t ) ϕ n x 1 ( t ) , , x N ( t ) = Δ x n ( t ) U n ( t ) U ¯ ( t ) , n N .
Here, Δ represents the positive learning rate, which controls the speed at which the clients adapt to different strategies. Based on replicator dynamics, when the utility of a client is lower than the average, it will adjust its strategy to obtain a higher utility. Often, at the end of the game, an equilibrium point will be reached.

4. Game Analysis

In this section, we investigate the existence and stability of the replicator dynamics described by Equation (9). Regarding existence, we demonstrate that the game has at least one stable point (equilibrium). Regarding stability, we show that once the game reaches equilibrium, it can return to the equilibrium state even if disturbed. First, we prove the boundedness of Equation (9) in Lemma 1.
Lemma 1.
For all n , v { 1 , , N } , the first-order derivative of ϕ n ( x 1 , , x N ) with respect to x v ( t ) is bounded.
Proof of Lemma 1.
For ease of calculation, we omit t in the proof. Then, the coefficients of the first, second, and third terms in Equation (3) are constants. The sum of the first and second terms is denoted by B, and the coefficient of the third term is denoted by C. Thus, Equation (3) can be rewritten as
L g = B + C · i = 1 N ( σ n 2 · I · x i )
The derivative of ϕ n ( x 1 , , x N ) with respect to x v can be given as follows:
d ϕ n ( x 1 , , x N ) d x v = Δ d x n x v U n U ¯ + x n d U n d x v d U ¯ d x v .
This equation represents the partial derivative of the replicator dynamics equation ϕ n ( x 1 , , x N ) with respect to the population share of strategy v, denoted as x v . It is derived using the product rule of differentiation. Specifically, considering ϕ n as a product of two terms, Δ x n and ( U n U ¯ ) , its derivative with respect to x v is given by d ϕ n d x v = d ( Δ x n ) d x v ( U n U ¯ ) + ( Δ x n ) d ( U n U ¯ ) d x v . To simplify the notation, we define
G ( x 1 , , x N ) = i = 1 N ( σ n 2 · I · x i ) , H ( x 1 , , x N ) = I · i = 1 N x i σ i .
Thus, we can obtain
d U n d x v = r · d G ( x 1 , , x N ) d x v B + c · G ( x 1 , , x N ) 2 · 1 σ n H ( x 1 , , x N ) + r B + c · G ( x 1 , , x N ) · d H ( x 1 , , x N ) d x v σ n H ( x 1 , , x N ) 2 .
Since d G ( x 1 , , x N ) d x v = σ v 2 · I , d H ( x 1 , , x N ) d x v = I σ v , we can obtain that d U n d x v is bounded. Hence, d ϕ n ( x 1 , , x N ) d x v is also bounded. □
Then, we can prove that the solution of (9) exists.
Theorem 1.
For all n { 1 , , N } , the solution of (9) exists.
Proof of Theorem 1.
From Lemma 1, we can obtain that d ϕ n ( x 1 , , x N ) d x v is bounded. In addition, it is straightforward to see that ϕ n ( x 1 , , x N ) and d ϕ n ( x 1 , , x N ) d x v are both continuous. Therefore, ϕ n ( x 1 , , x N ) satisfies the Lipschitz condition with respect to x 1 , , x N . According to the Picard-Lindelöf theorem, for any given initial conditions, the system of differential equations has a unique solution in the region under consideration. □
Now, we analyze the stability of this evolutionary game. We begin with the definition of Nash equilibrium.
Definition 1.
The population state x * = [ x 1 * , , x N * ] is called a Nash equilibrium of the game if for all feasible states x = [ x 1 , , x N ] in the population, the inequality ( x x * ) T ϕ ( x * ) 0 holds, where ϕ ( x * ) = [ ϕ 1 ( x * ) , , ϕ N ( x * ) ] [38].
In simpler terms, a Nash equilibrium is a fixed point of the ordinary differential equation, i.e.,
n N , x ˙ n ( t ) = 0 .
Suppose there exists another population state, x , that attempts to invade the state x * by attracting a fraction of players λ ( 0 , 1 ) . If the following condition holds for all λ ( 0 , 1 ) , then x * is an evolutionarily stable strategy (ESS). The definition is given as follows.
Definition 2.
If there exists a neighborhood B X such that
( x x * ) T ϕ ( x * ) = 0 , ( x * x ) T ϕ ( x ) 0 , x B
then the population state x * is an ESS of the game.
In practice, when N is very large, it is difficult to find a closed-form solution for x * . Thus, we study the case of two strategies ( N = 2 ) to illustrate stability in the game. Other cases can also be analyzed using a similar method, but due to the complexity of the calculations, they are not listed in this paper. In the numerical simulation section, cases with more than two strategies are still considered. Taking two strategies as an example, the share of choosing strategy 1 is x 1 , and the share of choosing strategy 2 is x 2 . Based on the replicator dynamics and the constraint conditions, the following equations can be listed:
Δ x 1 U 1 U ¯ = 0 , Δ x 2 U 2 U ¯ = 0 , x 1 + x 2 = 1 .
The utility for selecting strategy S 1 , U 1 can be expressed as
U 1 = r B + C · σ 1 2 I x 1 + σ 2 2 I x 2 · 1 σ 1 2 I x 1 σ 1 + x 2 σ 2 η 2 E 1 2 2 σ 1 2 + A
When I is sufficiently large, the influence of B on U 1 can be ignored, and we let
a = r C I 2 , b 1 = η 2 E 1 2 2 σ 1 2 , b 2 = η 2 E 2 2 2 σ 2 2 .
By substituting the above into Equation (15), we obtain the simplified expression of U 1 as follows:
U 1 = a σ 1 2 x 1 + σ 2 2 x 2 · x 1 σ 1 + x 2 σ 2 σ 1 b 1 + A .
By substituting Equation (17) into Equation (14), we obtain that
x 1 * σ 1 2 σ 2 2 σ 1 σ 2 β 2 σ 1 2 σ 2 2 , 0 , 1 , σ 1 2 σ 2 2 σ 1 σ 2 + β 2 σ 1 2 σ 2 2 ,
where
β = ( ( 4 a σ 1 2 4 a σ 2 2 + b 1 σ 1 4 + b 1 σ 2 4 b 2 σ 1 4 b 2 σ 2 4 + 2 b 1 σ 1 σ 2 3 + 2 b 1 σ 1 3 σ 2 2 b 2 σ 1 σ 2 3 2 b 2 σ 1 3 σ 2 + 3 b 1 σ 1 2 σ 2 2 3 b 2 σ 1 2 σ 2 2 ) / ( b 1 b 2 ) ) 0.5 .
Thus, x 2 * = 1 x 1 * , and in all solutions, we should have x 1 * [ 0 , 1 ] .
Further, we will study the evolutionary stability of four fixed points. In the cases of x 1 * = 0 and x 1 * = 1 , the shares of the two strategies are ( 0 , 1 ) and ( 1 , 0 ) , respectively. In these two cases, the forms are similar. Therefore, we only need to check the case of x 1 * = 0 .
The ordinary differential equation in Equation (9) asymptotically stabilizes to an ESS when the replicator dynamics are continuous in time. If the Jacobian matrix of the ordinary differential equation is negative definite at the steady state, then the point is asymptotically stable. The Jacobian matrix of the ordinary differential equations for this two-strategy case can be given as
J J 11 J 12 J 21 J 22 = ϕ 1 ( x 1 , x 2 ) x 1 ϕ 1 ( x 1 , x 2 ) x 2 ϕ 2 ( x 1 , x 2 ) x 1 ϕ 2 ( x 1 , x 2 ) x 2 | x 1 = x 1 * , x 2 = 1 x 1 *
To simplify the calculation process and make the final result simpler, we assume I + so that
lim I a = lim I r C I = 0 .
By substituting a = 0 into Equation (20), we can obtain that
J = 2 b 1 x 1 b 1 + b 2 x 2 b 2 x 1 b 1 x 2 b 1 x 1 b 2 + 2 b 2 x 2 .
Then, we can obtain the following:
  • When x 1 * = σ 1 2 σ 2 2 σ 1 σ 2 β 2 ( σ 1 2 σ 2 2 ) is an ESS, then the following inequalities must be satisfied:
    det ( J 11 ) = ( 2 b 1 b 2 ) γ 2 ( σ 1 2 σ 2 2 ) + b 1 < 0 , det ( J ) = ( b 1 b 2 ) ( σ 1 2 σ 2 2 + γ ) ( 2 b 1 σ 1 2 2 b 1 σ 2 2 + b 1 γ b 2 γ ) 2 ( σ 1 2 σ 2 2 ) 2 > 0 ,
    where γ = β | a = 0 + σ 1 σ 2 σ 1 2 + σ 2 2 .
  • When x 1 * = σ 1 2 σ 2 2 σ 1 σ 2 + β 2 ( σ 1 2 σ 2 2 ) is an ESS, then the following inequalities must be satisfied:
    det ( J 11 ) = ( b 2 2 b 1 ) γ 2 ( σ 1 2 σ 2 2 ) + b 1 < 0 , det ( J ) = ( b 1 b 2 ) ( σ 2 2 σ 1 2 + γ ) ( 2 b 1 σ 2 2 2 b 1 σ 1 2 + b 1 γ b 2 γ ) 2 ( σ 1 2 σ 2 2 ) 2 > 0 .
  • When x 1 * = 1 or x 1 * = 0 , the condition that the Jacobian matrix of the ordinary differential equation is negative definite cannot be satisfied. Taking x 1 * = 0 as an example, the following two conditions need to be satisfied:
    b 2 b 1 < 0 , b 2 2 b 1 b 2 > 0 .
    However, since b 1 and b 2 are both greater than 0, there is no solution that satisfies the conditions. Therefore, x 1 * = 1 or x 1 * = 0 are not stable points.

5. Numerical Results

In this section, we present numerical results to further substantiate our analysis and provide additional insights.

5.1. Utilities

In this subsection, we investigate the impact of different Gaussian noise powers on the utilities of various strategies in federated learning scenarios. Specifically, we analyze both two-strategy and four-strategy scenarios to understand how varying levels of noise affect the overall utility of the participating clients.

5.1.1. Two-Strategy Scenario

We first consider the model under a two-strategy scenario, with simulations examining the impact of varying Gaussian noise powers on the utilities. The parameters are set as follows: a = 10 , b 1 = 0.02 , and b 2 = 0.02 . The basic incentive for training is set as A = 5 , with a fixed population share of [ 0.5 , 0.5 ] . Additionally, σ 1 is held constant at 0.1 , while the initial value of σ 2 is set at 0.1 and incremented by 0.02 up to 2.5 . The values of U 1 and U 2 are observed in the left figure of Figure 4.
Figure 4. Utilities for different scenarios: (a) Two-strategy scenario. (b) Four-strategy scenario.
As shown in the left figure of Figure 4, the utilities of the two parties are initially identical. However, as σ 2 increases, the utilities of both parties begin to decrease, with a more pronounced impact on U 2 . This decline occurs because the continuous addition of Gaussian noise to S 2 affects the accuracy of model training, which, in turn, influences the size of the total incentive pool, ultimately leading to a reduction in the utilities of both parties.

5.1.2. Four-Strategy Scenario

In the right figure of Figure 4, we consider a scenario with four strategies. The parameters are set as a = 10 , b 1 = 0.02 , b 2 = 0.02 , σ 1 = 0.1 , σ 2 = 0.1 , σ 3 = 0.2 , and σ 4 = 0.3 . The shares are set to [ 0.25 , 0.25 , 0.25 , 0.25 ] , and the basic incentive for training is A = 5 . Throughout the simulation, the values of σ 1 , σ 3 , and σ 4 remain constant. We start with an initial value of σ 2 = 0.1 and increment it by 0.02 up to 2.5 . By examining the utilities of different strategies, as shown in the figure, it is clear that the greater the Gaussian noise added to S 2 , the lower the utilities.
Overall, the results demonstrate that the utility of each strategy is significantly influenced by the level of Gaussian noise added during the training process. As the noise increases, the utility of the strategies decreases, highlighting the trade-off between privacy protection and model performance.

5.2. Stability Analysis

In this section, we conduct a comprehensive stability analysis to verify the theoretical results presented in the previous section. Through numerical simulations, we examine the stability of the system under different scenarios, including two-strategy, four-strategy, and eight-strategy cases. The goal is to demonstrate that the system converges to an ESS under various initial conditions and parameter settings.

Two-Strategy Scenario

We begin with a two-strategy scenario, where the system’s stability is analyzed under varying Gaussian noise powers. The parameters are set as follows: I = 80 , 000 , N = 2 , σ 1 = 0.1 , E 1 = 100 , σ 2 = 1.1 , E 2 = 695 , η = 0.0002 , and A = 5 . Additional parameters include a = 0.02 , b 1 = 0.008 , b 2 = 0.004 , and initial shares [ 0.4 , 0.6 ] . The evolution of shares and utilities is shown in Figure 5.
Figure 5. Population shares and utilities for two-strategy scenario (Equation (23) holds): (a) Population shares. (b) Utilities.
The Jacobian matrix satisfies Equation (23), indicating that it is negative definite. The equilibrium point is calculated as x * = [ 0.7181 , 0.2813 ] . This implies that regardless of the initial shares, the system will eventually converge to this ESS. As shown in Figure 5, when the initial conditions satisfy Equation (23), the simulation results are consistent with the analytical results, thereby confirming the correctness of our previous analysis. From the right figure of Figure 5, we observe that when the system converges to the ESS, the utilities no longer change.
To further validate the stability analysis, we conducted a simulation for the second case where Equation (24) is satisfied. The number of clients is still set as I = 80 , 000 . For Strategy S 1 , the Gaussian noise is set to σ 1 = 0.4 with E 1 = 1265 iterations. For Strategy S 2 , the Gaussian noise is σ 2 = 1.1 with E 2 = 778 iterations. The learning rate is η = 0.0002 , and the basic incentive for training is still set as A = 5 . Other parameters are set as a = 0.1 , b 1 = 0.2 , and b 2 = 0.01 . The initial shares for selecting strategies were set to [ 0.4 , 0.6 ] . In this scenario, the Jacobian matrix satisfies Equation (24), indicating that it is negative definite. Through computation, we obtained the equilibrium point x * = [ 0.2180 , 0.7820 ] .
As shown in the left figure of Figure 6, when the initial conditions satisfy Equation (24), the simulation results are consistent with the computed results, thereby verifying our analysis. From the right figure of Figure 6, it is evident that when the system converges to the ESS, the utilities no longer change.
Figure 6. Population shares and utilities for two-strategy scenario (Equation (24) holds): (a) Population shares. (b) Utilities.

5.3. Beyond Two-Strategy Scenario

Next, we consider a four-strategy scenario to further validate the stability analysis. The parameters are set as a = 0.01 , b 1 = 0.03551 , b 2 = 0.0256 , b 3 = 0.0164 , and b 4 = 0.025 . The values of σ 1 to σ 4 form an arithmetic sequence starting at 0.21 with a common difference of 0.01 . The initial shares for selecting strategies were set to [ 0.4 , 0.1 , 0.15 , 0.35 ] . The numerical simulation results for this four-strategy scenario are shown in Figure 7. The equilibrium point is calculated as x * = [ 0.604 , 0.246 , 0.148 , 0.002 ] . The simulation result demonstrates that the system also reaches an evolutionarily stable strategy in the case of the four-strategy scenario.
Figure 7. Population shares for four-strategy scenario.
Finally, we extend the stability analysis to an eight-strategy scenario. The parameters are set as a = 0.0009 , b 1 = 0.007 × 10 5 , b 2 = 1 × 10 5 , b 3 = 1 × 10 5 , b 4 = 1 × 10 5 , b 5 = 0.002 , b 6 = 0.0011 , b 7 = 0.003 , and b 8 = 0.0031 . The values of σ 1 to σ 8 form an arithmetic sequence starting at 0.05 with a common difference of 0.001 . The initial shares for selecting strategies are set to [ 0.18 , 0.08 , 0.11 , 0.16 , 0.105 , 0.1 , 0.19 , 0.0705 ] . The simulation results for this eight-strategy scenario are shown in Figure 8. The equilibrium point is x * = [ 0.6817 , 0.3105 , x 3 , , x 8 ] , where x 3 to x 8 are all 9.75 × 10 4 . This simulation result demonstrates that the system also reaches an evolutionarily stable strategy in the eight-strategy case.
Figure 8. Population shares for eight-strategy scenario.
These results further confirm that the system reaches an evolutionarily stable strategy, even with a larger number of strategies.

6. Conclusions

This study integrates the differential privacy afforded by the Gaussian mechanism into federated learning, employing both global loss functions and privacy leakage metrics to quantify the added noise. This approach helps to create a more realistic modeling scenario. We propose an incentive allocation algorithm tailored to the DPFLM framework. This algorithm is designed to consider the noise levels introduced by each client. It also takes into account the number of iterations. Crucially, our DPFLM, analyzed through the lens of evolutionary game theory, provides a valuable framework to understand and address the critical challenges of privacy and incentive design in federated learning.
Through theoretical analysis, particularly focusing on a two-strategy scenario, we derive detailed analytical results regarding the ESS. This analysis demonstrates how clients, acting strategically, balance their privacy concerns with the desire for incentives and how the system can evolve towards a stable equilibrium of privacy strategies. Extensive simulations across two-, four-, and eight-strategy scenarios further validate our theoretical findings and demonstrate the practical applicability of our approach in diverse settings. Our model supports federated learning by offering a framework to analyze the privacy-utility trade-off, providing insights into the dynamic evolution of client privacy strategies, and informing the design of effective incentive mechanisms that promote both client participation and privacy protection.
For future work, we suggest exploring the incorporation of additional factors, such as propagation delay and dataset size, which could significantly influence the model’s behavior. Furthermore, integrating federated learning with other advanced deep learning techniques could provide valuable insights into the impact of these factors and open up new avenues for research.

Author Contributions

Conceptualization, Z.N. and Q.Z.; methodology, Q.Z.; software, Q.Z.; validation, Z.N.; formal analysis, Z.N.; investigation, Q.Z.; resources, Q.Z.; data curation, Q.Z.; writing—original draft preparation, Z.N.; writing—review and editing, Z.N.; visualization, Q.Z.; supervision, Q.Z.; project administration, Q.Z.; funding acquisition, Z.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Zhejiang Provincial Natural Science Foundation of China under Grant LQ22F010008.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, B.; Qi, P.; Liu, B.; Di, S.; Liu, J.; Pei, J.; Yi, J.; Zhou, B. Trustworthy AI: From principles to practices. ACM Comput. Surv. 2023, 55, 1–46. [Google Scholar] [CrossRef]
  2. Kaissis, G.A.; Makowski, M.R.; Rückert, D.; Braren, R.F. Secure, privacy-preserving and federated machine learning in medical imaging. Nat. Mach. Intell. 2020, 2, 305–311. [Google Scholar] [CrossRef]
  3. Liu, B.; Ding, M.; Shaham, S.; Rahayu, W.; Farokhi, F.; Lin, Z. When machine learning meets privacy: A survey and outlook. ACM Comput. Surv. (CSUR) 2021, 54, 1–36. [Google Scholar] [CrossRef]
  4. Wang, T.; Zhang, Y.; Qi, S.; Zhao, R.; Xia, Z.; Weng, J. Security and privacy on generative data in aigc: A survey. ACM Comput. Surv. 2024, 57, 1–34. [Google Scholar] [CrossRef]
  5. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership inference attacks against machine learning models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–24 May 2017; IEEE: New York, NY, USA; pp. 3–18. [Google Scholar] [CrossRef]
  6. Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated learning: Challenges, methods, and future directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
  7. Wei, K.; Li, J.; Ding, M.; Ma, C.; Yang, H.H.; Farokhi, F.; Jin, S.; Quek, T.Q.; Poor, H.V. Federated learning with differential privacy: Algorithms and performance analysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3454–3469. [Google Scholar] [CrossRef]
  8. Hu, R.; Guo, Y.; Li, H.; Pei, Q.; Gong, Y. Personalized federated learning with differential privacy. IEEE Internet Things J. 2020, 7, 9530–9539. [Google Scholar] [CrossRef]
  9. Geyer, R.C.; Klein, T.; Nabi, M. Differentially private federated learning: A client level perspective. arXiv 2017, arXiv:1712.07557. [Google Scholar] [CrossRef]
  10. Owen, G. Game Theory; Emerald Group Publishing: New Delhi, India, 2013. [Google Scholar]
  11. Hofbauer, J.; Sigmund, K. Evolutionary game dynamics. Bull. Am. Math. Soc. 2003, 40, 479–519. [Google Scholar] [CrossRef]
  12. Sandholm, W.H. Evolutionary game theory. In Complex Social and Behavioral Systems: Game Theory and Agent-Based Models; Springer: Berlin/Heidelberg, Germany, 2020; pp. 573–608. [Google Scholar] [CrossRef]
  13. Szabó, G.; Fath, G. Evolutionary games on graphs. Phys. Rep. 2007, 446, 97–216. [Google Scholar] [CrossRef]
  14. Zheng, C.; Wang, L.; Xu, Z.; Li, H. Optimizing Privacy in Federated Learning with MPC and Differential Privacy. In Proceedings of the 2024 3rd Asia Conference on Algorithms, Computing and Machine Learning, Shanghai, China, 22–24 March 2024; pp. 165–169. [Google Scholar] [CrossRef]
  15. Khan, A.; Thij, M.t.; Thuijsman, F.; Wilbik, A. Incentive Allocation in Vertical Federated Learning Based on Bankruptcy Problem. arXiv 2023, arXiv:2307.03515. [Google Scholar] [CrossRef]
  16. Talajić, M.; Vrankić, I.; Pejić Bach, M. Strategic Management of Workforce Diversity: An Evolutionary Game Theory Approach as a Foundation for AI-Driven Systems. Information 2024, 15, 366. [Google Scholar] [CrossRef]
  17. He, Y.; Wu, H.; Wu, A.Y.; Li, P.; Ding, M. Optimized shared energy storage in a peer-to-peer energy trading market: Two-stage strategic model regards bargaining and evolutionary game theory. Renew. Energy 2024, 224, 120190. [Google Scholar] [CrossRef]
  18. Wang, Y.; Zhou, Y.; Huang, P.Q. A novel incentive mechanism for federated learning over wireless communications. IEEE Trans. Artif. Intell. 2024, 5, 5561–5574. [Google Scholar] [CrossRef]
  19. Lim, W.Y.B.; Ng, J.S.; Nie, J.; Hu, Q.; Xiong, Z.; Niyato, D.; Miao, C. Evolutionary Model Owner Selection for Federated Learning with Heterogeneous Privacy Budgets. In Proceedings of the ICC 2022-IEEE International Conference on Communications, Seoul, Republic of Korea, 16–20 May 2022; IEEE: New York, NY, USA; pp. 980–985. [Google Scholar] [CrossRef]
  20. Du, J.; Jiang, C.; Chen, K.C.; Ren, Y.; Poor, H.V. Community-structured evolutionary game for privacy protection in social networks. IEEE Trans. Inf. Forensics Secur. 2017, 13, 574–589. [Google Scholar] [CrossRef]
  21. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar] [CrossRef]
  22. Yang, C.; Qi, J.; Zhou, A. Wasserstein Differential Privacy. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 16299–16307. [Google Scholar] [CrossRef]
  23. Triastcyn, A.; Faltings, B. Federated learning with bayesian differential privacy. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; IEEE: New York, NY, USA; pp. 2587–2596. [Google Scholar] [CrossRef]
  24. Triastcyn, A.; Faltings, B. Bayesian differential privacy for machine learning. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 9583–9592. [Google Scholar]
  25. Dong, J.; Roth, A.; Su, W.J. Gaussian differential privacy. J. R. Stat. Soc. Ser. Stat. Methodol. 2022, 84, 3–37. [Google Scholar] [CrossRef]
  26. Bu, Z.; Dong, J.; Long, Q.; Su, W.J. Deep learning with gaussian differential privacy. Harv. Data Sci. Rev. 2020, 2020, 10–1162. [Google Scholar] [CrossRef]
  27. Chuanxin, Z.; Yi, S.; Degang, W. Federated learning with Gaussian differential privacy. In Proceedings of the 2020 2nd International Conference on Robotics, Intelligent Control and Artificial Intelligence, Shanghai, China, 17–19 October 2020; pp. 296–301. [Google Scholar] [CrossRef]
  28. Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends® Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
  29. Liu, T.; Di, B.; Wang, B.; Song, L. Loss-privacy tradeoff in federated edge learning. IEEE J. Sel. Top. Signal Process. 2022, 16, 546–558. [Google Scholar] [CrossRef]
  30. ur Rehman, M.H.; Salah, K.; Damiani, E.; Svetinovic, D. Towards blockchain-based reputation-aware federated learning. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Beijing, China, 27–30 April 2020; IEEE: New York, NY, USA; pp. 183–188. [Google Scholar] [CrossRef]
  31. Wang, Y.; Su, Z.; Luan, T.H.; Li, R.; Zhang, K. Federated learning with fair incentives and robust aggregation for UAV-aided crowdsensing. IEEE Trans. Netw. Sci. Eng. 2021, 9, 3179–3196. [Google Scholar] [CrossRef]
  32. Weng, J.; Weng, J.; Zhang, J.; Li, M.; Zhang, Y.; Luo, W. Deepchain: Auditable and privacy-preserving deep learning with blockchain-based incentive. IEEE Trans. Dependable Secur. Comput. 2019, 18, 2438–2455. [Google Scholar] [CrossRef]
  33. Zou, Y.; Feng, S.; Niyato, D.; Jiao, Y.; Gong, S.; Cheng, W. Mobile device training strategies in federated learning: An evolutionary game approach. In Proceedings of the 2019 International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Atlanta, GA, USA, 14–17 July 2019; IEEE: New York, NY, USA; pp. 874–879. [Google Scholar] [CrossRef]
  34. Cheng, Z.; Wang, B.; Pan, Y.; Liu, Y. Strategic Analysis of Participants in BCFL-Enabled Decentralized IoT Data Sharing. Mathematics 2023, 11, 4520. [Google Scholar] [CrossRef]
  35. Hammoud, A.; Mourad, A.; Otrok, H.; Wahab, O.A.; Harmanani, H. Cloud federation formation using genetic and evolutionary game theoretical models. Future Gener. Comput. Syst. 2020, 104, 92–104. [Google Scholar] [CrossRef]
  36. Houda, Z.A.E.; Brik, B.; Ksentini, A.; Khoukhi, L.; Guizani, M. When Federated Learning Meets Game Theory: A Cooperative Framework to Secure IIoT Applications on Edge Computing. IEEE Trans. Ind. Inform. 2022, 18, 7988–7997. [Google Scholar] [CrossRef]
  37. Weibull, J.W. Evolutionary Game Theory; MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
  38. Hofbauer, J.; Sandholm, W.H. Stable games and their dynamics. J. Econ. Theory 2009, 144, 1665–1693. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.