Next Article in Journal
Ranking of Autonomous Technologies for Sustainable Logistics Activities in the Confectionery Industry
Previous Article in Journal
Uncertain Numbers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convergence Analysis for Differentially Private Federated Averaging in Heterogeneous Settings †

1
Fujian Key Laboratory of Communication Network and Information Processing, Xiamen University of Technology, Xiamen 361024, China
2
National Key Laboratory of Wireless Communications, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
This article is an expanded version of a paper entitled “Secure federated averaging algorithm with differential privacy”, which was presented at IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Espoo, Finland, 17–20 September 2020.
Mathematics 2025, 13(3), 497; https://doi.org/10.3390/math13030497
Submission received: 15 January 2025 / Revised: 26 January 2025 / Accepted: 27 January 2025 / Published: 2 February 2025
(This article belongs to the Section E1: Mathematics and Computer Science)

Abstract

:
Federated learning (FL) has emerged as a prominent approach for distributed machine learning, enabling collaborative model training while preserving data privacy. However, the presence of non-i.i.d. data and the need for robust privacy protection introduce significant challenges in theoretically analyzing the performance of FL algorithms. In this paper, we present novel theoretical analysis on typical differentially private federated averaging (DP-FedAvg) by judiciously considering the impact of non-i.i.d. data on convergence and privacy guarantees. Our contributions are threefold: (i) We introduce a theoretical framework for analyzing the convergence of DP-FedAvg algorithm by considering different client sampling and data sampling strategies, privacy amplification and non-i.i.d. data. (ii) We explore the privacy–utility tradeoff and demonstrate how client strategies interact with differential privacy to affect learning performance. (iii) We provide extensive experimental validation using real-world datasets to verify our theoretical findings.

1. Introduction

With the rapid growth of the Internet of Things (IoT) in recent years, the number of intelligent devices has surged globally. Many of these devices are equipped with various sensors and increasingly powerful hardware, enabling them to collect and process data at unprecedented scales [1,2,3]. However, this growth has been accompanied by rising concerns about data privacy. Federated learning (FL) has emerged as a key solution in this context, which enables collaborative machine learning (ML) by allowing a parameter server (PS) to coordinate the training of a global model across distributed clients, thus eliminating the need to centralize sensitive data [4,5]. However, FL systems may involve millions of clients, and clients collect their own data based on their local environments and usage patterns; the size and distribution of local datasets vary significantly across clients [6]. This variation leads to data heterogeneity, as data generated by different clients are typically non-independent and identically distributed (non-i.i.d.) [7]. Data heterogeneity is also reflected in the diverse dynamics exhibited by different devices and their communication network.
Federated Averaging (FedAvg), proposed by [5], is one of the most popular FL algorithms. In FedAvg, a small subset of clients is randomly selected to perform local model updates at each communication round, which are then aggregated by the PS to update the global model. Due to its simplicity and effectiveness, FedAvg serves as the foundation for most subsequent FL algorithms. Despite its widespread adoption, the convergence behavior of FedAvg under data heterogeneity remains poorly understood. Theoretical results, such as those in [8,9], suggest that data heterogeneity significantly impacts FedAvg, requiring many more communication rounds to converge as client gradients diverge. These findings are consistent with observations on synthetic or artificially partitioned datasets [10,11]. However, in many real-world FL tasks, FedAvg often performs remarkably well [5], contrary to these theoretical predictions. Furthermore, advanced methods designed to address data heterogeneity, such as SCAFFOLD [12], demonstrate significant advantages on synthetic heterogeneous datasets but exhibit similar performance to FedAvg on realistic datasets. This inconsistency between theory and real-world performance raises questions about the practical relevance of data heterogeneity and the adequacy of existing theoretical models for real-world FL systems.
On the other hand, FedAvg still faces significant challenges, particularly in ensuring robust privacy protections. Data privacy in FL is inherently at risk during the exchange of model parameters between clients and the PS, as adversaries may attempt to infer sensitive information from these exchanges [13]. Differential privacy (DP) has emerged as a key tool to address this risk by introducing calibrated noise into the learning process [14], offering mathematically rigorous privacy guarantees. However, the added DP noise can negatively affect model accuracy and algorithm convergence, requiring a careful balance between privacy and utility [15,16].
Currently, there are numerous works focused on analyzing the impact of data heterogeneity in FedAvg [17,18]. However, to the best of our knowledge, none of the existing works have provided a complete theoretical analysis on FedAvg for handling data heterogeneity together with the consideration of client and data sampling strategies as well as privacy amplification analysis, hence motivating us to develop an advanced theoretical framework for a DP-based FedAvg algorithm (DP-FedAvg) over non-i.i.d. data.

1.1. Related Work

This section provides a focused review of the literature to contextualize our work within the broader scope of FL research:
  • FL in the presence of data heterogeneity: A substantial body of research has tackled the challenges posed by non-i.i.d. data in FL. Studies such as [6,19] indicate that the convergence rate of FL models in non-i.i.d. settings is significantly lower than in i.i.d. scenarios. These works also provide theoretical analyses suggesting that effective client selection schemes can improve convergence rates in non-i.i.d. scenarios. Many existing FL frameworks [5,11,20] rely on random client selection strategies [19]. Additionally, some works [21,22] consider device heterogeneity by taking into account factors like wireless channel capacity and computational power. However, a unified theoretical analysis that examines both non-i.i.d. data effects and client sampling strategies in a single framework remains unexplored.
  • FL with privacy considerations: Privacy protection has become a critical focus in the FL system. The works [23,24] have explored DP as a robust mechanism to protect data privacy in FL, demonstrating its effectiveness in achieving privacy guarantees without incurring excessive computational overhead. Numerous privacy protection techniques based on DP algorithm in FL have been investigated, including DP-SGD [25], DP-FedAvg [1,26] and the DP-based primal-dual method (DP-PDM) [27], where additive noise is applied to local gradients/models. Notably, the impact of this additive noise on learning performance can be quantified through convergence analysis, and total privacy loss can be readily tracked during the training process. However, few of these studies consider privacy amplification techniques [28,29] to reduce the adverse effect of DP noise, which could potentially improve the privacy–utility tradeoff.
To the best of our knowledge, none of the existing works have fully explored the interplay between DP noise and non-i.i.d. data on the convergence of the DP-FedAvg algorithm. Although various attempts have been made to integrate DP into FL algorithms, such as DP-SGD [1,26], DP-Prox [30] and DP-SCAFFOLD [31], the privacy–utility tradeoffs achieved by these methods remain suboptimal in non-i.i.d. settings, especially, they did not include the privacy amplification analyses. This poor privacy–utility tradeoff can be attributed to both data heterogeneity and the added DP noise, as existing FL works do not comprehensively integrate DP noise, privacy amplification, data heterogeneity, client selection and data sampling strategies within one work. This work addresses these gaps, providing robust privacy protections through DP while offering strong theoretical guarantees.

1.2. Contributions

The main contributions are summarized as follows:
  • Theoretical framework: We propose a novel theoretical framework to analyze the convergence of DP-FedAvg under various client sampling strategies, with a particular focus on the impact of privacy protection and non-i.i.d. data. This framework enables us to quantify how client sampling methods and data heterogeneity influence the algorithm’s convergence rate and privacy guarantees.
  • Privacy–utility tradeoff: We explore the tradeoff between privacy protection and learning performance, taking into account the effects of privacy amplification. Our theoretical results demonstrate how various system parameters influence this tradeoff and highlights that the privacy protection level and gradient clipping play a crucial role in determining the privacy–utility tradeoff in practical FL systems.
  • Empirical validation: We empirically validate our theoretical findings through extensive experiments on real-world datasets, demonstrating the behavior of DP-FedAvg under non-i.i.d. data conditions.
Synopsis: For ease of the ensuing presentation, all the mathematical notations used are listed in Table 1. Section 2 introduces the preliminaries of FL, DP and the existing DP-FedAvg algorithm. Section 3 provides the convergence analysis and privacy analysis for the DP-FedAvg algorithm. Experimental results are presented in Section 4, and Section 5 concludes the paper.

2. Preliminaries

2.1. Federated Learning

We consider a vanilla FL system that consists of one parameter server (PS) and N clients in which each client has its local dataset D k : = x k , j , y k , j , k [ N ] , j n k , where n k and n = k = 1 N n k denote the number of k-th client’s training samples and its number of training samples, respectively. x k , j denotes the training data sample and y k , j is the corresponding label, while [ N ] denotes the set { 1 , 2 , , N } . The FL algorithm is to solve the following distributed optimization problem:
min w F ( w ) k = 1 N p k F k w
where w is the global model and p k = n k / n is the weight of the k-th client and k = 1 N p k = 1 ; the F k is the local objective function for client k, which is defined by
F k w = 1 n k j = 1 n k w ; x k , j , y k , j ,
where ( · ) is a user-specified loss function. Problem (1) can be solved using the well-known FedAvg algorithm [5]. Specifically, FedAvg involves T communication rounds, and each round contains the following steps:
1.
First, at the beginning of round t { 0 , , T 1 } , the PS selects a subset of clients S t [ N ] to participate and sends them the latest global model w t .
2.
Each client i S t initializes its local model w t + 1 k to be the global model w t and then performs Q iterations of stochastic gradient descent (SGD) on its local dataset.
w t + 1 k = w t ,
w t + j k = w t + j 1 k η t + j 1 F k w t + j 1 k , B t + j 1 k , j [ Q ] , k S t ,
where η t + j 1 represents the learning rate, and B t + j 1 k D k is the mini-batch data with | B t + j 1 k | = b ; F k w t + j 1 k , B t + j 1 k denotes the stochastic gradient.
3.
Each client i S t uploads the final local model update w t + Q k to the PS.
4.
The PS aggregates the local model updates from all participating clients and improves the global model as
w t + Q = k S t p k w t + Q k .
The same procedure repeats for the next round until the algorithm converges.

2.2. Differential Privacy

Although FedAvg avoids the direct information leakage by keeping the data locally, the intermediate updates exchanged during the collaboration process such as w t + Q k and w t + Q could still leak private information about the local data as demonstrated in recent advanced attacks such as model inversion attacks and membership attacks [32]. Figure 1 depicts the FL system model under attacks from an adversary. The adversary in this context can be either an “honest-but-curious” aggregation PS or clients within the system. The aggregation PS is assumed to honestly follow the designed training protocol but may be curious about clients’ private data and attempt to infer it from the shared messages. To provide strong privacy protection for local data, we adopt ( ϵ , δ ) -DP, which is defined as follows.
Definition 1
( ( ϵ , δ ) -DP [14]). A randomized mechanism M satisfies ( ϵ , δ ) -DP if for any two adjacent datasets D and D differing in only one record, and for any subsets of outputs S range ( M ) ,
P M ( D ) S exp ( ϵ ) · P M D S + δ ,
where ϵ > 0 accounts for privacy protection level, and  δ 0 is the probability threshold to break ( ϵ , 0 ) -DP.
For a query function g : X R , the ( ϵ , δ ) -DP mechanism can be implemented by adding artificial Gaussian noise to function g as follows [14]:
M ( D ) g ( D ) + N 0 , Δ 2 · σ 2 ,
where Δ is the sensitivity of function g, defined by   
Δ = max D , D g ( D ) g D ,
in which D , D X are neighboring datasets. σ 2 denotes the minimal required “noise scale” for achieving ( ϵ , δ ) -DP given by the following lemma in [14].
Lemma 1.
Suppose that the query function g accesses the dataset D via randomized mechanism M . Then, M satisfies ( ϵ , δ ) -DP if
σ 2 2 Δ 2 ln ( 1.25 / δ ) ϵ 2 .
Definition 2
(Privacy loss [14]). Suppose that a randomized mechanism M satisfies ( ϵ , δ ) -DP. Let D and D be two neighboring datasets and o be a possible output of M ( D ) and M ( D ) . The privacy loss is defined by
c ( o ) = ln P M ( D ) = o P M ( D ) = o .
It has been known that, running on a randomly generated subset of a dataset, the DP mechanism can yield stronger privacy protection than running on the entire dataset. This fact implies that the DP noise required for achieving a predefined DP level can be reduced when partial data are randomly selected at each local update. For the ease of later analysis, we restate the privacy amplification theorem as follows [7].
Theorem 1.
Suppose that a mechanism M is ( ϵ , δ ) -DP over a given dataset D with size | D | = n . Consider the subsampling mechanism that outputs a random sample uniformly over all subsets D s D with size | D s | = b . Then, when ϵ 1 , executing M mechanism on the subset D s guarantees ( ϵ , δ ) -DP, where ϵ and δ are given by
ϵ = min 2 q ϵ , ϵ ,
δ = q δ ,
where q = b / n is the data sampling ratio when data are sampled without replacement.
According to Theorem 1, the privacy would be amplified when q 1 / 2 . Note that the privacy amplification is pervasively adopted in the existing literature on FL [1,7,23] since only a small portion of data are currently being used in local SGD.

2.3. The Typical DP-FedAvg Algorithm

In this subsection, we present the typical DP-FedAvg, which is summarized in Algorithm 1. The main steps follow the FedAvg framework outlined in Section 2.1. The key distinctions between DP-FedAvg and FedAvg are twofold: ( i ) Gaussian noise is added to local gradients to ensure robust privacy protection and ( i i ) gradient clipping is applied to the locally updated model. Specifically, to protect the privacy of the local model w t k , we apply DP by adding Gaussian noise ξ t + 1 k N ( 0 , ( σ t + 1 k ) 2 I d ) to the local gradient as follows:
w t + 1 k = w t k η t g ˜ t k + ξ t + 1 k ,
where g ˜ t k is the clipped local gradient, defined by
g ˜ t k = g t k / max 1 , g t k G ,
in which g t k is the local stochastic gradient, given as follows:
g t k = F k w t k , B t k .
It is important to note that determining the 2 -sensitivity requires bounded local gradients. Therefore, gradient clipping is applied to the locally updated model. Specifically, if g t + 1 k G , the gradient is scaled down to have a norm of G ; otherwise, it remains unchanged. Gradient clipping is a standard operation in SGD with DP, particularly for determining sensitivity in FL systems.
Algorithm 1 Proposed DP-FedAvg algorithm
1:
Input : Initial values of w 0 1 , , w 0 N at the clients, and initial value of w 0 at the server.
2:
for   t = 0 , 1 , , T 1   do
3:
    Client side :
4:
   for  k S t in parallel do
5:
     Calculate local gradients: g t k via (15).
6:
     Perform gradient clipping: g ˜ t k via (14).
7:
     if  t + 1 z Q , z = 1 , 2 ,  then
8:
        Update w t + 1 k = w t k η t g t k
9:
     else if   t + 1 = z Q , z = 1 , 2 , then
10:
      Update  w t + 1 k via (13).
11:
    end if
12:
     Server side :
13:
    Update w t + 1 = k = 1 K p k w t + 1 k .
14:
    Update the subset of clients S t + 1 [ N ] through uniformly sampling, and broadcast x 0 t to all clients.
15:
   end for
16:
end for

3. Theoretical Analysis for DP-FedAvg

In this section, we give the formal privacy guarantee and rigorous convergence analysis of DP-FedAvg. Before stating our results, we make the following assumptions.

3.1. Assumptions

Assumption 1.
F 1 , F 2 ,…, F N are all L-smooth: for all w and v , F k ( v ) F k ( w ) + ( v w ) F k ( w ) + L 2 v w 2 , k [ N ] .
Assumption 2.
F 1 , F 2 ,…, F N are all μ-strong convex: for all w and v , F k ( v ) F k ( w ) + ( v w ) F k ( w ) + μ 2 v w 2 , k [ N ] .
Assumption 3.
Denote B t k as the mini-batch data from the dataset in client k with size | B t k | = b ; then, the variance of stochastic gradients in each client is bounded, that is, E F k ( w t k , B t k ) F k ( w t k ) 2 ϕ k 2 / b , k [ N ] .
Assumption 4.
The local gradient is bounded, i.e., F k ( w t k , B t k ) G .
Quantifying the degree of non i . i . d . data . Inspired by [6], we quantify data heterogeneity through the degree of non-i.i.d. data, defined as
Γ = F 🟉 k = 1 K p k F k 🟉 ,
where F 🟉 and F k 🟉 represent the minimum values of F and F k , respectively. A larger value of Γ indicates a higher degree of data heterogeneity. Specifically, when the data are i.i.d., Γ approaches zero.

3.2. Privacy Analysis

We firstly estimate the sensitivity on client perspective. The 2 -norm sensitivity for each local uploaded model is given in following Lemma.
Lemma 2.
Supposed that the Assumptions 1–4 hold and mod ( t + 1 , Q ) = 0 . The 2 -norm sensitivity is given by
t + 1 = max D k , D k w t + 1 , D k k w t + 1 , D k k = 4 Q G η t + 1 ,
where w t + 1 , D k k and w t + 1 , D k k denote the local model parameters updated from two neighboring datasets D k and D k , respectively.
Proof. 
See Appendix A. □
Lemma 3.
According to the privacy amplification theorem in Theorem 1, the required privacy protection level to guarantee ( ϵ , δ ) -DP is given by
ϵ = ϵ 2 q ,
δ = δ q ,
where q is the data sampling ratio.
By Lemma 2 and Lemma 3, the total DP noise scale for t-th iteration is
σ t 2 = k = 1 K p k ( σ t k ) 2 = 128 η t 2 Q 4 b 2 G 2 n ϵ 2 k = 1 K log ( 1.25 Q b / n k δ ) n k .
One can observe from (20) that the increasing Q promptly increases the noise scale, which is consistent the result in Lemma 2. According to (20), the DP noise scale is primarily determined by the privacy protection level ϵ , the gradient clipping threshold G and other algorithm-specific parameters, such as Q, b and η t . Thus, in practical FL systems, achieving a balance between privacy protection and learning performance relies on effectively controlling the key parameters ϵ and G .
Then, the accumulated noise over the entire training process with total T / Q communication rounds can be obtained by
σ t o t a l 2 = t = 1 T / Q k = 1 K p k ( σ t k ) 2 = 128 T Q 3 b 2 G 2 n ϵ 2 t = 1 T / Q η t 2 k = 1 K log ( 1.25 Q b / n k δ ) n k .
By considering privacy amplification, the total privacy loss after T communication rounds has been explored in [27]. For the sake of clarity and ease of later analysis, we restate it here as follows.
Theorem 2.
Suppose that any client i , i [ N ] in Algorithm 1 is uniformly sampled with probability p i and each communication round guarantees ( ϵ , δ ) -DP. The total privacy loss for client i after T communication rounds satisfies
ϵ ¯ i = c 0 q i ϵ p i T 1 q i , i [ N ] ,
where c 0 > 0 is a constant and q i is given by
q i = Q b | D i | , d a t a s a m p l i n g W O R , 1 ( 1 1 | D i | ) Q b , d a t a s a m p l i n g W R ,
where WOR and WR denote data sampling without and with replacement, respectively.
Remark 1.
By Theorem 2, the total privacy loss is significantly impacted by data sampling strategies. Specifically, the total privacy loss for data sampling with replacement is smaller than that for data sampling without replacement due to 1 ( 1 1 / | D i | ) Q b Q b / | D i | .

3.3. Convergence Analysis for DP-FedAvg

In this section, we analyze the convergence performance of the DP-FedAvg algorithm. Let T denote the total number of iterations and w 🟉 the optimal solution of F ( w ) . The convergence performance is measured by E F ( w T ) F ( w 🟉 ) , where the expectation is taken over the algorithm’s randomness. This metric represents the gap between the achieved objective value and the optimal objective value after T iterations. Notably, when E F ( w T ) F ( w 🟉 ) = 0 , the algorithm converges to the optimal solution.
Theorem 3.
For the full-participation case, assume that Assumption 1, Assumption 2 and Assumption 3 hold. Let κ = L / μ , α = max { 8 κ , Q } , and the learning rate η t = 2 / ( μ ( α + t ) ) . Then, the following inequality holds:
E F ( w ¯ t ) F 🟉 2 κ t + α A + B μ + 2 L w 0 w 🟉 2 ,
where
A = k = 1 K p k 2 ϕ k 2 / b + 8 ( Q 1 ) 2 G 2 + 6 L Γ ,
B = 8 d Q 4 b 2 G 2 n L 2 ϵ 2 log ( n ) log ( 1 . 25 2 Q 2 b 2 / n δ 2 ) 2 ( Q 1 ) 2 + 1 .
Proof. 
See Appendix B. □
Remark 2.
In Theorem 3, since w 0 w 🟉 2 1 μ 2 F ( w ) F ( w 🟉 ) 2 4 G 2 μ 2 , the dominating term in (24) is O A + B μ T . Therefore, the DP-FedAvg algorithm achieves a convergence rate in the order of
O k = 1 K p k 2 ϕ k 2 / b + Q 2 G 2 + L Γ μ T + d Q 4 b 2 G 2 log ( n ) log ( Q 2 b 2 / n δ 2 ) μ T n L 2 ϵ 2 .
Therefore, an increase in Q can significantly increase the noise scale, thereby deteriorating the convergence performance.
Next, we present the convergence analysis by considering client sampling. Inspired by [6], we make the following assumption.
Assumption 5.
Assume the total number of clients is N and the size of the selected client set is | S t | = K . If we uniformly sample the participating clients with replacement with sampling probabilities p 1 , , p N , then the aggregation step on the server side performs w t = 1 K k S t w t k . Otherwise, if we uniformly sample the participated clients without replacement from [ N ] , then the aggregation step on the PS performs w t = N K k S t p k w t k .
Theorem 4.
Suppose that Assumptions 1, 2, 3 and 5 hold. Let κ, α, η t , A and B be defined in Theorem 3. Then, the following inequality holds:
E F ( w ¯ t ) F 🟉 2 κ t + α A + B + H μ + 2 L w 0 w 🟉 2 ,
if sampling the clients with replacement:
H = 4 K Q 2 G 2 ,
and if sampling the clients without replacement:
H = N K N 1 4 K Q 2 G 2 .
Proof. 
See Appendix D. □
Theorem 4 provides an upper bound of the average total local SGDs over t iterations. A smaller value indicates a higher convergence rate and a lower cost function value in (28). Based on Theorems 3 and 4, we have following remarks.
Remark 3. 
(The impact of Q on communication efficiency) By denoting T ε as the number of required iterations for DP-FedAvg to achieve an ε accuracy. Based on the result in (27), the number of required communication rounds T ε / Q is
T ε Q O d b 2 G 2 log ( n ) log ( Q 2 b 2 / n δ 2 ) μ T n L 2 ϵ 2 Q 3 + G 2 Q + k = 1 K p k 2 ϕ k 2 / b + L Γ Q .
From (31), we observe that the communication round T ε / Q is the function of Q, and T ε / Q first decreases and then increases, i.e., there exists an optimal Q such that the communication cost is optimal. Furthermore, if Q is too large, the w k will promptly achieve local optimal, making FedAvg behave like one-shot average [33]. However, the one-shot averaging does not perform well in the non-i.i.d. data case as demonstrated in [6].
Remark 4. 
(The impact of Q on learning performance) The DP noise scale increases significantly with larger values of Q, which inevitably degrades learning performance. However, a larger Q can also reduce the number of communication rounds, as stated in Remark 3. Therefore, there exists an optimal value of Q that balances learning performance and communication efficiency.
Remark 5. 
(The impact of gradient clipping bound level G ) The convergence performance is influenced by gradient clipping level G . However, limiting the gradient norm has the following shortcomings: ( i ) Gradient clipping destroys the unbiasedness of the gradient estimate. ( i i ) If G is too small, the average of the clipped gradients may point in a direction significantly different from the true gradient, potentially leading to poor convergence. ( i i i ) If G is too large, the gradients will carry more DP noise, as G determines the sensitivity.

4. Numerical Experiments

4.1. Experimental Setting

In this section, we evaluate the performance of DP-FedAvg under a non-i.i.d. data setting. Denote the ( w ; x i ) as the prediction model with parameter w = ( W , b ) , then the form for logistic regression and softmax regression are ( w ; x i , y i ) = logisitc ( W x + b ) and ( w ; x i , y i ) = softmax ( W x + b ) , respectively. The loss function for both problems is given by
F k ( w ) = 1 n k j = 1 n k CrossEntropy ( w ; x j ) , y j .
Note that this is a convex and smooth problem. The testing accuracy (learning accuracy over the testing data) is evaluated as the learning performance in the experiments.
Datasets: Two benchmark datasets, Adult [34] and MNIST [35], were considered for performance evaluation. The MNIST dataset consists of 60,000 training samples and 10,000 testing samples, while the Adult dataset consists of 32,561 training samples and 16,281 testing samples. To simulate the FL system, we distributed all training samples across N = 100 clients in a non-i.i.d. manner, ensuring that each client possessed partial labels for the data:
  • For the MNIST dataset, following the heterogeneous data partition method in [6], each client was allocated data samples of only four different labels with | D i | = 600 . This led to a high degree of non-i.i.d. datasets among clients.
  • For the Adult dataset, following [1], all the training samples were uniformly distributed among a total N = 100 clients such that each client only contained data from one class with | D i | = 325 .
Parameter setting : For all experiments, we maintained the following parameter settings: K = 10 , b = 10 and δ = 10 5 . The choice of the learning rate η t depended on the dataset under consideration.
For the MNIST dataset, the local learning rate was set to η t = 0.04 / 1 + t , while for the Adult dataset, η t = 0.01 / 1 + t . Following the rule of gradient clipping in [25], we selected the value of G by taking the median of the norms of the unclipped gradients over the training process.

4.2. Impact of the Privacy Protection Level ϵ

To validate the learning performance at various privacy protection levels, the parameters were set as K = 10 , Q = 5 and δ = 10 5 , and the testing accuracy was computed with respect to communication rounds ( T / Q ). We compared the accuracy with various values of ϵ under both i.i.d. and non-i.i.d. scenarios. As illustrated in Figure 2, both the MNIST and Adult datasets provided strong privacy protection without a significant decrease in testing accuracy. Moreover, we find that MNIST has a higher tolerance to DP noise than the Adult dataset. Specifically, the difference in testing accuracy between the non-private and private models for both i.i.d. and non-i.i.d. datasets is much smaller than the corresponding drop in accuracy for the Adult dataset. The reason for the MNIST dataset’s robustness to DP noise compared with the Adult dataset may have been due to the following reasons:
  • Impact of non-i.i.d. data: The presence of non-i.i.d. data across clients in FL amplifies the impact of DP noise. The Adult dataset, which contains categorical and numerical features, is more sensitive to such heterogeneity. Conversely, the MNIST dataset, consisting of grayscale images of handwritten digits with simpler patterns, is inherently more resilient to non-i.i.d. data settings.
  • Inherent data structure: The MNIST dataset benefits from the spatial coherence of pixel intensities, where patterns such as the shapes of digits remain discernible when DP noise is introduced. This property of image data enables models to identify meaningful features despite perturbations. On the other hand, the Adult dataset, being tabular with categorical and numerical features, lacks such spatial structure. DP noise in tabular data can obscure critical relationships between features and labels, making it harder for the FL model to extract meaningful patterns under noisy conditions.
Figure 2. The testing accuracy versus communication rounds of the DP-FedAvg algorithm on various privacy protection levels under Adult and MNIST datasets.
Figure 2. The testing accuracy versus communication rounds of the DP-FedAvg algorithm on various privacy protection levels under Adult and MNIST datasets.
Mathematics 13 00497 g002

4.3. Impact of Local Epoch Length Q

In this experiment, we compared the learning performance under the parameter settings ϵ = 0.5 , K = 10 and δ = 10 5 for various values of Q. As shown in Figure 3, we observed that both excessively large and excessively small values of Q reduced the convergence rate on the Adult dataset. However, for the MNIST dataset, a large Q had a negligible effect on convergence. This may have been due to two reasons: ( i ) The optimal value of Q for MNIST is significantly larger than for the Adult dataset, and ( i i ) MNIST exhibits a higher tolerance to noise compared to the Adult dataset. Although a large Q alleviates the communication burden, it substantially increases the noise scale, thereby degrading learning performance as discussed in Remark 4. Additionally, our simulation results indicate that the optimal value of Q is sensitive to both the parameter settings and the training dataset. Furthermore, for both datasets, the convergence performance on i.i.d. data is better than on non-i.i.d. data under the same parameter settings.

5. Conclusions and Future Works

In this paper, we conducted a thorough theoretical analysis of DP-FedAvg under different client and data sampling strategies, with a particular focus on the impact of non-i.i.d. data on the convergence and privacy protection of FL algorithms. Our work fills a significant gap in the literature by providing a unified framework that integrates data heterogeneity with DP. Through our analysis, we demonstrated that the DP-FedAvg algorithm maintains robustness toward data heterogeneity and DP noise, achieving a sublinear convergence rate for convex FL problems. Additionally, our convergence and privacy analyses offer valuable insight and perspectives into how various system parameters in the FL framework influence learning performance, which can be used as guidelines for the practical FL algorithm design. Our theoretical findings are supported by extensive numerical experiments on real-world datasets, which are consistent with our theoretical findings.
While our theoretical analysis provides useful insights for DP-FedAvg under the data heterogeneity scenario, several promising directions remain for future research. First, our work is based on the typical star network framework, consisting of one server and multiple clients. A promising direction would be to explore hierarchical FL systems, where multiple servers collaborate to train a global model. This framework could involve more clients and larger datasets, potentially improving learning performance. Second, another promising avenue for future work involves examining the tradeoffs between privacy, utility and computational overhead when implementing advanced methods in resource-constrained FL systems, such as those in mobile edge computing environments. Finally, while the DP mechanism offers strong privacy guarantees, more sophisticated approaches could be employed to further enhance privacy protection. Techniques such as homomorphic encryption could be integrated to strengthen the security of DP-FedAvg. Addressing these challenges and exploring innovative theoretical analyses will be an exciting and valuable direction for future research.

Author Contributions

Methodology, Y.L. and S.W.; Formal analysis, Y.L., S.W. and Q.W.; Investigation, Y.L. and Q.W.; Writing—original draft preparation, Y.L.; Writing—review and editing, Y.L., S.W. and Q.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proof of the Lemma 2

Proof. 
Since there is only one different record between D k , D k , and each client performs Q steps of SGD per communication round, the local update scheme is w t + 1 , D k k = w t + 1 Q , D k k τ = 0 Q 1 η t + 1 Q + τ g t + 1 Q + τ , D k k . Thus, we have
w t + 1 , D k k w t + 1 , D k k 2 = τ = 0 Q 1 η t + 1 Q + τ g t + 1 Q + τ , D k k τ = 0 Q 1 η t + 1 Q + τ g t + 1 Q + τ , D k k 2 ( a ) 4 η t + 1 2 τ = 0 Q 1 g t + 1 Q + τ , D k k g t + 1 Q + τ , D k k 2 4 Q 2 η t + 1 2 τ = 0 Q 1 g t + 1 Q + τ , D k k g t + 1 Q + τ , D k k 2 ( b ) 16 Q 2 G 2 η t + 1 2 ,
where ( a ) holds due to η t 2 η t + 1 ; ( b ) follows from Assumption 4. □

Appendix B. Proof of Theorem 3

Proof. 
Inspired by the perturbed iterate framework in [6], we define I Q = z Q | z = 1 , 2 , as the set of global synchronization steps. When t + 1 I Q , it is the time that clients run the local model update; otherwise, it is the time step to communication. Then, the update rule of DP-FedAvg can be described as
v t + 1 k = w t k η t F k w t k , B t k if t + 1 I Q w t k η t F k w t k , B t k + ξ t k if t + 1 I Q
w t + 1 k = v t + 1 k if t + 1 I Q k = 1 K p k v t + 1 k if t + 1 I Q
where ξ t k is the Gaussian noise added to gradients, v t + 1 k denotes one-step SGD update from w t k and w t + 1 k can be interpreted as the result after communication steps.
Motivated by [6], we define two virtual sequences, w ¯ t = k = 1 K p k w t k and v ¯ t = k = 1 K p k v t k . For simplicity of analysis, when t + 1 I Q , define
g ¯ t = k = 1 K p k F k w t k + ξ t k ,
g t = k = 1 K p k F k w t k , B t k + ξ t k .
When t + 1 I Q , define
g t = k = 1 K p k F k w t k , B t k ,
g ¯ t = k = 1 K p k F k w t k .
Note that, for both cases, we always have v ¯ t + 1 = w ¯ t η t g t and E [ g t ] = g ¯ t . □

Appendix B.1. Key Lemmas

To clearly convey the proof of Theorem 3, we need the following key lemmas.
Lemma A1.
(one-step SGD) If Assumption 1 and Assumption 2 hold and η t 1 4 L . Then,
w h e n   t + 1 I Q ,
E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + η t 2 E g t g ¯ t 2 + 2 E k = 1 K p k w ¯ t w t k 2 + 6 L η t 2 Γ ,
when t + 1 I Q ,
E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + η t 2 E g t g ¯ t 2 + 2 E k = 1 K p k w ¯ t w t k 2 + 6 L η t 2 Γ + 2 d η t 2 E k = 1 K p k ξ t k 2 .
Lemma A2.
Suppose that Assumption 3 holds. Then,
E g t g ¯ t 2 k = 1 K p k 2 ϕ k 2 / b .
Lemma A3.
Assume η t is non-increasing and η t 2 η t + Q , then,
when t + 1 I Q ,
E k = 1 K p k w ¯ t w t k 2 4 ( Q 1 ) 2 η t 2 G 2 + d k = 1 K p k σ t , k 2 .
When t + 1 I Q ,
E k = 1 K p k w ¯ t w t k 2 4 ( Q 1 ) 2 G 2 η t 2 .
Lemma A4.
Assume η t is non-increasing and η t 1 4 L , then the total noise in each communication round is given by
E k = 1 K p k ξ t k 2 4 Q 4 b 2 G 2 n L 2 ϵ 2 log ( n ) log ( 1 . 25 2 Q 2 b 2 / n δ 2 ) .

Appendix B.2. Completing the Proof of Theorem 3

Proof. 
Let t + 1 = E w ¯ t + 1 w 🟉 2 . Based on the results of Lemma A1–A4, when t + 1 I Q , it follows that
E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + A η t 2 ,
where
A = k = 1 K p k 2 ϕ k 2 / b + 8 ( Q 1 ) 2 G 2 + 6 L Γ .
When t + 1 I Q ,
E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + ( A + B ) η t 2 ,
where
B = 8 d Q 4 b 2 G 2 log ( n ) log ( 1 . 25 2 Q 2 b 2 / n δ 2 ) n L 2 ϵ 2 2 ( Q 1 ) 2 + 1 .
Then, (A16) becomes
t + 1 1 μ η t t + η t 2 ( A + B ) .
For a diminishing stepsize, η t = β t + α for some β > 1 μ and α 0 such that η 0 min { 1 μ , 1 4 L } = 1 4 L and η t 2 η t + Q , this also implies that α Q :
t + 1 1 μ η t t + η t 2 ( A + B ) = 1 μ β t + α ν t + α + β 2 ( t + α ) 2 ( A + B ) 1 μ β t + α ν t + α + β 2 ( A + B ) ( t + α ) 2 t + α 1 ( t + α ) 2 ν + β 2 ( A + B ) ( t + α ) 2 ( μ β 1 ) ν ( t + α ) 2 t + α 1 ( t + α ) 2 1 ν = ν t + α + 1 ,
where ν β 2 ( A + B ) / ( μ β 1 ) . By the L-smoothness of F ( · ) , we have
E F ( w ¯ t ) F 🟉 L 2 E w ¯ t w 🟉 2 = L 2 t L 2 ν t + α .
Note that, from (A20), we have ν α 0 . Thus, we obtain ν = max β 2 ( A + B ) μ β 1 , α 0 . If we choose β = 2 / μ and then η t = 2 μ 1 α + t , where α max { 8 κ , Q } , κ = L μ . The range of α due to η 0 = 2 μ 1 α 1 4 L implies that α 8 κ . Then, we have
E F ( w ¯ t ) F 🟉 2 κ t + α A + B μ + 2 L 0 ,
where if β 2 ( A + B ) μ β 1 α 0 , then
E F ( w ¯ t ) F 🟉 L 2 1 t + α β 2 A μ β 1 = 2 κ t + α A + B μ ;
otherwise, if β 2 ( A + B ) μ β 1 < α 0 , and in this case, α = 8 κ , then
E F ( w ¯ t ) F 🟉 L 2 1 t + α α 0 = 4 κ L 0 t + α .
By combining (A21), (A22) and (A23), we obtain the result. □

Appendix C. Proofs of Key Lemmas in Theorem 3

Appendix C.1. Proof of Lemma A1

Proof. 
Since we have v ¯ t + 1 = w ¯ t η t g t , then
v ¯ t + 1 w 🟉 2 = w ¯ t η t g t w 🟉 2 = w ¯ t η t g t w 🟉 η t g ¯ t + η t g ¯ t 2 = w ¯ t w 🟉 η t g ¯ t 2 A 1 + 2 η t w ¯ t w 🟉 η t g ¯ t , g ¯ t g t + η t 2 g t g ¯ t 2 A 2 .
Then, E w ¯ t w 🟉 η t g ¯ t , g ¯ t g t = 0 due to E [ g t ] = g ¯ t . Then, for the A 1 term, we have
w ¯ t w 🟉 η t g ¯ t 2 = w ¯ t w 🟉 2 + η t 2 g ¯ t 2 2 η t w ¯ t w 🟉 , g ¯ t ( a ) w ¯ t w 🟉 2 + η t 2 k = 1 K p k F k w t k + ξ t k 2 2 η t k = 1 K p k w ¯ t w t k + w t k w 🟉 , F k w t k + ξ t k = w ¯ t w 🟉 2 2 η t k = 1 K p k w t k w 🟉 , F k w t k + ξ t k 2 η t k = 1 K p k w ¯ t w t k , F k w t k + ξ t k + η t 2 k = 1 K p k F k w t k + ξ t k 2 ,
where ( a ) follows from Jensen’s inequality of function · 2 . Then, by the L-smoothness of F k ( · ) , we have F k w t k 2 2 L F k w t k F k 🟉 . Then,
η t 2 k = 1 K p k Π G F k w t k + ξ t k 2 η t 2 k = 1 K p k F k w t k + ξ t k 2 = η t 2 k = 1 K p k F k w t k 2 + 2 F k w t k ξ t k + ξ t k 2 2 L η t 2 k = 1 K p k F k w t k F k 🟉 + η t 2 k = 1 K p k ξ t k 2 + 2 η t 2 k = 1 K p k F k w t k ξ t k ,
where Π G ( · ) denotes the gradient clipping operation. By μ -strong convex of F k ( · ) , we have F k w t k w t k w 🟉 F k w t k F k ( w 🟉 ) + μ 2 w t k w 🟉 2 , then
w t k w 🟉 , F k w t k + G · ξ t k = w t k w 🟉 , F k w t k G w t k w 🟉 , ξ t k F k w t k F k ( w 🟉 ) μ 2 w t k w 🟉 2 G w t k w 🟉 , ξ t k .
By the AM-GM inequality, i.e., 2 x , y ζ x 2 + ζ 1 y 2 , for any ζ > 0 , we have,
2 w ¯ t w t k , F k w t k + ξ t k 1 η t w ¯ t w t k 2 + η t F k w t k + ξ t k 2 1 η t w ¯ t w t k 2 + η t F k w t k 2 + 2 η t F k w t k ξ t k + η t ξ t k 2 1 η t w ¯ t w t k 2 + 2 L η t F k w t k F k 🟉 + 2 η t F k w t k ξ t k + η t ξ t k 2 .
By applying these estimates, we have
w ¯ t η t g ¯ t w 🟉 2 w ¯ t w 🟉 2 + 4 L η t 2 k = 1 K p k F k w t k F k 🟉 2 η t k = 1 K p k F k w t k F k ( w 🟉 ) + 4 η t 2 k = 1 K p k F k w t k ξ t k + 2 η t 2 k = 1 K p k ξ t k 2 + k = 1 K p k w ¯ t w t k 2 μ η t k = 1 K p k w t k w 🟉 2 2 η t k = 1 K p k w t k w 🟉 ξ t k ( a ) 1 μ η t w ¯ t w 🟉 2 + 4 L η t 2 k = 1 K p k F k w t k F k 🟉 2 η t k = 1 K p k F k w t k F k ( w 🟉 ) + 2 η t 2 k = 1 K p k ξ t k 2 + k = 1 K p k w ¯ t w t k 2 ,
where in ( a ) , we apply the fact that w t k is independent of the ξ t k and thus, E [ ξ t k ] = 0 . Furthermore, we also use the inequality k = 1 K p k w t k w 🟉 2 w ¯ t w 🟉 2 , which can be proved by
k = 1 K p k w t k w 🟉 2 k = 1 K p k w t k 2 2 w ¯ t w 🟉 + w 🟉 2 k = 1 K p k w t k 2 2 w ¯ t w 🟉 + w 🟉 2 = w ¯ t w 🟉 2 .
Then, by letting Ψ ( t ) 4 L η t 2 k = 1 K p k F k ( w t k ) F k 🟉 2 η t k = 1 K p k F k ( w t k ) F k ( w 🟉 ) , we have
Ψ ( t ) = 4 L η t 2 k = 1 K p k F k ( w t k ) F k 🟉 2 η t k = 1 K p k F k ( w t k ) F k ( w 🟉 ) = 4 L η t 2 2 η t k = 1 K p k F k ( w t k ) F k ( w 🟉 ) + 4 L η t 2 k = 1 K p k F k ( w 🟉 ) F k 🟉 = 4 L η t 2 2 η t k = 1 K p k F k ( w t k ) F 🟉 + 4 L η t 2 F 🟉 k = 1 K p k F k 🟉 = 4 L η t 2 2 η t k = 1 K p k F k ( w t k ) F 🟉 + 4 L η t 2 Γ .
Since η t 1 4 L , which implies that 4 L η t 2 2 η t 0 , then, k = 1 K p k F k ( w t k ) F 🟉 can be bounded by
k = 1 K p k F k ( w t k ) F 🟉 = k = 1 K p k F k ( w t k ) F k w ¯ t + k = 1 K p k F k w ¯ t F 🟉 ( a ) k = 1 K p k F k w ¯ t , w ¯ t k w ¯ t + F w ¯ F 🟉 ( b ) 1 2 k = 1 K p k η t F k w ¯ t 2 + 1 η t w t k w ¯ t 2 + F w ¯ F 🟉 ( c ) L η t k = 1 K p k F k w ¯ t F k 🟉 1 2 η t k = 1 K p k w t k w ¯ t 2 + F w ¯ F 🟉 ,
where ( a ) follows from the convexity of F k ( · ) ; ( b ) holds due to the AM-GM inequality; in ( c ) , we invoke the L-smoothness of F k ( · ) . Thus, by plugging (A32) into (A31), we have
Ψ ( t ) L η t 2 η t 4 L η t 2 k = 1 K p k F k w ¯ t F k 🟉 + ( 1 2 L η t ) k = 1 K p k w t k w ¯ t 2 2 η t 4 L η t 2 F w ¯ F 🟉 + 4 L η t 2 Γ ( L η t 1 ) 2 η t 4 L η t 2 F w ¯ F 🟉 + ( 1 2 L η t ) k = 1 K p k w t k w ¯ t 2 + 4 L η t 2 + L η t 2 η t 4 L η t 2 Γ ( a ) k = 1 K p k w t k w ¯ t 2 + 6 L η t 2 Γ ,
where ( a ) follows from L η t 1 < 0 and 2 η t 4 L η t 2 2 η t . Then, combining the results of (A29), (A31)–(A33), we have
E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + 2 E k = 1 K p k w ¯ t w t k 2 + 2 p η t 2 E k = 1 K p k ξ t k 2 + 6 L Γ η t 2 .
When t + 1 I Q , we have
g t = k = 1 K p k F k w t k , B t k ,
g ¯ t = k = 1 K p k F k w t k .
By following the same procedure, we obtain the similar result as follows:
E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + 2 E k = 1 K p k w ¯ t w t k 2 + 6 L Γ η t 2 .
Thus, we complete the proof. □

Appendix C.2. Proof of Lemma A2

Proof. 
According to Assumption 3, the variance of stochastic gradients in each client is bounded by ϕ k 2 / b . Then, the A 2 term in (A24) can be bounded by
E g t g ¯ t 2 = E k = 1 K p k Π G F k w t k , B t k + ξ t k k = 1 K p k Π G F k w t k + ξ t k 2 = k = 1 K p k 2 E F k w t k , B t k F k w t k 2 k = 1 K p k 2 ϕ k 2 / b .
Thus, the proof is completed. □

Appendix C.3. Proof of Lemma A3

Proof. 
For any t [ 0 , Q 1 ] , there exists t 0 t such that t t 0 Q 1 and w ¯ t 0 = w t 0 k for k [ 0 , N ] . Assume η t 0 2 η t and η t is non-increasing, i.e., η t η t 0 . When t + 1 I Q , we have
E k = 1 K p k w ¯ t w t k 2 = E k = 1 K p k w t k w ¯ t 0 w ¯ t w ¯ t 0 2 = E k = 1 K p k w t k w ¯ t 0 2 E k = 1 K p k w t k w ¯ t 0 2 ( a ) E k = 1 K p k w t k w ¯ t 0 2 ( b ) k = 1 K p k E τ = t 0 t ( Q 1 ) η τ 2 F k w τ k , B τ k + z τ k 2 k = 1 K p k τ = t 0 t ( Q 1 ) η t 0 2 G 2 + d σ t , k 2 4 ( Q 1 ) 2 η t 2 G 2 + d k = 1 K p k σ t , k 2 ,
where ( a ) follows from the convexity of · 2 , and ( b ) holds since E X E X 2 = E X 2 E [ X ] 2 . Similarly, when t + 1 I Q , we can derive a similar result:
E k = 1 K p k w ¯ t w t k 2 4 ( Q 1 ) 2 G 2 η t 2 .
Thus, the proof is completed. □

Appendix C.4. Proof of Lemma A4

Proof. 
When t + 1 I Q , the DP noise in each round is bounded by
E k = 1 K p k ξ t k 2 = k = 1 K p k ( σ t k ) 2 = 128 η t 2 Q 4 b 2 G 2 n ϵ 2 k = 1 K log ( 1.25 Q b / n k δ ) n k ( a ) 128 η t 2 Q 4 b 2 G 2 n ϵ 2 k = 1 n log ( 1.25 Q b / k δ ) k d k = 128 η t 2 Q 4 b 2 G 2 n ϵ 2 k = 1 n log ( k δ / 1.25 Q b ) d log ( k δ / 1.25 Q b ) = 64 η t 2 Q 4 b 2 G 2 n ϵ 2 log 2 ( n δ / 1.25 Q b ) log 2 ( δ / 1.25 Q b ) = 64 η t 2 Q 4 b 2 G 2 n ϵ 2 log ( n ) log ( 1 . 25 2 Q 2 b 2 / n δ 2 ) ( b ) 4 Q 4 b 2 G 2 n L 2 ϵ 2 log ( n ) log ( 1 . 25 2 Q 2 b 2 / n δ 2 ) ,
where ( a ) follows k = 1 K log ( 1.25 Q b / n k δ ) / n k k = 1 n log ( 1.25 Q b / k δ ) / k d k , and ( b ) holds due to η t 1 / 4 L . □

Appendix D. Proof of Theorem 4

As defined in Appendix A, let I Q = z Q | z = 1 , 2 , be the set of global synchronization steps. Note that, for each communication round, the PS randomly activates a subset of clients S t according to some sampling schemes, and then FedAvg only performs updates on selected clients. Thus, the S t varies in each communication round. Additionally, we define two virtual sequences w ¯ t = k = 1 K p k w t k and v ¯ t = k = 1 K p k v t k ; when t + 1 I Q , we define
g ¯ t = k = 1 K p k F k ( w t k ) + ξ t k ,
g t = k = 1 K p k F k ( w t k , B t k ) + ξ t k .
When t + 1 I Q , we define
g t = k = 1 K p k F k ( w t k , B t k ) ,
g ¯ t = k = 1 K p k F k ( w t k ) .
Thus, we always have v ¯ t + 1 = w ¯ t η t g t and E [ g t ] = g ¯ t . Then, the update scheme of DP-FedAvg with PCP being taken into consideration is
v t + 1 k = w t k η t F k ( w t k , B t k ) , if t + 1 I Q , w t k η t F k ( w t k , B t k + ξ t k ) , if t + 1 I Q ,
w t + 1 k = v t + 1 k if t + 1 I Q , samples S t + 1 and average v t + 1 k k S t + 1 if t + 1 I Q ,
where v t + 1 k denotes a one-step SGD update from w t k and w t + 1 k can be interpreted as the result after the communication steps.

Appendix D.1. Key Lemmas

Motivated by [6], when t + 1 I Q , we still have w ¯ t + 1 = v ¯ t + 1 . However, when t + 1 I Q , we aim to establish this relation in the sense of expectation, i.e., E S t + 1 w ¯ t + 1 = v ¯ t + 1 . To achieve this, the sampling strategy on the PS must be unbiased. We identify two sampling and averaging schemes that satisfy this requirement and provide convergence guarantees.
(I) If the PS performs the sampling with replacement to establish S t + 1 , then each element in S t + 1 may occur multiple times, with each client being sampled with equal probability. In this case, the PS averages the parameters by w t + 1 k = 1 K k S t + 1 v t + 1 k .
(II) If the PS performs the sampling without replacement to establish S t + 1 , then each element in S t + 1 can appear only once. In this case, the PS averages the parameters by w t + 1 k = N K k S t + 1 p k v t + 1 k .
Due to the fact that S t + 1 varies in each communication round, we assume that the DP-FedAvg algorithm activates all devices at the beginning of each communication round. The updated parameters depend only on the sampled clients, and the next-round parameters are broadcasted to all clients. This approach simplifies our theoretical analysis. As expected, w t + 1 k is equal to the weighted average of the parameters across all devices after the SGD updates.
Lemma A5.
If t + 1 I Q , for both client sampling with/without replacement schemes, we have
E S t + 1 w ¯ t + 1 = v ¯ t + 1 .
Proof. 
Please refer to the proof of Lemma 4 in [6]. □
Lemma A6.
For t + 1 I Q , assume η t is non-increasing and η t 2 η t + Q ; then, for client sampling with the replacement scheme,
E S t + 1 w ¯ t + 1 w ¯ t + 1 2 = 4 K η t 2 Q 2 G 2 ,
and for client sampling without the replacement scheme,
E S t + 1 w ¯ t + 1 w ¯ t + 1 2 = N K N 1 4 K η t 2 Q 2 G 2 .
Proof. 
Please refer to the proof of Lemma 5 in [6]. □

Appendix D.2. Completing the Proof of Theorem 4

Proof. 
w ¯ t + 1 w 🟉 2 = w ¯ t + 1 v ¯ t + 1 + v ¯ t + 1 w 🟉 2 = w ¯ t + 1 v ¯ t + 1 2 + v ¯ t + 1 w 🟉 2 + 2 w ¯ t + 1 v ¯ t + 1 , v ¯ t + 1 w 🟉 A 3 ,
where the A 3 term will be 0 due to Lemma A5. When t + 1 I Q , the first term will vanish due to w ¯ t + 1 = v ¯ t + 1 . Then, we have
E w ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + A η t 2 .
When t + 1 I Q , we use Lemma A6 to bound the second term in the second equality:
E w ¯ t + 1 w 🟉 2 = E w ¯ t + 1 v ¯ t + 1 2 + E v ¯ t + 1 w 🟉 2 1 μ η t E w ¯ t w 🟉 2 + A + B + H η t 2 ,
where H is defined in Theorem 4. The main difference between (A16) and (A52) is the additional term H. Thus, we can follow the same progress and argument to prove the theorem. To be specific, we choose ν = max β 2 ( A + B + H ) μ β 1 , α 0 , β = 2 μ and η t = 2 μ 1 α + t , where α max { 8 κ , Q } , κ = L μ . By the L-smoothness of F ( · ) , we have
E F ( w ¯ t ) F 🟉 L 2 E w ¯ t w 🟉 2 L 2 ν t + α ,
then,
E F ( w ¯ t ) F 🟉 2 κ t + α A + B + H μ + 2 L w 1 w 🟉 2 .
Thus, we complete the proof. □

References

  1. Li, Y.; Chang, T.H.; Chi, C.Y. Secure Federated Averaging Algorithm with Differential Privacy. In Proceedings of the IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Espoo, Finland, 17–20 September 2020; pp. 1–6. [Google Scholar]
  2. Sattler, F.; Wiedemann, S.; Müller, K.R.; Samek, W. Robust and communication-efficient federated learning from non-iid data. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 3400–3413. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, S.; Chang, T.-H. Federated Matrix Factorization: Algorithm Design and Application to Data Clustering. IEEE Trans. Signal Process. 2022, 70, 1625–1640. [Google Scholar] [CrossRef]
  4. Wang, S.; Xu, Y.; Yuan, Y.; Quek, T.Q.S. Toward Fast Personalized Semi-Supervised Federated Learning in Edge Networks: Algorithm Design and Theoretical Guarantee. IEEE Trans. Wireless Commun. 2024, 23, 1170–1183. [Google Scholar] [CrossRef]
  5. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  6. Li, X.; Huang, K.; Yang, W.; Wang, S.; Zhang, Z. On the Convergence of FedAvg on non-IID Data. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26–30 April 2020; pp. 1–26. [Google Scholar]
  7. Li, Y.; Wang, S.; Chi, C.Y.; Quek, T.Q.S. Differentially Private Federated Clustering Over Non-IID Data. IEEE Internet Things J. 2024, 11, 6705–6721. [Google Scholar] [CrossRef]
  8. Glasgow, M.R.; Yuan, H.; Ma, T. Sharp bounds for federated averaging (local SGD) and continuous perspective. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Virtual, 28–30 March 2022; pp. 9050–9090. [Google Scholar]
  9. Woodworth, B.E.; Patel, K.K.; Srebro, N. Minibatch vs local SGD for heterogeneous distributed learning. In Proceedings of the Advances in Neural Information Processing Systems 33 (NeurIPS 2020), Virtual, 6–12 December 2020; pp. 6281–6292. [Google Scholar]
  10. Hsu, T.M.H.; Qi, H.; Brown, M. Measuring the effects of non-identical data distribution for federated visual classification. arXiv 2019, arXiv:1909.06335. [Google Scholar]
  11. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2020, 2, 429–450. [Google Scholar]
  12. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.J.; Stich, S.U.; Suresh, A.T. Scaffold: Stochastic controlled averaging for on-device federated learning. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26–30 April 2020; pp. 5132–5143. [Google Scholar]
  13. Geiping, J.; Bauermeister, H.; Dröge, H.; Moeller, M. Inverting gradients-how easy is it to break privacy in federated learning? In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Virtual, 6–12 December 2020; pp. 16937–16947. [Google Scholar]
  14. Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends® Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
  15. Li, Y.; Huang, C.W.; Wang, S.; Chi, C.Y.; Quek, T.Q. Privacy-Preserving Federated Primal-Dual Learning for Non-Convex and Non-Smooth Problems With Model Sparsification. IEEE Internet Things J. 2024, 11, 25853–25866. [Google Scholar] [CrossRef]
  16. Wang, X.; Wang, S.; Li, Y.; Fan, F.; Li, S.; Lin, X. Differentially Private and Heterogeneity-Robust Federated Learning with Theoretical Guarantee. IEEE Trans. Artif. Intell. 2024, 5, 6369–6384. [Google Scholar] [CrossRef]
  17. Li, Z.; He, Y.; Yu, H.; Kang, J.; Li, X.; Xu, Z.; Niyato, D. Data heterogeneity-robust federated learning via group client selection in industrial IoT. IEEE Internet Things J. 2022, 9, 17844–17857. [Google Scholar] [CrossRef]
  18. Wang, S.; Xu, Y.; Wang, Z.; Chang, T.-H.; Quek, T.Q.S.; Sun, D. Beyond ADMM: A Unified Client-Variance-Reduced Adaptive Federated Learning Framework. In Proceedings of the AAAI, Washington, DC, USA, 7–14 February 2023; pp. 10175–10183. [Google Scholar]
  19. Wu, H.; Wang, P. Fast-convergent federated learning with adaptive weighting. IEEE Trans. Cogn. Commun. Netw. 2021, 7, 1078–1088. [Google Scholar] [CrossRef]
  20. Wu, H.; Tang, X.; Zhang, Y.J.A.; Gao, L. Incentive Mechanism for Federated Learning With Random Client Selection. IEEE Trans. Netw. Sci. Eng. 2024, 11, 1922–1933. [Google Scholar] [CrossRef]
  21. Saha, R.; Misra, S.; Chakraborty, A.; Chatterjee, C.; Deb, P.K. Data-centric client selection for federated learning over distributed edge networks. IEEE Trans. Parallel Distrib. Syst. 2022, 34, 675–686. [Google Scholar] [CrossRef]
  22. Pang, J.; Yu, J.; Zhou, R.; Lui, J.C. An incentive auction for heterogeneous client selection in federated learning. IEEE Trans. Mob. Comput. 2022, 22, 5733–5750. [Google Scholar] [CrossRef]
  23. Shen, X.; Liu, Y.; Zhang, Z. Performance-enhanced Federated Learning with Differential Privacy for Internet of Things. IEEE Internet Things J. 2022, 9, 24079–24094. [Google Scholar] [CrossRef]
  24. Li, Y.; Wang, S.; Chi, C.Y.; Quek, T.Q. Differentially Private Federated Learning in Edge Networks: The Perspective of Noise Reduction. IEEE Netw. 2022, 36, 167–172. [Google Scholar] [CrossRef]
  25. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar]
  26. McMahan, H.B.; Ramage, D.; Talwar, K.; Zhang, L. Learning Differentially Private Recurrent Language Models. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
  27. Huang, Z.; Hu, R.; Guo, Y.; Chan-Tin, E.; Gong, Y. DP-ADMM: ADMM-based distributed learning with differential privacy. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1002–1012. [Google Scholar] [CrossRef]
  28. Erlingsson, Ú.; Feldman, V.; Mironov, I.; Raghunathan, A.; Talwar, K.; Thakurta, A. Amplification by shuffling: From local to central differential privacy via anonymity. In Proceedings of the ACM-SIAM Symposium on Discrete Algorithms, San Diego, CA, USA, 6–9 January 2019; pp. 2468–2479. [Google Scholar]
  29. Balle, B.; Barthe, G.; Gaboardi, M. Privacy amplification by subsampling: Tight analyses via couplings and divergences. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 3–8 December 2018; pp. 6277–6287. [Google Scholar]
  30. Wei, K.; Li, J.; Ding, M.; Ma, C.; Yang, H.H.; Farokhi, F.; Jin, S.; Quek, T.Q.S.; Poor, H.V. Federated Learning With Differential Privacy: Algorithms and Performance Analysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3454–3469. [Google Scholar] [CrossRef]
  31. Noble, M.; Bellet, A.; Dieuleveut, A. Differentially Private Federated Learning on Heterogeneous Data. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Virtual, 28–30 March 2022; pp. 10110–10145. [Google Scholar]
  32. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership inference attacks against machine learning models. In Proceedings of the IEEE symposium on security and privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 3–18. [Google Scholar]
  33. Zhang, Y.; Duchi, J.C.; Wainwright, M.J. Communication-efficient algorithms for statistical optimization. J. Mach. Learn. Res. 2013, 14, 3321–3363. [Google Scholar]
  34. Blake, C.L.; Merz, C.J. UCI Repository of Machine Learning Databases. 1996. Available online: https://archive.ics.uci.edu/dataset/2/adult (accessed on 30 April 1996).
  35. LeCun, Y.; Cortes, C.; Burges, C. The MNIST Database. 1998. Available online: http://yann.lecun.com/exdb/mnist (accessed on 1 November 1998).
Figure 1. A typical FL system.
Figure 1. A typical FL system.
Mathematics 13 00497 g001
Figure 3. The testing accuracy versus communication rounds of the DP-FedAvg algorithm with various Q under Adult and MNIST datasets.
Figure 3. The testing accuracy versus communication rounds of the DP-FedAvg algorithm with various Q under Adult and MNIST datasets.
Mathematics 13 00497 g003
Table 1. Notations.
Table 1. Notations.
NotationDefinition
· Euclidean norm.
x , y = x y The inner product operator.
ϵ , δ The DP parameter.
The 2 -norm sensitivity.
NThe total number of clients.
TThe total number of iterations.
QThe number of local SGD updates.
| D | The size of set D .
w Transpose of vector w .
w t The global model at t-th iteration.
w t k The local model for k-th client at t-th iteration.
F k w The objective function for the client k.
F k w t k , B t k The mini-batch gradient using the data B t k at client k.
Γ The degree of non-i.i.d. data.
G The gradient clipping level.
n, n k The number of data samples for all client and the client k.
bThe size of mini-batch data.
dThe dimension of model parameters.
ϕ k The variance of stochastic gradients for the k-th client.
LL-smooth parameter.
μ μ -strong convex parameter.
E [ · ] , P [ · ] The statistical expectation and the probability function, respectively.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Wang, S.; Wu, Q. Convergence Analysis for Differentially Private Federated Averaging in Heterogeneous Settings. Mathematics 2025, 13, 497. https://doi.org/10.3390/math13030497

AMA Style

Li Y, Wang S, Wu Q. Convergence Analysis for Differentially Private Federated Averaging in Heterogeneous Settings. Mathematics. 2025; 13(3):497. https://doi.org/10.3390/math13030497

Chicago/Turabian Style

Li, Yiwei, Shuai Wang, and Qilong Wu. 2025. "Convergence Analysis for Differentially Private Federated Averaging in Heterogeneous Settings" Mathematics 13, no. 3: 497. https://doi.org/10.3390/math13030497

APA Style

Li, Y., Wang, S., & Wu, Q. (2025). Convergence Analysis for Differentially Private Federated Averaging in Heterogeneous Settings. Mathematics, 13(3), 497. https://doi.org/10.3390/math13030497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop