Next Article in Journal
Magnetohydrodynamic Flow and Transport Behaviors of Blood-Based Ternary Nanofluids in Stenosed Arteries with Axial Symmetry: Effects of Thermal Radiation and Caputo Fractional Derivatives
Previous Article in Journal
Strategic Scenario Interaction: A Computational Framework Based on Game-Theoretic and Quantum-Inspired Modeling
Previous Article in Special Issue
An Efficient and Effective Model for Preserving Privacy Data in Location-Based Graphs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

APDP-FL: Personalized Federated Learning Based on Adaptive Differential Privacy

1
School of Computer Science & Engineering, LinYi University, Linyi 276000, China
2
Shanxi Key Laboratory of Cryptography and Data Security, Shanxi Normal University, Taiyuan 030031, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(12), 2023; https://doi.org/10.3390/sym17122023
Submission received: 14 October 2025 / Revised: 13 November 2025 / Accepted: 14 November 2025 / Published: 24 November 2025

Abstract

Frequent gradient exchange and heterogeneous data distribution in federated learning can lead to serious privacy leakage risks. Traditional privacy-preserving strategies fail to meet the personalized privacy needs from different users and may cause a decrease in model accuracy and convergence difficulties. The symmetry of federated learning may lead to the insufficiency of contribution evaluation mechanisms in protecting the privacy of sensitive data holders. However, federated learning avoids the risk of privacy leakage caused by data centralization because the raw data is always stored on the local device during the training process, and only encrypted model parameters or gradient updates are exchanged. To address these issues, this paper proposes an adaptive personalized differential privacy federated learning scheme APDP-FL. First, we propose an adaptive noise addition method that scores each round of training based on the parameters generated during training and dynamically adjusts the noise level for the next round. This method adds larger noise scales in the early stages of training, consuming less privacy budget, and gradually reduces noise addition during training to accelerate model convergence. Second, we design a personalized privacy protection strategy that adds noise tailored to individual needs for participating clients based on their privacy preferences. This solves the problem of insufficient or excessive privacy protection for some participants due to identical privacy budget sets for all clients, achieving personalized privacy protection for clients. Finally, we conduct extensive experimental simulations, comparisons, and analyses on three real federated datasets, MNIST, FMNIST, and CIFAR-10, verifying the advantages of APDP-FL in terms of privacy protection, model accuracy, and convergence speed.

1. Introduction

The development of large language models has led to an explosion in the applications of artificial intelligence. In particular, the branch of machine learning [1] has been widely utilized in areas such as computer vision [2], natural language processing [3], and image generation [4]. Traditional centralized machine learning usually requires collecting a large amount of training data and handing it over to a third party for cleaning and aggregation; then, the model can be trained and published. However, the loss of participants’ control over their own data during the training process, coupled with the increase in the amount of training data, makes the risk of privacy leakage increase, which leads to limitations in the further development of machine learning.
Federated learning [5] (FL), as a distributed machine learning framework, is a breakthrough technology for addressing the privacy protection of machine learning. Federated learning adopts a “data localization and parameter exchange mechanism”, where each participant uses their own data to train the model locally, encrypting only the model parameters (such as gradients or weights) and uploading them to the central server for aggregation, rather than transmitting the original data. This design ensures that the raw data is always kept locally, greatly reducing the risk of data leakage. Currently, federated learning has been widely adopted in domains such as intelligent transportation, healthcare, and energy/power systems [6,7,8].
Although federated learning can realize model training with “data localization” and effectively reduce the risk of privacy leakage (e.g., sensitive information of raw data in machine learning may be leaked through model parameters, gradient updates, or parameter exchange processes), it is not able to completely resist external privacy attacks, and many studies have shown that federated learning strategies still have some privacy leakage and data security problems [9]. First, due to the distributed nature of federated learning, it is difficult to ensure that the local models uploaded by clients are non-malicious, and once a malicious client uploads a malicious model, it will interfere with the global model of federated learning. Second, the model gradient uploaded by the client also contains some information of the local data, and the attacker can reverse launch the client’s original data through the model gradient. Therefore, it is urgent to construct a federated learning scheme that is secure and efficient and ensures data security and privacy.
In order to solve the FL privacy protection problem, researchers have proposed a large number of solutions. The mainstream solutions are mainly based on homomorphic encryption (HE) [10], secure multi-party computation (SMC) [11], and differential privacy (DP) [12]. HE can train and predict models in the ciphertext state, but the encryption and decryption operations and ciphertext transmission will place huge computational and communication pressure on the FL system. SMC can complete model training without disclosing client privacy, but it will introduce huge communication overhead to the FL. DP can provide rigorous mathematical proofs to guarantee the validity of the privacy protection, and it is suitable for various types of data analysis tasks, as well as being able to flexibly balance privacy protection and data availability. In summary, DP has strict mathematical definitions, adjustable privacy budgets, and lightweight computation, making it more suitable for FL distributed environments. However, the application of DP in FL still faces the problems of trade-offs between privacy protection and model performance, poor model accuracy, and inefficient communication, in addition to the challenges exacerbated by the non-independent and identically distributed (non-IID) nature of local data.
The current research on personalized differential privacy in federated learning mainly focuses on dynamic optimization and localized protection. Researchers propose an adaptive noise adjustment mechanism that dynamically optimizes the noise level of each client through the design of a noise scaler under a given privacy budget. At the same time, the direction matrix is used to correct the model’s update direction, effectively alleviating the model drift problem caused by noise introduction. In terms of technical implementation, reparameterization training is used to select personalized client update information, combined with a dynamic adaptive norm control model update range, significantly improving the targeting of noise injection. In terms of cross-domain applications, the privacy budget allocation scheme based on information entropy provides differentiated privacy protection strengths for different users according to the distribution differences of user data, achieving the optimal balance between privacy and utility.
Therefore, in order to explore how to improve FL model training efficiency and accuracy, as well as communication efficiency, under DP protection, we propose a personalized federated learning framework (APDP-FL) based on self-adaptive adapted differential privacy. Our contributions to this work can be summarized as follows:
(1)
This study propose the APDP-FL architecture to solve the three problems of personalized privacy requirements, accuracy reduction, and model convergence difficulties in FL.
(2)
An adaptive noise addition method is proposed to score each parameter for each round of the training process, and then, it dynamically adjusts the size of the added noise for the next round, which consumes less of the privacy budget and speeds up the model’s convergence.
(3)
A personalized privacy protection strategy is designed, which can add noise for the client; it meets its own needs according to the user’s privacy preference, solving the problem of insufficient or excessive privacy protection for some clients due to the same set of privacy budgets, and it realizes the personalized privacy protection of the client.
(4)
Extensive experiments are carried out on three real datasets: MNIST, FMNIST, and CIFAR-10. The results show that APDP-FL is superior to state-of the-art baselines in terms of privacy protection, model accuracy, and convergence speed.
The rest of this paper is organized as follows. In Section 2, we present a literature review of related work. Section 3 describes the background and problem setting, including FL, differential privacy, system model, attack model, and design objectives. Section 4 details the proposed APDP-FL scheme. Section 5 provides the performance analysis of APDP-FL. Section 6 provides the experimental results and analyses. The conclusions are drawn in Section 7.

2. Related Work

Although the “Data Localization” training mode in federated learning can effectively reduce the risk of local data privacy leakage, it is still unable to completely provide protection against privacy attacks. Currently, the academic solutions that address the privacy risk of federated learning are mainly based on homomorphic encryption (HE), secure multi-party computation (SMC), and differential privacy (DP).

2.1. Homomorphic Encryption

HE allows model training and predictions in ciphertext without decrypting data through data encryption, ciphertext computation, model parameter encryption, and secure aggregation. Hu et al. [13] proposed MaskCrypt to balance the trade-off between security and practicality when homomorphic encryption is used. Rather than encrypting model updates in their entirety, MaskCrypt applies an encryption mask to sift out a small portion of the updates for encryption. Kumbhar et al. [14] designed puzzle Archimedes optimization algorithm (PAOA)-based multi-key HE. Healthcare parties utilized noise-free HE to protect sensitive data and enhance communication security. Wang et al. [15] proposed a secure and privacy-preserving medical data federated learning architecture based on Paillier homomorphic encryption; the shared training model of the client is encrypted to ensure security and privacy. Cai et al. [16] proposed a secure and efficient federated learning scheme (SecFed) based on multi-key HE to preserve user privacy and delegated some operations to TEE to improve efficiency while ensuring security. Hu et al. [13] proposed MaskCrypt, a new mechanism designed to balance the trade-off between security and practicality when homomorphic encryption is used. However, the encryption and decryption operations and ciphertext transmission of HE can place huge computational and communication pressure on the FL system and cannot handle complex operations.

2.2. Secure Multi-Party Computation

SMC ensures that all parties in FL work together on model training and prediction without disclosing their private data through secret sharing, obfuscated circuits, homomorphic encryption, and inadvertent transmission. Tan et al. [17] proposed a privacy-preserving personalized federated learning scheme under secure multi-party computation (SMC-PPFL), which preserved privacy while obtaining a local personalized model with high prediction accuracy. Ciucanu et al. [18] proposed three secure protocols relying on secure multi-party computation, which guarantee desirable security properties for the following: input data, intermediate data, and output data. Kalapaaking et al. [19] proposed a convolutional neural network (CNN)-based federated learning framework that combined secure multi-party computation (SMC)-based aggregation and encrypted inference methods, all within the context of 6G and IoMT. Xiao et al. [20] proposed a practical secure federated learning system named PrSeFL, which implemented secure multiparty computation-based secure aggregation in blockchain environments. Tan et al. [17] proposed a privacy-preserving personalized federated learning scheme under secure multi-party computation (SMC-PPFL), which could preserve privacy while obtaining a local personalized model with high prediction accuracy. However, SMC usually involves a large number of data interactions and computations, which incurs a huge communication overhead and may become a performance bottleneck.

2.3. Differential Privacy

DP is able to flexibly balance the privacy and availability of local data by adding random noise to local data or local models, which makes it impossible for an attacker to infer the client’s privacy information from the data analysis results. Lyu et al. [21] proposed a scheme that combines differential privacy with federated learning to realize privacy-preserving outcomes. Tang et al. [22] proposed a novel privacy-preserving FL system (ReFL) that enforced DP and avoided accuracy reductions in global models caused by the excessive DP noise of rogue clients. Fukami et al. [23] proposed a primal–dual DP algorithm with denoising normalization (DP-Norm) for less sensitivity to noise/interference, such as DP diffusion and heterogeneous data allocation. The above schemes perform poorly in FL with complex datasets due to the trade-off between privacy budget and performance. Sun et al. [24] provide a localized differential privacy–federated learning (LDP-FL) solution, solving the problem above, but it still exhibits poor accuracy in large model testing with a small number of users. Bhowmick et al. [25] were the first to combine localized differential privacy with federated learning for deep learning models, proposing a practical approach to large-scale local privacy model training, and this method is suitable for large-scale image classification with almost no decrease in model utility. Xie et al. [26] proposed a federated multitask learning (FedMTL) approach, which was introduced to reformulate the FL model as a multiobjective optimization problem, resulting in a federated multigradient descent algorithm (FedMGDA) with better model personalization against data heterogeneity and Byzantine attacks.

3. Background and Problem Setting

3.1. Federated Learning

The federated learning framework [5] consists of a central server and n clients, each of which has unique access to its private local data D k , where k is a variable and k { 1 , 2 , 3 , , n } . In each learning round t, the server randomly selects several clients to share the current global model W g l o b a l t . Each selected client updates its local data at D k for each local round to obtain a new local model sent to the server at D k t + 1 . The aggregation process usually uses the federated averaging algorithm to obtain the global model for the next round D g l o b a l t + 1 via Equation (1):
W g l o b a l t + 1 = W g l o b a l t + η n k = 1 n W k t + 1 W g l o b a l t
where η is the learning rate set by the server when the global model is updated.
There are a total of one server S and n clients in a generic federated learning system, and the server S collaborates with all clients to accomplish the following optimization objectives throughout the federated learning process:
arg   min   w R d f ( w ) = 1 N n = 1 N f n ( w )
where f n is the loss function of the client n, w is the local model, and w R d means that the variable w is a d-dimensional real vector. In the neural network, empirical risk minimization with the cross-entropy function is used as its loss function:
f n w = 1 | D n | j = 1 | D n | L x j , y j , w
where | D n | is the total number of samples owned by client n, and x j , y j is the j- t h sample in client n.
In each training round of federated learning t 1 , 2 , 3 , , T , the server sends the current global model W g l o b a l t to each client. Clients train their own local model W k t + 1 locally using their own private data D k D 1 , D 2 , D 3 , , D K . Then, the client computes its own model update gradient as Δ k t + 1 = W k t + 1 W g l o b a l t and sends it to the server. After the server receives the updates sent by the clients, it aggregates the model’s updates using the aggregation function and then updates the global model using Equations (2)–(4) to obtain the global model W g l o b a l t + 1 , which is used for the next round of training:
W g l o b a l t + 1 = W g l o b a l t + η · g Δ t + 1
where η is the learning rate of the server in federated learning; g ( · ) is the aggregation function, which is chosen according to the aggregation strategy; and Δ t + 1 = k C k Δ k t + 1 . The training process continues until the global model reaches the desired performance metrics on the validation dataset maintained by the server or reaches the maximum number of iteration rounds. All clients keep their private data confidential throughout the training process; for example, the attacker and the server do not have access to the client’s private data.

3.2. Differential Privacy

Differential privacy solves the core contradiction between privacy protection and data availability in the era of big data, providing a verifiable and implementable privacy protection framework for governments, enterprises, and individuals. Especially in the context of the proliferation of sensitive data and the tightening of privacy regulations, its technology has become the cornerstone of data security. Differential privacy is achieved by adding noise to the data that matches a specific distribution; even if a single piece of datum in the dataset is modified, it will not significantly change the statistical results, so it can effectively defend against differential attacks and thus protect personal privacy.
Definition 1
(Differential Privacy in Deep Learning). Let the mechanism of a deep learning algorithm be M. The training dataset is taken as input and trained using gradient descent; the output is the trained model’s parameters; P M is the set consisting of all possible outputs. Then, for any two neighboring training datasets D and D , the parameter distribution of the output satisfies Equation (5):
P M M D S M e x p ε × P M M D S M + δ
It can be said that the deep learning algorithm mechanism M satisfies ε , δ -DP (differential privacy), the parameter ε is the privacy-preserving budget, and the size of ε is negatively correlated with the privacy-preserving effect. δ represents the relaxation factor, which represents the probability of violating strict differential privacy. When ε is larger, this means that the added noise is small, and the privacy protection level is low. Conversely, when it is smaller, the added noise is large, and the privacy protection level is high. However, in the limit, the privacy-preserving budget is ε = 0 , which will have the highest privacy-preserving effect, and the difference between the outputs of the algorithm mechanism M on two neighboring datasets will be completely indistinguishable. However, because adding too much noise to the dataset will result in a large amount of interference when using the dataset, rendering the data unusable, it is necessary to control ε > 0 in practice. δ is called the relaxation term, which indicates the probability of being able to tolerate illegal strict differential privacy, and it is typically set to 1 | D | .
Definition 2
(Rényi Differential Privacy). Rényi differential privacy is obtained by restricting Rényi scattering, For two probability distributions P and Q, the Rényi divergence of order α is defined as follows:
D α P | | Q = 1 α 1 log E x Q P x Q x α
where the P x and Q x distributions represent the probability density function of x on the distributions P and Q. And for the mechanism M described in definition 1, let M D w represent the probability density of training on the dataset D to obtain the result w. If the mechanism M satisfies α , ε -RDP for any α 1 , , then the mechanism M needs to satisfy the following:
D α M D | | M D ε

3.3. System Model

In this paper, the differential privacy federal average algorithm is used to provide different privacy budget levels, and participants can achieve personalized privacy protection by choosing the privacy level that meets their needs. Due to the use of a uniform noise addition strategy throughout the process, this will lead to the addition of excessive noise at the late stage of training, which affects the model accuracy. Therefore, adaptive noise addition technology is also introduced in the scheme, where noise is added adaptively during the training process to minimize the impact on the global model’s accuracy. The framework of the adaptive personalized differential privacy federated learning algorithm is shown in Figure 1.
The system architecture consists of a centralized parameter server and multiple participating clients, achieving a balance between privacy and model performance through “data localization+model collaboration”. Firstly, the server distributes the initialized global model W t 1 (including model structure and initial weight parameters) and preset noise parameter σ i (such as noise scales in differential privacy, privacy budget ε , etc.) to all participating learning clients. After receiving the initial model and noise parameters issued by the server, the client iteratively trains the model using their own private dataset D i (which is always kept locally and not transmitted externally). After training is completed, the client needs to perform critical local privacy protection processing: Perturb the locally trained model parameters based on their personalized privacy preference r i and the noisy parameters distributed by the server. After completing noise processing, the client uploads the perturbed model parameters (rather than the original model or local data) to the server. After collecting the encrypted model parameters from all clients, the server performs weighted averaging on the model parameters from different clients to integrate knowledge from all parties and update the weights of the global model. Subsequently, the server uses the reserved validation dataset to evaluate the performance of the aggregated global model, calculates key indicators such as accuracy and loss function values, and dynamically adjusts the denoising coefficient for the next round of training based on the evaluation score. If the model’s performance does not meet expectations, the noise intensity can be appropriately reduced to improve model accuracy. If the privacy risk is high (for example, if it is found through privacy budget monitoring that the consumption of ε is too fast), the amplitude of noise enhancement should be increased to strengthen protection. This dynamic adjustment mechanism achieves the automated optimization of privacy utility trade-off through a feedback loop. Finally, the server distributes the updated global model (which incorporates multi-party knowledge but hides individual data features) and the recalculated new round of denoising coefficients to all participating clients, initiating the next round of federated learning iteration.

3.4. Attack Model

We assume that the central server is honest and curious, will perform the federated learning process strictly and accurately, and return correct results to all clients after the global aggregation of the received intermediate parameters. However, the central server is curious about the private data of participating users and tries to obtain private data by analyzing the model parameters submitted by users. All users adhere to the model training process with respect to the same central server, but they may try to obtain other users’ private information through the global model or public channels. Assuming that there is an active attacker A in the training process with the goal of inferring the private information of users using intermediate parameters or aggregation results, attacker A has the following capabilities:
(1)
The ability to eavesdrop the intermediate parameters of the interaction between the central server and the user through the public channel to infer the private information of the user.
(2)
The ability to hack the central server and use the intermediate parameters or aggregation results of the server to infer the private information of the user.
(3)
The ability to hack into one or more users and use their information to infer private data about other users but not the ability to manipulate all users at once.

3.5. Design Objectives

The proposed scheme should ensure the data privacy of participating users and provide personalized privacy protection for participants, and it should also ensure that the accuracy of the model is not degraded:
(1)
Data Privacy: Ensure that Attacker A cannot directly or indirectly use the local model parameters and the global model aggregated by the central server to infer the private data of users.
(2)
Personalized Privacy Protection: Different users possess different sensitivities to data, and the scheme should ensure that all participants can set their own privacy parameters at the beginning of training to meet their personalized privacy protection needs.
(3)
Model Accuracy: The server needs to provide noise addition coefficients for participating users, and the scheme needs to ensure that the noise added in each round meets privacy protection needs while not affecting the model’s accuracy as much as possible.

4. Adaptive Personalized Differential Privacy Federated Learning Algorithm (APDP-FL)

APDP-FL mainly consists of two core parts: adaptive noise addition and personalized privacy protection. According to the various parameters generated during the training process to score each round of training and the scoring results that dynamically adjust the size of the next noise-adding round, the client’s local training will be based on their own privacy preferences, adjusting the level of local privacy protection to achieve the personalized privacy protection of the client.

4.1. Adaptive Noise Addition

As training progresses, the gradient of the model uploaded by the client will gradually decrease, and adding noise at the same level as before when the model is about to converge will affect the accuracy of the model. Therefore, adding more noise in the early stage of training and reducing the added noise in the later stage can reduce the impact on the model’s accuracy. In the case of limited privacy budget, the parameter σ of adding noise dynamically according to the indexes in the training process can reduce the impact on the model’s accuracy with the same privacy budget.
In order to realize the adaptive addition of noise in FL, this paper fuses four metrics, gradient size, training loss, model accuracy, and training rounds, to design a scoring function. When the score exceeds 50 points, the noise parameter σ is reduced to p σ , where p 0 < p < 1 is a constant, and the server calculates the score of the round during the training process and then dynamically adjusts the noise parameter σ sent to the client according to the score. Each computational module of the scoring function is as follows:
(1)
Gradient: Let g t be the gradient biparadigm at the time of global model update in the t-th round of communication, g represents gradient, and let g m a x be the maximum value of the historical gradient biparadigm. The scoring of s c o r e g with the gradient as the scoring criterion is calculated using the following formula:
s c o r e g = m i n m a x 2 1 2 g t g m a x , 0 , 1 × 100 %
To minimize added noise when the global model gradient paradigm is small, the opposite of the ratio of g t to g m a x plus 1 was used as the scoring criterion for the gradient, and the size of s c o r e g was limited to no more than 100 % .
When the global model gradient paradigm is small, the model approaches convergence, and noise needs to be reduced to avoid interfering with optimization. Using the inverse of the ratio of the current gradient, two normal forms were used relative to the historical average plus 1 as the score. Dynamic denoising is as follows: The smaller the gradient, the higher the score, and the stronger the denoising. Limit the rating to ≤100% to ensure privacy protection and prevent noise from disappearing.
(2)
Training Loss: Let L o s s t be the loss on the test set after the t-th round of communication, and let L o s s t 1 be the loss on the t 1 round. This module’s score s c o r e l is calculated as follows:
s c o r e l = 100 % , i f l o s s t 1 < l o s s t 0 , e l s e
If the training loss of the current round is greater than the training loss of the previous round, this means that the noise added during the training of this round may be too large, so during the next round of training, as little noise as possible should be added.
(3)
Model Accuracy: Let A c c t be the accuracy of the global model on the test set after the t-th round of communication, and let A c c t 1 , A c c t 2 , A c c t 3 , and A c c t 4 be the accuracy of the model in the first four rounds. The model accuracy module score s c o r e a is calculated as follows:
s c o r e a = 100 , % i f ( A c c t + A c c t 1 + A c c t 3 + A c c t 4 ) 5 > A c c t 0 , e l s e
s c o r e a is based on the mean value of the model accuracy of the last five times (a represents accuracy). If the model accuracy of the last round is lower than the mean value, this means that the model’s accuracy has shown a downward trend during the training process, so it is necessary to minimize the addition of noise and reduce the interference of added noise on the model’s accuracy.
(4)
Training Rounds: Set T m as the maximum number of communication rounds set. The training process needs to gradually reduce the addition of noise; thus, the training rounds that prevail in s c o r e t are calculated as follows:
s c o r e t = m i n 2 t T m , 1 × 100 %
Combining the above scoring formula and giving different weights to each module, the final scoring calculation method can be derived as follows:
s c o r e = 35 × s o c r e g + 20 × s c o r e l + 25 × s c o r e a + 20 × s c o r e t
The comprehensive scoring function dynamically adjusts noise intensities by the weighted fusion of individual scores, where the weight coefficients α = 35 ,   β = 20 , γ = 35 ,   δ = 20 , and delta satisfy α + β + γ + δ = 100 and dynamically allocate α ,   β ,   γ , and δ according to task requirements. The coefficients can be dynamically adjusted.
Relying on this scoring function, adaptive noise additions to the model during the training process can be realized, where the details of variations in the noise parameter σ are as follows:
σ t = p σ t 1 , s c o r e > 50 σ t 1 , s c o r e 50
where p 0 < p < 1 is a constant, representing the proportion of reduced noise addition. The process of calculating noise parameters is shown in Figure 2.

4.2. Personalized Privacy Protection

Personalized privacy protection strategy is added on the basis of adaptive noise addition, which allows participants to add noise that meets their own needs, and this solves the problem of insufficient or excessive privacy protection for some participants due to the same privacy budget set for each client.
The personalized differential privacy strategy works on the basis of the adaptive noise addition strategy, and the server sends Gaussian noise parameters σ to the participating users during each round of training, and if all users have the same privacy level, the noise addition process during the training is as follows:
g k = 1 r b r k g ¯ b + 2 C r N ( 0 , σ 2 I )
where g ¯ b is the gradient of the model after gradient clipping, I represents the identity matrix, and r k is the subset of the client’s k-th round of sampling, while r is the sample size, and C is the gradient clipping factor. The size of the added noise is determined by the Gaussian noise parameter σ . The larger σ is, the more noise is added to the gradient, which is good for privacy protection, but model utilization will be reduced. Therefore, the local privacy protection level can be controlled by controlling the size of the Gaussian noise parameter σ . Participating users need to quantify their privacy preferences, R = R 1 , R 2 , R 3 , , R n , and set the standard privacy protection level L. During local training, the client needs to adjust the Gaussian noise parameter σ according to the set privacy preference R i and the standard privacy protection level L before adding noise to the model gradient:
σ i = R i L σ
It is very difficult to compute the privacy overhead for each client since each client in each round adds a different amount of noise to the model’s gradient. Therefore, after one round of the learning process, this paper calculates the privacy overhead under the standard privacy protection level, and at the same time, since the noise added in each round is dynamic, it is necessary to determine whether the privacy overhead exceeds the privacy budget after the completion of each round of training, and if the privacy budget is consumed in advance, the federated training process needs to be terminated early.
The criteria for the standard privacy protection level L are as follows: The smaller the L value, the greater the strength of privacy protection. The larger the L value, the weaker the privacy protection strength. According to research practice, when L < 0.1 , this provides strong privacy protection. When L < 1 , generally acceptable privacy protection is provided.

4.3. APDP-FL Training Process

4.3.1. Server-Side Training Process

First, the server randomly selects the clients participating in this round of learning and sends the global model W t 1 and the noise parameter σ t 1 of this round to the selected clients. The selected clients use the global model for multiple rounds of local training, adding noise to the model according to the locally set privacy preference R i , the standard privacy protection level L, and the Gaussian noise parameter σ t 1 distributed by the server. Finally, the model parameters, w ˜ i t , are uploaded. The server receives the model parameters uploaded by the client and completes model aggregation to update the global model with the set learning rate. Then, calculate the baseline privacy overhead ε t for this round and determine whether ε t exceeds the set privacy budget. If the privacy budget reaches the upper limit, the training process needs to be aborted. Test the global model on the test set to obtain the training loss and model accuracy, calculate the training score for this round based on the scoring function, and generate a new noise parameter σ t based on the score. Repeat the above steps and stop the federated learning process if the privacy budget is exhausted or the number of training rounds reaches a pre-set value, T m . The algorithm of APDP-FL on the server side is as follows (Algorithm 1).
Algorithm 1 APDP-FL server-side algorithm
Input: 
Set of users participating in federated learning N, maximum number of communication rounds T m , user selection rate γ , 0 < γ < 1 , privacy budget ε , learning rate η ;
Output: 
global model W;
  1:
Initialize the global model W 0 ;
  2:
for   t = 1 , 2 , , T m   do
  3:
     Server randomly selects γ N users who will participate in this round of learning n t N ;
  4:
     Server sends model parameter W t 1 and noise parameter to client i n t σ t 1 ;
  5:
     for   i n t do                  ▹ Clients run simultaneously
       ▹ Client i executes local learning strategy to get local model for round t and adds noise.
  6:
    w ˜ t , i C l i e n t U p d a t e ( W t 1 , σ t 1 ) ;
  7:
   Client uploads local model w ˜ t , i ;
  8:
     end for
  9:
     Server collects model from client w t , i ;
10:
     Update the global model W t = W t 1 + η γ N i n t w ˜ t , i ;
11:
     Calculate the privacy overhead for this round ε t ;
12:
     if   ε t > ε then
13:
   End model training;
14:
     end if
15:
     Calculate the score for this round s c o r e ;
16:
     if   s o c r e > 50   then
17:
    σ t = p σ t 1 ;
18:
     end if
19:
end for

4.3.2. Client-Side Training Process

First, the client determines the number of local update iterations, E, and downloads the global model W t 1 and noise parameters, σ t 1 , from the server. Then, the data subset involved in training is obtained by equal sampling the local dataset D i with the size of d, and the local model is trained, where the local data needs to be partitioned into data subsets, b, according to the batch size; the data subsets are iterated to train the model gradient and perform gradient clipping. Finally, noise is added to the model gradient, and the model is updated according to the local privacy preferences, standard privacy levels, and noise parameters distributed by the server. The updated model is then uploaded to the server. The algorithm of APDP-FL on the client side is as follows (Algorithm 2):
Algorithm 2 APDP-FL client-side algorithm
Input: 
local dataset D i , minimum iterative packet size d, number of local trainings E, loss function L , standard privacy protection level L, learning rate α , privacy preference R i , global model W t 1 , noise parameters σ t 1 ;
Output: 
local model after noise addition w ˜ t , i ;
  1:
Download the global model from the server W t 1 ;
  2:
w j , i = W t 1 ;                    ▹Update the local model
  3:
for   j = 1 , 2 , , E   do
  4:
      Divide the local dataset D i into subsets b of size d;
  5:
      for each batch b D i  do
  6:
             g b = L w j , i ; b ;
  7:
             g ¯ b = g b / m a x ( 1 , g b 2 / C ) ;            ▹ Perform gradient clipping
  8:
      end for
  9:
       σ i = R i L σ ;
10:
       g k = 1 r x r k g ¯ b + 2 C r N 0 , σ i 2 I ;          ▹ add noise to the model gradient
11:
       w j , i = w j , i α g k ;
12:
end for
13:
w ˜ t , i = w j , i ;
14:
Submit the noisy local model w ˜ t , i to the server;

5. Performance Analysis

5.1. Privacy Protection

APDP-FL satisfies ε , δ -DP; the proof is as follows.
Theorem 1.
Given the client’s local data sampling parameter q; the current training round T; the number of clients picked in each training round u; the number of local iterations F; and the Gaussian mechanism M f ( d ) = f ( d ) + N ( 0 , σ t 2 ) , where σ t is the Gaussian noise parameter of the standard privacy-preserving level at the t-th round of training, at α 2 if σ t is satisfied, we have the following:
t = 1 T ε ( α , σ t ) + log 1 δ α 1 ε
Then, APDP-FL satisfies ε , δ -DP.
Proof of Theorem 1. 
According to the RDP Gaussian mechanism [27], the Gaussian mechanism M f d = f d + N 0 , σ t 2 satisfies α , α 2 σ t 2 -RDP when the Gaussian noise parameter is σ t , which means ε α , σ t α 2 σ t 2 . Also, according to the subsampled mechanism of the RDP algorithm [28], the privacy overhead at each subsampled mechanism during a training session is as follows:
ε i ( α , σ t ) 1 α 1 log { 1 + q 2 α 2 m i n { 4 ( e 1 σ t 2 1 ) , e 1 σ t 2 m i n [ 2 , ( e 1 ) 2 ] } + j = 3 α q j α j e j ( j 1 ) 2 σ t 2 m i n [ 2 , e 1 j ] }
Since e has no upper bound, m i n [ 2 , e 1 2 ] 2 , m i n [ 2 , e 1 j ] 2 , and the equation above can be simplified into the following:
ε i ( α , σ t ) 1 α 1 log { 1 + q 2 α 2 m i n [ 4 ( e 1 σ t 2 1 ) , 2 e 1 σ t 2 ] + 2 j = 3 α q j α j e j ( j 1 ) 2 σ t 2 }
Since all subsampled operations satisfy α , ε i α , σ t -RDP, according to the RDP combination theorem, the privacy overhead after F rounds of local training is calculate as follows:
ε i ( α , σ t ) F α 1 log { 1 + q 2 α 2 m i n [ 4 ( e 1 σ t 2 1 ) , 2 e 1 σ t 2 ] + 2 j = 3 α q j α j e j ( j 1 ) 2 σ t 2 }
Therefore, when the number of clients is u, the privacy overhead of APDP-FL for one round of communication over the global dataset D is as follows:
ε i ( α , σ t ) F α 1 log { 1 + q 2 α 2 m i n [ 4 ( e 1 σ t 2 1 ) , 2 e 1 u σ t 2 ] + 2 j = 3 α q j α j e j ( j 1 ) 2 u σ t 2 }
After the T-round of training, the total privacy budget consumed by the entire global model is as follows:
t = 1 T F α 1 l o g { 1 + q 2 α 2 m i n [ 4 ( e 1 u σ t 2 1 ) , 2 e 1 u σ t 2 ] + 2 j = 3 α q j α j e j j 1 2 u σ t 2 } + l o g 1 δ α 1 ε
Let α , t = 1 T ε α , σ t be represented by DP. If σ t satisfies
t = 1 T ε α , σ t + l o g 1 δ α 1 ε
then APDP-FL satisfies
t = 1 T ε α , σ t + l o g 1 δ α 1 , δ - DP
If APDP-FL satisfies ε , δ -DP, the proof is complete. □
Although it has been proven that the proposed scheme satisfies ( ε , δ ) - differential privacy , attackers may still launch multiple attacks using model parameters, gradients, or aggregation results. The following is an analysis of the attack model:
(1)
Membership Inference Attack (MIA)
Attackers use model outputs (such as prediction confidence) to determine whether a certain piece of datum participated in training. For example, in medical federated learning, attackers may infer whether a specific patient participated in the training set of a disease study. Differential privacy reduces the success rate of MIA by adding noise to blur data correlations. However, attacks may still infer member identities based on the differences in statistical model outputs.
(2)
Reconstruction Attack
Attackers use model gradients or parameters to infer the original training data. For example, gradient inversion attack utilizes optimization algorithms to reconstruct image data from gradients, and gradient cropping and noise addition can increase the difficulty of reconstruction. But if the noise intensity is insufficient, attackers may still approach the real data through multiple iterations of optimization.
(3)
Model Inversion Attack
Attackers generate adversarial samples by querying the model’s API and inferring the distribution of training data. For example, in medical federated learning, attackers may use GANs to generate fake images that are similar to training data. However, although differential privacy reduces data distinguishability, it cannot completely prevent attackers from using generative models to simulate data distribution.
(4)
Backdoor Attack
The attacker implants specific triggers in the training data, causing the model to output preset results under specific inputs. For example, in image classification, attackers may add tiny markers in the corners of the image, forcing the model to classify it into a specific category. Differential privacy has no direct defense against backdoor attacks, and attackers may bypass noise interference by poisoning data, especially when triggers overlap with normal data distributions.

5.2. Complexity Analysis

Typically, the time complexity of federated learning algorithms is mainly related to several of its main processes, such as communication overhead, model aggregation, and client-side local training. In the APDP-FL proposed in this paper, assume that the communication overhead between the server and each client is O T during each round of learning, the maximum number of training rounds in the training process is R, and the server communicates with μ N users selected in each round. The time complexity overhead with respect to communication during the whole training process is O T × R × μ N . In the process of locally training the client, assume that the amount of training data of each client is D, the number of local number of iterations is F, and the size of the model parameters is M. The time overhead spent with respect to the client local is mainly a result of local training, adding Gaussian noise, and gradient clipping, and the time complexity is O D × M × F , O M , and O 1 respectively. In the server update model stage, the biparadigm of the model gradient needs to be computed when calculating the score, and its time complexity is proportional to the size of the model parameters, and its time complexity is O M × R . Thus, the total time complexity of the APDP-FL algorithm is O T × R × μ N + O D × M × F + O M + O M × R . The time complexity of the algorithm is within the acceptable range and has strong application value.

6. Experiment

The hardware environment for the experiments comprises AMD Ryzen 7 5800H 3.20 GHz, 16.00 RAM, Win11 64-bit, and an NVIDIA GeForce RTX 3050 GPU, and the software environment is as follows: compiled language: Python 3.9 version; IDE: Pycharm2023.3 version; deep learning framework and version: PyTorch1.8.0 and CUDA11.1.

6.1. Experimental Settings

(1)
Dataset Introduction
We set up a federated learning system with multiple clients and a central server for collaborative training until the global model converged; then, training stopped. In this paper, we use three commonly used image classification datasets, MNIST, Fashion-MNIST, and CIFAR-10, for the image classification task. The MNIST dataset consists of 60,000 training samples with 10,000 test samples, the size of each image is 28 × 28 , and the content of the image is handwritten numbers from 0 to 9. Fashion-MNIST similarly consists of 60,000 training samples with 10,000 test samples, the size of each image is 28 × 28 , and the image content includes 10 different kinds of goods. CIFAR-10 includes 60,000 natural images divided into 10 classes, and each image is 3 × 32 × 32 .
(2)
Neural Network Architecture
The experiment used a two-layer convolutional neural network on both the FMNIST dataset and the CIFAR10 dataset, as shown in Figure 3. The model structures on the two datasets are the same, but due to different inputs and outputs, the number of parameters in the model is slightly different.
(3)
Hyperparameter setting
The privacy preference levels selected by users are set to conform to normal distributions. The initial Gaussian noise parameters, σ , are 5 and 10; the learning rate α is 0.05; the number of participating users is 100; the noise reduction coefficient q is 0.90/0.94/0.98 (MNIST/Fashion-MNIST/CIFAR-10); the client’s selection rate in each round is 0.1. The maximum number of training rounds is 100, and the number of local training iterations is 5/10/20 (MNIST/Fashion-MNIST/CIFAR-10).
(4)
Test metrics
Accuracy is used as the evaluation index of model utility for comparison, which is calculated as follows:
A c c = A c c u r a t e n u m T o t a l n u m × 100 %
where A c c u r a t e n u m is the number of correctly predicted samples, and T o t a l n u m is the number of samples in the whole test set. In order to visualize the progress of federated learning in the process of privacy budget consumption, this paper stretches the privacy budget consumed by APDP-FL and the comparison scheme to a uniform scale for graphical comparison.
(5)
Baselines
The comparison schemes in the experiment include NoDP-FL, DP-FL, and DP-FLProx, where NoDP-FL [29] represents the federated learning scheme without adding noise; NoDP-FL represents no privacy budget consumption; DP-FL [30] represents the generalized federated learning scheme for differential privacy, which adds a unified noise; and DP-FLProx [31] represents the privacy-preserving scheme using differential privacy on FedProx.
(6)
Evaluation Metrics: Main task accuracy (MTA) [32] indicates the accuracy of a model with respect to its main (benign) task, which is used in this experiment to measure the performance of the defense method. The formula is as follows:
M T A = ( x , y ) M n M n × 100 %
where ( x , y ) represents the labels that are categorized correctly, and M n denotes the benign model. The larger the M T A , the better the performance of the method.

6.2. Experimental Performance Analysis

6.2.1. Experiments on the MNIST Dataset

The privacy preference levels selected by users are set to conform to the normal distribution. The initial Gaussian noise parameters, σ , are 5 and 10; the learning rate α is 0.05 ; the number of participating users is 100; the noise reduction coefficient q is 0.90 ; the client’s selection rate in each round is 0.1 . The maximum number of training rounds is 100, and the number of local training iterations is 5. The results of the experiments are shown as Figure 4 and Figure 5.
In the MNIST dataset, with the initial Gaussian noise parameter σ of 5 and the privacy budget ε of 2.5 , after exhausting the privacy budget, APDP-FL reaches a final model accuracy of ( 96.31 ± 0.17 ) %, while DP-FL reaches 94.27 % and DP-FedProx reaches 95.18 % . If no noise is added during the training process, the model’s accuracy can eventually reach 97.21 % after the model converges. From the communication cost perspective, APDP-FL has a total of 63 rounds of training and achieves 96.31 % before exhausting the privacy budget, which is just 0.9 % lower than NoDP-FL, which does not add noise and carries out training to make the model converge in its entirety.
With an initial Gaussian noise parameter σ of 10 and privacy budget ε of 2, APDP-FL can eventually reach ( 95.16 ± 0.14 ) % model accuracy after exhausting the privacy budget. DP-FL can eventually reach 93.11 % model accuracy after exhausting the privacy budget. DP-FedProx can eventually reach 94.23 % model accuracy after exhausting the privacy budget. From the perspective of communication cost, APDP-FL has a total of 76 rounds of training before exhausting the privacy budget and achieves 95.16 % , which is 2.05 % more accurate than the model trained by the DP-FL scheme under the addition of uniform noise during the training process.

6.2.2. Experiments on the Fashion-MNIST Dataset

The user-selected privacy preference levels are set to conform to the normal distribution. The initial Gaussian noise parameters, σ , are 5 and 10; the learning rate α is 0.05 ; the number of participating users is 100; the noise reduction coefficient q is 0.94; the selection rate of the client in each round is 0.1. The maximum number of training rounds is 100, and the number of local training iterations is 10. The experimental results are shown in Figure 6 and Figure 7.
In the Fashion-MNIST dataset, with the initial Gaussian noise parameter σ set to 5 and the privacy budget ε set to 2.5, the model accuracy of APDP-FL can finally reach ( 80.39 ± 0.25 ) % after exhausting the privacy budget. The model accuracy of DP-FL can finally reach 77.18% after exhausting the privacy budget. The model accuracy of DP-FedProx can finally reach 78.18% after exhausting the privacy budget. Model accuracy can finally reach 78.76%, and if no noise is added during the training process, the model’s accuracy can finally reach 82.17% after the model converges. From a communication cost perspective, APDP-FL has 72 rounds of training and achieves 80.39% before exhausting the privacy budget, which is only 1.78% lower than NoDP-FL; this does not add noise and carries out training so that the model converges in its entirety.
Setting the initial Gaussian noise parameter σ as 10 and the privacy budget ε as 2, the model accuracy of APDP-FL can finally reach ( 78.55 ± 0.22 ) % after exhausting the privacy budget. DP-FL can finally reach 75.92% after exhausting the privacy budget. DP-FedProx can finally reach 77.06% after exhausting the privacy budget. From the perspective of communication cost, APDP-FL performed a total of 80 rounds of training before exhausting the privacy budget and achieved 78.55%, which is 2.63% more accurate than the model trained by the DP-FL scheme under the addition of uniform noise during the training process.

6.2.3. Experiments on the CIFAR-10 Dataset

The privacy preference level selected by users is set to conform to the normal distribution. The initial Gaussian noise parameters, σ , are 5 and 10; the learning rate α is 0.05; the participating users are 100; the noise reduction coefficient q is 0.98; the selection rate of the clients in each round is 0.1. The maximum number of training rounds is 100, and the number of local training iterations is 20. The results of the experiments are shown in Figure 8 and Figure 9.
In the CIFAR-10 dataset, with the initial Gaussian noise parameter σ set to 5 and the privacy budget ε set to 2.5, the model accuracy of APDP-FL can finally reach ( 70.74 ± 0.26 ) % after exhausting the privacy budget. DP-FL can finally reach 69.11% after exhausting the privacy budget. DP-FedProx can finally reach 70.07% after exhausting the privacy budget, and if no noise is added during the training process, the model accuracy can eventually reach 74.06% after the model converges. From a communication cost perspective, APDP-FL underwent a total of 88 rounds of training and achieved 70.74% before exhausting the privacy budget, only 3.32% lower than NoDP-FL, which was trained without adding noise and converged the model in its entirety.
Setting the initial Gaussian noise parameter σ as 10 and the privacy budget ε as 2, APDP-FL can finally reach ( 66.32 ± 0.13 ) % model accuracy after exhausting the privacy budget (repeat the experiment five runs to obtain the average value and range of values). DP-FL can finally reach 63.53% model accuracy after exhausting the privacy budget. DP-FedProx can finally reach 65.66% model accuracy after exhausting the privacy budget. From a communication cost perspective, APDP-FL has a total of 96 rounds of training before exhausting the privacy budget, and it achieves 66.32%, which is 2.79% more accurate than the model trained by the DP-FL scheme with the addition of a uniform amount of noise during the training process.
We analyze APDP-FL’s experimental results to investigate how Gaussian noise parameter σ and the privacy budget ε impact model accuracy across varying levels of data heterogeneity. The experimental results demonstrate certain advantages compared to other similar methods after exhausting the privacy budget ε , which is about 2.05–2.79% more accurate than scheme using differential privacy. The specific quantitative of the results are shown in Table 1.

6.2.4. Comparison with Other Methods

This section evaluates the effectiveness of the APDP-FL method by comparing it with existing state-of-the-art approaches. Non-independent identically distributed (non-IID) data are generated on the CIFAR-10 and CIFAR-100 datasets, with data heterogeneity measured using the γ . The experimental comparison schemes are as follows: (1) LDP-FedProx [31], where FedProx incorporates local differential privacy; (2) LDP-FedGA [33], federated learning of generative adversarial networks with local differential privacy; (3) LDP-FedTweet [34], two-stage knowledge distillation with local differential privacy; (4) LDP-DKD-pFed [35], federated learning of personalized models with differential privacy. CIFAR-10 and CIFAR-100 employ Dirichlet coefficients of 0.05 and 0.1, respectively, to simulate heterogeneous datasets.
The experimental results demonstrate that APDP-FL exhibits a more pronounced advantage in model accuracy (as shown in Figure 10). On the heterogeneous CIFAR-10 dataset, APDP-FL can finally reach ( 74.52 ± 0.12 ) % model accuracy, and it achieves accuracy improvements ranging from 1.9% to 4.52% compared to other methods. On the heterogeneous CIFAR-100 dataset, APDP-FL can finally reach ( 44.98 ± 0.18 ) % model accuracy. Its improvement spans 0.89% to 3.96%, while APDP-FL requires significantly fewer communication rounds than alternative approaches. This fully demonstrates that, in a federated learning environment with the same number of communication rounds, APDP-FL achieves higher model accuracy.
Under identical privacy budgets in heterogeneous data scenarios, despite incorporating multiple complex techniques, APDP-FL achieves superior model accuracy due to its optimized adaptive local differential privacy mechanism and the synergistic effects of knowledge distillation and GANs on model training. Regarding computational overhead, while APDP-FL’s technical combination introduces increased computational complexity, its rapid convergence properties ensure that overall time consumption does not significantly increase compared to other methods when processing heterogeneous CIFAR-10 and CIFAR-100 datasets. It even demonstrates advantages in the complex heterogeneous CIFAR-100 training scenario, further highlighting APDP-FL’s comprehensive strengths in heterogeneous data federated learning.

6.2.5. Model Convergence

In order to evaluate the convergence speed and stability of the model during the training process, five methods (APDP-FL, LDP-FedProx, LDP-FedGA, LDP -FedTweet, and LDP-DKD-pFeed) were trained for 200 rounds, and the corresponding loss function values for each round were recorded. The loss curves were plotted, as shown in Figure 11.
The convergence experiment results show that APDP-FL can converge faster and maintain lower final loss values on all three datasets. Compared to methods such as LDP-FedProx and LDP-FedTweet, its training curve is smoother, indicating that this method has better optimization ability and convergence stability.

6.2.6. Computational Overhead and Communication Rounds

(1)
Comparison of computational overhead between APDP-FL and DP-FL
From Table 2, it can be seen that, in non-IID data, APDP-FL obtains the global model through local model weighting, and the optimization process requires more local iterations, resulting in slightly higher computational overhead than DP-FL. As the client data size increases (from 250 to 1000 client samples) or the local training epochs increase (from 1 to 10), the efficiency difference between the two methods gradually narrows. For example, when the client data volume is N = 1000, the computational cost of APDP-FL is only about 3–8% higher than that of DP-FL.
(2)
Comparison of communication rounds between baselines
From Table 3, it can be seen that due to the addition of noise parameters, the communication cost of the privacy scheme is about 5% higher than that of the non-privacy scheme (420 vs. 400 MB). The personalized plan increases costs by approximately 4.8% (440 vs. 420 MB) due to the transmission of more local adjustment information. If personalized needs need to be taken into account, personalized solutions may have slightly higher costs, but privacy protection is more precise.

7. Conclusions

In this paper, the APDP-FL algorithm is proposed to address the problems of reduced model accuracy and inability to meet the personalized privacy requirements of traditional differential privacy techniques in order to protect federated learning. In terms of theoretical contribution, the algorithm first scores each round of training according to the parameters generated during the training process, and then, it automatically adjusts the noise scale added in the next round according to the scoring results. Meanwhile, when the client locally trains, it will adjust the local privacy protection level according to its own privacy preference, which meets the personalized privacy needs of different users, and the usability and superiority of the algorithm are proven through a series of experiments. In terms of technological innovation, the algorithm introduces personalized privacy protection strategies based on adaptive noise mechanisms. By adjusting noise, participants can adjust the Gaussian noise parameters according to their own privacy preferences, solving the problem of insufficient or excessive protection caused by a unified privacy budget setting. The experimental results have shown that this strategy has a 0.9% lower accuracy than the noise free approach on the MNIST dataset (96.31% vs. 97.21%), and the difference is also controlled within 3.32% on the CIFAR-10 dataset.
Although APDP-FL provides different privacy levels, it does not take into account the accurate privacy budget for different users. However, the problem of accurate privacy budgets can be solved through adaptive noise allocation, personalized dynamic adjustment, and optimization strategies such as gradient clipping. Future directions include multidimensional evaluation systems, personalized measurement, cross-data source budget management, graph data privacy allocation, and intelligent dynamic adjustment.
Therefore, future work needs to provide accurate privacy budgets directly to the participating clients and calculate the coefficients to adjust the Gaussian noise parameters to provide more accurate personalized privacy protection services. It is also very important to analyze whether the personalized approach actually achieves the desired, varied levels of privacy protection for different clients as intended.

Author Contributions

Conceptualization, F.G. and R.W.; methodology, F.G.; software, R.W.; validation, R.W., J.W. and H.L.; formal analysis, H.L.; investigation, F.G.; data curation, J.W.; writing—original draft preparation, F.G.; writing—review and editing, F.G., R.W., C.Y. and Z.L.; supervision, J.W. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Natural Science Foundation of China under grant no. 61702316 and the Natural Science Foundation of Shanxi Province, China, under grant no. 20210302123338.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author, L.H.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sayal, A.; Jha, J.; Gupta, V.; Gupta, A.; Gupta, O.; Memoria, M. Neural networks and machine learning. In Proceedings of the 2023 IEEE 5th International Conference on Cybernetics, Cognition and Machine Learning Applications (ICCCMLA), Hamburg, Germany, 7–8 October 2023; IEEE: New York, NY, USA, 2023; pp. 58–63. [Google Scholar] [CrossRef]
  2. Peddiraju, V.; Pamulaparthi, R.R.; Adupa, C.; Thoutam, L.R. Design and Development of Cost-Effective Child Surveillance System using Computer Vision Technology. In Proceedings of the 2022 International Conference on Recent Trends in Microelectronics, Automation, Computing and Communications Systems (ICMACC), Hyderabad, India, 28–30 December 2022; IEEE: New York, NY, USA, 2022; pp. 119–124. [Google Scholar]
  3. Liu, P.; Yuan, W.; Fu, J.; Jiang, Z.; Hayashi, H.; Neubig, G. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 2023, 55, 1–35. [Google Scholar] [CrossRef]
  4. Li, J.; Li, D.; Xiong, C.; Hoi, S. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 12888–12900. [Google Scholar]
  5. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  6. Zhang, T.; Gao, L.; He, C.; Zhang, M.; Krishnamachari, B.; Avestimehr, A.S. Federated learning for the internet of things: Applications, challenges, and opportunities. IEEE Internet Things Mag. 2022, 5, 24–29. [Google Scholar] [CrossRef]
  7. Alazab, M.; Rm, S.P.; Maddikunta, P.K.R.; Gadekallu, T.R.; Pham, Q.V. Federated learning for cybersecurity: Concepts, challenges, and future directions. IEEE Trans. Ind. Inform. 2021, 18, 3501–3509. [Google Scholar] [CrossRef]
  8. Al-Huthaifi, R.; Li, T.; Huang, W.; Gu, J.; Li, C. Federated learning in smart cities: Privacy and security survey. Inf. Sci. 2023, 632, 833–857. [Google Scholar] [CrossRef]
  9. Geiping, J.; Bauermeister, H.; Dröge, H.; Moeller, M. Inverting gradients-how easy is it to break privacy in federated learning? Adv. Neural Inf. Process. Syst. 2020, 33, 16937–16947. [Google Scholar]
  10. Paillier, P. Public-key cryptosystems based on composite degree residuosity classes. In Proceedings of the International Conference on the Theory and Applications of Cryptographic Techniques, Prague, Czech Republic, 2–6 May 1999; Springer: Berlin/Heidelberg, Germany, 1999; pp. 223–238. [Google Scholar]
  11. Zhao, C.; Zhao, S.; Zhao, M.; Chen, Z.; Gao, C.Z.; Li, H.; Tan, Y.a. Secure multi-party computation: Theory, practice and applications. Inf. Sci. 2019, 476, 357–372. [Google Scholar] [CrossRef]
  12. Li, Y.; Yin, Y.; Gao, H.; Jin, Y.; Wang, X. Survey on privacy protection in non-aggregated data sharing. J. Commun. 2021, 42, 195–212. [Google Scholar]
  13. Hu, C.; Li, B. Maskcrypt: Federated learning with selective homomorphic encryption. IEEE Trans. Dependable Secur. Comput. 2024, 22, 221–233. [Google Scholar] [CrossRef]
  14. Kumbhar, H.R.; Rao, S.S. Federated learning enabled multi-key homomorphic encryption. Expert Syst. Appl. 2025, 268, 126197. [Google Scholar] [CrossRef]
  15. Wang, B.; Li, H.; Wang, J.; Guo, Y. Federated learning scheme for privacy-preserving of medical data. J. Xidian Univ. 2023, 50, 166–177. [Google Scholar]
  16. Cai, Y.; Ding, W.; Xiao, Y.; Yan, Z.; Liu, X.; Wan, Z. Secfed: A secure and efficient federated learning based on multi-key homomorphic encryption. IEEE Trans. Dependable Secur. Comput. 2023, 21, 3817–3833. [Google Scholar] [CrossRef]
  17. Tan, Z.; Le, J.; Yang, F.; Huang, M.; Xiang, T.; Liao, X. Secure and accurate personalized federated learning with similarity-based model aggregation. IEEE Trans. Sustain. Comput. 2024, 10, 132–145. [Google Scholar] [CrossRef]
  18. Ciucanu, R.; Delabrouille, A.; Lafourcade, P.; Soare, M. Secure Protocols for Best Arm Identification in Federated Stochastic Multi-Armed Bandits. IEEE Trans. Dependable Secur. Comput. 2022, 20, 1378–1389. [Google Scholar] [CrossRef]
  19. Kalapaaking, A.P.; Stephanie, V.; Khalil, I.; Atiquzzaman, M.; Yi, X.; Almashor, M. Smpc-based federated learning for 6g-enabled internet of medical things. IEEE Netw. 2022, 36, 182–189. [Google Scholar] [CrossRef]
  20. Xiao, Y.; Xu, L.; Wu, Y.; Sun, J.; Zhu, L. PrSeFL: Achieving Practical Privacy and Robustness in Blockchain-Based Federated Learning. IEEE Internet Things J. 2024, 11, 40771–40786. [Google Scholar] [CrossRef]
  21. Lyu, L.; Yu, H.; Ma, X.; Chen, C.; Sun, L.; Zhao, J.; Yang, Q.; Yu, P.S. Privacy and robustness in federated learning: Attacks and defenses. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 8726–8746. [Google Scholar] [CrossRef]
  22. Tang, X.; Peng, L.; Weng, Y.; Shen, M.; Zhu, L.; Deng, R.H. Enforcing Differential Privacy in Federated Learning via Long-Term Contribution Incentives. IEEE Trans. Inf. Forensics Secur. 2025, 20, 3102–3115. [Google Scholar] [CrossRef]
  23. Fukami, T.; Murata, T.; Niwa, K.; Tyou, I. Dp-norm: Differential privacy primal-dual algorithm for decentralized federated learning. IEEE Trans. Inf. Forensics Secur. 2024, 19, 5783–5797. [Google Scholar] [CrossRef]
  24. Sun, L.; Qian, J.; Chen, X. LDP-FL: Practical private aggregation in federated learning with local differential privacy. arXiv 2020, arXiv:2007.15789. [Google Scholar]
  25. Bhowmick, A.; Duchi, J.; Freudiger, J.; Kapoor, G.; Rogers, R. Protection against reconstruction and its applications in private federated learning. arXiv 2018, arXiv:1812.00984. [Google Scholar]
  26. Xie, R.; Li, C.; Yang, Z.; Xu, Z.; Huang, J.; Dong, Z. Differential privacy enabled robust asynchronous federated multitask learning: A multigradient descent approach. IEEE Trans. Cybern. 2025, 55, 3546–3559. [Google Scholar] [CrossRef]
  27. Mironov, I. Rényi differential privacy. In Proceedings of the 2017 IEEE 30th Computer Security Foundations Symposium (CSF), Santa Barbara, CA, USA, 21–25 August 2017; IEEE: New York, NY, USA, 2017; pp. 263–275. [Google Scholar]
  28. Wang, Y.X.; Balle, B.; Kasiviswanathan, S.P. Subsampled rényi differential privacy and analytical moments accountant. In Proceedings of the The 22nd International Conference on Artificial Intelligence and Statistics, PMLR, Okinawa, Japan, 16–18 April 2019; pp. 1226–1235. [Google Scholar]
  29. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial intelligence and Statistics, PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  30. Arachchige, P.C.M.; Bertok, P.; Khalil, I.; Liu, D.; Camtepe, S.; Atiquzzaman, M. Local differential privacy for deep learning. IEEE Internet Things J. 2019, 7, 5827–5842. [Google Scholar] [CrossRef]
  31. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2020, 2, 429–450. [Google Scholar]
  32. Zhuang, H.; Yu, M.; Wang, H.; Hua, Y.; Li, J.; Yuan, X. Backdoor federated learning by poisoning backdoor-critical layers. arXiv 2023, arXiv:2308.04466. [Google Scholar]
  33. Cong, Y.; Zeng, Y.; Qiu, J.; Fang, Z.; Zhang, L.; Cheng, D.; Liu, J.; Tian, Z. Fedga: A greedy approach to enhance federated learning with non-iid data. Knowl.-Based Syst. 2024, 301, 112201. [Google Scholar] [CrossRef]
  34. Wang, Y.; Wang, W.; Wang, X.; Zhang, H.; Wu, X.; Yang, M. FedTweet: Two-fold knowledge distillation for non-IID federated learning. Comput. Electr. Eng. 2024, 114, 109067. [Google Scholar] [CrossRef]
  35. Su, L.; Wang, D.; Zhu, J. DKD-pFed: A novel framework for personalized federated learning via decoupling knowledge distillation and feature decorrelation. Expert Syst. Appl. 2025, 259, 125336. [Google Scholar] [CrossRef]
Figure 1. APDP-FL system architecture diagram.
Figure 1. APDP-FL system architecture diagram.
Symmetry 17 02023 g001
Figure 2. Noise parameter calculation.
Figure 2. Noise parameter calculation.
Symmetry 17 02023 g002
Figure 3. Structure of the neural network model used in the experiment.
Figure 3. Structure of the neural network model used in the experiment.
Symmetry 17 02023 g003
Figure 4. Performance of APDP-FL on the MNIST dataset with initial Gaussian noise parameter σ = 5 and privacy budget ε = 2.5 .
Figure 4. Performance of APDP-FL on the MNIST dataset with initial Gaussian noise parameter σ = 5 and privacy budget ε = 2.5 .
Symmetry 17 02023 g004
Figure 5. Performance of APDP-FL on the MNIST dataset with initial Gaussian noise parameter σ = 10 and privacy budget ε = 2.0 .
Figure 5. Performance of APDP-FL on the MNIST dataset with initial Gaussian noise parameter σ = 10 and privacy budget ε = 2.0 .
Symmetry 17 02023 g005
Figure 6. Performance of APDP-FL on Fashion-MNIST with initial Gaussian noise parameter σ = 5 and privacy budget ε = 2.5 .
Figure 6. Performance of APDP-FL on Fashion-MNIST with initial Gaussian noise parameter σ = 5 and privacy budget ε = 2.5 .
Symmetry 17 02023 g006
Figure 7. Performance of APDP-FL on Fashion-MNIST with initial Gaussian noise parameter σ = 10 and privacy budget ε = 2.0 .
Figure 7. Performance of APDP-FL on Fashion-MNIST with initial Gaussian noise parameter σ = 10 and privacy budget ε = 2.0 .
Symmetry 17 02023 g007
Figure 8. Performance of APDP-FL on CIFAR-10 with initial Gaussian noise parameter σ = 5 and privacy budget ε = 2.5 .
Figure 8. Performance of APDP-FL on CIFAR-10 with initial Gaussian noise parameter σ = 5 and privacy budget ε = 2.5 .
Symmetry 17 02023 g008
Figure 9. Performance of APDP-FL on Fashion-MNIST with initial Gaussian noise parameter σ = 10 and privacy budget ε = 2.0 .
Figure 9. Performance of APDP-FL on Fashion-MNIST with initial Gaussian noise parameter σ = 10 and privacy budget ε = 2.0 .
Symmetry 17 02023 g009
Figure 10. Comparison between our method and SOTA methods on CIFAR-10 and CIFAR-100 datasets.
Figure 10. Comparison between our method and SOTA methods on CIFAR-10 and CIFAR-100 datasets.
Symmetry 17 02023 g010
Figure 11. Comparison of model convergence between APDP-FL and baselines.
Figure 11. Comparison of model convergence between APDP-FL and baselines.
Symmetry 17 02023 g011
Table 1. Test accuracy (%) of APDP-FL on different datasets for different initial Gaussian noise parameters, σ , and privacy budgets, ε .
Table 1. Test accuracy (%) of APDP-FL on different datasets for different initial Gaussian noise parameters, σ , and privacy budgets, ε .
MethodMNISTFashion-MNISTCIFAR-10
σ = 5 , ε = 2.5 σ = 10 , ε = 2.0 σ = 5 , ε = 2.5 σ = 10 , ε = 2.0 σ = 5 , ε = 2.5 σ = 10 , ε = 2.0
NoDP-FL97.2196.0882.1779.9674.0675.62
DP-FL94.2793.1177.1875.9269.1163.55
DP-FedProx95.1894.2378.1877.0670.0765.66
APDP-FL 96.31 ± 0.17 95.16 ± 0.14 80.39 ± 0.25 78.55 ± 0.22 70.74 ± 0.26 66.32 ± 0.13
Table 2. Comparison of computational overhead under different client numbers (seconds/rounds).
Table 2. Comparison of computational overhead under different client numbers (seconds/rounds).
Clients NumberDP-FLAPDP-FLOverhead Variance Rate
25012.413.15.5%
50010.811.34.4%
100010.410.83.7%
Table 3. Comparison of computational overhead under different client numbers (seconds/rounds).
Table 3. Comparison of computational overhead under different client numbers (seconds/rounds).
SchemesClient Upload Data SizeServer Distributed Data SizeTotal Communication CostAdditional Cost Sources
APDP-FL11.0 MB × 20 = 220 MB11.0 MB × 20 = 220 MB440 MBPersonalized noise parameters + model adjustment
FedAvg10 MB × 20 = 200 MB10 MB × 20 = 200 MB400 MBModel parameter transmission
NoDP-FL10.5 MB × 20 = 210 MB10.5 MB × 20 = 210 MB420 MBEncryption or redundant parameters
DP-FL10.5 MB × 20 = 210 MB10.5 MB × 20 = 210 MB420 MBAdditional noise parameters
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, F.; Wang, R.; Wang, J.; Yang, C.; Liu, Z.; Li, H. APDP-FL: Personalized Federated Learning Based on Adaptive Differential Privacy. Symmetry 2025, 17, 2023. https://doi.org/10.3390/sym17122023

AMA Style

Guo F, Wang R, Wang J, Yang C, Liu Z, Li H. APDP-FL: Personalized Federated Learning Based on Adaptive Differential Privacy. Symmetry. 2025; 17(12):2023. https://doi.org/10.3390/sym17122023

Chicago/Turabian Style

Guo, Feng, Ruoxu Wang, Jiuru Wang, Chen Yang, Zhuo Liu, and Hongtao Li. 2025. "APDP-FL: Personalized Federated Learning Based on Adaptive Differential Privacy" Symmetry 17, no. 12: 2023. https://doi.org/10.3390/sym17122023

APA Style

Guo, F., Wang, R., Wang, J., Yang, C., Liu, Z., & Li, H. (2025). APDP-FL: Personalized Federated Learning Based on Adaptive Differential Privacy. Symmetry, 17(12), 2023. https://doi.org/10.3390/sym17122023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop