Next Article in Journal
A Principal Component Analysis-Based Feature Optimization Network for Few-Shot Fine-Grained Image Classification
Previous Article in Journal
Solution Methods for the Multiple-Choice Knapsack Problem and Their Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Differential Privacy Cox-MLP Model Based on Federated Learning

1
Yunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming 650500, China
2
School of Mathematics and Computer Science, Yunnan Minzu University, Kunming 650504, China
*
Authors to whom correspondence should be addressed.
Current address: School of Mathematics and Statistics, Yunnan University, Kunming 650106, China.
Mathematics 2025, 13(7), 1096; https://doi.org/10.3390/math13071096
Submission received: 2 March 2025 / Revised: 21 March 2025 / Accepted: 24 March 2025 / Published: 27 March 2025

Abstract

:
In the data-driven healthcare sector, balancing privacy protection and model performance is critical. This paper enhances accuracy and reliability in survival analysis by integrating differential privacy, deep learning, and the Cox proportional hazards model within a federated learning framework. Traditionally, differential privacy’s noise injection often degrades model performance. To address this, we propose two adaptive privacy budget allocation strategies considering weight changes across neural network layers. The first, LS-ADP, utilizes layer sensitivity to assess the influence of individual layer weights on model performance and develops an adaptive differential privacy algorithm. The second, ROW-DP, comprehensively assesses weight variations and absolute values to propose a random one-layer weighted differential privacy algorithm. These algorithms provide differentiated privacy protection for various weights, mitigating privacy leakage while ensuring model performance. Experimental results on simulated and clinical datasets demonstrate improved predictive performance and robust privacy protection.

1. Introduction

With the advent of the big data era, various sectors are actively exploring the tremendous potential of data. In clinical diagnostics, accurately estimating the survival time of patients is a critical concern for both patients and healthcare professionals. Survival analysis is usually applied to address this issue, with the Cox proportional hazards model [1] serving as a classic tool having long played an important role in evaluating disease risks and predicting patients’ survival time. However, due to its relatively simple structure, the traditional Cox proportional hazards model is limited in its ability to capture complex relationships among influencing factors [2]. To overcome this limitation, Kvamme et al. [3] proposed a hybrid model that integrates the Cox proportional hazards model with a neural network. This approach retains the advantages of the Cox model while leveraging deep learning techniques to capture complex nonlinear relationships, thereby improving the prediction accuracy. Nevertheless, the above studies predominantly relied on all data being stored in a certain central location.
Due to the decentralized and sensitive nature of healthcare data, such data are often isolated across different institutions, complicating effective sharing and integration [4]. Federated learning, as an emerging distributed machine learning framework, offers the possibility to solve this problem. Federated learning allows different “centers”, a term used in this context to refer to medical institutions, to jointly train models without directly sharing the original data, thus enabling data collaboration while protecting privacy [5]. Jung et al. [6] proposed the LAFD framework, which consists of two parts: LFL-DFALS and AFL-LMTGR. LFL-DFALS reduces the computational and communication costs during the federated learning training process through direct feedback alignment and layer sampling. AFL-LMTGR handles the problem of stragglers through local model training and gradient rebalancing while ensuring data privacy. Amit Khanna et al. [7] explored how to utilize federated learning to accelerate the drug development for Parkinson’s disease. They achieved this by effectively sharing and analyzing data while protecting patients’ privacy. They emphasized the potential of digital health technologies in improving the assessment and monitoring of Parkinson’s disease. Amandeep Singh Bhatia et al. [8] proposed a federated learning framework based on quantum tensor networks, named FedQTN, for processing medical image data, demonstrating significant improvements in model performance while ensuring privacy. G. Yue et al. [9] introduced a specificity-aware federated learning framework that employs an adaptive aggregation mechanism (AdapAM) and a dynamic feature fusion strategy (DFFS) to deal with the model specificity and class imbalance issues in medical image classification. Although federated learning provides a new approach for data privacy protection, it also has some security risks. Studies by Zhu et al. [10] and Geiping et al. [11] demonstrated that local data can be inferred by analyzing parameter gradients, which implies that relying solely on federated learning is not sufficient to fully protect data privacy.
Currently, differential privacy, as a rigorous privacy protection model, achieves a balance between privacy protection and model performance by adding noise to the data and is a primary means of privacy protection within the federated learning framework [12]. The user-level differential privacy federated learning framework proposed by Geyer et al. [13] was further optimized by Zhao et al. [14] and Nuria et al. [15]. For the federated averaging algorithm (FedAvg), noise is added during the model aggregation stage to implement differential privacy. Subsequently, Abadi et al. [16] achieved significant improvement in differential privacy algorithms by proposing an adaptive noise addition mechanism that dynamically adjusts noise levels according to the sensitivity of model parameters. Dwork et al. [17] provided a significant basis for understanding the theoretical boundaries of the privacy–utility trade-off. Rieke et al. [18] focused on differential privacy protection in the medical field, where sensitive information such as genetic sequences and medical histories is prevalent. They proposed a hierarchical privacy protection strategy that processes the features of medical data with different sensitivity levels during the data preprocessing stage, applying strict differential privacy to genetic data while employing relatively looser protections for clinical symptom data. This strategy effectively protects patients’ privacy and improves the accuracy of disease prediction models. The Adap DPFL scheme proposed by Fu et al. [19] adaptively clips the gradients according to the gradient heterogeneity of training parameters and adds decreasing noise based on the gradient convergence. Experimental results indicated that this scheme outperforms previous methods on real datasets. The AdaDpFed algorithm proposed by Zhao et al. [20] is an adaptive federated differential privacy protocol designated for a non-independent and identically distributed environment. It can adaptively adjust perturbation parameters, client aggregation, and sampling size according to individual data distribution changes. The proof of convergence indicates the convergence rate of O ( 1 / T ) , and comparative experiments show that its performance is superior to other advanced protocols, providing an effective privacy protection solution for federated learning in such environment. The adaptive privacy protection framework proposed by Hu et al. [21] considers dynamic allocation of privacy budgets, focusing on the impact of weight changes in the input layer on model performance to dynamically adjust privacy budgets. However, the above methods did not consider the historical information of gradients and the impact of all parameter changes on the performance of data models during the training process.
To address these limitations, this paper proposes an adaptive differential privacy framework applicable to the deep learning Cox model in a federated learning environment. Our contributions are as follows.
  • We integrate deep learning, differential privacy, and the Cox proportional hazards model into a federated learning framework, proposing a novel privacy-preserving method for survival analysis. By ensuring data privacy, it addresses challenges such as uneven data distribution, difficulties in data sharing, and limited computational resources, thereby effectively supporting medical research and clinical decision making.
  • We comprehensively consider both the degree of dispersion and the magnitude of the absolute values of all weights. For weights that change significantly or have higher absolute values during the training process, we propose to add less noise; conversely, more noise can be added.
  • We quantify the impact of weight changes in different layers on model performance and allocate privacy budgets adaptively based on the impact of weight. This enables us to maintain the performance of the model as much as possible while protecting privacy.
The rest of this paper is organized as follows: In Section 2, we discuss the Cox-MLP algorithm based on federated learning with differential privacy. In Section 3, we introduce the methods for implementing adaptive privacy budget allocation in federated learning (FL). Section 4 and Section 5 present the experimental evaluation and results, and Section 6 gives the summary and conclusions.

2. Model and Notation

In this section, we delve into the integration of federated learning algorithms with differential privacy into the Cox-MLP model. First, we give some notations in Table 1.
In parametric models, the essence of data privacy fundamentally resides in the protection of model parameters. We consider a federated learning system comprising N centers, where each center i ( i = 1 , , N ) manages a local dataset D i = ( ξ i , T i , Δ i ) with n i samples. The survival times are denoted by T i = ( T i 1 , , T i n i ) , where T i j = min ( T i j * , C i j ) . Here, T i j * represents the true survival time, and C i j denotes the censoring time. The censoring indicators are given by Δ i = ( δ i 1 , , δ i n i ) , with δ i j = I ( T i j * C i j ) serving as the indicator function to distinguish between observed and censored survival times. Additionally, ξ i = { ξ i 1 , ξ i 2 , , ξ i n i } represents the covariates associated with center i, where ξ ij = { ξ i j 1 , ξ i j 2 , , ξ i j p } .

2.1. Cox-MLP Model

We first consider that, in the traditional Cox model, it is assumed that the relationship between the log hazard ratio and the covariates is linear. The hazard function of center i [1] is
h ( t | ξ ij ) = h 0 ( t ) exp g ( ξ ij ) , g ( ξ ij ) = β T ξ ij ,
where h 0 ( t ) is the baseline hazard function, exp g ( ξ ij ) is the relative hazard function, the covariate ξ ij = { ξ i j 1 , ξ i j 2 , , ξ i j p } is the covariate that affects the survival time, and β = ( β 1 , , β p ) T is the parameter vector. The likelihood function of center i is
L i = f T i j | ξ i j δ i j S T i j | ξ i j 1 δ i j = h T i j | ξ i j δ i j S T i j | ξ i j .
However, in complex problems, the Cox model will fail to meet the precision requirements. We consider achieving precise fitting of complex structures through neural networks. Katzman et al. [22] indicate that the Cox model combined with deep neural networks has better flexibility and outperforms the traditional Cox model in terms of the concordance index.
Similar to Kvamm et al. [3], we process the entire dataset in batches and introduce weights q j . Then the risk set R j is restricted to only include the individuals in the current batch, which is denoted as R ˜ j . At this time, the partial likelihood function and the loss function of the traditional Cox model of center i are redefined as
L c o x i = j exp g ( ξ ij ) q j r R ˜ j exp g ( ξ i r ) δ i j ,
l o s s i = 1 n i j = 1 l o g r R ˜ j exp g ( ξ i j ) g ( ξ i r ) .
For the combination of neural networks and the Cox proportional hazards model, after determining the loss function, the aim is to enhance the robustness of parameter estimation. A penalty term as (5) is incorporated to constrain the complexity of the model. This leads to a more precise estimation of g ( ξ i j ) ,
p e n a l t y i = λ j = 1 r R ˜ j | g ( ξ i r ) | .
The proportionality assumption of the Cox model is quite strict. However, the neural network parameterized Cox model cannot relax this constraint. To avoid this limitation, it is common to group the data according to categorical covariates and apply the stratified version of the Cox model [23]. Kvamm et al. [3] proposed an innovative method that does not require grouping, making the relative risk function exp [ g ( ξ i j ) ] depend on the covariates ξ i j and time t, thus constructing a more flexible interaction model of center i
h ( t | ξ i j ) = h 0 ( t ) exp [ g ( t , ξ i j ) ] .
Model (6) is regarded as a non-proportional relative risk model under dynamic covariate adjustment, and thus the loss function of center i is redefined as
L i ( θ ) = 1 n i j = 1 l o g r R ˜ j exp g ( T i j , ξ ij ) g ( T i r , ξ ir ) .
The above loss function still needs to take the penalty term (5) into consideration. By combining the Cox model with neural networks to estimate parameters and performing iterative optimization on the loss function (7), the parameter estimate g ^ ( T i j , ξ ij ) is obtained. Furthermore, the cumulative baseline hazard rate is accurately estimated through the Breslow estimator as
H ^ ( t | ξ ij ) = T i j t Δ H ^ 0 ( T i j ) exp g ^ ( T i j , ξ ij ) ,
H ^ 0 ( T i j ) = δ i j r R j exp g ^ ( T i r , ξ ir ) .
If it is necessary to evaluate the survival probability of patients, the survival function can be characterized by the cumulative baseline hazard rate as
S ^ ( t ) = exp [ H ^ ( t ) ] .

2.2. Federated Learning

Federated learning is a distributed machine learning framework designed to enable the training of local models using datasets distributed across multiple centers. These locally trained models are then aggregated to collaboratively construct a global model. Throughout the training process, only model parameters are exchanged among the participating centers, while the local datasets D i remain securely stored at their respective sources. This approach ensures the privacy of local data, enhances data security, and significantly mitigates the risk of privacy breaches. A typical federated learning system is illustrated in Figure 1. In this setup, centers with identical attribute spaces collaborate under the coordination of a parameter aggregation server to optimize and develop a high-performing global model [16].
Within this architecture, the algorithm iteratively processes data until predefined convergence criteria are met, or until the maximum number of iterations or training time is reached.The process of the federated learning algorithm we consider is shown in Algorithm 1.
Algorithm 1 Federated learning algorithm
Input: Server: Initial global model w 0 , total rounds T
   Centers ( E ) : Local datasets { D i } i = 1 N , local epochs E
Output: Optimized global model w T
for round t = 0 , 1 , , T 1  do
   Server executes:
      Broadcast current global model w t to all centers
   Centers parallel execute:
   Local training:
      Initialize local model: w t , 0 i w t
      for epoch e = 1  to E do
         Split D i into batches B k of size B
         for each batch B k  do
             w t , e i w t , e 1 i η L ( w t , e 1 i ; B k )
       Δ w t i = w t , E i w t
   Server aggregates:
      Receive updates { Δ w t i } i E
      Compute weighted average:
          w t + 1 w t + 1 N i E | D i | | D | Δ w t i
end

2.3. Differential Privacy

In this paper, we primarily focus on global differential privacy within the specific context of federated learning. The implementation of the Laplacian mechanism [20] is relatively intricate within this framework, whereas the Gaussian mechanism is considered more suitable due to its advantageous additivity property. We emphasize differential privacy under the Gaussian mechanism. To introduce the mechanism, we first briefly discuss the concept of ( ε , δ )-approximate differential privacy.
Definition 1
( ( ε , δ ) -approximate differential privacy). For neighboring datasets that differ in exactly one data record, there exists an algorithm g that its any output Y ( Y R a n g e ( g ) ) satisfies the inequality
P r g ( D ) = Y e ε P r g ( D ) = Y + δ ;
then the algorithm g satisfies ( ε , δ ) differential privacy, where P r denotes the probability of an event occurring, e is the natural base, and ε is a privacy budget.
Approximate differential privacy introduces an additional privacy parameter, δ , compared with strict differential privacy. This parameter controls the failure rate of differential privacy: the smaller the value, the higher the privacy. In general, a smaller value of δ will be taken to improve the usability of the algorithm under the same privacy protection conditions.
Definition 2
(Gaussian mechanism). The core concept of the Gaussian mechanism is to achieve differential privacy protection by adding a random noise that follows a specific Gaussian distribution to the output result of a function. For a given dataset D and an arbitrary function g : D R d , where the L 2 global sensitivity of g is Δ g , the algorithm
G ( D ) = g ( D ) + N ( 0 , σ 2 )
satisfies ( ε , δ ) differential privacy with ε < 1 , σ = Δ g 2 ln ( 1.25 δ ) ε

3. The Cox-MLP Based on FL with Differential Privacy

We consider the client–server architecture of the federated learning system, whose structure is shown in Figure 1. In such a system, N centers with the same attribute space, with the coordination and assistance of the parameter aggregation server [24], jointly collaborate to optimize the loss
l ( θ ) = φ i L i ( θ ) .
To ensure the privacy of individual data during the training process, similar to Wei [25], we adopt the differential privacy mechanism when uploading parameters from centers to the server. Specifically, to achieve ( ε , δ ) differential privacy, that is, for neighboring datasets D and D with exactly one different data record, any output result Y ( Y R a n g e ( g ) ) of the algorithm g satisfies inequality Pr [ g ( D ) = Y ] e ε Pr [ g ( D ) = Y ] + δ . We need to add noise from the normal distribution N 0 , σ U 2 to each parameter. The calculation formula for the standard deviation of the noise is
σ U = c L Δ g U ε ,
where c is a constant, and c 2 ln ( 1.25 / δ ) . L represents the maximum number of exposure times during the process of uploading local parameters to the server, and Δ g U is the global sensitivity. If C is the clipping scale of the gradient during the training process in the center, and the minimum value among the data amounts contained in each center is m, then the sensitivity Δ g U = 2 c m . However, Wei [24] shows that as the times of iterations increases, the risk of privacy leakage will also increase. Therefore, it is necessary to perturb the parameters before the centers upload them, and then the server aggregates the received parameters. In order to maintain the global ( ε , δ )-differential privacy, the server also needs to add Gaussian noise to the parameters according to the training process and then distribute the randomly perturbed parameters to each center. The added noise follows the following normal distribution N ( 0 , σ D 2 ) , where the standard deviation σ D is calculated according to the global iteration number T and the number of centers participating in the training N as
σ D = 2 c C T 2 L 2 N m N ε T > L N , 0 T L N .
In this manner, the Cox-MLP model demonstrates its superiority in not only fitting complex data relationships better but also protecting data privacy throughout the entire distributed training process. During the training process, the noise added for privacy protection will inevitably affect the performance of the model. To alleviate this problem, we then propose a method based on weight importance to dynamically adjust the magnitude of the noise in differential privacy.

3.1. Adaptive Differential Privacy Based on Layer Sensitivity (LS-ADP)

To balance privacy protection and model performance, in this paper, we propose two methods based on weight importance. The key to the method lies in quantifying the impact of each layer’s weights on the model performance and dynamically adjusting the magnitude of the noise in differential privacy according to this metric.
We consider adopting the idea of random perturbation. Our method evaluates the impact of each weight change on the model’s performance and adjusts the privacy budget allocation strategy dynamically based on this evaluation.
To reduce the computational load, instead of adding random perturbations to each weight w i j , we apply the same random noise to all weights in a layer, and add the same random perturbation to all the weights in this layer. Then we calculate the change in the model performance to evaluate the impact of the weight change in this layer. Specifically, after the local training is completed, we first calculate the evaluation index value p of the model without random perturbations as a benchmark. Under the condition that other weights remain unchanged, we add random noise x i to all the weights w i in the i-th layer of the model to obtain the weights w i , that is,
w i = w i + x i ,
x i N ( 0 , σ U 2 I ) ,
σ U = c L Δ g U ε ,
where w i = ( w i 1 , w i 2 , , w i m i ) , x i = ( x i 1 , x i 2 , , x i m i ) , and m i represents the number of weights in the i-th layer. c is a constant, and c 2 ln ( 1.25 / δ ) , ensuring that the Gaussian mechanism satisfies ( ε , δ ) -DP requirements. L is the maximum times of exposures during the process of uploading local parameters to the server. Δ g U is the global sensitivity, ε is the privacy budget, and δ is the probability of privacy leakage. Through the perturbed model, to re-evaluate the model performance, new evaluation index values p i ( i = 1 , 2 , 3 , K ) are obtained, where K is the number of layers of weights in the neural network. Calculate the performance change by
V i . = | p i p | i = 1 K | p i p |
to measure the impact of weight changes in this layer on model performance. The numerator | p i p | quantifies the absolute change in performance caused by the perturbation. This value reflects the sensitivity of the model performance to changes in that particular layer. Based on V i . , the privacy budget of weights in the ith layer can be obtained as
ε i . = ε V i . ,
where ε represents the total privacy budget. If V i . is relatively large, it indicates that even a slight change in the weights of this layer will lead to a significant change on the model performance. Therefore, a larger privacy budget should be allocated to the weights of this layer so that less noise will be added. Conversely, if V i . is relatively small, it means that the change in the weights of the layer causes a minor impact on the model performance. Thus, a smaller privacy budget is allocated; that is, more noise can be added.

3.2. Random One-Layer Weighted Differential Privacy Algorithm (ROW-DP)

During the training process of neural networks, the gradient descent method is commonly employed to update the weights. Weights that have a significant impact on the model performance usually undergo relatively large changes. Therefore, we take the degree of dispersion of each weight during the local iteration process as an evaluation indicator to assess its importance. Commonly used metrics include variance (Var), standard deviation (Std), range (Range), interquartile range (IQR), and mean absolute error (MAE). Furthermore, some weights may already be near optimal at initialization, exhibiting low dispersion and potentially being underestimated in their contribution to model performance.
To this end, a comprehensive importance measure V i j for the j-th weight w i j of the i-th layer is defined as
V i j = f ( w i j ) × | w i j m | f ( w i j ) × | w i j m | ,
where f ( w i j ) represents the measures of dispersion of the weight w i j , which indicates how much the weight fluctuates during the local iterations. w ij = ( w i j 1 , , w i j m ) , with m denoting the number of local iterations, and | w i j m | is the absolute value of the weight obtained from the last local iteration. Larger magnitudes can contribute more significantly to the model’s predictions. Based on the value of V i j , an adaptive privacy budget allocation can be achieved. Specifically, the privacy budget allocated to the weight w i j is
ε i j = ε V i j ,
where ε represents the total privacy budget. If V i j is relatively large, it indicates that the change of w i j has a significant impact on the model performance. Therefore, a larger privacy budget should be allocated to this weight, and less noise should be added. Conversely, a larger privacy budget should be allocated if V i j is relatively small.
When employing the above strategy for privacy protection, if weighted noise is added to the weights of all nodes, excessive noise may be added to the weights of some nodes, resulting in an unstable fitting process. Based on this, we propose a random one weighted layer method (ROW-DP) on top of the original algorithm. That is, the weights of one layer of the neural network are randomly selected to conduct the above adaptive privacy budget allocation, while the same Gaussian-mechanism differential privacy noise is added to the weights of other layers, enabling our algorithm to possess better stability during the training process while increasing randomness. The privacy budget allocation is as
ε i j = ε f ( w i j ) × w i j m f ( w i j ) × w i j m ,   if w i j in the selected layer ε 1 M i ,   if w i j not in the selected layer ,
where M i is the number of weights contained in the layer.
In this way, it can not only effectively protect data privacy but also minimize the impact of privacy protection measures on the model’s predictive performance. This adaptive noise addition strategy can achieve a balance between privacy protection and model utility. The specific steps are as shown in Algorithm 2, as detailed below, with the proofs provided in Appendix A.
Algorithm 2 Adaptive hierarchical differential privacy optimization framework
Input: Server: Initial global model w 0 , total rounds T
   Centers ( E ) : Local datasets { D i } i = 1 N ,local epochs E
Output: Optimized global model w T
for round t = 0 , 1 , , T 1  do
   Server executes:
      Broadcast current global model w t to all centers
   Centers parallel execute:
   Local training:
      Initialize local model: w t , 0 i w t
      for epoch e = 1  to E do
         Split D i into batches B k of size B
         for each batch B k  do
             w t , e i w t , e 1 i η L ( w t , e 1 i ; B k )
            For each weight w t i j in the neural network w t , E i :
               Calculate V t i j according to (14) or (16).
                ε t i j = ε · V t i j .
                w ( t + 1 ) i j w t i j + N ( 0 , c L Δ g U ε t i j )
            End
         End
   Server aggregates:
      Receive updates { w t + 1 i } i E
      Compute weighted average:
          w t + 1 w t + 1 N i E | D i | | D | w t + 1 i
         The noise variance σ 2 2 is calculated by (13), w t + 1 = w t + 1 + N ( 0 , σ 2 2 I ) ;
end
In the next section, we will validate the performance of our algorithm on both simulated and real datasets, comparing it with the NbAFL algorithm. NbAFL assigns equal privacy budgets to each weight before aggregation and broadcasting, without considering the differences between weights. Our algorithm, on the other hand, takes into account the importance of each weight, resulting in better performance in the numerical experiments.

4. Simulation Studies

In this section, we employ simulated data generation to validate the efficacy of our algorithm. We randomly generate the variables X 1 and X 2 from normal distributions, X 3 from a binomial distribution, and X 4 from uniform distributions, forming the covariate vector X = ( X 1 , X 2 , X 3 , X 4 ) T . Subsequently, we sample survival times from the hazard function h ( t | x ) = h 0 ( t ) exp [ g ( x ) ] and censored times from 3 7 h 0 ( t ) exp [ g ( x ) ] , resulting in data with a 30% censored rate.
For the simulation setup, we assume a baseline hazard function h 0 ( t ) = 1 and parameters β = [ 2.2 , 5.5 , 8.8 , 0 ] , with g ( x ) = β T x . We generate 50,000 data points and partition them into a training set (70%) and a testing set (30%). We allocate average privacy budgets of 0.6, 0.7, 0.8, and 0.9 to the weights. Using centers N = 5 and global iterations T = 50 , we train our model using both the random one-layer weighted differential privacy (ROW-DP) and the adaptive differential privacy based on layer sensitivity (LS-ADP) algorithms. To assess the performance of our proposed algorithm, we conducted 100 independent experiments and compared the results with the NbAFL algorithm [25]. The NbAFL algorithm ensures privacy protection by assigning an equal privacy budget to each weight during both the upload and broadcast phases. Furthermore, we used the Cox-MLP algorithm [22] without differential privacy (Npdp) as a baseline for comparison.
Given the complexities inherent in neural networks, the coefficients associated with each covariate cannot be directly obtained. Therefore, we primarily evaluate our model using the Brier Score and the Integrated Negative Log-Likelihood (INLL). The Brier Score (BS), which ranges from 0 to 1, measures the Mean Squared Error (MSE) between the predicted survival probabilities and the true survival status, with 0 indicating perfect prediction and 1 indicating no predictive value. In survival analysis, for each time point t, the Negative Log-Likelihood gauges the discrepancy between the model’s predicted survival probability and the observed outcome. The Integrated Negative Log-Likelihood, by integrating the Negative Log-Likelihood over all time points, provides a comprehensive evaluation metric, where a lower value implies superior model performance.
Figure 2 and Table 2 and Table 3 demonstrate that an increase in privacy budget leads to a gradual reduction in noise added during training, and then the influence of differential privacy on models diminishes progressively. Notably, when ε ≥ 0.8, the difference in INLL between ROW-DP/LS-ADP and Nopd narrows to within 0.5%, validating that the impact of DP on model performance becomes negligible at this stage. Under a low privacy budget ( ε = 0.6), the LS-ADP method achieves an INLL of 0.2560 and a BS of 0.0551, which are notably superior to NbAFL’s INLL of 0.2753 and BS of 0.0592. This performance advantage is primarily attributable to LS-ADP’s incorporation of weight importance, which enables a more nuanced and refined allocation of the privacy budget. ROW-DP, by employing random single-layer weighting, concentrates noise on layers with lower sensitivity, thereby reducing perturbations to critical parameters. For example, ROW-DP(Var) achieves an INLL of 0.2681 at ε = 0.6, representing a 2.6% (0.2681 vs. 0.2753)) improvement over NbAFL. LS-ADP, which dynamically adjusts noise based on layer sensitivity, further optimizes the allocation of the privacy budget. At ε = 0.6, LS-ADP’s BS is reduced by 3.2% (from 0.0569 to 0.0551) compared with ROW-DP(Var), indicating that its finer control over the distribution of noise provides a distinct advantage. The impact of different dispersion measurement methods on the performance of ROW-DP is significant. For example, at ε = 0.7, ROW-DP(MAE) achieves an INLL of 0.2683, which is 5.196 higher than ROW-DP(Std)(0.2553), indicating that MAE is more sensitive to noise. In contrast, ROW-DP(IQR) achieves an INLL of 0.2365 at ε = 1, which is close to Nopd (0.2309), suggesting that IQR effectively balances noise and model performance under high privacy budgets. This phenomenon suggests that selecting robust dispersion metrics, such as IQR, can further enhance the algorithm’s performance under stringent privacy constraints. Moreover, across all privacy budgets tested, ROW-DP and LS-ADP exhibited performance comparable to or superior to the NbAFL algorithm. Particularly with smaller privacy budgets, when noise levels are higher, ROW-DP and LS-ADP significantly surpass the NbAFL algorithm. This phenomenon can be primarily attributed to an enhanced privacy allocation strategy that mitigates the effects of differential privacy on the performance of the model.

5. A Real Example of LS-ADP and ROW-DP

To validate the utility of our algorithm, we utilize real-world data, the Free Light Chain (FLCHAIN) dataset compiled by T. Therneau et al. The FLCHAIN dataset encompasses 7874 entries, including 9 covariates, as well as survival time and survival status information at the last known contact for each patient. Additionally, 28% of these data points feature censored survival times. Initial data preprocessing includes handling missing covariate values through multiple imputation and subsequent standardization.
To gauge the effectiveness of our algorithm, we employ the C-index, as proposed by Harrell et al. [26]. This index measures the proportion of patient pairs for whom all available predictions align with actual outcomes, indicating the proportion of patients with lower predicted risks who actually exhibit longer survival times. However, the presence of censored data within the dataset can render certain patient pairs ineligible for C-index computation. Specifically, such pairs are deemed unusable and excluded if both patients are censored or if one patient’s censored time precedes the other’s survival time. The C-index ranges from 0 to 1, with 0.5 suggesting model performance equivalent to random guessing, indicating a lack of predictive ability beyond chance. For predictive evaluation, the dataset is randomly divided into an 80% training set and a 20% test set. The average privacy budgets are 3, 4, 5, and 7, and the experiment was conducted 100 times.
Table 4 and Figure 3 demonstrate that the addition of noise in differential privacy diminishes with an increasing privacy budget, which leads to a decreasing influence on model performance and a more stable convergence of the C-index. A comparison of various noise addition methods reveals that, in experiments utilizing instance survival data, the performance discrepancy between the NbAFL and ROW-DP algorithms is negligible. LS-ADP achieves 1.7% higher C-index than NbAFL at ε = 3 (0.7627 vs. 0.7361), narrowing to 0.3% at ε = 7 (0.7701 vs. 0.7676). For ε 5 , all algorithms reach > 98 % of Nopd’s performance (0.7701). LS-ADP maintains a > 0.762 C-index across all ε , outperforming ROW-DP variants by 2.1–5.3%. In contrast, the LS-ADP algorithm achieves superior performance, particularly in terms of convergence speed and stability when applied across diverse privacy budgets. Notably, when the privacy budget is low, the advantages of LS-ADP become more pronounced, enabling the accommodation of a higher-performance model while ensuring adequate privacy protection.

6. Conclusions

In this paper, we propose a novel privacy budget allocation mechanism. Relying on the relationship between historical changes in parameters and the model performance, we develop a new federated learning algorithm with differential privacy. The innovation of it lies in achieving adaptive privacy budget allocation based on the relationship between parameter changes and model performance. Our results demonstrate that, while maintaining a fixed total privacy budget, both proposed methods effectively balance the trade-off between global differential privacy protection and global model performance. Notably, under smaller privacy budgets, the ROW-DP and LS-ADP algorithms outperform existing methods, such as NbAFL, which lack well-considered differential privacy budget allocation mechanisms, in metrics such as Integrated Negative Log-Likelihood and Brier Score. Furthermore, LS-ADP demonstrates superior accuracy and stability, making it more suitable for scenarios requiring high model performance. On the other hand, ROW-DP, with its simplicity and lower computational overhead, is better suited for resource-constrained applications. The distinctive characteristics of these two methods offer flexible solutions to meet different practical demands. However, this study does not fully address the adaptability of privacy budget allocation in more complex network structures and in scenarios involving non-independent and identically distributed (Non-IID) data. Currently, the differential privacy protection mechanisms we are discussing are far from sufficient for ensuring privacy protection. Attackers may carry out membership inference attacks, model inversion attacks, attribute inference attacks, data re-identification attacks, and data poisoning attacks. Overall, although research on Cox-MLP differential privacy provides a powerful tool for protecting specific models, in the face of various potential privacy threats, we still need to comprehensively utilize multiple technical means (such as encryption, secure multi-party computation, and access control) to build a more comprehensive and robust data protection system. These limitations will be explored and addressed in future research.

Author Contributions

Conceptualization, J.N.; project administration, J.N.; methodology, J.N. and R.H.; formal analysis, J.N.; software, R.H. and J.N.; validation, J.N., R.H. and Q.Z.; visualization, W.L. and Q.Z.; writing—original draft, J.N., Q.Z., R.J., R.H. and W.L.; writing—review and editing, D.C., H.L., R.H. and J.N.; supervision, D.C.; resources, D.C. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key R&D Program of China (Grant No. 2022YFA1003701); Open Research Fund of Yunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University (Grant No. SMDAYB2023004); and Research and Innovation Project for Exempt Postgraduate Students, Yunnan University (Grant No. TM-23236998).

Data Availability Statement

The data presented in this study are available on request from the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proofs

The appendices provide a proof demonstrating the satisfaction of differential privacy by Algorithm 2.
We establish the proof that Algorithm 2 adheres to differential privacy. However, prior to presenting the proof, it is essential to introduce the following Lemma A1.
Lemma A1.
During the aggregation of federated learning models, the clipping scale of model gradients is denoted by C, while the amount of local data at the central hub is m. There are a total of N central hubs, each with equal aggregate weights. The global sensitivity of the model is expressed as
Δ g D = 2 C m 1 N .
Assume that the total privacy budget is ε ; for each weight between neural network nodes, the privacy budget assigned to each is ε i , resulting in ε = i = 1 M ε i , where i = 1 , 2 , M , and M represents the total number of weights.
Leveraging Lemma A1, the global differential privacy sensitivity is Δ w i D = 2 C m 1 N . For a given weight w i , during the noise addition for differential privacy, we have the process G i ( D ) = w i ( D ) + x i , with x i N ( 0 , σ i 2 ) as the Gaussian noise added.
Consequently,
Pr G i ( D ) = t Pr G i ( D ) = t = Pr w i ( D ) + x i = t Pr w i ( D ) + x i = t = Pr x = t w i ( D ) Pr x = t w i ( D ) = 1 2 π σ i e 1 2 σ i 2 ( t w i ( D ) ) 2 1 2 π σ i e 1 2 σ i 2 ( t w i ( D ) ) 2 = e 1 2 σ i 2 ( t w i ( D ) ) 2 e 1 2 σ i 2 ( t w i ( D ) ) 2 .
Given the objective function as a real-valued function, Δ w i ( D ) = Δ w i 1 ( D ) = Δ w i 2 ( D ) , where Δ w i 1 ( D ) and Δ w i 2 ( D ) represent L1 and L2 sensitivities, respectively. Hence, x i = t w i ( D ) , so that t w i ( D ) = x i + Δ S D , and Equation (A2) can be equivalently written as
Pr G i ( D ) = t Pr G i ( D ) = t = e 1 2 σ i 2 x i 2 e 1 2 σ i 2 ( x i + Δ S D ) 2 = e 1 2 σ i 2 ( 2 x i Δ w i D + Δ w i D 2 ) ;
hence, the condition e 1 2 σ i 2 ( 2 x i Δ w i D + Δ w i D 2 ) e ε i is required, where x i N ( 0 , σ i 2 ) . Consequently, strict differential privacy adherence is not feasible, leading to the adoption of approximate differential privacy, which allows for a degree of privacy leakage. From e 1 2 σ i 2 2 x i Δ w i D + Δ w i D 2 e ε i , we obtain
x i ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2
to manage the risk of privacy leakage within δ , ensuring that
Pr | x i | ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 δ .
Considering the symmetry of the normal distribution, we set x i 0 , which leads to
Pr x i ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 δ 2 .
Define t i = ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 and Pr x i t i δ 2 , which implies that
Pr x i t i = t i + 1 2 π σ i e 1 2 σ i 2 x i 2 d x i 1 2 π σ i t i + x i t e 1 2 σ i 2 x i 2 d x i .
Therefore,
Pr x i t i = σ i 2 π t i e 1 2 σ i 2 t i 2 δ 2 .
These results simplify to
ln t i σ i + t i 2 2 σ i 2 ln 2 2 π δ .
Substituting t i = ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 into Equation (A3) yields
ln ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 σ + ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 2 2 σ 2 ln 2 π 1 δ = ln 2 π 1 δ .
Consequently,
ln T 2 ln 1.25 δ ε i T 2 ln 1.25 δ + 1 2 2 T 2 ln 1.25 δ ε i + ε i 2 8 T 2 ln 1.25 δ ln 2 π 1 δ .
Let T 2 ln ( 1.25 δ 1.25 ) = b ; then Equation (A5) is equivalent to
ln b ε i b + 1 2 b 2 ε i + ε i 2 4 b 2 ln 2 π 1 δ .
For the first equation in (A6), if ε 10 , with T 1 , δ 1 , b 2 ln ( 1.25 ) 1 2 , and b > 16 , ln b ε i b > 0 . For the second equation in (A6), b 2 ε i + ε i 2 4 b 2 0 , given ε 10 , b 16 , we have b 2 ε i + ε i 2 4 b 2 > b 2 9.902 . Since b 2 2 ln 2 π 1 δ = 2 T 2 ln 1.25 δ 2 ln 2 π 1 δ > 2 T 2 ln 1.25 δ 2 ln 1.25 δ , to make b > 16 valid, we require T 5 . Thus,
2 T 2 ln 1.25 δ 2 ln 1.25 δ 32.064 ,
which is b 2 2 ln 2 π 1 δ > 32.064 ; thus, b 2 32.064 > 2 ln 2 π 1 δ . Then,
b 2 ε i + ε i 2 4 b 2 > b 2 9.901 > b 2 32.064 > 2 ln 2 π 1 δ .
Thus, 1 2 ( b 2 ε i + ε i 2 4 b 2 ) > ln 2 π 1 δ , and ln T 2 ln 1.25 δ ε i T 2 ln 1.25 δ > 0 , implying
ln T 2 ln 1.25 δ ε i T 2 ln 1.25 δ + 1 2 2 T 2 ln 1.25 δ ε i + ε i 2 8 T 2 ln 1.25 δ ln 2 π 1 δ
holds true. That is, when T 5 , ε i < 10 , Pr | x i | ε i σ i 2 Δ w i ( D ) Δ w i ( D ) 2 δ is valid. Furthermore, it can be said that Pr Pr G i ( D ) = t Pr G i ( D ) = t e ε i δ is valid, satisfying approximate differential privacy.
Therefore, for Algorithm 2, given that the added noise x i is independently and identically distributed, for all weights in the neural network, there exists
Pr Pr G D ¯ = t Pr G D ¯ = t = i = 1 M Pr G i ( D ) = t Pr G i ( D ) = t i = 1 M e E i = e i = 1 M E i = e E δ M δ ,
so Algorithm 2 achieves ( ε , δ )-differential privacy.

References

  1. Cox, D.R. Regression Models and Life-Tables. J. R. Stat. Soc. Ser. B Methodol. 1972, 34, 187–220. [Google Scholar] [CrossRef]
  2. Therneau, T.M.; Grambsch, P.M. Modeling survival data: Extending the Cox model. In Statistics for Biology and Health; Springer: New York, NY, USA, 2000. [Google Scholar]
  3. Kvamme, H.; Borgan, Ø.; Scheel, I. Time-to-event prediction with neural networks and Cox regression. J. Mach. Learn. Res. 2019, 20, 1–30. [Google Scholar]
  4. Yin, L.; Feng, J.; Xun, H.; Sun, Z.; Cheng, X. A privacy-preserving federated learning for multiparty data sharing in social IoTs. IEEE Trans. Netw. Sci. Eng. 2021, 8, 2706–2718. [Google Scholar]
  5. Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res. 2021, 5, 1–19. [Google Scholar] [PubMed]
  6. Jung, K.; Baek, I.; Kim, S.; Chung, Y.D. LAFD: Local—Differentially Private and Asynchronous Federated Learning with Direct Feedback Alignment. IEEE Access 2023, 11, 86754–86769. [Google Scholar] [CrossRef]
  7. Khanna, A.; Adams, J.; Antoniades, C.; Bloem, B.R.; Carroll, C.; Cedarbaum, J.; Cosman, J.; Dexter, D.T.; Dockendorf, M.F.; Edgerton, J.; et al. Accelerating Parkinson’s Disease drug development with federated learning approaches. NPJ Park. Dis. 2024, 10, 225. [Google Scholar]
  8. Bhatia, A.S.; Neira, D.E.B. Federated learning with tensor networks: A quantum AI framework for healthcare. Mach. Learn. Sci. Technol. 2024, 5, 045035. [Google Scholar] [CrossRef]
  9. Yue, G.; Wei, P.; Zhou, T.; Song, Y.; Zhao, C.; Wang, T.; Lei, B. Specificity-Aware Federated Learning with Dynamic Feature Fusion Network for Imbalanced Medical Image Classification. IEEE J. Biomed. Health Inform. 2024, 28, 6373–6383. [Google Scholar] [CrossRef] [PubMed]
  10. Zhu, L.; Liu, Z.; Han, S. Deep leakage from gradients. Adv. Neural Inf. Process. Syst. 2019, 32, 1–11. [Google Scholar]
  11. Geiping, J.; Bauermeister, H.; Dröge, H.; Moeller, M. Inverting gradients-how easy is it to break privacy in federated learning? Adv. Neural Inf. Process. Syst. 2020, 33, 16937–16947. [Google Scholar]
  12. Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical Secure Aggregation for Privacy-Preserving Machine Learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 1132–1147. [Google Scholar]
  13. Geyer, R.C.; Klein, T.; Nabi, M. Differentially Private Federated Learning: A Client Level Perspective. arXiv 2017, arXiv:1712.07557. [Google Scholar]
  14. Zhao, C.X.; Sun, Y.; Wang, D.G. Federated learning with Gaussian differential privacy. In Proceedings of the IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; pp. 296–301. [Google Scholar]
  15. Rodríguez-Barroso, N.; Stipcich, G.; Jiménez-López, D.; Ruiz-Millán, J.A.; Martínez-Cámara, E.; González-Seco, G.; Luzón, M.V.; Veganzones, M.A.; Herrera, F. Federated Learning and Differential Privacy: Software tools analysis, the Sherpa.ai FL framework and methodological guidelines for preserving data privacy. Inf. Fusion 2020, 64, 270–292. [Google Scholar] [CrossRef]
  16. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar]
  17. Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
  18. Rieke, N.; Hancox, J.; Li, W.; Milletari, F.; Roth, H.R.; Albarqouni, S.; Bakas, S.; Galtier, M.N.; Landman, B.A.; Maier-Hein, K.; et al. The future of digital health with federated learning. NPJ Digit. Med. 2020, 3, 1–7. [Google Scholar]
  19. Fu, J.; Chen, Z.; Han, X. Adap DP-FL: Differentially Private Federated Learning with Adaptive Noise. In Proceedings of the 2022 IEEE International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Wuhan, China, 9–11 December 2022; pp. 1–6. [Google Scholar]
  20. Zhao, Z.; Sun, Y.; Bashir, A.K.; Lin, Z. AdaDpFed: A Differentially Private Federated Learning Algorithm with Adaptive Noise on Non-IID Data. IEEE Trans. Consum. Electron. 2024, 70, 2536–2545. [Google Scholar]
  21. Hu, J.; Wang, Z.; Shen, Y.; Lin, B.; Sun, P.; Pang, X.; Liu, J.; Ren, K. Shield Against Gradient Leakage Attacks: Adaptive Privacy-Preserving Federated Learning. IEEE/ACM Trans. Netw. 2023, 32, 1407–1422. [Google Scholar] [CrossRef]
  22. Katzman, J.L.; Shaham, U.; Cloninger, A.; Bates, J.; Jiang, T.; Kluger, Y. DeepSurv: Personalized treatment recommender system using a Cox proportional hazards deep neural network. BMC Med. Res. Methodol. 2018, 18, 1–12. [Google Scholar] [CrossRef] [PubMed]
  23. Klein, J.P.; Moeschberger, M.L. Survival Analysis: Techniques for Censored and Truncated Data, 2nd ed.; Springer: New York, NY, USA, 2003. [Google Scholar]
  24. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2019, 10, 1–19. [Google Scholar]
  25. Wei, K.; Li, J.; Ding, M.; Ma, C.; Yang, H.H.; Farokhi, F.; Jin, S.; Quek, T.Q.; Poor, H.V. Federated learning with differential privacy: Algorithms and performance analysis. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3454–3469. [Google Scholar]
  26. Harrell, F.E.; Lee, K.L.; Mark, D.B. Multivariable prognostic models: Issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat. Med. 1996, 15, 361–387. [Google Scholar] [PubMed]
Figure 1. Client–server architecture.
Figure 1. Client–server architecture.
Mathematics 13 01096 g001
Figure 2. Under different privacy budgets ( ε = 0.6, 0.7, 0.8, 0.9, 1.0), Integrated Negative Log-Likelihood and Brier Score for Cox-MLP (Nopd) algorithm, LS-ADP algorithm, and ROW-DP algorithm based on five discrete degree measurement methods (Var, Range, Std, MAE, IQR).
Figure 2. Under different privacy budgets ( ε = 0.6, 0.7, 0.8, 0.9, 1.0), Integrated Negative Log-Likelihood and Brier Score for Cox-MLP (Nopd) algorithm, LS-ADP algorithm, and ROW-DP algorithm based on five discrete degree measurement methods (Var, Range, Std, MAE, IQR).
Mathematics 13 01096 g002aMathematics 13 01096 g002b
Figure 3. Function of C-index of interaction count using seven different colors to represent various privacy budget allocation methods.
Figure 3. Function of C-index of interaction count using seven different colors to represent various privacy budget allocation methods.
Mathematics 13 01096 g003
Table 1. Notations.
Table 1. Notations.
NotationDescriptionNotationDescription
NNumber of centers m i Number of weights in the i-th layer
D i Local dataset of center i w i Perturbed weights after adding random noise to the i-th layer
n i Number of samples in center i’s dataset Δ g Global sensitivity
ξ i Matrix of covariates associated with samples in center i’s dataset ε Privacy budget parameter
T i Vector of survival times for each sample in center i’s dataset δ Probability of privacy leakage
Δ i Vector of censoring indicators for each sample in center i’s dataset p i New evaluation index values
C i j Censoring time for sample j at center iKNumber of layers of weights in the neural network
β Parameter vector V i Privacy budget of weights in the ith layer
h 0 ( t ) Baseline hazard function V i j The jth weight w i j of the ith layer
R ˜ i Risk set restricted to only include individuals in the current batch w ij Measures of dispersion of the weight
q j Weight of dataset of Cox-MLPmNumber of local iterations
λ Regularization parameter w i j m Weight value obtained from the last local iteration
TTotal number of training rounds | w i j m | Absolute value of the local weight
η Learning rate for updates σ 2 2 Variance of the noise added to the weights
x i Random noise added to the i-th layer’s weights ϕ i Weight assigned to the update from center i
Table 2. Integrated Negative Log-Likelihood for Cox-MLP (Nopd), LS-ADP, and ROW-DP algorithms under different privacy budgets ( ε ) based on five dispersion measurement methods (Var, Range, Std, MAE, IQR).
Table 2. Integrated Negative Log-Likelihood for Cox-MLP (Nopd), LS-ADP, and ROW-DP algorithms under different privacy budgets ( ε ) based on five dispersion measurement methods (Var, Range, Std, MAE, IQR).
Privacy Budget0.60.70.80.91
NbAFL0.27530.26130.25810.24460.2359
ROW-DP(Std)0.27260.25530.25410.24980.2476
ROW-DP(Var)0.26810.25960.25720.25300.2456
ROW-DP(Range)0.26860.25940.25670.24640.2455
ROW-DP(MAE)0.26920.26830.25930.25150.2450
ROW-DP(IQR)0.26900.26000.25540.25280.2365
LS-ADP0.25600.25510.25500.24640.2455
Nopd0.23090.23090.23090.23090.2309
Table 3. Brier Score log-likelihood for Cox-MLP(Nopd), LS-ADP, and ROW-DP algorithms under different privacy budgets ( ε ) based on five dispersion measurement methods (Var, Range, Std, MAE, IQR).
Table 3. Brier Score log-likelihood for Cox-MLP(Nopd), LS-ADP, and ROW-DP algorithms under different privacy budgets ( ε ) based on five dispersion measurement methods (Var, Range, Std, MAE, IQR).
Privacy Budget0.60.70.80.91
NbAFL0.05920.05720.05490.05480.0542
ROW-DP(Std)0.05770.05600.05530.05430.0545
ROW-DP(Var)0.05690.05620.05550.05520.0546
ROW-DP(Range)0.05760.05670.05530.05480.0547
ROW-DP(MAE)0.05610.05580.05530.05480.0543
ROW-DP(IQR)0.05690.05690.05450.05450.0544
LS-ADP0.05510.05470.05420.05430.0543
Nopd0.05400.05400.05400.05400.0540
Table 4. C-index log-likelihood for Cox-MLP(Nopd), LS-ADP, and ROW-DP algorithms under different privacy budgets ( ε ) based on five dispersion measurement methods (Var, Range, Std, MAE, IQR).
Table 4. C-index log-likelihood for Cox-MLP(Nopd), LS-ADP, and ROW-DP algorithms under different privacy budgets ( ε ) based on five dispersion measurement methods (Var, Range, Std, MAE, IQR).
Privacy Budget3457
NbAFL0.73610.75600.76560.7676
Std0.73010.74840.75380.7550
Var0.72920.74380.74330.7573
Range0.72510.75220.75300.7568
MAE0.72310.74850.75150.7599
LS-ADP0.76270.76530.76780.7701
IQR0.72480.75150.75850.7593
Nopd0.77010.77010.77010.7701
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Niu, J.; He, R.; Zhou, Q.; Li, W.; Jiang, R.; Li, H.; Chen, D. Adaptive Differential Privacy Cox-MLP Model Based on Federated Learning. Mathematics 2025, 13, 1096. https://doi.org/10.3390/math13071096

AMA Style

Niu J, He R, Zhou Q, Li W, Jiang R, Li H, Chen D. Adaptive Differential Privacy Cox-MLP Model Based on Federated Learning. Mathematics. 2025; 13(7):1096. https://doi.org/10.3390/math13071096

Chicago/Turabian Style

Niu, Jie, Runqi He, Qiyao Zhou, Wenjing Li, Ruxian Jiang, Huimin Li, and Dan Chen. 2025. "Adaptive Differential Privacy Cox-MLP Model Based on Federated Learning" Mathematics 13, no. 7: 1096. https://doi.org/10.3390/math13071096

APA Style

Niu, J., He, R., Zhou, Q., Li, W., Jiang, R., Li, H., & Chen, D. (2025). Adaptive Differential Privacy Cox-MLP Model Based on Federated Learning. Mathematics, 13(7), 1096. https://doi.org/10.3390/math13071096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop