Privacy Auditing in Differential Private Machine Learning: The Current Trends
Abstract
:1. Introduction
- The implementation of differential privacy in consumer-use cases makes greater privacy awareness necessary, thus raising both data-related and technical concerns. As a result, privacy auditors are looking for scalable, transparent, and powerful auditing methods that enable accurate privacy assessment under realistic conditions.
- Auditing methods and algorithms have been researched and proven effective for DPML models. In general, auditing methods can be categorized according to privacy attacks. However, the implementation of sophisticated privacy auditing requires a comprehensive privacy-auditing methodology.
- Existing privacy-auditing techniques are not yet well adapted to specific tasks and models, as there is no clear consensus on the privacy loss parameters to be chosen, such as ϵ, algorithmic vulnerabilities, and complexity issues. Therefore, there is an urgent need for effective auditing schemes that can provide empirical guarantees for privacy loss.
- We systematically present types and techniques for privacy attacks in the context of differential privacy machine-learning modeling. Recent research on privacy attacks for privacy auditing is categorized into five main categories: membership inference attacks, data-poisoning attacks, model inversion attacks, model extraction attacks, and property inference.
- A structured literature review of existing approaches to privacy auditing in differential privacy is conducted with examples from influential research papers. The comprehensive process of proving auditing schemes is presented. An in-depth analysis of auditing schemes is provided, along with an abridged description paper of the papers.
2. Preliminaries
2.1. Differential Privacy Fundamentals
2.2. Differential Privacy Composition
2.3. Centralized and Local Models of Differential Privacy
2.4. Noise Injecting
3. Privacy Attacks
3.1. Overview
3.2. White-Box vs. Black-Box Attacks
3.3. Type of Attacks
3.3.1. Membership Inference Attack
- White-box membership inference.
- Black-box membership inference.
- Label-only membership inference.
- Transfer membership inference.
3.3.2. Data-Poisoning Attack
3.3.3. Model Inversion Attack
- Learning-based methods.
- White-box inversion attacks.
- Black-box inversion attacks.
- The gradient-based inversion attacks.
3.3.4. Model Extraction Attack
3.3.5. Property Inference Attacks
4. Privacy Auditing Schemes
4.1. Privacy Auditing in Differential Privacy
4.2. Privacy Auditing Techniques
4.3. Privacy Audits
- Membership inference auditing.
- Poisoning auditing.
- Model inversion auditing.
- Model extraction auditing.
- Property inference auditing.
4.3.1. Differential Privacy Auditing Using Membership Inference
- Black-box inference membership auditing: This approach relies solely on assessing the privacy guarantees of machine-learning models by evaluating their vulnerability to membership inference attacks (MIAs) without accessing the internal workings of the model.
4.3.2. Differential Privacy Auditing with Data Poisoning
4.3.3. Differential Privacy Auditing with Model Inversion
4.3.4. Differential Privacy Auditing Using Model Extraction
4.3.5. Differential Privacy Auditing Using Property Inference
5. Discussion and Future Research
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A
Privacy-Attack Methodology | Reference | Privacy Guarantees | Methodology and the Main Contribution |
---|---|---|---|
Membership inference auditing | |||
Black-box membership inference auditing | Song et al. [175] | Membership inference attack analysis: Investigates the vulnerability of adversarial robust DL to MIAs and shows that there are significant privacy risks despite the claimed robustness. | Methodology: Performs a comprehensive analysis of MIAs targeting robust models proposing new benchmark attacks that improve existing methods by leveraging prediction entropy and other metrics to evaluate privacy risks. Empirical evaluations show that even robust models can leak sensitive information about training data. Contribution: Reveals that adversarial robustness does not inherently protect against MIAs and challenges the assumption that such protection is sufficient for privacy. Introduces the privacy risk score, a new metric that quantifies the likelihood of an individual sample being part of the training set providing a more nuanced understanding of privacy vulnerabilities in ML models. |
Carlini et al. [103] | Analyzes the effectiveness of MIAs against ML models: Shows that existing metrics may underestimate the vulnerability of a model to MIAs. | Methodology: Introduces a new attack framework based on quantile regression of models’ confidence scores. Proposes a likelihood ratio attack (LiRA) that significantly improves TPR at low FNR. Contribution: Establishes a more rigorous evaluation standard for MIAs and presents a likelihood ratio attack (LiRA) method to increase the effectiveness of MIAs by improving the accuracy in identifying training data members. | |
Lu et al. [172] | Introduces a black-box estimator for DP: Allows domain experts to empirically estimate the privacy of arbitrary mechanisms without requiring detailed knowledge of these mechanisms. | Methodology: Combines different estimates of DP parameters with Bayes optimal classifiers. Proposes a relative DP framework that defines privacy with respect to a finite input set, T, which improves scalability and robustness. Contribution: Establishes a theoretical foundation for linking black-box poly-time parameter estimates to classifier performance and demonstrates the ability to handle large output spaces with tight accuracy bounds, thereby improving the understanding of privacy risks. Introduces a distributional DP estimator and compares its performance on different mechanisms. | |
Kazmi et al. [174] | Measuring privacy violations in DPML models: Introduces a framework for through MIAs without the need to retrain or modify the model. | Methodology: PANORAMIA uses generated data from non-members to assess privacy leakage, eliminating the dependency on in-distribution non-members included in the distribution from the same dataset. This approach enables privacy measurement with minimal access to the training dataset. Contribution: The framework was evaluated with various ML models for image and tabular data classification, as well as with large-scale language models, demonstrating its effectiveness in auditing privacy without altering existing models or their training processes. | |
Koskela et al. [176] | DP: Proposes a method for auditing require prior knowledge of the noise distribution or subsampling ratio in black-box settings. | Methodology: Uses a histogram-based density estimation technique to compare lower bounds for the total variance distance (TVD) between outputs from two neighboring datasets. Contribution: The method generalizes existing threshold-based membership inference auditing techniques and improves prior approaches, such as f-DP auditing, by addressing the challenges of accurately auditing the subsampled Gaussian mechanism. | |
Kutta et al. [185] | Rényi DP: Establishes new lower bounds for Rényi DP in black-box settings providing statistical guarantees for privacy leakage that hold with high probability for large sample sizes. | Methodology: Introduces a novel estimator for the Rényi divergence between the output distributions of algorithms. This estimator is converted into a statistical lower bound that is applicable to a wide range of algorithms. Contribution: The work pioneers the treatment of Rényi DP in black-box scenarios and demonstrates the effectiveness of the proposed method by experimenting with previously unstudied algorithms and privacy enhancement techniques. | |
Domingo-Enrich et al. [186] | DP: Proposes auditing procedures for different DP guarantees: -DP, -DP, and -Rényi DP. | Methodology: The regularized kern Rényi divergence can be estimated from random samples, which enables effective auditing even in high-dimensional settings. Contribution: Introduces relaxations of DP using the kernel Rényi divergence and its regularized version. | |
White-box membership inference auditing | Leino and Fredrikson [107] | Membership inference attack analysis: Introduces a calibrated attack that significantly improves the precision of membership inference | Methodology: Exploits the internal workings of deep neural networks to develop a white-box membership inference attack. Contribution: Demonstrates how MIAs can be utilized as a tool to quantify the privacy risks associated with ML models. |
Chen et al. [177] | DP: Evaluates the effectiveness of differential privacy as a defense mechanism by perturbating the model weights. | Methodology: Evaluate the differential private convolutional neural networks (CNNs) and Lasso regression model with and without sparsity. Contribution: Investigate the impact of sparsity on privacy guarantees in CNNs and regression models and provide insights into model design for improved privacy. | |
Black- and white-box membership inference auditing | Nasr et al. [47] | DP: Determines lower bounds on the effectiveness of MIAs against DPML models and shows that existing privacy guarantees may not be as robust as previously thought. | Methodology: Instantiates a hypothetical attacker that is able to distinguish between two datasets that differ only by a single example. Develops two algorithms, one for crafting these datasets and another for predicting which dataset was used to train a particular model. This approach allows users to analyze the impact of the attacker’s capabilities on the privacy guarantees of DP mechanisms such as DP-SGD. Contribution: Provides empirical and theoretical insights into the limitations of DP in practical scenarios. It is shown that existing upper bounds may not hold up under stronger attacker conditions, and it is suggested that better upper bounds require additional assumptions on the attacker’s capabilities. |
Tramèr et al. [49] | DP: Investigates the reliability of DP guarantees in an open-source implementation of a DL algorithm. | Methodology: Explores auditing techniques inspired by recent advances in lower bound estimation for DP algorithms. Performs a detailed audit of a specific implementation to assess whether it satisfies the claimed DP guarantees. Contribution: Shows that the audited implementation does not satisfy the claimed differential privacy guarantee with 99.9% confidence. This emphasizes the importance of audits in identifying errors in purported DP systems and shows that even well-established methods can have critical vulnerabilities. | |
Nasr et al. [42] | DP: Provides tight empirical privacy estimates. | Methodology: Adversary instantiation to establish lower bounds for DP. Contribution: Develops techniques to evaluate the capabilities of attackers, providing lower bounds that inform practical privacy auditing. | |
Sablayrolles et al. [22] | Membership inference attack analysis: Analyzes MIAs in both white-box and black-box settings and shows that optimal attack strategies depend primarily on the loss function and not on the model architecture or access type. | Methodology: Derives the optimal strategy for membership inference under certain assumptions about parameter distributions and shows that both white-box and black-box settings can achieve similar effectiveness by focusing on the loss function. Provides approximations for the optimal strategy, leading to new inference methods. Contribution: Establishes a formal framework for MIAs and presents State-of-the-Art results for various ML models, including logistic regression and complex architectures such as ResNet-101 on datasets such as ImageNet. | |
Shadow modeling membership inference auditing | Shokri et al. [52] | Membership inference attack analysis: Develop a MIA that utilizes a shadow training technique. | Methodology: Investigates membership inference attacks using black-box access to models. Contribution: Quantitatively analyzes how ML models leak membership information and introducing a shadow training technique for attacks. |
Salem et al. [112] | Membership inference attack analysis: Demonstrates that MIAs can be performed without needing to know the architecture of the target model or the distribution of the training data, highlighting a broader vulnerability in ML models. | Methodology: Introduces a new approach called “shadow training”. This involves training multiple shadow models that mimic the behavior of the target model using similar but unrelated datasets. These shadow models are used to generate outputs that inform an attack model designed to distinguish between training and non-training data. Contribution: Presents a comprehensive assessment of membership inference attacks across different datasets and domains, highlighting the significant privacy risks associated with ML models. It also suggests effective defenses that preserve the benefits of the model while mitigating these risks. | |
Memorization auditing | Yeom et al. [10] | Membership inference and attribute inference analysis: Analyzes how overfitting and influence can increase the risk of membership inference and attribute inference attacks on ML models, highlighting that overfitting is sufficient but not necessary for these attacks. | Methodology: Conducts both formal and empirical analyses to examine the relationship between overfitting, influence, and privacy risk. Introduces quantitative measures of attacker advantage that attempt to infer training data membership or attributes of training data. The study evaluates different ML algorithms to illustrate how generalization errors and influential features impact privacy vulnerability. Contribution: This work provides new insights into the mechanisms behind membership and attribute inference attacks. It establishes a clear connection between model overfitting and privacy risks, while identifying other factors that can increase an attacker’s advantage. |
Carlini et al. [102] | Membership inference attack analysis: Identifies the risk of unintended memorization in neural networks, especially in generative models trained on sensitive data, and shows that unique sequences can be extracted from the models. | Methodology: Develops a testing framework to quantitatively assess the extent of memorization in neural networks. It uses exposure metrics to assess the likelihood that specific training sequences will be memorized and subsequently extracted. The study includes hands-on experiments with Google’s Smart Compose system to illustrate the effectiveness of their approach. Contribution: It becomes clear that unintentional memorization is a common problem with different model architectures and training strategies, and it occurs early in training and is not just a consequence of overfitting. Strategies to mitigate the problem are also discussed. These include DP, which effectively reduces the risk of memorization but may introduce utility trade-offs. | |
Label-only membership inference auditing | Malek et al. [178] | Label differential privacy: Proposes two new approaches—PATE (Private Aggregation of Teacher Ensembles) and ALIBI (additive Laplace noise coupled with Bayesian inference)—to achieve strong label differential privacy (LDP) guarantees in machine-learning models. | Methodology: Analyzes and compares the effectiveness of PATE and ALIBI in the delivering LDP. It demonstrates how PATE leverages a teacher–student framework to ensure privacy, while ALIBI is more suitable for typical ML tasks by adding Laplacian noise to the model outputs. The study includes a theoretical analysis of privacy guarantees and empirical evaluations of memorization properties for both approaches. Contribution: It demonstrates that traditional comparisons of algorithms based solely on provable DP guarantees can be misleading, advocating for a more nuanced understanding of privacy in ML. Additionally, it illustrates how strong privacy can be achieved with the proposed methods in specific contexts. |
Choquette-Choo et al. [179] | Membership inference attack analysis: Introduces attacks that infer membership inference based only on labels and evaluate model predictions without access to confidence scores and shows that these attacks can effectively infer membership status. | Methodology: It proposes a novel attack strategy that evaluates the robustness of a model’s predicted labels in the presence of input perturbations such as data augmentation and adversarial examples. It is empirically confirmed that their label-only attacks are comparable to traditional methods that require confidence scores. Contribution: The study shows that existing protection mechanisms based on confidence value masking are insufficient against label-only attacks. The study also highlights that training with DP or strong L2 regularization is a currently effective strategy to reduce membership leakage, even for outlier data points. | |
Single-training membership inference auditing | Steinke et al. [43] | DP: Proposes a novel auditing scheme for DPML systems that can be performed with a single training run and increases the efficiency of privacy assessments. | Methodology: It utilizes the ability to independently add or remove multiple training examples during a single training run. It analyzes the relationship between DP and statistical generalization to develop its auditing framework. This approach can be applied in both black-box and white-box settings with minimal assumptions about the underlying algorithm. Contribution: It provides a practical solution for privacy auditing in ML models without the need for extensive retraining. This reduces the computational burden while ensuring robust privacy assessment. |
Andrew et al. [118] | DP: Introduces a novel “one-shot” approach for estimating privacy loss in federated learning. | Methodology: Develops a one-shot empirical privacy evaluation method for federated learning. Contribution: Provides a method for estimating privacy guarantees in federated learning environments using a single training run, improving the efficiency of privacy auditing in decentralized environments without a priori knowledge of the model architecture, tasks or DP training algorithm. | |
Annamalai et al. [109] | DP: Proposes an auditing procedure for the Differentially Private Stochastic Gradient Descent (DP-SGD) algorithm that provides tighter empirical privacy estimates compared to previous methods, especially in black-box settings. | Methodology: It introduces a novel auditing technique that crafts worst-case initial model parameters, which significantly affects the privacy analysis of DP-SGD. Contribution: This work improves the understanding of how the initial parameters affect the privacy guarantees in DP-SGD and provides insights for detecting potential privacy violations in real-world implementations, improving the robustness of differential privacy auditing. | |
Loss-based membership inference auditing | Wang et al. [111] | DP: Introduces a new differential privacy paradigm called estimate–verify–release (EVR). | Methodology: Develops a randomized privacy verification procedure using Monte Carlo techniques and proposes an estimate–verify–release (EVR) paradigm. Contribution: Introduces a tight and efficient auditing procedure that converts estimates of privacy parameters into formal guarantees, allowing for effective privacy accounting with only one training run and averages the concept of Privacy Loss Distribution (PLD) to more accurately measure and track the cumulative privacy loss through a sequence of computations. |
Confidence score membership inference auditing | Askin et al. [182] | DP: Introduces a statistical method for quantifying differential privacy in a black-box setting, providing estimators for the optimal privacy parameter and confidence intervals. | Methodology: Introduces a local approach for the statistical quantification of DP in a black-box setting. Contribution: Develops estimators and confidence intervals for optimal privacy parameters, avoiding event selection issues and demonstrating fast convergence rates through experimental validation. |
Metric-based membership inference auditing | Rahman et al. [180] | DP: Examines the effectiveness of differential privacy in protecting deep-learning models against membership inference attacks. | Methodology: Investigates MIAs on DPML models through membership inference. Contribution: Analyzes the vulnerability of DP models to MIAs and shows that they can still leak information about training data under certain conditions, using accuracy and F-score as privacy leakage metrics. |
Liu et al. [169] | DP: Focuses on how differential privacy can be understood through hypothesis testing. | Methodology: Explores statistical privacy frameworks through the lens of hypothesis testing. Contribution: Provides a comprehensive analysis of privacy frameworks, emphasizing the role of hypothesis testing in evaluating privacy guarantees in ML models, linking precision, recall, and F-score metrics to the privacy parameters; and uses hypothesis testing techniques. | |
Balle et al. [170] | Rényi DP: Explores the relationship between differential privacy and hypothesis testing interpretations. | Methodology: Examines hypothesis testing interpretations in relation to Rényi DP. Contribution: Establishes connections between statistical hypothesis testing and Rényi differential privacy, improving the theoretical understanding of privacy guarantees in the context of ML. | |
Humphries et al. [181] | Membership inference attack analysis: Conducts empirical evaluations of various DP models across multiple datasets to assess their vulnerability to membership inference attacks. | Methodology: Analyzes the limitations of DP in the bounding of MIAs. Contribution: Shows that DP does not necessarily prevent MIAs and points out vulnerabilities in current privacy-preserving techniques. | |
Ha et al. [41] | DP: Investigates how DP can be affected by MIAs. | Methodology: Analyzes the impact of MIAs on DP mechanisms. Contribution: Examines how MIAs can be used as an audit tool to quantify training data leaks in ML models and proposes new metrics to assess vulnerability disparities across demographic groups. | |
Data augmentation-based auditing | Kong et al. [184] | Membership inference attack analysis: Investigates the relationship between forgeability in ML models and the vulnerability to MIAs and uncovers vulnerabilities that can be exploited by attackers. | Methodology: It proposes a framework to analyze forgeability—defined as the ability of an attacker to generate outputs that mimic a model’s behavior—and its connection to membership inference. It conducts empirical evaluations to show how certain model properties influence both forgeability and the risk of MIAs. Contribution: It shows how the choice of model design can inadvertently increase vulnerability to MIAs. This suggests that understanding forgeability can help in the development of secure ML systems. |
Data-poisoning auditing | |||
Influence-function analysis | Koh and Ling [187] | Model Interpretation: Investigates how influence functions can be used to trace predictions back to training data and thus gain insight into the behavior of the model without direct access to the internal workings of the model. | Methodology: Uses influence functions from robust statistics to find out which training points have a significant influence on a particular prediction. Develops an efficient implementation that only requires oracle access to gradients and Hessian-vector products, allowing scalability in modern ML contexts. Contribution: Demonstrates the usefulness of influence functions for various applications, including understanding model behavior, debugging, detecting dataset errors, and creating attacks on training sets, improving the interpretability of black-box models. |
Jayaraman and Evans [21] | DP: Investigates the limitations of DPML, particularly focusing on the impact of the privacy parameter on privacy leakage. | Methodology: Evaluates the practical implementation of differential privacy in machine-learning systems. Contribution: Conducts an empirical analysis of differentially private machine-learning algorithms, assessing their performance and privacy guarantees in real-world applications. | |
Lu et al. [61] | DP: Focuses on the auditing of DPML models for the empirical evaluation of privacy guarantees. | Methodology: Proposes a general framework for auditing differentially private machine-learning models. Contribution: Introduces a comprehensive tight auditing framework that assesses the effectiveness and robustness of differential privacy mechanisms in various machine-learning contexts. | |
Gradient manipulation in DP training. | Chen et al. [188] | Gradient leakage analysis: Investigates the potential for training data leakage from gradients in neural networks, highlighting that gradients can be exploited to reconstruct training images. | Methodology: Analyzes training-data leakage from gradients in neural networks for image classification. Contribution: Provides a theoretical framework for understanding how training data can be reconstructed from gradients, proposing a metric to measure model security against such attacks. |
Xie et al. [189] | Generalization improvement: Focuses on improving generalization in DL models through the manipulation of stochastic gradient noise (SGN). | Methodology: Introduces Positive–Negative Momentum (PNM) to manipulate stochastic gradient noise for improved generalization in machine-learning models. Contribution: Proposes a novel approach that demonstrates the convergence guarantees and generalization of the model using PNM approach that leverages stochastic gradient noise more effectively without increasing computational costs. | |
Ma et al. [54] | DP: Investigates the resilience of differentially private learners against data-poisoning attacks. | Methodology: Designs specific attack algorithms targeting two common approaches in DP, objective perturbation and output perturbation. Contribution: Analyzes vulnerabilities of differentially private models to data-poisoning attacks and proposes defensive strategies to mitigate these risks. | |
Jagielski et al. [46] | DP: Investigates the practical privacy guarantees of Differentially Private Stochastic Gradient Descent (DP-SGD). | Methodology: Audits differentially private machine-learning models, specifically examining the privacy guarantees of stochastic gradient descent (SGD). Contribution: Evaluates the effectiveness of differential privacy mechanisms in SGD, providing insights into how private the training process really is under various conditions. | |
Empirical evaluation of privacy loss. | Steinke and Ullman [191] | DP: Establishes a new lower bound on the sample complexity of differentially private algorithms for accurately answering statistical queries. | Methodology: Derives a necessary condition for the number of records, n, required to satisfy differential privacy while achieving a specified accuracy. Contribution: Introduces a framework that interpolates between pure and approximate differential privacy, providing optimal sample size requirements for answering statistical queries in high-dimensional databases. |
Kairouz et al. [192] | DP: Presents a new approach for training DP models without relying on sampling or shuffling, addressing the limitations of Differentially Private Stochastic Gradient Descent (DP-SGD). | Methodology: Proposes a method for practical and private deep learning without relying on sampling through shuffling techniques. Contribution: Develops auditing procedure for evaluating the effectiveness of shuffling in DPML models by leveraging various network parameters and likelihood ratio functions. | |
Privacy violation | Li et al. [193] | Information privacy: Reviews various theories related to online information privacy, analyzing how they contribute to understanding privacy concerns. | Methodology: Conducts a critical review of theories in online information privacy research and proposes an integrated framework. Contribution: Conducts a critical review of theories in online information privacy research and proposes an integrated framework. |
Hay et al. [194] | DP: Emphasizes the importance of rigorous evaluation of DP algorithms. | Methodology: Develops DPBench, a benchmarking suite for evaluating differential privacy algorithms. Contribution: Propose a systematic benchmarking methodology that includes various metrics to evaluate the privacy loss, utility, and robustness of algorithms with different privacy. | |
Ding et al. [45] | DP: Addresses the issue of verifying whether algorithms claiming DP actually adhere to their stated privacy guarantees. | Methodology: Develops a statistical approach to detect violations of differential privacy in algorithms. Contribution: Proposes the first counterexample generator that produces human-understandable counterexamples specifically designed to detect violations to DP in algorithms. | |
Wang et al. [195] | DP: Introduces CheckDP, an automated framework designed to prove or disprove claims of DP for algorithms. | Methodology: Utilizes a bidirectional Counterexample-Guided Inductive Synthesis (CEGIS) approach embedded in CheckDP, allowing it to generate proofs for correct systems and counterexamples for incorrect ones. Contribution: Presents an integrated approach that automates the verification process for differential privacy claims, enhancing the reliability of privacy-preserving mechanisms. | |
Barthe et al. [196] | DP: Addresses the problem of deciding whether probabilistic programs satisfy DP when restricted to finite inputs and outputs. | Methodology: Develops a decision procedure that leverages type systems and program analysis techniques to check for differential privacy in a class of probabilistic computations. Contribution: Explores theoretical aspects of differential privacy, providing insights into the conditions under which differential privacy can be effectively decided in computational settings. | |
Niu et al. [165] | DP: Presents DP-Opt, a framework designed to identify violations of DP in algorithms by optimizing for counterexamples. | Methodology: Utilizes optimization techniques to search for counterexamples that demonstrate when the lower bounds on differential privacy exceed the claimed values. Contribution: Develops a disprover that searches for counterexamples where the lower bounds on differential privacy exceed claimed values, enhancing the ability to detect and analyze privacy violations in algorithms. | |
Lokna et al. [48] | DP: pairs can be grouped, as they result in the same algorithm. | Contribution: Introduces Delta-Siege, an auditing tool that efficiently discovers violations of differential privacy across multiple claims simultaneously, demonstrating superior performance compared to existing tools and providing insights into the root causes of vulnerabilities. | |
Model inversion auditing | |||
Sensitivity analysis. | Frederikson et al. [100] | Model inversion attack analysis: Explores vulnerabilities in ML models through model inversion attacks that exploit confidence information and pose significant risks to user privacy. | Methodology: A new class of model inversion attacks is developed that exploits the confidence values given next to the predictions. It empirically evaluates these attacks in two contexts: decision trees for lifestyle surveys and neural networks for face recognition. The study includes experimental results that show how attackers can infer sensitive information and recover recognizable images based solely on model outputs. Contribution: It demonstrates the effectiveness of model inversion attacks in different contexts and presents basic countermeasures, such as training algorithms that obfuscate confidence values, that can mitigate the risk of these attacks while preserving the utility. |
Wang et al. [136] | DP: Proposes a DP regression model that aims to protect against model inversion attacks while preserving the model utility. | Methodology: A novel approach is presented that utilizes the functional mechanism to perturb the coefficients of the regression model. It analyzes how existing DP mechanisms cannot effectively prevent model inversion attacks. It provides a theoretical analysis and empirical evaluations showing that their approach can balance privacy for sensitive and non-sensitive attributes while preserving model performance. Contribution: It demonstrates the limitations of traditional DP in protecting sensitive attributes in model inversion attacks and presents a new method that effectively mitigates these risks while ensuring that the utility of the regression model is preserved. | |
Hitaj et al. [197] | Information leakage analysis: Investigates vulnerabilities in collaborative DL models and shows that these models are susceptible to information leakage despite attempts to protect privacy through parameter sharing and DP. | Methodology: Develops a novel attack that exploits the real-time nature of the learning process in collaborative DL environments. They show how an attacker can train a generative adversarial network (GAN) to generate prototypical samples from the private training data of honest participants. It criticizes existing privacy-preserving techniques, particularly record-level DP at the dataset level, and highlights their ineffectiveness against their proposed attack. Contribution: Reveals fundamental flaws in the design of collaborative DL systems and emphasizes that current privacy-preserving measures do not provide adequate protection against sophisticated attacks such as those enabled by GANs. It calls for a re-evaluation of privacy-preserving strategies in decentralized ML settings. | |
Song et al. [198] | Model inversion attack analysis: Investigates the risks of overfitting in ML models and shows that models can inadvertently memorize sensitive training data, leading to potential privacy violations. | Methodology: Analyzes different ML models to assess their vulnerability to memorization attacks. Introduces a framework to quantify the amount of information a model stores about its training data and conduct empirical experiments to illustrate how certain models can reconstruct sensitive information from their outputs. Contribution: The study highlights the implications of model overfitting on privacy, showing that even well-regulated models can leak sensitive information. The study emphasizes the need for robust privacy-preserving techniques in ML to mitigate these risks. | |
Fang et al. [135] | DP: Provides a formal guarantee that the output of the analysis will not change significantly if an individual’s data are altered. | Methodology: Utilizes a functional mechanism that adds calibrated noise to the regression outputs, balancing privacy protection with data utility. Contribution: Introduces a functional mechanism for regression analysis under DP. Evaluates the performance of the model in terms of noise reduction and resilience to model inversion attacks. | |
Cummings et al. [199] | DP: Ensures that the output of the regression analysis does not change significantly when the data of a single individual are changed. | Methodology: Introduces individual sensitivity preprocessing techniques for enhancing data privacy. Contribution: Proposes preprocessing methods that adjust data sensitivity on an individual level, improving privacy protection while allowing for meaningful data analysis. Introduces an individual sensitivity metric technique to improve the accuracy of private data. | |
Gradient and weight analyses | Zhu et al. [200] | Model inversion attack analysis: Utilizes gradients to reconstruct inputs from model outputs. | Methodology: Explores model inversion attacks enhanced by adversarial examples in ML models. Contribution: Demonstrates how adversarial examples can significantly boost the effectiveness of model inversion attacks, providing insights into potential vulnerabilities in machine-learning systems. |
Zhu et al. [201] | Gradient leakage analysis: Exchanges gradients that lead to the leakage of private training data. | Methodology: Investigates deep leakage from gradients in machine-learning models. Contribution: Analyzes how gradients can leak sensitive information about training data, contributing to the understanding of privacy risks associated with model training. | |
Huang et al. [202] | Gradient inversion attack analysis: Evaluates gradient inversion attacks in federated learning. | Methodology: Explores model inversion attacks enhanced by adversarial examples in ML models. Contribution: Assesses the effectiveness of gradient inversion attacks in federated learning settings and proposes defenses to mitigate these vulnerabilities. | |
Wu et al. [203] | Gradient inversion attack analysis: Introduces a new gradient inversion method, Learning to Invert (LIT). | Methodology: Develops adaptive attacks for gradient inversion in federated learning environments. Contribution: Introduces simple adaptive attack strategies to enhance the success rate of gradient inversion attacks (gradient compression), highlighting the risks in federated learning scenarios. | |
Zhu et al. [204] | Gradient inversion attack analysis: Proposes a generative gradient inversion attack (GGI) in federated learning contexts. | Methodology: Utilizes generative models to perform gradient inversion without requiring prior knowledge of the data distribution. Contribution: Presents a novel attack that utilizes generative models to enhance gradient inversion attacks, demonstrating new avenues for information leakage in collaborative settings. | |
Empirical privacy loss | Yang et al. [205] | DP: Proposes a method to enhance privacy by purifying predictions. | Methodology: Proposes a defense mechanism against model inversion and membership inference attacks through prediction purification. Contribution: Demonstrates that a purifier dedicated to one type of attack can effectively defend against the other, establishing a connection between model inversion and membership inference vulnerabilities, employing a prediction purification technique. |
Zhang et al. [206] | DP: Incorporates additional noise mechanisms specifically designed to counter model inversion attacks. | Methodology: Broadens differential privacy frameworks to enhance protection against model inversion attacks in deep learning. Contribution: Introduces new techniques to strengthen differential privacy guarantees specifically against model inversion, improving the robustness of deep-learning models against such attacks, and propose class and subclass DP within context of random forest algorithms. | |
Reconstruction test | Manchini et al. [207] | DP: Use differential privacy in regression models that accounts for heteroscedasticity. | Methodology: Proposes a new approach to data differential privacy using regression models under heteroscedasticity. Contribution: Develops methods to enhance differential privacy in regression analysis, particularly for datasets with varying levels of noise, improving privacy guarantees for ML applications. |
Park et al. [139] | DP: Evaluates the effectiveness of differentially private learning models against model inversion attacks. | Methodology: Evaluates differentially private learning against model inversion attacks through an attack-based evaluation method. Contribution: Introduces an evaluation framework that assesses the robustness of differentially private models against model inversion attacks, providing insights into the effectiveness of privacy-preserving techniques. | |
Model extraction auditing | |||
Query analysis | Carlini et al. [101] | Model extraction attack analysis: Demonstrates that large language models, such as GPT-2, are vulnerable to training data-extraction attacks. | Methodology: Employs a two-stage approach for training data extraction, suffix generation and suffix ranking. Contribution: Shows that attackers can recover individual training examples from large language models by querying them, highlighting vulnerabilities in model training processes and discussing potential safeguards. |
Dziedzic et al. [209] | Model extraction attack analysis: Addresses model extraction attacks, where attackers can steal ML models by querying them. | Methodology: Proposes a calibrated proof of work mechanism to increase the cost of model extraction attacks. Contribution: Introduces a novel approach, BDPL (Boundary Differential Private Layer), that raises the resource requirements for adversaries attempting to extract models, thereby enhancing the security of machine-learning systems against such attacks. | |
Li et al. [210] | Local DP: Introduces a personalized local differential privacy (PLDP) mechanism designed to protect regression models from model extraction attacks. | Methodology: Uses a novel perturbation mechanism that adds high-dimensional Gaussian noise to the model outputs based on personalized privacy parameters. Contribution: Personalized local differential privacy (PLDP) ensures that individual user data are perturbed before being sent to the model, thereby protecting sensitive information from being extracted through queries. | |
Li et al. [146] | Model extraction attack analysis: Proposes a framework designed to protect object detection models from model extraction attacks by focusing on feature space coverage. | Methodology: Uses a novel detection framework that identifies suspicious users based on their query traffic and feature coverage. Contribution: Develops a detection framework that identifies suspicious users based on feature coverage in query traffic, employing an active verification module to confirm potential attackers, thereby enhancing the security of object detection models and distinguishing between malicious and benign queries. | |
Zheng et al. [211] | Boundary Differential Privacy (-BDP): Introduces Boundary Differential Privacy (ϵ-BDP), which protects against model extraction attacks by obfuscating prediction responses near the decision boundary. | Methodology: Uses a perturbation algorithm called boundary randomized response, which achieves ϵ-BDP by adding noise to the model’s outputs based on their proximity to the decision boundary. Contribution: Introduces a novel layer that obfuscates prediction responses near the decision boundary to prevent adversaries from inferring model parameters, demonstrating effectiveness through extensive experiments. | |
Yan et al. [212] | DP: Proposes a monitoring-based differential privacy (MDP) mechanism that enhances the security of machine-learning models against query flooding attacks. | Methodology: Introduces a novel real-time model extraction status assessment scheme called “Monitor”, which evaluates the model’s exposure to potential extraction based on incoming queries. Contribution: Proposes a mechanism that monitors query patterns to detect and mitigate model extraction attempts, enhancing the resilience of machine-learning models against flooding attacks. | |
Property inference auditing | |||
Evaluating property sensitivity with model outputs. | Suri et al. [213] | Distribution inference attack analysis: Investigates distribution inference attacks, which aim to infer statistical properties of the training data used by ML models. | Methodology: Introduces a distribution inference attack that infers statistical properties of training data using a KL divergence approach. Contribution: Develops a novel black-box attack that outperforms existing white-box methods, evaluating the effectiveness of various defenses against distribution inference risks; performs disclosure at three granularities, namely distribution, user, and record levels; and proposes metrics to quantify observed leakage from models under attack. |
Property inference framework | Ganju et al. [214] | Property inference attack analysis: Explores property inference attacks on fully connected neural networks (FCNNs), demonstrating that attackers can infer global properties of the training data. | Methodology: Leverages permutation invariant representations to reduce the complexity of inferring properties from FCNNs. Contribution: Analyzes how permutation invariant representations can be exploited to infer sensitive properties of training data, highlighting vulnerabilities in neural network architectures. |
Melis et al. [215] | Feature leakage analysis: Reveals that collaborative learning frameworks inadvertently leak sensitive information about participants’ training data through model updates. | Methodology: Uses both passive and active inference attacks to exploit unintended feature leakage. Contribution: Examines how collaborative learning frameworks can leak sensitive features, providing insights into the risks associated with sharing models across different parties. | |
Empirical evaluation of linear queries | Huang and Zhou [216] | DP: Discusses how DP mechanisms can inadvertently leak sensitive information when linear queries are involved. | Methodology: Studies unexpected information leakage in differential privacy due to linear properties of queries. Contribution: Analyzes how certain (linear) query structures can lead to information leakage despite differential privacy guarantees, suggesting improvements for privacy-preserving mechanisms. |
Analysis of DP implementation | Ben Hamida et al. [217] | DP: Discusses how differential privacy (DP) enhances the privacy of machine-learning models by ensuring that individual data contributions do not significantly affect the model’s output. | Methodology: Explore various techniques for implementing DPML, including adding noise to gradients during training and employing mechanisms that ensure statistical outputs mask individual contributions. Contribution: Explores the interplay between differential privacy techniques and their effectiveness in enhancing model security against various types of attacks. |
Song et al. [218] | Privacy risk evaluation: | Methodology: Conducts a systematic evaluation of privacy risks in machine-learning models across different scenarios. Contribution: Provides a comprehensive framework for assessing the privacy risks associated with machine-learning models, identifying key vulnerabilities and suggesting mitigation strategies. |
References
- Choudhury, O.; Gkoulalas-Divanis, A.; Salonidis, T.; Sylla, I.; Park, Y.; Hsu, G.; Das, A. Differential Privacy-Enabled Federated Learning For Sensitive Health Data. arXiv 2019, arXiv:1910.02578. Available online: https://arxiv.org/abs/1910.02578 (accessed on 1 December 2024).
- Dwork, C.; McSherry, F.; Nissim, K.; Smith, A. Calibrating Noise To Sensitivity In Private Data Analysis. In Theory of Cryptography; Halevi, S., Rabin, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 265–284. [Google Scholar]
- Williamson, S.M.; Prybutok, V. Balancing Privacy and Progress: A Review of Privacy Challenges, Systemic Oversight, and Patient Perceptions. In: AI-Driven Healthcare. Appl. Sci. 2024, 14, 675. [Google Scholar] [CrossRef]
- Barbierato, E.; Gatti, A. The Challenges of Machine Learning: A Critical Review. Electronics 2024, 13, 416. [Google Scholar] [CrossRef]
- Noor, M.H.M.; Ige, A.O. A Survey on State-of-the-art Deep Learning Applications and Challenges. arXiv 2024, arXiv:2403.17561. Available online: https://arxiv.org/abs/2403.17561 (accessed on 1 December 2024).
- Du Pin Calmon, F.; Fawaz, N. Privacy Against Statistical Inference. In Proceedings of the 2012 50th Annual Allerton Conference on Communication, Control, and Computing, Allerton, Monticello, IL, USA, 1–5 October 2012; pp. 1401–1408. [Google Scholar]
- Dehghani, M.; Azarbonyad, H.; Kamps, J.; de Rijke, M. Share your Model of your Data: Privacy Preserving Mimic Learning for Mimic Learning for Ranking. arXiv 2017, arXiv:1707.07605. Available online: https://arxiv.org/abs/1707.07605 (accessed on 1 December 2024).
- Bouke, M.; Abdullah, A. An Empirical Study Of Pattern Leakage Impact During Data Preprocessing on Machine Learning-Based Intrusion Detection Models Reliability. Expert Syst. Appl. 2023, 230, 120715. [Google Scholar] [CrossRef]
- Xu, J.; Wu, Z.; Wang, C.; Jia, X. Machine Unlearning: Solutions and Challenges. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 2150–2168. [Google Scholar] [CrossRef]
- Yeom, S.; Giacomelli, I.; Fredrikson, M.; Jha, S. Privacy risk in machine learning: Analyzing the connection to overfitting. In Proceedings of the 2018 IEEE 31st Computer Security Foundations Symposium (CSF), Oxford, UK, 9–12 July 2018; pp. 268–282. [Google Scholar]
- Li, Y.; Yan, H.; Huang, T.; Pan, Z.; Lai, J.; Zhang, X.; Chen, K.; Li, J. Model Architecture Level Privacy Leakage In Neural Networks. Sci. China Inf. Sci. 2024, 67, 3. [Google Scholar] [CrossRef]
- Del Grosso, G.; Pichler, G.; Palamidessi, C.; Piantanida, P. Bounding information leakage in machine learning. Neurocomputing 2023, 534, 1–17. [Google Scholar] [CrossRef]
- McSherry, F.; Talwar, K. Mechanism Design via Differential Privacy. In Proceedings of the 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2007), Providence, RI, USA, 20–23 October 2007; pp. 94–103. [Google Scholar]
- Mulder, V.; Humbert, M. Differential privacy. In Trends in Data Protection and Encryption Technologies; Springer: Berlin/Heidelberg, Germany, 2023; pp. 157–161. [Google Scholar]
- Gong, M.; Xie, Y.; Pan, K.; Feng, K.; Qin, A. A Survey on Differential Private Machine Learning. IEEE Comput. Intell. Mag. 2020, 15, 49–64. [Google Scholar] [CrossRef]
- Liu, B.; Ding, M.; Shaham, S.; Rahayu, W.; Farokhi, F.; Lin, Z. When Machine Learning Meets Privacy: A Survey and Outlook. ACM Comput. Surv. 2021, 54, 1–31. [Google Scholar] [CrossRef]
- Blanco-Justicia, A.; Sanchez, A.; Domingo-Ferrer, J.; Muralidhar, K.A. Critical Review on the Use (and Misuse) of Differential Privacy in Machine Learning. ACM Comput. Surv. 2023, 55, 1–16. [Google Scholar] [CrossRef]
- Zheng, H.; Ye, Q.; Hu, H.; Fang, C.; Shi, J. Protecting Decision Boundary of Machine Learning Model With Differential Private Perturbation. IEEE Trans. Dependable Secur. Comput. 2022, 19, 2007–2022. [Google Scholar] [CrossRef]
- Ponomareva, N.; Hazimeh, H.; Kurakin, A.; Xu, Z.; Denison, C.; McMahan, H.B.; Vassilvitskii, S.; Chien, S.; Thakurta, A.G. A Practical Guide to Machine Learning with Differential Privacy. J. Artif. Intell. Res. 2023, 77, 1113–1201. [Google Scholar] [CrossRef]
- Choquette-Choo, C.A.; Dullerud, N.; Dziedzic, A.; Zhang, Y.; Jha, S.; Papernot, N.; Wang, X. CaPC Learning: Confidential and Private Collaborative Learning. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 4 May 2021. [Google Scholar]
- Jayaraman, B.; Evans, D. Evaluating Differentially Private Machine Learning in Practice. In Proceedings of the 28th USENIX Security Symposium (USENIX Security 19), Santa Clara, CA, USA, 14–16 August 2019; pp. 1895–1912. [Google Scholar] [CrossRef]
- Sablayrolles, A.; Douze, M.; Schmid, C.; Ollivier, Y.; Jégou, H. White-box vs black-box: Bayes optimal strategies for membership inference. In Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 5558–5567. [Google Scholar]
- Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep Learning with Differential Privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (CCS), Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar] [CrossRef]
- Bagdasaryan, E. Differential Privacy Has Disparate Impact on Model Accuracy. Adv. Neural Inf. Process. Syst. 2019, 32, 161263. [Google Scholar] [CrossRef]
- Tran, C.; Dinh, M.H. Differential Private Empirical Risk Minimization under the Fairness Lens. Adv. Neural Inf. Process. Syst. 2021, 33, 27555–27565. [Google Scholar] [CrossRef]
- Bichsel, B.; Stefen, S.; Bogunovic, I.; Vechev, M. Dp-Sniper: Black-Box Discovery Of Differential Privacy Violations Using Classifiers. In Proceedings of the 2021 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; pp. 391–409. [Google Scholar] [CrossRef]
- Dwork, C. Differential Privacy. In Automata, Languages and Programming; Bugliesi, M., Preneel, B., Sassone, V., Wegener, I., Eds.; Lecture Notes in Computer Science; Springer: Berlin, Germany, 2006; pp. 1–12. [Google Scholar] [CrossRef]
- He, J.; Cai, L.; Guan, X. Differential Private Noise Adding Mechanism and Its Application on Consensus Algorithm. IEEE Trans. Signal Process. 2020, 68, 4069–4082. [Google Scholar] [CrossRef]
- Wang, R.; Fung, B.C.M.; Zhu, Y.; Peng, Q. Differentially Private Data Publishing for Arbitrary Partitioned Data. Inf. Sci. 2021, 553, 247–265. [Google Scholar] [CrossRef]
- Baraheem, S.S.; Yao, Z. A Survey on Differential Privacy with Machine Learning and Future Outlook. arXiv 2022, arXiv:2211.10708. Available online: https://arxiv.org/abs/2211.10708 (accessed on 1 December 2024).
- Dwork, C.; Roth, A. The Algorithmic Foundations Of Differential Privacy. Found. Trends Theor. Comput. Sci. 2014, 9, 211–407. Available online: https://www.nowpublishers.com/article/Details/TCS-042 (accessed on 1 December 2024). [CrossRef]
- Chadha, K.; Jagielski, M.; Papernot, N.; Choquette-Choo, C.A.; Nasr, M. Auditing Private Prediction. arXiv 2024, arXiv:2402.0940. [Google Scholar] [CrossRef]
- Papernot, N.; Abadi, M.; Erlingsson, Ú.; Goodfelow, I.; Talwar, K. Semi-Supervise Knowledge Transfer for Deep Learning from Private Training Data. International Conference on Learning Representations. 2016. Available online: https://openreview.net/forum?id=HkwoSDPgg (accessed on 1 December 2024).
- Bernau, D.; Robl, J.; Grassal, P.W.; Schneider, S.; Kerschbaum, F. Comparing Local and Central Differential Privacy Using Membership Inference Attacks. In IFIP Annual Conference on Data and Applications Security and Privacy; Springer: Berlin/Heidelberg, Germany, 2021; pp. 22–42. [Google Scholar] [CrossRef]
- Hsu, J.; Gaboardi, M.; Haeberlen, A.; Khanna, S.; Narayan, A.; Pierce, B.C.; Roth, A. Differential Privacy: An economic methods for choosing epsilon. In Proceedings of the Computer Security Foundations Workshop, Vienna, Austria, 19–22 July 2014; pp. 398–410. [Google Scholar] [CrossRef]
- Mehner, L.; Voigt, S.N.V.; Tschorsch, F. Towards Explaining Epsilon: A Worst-Case Study of Differential Privacy Risks. In Proceedings of the 2021 IEEE European Symposium on Security and Privacy Workshop, Euro S and PW, Virtual, 6–10 September 2021; pp. 328–331. [Google Scholar]
- Busa-Fekete, R.I.; Dick, T.; Gentile, C.; Medina, A.M.; Smith, A.; Swanberg, M. Auditing Privacy Mechanisms via Label Inference Attacks. arXiv 2024, arXiv:2406.01797. Available online: https://arxiv.org/abs/2406.02797 (accessed on 1 December 2024).
- Desfontaines, D.; Pejó, B. SoK: Differential Privacies. arXiv 2022, arXiv:1906.01337. Available online: https://arxiv.org/abs/1906.01337 (accessed on 1 December 2024). [CrossRef]
- Lycklama, H.; Viand, A.; Küchler, N.; Knabenhans, C.; Hithnawi, A. Holding Secrets Accountable: Auditing Privacy-Preserving Machine Learning. arXiv 2024, arXiv:2402.15780. Available online: https://arxiv.org/abs/2402.15780 (accessed on 1 December 2024).
- Kong, W.; Medina, A.M.; Ribero, M.; Syed, U. DP-Auditorium: A Large Scale Library for Auditing Differential Privacy. arXiv 2023, arXiv:2307.05608. Available online: https://arxiv.org/abs/2307.05608 (accessed on 1 December 2024).
- Ha, T.; Vo, T.; Dang, T.K. Differential Privacy Under Membership Inference Attacks. Commun. Comput. Inf. Sci. 2023, 1925, 255–269. [Google Scholar]
- Nasr, M.; Hayes, J.; Steinke, T.; Balle, B.; Tramer, F.; Jagielski, M.; Carlini, N.; Terzis, A. Tight Auditing of Differentially Private Machine Learning. In Proceedings of the 32nd USENIX Security Symposium (USENIX Security 23), San Francisco, CA, USA, 9–11 August 2023; pp. 1631–1648. [Google Scholar]
- Steinke, T.; Nasr, M.; Jagielski, M. Privacy Auditing with One (1) Training Run. arXiv 2023, arXiv:2305.08846. Available online: https://arxiv.org/abs/2305.08846 (accessed on 1 December 2024).
- Wairimu, S.; Iwaya, L.H.; Fritsch, L.; Lindskog, S. Assessment and Privacy Risk Assessment Methodologies: A Systematic Literature Review. IEEE Access 2024, 12, 19625–19650. [Google Scholar] [CrossRef]
- Ding, Z.; Wang, Y.; Wang, G.; Zhang, D.; Kifer, D. Detecting Violations Of Differential Privacy. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, Toronto, ON, Canada, 15–19 October 2018; pp. 475–489. [Google Scholar] [CrossRef]
- Jagielski, M.; Ullman, J.; Oprea, A. Auditing Differentially Private Machine Learning: How Private is Private sgd? Adv. Neural Inf. Process. Syst. 2020, 33, 22205–22216. [Google Scholar] [CrossRef]
- Nasr, M.; Songi, S.; Thakurta, A.; Papernot, N.; Carlin, N. Adversary instantiation: Lower bounds for differentially private machine learning. In Proceedings of the 2021 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; pp. 866–882. [Google Scholar]
- Lokna, J.; Paradis, A.; Dimitrov, D.I.; Vechev, M. Group and Attack: Auditing Differential Privacy. In Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security (CCS ’23), Copenhagen, Denmark, 26–30 November 2023; ACM: New York, NY, USA, 2023; pp. 1–22. Available online: https://dl.acm.org/doi/10.1145/3576915.3616607 (accessed on 1 December 2024).
- Tramèr, F.; Terzis, A.; Steinke, T.; Song, S.; Jagielski, M.; Carlini, N. Debugging differential privacy: A case study for privacy auditing. arXiv 2022, arXiv:2202.12219. Available online: https://arxiv.org/abs/2202.12219 (accessed on 1 December 2024).
- Kifer, D.; Messing, S.; Roth, A.; Thakurta, A.; Zhang, D. Guidelines for Implementing and Auditing Differentially Private Systems. arXiv 2020, arXiv:2002.04049. Available online: https://arxiv.org/abs/2002.04049 (accessed on 1 December 2024).
- Homer, N.; Szelinger, S.; Redman, M.; Duggan, D.; Tembe, W.; Muehling, J.; Pearson, J.V.; Stephan, D.A.; Nelson, S.F.; Craig, D.W. Resolving individuals contributing trace amounts of DNA to highly complex mixtures using high-density SNP genotyping microarrays. PLoS Genet. 2008, 4, e1000167. [Google Scholar] [CrossRef] [PubMed]
- Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership Inference Attacks against Machine Learning Models. In Proceedings of the 2017 IEEE Symposium on Security and Priavcy (S&P), San Jose, CA, USA, 22–26 May 2017; pp. 3–18. [Google Scholar]
- Cui, G.; Ge, L.; Zhao, Y.; Fang, T. A Membership Inference Attack Defense Methods Based on Differential Privacy and Data Enhancement. In Proceedings of the Communication in Computer and Information Science, Manchester, UK, 9–11 September 2024; Volume 2015 CCIS, pp. 258–270. [Google Scholar]
- Ma, Y.; Zhu, X.; Hsu, J. Data Poisoning against Differentially-Private Learners: Attacks and Defences. arXiv 2019, arXiv:1903.09860. Available online: https://arxiv.org/abs/1903.09860 (accessed on 1 December 2024).
- Cinà, A.E.; Grosse, K.; Demondis, A.; Biggo, B.; Roli, F.; Pelillo, M. Machine Learning Security Against Data Poisoning: Are We There Yet? Computer 2024, 7, 26–34. [Google Scholar] [CrossRef]
- Cheng, Z.; Li, Z.; Zhang, L.; Zhang, S. Differentially Private Machine Learning Model against Model Extraction Attack. In Proceedings of the IEEE 2020 International Conferences on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData) and IEEE Congress on Cybermatics (Cybermatics), Rhodes, Greece, 2–6 November 2020; pp. 722–728. Available online: https://ieeexplore.ieee.org/document/9291542 (accessed on 3 December 2024).
- Miura, T.; Hasegawa, S.; Shibahara, T. MEGEX: Data-free model extraction attack against gradient-based explainable AI. arXiv 2021, arXiv:2107.08909. Available online: https://arxiv.org/abs/2107.08909 (accessed on 1 December 2024).
- Ye, Z.; Luo, W.; Naseem, M.L.; Yang, X.; Shi, Y.; Jia, Y. C2FMI: Corse-to-Fine Black-Box Model Inversion Attack. IEEE Trans. Dependable Secur. Comput. 2024, 21, 1437–1450. Available online: https://ieeexplore.ieee.org/document/10148574 (accessed on 3 December 2024). [CrossRef]
- Qiu, Y.; Yu, H.; Fang, H.; Yu, W.; Chen, B.; Wang, X.; Xia, S.-T.; Xu, K. MIBench: A Comprehensive Benchmark for Model Inversion Attack and Defense. arXiv 2024, arXiv:2410.05159. Available online: https://arxiv.org/abs/2410.05159 (accessed on 1 December 2024).
- Stock, J.; Lange, L.; Erhard, R.; Federrath, H. Property Inference as a Regression Problem: Attacks and Defense. In Proceedings of the International Conference on Security and Cryptography, Bengaluru, India, 18–19 April 2024; pp. 876–885. Available online: https://www.scitepress.org/publishedPapers/2024/128638/pdf/index.html (accessed on 3 December 2024).
- Lu, F.; Munoz, J.; Fuchs, M.; LeBlond, T.; Zaresky-Williams, E.; Raff, E.; Ferraro, F.; Testa, B. A General Framework for Auditing Differentially Private Machine Learning. In Advances in Neural Information Processing Systems; Oh, A.H., Belgrave, A., Cho, K., Eds.; The MIT Press: Cambridge, MA, USA, 2022; Available online: https://openreview.net/forum?id=AKM3C3tsSx3 (accessed on 1 December 2024).
- Zanella-Béguelin, S.; Wutschitz, L.; Tople, S.; Salem, A.; Rühle, V.; Paverd, A.; Naseri, M.; Köpf, B.; Jones, D. Bayesian Estimation Of Differential Privacy. In Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; Volume 202, pp. 40624–40636. [Google Scholar]
- Cowan, E.; Shoemate, M.; Pereira, M. Hands-On Differential Privacy; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2024; ISBN 9781492097747. [Google Scholar]
- Bailie, J.; Gong, R. Differential Privacy: General Inferential Limits via Intervals of Measures. Proc. Mach. Learn. Res. 2023, 215, 11–24. Available online: https://proceedings.mlr.press/v215/bailie23a/bailie23a.pdf (accessed on 3 December 2024).
- Kilpala, M.; Kärkäinen, T. Artificial Intelligence and Differential Privacy: Review of Protection Estimate Models. In Artificial Intelligence for Security: Enhancing Protection in a Changing World; Springer Nature: Cham, Switherland, 2024; pp. 35–54. [Google Scholar]
- Balle, B.; Wang, Y.-X. Improving the Gaussian Mechanism for Differential Privacy: Analytical Calibration and Optimal Denoising. In Proceedings of the 35th International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018; pp. 394–403. Available online: http://proceedings.mlr.press/v80/balle18a/balle18a.pdf (accessed on 3 December 2024).
- Chen, B.; Hale, M. The Bounded Gaussian Mechanism for Differential Privacy. J. Priv. Confidentiality 2024, 14, 1. [Google Scholar] [CrossRef]
- Zhang, K.; Zhang, Y.; Sun, R.; Tsai, P.-W.; Ul Hassan, M.; Yuan, X.; Xue, M.; Chen, J. Bounded and Unbiased Composite Differential Privacy. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 19–23 May 2024; pp. 972–990. [Google Scholar]
- Nanayakkara, P.; Smart, M.A.; Cummings, R.; Kaptchuk, G. What Are the Chances? Explaining the Epsilon Parameter in Differential Privacy. In Proceedings of the 32nd USINEX Security Symposium, Anaheim, CA, USA, 9–11 August 2023; Volume 3, pp. 1613–1630. [Google Scholar] [CrossRef]
- Cannone, C.; Kamath, G.; McMillan, A.; Smith, A.; Ullman, J. The Structure Of Optimal Private Tests For Simple Hypotheses. In Proceedings of the Annual ACM Symposium on Theory of Computing, Phoenix, AZ, USA, 23–26 June 2019; pp. 310–321. Available online: https://arxiv.org/abs/1811.11148 (accessed on 3 December 2024).
- Dwork, C.; Feldman, V. Privacy-preserving Prediction. arXiv 2018, arXiv:1803.10266. Available online: https://arxiv.org/abs/1803.10266 (accessed on 1 December 2024).
- Mironov, I. Rényi Differential Privacy. In Proceedings of the 30th IEEE Computer Security Foundations Symposium, CSF, Santa Barbara, CA, USA, 21–25 August 2017; pp. 263–275. [Google Scholar] [CrossRef]
- Sarathy, R.; Muralidhar, K. Evaluating Laplace noise addition to satisfy differential privacy for numeric data. Trans. Data Priv. 2011, 4, 1–17. [Google Scholar] [CrossRef]
- Kumar, G.S.; Premalatha, K.; Uma Maheshwari, G.; Rajesh Kanna, P.; Vijaya, G.; Nivaashini, M. Differential privacy scheme using Laplace mechanism and statistical method computation in deep neural network for privacy preservation. Eng. Appl. Artif. Intell. 2024, 128, 107399. [Google Scholar] [CrossRef]
- Liu, F. Generalized Gaussian Mechanism for Differential Privacy. IEEE Trans. Knowl. Data Eng. 2018, 31, 747–756. [Google Scholar] [CrossRef]
- Dong, J.; Roth, A.; Su, W.J. Gaussian Differential privacy. arXiv 2019, arXiv:1905.02383. Available online: https://arxiv.org/abs/1905.02383 (accessed on 1 December 2024). [CrossRef]
- Geng, Q.; Ding, W.; Guo, R.; Kumar, S. Tight Analysis of Privacy and Utility Tradeoff in Approximate Differential Privacy. Proc. Mach. Lerning Res. 2020, 108, 89–99. Available online: http://proceedings.mlr.press/v108/geng20a/geng20a.pdf (accessed on 3 December 2024).
- Whitehouse, J.; Ramdas, A.; Rogers, R.; Wu, Z.S. Fully-Adaptive Composition in Differential Privacy. arXiv 2023, arXiv:2203.05481. Available online: https://arxiv.org/abs/2203.05481 (accessed on 1 December 2024).
- Dwork, C.; Kenthapadi, K.; McSherry, F.; Mironov, I.; Naor, M. Our Data, Ourselves: Privacy Via Distributed Noise Generation. In Advances in Cryptology—EUROCRYPT; Vaudenay, S., Ed.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 486–503. [Google Scholar]
- Zhu, K.; Fioretto, F.; Van Hentenryck, P. Post-processing of Differentially Private Data: A Fairness Perspective. In Proceedings of the 31st International Joint Conference on Artificial Intelligence (IJCAI), Vienna, Austria, 23–29 July 2022; pp. 4029–4035. [Google Scholar] [CrossRef]
- Ganev, G.; Annamalai, M.S.M.S.; De Cristofaro, E. The Elusive Pursuit of Replicating PATE-GAN: Benchmarking, Auditing, Debugging. arXiv 2024, arXiv:2406.13985. Available online: https://arxiv.org/abs/2406.13985 (accessed on 1 December 2024).
- Naseri, M.; Hayes, J.; De Cristofaro, E. Local and Central Differential Privacy for Robustness and Privacy in Federated Learning. arXiv 2022, arXiv:2009.03561. Available online: https://arxiv.org/abs/2009.03561 (accessed on 1 December 2024).
- Babesne, B. Local Differential Privacy: A tutorial. arXiv 2019, arXiv:1907.11908. Available online: https://arxiv.org/abs/1907.11908 (accessed on 1 December 2024).
- Nasr, M.; Shokri, R.; Houmandsadr, A. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning. arXiv 2020, arXiv:1812.00910. Available online: https://arxiv.org/abs/1812.00910 (accessed on 1 December 2024).
- Galli, F.; Biswas, S.; Jung, K.; Cucinotta, T.; Palamidessi, C. Group privacy for personalized federated learning. arXiv 2022, arXiv:2206.03396. Available online: https://arxiv.org/abs/2206.03396 (accessed on 1 December 2024).
- Cormode, G.; Jha, S.; Kulkarni, T.; Li, N.; Srivastava, D.; Wang, T. Privacy At Scale: Local Differential Privacy in Practice. In Proceedings of the ACM SIGMOD International Conference on Management of Data, Houston, TX, USA, 10–15 June 2018; pp. 1655–1658. [Google Scholar] [CrossRef]
- Yang, M.; Guo, T.; Zhu, T.; Tjuawinata, I.; Zhao, J.; Lam, K.-Y. Local Differential Privacy And Its Applications: A Comprehensive Survey. Comput. Stand. Interfaces 2024, 89, 103827. [Google Scholar] [CrossRef]
- Duchi, J.; Wainwright, M.J.; Jordan, M.I. Local Privacy And Minimax Bounds: Sharp Rates For Probability Estimation. Adv. Neural Inf. Process. Syst. 2013, 26, 1529–1537. [Google Scholar] [CrossRef]
- Ruan, W.; Xu, M.; Fang, W.; Wang, L.; Wang, L.; Han, W. Private, Efficient, and Accurate: Protecting Models Trained by Multi-party Learning with Differential Privacy. In Proceedings of the—IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 21–25 May 2023; pp. 1926–1943. [Google Scholar]
- Pan, K.; Ong, Y.-S.; Gong, M.; Li, H.; Qin, A.K.; Gao, Y. Differential privacy in deep learning: A literature review. Neurocomputing 2024, 589, 127663. [Google Scholar] [CrossRef]
- Kang, Y.; Liu, Y.; Niu, B.; Tong, X.; Zhang, L.; Wang, W. Input Perturbation: A New Paradigm between Central and Local Differential Privacy. arXiv 2020, arXiv:2002.08570. Available online: https://arxiv.org/abs/2002.08570 (accessed on 1 December 2024).
- Chaudhuri, K.; Monteleoni, C.; Sarwate, A.D. Differentially Private Empirical Risk Minimization. J. Mach. Learn. Res. 2011, 12, 1069–1109. [Google Scholar] [CrossRef]
- De Cristofaro, E. Critical Overview of Privacy in Machine Learning. IEEE Secur. Priv. 2021, 19, 19–27. [Google Scholar] [CrossRef]
- Shen, Z.; Zhong, T. Analysis of Application Examples of Differential Privacy in Deep Learning. Comput. Intell. Neurosci. 2021, 2021, e4244040. [Google Scholar] [CrossRef]
- Rigaki, M.; Garcia, S. A Survey of Privacy Attacks in Machine Learning. ACM Comput.Surv. 2023, 56, 101. [Google Scholar] [CrossRef]
- Wu, D.; Qi, S.; Li, Q.; Cai, B.; Guo, Q.; Cheng, J. Understanding and Defending against White-Box Membership Inference Attack in Deep Learning. Knowl. Based Syst. 2023, 259, 110014. [Google Scholar] [CrossRef]
- Fang, H.; Qiu, Y.; Yu, H.; Yu, W.; Kong, J.; Chong, B.; Chen, B.; Wang, X.; Xia, S.-T. Privacy Leakage on DNNs: A Survey of Model Inversion Attacks and Defenses. arXiv 2024, arXiv:2402.04013. Available online: https://arxiv.org/abs/2402.04013 (accessed on 1 December 2024).
- He, X.-M.; Wang, X.S.; Chen, H.-H.; Dong, Y.-H. Study on Choosing the Parameter ε in Differential Privacy. Tongxin Xuebo/J. Commun. 2015, 36, 12. [Google Scholar]
- Mazzone, F.; Al Badawi, A.; Polyakov, Y.; Everts, M.; Hahn, F.; Peter, A. Investigating Privacy Attacks in the Gray-Box Settings to Enhance Collaborative Learning Schemes. arXiv 2024, arXiv:2409.17283. Available online: https://arxiv.org/abs/2409.17283 (accessed on 1 December 2024).
- Fredrikson, M.; Jha, S.; Ristenpart, T. Model inversion attacks that exploit confidence information and basic countermeasures. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, ACM, Denver, CO, USA, 12–16 October 2015; pp. 1322–1333. [Google Scholar]
- Carlini, N.; Tramèr, F.; Wallace, E.; Jagielski, M.; Herbert-Voss, A.; Lee, K.; Roberts, A.; Brown, T.; Song, D.; Erlingsson, U.; et al. Extracting training data from large language models. arXiv 2020, arXiv:2012.07805. Available online: https://arxiv.org/abs/2012.07805 (accessed on 1 December 2024).
- Carlini, N.; Liu, C.; Erlingsson, Ś.; Kos, J.; Song, D. The secret sharer: Evaluating and testing unintended memorization in neural networks. In Proceedings of the 28th USENIX Security Symposium (USENIX Security 19), Santa Clara, CA, USA, 14–16 August 2019; pp. 267–284. [Google Scholar]
- Carlini, N.; Chien, S.; Nasr, M.; Song, S.; Terzis, A.; Tramèr, F. Membership Inference Attacks from First Principles. arXiv 2021, arXiv:2112.03570. Available online: https://arxiv.org/abs/2112.03570 (accessed on 1 December 2024).
- Hu, H.; Salcic, Z.; Sun, L.; Dobbie, G.; Yu, P.S.; Zhang, X. Membership Inference Attacks on Machine Learning: A Survey. arXiv 2022, arXiv:2103.07853. Available online: https://arxiv.org/abs/2103.07853 (accessed on 1 December 2024). [CrossRef]
- Zarifzadeh, S.; Liu, P.; Shokri, R. Low-Cost High-Power Membership Inference Attacks. arXiv 2023, arXiv:2312.03262. Available online: https://arxiv.org/abs/2312.03262 (accessed on 1 December 2024).
- Aubinais, E.; Gassiat, E.; Piantanida, P. Fundamental Limits of Membership Inference attacks on Machine Learning Models. arXiv 2024, arXiv:2310.13786. Available online: https://arxiv.org/html/2310.13786v4 (accessed on 1 December 2024).
- Leino, K.; Fredrikson, M. Stolen memories: Leveraging model memorization for calibrated white box membership inference. In Proceedings of the 29th {USENIX} Security Symposium {USENIX} Security 20, Online, 12–14 August 2020; pp. 1605–1622. Available online: https://www.usenix.org/conference/usenixsecurity20/presentation/leino (accessed on 2 December 2024).
- Liu, R.; Wang, D.; Ren, Y.; Wang, Z.; Guo, K.; Qin, Q.; Liu, X. Unstoppable Attack: Label-Only Model Inversion via Conditional Diffusion Model. IEEE Trans. Inf. Forensics Secur. 2024, 19, 3958–3973. [Google Scholar] [CrossRef]
- Annamalai, M.S.M.S. Nearly Tight Black-Box Auditing of Differential Private Machine Learning. arXiv 2024, arXiv:2405.14106. Available online: https://arxiv.org/abs/2405.14106 (accessed on 1 December 2024).
- Lin, S.; Bun, M.; Gaboardi, M.; Kolaczyk, E.D.; Smith, A. Differential Private Confidence Intervals for Proportions Under Stratified Random Sampling. Electron. J. Stat. 2024, 18, 1455–1494. [Google Scholar] [CrossRef]
- Wang, J.T.; Mahloujifar, S.; Wu, T.; Jia, R.; Mittal, P. A Randomized Approach to Tight Privacy Accounting. arXiv 2023, arXiv:2304.07927. Available online: https://arxiv.org/abs/2304.07927 (accessed on 1 December 2024).
- Salem, A.; Zhang, Y.; Humbert, M.; Fritz, M.; Backes, M. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models. arXiv 2019, arXiv:1806.01246. Available online: https://arxiv.org/abs/1806.01246 (accessed on 1 December 2024).
- Ye, D.; Shen, S.; Zhu, T.; Liu, B.; Zhou, W. One Parameter Defense—Defending against Data Inference Attacks via Differential Privacy. arXiv 2022, arXiv:2203.06580. Available online: https://arxiv.org/abs/2203.06580 (accessed on 1 December 2024). [CrossRef]
- Cummings, R.; Desfontaines, D.; Evans, D.; Geambasu, R.; Huang, Y.; Jagielski, M.; Kairouz, P.; Kamath, G.; Oh, S.; Ohrimenko, O.; et al. Advancing Differential Privacy: Where We are Now and Future Directions. Harv. Data Sci. Rev. 2024, 6, 475–489. [Google Scholar] [CrossRef]
- Zhang, G.; Liu, B.; Zhu, T.; Ding, M.; Zhou, W. Label-Only Membership Inference attacks and Defense in Semantic Segmentation Models. IEEE Trans. Dependable Secur. Comput. 2023, 20, 1435–1449. [Google Scholar] [CrossRef]
- Wu, Y.; Qiu, H.; Guo, S.; Li, J.; Zhang, T. You Only Query Once: An Efficient Label-Only Membership Inference Attack. In Proceedings of the 12th International Conference on Learning Representations, ICLR 2024, Hybrid, Vienna, 7–11 May 2024; Available online: https://openreview.net/forum?id=7WsivwyHrS¬eId=QjoAoa8UVW (accessed on 3 December 2024).
- Li, N.; Qardaji, W.; Su, D.; Wu, Y.; Yang, W. Membership privacy: A Unifying Framework for Privcy Definitions. In Proceedings of the ACM Conference on Computers and Communication Security (CCS), Berlin, Germany, 4–8 November 2013; pp. 889–900. [Google Scholar] [CrossRef]
- Andrew, G.; Kairouz, P.; Oh, S.; Oprea, A.; McMahan, H.B.; Suriyakumar, V. One-shot Empirical Privacy for Federated Learning. arXiv 2024, arXiv:2302.03098. [Google Scholar]
- Patel, N.; Shokri, R.; Zick, Y. Model Explanations with Differential Privacy. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (FaccT ’22), Seoul, Republic of Korea, 21–24 June 2022; ACM: New York, NY, USA, 2022. [Google Scholar] [CrossRef]
- Ding, Z.; Tian, Y.; Wang, G.; Xiong, J. Regularization Mixup Adversarial Training: A Defense Strategy for Membership Privacy with Model Availbility Assurance. In Proceedings of the 2024 2nd Interntional Conference on Big Data and Privacy Computing, BDPC, Macau, China, 10–12 January 2024; pp. 206–212. [Google Scholar]
- Qui, W. A Survey on Poisoning Attacks Against Supervised Machine Learning. arXiv 2022, arXiv:2202.02510. Available online: https://arxiv.org/abs/2202.02510 (accessed on 1 December 2024).
- Zhao, B. Towards Class-Oriented Poisoning Attacks Against Neural Networks. In Proceedings of the 2022 IEEE/CVF Winter Conference on Application of Computer Vision, WACV, Waikoloa, HI, USA, 3–8 January 2022; pp. 2244–2253. [Google Scholar] [CrossRef]
- Koh, P.W.; Steinhardt, J.; Liang, P. Stronger data poisoning attacks data sanitization defenses. arXiv 2021, arXiv:1811.00741. Available online: https://arxiv.org/abs/1811.00741 (accessed on 1 December 2024). [CrossRef]
- Zhang, R.; Gou, S.; Wang, J.; Xie, X.; Tao, D. A Survey on Gradient Inversion Attacks, Defense and Future Directions. In Proceedings of the 31st Joint Conference on Artificial Intelligence (IJCAI-22), Vienna, Austria, 23–29 July 2022; pp. 5678–5685. Available online: https://www.ijcai.org/proceedings/2022/0791.pdf (accessed on 1 December 2024).
- Yan, H.; Wang, Y.; Yao, L.; Zhong, X.; Zhao, J. A Stacionary Random Process based Privacy-Utility Tradeoff in Differential Privacy. In Proceedings of the 2023 International Confernce on High Performance Big Data and Intelligence Systems, HDIS 2023, Macau, China, 6–8 December 2023; pp. 178–185. [Google Scholar]
- D’Oliveira, R.G.L.; Salamtian, S.; Médard, M. Low Influence, Utiltiy, and Independence in Differential Privacy: A Curious Case of (32). IEEE J. Sel. Areas Inf. Theory 2021, 2, 240–252. [Google Scholar] [CrossRef]
- Chen, M.; Liu, C.; Li, B.; Lu, K.; Song, D. Targeted Backdoor attacks on deed learning systems using data poisoning. arXiv 2017, arXiv:1712.05526. Available online: https://arxiv.org/abs/1712.05526 (accessed on 1 December 2024).
- Feng, S.; Tramèr, F. Privacy Backdoors: Stealing Data with Corrupted Pretrained Models. arXiv 2024, arXiv:2404.00473. Available online: https://arxiv.org/abs/2404.00473 (accessed on 1 December 2024).
- Gu, T.; Dolan-Gavitt, B.; Garg, S. BadNets: Identifying vulnerabilities in the machine learning model supply chain. arXiv 2019, arXiv:1708.06733. Available online: https://arxiv.org/abs/1708.06733 (accessed on 1 December 2024).
- Demelius, L.; Kern, R.; Trügler, A. Recent Advances of Differential Privacy in Centralized Deep Learning: A Systematic Survey. arXiv 2023, arXiv:2309.16398. Available online: https://arxiv.org/abs/2309.16398 (accessed on 1 December 2024).
- Oprea, A.; Singhal, A.; Vassilev, A. Poisoning attacks against machine learning: Can machine learning be trustworthy? Computer 2022, 55, 94–99. Available online: https://ieeexplore.ieee.org/document/9928202 (accessed on 1 December 2024). [CrossRef]
- Salem, A.; Wen, R.; Backes, M.; Ma, S.; Zhang, Y. Dynamic Backdoor Attacks Against Machine Learning Models. In Proceedings of the IEEE European Symposium Security Privacy (EuroS&P), Genoa, Italy, 6–10 June 2022; pp. 703–718. [Google Scholar] [CrossRef]
- Xu, X.; Chen, Y.; Wang, B.; Bian, Z.; Han, S.; Dong, C.; Sun, C.; Zhang, W.; Xu, L.; Zhang, P. CSBA: Covert Semantic Backdoor Attack Against Intelligent Connected Vehicles. IEEE Trans. Veh. Technol. 2024, 73, 17923–17928. [Google Scholar] [CrossRef]
- Li, X.; Li, N.; Sun, W.; Gong, N.Z.; Li, H. Fine-grained Poisoning attack to Local Differential Privacy Protocols for Mean and Variance Estimation. In Proceedings of the 32nd USENIX Security Symposium (USINEX Security), Anaheim, CA, USA, 9–11 August 2023; Volume 3, pp. 1739–1756. Available online: https://www.usenix.org/conference/usenixsecurity23/presentation/li-xiaoguang (accessed on 3 December 2024).
- Fang, X.; Yu, F.; Yang, G.; Qu, Y. Regression Analysis with Differential Privacy Preserving. IEEE Access 2019, 7, 129353–129361. [Google Scholar] [CrossRef]
- Wang, Y.; Si, C.; Wu, X. Regression Model Fitting under Differential Privacy and Model Inversion Attack. In Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI), Buenos Aires, Argentina, 25–31 July 2015; pp. 1003–1009. [Google Scholar]
- Dibbo, S.V. SoK: Model Inversion Attack Landscape: Taxonomy, Challenges, and Future Roadmap. In Proceedings of the IEEE 36th Computer Security Foundations Symposium (CSF), Dubrovnik, Croatia, 10–14 July 2023; Available online: https://ieeexplore.ieee.org/document/10221914 (accessed on 1 December 2024).
- Wu, X.; Fredrikson, M.; Jha, S.; Naughton, J.F. A methodology for formalizing model-inversion attacks. In Proceedings of the 2016 IEEE 29th Computer Security Foundations Symposium (CSF), Lisbon, Portugal, 27 June–1 July 2016; pp. 355–370. Available online: https://ieeexplore.ieee.org/document/7536387 (accessed on 3 December 2024).
- Park, C.; Hong, D.; Seo, C. An Attack-Based Evaluation Method for Differentially Private Learning Against Model Inversion Attack. IEEE Access 2019, 7, 124988–124999. [Google Scholar] [CrossRef]
- Zhao, J.; Chen, Y.; Zhang, W. Differential Privacy Preservation in Deep Learning: Challenges, Opportunities and Solutions. IEEE Access 2019, 7, 48901–48911. [Google Scholar] [CrossRef]
- Yang, Z.; Zhang, J.; Chang, E.-C.; Liang, Z. Neural Network in Adversarial Setting via Background Knowledge Alignment. In Proceedings of the 2019 ACM SIGSAC Conf. on Computing and Communication Security, London, UK, 11–15 November 2019; pp. 225–240. [Google Scholar] [CrossRef]
- Han, G.; Choi, J.; Lee, H.; Kim, J. Reinforcement Learning-Based Black-Box Model Inversion Attacks. arXiv 2023, arXiv:2304.04625. Available online: https://arxiv.org/abs/2304.04625 (accessed on 1 December 2024).
- Han, G.; Choi, J.; Lee, H.; Kim, J. Reinforcement Learning-Based Black-Box Model Inversion Attacks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 20504–20513. [Google Scholar] [CrossRef]
- Bekman, T.; Abolfathi, M.; Jafarian, H.; Biswas, A.; Banaei-Kashani, F.; Das, K. Practical Black Box Model Inversion Attacks Against Neural Nets. Commun. Comput. Inf. Sci. 2021, 1525, 39–54. [Google Scholar] [CrossRef]
- Du, J.; Hu, J.; Wang, Z.; Sun, P.; Gong, N.Z.; Ren, K. SoK: Gradient Leakage in Federated Learning. arXiv 2024, arXiv:2404.05403. Available online: https://arxiv.org/abs/2404.05403 (accessed on 1 December 2024).
- Li, Z.; Pu, Y.; Zhang, X.; Li, Y.; Li, J.; Ji, S. Protecting Object Detection Models From Model Extraction Attack via Feature Space Coverage. In Proceedings of the 33rd International Joint Conference on Artificial Intelligence (IJCAI), Jeju, Republic of Korea, 3–9 August 2024; pp. 431–439. [Google Scholar] [CrossRef]
- Tramér, F.; Zhang, F.; Juels, A.; Reiter, M.K.; Ristenpart, T. Stealing Machine Learning Models and Prediction APIs. In Proceedings of the USENIX Security Symposium (SEC), Austin, TX, USA, 10–12 August 2016; Available online: https://www.usenix.org/conference/usenixsecurity16/technical-sessions/presentation/tramer (accessed on 3 December 2024).
- Liang, J.; Pang, R.; Li, C.; Wang, T. Model Extraction Attacks Revisited. arXiv 2023, arXiv:2312.05386. Available online: https://arxiv.org/abs/2312.05386 (accessed on 1 December 2024).
- Liu, S. Model Extraction Attack and Defense on Deep Generative Models. J. Phys. Conf. Ser. 2022, 2189, 012024. [Google Scholar] [CrossRef]
- Parisot, M.P.M.; Pejo, B.; Spagnuelo, D. Property Inference Attacks on Convolution Neural Networks: Influence and Implications of Target Model’s Complexity. arXiv 2021, arXiv:2104.13061. Available online: https://arxiv.org/abs/2104.13061 (accessed on 1 December 2024).
- Zhang, W.; Tople, S.; Ohrimenko, O. Leakage of dataset properties in Multi-Party machine learning. In Proceedings of the 30th USINEX Security Symposium (USINEX Security), virtual, 11–13 August 2021; USINEX Association: Berkeley, CA, USA, 2021; pp. 2687–2704. Available online: https://www.usenix.org/conference/usenixsecurity21/presentation/zhang-wanrong (accessed on 1 December 2024).
- Mahloujifar, S.; Ghosh, E.; Chase, M. Property Inference from Poisoning. In Proceedings of the 2022 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 22–26 May 2022; pp. 1120–1137. [Google Scholar] [CrossRef]
- Horigome, H.; Kikuchi, H.; Fujita, M.; Yu, C.-M. Robust Estimation Method against Poisoning Attacks for Key-Value Data Local Differential Privacy. Appl. Sci. 2024, 14, 6368. [Google Scholar] [CrossRef]
- Parisot, M.P.M.; Pejó, B.; Spagnuelo, D. Property Inference Attacks on Convolutional Neural Networks: Influence and Implications of Target Model’s Complexity. In Proceedings of the 18th International Conference on Security and Cryptography, SECRYPT, Online, 6–8 July 2021; pp. 715–721. [Google Scholar] [CrossRef]
- Chase, M.; Ghosh, E.; Mahloujifar, S. Property Inference from Poisoning. arXiv 2021, arXiv:2101.11073. Available online: https://arxiv.org/abs/2101.11073 (accessed on 1 December 2024).
- Liu, X.; Xie, L.; Wang, Y.; Zou, J.; Xiong, J.; Ying, Z.; Vasilakos, A.V. Privacy and Security Issues in Deep Learning: A Survey. In IEEE Access 2020, 9, 4566–4593. [Google Scholar] [CrossRef]
- Gilbert, A.C.; McMillan, A. Property Testing for Differential Privacy. In Proceedings of the 56th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 2–5 October 2018; pp. 249–258. [Google Scholar] [CrossRef]
- Liu, X.; Oh, S. Minimax Optimal Estimation of Approximate Differential Privacy on Neighbouring Databases. In Proceedings of the 33rd International Conference on Neural Information Processing Systems (NIPS’19), Vancouver, BC, Canada, 8 December 2019; pp. 2417–2428. Available online: https://dl.acm.org/doi/10.5555/3454287.3454504 (accessed on 1 December 2024).
- Tschantz, M.C.; Kaynar, D.; Datta, A. Formal Verification of Differential Privacy for Interactive Systems (Extended Abstract). Electron. Notes Theor. Comput. Sci. 2011, 276, 61–79. [Google Scholar] [CrossRef]
- Pillutla, K.; McMahan, H.B.; Andrew, G.; Oprea, A.; Kairouz, P.; Oh, S. Unleashing the Power of Randomization in Auditing Differential Private ML. Adv. Neural Inf. Process. Syst. 2023, 36, 198465. Available online: https://arxiv.org/abs/2305.18447 (accessed on 3 December 2024).
- Cebere, T.; Bellet, A.; Papernot, N. Tighter Privacy Auditing of DP-SGD in the Hidden State Threat Model. arXiv 2024, arXiv:2405.14457. Available online: https://arxiv.org/abs/2405.14457 (accessed on 1 December 2024).
- Zhang, J.; Das, D.; Kamath, G.; Tramèr, F. Membership Inference Attacks Cannot Prove that a Model Was Trained On Your Data. arXiv 2024, arXiv:2409.19798. Available online: https://arxiv.org/abs/2409.19798 (accessed on 3 December 2024).
- Yin, Y.; Chen, K.; Shou, L.; Chen, G. Defending Privacy against More Knowledge Membership Inference Attackers. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Singapore, 14–18 August 2021; pp. 2026–2036. [Google Scholar] [CrossRef]
- Bichsel, B.; Gehr, T.; Drachsler-Cohen, D.; Tsankov, P.; Vechev, M. DP-Finder: Finding Differential Privacy Violations, by Sampling and Optimization. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security (CCS ’18), Toronto, ON, Canada, 15–19 October 2018; ACM: New York, NY, USA, 2018. [Google Scholar] [CrossRef]
- Niu, B.; Zhou, Z.; Chen, Y.; Cao, J.; Li, F. DP-Opt: Identify High Differential Privacy Violation by Optimization. In Wireless Algorithms, Systems, and Applications. WASA 2022; Wang, L., Segal, M., Chen, J., Qiu, T., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2022; Volume 13472. [Google Scholar] [CrossRef]
- Birhane, A.; Steed, R.; Ojewale, V.; Vecchione, B.; Raji, I.D. AI auditing: The broken bus on the road to AI accountability. arXiv 2024, arXiv:2401.14462. Available online: https://arxiv.org/abs/2401.14462 (accessed on 1 December 2024).
- Dwork, C. A Firm Foundation for Private Data Analysis. Commun. ACM 2011, 54, 86–95. [Google Scholar] [CrossRef]
- Dwork, C.; Su, W.J.; Zhang, L. Differential Private False Discovery Rate. J. Priv. Confidentiality 2021, 11, 2. [Google Scholar] [CrossRef]
- Liu, C.; He, X.; Chanyaswad, T.; Wang, S.; Mittal, P. Investigating Statistical Privacy Frameworks from the Perspective of Hypothesis Testing. Proc. Priv. Enhancing Technol. (PoPETs) 2019, 2019, 234–254. [Google Scholar] [CrossRef]
- Balle, B.; Barthe, G.; Gaboardi, M.; Hsu, J.; Sato, T. Hypothesis Testing Interpretations and Rényi Differential Privacy. In Proceedings of the 23rd International Conference on Artificial Intelligence and Statisitcs (AISTATS), Online, 26–28 August 2020; Volume 108, pp. 2496–2506. [Google Scholar]
- Kairouz, P.; Oh, S.; Viswanath, P. The Composition Theorem for Differential Privacy. In Proceedings of the 32nd International Conference on Machine Learning, ICML, Lille, France, 6–11 July 2015; pp. 1376–1385. Available online: https://proceedings.mlr.press/v37/kairouz15.html (accessed on 3 December 2024).
- Lu, Y.; Magdon-Ismail, M.; Wei, Y.; Zikas, V. Eureka: A General Framework for Black-box Differential Privacy Estimators. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 19–23 May 2024; pp. 913–931. [Google Scholar]
- Shamsabadi, A.S.; Tan, G.; Cebere, T.I.; Bellet, A.; Haddadi, H.; Papernot, N.; Wang, X.; Weller, A. Confident-Dproof: Confidential Proof of Differential Private Training. In Proceedings of the 12th International Conference on Learning Representations, ICLR, Hybrid, Vienna, 7–11 May 2024; Available online: https://openreview.net/forum?id=PQY2v6VtGe#tab-accept-oral (accessed on 1 December 2024).
- Kazmi, M.; Lautraite, H.; Akbari, A.; Soroco, M.; Tang, Q.; Wang, T.; Gambs, S.; Lécuyer, M. PANORAMIA: Privacy Auditing of Machine Learning Models without Retraining. arXiv 2024, arXiv:2402.09477. Available online: https://arxiv.org/abs/2402.09477 (accessed on 1 December 2024).
- Song, L.; Shokri, R.; Mittal, P. Membership Inference Attacks Against Adversarially Robust Deep Learning Models. In Proceedings of the 30th USENIX Security Symposium (USENIX Security 21), San Francisco, CA, USA, 19–23 May 2019. [Google Scholar]
- Koskela, A.; Mohammadi, J. Black Box Differential Privacy Auditing Using Total Variation Distance. arXiv 2024, arXiv:2406.04827. Available online: https://arxiv.org/abs/2406.04827 (accessed on 1 December 2024).
- Chen, J.; Wang, W.H.; Shi, X. Differential Privacy Protection Against Membership Inference Attack on Machine Learning for Genomic Data. Pac. Symp. Biocomput. 2021, 26, 26–37. [Google Scholar] [CrossRef]
- Malek, M.; Mironov, I.; Prasad, K.; Shilov, I.; Tramèr, F. Antipodes of Label Differential Privacy: PATE and ALBI. arXiv 2021, arXiv:2106.03408. Available online: https://arxiv.org/abs/2106.03408 (accessed on 1 December 2024).
- Choquette-Choo, C.A.; Tramèr, F.; Carlini, N.; Papernot, N. Label-only Membership Inference Attacks. In Proceedings of the 38th International Conference on Machine Learning (ICML), Online, 18–24 July 2021; pp. 1964–1974. [Google Scholar]
- Rahman, M.A.; Rahman, T.; Laganière, R.; Mohammed, N.; Wang, Y. Membership Inference Attack against Differentially Private Deep Learning Models. Trans. Data Priv. 2018, 11, 61–79. [Google Scholar]
- Humphries, T.; Rafuse, M.; Lindsey, T.; Oya, S.; Goldberg, I.; Kerschbaum, F. Differential Private Learning does not Bound Membership Inference. arXiv 2020, arXiv:2010.12112. Available online: http://www.arxiv.org/abs/2010.12112v1 (accessed on 1 December 2024).
- Askin, Ö.; Kutta, T.; Dette, H. Statistical Quantification of Differential Privacy. arXiv 2022, arXiv:2108.09528. Available online: https://arxiv.org/abs/2108.09528 (accessed on 1 December 2024).
- Aerni, M.; Zhang, J.; Tramèr, F. Evaluation of Machine Learning Privacy Defenses are Misleading. arXiv 2024, arXiv:2404.17399. Available online: https://arxiv.org/abs/2404.17399 (accessed on 1 December 2024).
- Kong, Z.; Chowdhury, A.R.; Chaudhurury, K. Forgeability and Membership Inference Attacks. In Proceedings of the 15th ACM Workshop on Artificial Intelligence and Security (AISec ’22), Los Angeles, CA, USA, 11 November 2022. [Google Scholar] [CrossRef]
- Kutta, T.; Askin, Ö.; Dunsche, M. Lower Bounds for Rényi Differential Privacy in a Black-Box Settings. arXiv 2022, arXiv:2212.04739. Available online: https://arxiv.org/abs/2212.04739 (accessed on 1 December 2024).
- Domingo-Enrich, C.; Mroueh, Y. Auditing Differential Privacy in High Dimensions with the Kernel Quantum Rényi Divergence. arXiv 2022, arXiv:2205.13941. Available online: https://arxiv.org/abs/2205.13941 (accessed on 1 December 2024).
- Koh, P.W.; Ling, P. Understanding Black-box Predictions via Influence Functions. arXiv 2017, arXiv:1703.04730. Available online: https://arxiv.org/abs/1703.04730 (accessed on 1 December 2024).
- Chen, C.; Campbell, N.D. Understanding training-data leakage from gradients in neural networks for image classification. arXiv 2021, arXiv:2111.10178. Available online: https://arxiv.org/abs/2111.10178 (accessed on 1 December 2024).
- Xie, Z.; Yan, L.; Zhu, Z.; Sugiyama, M. Positive-Negative Momentum: Manipulating Stochastic Gradient Noise to Imrove Generalization. arXiv 2021, arXiv:2103.17182. Available online: https://arxiv.org/abs/2103.17182 (accessed on 2 December 2024).
- Liu, F.; Zhao, X. Disclosure Risk from Homogeneity Attack in Differntial Private Frequency Distribution. arXiv 2021, arXiv:2101.00311. Available online: https://arxiv.org/abs/2101.00311 (accessed on 1 December 2024).
- Steinke, T.; Ullman, J. Between Pure and Approximate Differential Privacy. arXiv 2015, arXiv:1501.06095. Available online: https://arxiv.org/abs/1501.06095 (accessed on 1 December 2024). [CrossRef]
- Kairouz, P.; McMahan, B.; Song, S.; Thakkar, O.; Xu, Z. Practical and Private (Deep) Learning without Sampling on Shuffling. In Proceedings of the 38th International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 5213–5225. Available online: https://proceedings.mlr.press/v139/kairouz21b.html (accessed on 3 December 2024).
- Li, Y. Theories in Online Information Privacy Research: A Critical Review and an Integrated Framework. Decis. Support. Syst. 2021, 54, 471–481. [Google Scholar] [CrossRef]
- Hay, M.; Machanavajjhala, A.; Miklau, G.; Chen, Y.; Zhang, D. Principled evaluation of differential private algorithms using DPBench. In Proceedings of the ACM SIGMOD Conference on Management of Data, San Francisco, CA, USA, 26 June–1 July 2016; pp. 919–938. [Google Scholar] [CrossRef]
- Wang, Y.; Ding, Z.; Kifer, D.; Zhang, D. Checkdp: An Automated and Integrated Approach for Proving Differential Privacy or Finding Precise Counterexamples. In Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, 9–13 November 2020; pp. 919–938. [Google Scholar] [CrossRef]
- Barthe, G.; Chadha, R.; Jagannath, V.; Sistla, A.P.; Viswanathan, M. Deciding Differential Privacy for Programming with Finite Inputs and Outpus. arXiv 2022, arXiv:1910.04137. Available online: https://arxiv.org/abs/1910.04137 (accessed on 2 December 2024).
- Hitaj, B.; Ateniese, G.; Perez-Cruz, F. Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning. arXiv 2017, arXiv:1702.07464. Available online: https://arxiv.org/abs/1702.07464 (accessed on 1 December 2024).
- Song, C.; Ristenpart, T.; Shmatikov, V. Machine Learning Models that Remember Too Much. In Proceedings of the ACM SIGSAC Conference on Computer and Communications Security (CCS), Dallas, TX, USA, 30 October–3 November 2017; pp. 587–601. [Google Scholar] [CrossRef]
- Cummings, R.; Durfee, D. Individual Sensitivity Preprocessing for Data Privacy. In Proceedings of the Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), Salt Lake City, UT, USA, 5–8 January 2020; pp. 528–547. [Google Scholar]
- Zhou, S.; Zhu, T.; Ye, D.; Yu, X.; Zhou, W. Boosting Model Inversion Attacks With Adversarial Examples. IEEE Trans. Dependable Secur. Comput. 2023, 21, 1451–1468. [Google Scholar] [CrossRef]
- Zhu, L.; Liu, Z.; Han, S. Deep Leakage from Gradients. arXiv 2019, arXiv:1906.08935. Available online: https://arxiv.org/abs/1906.08935 (accessed on 1 December 2024).
- Huang, Y.; Gupta, S.; Song, Z.; Li, K.; Arora, S. Evaluating Gradient Inversion Attacks and Defenses in Federated Learning. Adv. Neural Netw. Inf. Process. Syst. 2021, 9, 7232–7241. Available online: https://proceedings.neurips.cc/paper_files/paper/2021/hash/3b3fff6463464959dcd1b68d0320f781-Abstract.html (accessed on 3 December 2024).
- Wu, R.; Chen, X.; Guo, C.; Weinberger, K.Q. Learning to Invert: Simple Adaptive Attacks for Gradient Inversion in Federated Learning. In Proceedings of the 39th Conferrence on Uncertainty in Artificial Intelligence (UAI), Pittsburgh, PA, USA, 31 July–4 August 2023; Volume 216, pp. 2293–2303. Available online: https://proceedings.mlr.press/v216/wu23a.html (accessed on 3 December 2024).
- Zhu, H.; Huang, L.; Xie, Z. GGI: Generative Gradient Inversion Attack in Federated Learning. In Proceedings of the 6th International Conference on Data-Driven Optimization of Complex Systems(DOCS), Hangzhou, China, 16–18 August 2024; pp. 379–384. Available online: http://arxiv.org/pdf/2405.10376.pdf (accessed on 3 December 2024).
- Yang, Z.; Zhang, B.; Chen, G.; Li, T.; Su, D. Defending Model Inversion and Membership Inference Attacks vi Prediction Purification. In Proceedings of the IEEE/CVF Conference on Computing Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1234–1243. [Google Scholar]
- Zhang, Q.; Ma, J.; Xiao, Y.; Lou, J.; Xiong, L. Broadening Differential Privacy for Deep Learning against Model Inversion Attacks. In Proceedings of the 2020 IEEE International Conference on Big Data, Atlanta, GA, USA, 10–13 December 2020; pp. 1061–1070. [Google Scholar] [CrossRef]
- Manchini, C.; Ospina, R.; Leiva, V.; Martin-Barreiro, C. A new approach to data differential privacy based on regression models under heteroscedasticity with applications to machine learning repository data. Inf. Sci. 2023, 627, 280–300. [Google Scholar] [CrossRef]
- Zhang, Z.; Liu, Q.; Huang, Z.; Wang, H.; Lu, C.; Liu, C.; Chen, E. GraphMI: Extracting Private Graph Data from Graph Neural Networks. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Montreal, QC, Canada, 19–27 August 2021; pp. 3749–3755. [Google Scholar] [CrossRef]
- Dziedzic, A.; Kaleem, M.A.; Lu, Y.S.; Papernot, N. Increasing the Cost of Model Extraction with Calibrated Proof of Work. In Proceedings of the 10th International Conference on Learning Representations (ICLR), Online, 25 April 2022; Available online: https://openreview.net/forum?id=EAy7C1cgE1L (accessed on 3 December 2024).
- Li, X.; Yan, H.; Cheng, Z.; Sun, W.; Li, H. Protecting Regression Models with Personalized Local Differential Privacy. IEEE Trans. Dependable Secur. Comput. 2023, 20, 960–974. [Google Scholar] [CrossRef]
- Zheng, H.; Ye, Q.; Hu, H.; Fang, C.; Shi, J. BDPL: A Boundary Differential Private Layer Against Machine Learning Model Extraction Attacks. In Computer Security—ESORICS 2019; Sako, K., Schneider, S., Ryan, P., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2019; Volume 11735. [Google Scholar] [CrossRef]
- Yan, H.; Li, X.; Li, H.; Li, J.; Sun, W.; Li, F. Monitoring-Based Differential Privacy Mechanism Against Query Flooding-based Model Extraction Attack. IEEE Trans. Dependable Secur. Comput. 2022, 19, 2680–2694. [Google Scholar] [CrossRef]
- Suri, A.; Lu, Y.; Chen, Y.; Evans, D. Dissecting Distribution Inferrence. In Proceedings of the 2023 IEEE Confernce Security and Trustworthy Machine Learning (SaTML), Raleigh, NC, USA, 8–10 February 2023; pp. 150–164. [Google Scholar]
- Ganju, K.; Wang, Q.; Yang, W.; Gunter, C.A.; Borisov, N. Property Inference Attacks on Fully Connected Neural Networks using Permutation Invariant Representations. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communication Security, Toronto, ON, Canada, 15–19 October 2018; pp. 619–633. Available online: https://dl.acm.org/doi/10.1145/3243734.3243834 (accessed on 1 December 2024).
- Melis, L.; Song, C.; De Cristofaro, E.; Shmatikov, V. Exploiting Unintended Feature Leakage in Collaborative Learning. In Proceedings of the Symposium on Security and Privacy (SP), San Francisco, CA, USA, 19–23 May 2019; pp. 691–706. [Google Scholar]
- Huang, W.; Zhou, S. Unexpected Information Leakage of Differential Privacy Due to the Linear Properties of Queries. IEEE Trans. Inf. Forensics Secur. 2021, 16, 3123–3137. [Google Scholar] [CrossRef]
- Ben Hamida, S.; Hichem, M.; Jemai, A. How Differential Privacy Reinforces Privacy of Machine Learning Modeles? In Proceedings of the International Conference on Computational Collective Intelligence (ICCI), Leipzig, Germany, 9–11 September 2024. [Google Scholar]
- Song, L.; Mittal, P.; Gong, N.Z. Systematic Evaluation of Privacy Risks in Machine Learning Models. In Proceedings of the ACM on Asian Conference on Computer and Communication Security, Taipei, Taiwan, 5–9 October 2020. [Google Scholar]
Attack | DPML Stages Impact | Type of Attack | Attack Techniques |
---|---|---|---|
Membership inference | Training data | White-box membership inference attack | Gradient-based approaches: Exploiting gradients whether specific data points were part of the training dataset. |
Activation analysis: Exploiting the activations for training data based on the assumption that they differ in certain layers from the activations for non-training data in certain layers. | |||
Black-box membership inference attack | Training shadow models: Creating and training a set of models that mimic the behavior of the targeted model. | ||
Confidence score analysis: Construct and analyze confidence scores or confidence intervals. | |||
Label-only membership inference attack | Adaptive querying: Modifying the inputs to answer queries that are individually selected, where each query depends on the answer to the previous query when the model changes the label. | ||
Meta-classification: Training a secondary model to distinguish between the labels of training and non-training data. | |||
Transfer membership inference attack | Model approximation: Using approximation algorithms to test the decision boundaries of the target model. | ||
Adversarial examples: Using adversarial techniques to evaluate privacy guarantees. | |||
Data poisoning | Training phase/model, data | Gradient manipulation attack | The gradients are intentionally altered during the model training process. |
Targeted label flipping | Label modification of certain data points in the training data without changing the data themselves. | ||
Backdoor poisoning | Inserting a specific trigger or “backdoor”. | ||
Data injection | Injecting malicious data samples that are designed to disrupt the model’s training. | ||
Adaptive querying and poisoning | Injecting a slightly modified version of data points and analyzing how these changes affect label predictions. | ||
Model inversion | Model | White-box inversion attacks | The attacker uses detailed insights into the model’s structure and parameters (e.g., model weights or gradients) to recover private training data. |
Black-box inversion attacks | The attacker iteratively queries the model and uses the outputs to infer sensitive information without access to the model’s internals. | ||
Inferring sensitive attributes from the model | Balancing the privacy budget for sensitive and non-sensitive attributes. | ||
Gradient-based inversion attacks | The attacker tries to recover private training data from shared gradients. | ||
Model extraction | Model | Adaptive Query-Flooding Parameter Duplication (QPD) attack | Allow the attacker to infer model information with black-box access and no prior knowledge of model parameters or training data. |
Equation-solving attack | Targets regression models by adding high-dimensional Gaussian noise to model coefficients. | ||
Membership-based property inference | Combines membership inference with property inference, targeting specific subpopulations with unique features. |
Privacy Auditing Scheme | Privacy Attack | Auditing Methodology |
---|---|---|
Membership inference audits | White-box membership inference auditing | Auditors analyze gradients, hidden layers, intermediate activations measuring how training data influences model behavior. |
Black-box membership inference auditing | Auditors observe probability distributions and confidence scores by analyzing these outputs and assessing the likelihood that certain samples were part of the training data. | |
Shadow model membership auditing | Auditors use “shadow models” to mimic the behavior of the target model. | |
Label-only membership inference auditing | Auditor evaluates the privacy guarantee leveraging only output labels, training shadow models, generating a separate classifier, and quantifying true-positive rate and accuracy. | |
Single-training membership inference run auditing | Auditor leverages the ability to add or remove multiple training examples independently during the run. This approach focuses on estimating the lower bounds of the privacy parameters without the need for extensive retraining of the models. | |
Metric-based membership inference auditing | Auditor assesses privacy guarantees directly evaluating metrics and statistics derived from the model’s outputs (precision, recall, and F1-score) on data points. | |
Data augmentation-based auditing | Auditor generates or augmented data samples similar to training set, testing whether these samples reveal membership risk. | |
Data poisoning auditing | Influence-function analysis | Evaluate privacy by introduction malicious data. |
Gradient manipulation in DP training | Auditor alters the training data using back-gradient optimization, gradient ascent poisoning, etc. | |
Empirical evaluation of privacy loss | Auditor conducts quantitative analyses of how the privacy budgets is affected. | |
Simulation of worst-case poisoning scenarios | Auditor constructs approximate upper bounds on the privacy loss. | |
Model inversion auditing | Sensitivity analysis | Auditor quantifies how much private information is embedded in the model outputs. |
Gradient and weight analyses | Auditor attempts to recreate input features or private data points form model outputs using gradient-based or optimization methods. | |
Empirical privacy loss | Auditor calculates the difference between theoretical and empirical privacy losses. | |
Embedding and reconstruction test | Auditor examines whether latent representations or embeddings could be reversed to obtain private data. | |
Model extraction auditing | Query analysis | Auditors simulate extraction attacks by extensively querying the model and analyzing how well they can replicate its outputs or decision boundaries. |
Property inference auditing | Evaluating property sensitivity with model outputs. | The auditor performs a test to infer whether certain properties can be derived from the model and whether the privacy parameters are sufficient to obscure such data properties. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Namatevs, I.; Sudars, K.; Nikulins, A.; Ozols, K. Privacy Auditing in Differential Private Machine Learning: The Current Trends. Appl. Sci. 2025, 15, 647. https://doi.org/10.3390/app15020647
Namatevs I, Sudars K, Nikulins A, Ozols K. Privacy Auditing in Differential Private Machine Learning: The Current Trends. Applied Sciences. 2025; 15(2):647. https://doi.org/10.3390/app15020647
Chicago/Turabian StyleNamatevs, Ivars, Kaspars Sudars, Arturs Nikulins, and Kaspars Ozols. 2025. "Privacy Auditing in Differential Private Machine Learning: The Current Trends" Applied Sciences 15, no. 2: 647. https://doi.org/10.3390/app15020647
APA StyleNamatevs, I., Sudars, K., Nikulins, A., & Ozols, K. (2025). Privacy Auditing in Differential Private Machine Learning: The Current Trends. Applied Sciences, 15(2), 647. https://doi.org/10.3390/app15020647