Opacus
, based on
PyTorch
.
DP-SGD adds noise to the gradients during training to preserve the privacy of individual training data records. More noise implies better privacy, but this comes with a utility cost: the model will be less accurate. The privacy is typically evaluated according to (ε,δ)-DP values; roughly speaking, they bound the amount of information leakage (ε), and the probability with which this could be violated (δ). These values do not assume any specific threat (i.e. the attacker's objective).
The interactive tool below gives guidance on how to select the DP-SGD parameters (e.g., amount of noise) to achieve a certain level of resilience against a specific threat: membership inference. It has been shown that, if a system is resilient against membership inference, it is also resilient against all other privacy attacks targeting individual records. This tool is based on recent research by G. Cherubin, B. Köpf, A. Paverd, S. Tople, L. Wutschitz, and S. Zanella-Béguelin, which shows that the level of privacy achieved by training with DP-SGD can be approximated via a closed-form expression.
The tool below calculates the resilience of a model against membership attacks in terms of Bayes security.
This metric indicates how much information an attacker can gain by using the model, compared to an attacker who has no access to the model and relies only on their prior knowledge.
It takes values between [0,1]
, where 1
(perfect security) means that the model does not provide any additional information to the attacker.
This metric can thus be seen as describing the risk of releasing a model, with respect to the chosen threat.
This tool is offered as a research prototype, and it should be used only as a guidance.
The colored bands in the above calculator provide an example of how the results could be interpreted for a typical scenario. In this example, the bands are defined as follows:
Note: this interpretation is given as an example; users must determine the acceptable level of risk for their own scenarios.
The above interactive tool is based on the following assumptions:
The above analysis is based on an approximate expression for the Bayes security of DP-SGD, which we observed to work well (up to 0.015 absolute error on Bayes security) for the parameters ranged that the above tool allows. For a more accurate analysis, you may want to use a moment accountant (e.g., PRV accountant or PLD accountant).
Bayes security measures the advantage of an attacker performing an attack with access to the trained model, versus an attacker who has no access to the model and relies only on their prior knowledge. This is based on the concept of advantage, which is widely used both in cryptography and in privacy-preserving ML. Bayes security can be converted to other commonly-used metrics:
True positive rate at chosen false positive rate (TPR @ FPR). One way of interpreting Bayes security is by matching it to the maximum achievable true positive rate that an attacker can achieve when aiming for a (low) false positive rate; this metric is called TPR @ FPR. Intuitively, a good attack is one that confidently makes good predictions. The following tool shows the maximum achievable TPR for various choices of FPR, for given values of Bayes security.
Bayes security:
Attacker success rate. Another useful way to interpret a value of Bayes security is by matching it to the probability that an attack will succeed (attacker success rate). Because Bayes security measures the relation between an attacker with and without access to the model, to compute the success rate we need to assume a prior; that is: what is the probability that an attacker will succeed in the attack without access to the model?
Bayes security:
Prior probability:
The attacker is at most % likely to succeed in guessing whether a data record is a member or not, assuming that members have the above prior.
The analysis above considers membership inference, in which the attacker is assumed to have access to a complete data record and uses the model to ascertain whether that record was included in the training dataset. This is depicted graphically below:
It has been shown that resilience against membership inference implies an equivalent level of resilience against all other privacy attacks that target individual data records. However, there may be scenarios in which membership inference is not a concern. In these scenarios it may be possible to obtain stronger resilience bounds for other types of privacy attacks.
For example, in attribute inference, the attacker is assumed to have partial knowledge of a data record and uses the model to infer the remainder of the data record. This is depicted graphically below:
We developed a tool that enables quantifying resilience against attribute inference. This analysis is data-dependent, meaning that the results depend on both the DP-SGD training parameters as well as the specific dataset used for training. If the dataset itself is inherently resilient against attribute inference, this will be reflected in the results.
The tool is based on the Opacus
implementation of DP-SGD, and is available at: https://github.com/microsoft/dpsgd-calculator.
If you have a usage question, have found a bug, or have a suggestion for improvement, please file a Github issue, or get in touch at gcherubin [at] microsoft.com.
This calculator is based on the closed-form expression of the Bayes security metric for DP-SGD. Please cite as:
@inproceedings{cherubin2024closed, title={Closed-Form Bounds for DP-SGD against Record-level Inference}, author={Cherubin, Giovanni and Köpf, Boris and Paverd, Andrew and Tople, Shruti and Wutschitz, Lukas and Zanella-Béguelin, Santiago}, booktitle={33rd USENIX Security Symposium (USENIX Security 24)}, year={2024} }