You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First, thanks for this excellent library and for publishing research experiments!
I have questions about the privacy amplification by iteration script. Could the authors provide a clear explanation of the following:
Which theorem are they using for privacy accounting?
How was the theorem implemented in tensorflow privacy?
As far as I understand from this file (but please correct me if I am wrong), TF Privacy is computing a average over clipped gradients, and then noise has a scale of sensitivity * noise_multiplier. So the updates rule is
Where $\eta$ is the learning rate, C is the sensitivity, and B is the batch size. To account for this, the authors correctly multiply the noise term with the batch size so they can derive the correct privacy amplification by iteration guarantees, rewriting the above term as:
For them to be comparable, shouldn't we scale $\sigma$ with $B$ when computing the RDP analysis for SGM here?
The text was updated successfully, but these errors were encountered:
tudorcebere
changed the title
Privacy guarantees of of privacy amplification by iteration notebook
Privacy guarantees of of privacy amplification by iteration example
Sep 4, 2023
Hi!
First, thanks for this excellent library and for publishing research experiments!
I have questions about the privacy amplification by iteration script. Could the authors provide a clear explanation of the following:
As far as I understand from this file (but please correct me if I am wrong), TF Privacy is computing a average over clipped gradients, and then noise has a scale of sensitivity * noise_multiplier. So the updates rule is
Where$\eta$ is the learning rate, C is the sensitivity, and B is the batch size. To account for this, the authors correctly multiply the noise term with the batch size so they can derive the correct privacy amplification by iteration guarantees, rewriting the above term as:
That's how we can observe a RDP coefficient of:
Now, this is neat, but I am not sure this is comparable with the analysis of DP-SGD from here, as they are considering an update rule of:
For them to be comparable, shouldn't we scale$\sigma$ with $B$ when computing the RDP analysis for SGM here?
The text was updated successfully, but these errors were encountered: