Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Federated DPFTRL and Adaptive Clipping noise injection. #519

Open
BucketHeadP65 opened this issue Sep 27, 2023 · 1 comment
Open

Federated DPFTRL and Adaptive Clipping noise injection. #519

BucketHeadP65 opened this issue Sep 27, 2023 · 1 comment

Comments

@BucketHeadP65
Copy link

BucketHeadP65 commented Sep 27, 2023

Dear developers of the privacy framework. I was checking the implementation of quantile_adaptive_clip_tree_query.py

I can see that a new tree is required to inject noise after aggregating the norm bit as the traditional QuantileEstimatorQuery
is wrapped from the TreeQuantileEstimatorQuery

In this paper Federated Learning of Gboard Language Models with Differential Privacy I see the paragraph highlighted in the image below.
Question: Does this mean that I can just keep the logic of restarting the tree and then updating the clipping norm of the DPFTRL skipping the second tree that injects noise on norm bits and just replace the whole noise of Federated DP-FTRL with Adaptive clipping setup with just a Tree based Gaussian noise with Z as stated in the figure? I am confused here.

Thanks you in advance.

image

@nightldj
Copy link
Contributor

Thanks for your interest. The noise multiplier is slightly inflated when using adaptive clipping as suggested in the theory. This can be done once similar to https://github.com/tensorflow/federated/blob/main/tensorflow_federated/python/aggregators/differential_privacy.py#L400

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants