Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some questions about model parameters. #1

Open
KoalaYan opened this issue Jan 18, 2024 · 1 comment
Open

Some questions about model parameters. #1

KoalaYan opened this issue Jan 18, 2024 · 1 comment

Comments

@KoalaYan
Copy link

Thank you for bringing attention to the code.

In the functions 'create_global_model_copy' and 'copy_params,' only the variables of ResNet are copied, excluding BatchNorm layer (BN) statistical information such as 'num_batches_tracked,' 'running_var,' and 'running_mean.' Does this imply that the server does not aggregate this information for the global model? If so, why does the global model from the 1900th iteration contain non-zero values for these statistics?

Additionally, the lines "model = self.helper.local_model" and "self.copy_params(model, global_model_copy)" suggest that the local model is refreshed with global model's variables, excluding BN statistical information. This raises a concern as it implies that all clients share their Batch Normalization statistical information, which appears unusual. Could you please clarify this aspect?

@Bast-V
Copy link

Bast-V commented Oct 7, 2024

Hello, thank you for having published your algorithm and the implementation. We are using A3FL for our works as a state-of-the-art attack on several type of architectures.
I have a question for attacker.py:L61-62 : the cosine similarity is computed with the gradient of the layers instead of the weights themselves as described in your paper (Section 4.1: A3FL setup). Is there a specific reason the cosine similarity is computed on the gradients ? Trying to compute the similarity on the weights seems to lead to similar results. Can you confirm that your attack still work when following the paper description (cos sim on the weights) or do you notice that using the gradient is more efficient ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants