Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Potential bug in gradient accumulation? #9

Open
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

dorarad
Copy link

@dorarad dorarad commented Mar 7, 2020

For the gradient accumulation case, currently the regularization for the discriminator performs multiple rounds on the same batch of images, instead of averaging the regularization loss over multiple different batches (as is done for the standard discriminator loss).
Is this a bug or is it arranged this way intentionally?

Thanks!

For the gradient accumulation case, fix Discriminator regularization to use a different batch of real images every round.
@ghost
Copy link

ghost commented Aug 6, 2020

@tkarras will the accumulation bug be addressed?
https://www.reddit.com/r/MachineLearning/comments/i3si0j/d_stylegan2_potential_gradient_accumulation/

@dorarad
Copy link
Author

dorarad commented Aug 6, 2020

While I believe that might be a bug, I tested the difference between with and without loading new data and they were quite small so probably that doesn't affect things significantly. It also doesn't affect the performance at all for the case of using multi-gpu because than there's anyway only one round.

patrickxia pushed a commit to patrickxia/stylegan2 that referenced this pull request Jan 10, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant