Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

master consideres batchnorm mean/std update as gradients #19

Open
vlimant opened this issue Apr 6, 2018 · 2 comments
Open

master consideres batchnorm mean/std update as gradients #19

vlimant opened this issue Apr 6, 2018 · 2 comments

Comments

@vlimant
Copy link
Owner

vlimant commented Apr 6, 2018

in the way the master receives the "update" from the workers, for the bn mean weight (the running mean) it would consider the diff as a gradient and do something with this, instead of applying something dedicated to a value that was not updated by gradient descent.
the gamma/beta weights of the bn are ok in this respect.
I fear that this is playing badly with svalleco#3
@duanders if you have any insights on how to modify the mpi-learn-optimizers to take this in consideration, please do tell

@duanders
Copy link
Contributor

duanders commented Apr 8, 2018

I see, glad you got to the bottom of it. Have you checked how the keras optimizers handle this?

@vlimant
Copy link
Owner Author

vlimant commented Apr 8, 2018

I was not able to track this down all the way

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants