-
Notifications
You must be signed in to change notification settings - Fork 218
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BatchNormsync with Adam Optimizer #35
Comments
@tarun005 Have you tested with SGD optimizer? Does it drive the training process to convergence? |
Yes, the model converges with SGD, but same model does not if I replace SGD with adam. |
@tarun005 Although I suppose that BN should be irrelevant to the optimization method, when I used the syncbn by just adding the folder |
Agree that BN shouldn't be relevant to optimization method, but I have read somewhere that Adam requires global statistics at every iteration, so the implementation of BNsync given here could be an issue. |
Is the bnsync code written specifically for SGD optimizer? The loss is not converging if I use and train the model with Adam optimizer.
The text was updated successfully, but these errors were encountered: