Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Adam, AdaMax and FTML cannot be used with Trainer(update_on_kv=False) #13752

Closed
eric-haibin-lin opened this issue Jan 1, 2019 · 0 comments · Fixed by #14377
Closed

Adam, AdaMax and FTML cannot be used with Trainer(update_on_kv=False) #13752

eric-haibin-lin opened this issue Jan 1, 2019 · 0 comments · Fixed by #14377

Comments

@eric-haibin-lin
Copy link
Member

These optimizers adjust the learning rate by the number of optimization steps, which is problematic if update_on_kv=False is set and multiple GPUs share the same optimizer object (leading to incorrect count of optimization steps). For example: https://github.com/apache/incubator-mxnet/blob/master/python/mxnet/optimizer/optimizer.py#L1093

cc @szha @sandeep-krishnamurthy

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant