Skip to content

Relate alpha, beta1, beta2 and epsilon to learning rate and momentum in adam_sgd optimizer

Chris Basoglu edited this page Dec 21, 2016 · 2 revisions

Here is a mapping :

  • Alpha is the learning_rate
  • Beta1 is momentum parameter
  • Beta2 is variance_momentum parameter

I don’t think you’re able to define the epsilon one.

Clone this wiki locally