forked from microsoft/CNTK
-
Notifications
You must be signed in to change notification settings - Fork 0
Relate alpha, beta1, beta2 and epsilon to learning rate and momentum in adam_sgd optimizer
Chris Basoglu edited this page Dec 21, 2016
·
2 revisions
Here is a mapping :
- Alpha is the learning_rate
- Beta1 is momentum parameter
- Beta2 is variance_momentum parameter
I don’t think you’re able to define the epsilon one.
- Home
- What's new
- Breaking changes
- CNTK installation
- Python API / Docs
- FAQ
- How do I...
- Troubleshoot CNTK
Getting Started
Additional Documentation
How to use CNTK
Using CNTK Models in Your Code
Advanced topics
- BrainScript via command line
- Top-level commands
- Working with sequences
- Plot Command
- ConvertDBN Command
- Baseline Metrics
Licenses
Source Code & Development