-
Notifications
You must be signed in to change notification settings - Fork 271
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Resume a training #444
Comments
I've already started to work on this here ApproachI've defined a new meta parameter called load_net "Test"In order to check if this approach work, I have modify the demo spec (DQN Cartpole) in order to split it into two parts, one part which trains until 3000 frames and the other part which loads the previous network and train for another 7000 frames. One-shot training
Two-shot trainingIn order to see better the improvement and as the number of frames is reduced, the evaluation and log frequency as been dropped from 500 to 100. First part
Second part
|
Hi @ingambe thanks for looking at this, and the fantastic showcase above! We did not implement a resume function yet but it should be relatively simple and clean, since the
If you open a PR I can also work with you to get the steps above implemented, or if you prefer to wait a bit I can also get to it sometime this week/next. |
Hi @kengz I've created a draft pull request #445 in order to work together on this. Thank you for your help |
Implemented resume mode, see the linked PR above. |
Are you requesting a feature or an implementation?
I would like to know if it is possible to load a previously trained experiment and continue the training (i.e. load the neural network and start a new training with the previously trained neural network as the initial network)
This would be useful in case our previous experiment didn't reach a plateau with the previously assigned number of step or maybe in order to use a reuse a previously trained neural network for a similar task
If you have any suggested solutions
Add a "resume_training" mode in order to continue the training
Or add the possibility to load a neural net model in a training spec
The text was updated successfully, but these errors were encountered: