Skip to content

Commit

Permalink
Merge pull request #862 from bghira/main
Browse files Browse the repository at this point in the history
schedulefree: return to previous stable settings and add a new preset for aggressive training
  • Loading branch information
bghira authored Aug 24, 2024
2 parents 413f774 + 76b377c commit e704678
Showing 1 changed file with 25 additions and 2 deletions.
27 changes: 25 additions & 2 deletions helpers/training/optimizer_param.py
Original file line number Diff line number Diff line change
Expand Up @@ -41,9 +41,32 @@
"override_lr_scheduler": True,
"can_warmup": True,
"default_settings": {
"betas": (0.9, 0.990),
"weight_decay": 1e-4,
"betas": (0.9, 0.999),
"weight_decay": 1e-2,
"eps": 1e-8,
},
"class": AdamWScheduleFreeKahan,
},
"adamw_schedulefree+aggressive": {
"precision": "any",
"override_lr_scheduler": True,
"can_warmup": True,
"default_settings": {
"betas": (0.9, 0.999),
"weight_decay": 1e-3,
"eps": 1e-6,
},
"class": AdamWScheduleFreeKahan,
},
"adamw_schedulefree+no_kahan": {
"precision": "any",
"override_lr_scheduler": True,
"can_warmup": True,
"default_settings": {
"betas": (0.9, 0.999),
"weight_decay": 1e-3,
"eps": 1e-6,
"use_kahan": False,
},
"class": AdamWScheduleFreeKahan,
},
Expand Down

0 comments on commit e704678

Please sign in to comment.