Skip to content

Conversation

@garrett361
Copy link
Contributor

Currently, the SFT learning rate is run all the way down to zero by default in finetune.py. This PR adds a --final_lr_ratio flag which allows the LR to run only down to a value of learning_rate * final_lr_ratio over the course of SFT. Qwen, for instance, follows this strategy of only decaying the lr down by a fixed factrion.

Only currently implemented for the linear scheduler.

@hamishivi hamishivi merged commit 8cbf07b into allenai:main Jul 2, 2025
3 checks passed
finbarrtimbers pushed a commit that referenced this pull request Jul 2, 2025
finbarrtimbers pushed a commit that referenced this pull request Jul 2, 2025
@fabianlim fabianlim mentioned this pull request Jul 10, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants