Skip to content

Fine-tune PLATO-2 #127

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
sdai654416 opened this issue Apr 12, 2022 · 2 comments
Open

Fine-tune PLATO-2 #127

sdai654416 opened this issue Apr 12, 2022 · 2 comments

Comments

@sdai654416
Copy link

  1. I download the 24L model, and run the finetune script bash ./scripts/local/job.sh ./projects/PLATO-2/finetune/24L_train.conf. I got nan for my loss at the very beginning of the fine-tuning. Am I missing any stages?

  2. If I do the pre-train script, the pretrain stage 1 does not store anything in the output/. I assume stage 2.1, and stage 2.2, requires stage 1's output right? How do I store stage 1? Thanks

Thanks!

@sserdoubleh
Copy link
Collaborator

sserdoubleh commented Apr 13, 2022

You can change: AMP setting in knover/core/model.py
https://github.com/PaddlePaddle/Knover/blame/develop/knover/core/model.py#L165

"custom_white_list": ["gelu"],

It seems that old models need to disable fp16 softmax / layer_norm.
Thanks for feedback!

@py703703
Copy link

py703703 commented May 3, 2022

As for your second question, I think the pre-training data is too small and the number of steps saved is too large. As a result, the training ends before the set number of steps is reached. Save_steps can be modified in/"projects/ PLATe-2 /pretrain/ 24l_train_stage-1.conf"

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants