-
Notifications
You must be signed in to change notification settings - Fork 637
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ChatLLaMA] RLHF Training: Prompt too long #299
Comments
Hi @swang99, thank you for reaching out! What model are you currently using? |
Hi @swang99, thanks for reaching out.
|
Thanks for the recommendations. The error still persists unfortunately. Can I simply increase the additional_prompt_tokens or would I need to save a new actor model? Below is my config.yaml
|
Hi @swang99 |
Hi @PierpaoloSorbellini thank you for rolling out the fixes. This might not be very specific, but although I was able to get further into training, around the 9th timestep the training stopped suddently due to a loss is NaN error. Has this been addressed in the past? |
I have the same problem. Did you fixed it? |
Hi @Mialiu91 @swang99,
|
I am getting the following error when doing RLHF training. I decreased the max_sequence_length in my actor configuration to 1024 because there were errors with training for me when set to 2048. Is my actor max_sequence_length too small, and does this mean I have to redo pre-training with a larger max sequence? There isn't a way to change the state_length to my knowledge.
ValueError: The prompt is too long w.r.t the model sequence length
max_sequence_length=1024
state_length=1024
min_tokens=100
max_tokens=2048
max_generation_possible=0
The text was updated successfully, but these errors were encountered: