-
Notifications
You must be signed in to change notification settings - Fork 306
Add 7B presets for Mistral #1436
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add 7B presets for Mistral #1436
Conversation
|
Can you show a full generation run with the causal lm class running in a colab? Loaded from kaggle. |
[skip ci]
Here: https://colab.research.google.com/drive/1508c8IY_nQQIsF33nUc_87l0mLo2BFw0?usp=sharing |
|
This looks great! Just made the Kaggle model public so we can re-run ci. I think there is one small thing to fix here. With the Kaggle rewrite our tokenizers can be created without a vocabulary, which is loaded after the fact. Basically loading will look like To deal with this fact, we only grab the special tokens from the tokenizer the first time a preprocessing layer Instead of this... You might be able to trigger a bug if you save a |
|
Thanks for noticing this @mattdangerw! Should be fixed now. Let me know if it looks good to you now! |
[skip ci]
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm!
Closes #1418
Adds
mistral_7b_enandmistral_instruct_7b_enpresets for Mistral. Also added tests for presets for all the components of Mistral.