-
Notifications
You must be signed in to change notification settings - Fork 4.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added support of google/flan models #2321
Conversation
|
||
def match(self, model_path: str): | ||
return "flan" in model_path.lower() | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What is the instruction template for flan? Do we need to set it here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @merrymercy, do you mean prompt template?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yes
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@wangzhen263 Any updates? IIRC, flan-t5 has a default prompt template.
bf7aa7e
to
a81a04c
Compare
Added AutoModelForSeq2SeqLM to init google/flan models in 8bit
Closed due to inactivity. Feel free to reopen after you address these problems.
|
Why are these changes needed?
Related issue number (if applicable)
Checks
format.sh
to lint the changes in this PR.