Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Avoid overly strict "transformers==4.34.0", #26

Open
pseudotensor opened this issue Oct 24, 2023 · 2 comments
Open

Avoid overly strict "transformers==4.34.0", #26

pseudotensor opened this issue Oct 24, 2023 · 2 comments

Comments

@pseudotensor
Copy link

pseudotensor commented Oct 24, 2023

Makes it hard to upgrade. Thanks!

@tomaarsen
Copy link
Owner

Makes sense! I hate strict requirements myself, but for attention_sinks I'm somewhat forced into it. This project works by overriding the forward method of the ...Attention classes in transformers, which are often updated, even in minor versions. Any mismatch can cause failures, so I can really only support one transformers version at a time. At least, until huggingface/transformers#26681 is merged and Attention Sinks can be implemented that way.

  • Tom Aarsen

@pseudotensor
Copy link
Author

ok thanks for consideration.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants