-
Notifications
You must be signed in to change notification settings - Fork 3.7k
[ONNX] Extend converter for Attention from Microsoft onnxruntime contrib opset #13797
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Thanks for contributing to TVM! Please refer to the contributing guidelines https://tvm.apache.org/docs/contribute/ for useful information and tips. Please request code reviews from Reviewers by @-ing them in a comment. Generated by tvm-bot |
echuraev
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
|
@tvm-bot rerun |
4a8b581 to
8e7462f
Compare
vvchernov
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. I belive that somewhere problems with magic number -10000 from ONNX Runtime will be resolved
|
Hello @AndrewZhaoLuo! |
…rib opset (apache#13797) * add type & shape checking * add base class for Attention converter * add support for 'past' input * add support for 'unidirectional' attribute * fix for 'huggingface implementation' * add common method for calculating Attention * expand test coverage for Attention
Earlier I added converter for
QAttention(see PR#13654).In this PR, I've added support for
pastinput andunidirectionalattribute inAttention. I've also moved the common code forQAttentionandAttentioninto a base class.