Update schema in test_flash3_bw_compatibility#2153
Update schema in test_flash3_bw_compatibility#2153guilhermeleobas wants to merge 2 commits intoDao-AILab:mainfrom
test_flash3_bw_compatibility#2153Conversation
|
Thanks, I do think the longterm solution if we were to fix the test is to update the schema. |
|
Isn't the test testing for backwards compatibility with Flash Attention 2? |
|
@guilhermeleobas do you mean this schema is from FA2? https://github.com/Dao-AILab/flash-attention/blob/main/hopper/test_flash_attn.py#L1244-L1250 I wasn't aware they were tied. |
|
Let me double-check that info, if FA3 schema is backward compatible with FA2. I'm not sure if that's feasible, but I can try to return empty tensors for flash-attention/hopper/flash_api_stable.cpp Line 1645 in 13696f2 |
I think you're right, FA2 and FA3 schemas are not tied. |
test_flash3_bw_compatibility as xfailtest_flash3_bw_compatibility
janeyx99
left a comment
There was a problem hiding this comment.
If we're okay BC breaking originally, this change now LGTM as long as the test passes locally.
I'm not sure if the alternative where we undo the BC break is better because it's already been a month or so and people have been building wheels with the current FA3. So changing it back will BC break those people 😬
This test will be broken until Dao-AILab/flash-attention#2153 is resolved, which doesn't seem soon. Pull Request resolved: #172452 Approved by: https://github.com/mikaylagawarecki
This test will be broken until Dao-AILab/flash-attention#2153 is resolved, which doesn't seem soon. Pull Request resolved: pytorch#172452 Approved by: https://github.com/mikaylagawarecki
xref: #1769 (comment)
Pull request #1769 breaks backward compatibility of the backward function as the C++
mha_bwdfunction returns some of its inputs (dq,dk,dv).cc: @janeyx99