Skip to content

Install FlashAttention on non-RDNA1/2 GPUs#2

Merged
patientx merged 2 commits intopatientx-cfz:masterfrom
0xDELUXA:add-flash-attn
Apr 2, 2026
Merged

Install FlashAttention on non-RDNA1/2 GPUs#2
patientx merged 2 commits intopatientx-cfz:masterfrom
0xDELUXA:add-flash-attn

Conversation

@0xDELUXA
Copy link
Copy Markdown

@0xDELUXA 0xDELUXA commented Apr 2, 2026

This PR adds automatic FlashAttention installation on GPUs except RDNA1/2 during setup.

See the results of SageAttn V1 vs FA2 Triton here: Dao-AILab/flash-attention#2400.
Ignore CK, as it’s very specific for now. The Triton backend works for all GPUs supported by aiter.

Based on the wheels available at: https://github.com/0xDELUXA/flash-attention/releases/tag/v2.8.4_win-rocm

@0xDELUXA 0xDELUXA marked this pull request as draft April 2, 2026 18:41
@0xDELUXA 0xDELUXA force-pushed the add-flash-attn branch 2 times, most recently from 55a8839 to 9fd94c2 Compare April 2, 2026 20:12
@0xDELUXA 0xDELUXA marked this pull request as ready for review April 2, 2026 20:14
@0xDELUXA 0xDELUXA marked this pull request as draft April 2, 2026 20:40
@0xDELUXA 0xDELUXA marked this pull request as ready for review April 2, 2026 21:48
@patientx patientx merged commit f8b2518 into patientx-cfz:master Apr 2, 2026
@0xDELUXA 0xDELUXA deleted the add-flash-attn branch April 3, 2026 14:08
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants