Skip to content

[Bugfix] [ROCm] [UX]: revert Flex attention backend#29371

Merged
DarkLight1337 merged 1 commit intovllm-project:mainfrom
EmbeddedLLM:fix-rocmattnselection-replenish
Nov 25, 2025
Merged

[Bugfix] [ROCm] [UX]: revert Flex attention backend#29371
DarkLight1337 merged 1 commit intovllm-project:mainfrom
EmbeddedLLM:fix-rocmattnselection-replenish

Conversation

@vllmellm
Copy link
Copy Markdown
Contributor

@vllmellm vllmellm commented Nov 25, 2025

Description This PR restores the FLEX_ATTENTION backend selection logic that was accidentally removed in PR #26980.

Changes

Re-added explicit check for AttentionBackendEnum.FLEX_ATTENTION in vllm/platforms/rocm.py.

Added a corresponding unit test case in tests/v1/attention/test_rocm_attention_backends_selection.py to ensure it is correctly selected.

Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com>
@mergify mergify bot added rocm Related to AMD ROCm v1 labels Nov 25, 2025
@vllmellm
Copy link
Copy Markdown
Contributor Author

/gemini review

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request correctly restores the selection logic for the FLEX_ATTENTION backend, which was accidentally removed. The change in vllm/platforms/rocm.py re-introduces the necessary check. Additionally, a new unit test has been added in tests/v1/attention/test_rocm_attention_backends_selection.py to ensure this backend can be selected, which is a good practice to prevent future regressions. The changes are correct and well-contained.

@vllmellm vllmellm marked this pull request as ready for review November 25, 2025 04:43
@vllmellm vllmellm requested a review from tjtanaa as a code owner November 25, 2025 04:43
@vllmellm
Copy link
Copy Markdown
Contributor Author

Hi @tjtanaa , I fixed this, pls review the code when you have time. Thanks!

@DarkLight1337 DarkLight1337 enabled auto-merge (squash) November 25, 2025 04:46
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Nov 25, 2025
@DarkLight1337 DarkLight1337 merged commit 64deead into vllm-project:main Nov 25, 2025
48 of 50 checks passed
devpatelio pushed a commit to SumanthRH/vllm that referenced this pull request Nov 29, 2025
Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com>
kitaekatt pushed a commit to kitaekatt/vllm that referenced this pull request Dec 1, 2025
Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm v1

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants