Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion vllm/platforms/rocm.py
Original file line number Diff line number Diff line change
Expand Up @@ -110,7 +110,8 @@ def use_rocm_custom_paged_attention(qtype: torch.dtype, head_size: int,

# rocm custom page attention not support on navi (gfx1*)
return (ON_MI250_MI300 and not ON_NAVI
and (sliding_window == 0 or sliding_window == (-1, -1))
and (not envs.VLLM_USE_V1 or sliding_window == 0
or sliding_window == (-1, -1))
Comment on lines +115 to +116
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do you mind adding a comment here? Something like: "custom paged attn always supported on V0, only [with(out)...] sliding window on V1" - where the [...] is a short description of why the sliding window checks are there

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Done

and (qtype == torch.half or qtype == torch.bfloat16)
and (head_size == 64 or head_size == 128)
and (block_size == 16 or block_size == 32)
Expand Down