Skip to content

Commit dd36f79

Browse files
committed
fix for navi
Signed-off-by: fsx950223 <[email protected]>
1 parent dc5bd7c commit dd36f79

File tree

1 file changed

+5
-3
lines changed

1 file changed

+5
-3
lines changed

vllm/attention/ops/chunked_prefill_paged_decode.py

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,7 @@
99

1010
import torch
1111

12+
from vllm import _custom_ops as ops
1213
from vllm.platforms import current_platform
1314
from vllm.platforms.rocm import use_rocm_custom_paged_attention
1415
from vllm.triton_utils import tl, triton
@@ -304,7 +305,7 @@ def chunked_prefill_paged_decode(
304305
)
305306
max_logits = torch.empty_like(exp_sums)
306307

307-
torch.ops.aiter.paged_attention_rocm(
308+
ops.paged_attention_rocm(
308309
output,
309310
exp_sums,
310311
max_logits,
@@ -315,9 +316,10 @@ def chunked_prefill_paged_decode(
315316
num_kv_heads,
316317
scale=sm_scale,
317318
block_tables=block_table,
318-
context_lens=seq_lens,
319+
seq_lens=seq_lens,
320+
query_start_loc=query_start_loc,
319321
block_size=block_size,
320-
max_context_len=max_seq_len,
322+
max_seq_len=max_seq_len,
321323
alibi_slopes=alibi_slopes,
322324
kv_cache_dtype=kv_cache_dtype,
323325
k_scale=k_scale,

0 commit comments

Comments
 (0)