Skip to content

Commit 62de2ea

Browse files
committed
fix num_heads_q
Signed-off-by: yuanyongjie.yyj <[email protected]> Signed-off-by: FENP <[email protected]>
1 parent 66bad29 commit 62de2ea

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/v1/attention/backends/flash_attn.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -317,7 +317,7 @@ def schedule(
317317
batch_size=batch_size,
318318
max_seqlen_q=max_query_len,
319319
max_seqlen_k=max_seq_len,
320-
num_heads_q=self.num_heads_q,
320+
num_heads_q=self.num_heads_q * self.dcp_world_size,
321321
num_heads_kv=self.num_heads_kv,
322322
headdim=self.headdim,
323323
cache_seqlens=seqlens,

0 commit comments

Comments
 (0)