We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4b299cb commit 17e0d0fCopy full SHA for 17e0d0f
tensorrt_llm/_torch/attention_backend/trtllm.py
@@ -634,7 +634,7 @@ def __post_init__(self) -> None:
634
self.block_ids_per_seq = None
635
self.kv_block_ids_per_seq = None
636
if self.enable_flash_mla:
637
- self.block_ids_per_seq = torch.empty(
+ self.block_ids_per_seq = torch.zeros(
638
[
639
self.kv_cache_manager.max_batch_size,
640
self.kv_cache_manager.max_blocks_per_seq
0 commit comments