Skip to content

[NVIDIA] Support Flashinfer TRTLLM FP8-q/kv/out Attention Kernel#21716

Merged
ProExpertProg merged 13 commits intovllm-project:mainfrom
elvischenv:elvischenv/fp8-trtllm-attn
Aug 19, 2025
Merged

[NVIDIA] Support Flashinfer TRTLLM FP8-q/kv/out Attention Kernel#21716
ProExpertProg merged 13 commits intovllm-project:mainfrom
elvischenv:elvischenv/fp8-trtllm-attn

Commits

Commits on Aug 18, 2025

Commits on Aug 19, 2025