Skip to content

[Bugfix] Enable attn quantization of Llama-4 by correctly permuting scales for rope (int8, fp8)#34243

Merged
mgoin merged 3 commits intovllm-project:mainfrom
eldarkurtic:fix-attn-quant-llama4
Feb 11, 2026
Merged

[Bugfix] Enable attn quantization of Llama-4 by correctly permuting scales for rope (int8, fp8)#34243
mgoin merged 3 commits intovllm-project:mainfrom
eldarkurtic:fix-attn-quant-llama4

Commits

Commits on Feb 10, 2026