[Bugfix] Enable attn quantization of Llama-4 by correctly permuting scales for rope (int8, fp8)#34243
Merged
mgoin merged 3 commits intovllm-project:mainfrom Feb 11, 2026
Merged
Commits
Commits on Feb 10, 2026
- committed
Your Name - committed
Your Name - authored