Skip to content

[Bugfix] Fix MLA attention crash with AWQ/GPTQ quantized models#34695

Merged
MatthewBonanni merged 3 commits intovllm-project:mainfrom
haosdent:fix-34561
Mar 13, 2026
Merged

[Bugfix] Fix MLA attention crash with AWQ/GPTQ quantized models#34695
MatthewBonanni merged 3 commits intovllm-project:mainfrom
haosdent:fix-34561

Commits

Commits on Feb 17, 2026

Commits on Mar 6, 2026

Commits on Mar 13, 2026