Skip to content

[W8A8 Block Linear Refactor][1/N] Keep all quantization types into QuantFP8 class.#33047

Merged
DarkLight1337 merged 21 commits intovllm-project:mainfrom
EmbeddedLLM:rfc-quant-fp8
Feb 1, 2026
Merged

[W8A8 Block Linear Refactor][1/N] Keep all quantization types into QuantFP8 class.#33047
DarkLight1337 merged 21 commits intovllm-project:mainfrom
EmbeddedLLM:rfc-quant-fp8

Commits

Commits on Jan 22, 2026

Commits on Jan 26, 2026

Commits on Jan 27, 2026

Commits on Jan 29, 2026

Commits on Jan 30, 2026

Commits on Jan 31, 2026

Commits on Feb 1, 2026