Skip to content

Add FP8 quantization ignored_layers support in llama#6592

Closed
cli99 wants to merge 2 commits intovllm-project:mainfrom
cli99:fp8-quant-ignore-layers
Closed

Add FP8 quantization ignored_layers support in llama#6592
cli99 wants to merge 2 commits intovllm-project:mainfrom
cli99:fp8-quant-ignore-layers

Commits

Commits on Jul 19, 2024