Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to deploy a quantized FP8 multimodel llm ? #681

Open
zhishao opened this issue Jan 22, 2025 · 1 comment
Open

How to deploy a quantized FP8 multimodel llm ? #681

zhishao opened this issue Jan 22, 2025 · 1 comment

Comments

@zhishao
Copy link

zhishao commented Jan 22, 2025

Do both the visual encoder and the language model need to be quantized to fp8?
How to modify the config.pbtxt?

@zhishao
Copy link
Author

zhishao commented Jan 23, 2025

When I launch an Internvl2 multimodal model, it can perform inference normally. However, after I quantize its language model component, it loads successfully but encounters errors during inference :

[TensorRT-LLM][ERROR] Request embedding table data type doesn't match model weight data type.
[TensorRT-LLM][ERROR] Encountered an error in forwardAsync function: Request embedding table data type doesn't match model weight data type.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant