Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FLOAT8] Add Hardware Compatibility Check for FP8 Quantization #1188

Open
drisspg opened this issue Oct 29, 2024 · 1 comment
Open

[FLOAT8] Add Hardware Compatibility Check for FP8 Quantization #1188

drisspg opened this issue Oct 29, 2024 · 1 comment
Assignees
Labels
float8 good first issue Good for newcomers

Comments

@drisspg
Copy link
Contributor

drisspg commented Oct 29, 2024

Add Hardware Compatibility Check for FP8 Quantization

Issue Summary

In our current implementation, we provide three APIs for model computation in FP8 format. However, for dynamic activation quant these FP8 computations are only supported on NVIDIA GPUs with SM89 and SM90 architectures. When models are quantized to FP8 on unsupported hardware, errors only occur during runtime, which can lead to confusion and wasted resources.

Proposed Solution

Check at the model quantization stage if the target hardware does not support FP8 computations and raise an error accordingly. This way, users are informed immediately if their hardware cannot handle FP8 quantization, rather than discovering it during runtime. Potentially point to weight-only quant which as more supported

Changes where to add errors:

    "float8_dynamic_activation_float8_weight",
    "float8_static_activation_float8_weight"
@drisspg drisspg added good first issue Good for newcomers float8 labels Oct 29, 2024
@petrex
Copy link
Collaborator

petrex commented Oct 31, 2024

good idea! I will add arch check for AMD GPUs as well.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
float8 good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

3 participants