-
Notifications
You must be signed in to change notification settings - Fork 786
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue with converting Whisper model to ONNX #1040
Comments
Thanks @AvivSham, I am able to reproduce the issue. Same thing happens with other variants of whisper. @echarlaix looks to be an issue with Optimum as I'm able to reproduce with |
bumping... |
@AvivSham In the meantime, can you try downgrade to |
We downgraded transformers to 4.38.2, and still none of the model versions (small/medium/large) meet the threshold. The trace looks more of the same for all versions:
However we see 2 differences:
|
@AvivSham Those differences are negligible and the model will produce similar results to the python version! Looks like we need to investigate what broke in a recent update to transformers. cc @echarlaix |
@echarlaix @xenova thanks! Do you know why the logs are less informative? I can recall from optimum that if layers meet the threshold they are also printed with a checkmark sign, here the logs are only for single layer. I also have a follow-up question which I also asked here - #917 (comment) |
System Info
Created new env following this requirement file:
system info:
MAC M2
converting using CPU device
Environment/Platform
Description
We are attempting to convert
whisper-small
using the HF modelopenai/whisper-small
by executing the command specified in the README file.python -m scripts.convert --quantize --model_id openai/whisper-small
We get the following trace:
None of the layers meet the default tolerance and in most layers, the difference is more than 3 orders of magnitude.
@xenova can you please help with this?
Thanks,
Reproduction
just run:
python -m scripts.convert --quantize --model_id openai/whisper-small
The text was updated successfully, but these errors were encountered: