-
Notifications
You must be signed in to change notification settings - Fork 785
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] type object 'TasksManager' has no attribute '_TASKS_TO_AUTOMODELS' #297
Comments
Hi there 👋 this was fixed yesterday (and was caused by a new version of optimum). If you use the current conversion script (located in the |
OK. Looks like it is running now :) The other issue I see is that
zsh: killed python -m scripts.convert --quantize --model_id meta-llama/Llama-2-7b-chat-hf |
It looks like the process ran out of memory. Could you try add Just note that at the moment, we do not support models larger than 2GB (due to the protobuf limit combined with onnxruntime-node not supporting the external file format). This will most likely be fixed in the next version of onnxruntime-node (>1.16), thanks to contributions from @dakenf! 🤗 |
Thanks
I am getting the following error: |
It is possible to pass filename to onnx runtime in node and it will load anything that fits into RAM/VRAM but you'll need to tamper with hub.js in node_modules, see #123 Also, you can run it on GPU with latest changes microsoft/onnxruntime#16050 |
OK. I am getting a segmentation fault and not sure why. I guess the original issue I opened this ticket for is resolved. Not sure where to go from here on this latest issue. app.get("/", async (req, res) => { |
Will open another issue. For latest troubles thanks. |
Please see the above comment: #297 (comment) Transformers.js does not yet support models larger than 2GB. This is due to the support lacking in onnxruntime 1.14.0 |
I'm now working to bypass 4gb limit in WASM FS that is required for SDXL (and other models like LLAMA), so soon JS/TS will rule the inference as the most accessible way for developers. So you can expect much more stars and issues hehe |
Thanks @dakenf that is great news! |
Describe the bug
I am trying to run
python -m scripts.convert --quantize --model_id bert-base-uncased
and I am getting the following errorHow to reproduce
Steps or a minimal working example to reproduce the behavior
python -m scripts.convert --quantize --model_id bert-base-uncased
Expected behavior
I should get an ONNX model I can use with "@xenova/transformers"
Environment
Additional context
I have tried other models with the same error e.g. meta-llama/Llama-2-7b-chat-hf
The text was updated successfully, but these errors were encountered: