Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update tch to 0.16+ #1765

Open
syl20bnr opened this issue May 13, 2024 · 6 comments
Open

Update tch to 0.16+ #1765

syl20bnr opened this issue May 13, 2024 · 6 comments
Assignees
Labels
bug Something isn't working dependencies Pull requests that update a dependency file

Comments

@syl20bnr
Copy link
Member

Update tch once the upstream fix is released in Pytorch 2.3.1 and tch is updated.

See compilation bug issue: LaurentMazare/tch-rs#870

@syl20bnr syl20bnr added bug Something isn't working dependencies Pull requests that update a dependency file labels May 13, 2024
@antimora
Copy link
Collaborator

The upstream issue in PyTorch seems to be fixed (pytorch/pytorch#124009).

@antimora
Copy link
Collaborator

@syl20bnr can we try updating it again and see if this is fixed? We should try it before the upcoming release.

@syl20bnr
Copy link
Member Author

Yep, I'll look into it next week,

@syl20bnr
Copy link
Member Author

Looking at it while I am refactoring our CI.

@syl20bnr syl20bnr self-assigned this Jul 29, 2024
@syl20bnr
Copy link
Member Author

The issue is still happening but with another DLL:

INTEL MKL ERROR: The specified module could not be found. mkl_vml_def.1.dll.
Intel MKL FATAL ERROR: cannot load mkl_vml_def.1.dll.

@oleid
Copy link

oleid commented Aug 11, 2024

It would appear tch 0.17 is now available. It features libtorch-2.4

FWIW: This is probably not related, I'm trying to get burn working using libtorch on my Radeon GPU. pyTorch-2.3 and 2.4 work fine, yet burn appers not to work when changing tch from 0.15 to 0.17.

Running benches/custom_gelu.rs (target/benchmarks/release/deps/custom_gelu-82b2276b553d5723)
thread 'main' panicked at /home/oleid/.cargo/registry/src/index.crates.io-6f17d22bba15001f/tch-0.17.0/src/wrappers/tensor_generated.rs:8361:40:
called `Result::unwrap()` on an `Err` value: Torch("Could not run 'aten::empty.memory_format' with arguments from the 'CUDA' backend. This could be because the operator doesn't exist for this backend, or was omitted during the selective/custom build process (if using custom build). If you are a Facebook employee using PyTorch on mobile, please visit https://fburl.com/ptmfixes for possible resolutions. 'aten::empty.memory_format' is only available for these backends: [CPU, Meta, QuantizedCPU, QuantizedMeta, MkldnnCPU, SparseCPU, SparseMeta, SparseCsrCPU, SparseCsrMeta, BackendSelect, Python, FuncTorchDynamicLayerBackMode, Functionalize, Named, Conjugate, Negative, ZeroTensor, ADInplaceOrView, AutogradOther, AutogradCPU, AutogradCUDA, AutogradHIP, AutogradXLA, AutogradMPS, AutogradIPU, AutogradXPU, AutogradHPU, AutogradVE, AutogradLazy, AutogradMTIA, AutogradPrivateUse1, AutogradPrivateUse2, AutogradPrivateUse3, AutogradMeta, AutogradNestedTensor, Tracer, AutocastCPU, AutocastXPU, AutocastCUDA, FuncTorchBatched, BatchedNestedTensor, FuncTorchVmapMode, Batched, VmapMode, FuncTorchGradWrapper, PythonTLSSnapshot, FuncTorchDynamicLayerFrontMode, PreDispatch, PythonDispatcher].\n
[...]

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working dependencies Pull requests that update a dependency file
Projects
None yet
Development

No branches or pull requests

3 participants