Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support loading ONNX models directly #141

Open
robertknight opened this issue May 4, 2024 · 1 comment
Open

Support loading ONNX models directly #141

robertknight opened this issue May 4, 2024 · 1 comment

Comments

@robertknight
Copy link
Owner

robertknight commented May 4, 2024

ONNX models currently have to be converted into the FlatBuffers-based .rten format to use them.

The .rten format is intended to support efficient loading, and have a small code footprint (it serves a similar role to ORT). However the need to convert models is a barrier to using this library, and inconvenient in projects that want to trial or mix different runtimes for various reasons. It would reduce friction if .onnx models could be loaded directly.

@robertknight
Copy link
Owner Author

Had a look at Rust protocol buffers runtimes:

  • Prost has the widest adoption, but it has downsides:
  • quick-protobuf has less adoption and hasn't been released in a while, but seems much more aligned with this project's priorities:
    • Minimizes allocations when deserializing
    • Does not require protoc installed

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant