Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

QUESTION: Hello, your code is well-written. Can I leverage the TensorRT Actor in a complimentary manner? #670

Closed
glide-the opened this issue Nov 21, 2023 · 3 comments
Labels
question Further information is requested
Milestone

Comments

@glide-the
Copy link

Hello, your code is well-crafted. It utilizes the xoscar plugin to achieve distributed scheduling. From an architectural standpoint, it employs only the actor_ref, which is more elegant compared to the HTTP interaction. You can refer to the specific code here: https://github.com/xorbitsai/inference/blob/6135eb66f1595d41a7210f9f64c3db97adf0364b/xinference/client/oscar/actor_client.py#L432C14-L432C14

Regarding the features of the RPC framework, approximately 40% of the code in xinference is dedicated to handling basic interactions: https://github.com/xorbitsai/inference/blob/main/xinference/core/supervisor.py and https://github.com/xorbitsai/inference/blob/main/xinference/core/worker.py

Hence, this framework seems destined to be unable to resolve conflicts that may arise when multiple LLMs collaborate: https://github.com/xorbitsai/inference/blob/main/xinference/model/core.py#L32

From the code, it appears that they are planning to place the TensorRT LLM in another project: https://github.com/xorbitsai/inference/blob/main/xinference/model/llm/core.py#L31

@glide-the glide-the added the question Further information is requested label Nov 21, 2023
@XprobeBot XprobeBot added this to the v0.6.4 milestone Nov 21, 2023
@glide-the glide-the changed the title QUESTION Hello, your code is well-written. Can I leverage the TensorRT Actor in a complimentary manner? QUESTION: Hello, your code is well-written. Can I leverage the TensorRT Actor in a complimentary manner? Nov 21, 2023
@aresnow1
Copy link
Contributor

Hence, this framework seems destined to be unable to resolve conflicts that may arise when multiple LLMs collaborate: https://github.com/xorbitsai/inference/blob/main/xinference/model/core.py#L32

It's OK to launch multiple models and won't be a problem.

From the code, it appears that they are planning to place the TensorRT LLM in another project: https://github.com/xorbitsai/inference/blob/main/xinference/model/llm/core.py#L31

What's this mean? Using TensorRT-LLM for inference is certainly possible and there's a draft PR #646.

@glide-the
Copy link
Author

Thank you for your reply

i just feel different llm in “family wrapper ” ,Code dependencies and collisions occurring when different modules are running at the same time

@aresnow1
Copy link
Contributor

Feel free to reopen this issue if have any questions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants