Can you run Langflow, (embeddings and LLM's) on CUDA? #318
Replies: 1 comment
-
The process depends on the embeddings. If the user is utilizing Hugging Face embeddings for inference, they can run the inference engine on a CUDA-based server. However, it is completely independent of Langflow. We simply utilize the base URLs or API URLs of the deployed services. To my knowledge, Langflow does not have any resources that directly use CUDA. We have transitioned most services to use API examples, like Ollama and Hugging Face inference. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Can you run Langflow, (embeddings and LLM's) on CUDA? If so, where can I find instructions to run this on Docker?
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions