-
Notifications
You must be signed in to change notification settings - Fork 323
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Custom openai compatible endpoint #1290
Comments
@riyajatar37003 - You could for example use the openai provider and then point the |
thanks could you share any doc link. where in the .toml i need to set this |
myconfig.toml [completion]
provider = "litellm"
concurrent_request_limit = 16
[completion.generation_config]
model = "openai/llama3.2" #add your model name here
temperature = 0.1
top_p = 1
max_tokens_to_sample = 1_024
stream = true
add_generation_kwargs = { } then you do |
Let's say I have a openai-like endpoint served locally under "http://localhost:8004" + it's called "custom-model". It is in line with OpenAI V1 API. How to connect it to r2r? |
I believe the correct environment variable is |
Also if you are using LiteLLM Proxy with R2R, then since it internally uses LiteLLM SDK, the name of the model in the r2r.toml config file should include so if for instance in Proxy you have a model named Assuming the name of your model in LiteLLM Proxy is [completion]
provider = "litellm"
concurrent_request_limit = 64
[completion.generation_config]
model = "openai/openai/llama3.3" Assuming your LiteLLM Proxy config looks like this: proxy_config:
litellm_settings:
drop_params: True
model_list:
# At least one model must exist for the proxy to start.
- model_name: "openai/llama3.3"
litellm_params:
model: "openai/llama3.3"
api_key: fake-key
api_base: "http://ollama.mywebsite.com:11434" Assuming you have an ollama app running on port It might be confusing, but in the r2r.toml when you see And the name prefix |
Hi,
I have custom llm and embedding deployment using triton server and also a wrapper around it which is openai compatible.
how can i use this in .toml config file.
I have tested it with litellm proxy server and its working.
The text was updated successfully, but these errors were encountered: