-
-
Notifications
You must be signed in to change notification settings - Fork 6.7k
fix(ovhcloud): Edit models capabilities in model_prices_and_context_window.json #22905
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -27224,208 +27224,6 @@ | |
| "max_tokens": 128000, | ||
| "mode": "chat" | ||
| }, | ||
| "ovhcloud/DeepSeek-R1-Distill-Llama-70B": { | ||
| "input_cost_per_token": 6.7e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 131000, | ||
| "max_output_tokens": 131000, | ||
| "max_tokens": 131000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 6.7e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/deepseek-r1-distill-llama-70b", | ||
| "supports_function_calling": true, | ||
| "supports_reasoning": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Llama-3.1-8B-Instruct": { | ||
| "input_cost_per_token": 1e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 131000, | ||
| "max_output_tokens": 131000, | ||
| "max_tokens": 131000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 1e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/llama-3-1-8b-instruct", | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Meta-Llama-3_1-70B-Instruct": { | ||
| "input_cost_per_token": 6.7e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 131000, | ||
| "max_output_tokens": 131000, | ||
| "max_tokens": 131000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 6.7e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/meta-llama-3-1-70b-instruct", | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": false, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "ovhcloud/Meta-Llama-3_3-70B-Instruct": { | ||
| "input_cost_per_token": 6.7e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 131000, | ||
| "max_output_tokens": 131000, | ||
| "max_tokens": 131000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 6.7e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/meta-llama-3-3-70b-instruct", | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Mistral-7B-Instruct-v0.3": { | ||
| "input_cost_per_token": 1e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 127000, | ||
| "max_output_tokens": 127000, | ||
| "max_tokens": 127000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 1e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/mistral-7b-instruct-v0-3", | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Mistral-Nemo-Instruct-2407": { | ||
| "input_cost_per_token": 1.3e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 118000, | ||
| "max_output_tokens": 118000, | ||
| "max_tokens": 118000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 1.3e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/mistral-nemo-instruct-2407", | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Mistral-Small-3.2-24B-Instruct-2506": { | ||
| "input_cost_per_token": 9e-08, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 128000, | ||
| "max_output_tokens": 128000, | ||
| "max_tokens": 128000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 2.8e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/mistral-small-3-2-24b-instruct-2506", | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
| "supports_vision": true | ||
| }, | ||
| "ovhcloud/Mixtral-8x7B-Instruct-v0.1": { | ||
| "input_cost_per_token": 6.3e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 32000, | ||
| "max_output_tokens": 32000, | ||
| "max_tokens": 32000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 6.3e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/mixtral-8x7b-instruct-v0-1", | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "ovhcloud/Qwen2.5-Coder-32B-Instruct": { | ||
| "input_cost_per_token": 8.7e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 32000, | ||
| "max_output_tokens": 32000, | ||
| "max_tokens": 32000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 8.7e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/qwen2-5-coder-32b-instruct", | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "ovhcloud/Qwen2.5-VL-72B-Instruct": { | ||
| "input_cost_per_token": 9.1e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 32000, | ||
| "max_output_tokens": 32000, | ||
| "max_tokens": 32000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 9.1e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/qwen2-5-vl-72b-instruct", | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false, | ||
| "supports_vision": true | ||
| }, | ||
| "ovhcloud/Qwen3-32B": { | ||
| "input_cost_per_token": 8e-08, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 32000, | ||
| "max_output_tokens": 32000, | ||
| "max_tokens": 32000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 2.3e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/qwen3-32b", | ||
| "supports_function_calling": true, | ||
| "supports_reasoning": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/gpt-oss-120b": { | ||
| "input_cost_per_token": 8e-08, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 131000, | ||
| "max_output_tokens": 131000, | ||
| "max_tokens": 131000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 4e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/gpt-oss-120b", | ||
| "supports_function_calling": false, | ||
| "supports_reasoning": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "ovhcloud/gpt-oss-20b": { | ||
| "input_cost_per_token": 4e-08, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 131000, | ||
| "max_output_tokens": 131000, | ||
| "max_tokens": 131000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 1.5e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/gpt-oss-20b", | ||
| "supports_function_calling": false, | ||
| "supports_reasoning": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "ovhcloud/llava-v1.6-mistral-7b-hf": { | ||
| "input_cost_per_token": 2.9e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 32000, | ||
| "max_output_tokens": 32000, | ||
| "max_tokens": 32000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 2.9e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/llava-next-mistral-7b", | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false, | ||
| "supports_vision": true | ||
| }, | ||
| "ovhcloud/mamba-codestral-7B-v0.1": { | ||
| "input_cost_per_token": 1.9e-07, | ||
| "litellm_provider": "ovhcloud", | ||
| "max_input_tokens": 256000, | ||
| "max_output_tokens": 256000, | ||
| "max_tokens": 256000, | ||
| "mode": "chat", | ||
| "output_cost_per_token": 1.9e-07, | ||
| "source": "https://endpoints.ai.cloud.ovh.net/models/mamba-codestral-7b-v0-1", | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "palm/chat-bison": { | ||
| "input_cost_per_token": 1.25e-07, | ||
| "litellm_provider": "palm", | ||
|
|
@@ -39203,5 +39001,170 @@ | |
| "supports_reasoning": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Qwen3Guard-Gen-8B": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 32768, | ||
| "max_input_tokens": 32768, | ||
| "max_output_tokens": 32768 | ||
| }, | ||
| "ovhcloud/Qwen3Guard-Gen-0.6B": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 32768, | ||
| "max_input_tokens": 32768, | ||
| "max_output_tokens": 32768 | ||
| }, | ||
eliasto marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| "ovhcloud/Meta-Llama-3_3-70B-Instruct": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 131072, | ||
| "max_input_tokens": 131072, | ||
| "max_output_tokens": 131072, | ||
| "input_cost_per_token": 7.4e-07, | ||
| "output_cost_per_token": 7.4e-07, | ||
| "supports_function_calling": true, | ||
|
Comment on lines
+39016
to
+39027
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
Every previously-existing OVHcloud entry had a Please restore the "source": "https://endpoints.ai.cloud.ovh.net/models/meta-llama-3-3-70b-instruct"This pattern applies to all twelve re-declared models (
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. No it's not necessary |
||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Llama-3.1-8B-Instruct": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 131072, | ||
| "max_input_tokens": 131072, | ||
| "max_output_tokens": 131072, | ||
| "input_cost_per_token": 1.1e-07, | ||
| "output_cost_per_token": 1.1e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Qwen2.5-VL-72B-Instruct": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 32768, | ||
| "max_input_tokens": 32768, | ||
| "max_output_tokens": 32768, | ||
| "input_cost_per_token": 1.01e-06, | ||
| "output_cost_per_token": 1.01e-06, | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false, | ||
| "supports_vision": true | ||
| }, | ||
| "ovhcloud/Qwen3-Coder-30B-A3B-Instruct": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 262144, | ||
| "max_input_tokens": 262144, | ||
| "max_output_tokens": 262144, | ||
| "input_cost_per_token": 7e-08, | ||
| "output_cost_per_token": 2.6e-07, | ||
| "supports_function_calling": true, | ||
eliasto marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
| "supports_reasoning": false | ||
| }, | ||
| "ovhcloud/Mistral-Small-3.2-24B-Instruct-2506": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 131072, | ||
| "max_input_tokens": 131072, | ||
| "max_output_tokens": 131072, | ||
| "input_cost_per_token": 1e-07, | ||
| "output_cost_per_token": 3.1e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
| "supports_vision": true | ||
| }, | ||
| "ovhcloud/Mistral-Nemo-Instruct-2407": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
|
Comment on lines
+39082
to
+39084
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Context window reduced for Two models have had their context windows significantly reduced compared to the previous entries:
Users who currently pass prompts between 65 K and 118 K / 127 K tokens to these models will start receiving context-limit errors after this change is deployed. Per the repo's guideline against backwards-incompatible changes without user-controlled flags, this should be documented in the PR description if it reflects a real OVHcloud endpoint limit change, or reverted if it does not. Context Used: Rule from Why: This breaks current ... (source)
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Yes it's normal |
||
| "max_tokens": 65536, | ||
| "max_input_tokens": 65536, | ||
| "max_output_tokens": 65536, | ||
| "input_cost_per_token": 1.4e-07, | ||
| "output_cost_per_token": 1.4e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/Mixtral-8x7B-Instruct-v0.1": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 32768, | ||
| "max_input_tokens": 32768, | ||
| "max_output_tokens": 32768, | ||
| "input_cost_per_token": 7e-07, | ||
| "output_cost_per_token": 7e-07, | ||
| "supports_function_calling": false, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": false | ||
| }, | ||
| "ovhcloud/Qwen3-32B": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 32768, | ||
| "max_input_tokens": 32768, | ||
| "max_output_tokens": 32768, | ||
| "input_cost_per_token": 9e-08, | ||
| "output_cost_per_token": 2.5e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
| "supports_reasoning": true | ||
| }, | ||
| "ovhcloud/DeepSeek-R1-Distill-Llama-70B": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 131072, | ||
| "max_input_tokens": 131072, | ||
| "max_output_tokens": 131072, | ||
| "input_cost_per_token": 7.4e-07, | ||
| "output_cost_per_token": 7.4e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
| "supports_reasoning": true | ||
| }, | ||
| "ovhcloud/gpt-oss-20b": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 131072, | ||
| "max_input_tokens": 131072, | ||
| "max_output_tokens": 131072, | ||
| "input_cost_per_token": 5e-08, | ||
| "output_cost_per_token": 1.8e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
eliasto marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| "supports_reasoning": true | ||
| }, | ||
| "ovhcloud/Mistral-7B-Instruct-v0.3": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 65536, | ||
| "max_input_tokens": 65536, | ||
| "max_output_tokens": 65536, | ||
| "input_cost_per_token": 1.1e-07, | ||
| "output_cost_per_token": 1.1e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true | ||
| }, | ||
| "ovhcloud/gpt-oss-120b": { | ||
| "litellm_provider": "ovhcloud", | ||
| "mode": "chat", | ||
| "max_tokens": 131072, | ||
| "max_input_tokens": 131072, | ||
| "max_output_tokens": 131072, | ||
| "input_cost_per_token": 9e-08, | ||
| "output_cost_per_token": 4.7e-07, | ||
| "supports_function_calling": true, | ||
| "supports_response_schema": true, | ||
| "supports_tool_choice": true, | ||
| "supports_reasoning": true | ||
| } | ||
eliasto marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| } | ||
| } | ||
eliasto marked this conversation as resolved.
Show resolved
Hide resolved
|
||
Uh oh!
There was an error while loading. Please reload this page.