diff --git a/docs/source/use-litellm-as-backend.mdx b/docs/source/use-litellm-as-backend.mdx index 83f744388..257da5d32 100644 --- a/docs/source/use-litellm-as-backend.mdx +++ b/docs/source/use-litellm-as-backend.mdx @@ -63,8 +63,8 @@ vllm serve HuggingFaceH4/zephyr-7b-beta --host 0.0.0.0 --port 8000 2. Configure LiteLLM to use the local server: ```yaml model_parameters: - provider: "openai" - model_name: "HuggingFaceH4/zephyr-7b-beta" + provider: "hosted_vllm" + model_name: "hosted_vllm/HuggingFaceH4/zephyr-7b-beta" base_url: "http://localhost:8000/v1" api_key: "" ```