A proxy server that transforms Anthropic API requests to OpenAI format and sends it to openrouter.ai. This enables you to use Anthropic's API format while connecting to OpenAI-compatible endpoints.
With this command, you can start the proxy server with your OpenRouter API key on port 3000:
OPENROUTER_API_KEY=your-api-key npx anthropic-proxyEnvironment variables:
OPENROUTER_API_KEY: Your OpenRouter API key (required when using OpenRouter)ANTHROPIC_PROXY_BASE_URL: Custom base URL for the transformed OpenAI-format message (default:openrouter.ai)PORT: The port the proxy server should listen on (default: 3000)REASONING_MODEL: The reasoning model to use (default:google/gemini-2.0-pro-exp-02-05:free)COMPLETION_MODEL: The completion model to use (default:google/gemini-2.0-pro-exp-02-05:free)DEBUG: Set to1to enable debug logging
Note: When ANTHROPIC_PROXY_BASE_URL is set to a custom URL, the OPENROUTER_API_KEY is not required.
To use the proxy server as a backend for Claude Code, you have to set the ANTHROPIC_BASE_URL to the URL of the proxy server:
ANTHROPIC_BASE_URL=http://0.0.0.0:3000 claudeLicensed under MIT license. Copyright (c) 2025 Max Nowack
Contributions are welcome. Please open issues and/or file Pull Requests.