OpenRouter¶
Configure HolmesGPT to use OpenRouter for access to multiple AI models through a single API.
Methods¶
Method 1: Native LiteLLM OpenRouter Provider (Recommended)¶
The simplest approach uses LiteLLM's native OpenRouter support. Only OPENROUTER_API_KEY is required. This method is preferred because HolmesGPT can automatically determine token limits and context window sizes for each model.
export OPENROUTER_API_KEY="sk-or-..." # your OpenRouter key
holmes ask "hello" --model="openrouter/anthropic/claude-opus-4.5" --no-interactive
Optional environment variables:
OPENROUTER_API_BASE- Custom API base URL (defaults tohttps://openrouter.ai/api/v1)OR_SITE_URL- Your site URL for OpenRouter rankingsOR_APP_NAME- Your app name for OpenRouter rankings
Method 2: OpenAI-Compatible Endpoint¶
Alternatively, you can use OpenRouter's OpenAI-compatible endpoint by setting the base URL and using OPENAI_API_KEY. Note the openai/ prefix instead of openrouter/.
Token Limits
With this method, HolmesGPT cannot automatically determine token limits for the model. You may need to set OVERRIDE_MAX_CONTENT_SIZE and OVERRIDE_MAX_OUTPUT_TOKEN environment variables manually.
export OPENAI_API_BASE="https://openrouter.ai/api/v1"
export OPENAI_API_KEY="sk-or-..." # your OpenRouter key
holmes ask "hello" --model="openai/anthropic/claude-opus-4.5" --no-interactive
Available Models¶
You can use any model available on OpenRouter. The model prefix depends on which method you use:
Method 1 (Native): Use openrouter/ prefix
openrouter/anthropic/claude-opus-4.5openrouter/openai/gpt-4oopenrouter/google/gemini-pro
Method 2 (OpenAI-Compatible): Use openai/ prefix
openai/anthropic/claude-opus-4.5openai/openai/gpt-4oopenai/google/gemini-pro
See the OpenRouter models page for a complete list of available models.