--- summary: "Model providers (LLMs) supported by OpenClaw" read_when: - You want to choose a model provider - You want quick setup examples for LLM auth + model selection title: "Model Provider Quickstart" --- # Model Providers OpenClaw can use many LLM providers. Pick one, authenticate, then set the default model as `provider/model`. ## Quick start (two steps) 1. Authenticate with the provider (usually via `openclaw onboard`). 2. Set the default model: ```json5 { agents: { defaults: { model: { primary: "anthropic/claude-opus-4-6" } } }, } ``` ## Supported providers (starter set) - [Anthropic (API + Claude Code CLI)](/providers/anthropic) - [Amazon Bedrock](/providers/bedrock) - [BytePlus (International)](/concepts/model-providers#byteplus-international) - [Chutes](/providers/chutes) - [Cloudflare AI Gateway](/providers/cloudflare-ai-gateway) - [GLM models](/providers/glm) - [MiniMax](/providers/minimax) - [Mistral](/providers/mistral) - [Moonshot AI (Kimi + Kimi Coding)](/providers/moonshot) - [OpenAI (API + Codex)](/providers/openai) - [OpenCode (Zen + Go)](/providers/opencode) - [OpenRouter](/providers/openrouter) - [Qianfan](/providers/qianfan) - [Qwen](/providers/qwen) - [StepFun](/providers/stepfun) - [Synthetic](/providers/synthetic) - [Vercel AI Gateway](/providers/vercel-ai-gateway) - [Venice (Venice AI)](/providers/venice) - [xAI](/providers/xai) - [Z.AI](/providers/zai) ## Additional bundled provider variants - `anthropic-vertex` - implicit Anthropic on Google Vertex support when Vertex credentials are available; no separate onboarding auth choice - `copilot-proxy` - local VS Code Copilot Proxy bridge; use `openclaw onboard --auth-choice copilot-proxy` - `google-gemini-cli` - unofficial Gemini CLI OAuth flow; use `openclaw onboard --auth-choice google-gemini-cli` or `openclaw models auth login --provider google-gemini-cli --set-default` For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, see [Model providers](/concepts/model-providers).