mirror of https://github.com/openclaw/openclaw.git
docs(providers): refresh model examples and env defaults
This commit is contained in:
parent
0609bf8581
commit
da3f5e9bca
|
|
@ -945,7 +945,7 @@ Time format in system prompt. Default: `auto` (OS preference).
|
|||
},
|
||||
pdfModel: {
|
||||
primary: "anthropic/claude-opus-4-6",
|
||||
fallbacks: ["openai/gpt-5-mini"],
|
||||
fallbacks: ["openai/gpt-5.4-mini"],
|
||||
},
|
||||
params: { cacheRetention: "long" }, // global default provider params
|
||||
pdfMaxBytesMb: 10,
|
||||
|
|
@ -994,7 +994,8 @@ Time format in system prompt. Default: `auto` (OS preference).
|
|||
| `opus` | `anthropic/claude-opus-4-6` |
|
||||
| `sonnet` | `anthropic/claude-sonnet-4-6` |
|
||||
| `gpt` | `openai/gpt-5.4` |
|
||||
| `gpt-mini` | `openai/gpt-5-mini` |
|
||||
| `gpt-mini` | `openai/gpt-5.4-mini` |
|
||||
| `gpt-nano` | `openai/gpt-5.4-nano` |
|
||||
| `gemini` | `google/gemini-3.1-pro-preview` |
|
||||
| `gemini-flash` | `google/gemini-3-flash-preview` |
|
||||
| `gemini-flash-lite` | `google/gemini-3.1-flash-lite-preview` |
|
||||
|
|
|
|||
|
|
@ -2247,7 +2247,8 @@ for usage/billing and raise limits as needed.
|
|||
- `opus` → `anthropic/claude-opus-4-6`
|
||||
- `sonnet` → `anthropic/claude-sonnet-4-6`
|
||||
- `gpt` → `openai/gpt-5.4`
|
||||
- `gpt-mini` → `openai/gpt-5-mini`
|
||||
- `gpt-mini` → `openai/gpt-5.4-mini`
|
||||
- `gpt-nano` → `openai/gpt-5.4-nano`
|
||||
- `gemini` → `google/gemini-3.1-pro-preview`
|
||||
- `gemini-flash` → `google/gemini-3-flash-preview`
|
||||
- `gemini-flash-lite` → `google/gemini-3.1-flash-lite-preview`
|
||||
|
|
|
|||
|
|
@ -60,7 +60,7 @@ Any model available on the gateway can be used with the `kilocode/` prefix:
|
|||
```
|
||||
kilocode/kilo/auto (default - smart routing)
|
||||
kilocode/anthropic/claude-sonnet-4
|
||||
kilocode/openai/gpt-5.2
|
||||
kilocode/openai/gpt-5.4
|
||||
kilocode/google/gemini-3-pro-preview
|
||||
...and many more
|
||||
```
|
||||
|
|
|
|||
|
|
@ -47,7 +47,7 @@ openclaw onboard --opencode-go-api-key "$OPENCODE_API_KEY"
|
|||
### Zen
|
||||
|
||||
- Runtime provider: `opencode`
|
||||
- Example models: `opencode/claude-opus-4-6`, `opencode/gpt-5.2`, `opencode/gemini-3-pro`
|
||||
- Example models: `opencode/claude-opus-4-6`, `opencode/gpt-5.4`, `opencode/gemini-3-pro`
|
||||
- Best when you want the curated OpenCode multi-model proxy
|
||||
|
||||
### Go
|
||||
|
|
|
|||
|
|
@ -90,7 +90,7 @@ Fallback details:
|
|||
defaults: {
|
||||
pdfModel: {
|
||||
primary: "anthropic/claude-opus-4-6",
|
||||
fallbacks: ["openai/gpt-5-mini"],
|
||||
fallbacks: ["openai/gpt-5.4-mini"],
|
||||
},
|
||||
pdfMaxBytesMb: 10,
|
||||
pdfMaxPages: 20,
|
||||
|
|
@ -150,7 +150,7 @@ Page-filtered fallback model:
|
|||
{
|
||||
"pdf": "https://example.com/report.pdf",
|
||||
"pages": "1-3,7",
|
||||
"model": "openai/gpt-5-mini",
|
||||
"model": "openai/gpt-5.4-mini",
|
||||
"prompt": "Extract only customer-impacting incidents"
|
||||
}
|
||||
```
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ describe("createOpenClawTools PDF registration", () => {
|
|||
const cfg: OpenClawConfig = {
|
||||
agents: {
|
||||
defaults: {
|
||||
pdfModel: { primary: "openai/gpt-5-mini" },
|
||||
pdfModel: { primary: "openai/gpt-5.4-mini" },
|
||||
},
|
||||
},
|
||||
};
|
||||
|
|
|
|||
|
|
@ -983,7 +983,7 @@ describe("applyMediaUnderstanding", () => {
|
|||
const cfg: OpenClawConfig = {
|
||||
tools: {
|
||||
media: {
|
||||
image: { enabled: true, models: [{ provider: "openai", model: "gpt-5.2" }] },
|
||||
image: { enabled: true, models: [{ provider: "openai", model: "gpt-5.4" }] },
|
||||
audio: { enabled: true, models: [{ provider: "groq" }] },
|
||||
video: { enabled: true, models: [{ provider: "google", model: "gemini-3" }] },
|
||||
},
|
||||
|
|
|
|||
|
|
@ -13,7 +13,7 @@ describe("resolveModelEntries", () => {
|
|||
const cfg: OpenClawConfig = {
|
||||
tools: {
|
||||
media: {
|
||||
models: [{ provider: "openai", model: "gpt-5.2" }],
|
||||
models: [{ provider: "openai", model: "gpt-5.4" }],
|
||||
},
|
||||
},
|
||||
};
|
||||
|
|
@ -38,7 +38,7 @@ describe("resolveModelEntries", () => {
|
|||
tools: {
|
||||
media: {
|
||||
image: {
|
||||
models: [{ provider: "openai", model: "gpt-5.2" }],
|
||||
models: [{ provider: "openai", model: "gpt-5.4" }],
|
||||
},
|
||||
},
|
||||
},
|
||||
|
|
|
|||
|
|
@ -61,5 +61,6 @@ describe("provider env vars", () => {
|
|||
expect(getProviderEnvVars("__proto__")).toEqual([]);
|
||||
expect(getProviderEnvVars("constructor")).toEqual([]);
|
||||
expect(getProviderEnvVars("openai")).toEqual(["OPENAI_API_KEY"]);
|
||||
expect(getProviderEnvVars("anthropic")).toEqual(["ANTHROPIC_OAUTH_TOKEN", "ANTHROPIC_API_KEY"]);
|
||||
});
|
||||
});
|
||||
|
|
|
|||
|
|
@ -8,7 +8,6 @@ const CORE_PROVIDER_AUTH_ENV_VAR_CANDIDATES = {
|
|||
} as const;
|
||||
|
||||
const CORE_PROVIDER_SETUP_ENV_VAR_OVERRIDES = {
|
||||
anthropic: ["ANTHROPIC_API_KEY", "ANTHROPIC_OAUTH_TOKEN"],
|
||||
"minimax-cn": ["MINIMAX_API_KEY"],
|
||||
} as const;
|
||||
|
||||
|
|
|
|||
Loading…
Reference in New Issue