revert: restore supportsUsageInStreaming=false default for non-native endpoints

Reverts #46500. Breaks Ollama, LM Studio, TGI, LocalAI, Mistral API -
these backends reject stream_options with 400/422.

This reverts commit bb06dc7cc9.
This commit is contained in:
Andrew Demczuk 2026-03-15 00:34:04 +01:00 committed by GitHub
parent 2806f2b878
commit b202ac2ad1
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 27 additions and 30 deletions

View File

@ -20,7 +20,6 @@ Docs: https://docs.openclaw.ai
- Slack/interactive replies: preserve `channelData.slack.blocks` through live DM delivery and preview-finalized edits so Block Kit button and select directives render instead of falling back to raw text. Thanks @vincentkoc. - Slack/interactive replies: preserve `channelData.slack.blocks` through live DM delivery and preview-finalized edits so Block Kit button and select directives render instead of falling back to raw text. Thanks @vincentkoc.
- CI/channel test routing: move the built-in channel suites into `test:channels` and keep them out of `test:extensions`, so extension CI no longer fails after the channel migration while targeted test routing still sends Slack, Signal, and iMessage suites to the right lane. (#46066) Thanks @scoootscooob. - CI/channel test routing: move the built-in channel suites into `test:channels` and keep them out of `test:extensions`, so extension CI no longer fails after the channel migration while targeted test routing still sends Slack, Signal, and iMessage suites to the right lane. (#46066) Thanks @scoootscooob.
- Agents/usage tracking: stop forcing `supportsUsageInStreaming: false` on non-native openai-completions endpoints so providers like DashScope, DeepSeek, and other OpenAI-compatible backends report token usage and cost instead of showing all zeros. (#46142)
- Node/startup: remove leftover debug `console.log("node host PATH: ...")` that printed the resolved PATH on every `openclaw node run` invocation. (#46411) - Node/startup: remove leftover debug `console.log("node host PATH: ...")` that printed the resolved PATH on every `openclaw node run` invocation. (#46411)
- Control UI/dashboard: preserve structured gateway shutdown reasons across restart disconnects so config-triggered restarts no longer fall back to `disconnected (1006): no reason`. (#46532) Thanks @vincentkoc. - Control UI/dashboard: preserve structured gateway shutdown reasons across restart disconnects so config-triggered restarts no longer fall back to `disconnected (1006): no reason`. (#46532) Thanks @vincentkoc.
- Feishu/topic threads: fetch full thread context, including prior bot replies, when starting a topic-thread session so follow-up turns in Feishu topics keep the right conversation state. Thanks @Coobiw. - Feishu/topic threads: fetch full thread context, including prior bot replies, when starting a topic-thread session so follow-up turns in Feishu topics keep the right conversation state. Thanks @Coobiw.

View File

@ -86,6 +86,14 @@ function expectSupportsDeveloperRoleForcedOff(overrides?: Partial<Model<Api>>):
const normalized = normalizeModelCompat(model as Model<Api>); const normalized = normalizeModelCompat(model as Model<Api>);
expect(supportsDeveloperRole(normalized)).toBe(false); expect(supportsDeveloperRole(normalized)).toBe(false);
} }
function expectSupportsUsageInStreamingForcedOff(overrides?: Partial<Model<Api>>): void {
const model = { ...baseModel(), ...overrides };
delete (model as { compat?: unknown }).compat;
const normalized = normalizeModelCompat(model as Model<Api>);
expect(supportsUsageInStreaming(normalized)).toBe(false);
}
function expectResolvedForwardCompat( function expectResolvedForwardCompat(
model: Model<Api> | undefined, model: Model<Api> | undefined,
expected: { provider: string; id: string }, expected: { provider: string; id: string },
@ -211,16 +219,11 @@ describe("normalizeModelCompat", () => {
}); });
}); });
it("leaves supportsUsageInStreaming at default for generic custom openai-completions provider", () => { it("forces supportsUsageInStreaming off for generic custom openai-completions provider", () => {
const model = { expectSupportsUsageInStreamingForcedOff({
...baseModel(),
provider: "custom-cpa", provider: "custom-cpa",
baseUrl: "https://cpa.example.com/v1", baseUrl: "https://cpa.example.com/v1",
}; });
delete (model as { compat?: unknown }).compat;
const normalized = normalizeModelCompat(model as Model<Api>);
// supportsUsageInStreaming is no longer forced off — pi-ai's default (true) applies
expect(supportsUsageInStreaming(normalized)).toBeUndefined();
}); });
it("forces supportsDeveloperRole off for Qwen proxy via openai-completions", () => { it("forces supportsDeveloperRole off for Qwen proxy via openai-completions", () => {
@ -270,7 +273,7 @@ describe("normalizeModelCompat", () => {
expect(supportsUsageInStreaming(normalized)).toBe(true); expect(supportsUsageInStreaming(normalized)).toBe(true);
}); });
it("forces supportsDeveloperRole off but leaves supportsUsageInStreaming unset for non-native endpoints", () => { it("still forces flags off when not explicitly set by user", () => {
const model = { const model = {
...baseModel(), ...baseModel(),
provider: "custom-cpa", provider: "custom-cpa",
@ -279,8 +282,7 @@ describe("normalizeModelCompat", () => {
delete (model as { compat?: unknown }).compat; delete (model as { compat?: unknown }).compat;
const normalized = normalizeModelCompat(model); const normalized = normalizeModelCompat(model);
expect(supportsDeveloperRole(normalized)).toBe(false); expect(supportsDeveloperRole(normalized)).toBe(false);
// supportsUsageInStreaming is no longer forced off — pi-ai default applies expect(supportsUsageInStreaming(normalized)).toBe(false);
expect(supportsUsageInStreaming(normalized)).toBeUndefined();
}); });
it("does not mutate caller model when forcing supportsDeveloperRole off", () => { it("does not mutate caller model when forcing supportsDeveloperRole off", () => {
@ -295,8 +297,7 @@ describe("normalizeModelCompat", () => {
expect(supportsDeveloperRole(model)).toBeUndefined(); expect(supportsDeveloperRole(model)).toBeUndefined();
expect(supportsUsageInStreaming(model)).toBeUndefined(); expect(supportsUsageInStreaming(model)).toBeUndefined();
expect(supportsDeveloperRole(normalized)).toBe(false); expect(supportsDeveloperRole(normalized)).toBe(false);
// supportsUsageInStreaming is not set by normalizeModelCompat — pi-ai default applies expect(supportsUsageInStreaming(normalized)).toBe(false);
expect(supportsUsageInStreaming(normalized)).toBeUndefined();
}); });
it("does not override explicit compat false", () => { it("does not override explicit compat false", () => {

View File

@ -52,16 +52,11 @@ export function normalizeModelCompat(model: Model<Api>): Model<Api> {
return model; return model;
} }
// The `developer` role is an OpenAI-native behavior that most compatible // The `developer` role and stream usage chunks are OpenAI-native behaviors.
// backends reject. Force it off for non-native endpoints unless the user // Many OpenAI-compatible backends reject `developer` and/or emit usage-only
// has explicitly opted in via their model config. // chunks that break strict parsers expecting choices[0]. For non-native
// // openai-completions endpoints, force both compat flags off — unless the
// `supportsUsageInStreaming` is NOT forced off — most OpenAI-compatible // user has explicitly opted in via their model config.
// backends (DashScope, DeepSeek, Groq, Together, etc.) handle
// `stream_options: { include_usage: true }` correctly, and disabling it
// silently breaks usage/cost tracking for all non-native providers.
// Users can still opt out with `compat.supportsUsageInStreaming: false`
// if their backend rejects the parameter.
const compat = model.compat ?? undefined; const compat = model.compat ?? undefined;
// When baseUrl is empty the pi-ai library defaults to api.openai.com, so // When baseUrl is empty the pi-ai library defaults to api.openai.com, so
// leave compat unchanged and let default native behavior apply. // leave compat unchanged and let default native behavior apply.
@ -70,22 +65,24 @@ export function normalizeModelCompat(model: Model<Api>): Model<Api> {
return model; return model;
} }
// Respect explicit user overrides. // Respect explicit user overrides: if the user has set a compat flag to
// true in their model definition, they know their endpoint supports it.
const forcedDeveloperRole = compat?.supportsDeveloperRole === true; const forcedDeveloperRole = compat?.supportsDeveloperRole === true;
const forcedUsageStreaming = compat?.supportsUsageInStreaming === true;
if (forcedDeveloperRole) { if (forcedDeveloperRole && forcedUsageStreaming) {
return model; return model;
} }
// Only force supportsDeveloperRole off. Leave supportsUsageInStreaming // Return a new object — do not mutate the caller's model reference.
// at whatever the user set or pi-ai's default (true).
return { return {
...model, ...model,
compat: compat compat: compat
? { ? {
...compat, ...compat,
supportsDeveloperRole: false, supportsDeveloperRole: forcedDeveloperRole || false,
supportsUsageInStreaming: forcedUsageStreaming || false,
} }
: { supportsDeveloperRole: false }, : { supportsDeveloperRole: false, supportsUsageInStreaming: false },
} as typeof model; } as typeof model;
} }