From 5801cce1b597da59acdbe4a20e0a5b38e7fde568 Mon Sep 17 00:00:00 2001 From: "opencode-agent[bot]" Date: Mon, 11 May 2026 01:18:39 +0000 Subject: [PATCH] chore: generate --- packages/llm/src/protocols/bedrock-converse.ts | 5 +---- packages/llm/test/provider/bedrock-converse.test.ts | 4 +--- packages/llm/test/recorded-scenarios.ts | 3 +-- 3 files changed, 3 insertions(+), 9 deletions(-) diff --git a/packages/llm/src/protocols/bedrock-converse.ts b/packages/llm/src/protocols/bedrock-converse.ts index 34bb13e61a..e2ba1ff3be 100644 --- a/packages/llm/src/protocols/bedrock-converse.ts +++ b/packages/llm/src/protocols/bedrock-converse.ts @@ -225,10 +225,7 @@ const lowerToolSpec = (tool: ToolDefinition): BedrockToolSpec => ({ }, }) -const lowerTools = ( - breakpoints: BedrockCache.Breakpoints, - tools: ReadonlyArray, -): BedrockTool[] => { +const lowerTools = (breakpoints: BedrockCache.Breakpoints, tools: ReadonlyArray): BedrockTool[] => { const result: BedrockTool[] = [] for (const tool of tools) { result.push(lowerToolSpec(tool)) diff --git a/packages/llm/test/provider/bedrock-converse.test.ts b/packages/llm/test/provider/bedrock-converse.test.ts index d72e7f3116..afadd89ac7 100644 --- a/packages/llm/test/provider/bedrock-converse.test.ts +++ b/packages/llm/test/provider/bedrock-converse.test.ts @@ -464,9 +464,7 @@ describe("Bedrock Converse route", () => { const prepared = yield* LLMClient.prepare( LLM.request({ model, - tools: [ - { name: "lookup", description: "lookup", inputSchema: { type: "object", properties: {} }, cache }, - ], + tools: [{ name: "lookup", description: "lookup", inputSchema: { type: "object", properties: {} }, cache }], messages: [ LLM.user("What's the weather?"), LLM.assistant([LLM.toolCall({ id: "call_1", name: "lookup", input: {} })]), diff --git a/packages/llm/test/recorded-scenarios.ts b/packages/llm/test/recorded-scenarios.ts index 2361bfdb78..8a02bc3a0a 100644 --- a/packages/llm/test/recorded-scenarios.ts +++ b/packages/llm/test/recorded-scenarios.ts @@ -12,8 +12,7 @@ export const weatherToolName = "get_weather" // a fixed sentence — the cassette replays bit-for-bit, so the exact text matters // only when re-recording with `RECORD=true`. export const LARGE_CACHEABLE_SYSTEM = (() => { - const sentence = - "You are a concise, factual assistant. Answer precisely and avoid filler. Cite numbers when known. " + const sentence = "You are a concise, factual assistant. Answer precisely and avoid filler. Cite numbers when known. " // ~100 chars per sentence × 250 repeats ≈ 25,000 chars ≈ 5k+ tokens, safely // above every provider's threshold. return sentence.repeat(250)