From 5dd96c9c8b5871bacf50ad6af30309abfdbdebbf Mon Sep 17 00:00:00 2001 From: Kevin Lin Date: Sat, 6 Dec 2025 14:01:28 -0800 Subject: [PATCH] fix: Increase `max_output_tokens` for gpt-5 and gpt-5.1 (#161) --- src/agent/modify.ts | 7 +++++++ src/models.json | 24 ++++++++++++++++-------- 2 files changed, 23 insertions(+), 8 deletions(-) diff --git a/src/agent/modify.ts b/src/agent/modify.ts index 785cfd3..c1dc3bb 100644 --- a/src/agent/modify.ts +++ b/src/agent/modify.ts @@ -43,6 +43,13 @@ function buildModelSettings( | "high", }; } + // If a max_output_tokens override is provided in the model's updateArgs, + // propagate it into the OpenAI model settings so the backend can raise + // the generation cap (prevents premature ResponseIncompleteEvent with + // reason='max_output_tokens'). + if (typeof updateArgs?.max_output_tokens === "number") { + openaiSettings.max_output_tokens = updateArgs.max_output_tokens as number; + } return openaiSettings; } diff --git a/src/models.json b/src/models.json index affaa37..33a6884 100644 --- a/src/models.json +++ b/src/models.json @@ -49,7 +49,8 @@ "updateArgs": { "reasoning_effort": "medium", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -60,7 +61,8 @@ "updateArgs": { "reasoning_effort": "none", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -71,7 +73,8 @@ "updateArgs": { "reasoning_effort": "low", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -83,7 +86,8 @@ "updateArgs": { "reasoning_effort": "medium", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -94,7 +98,8 @@ "updateArgs": { "reasoning_effort": "high", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -105,7 +110,8 @@ "updateArgs": { "reasoning_effort": "none", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -117,7 +123,8 @@ "updateArgs": { "reasoning_effort": "medium", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -128,7 +135,8 @@ "updateArgs": { "reasoning_effort": "high", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, {