From 4569382d20794b6f3341d601a9ffb526d80535c6 Mon Sep 17 00:00:00 2001 From: jnjpng Date: Fri, 20 Feb 2026 12:55:59 -0800 Subject: [PATCH] fix(models): set max_output_tokens for GPT-5 reasoning variants (#1064) Co-authored-by: Letta --- src/models.json | 18 ++++++++++++------ 1 file changed, 12 insertions(+), 6 deletions(-) diff --git a/src/models.json b/src/models.json index daadf37..2ec9ad0 100644 --- a/src/models.json +++ b/src/models.json @@ -629,7 +629,8 @@ "updateArgs": { "reasoning_effort": "minimal", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -640,7 +641,8 @@ "updateArgs": { "reasoning_effort": "low", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -651,7 +653,8 @@ "updateArgs": { "reasoning_effort": "medium", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -662,7 +665,8 @@ "updateArgs": { "reasoning_effort": "high", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -673,7 +677,8 @@ "updateArgs": { "reasoning_effort": "medium", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, { @@ -684,7 +689,8 @@ "updateArgs": { "reasoning_effort": "medium", "verbosity": "medium", - "context_window": 272000 + "context_window": 272000, + "max_output_tokens": 128000 } }, {