fix: Increase max_output_tokens for gpt-5 and gpt-5.1 (#161)

This commit is contained in:
Kevin Lin
2025-12-06 14:01:28 -08:00
committed by GitHub
parent cf5c738eb9
commit 5dd96c9c8b
2 changed files with 23 additions and 8 deletions

View File

@@ -43,6 +43,13 @@ function buildModelSettings(
| "high",
};
}
// If a max_output_tokens override is provided in the model's updateArgs,
// propagate it into the OpenAI model settings so the backend can raise
// the generation cap (prevents premature ResponseIncompleteEvent with
// reason='max_output_tokens').
if (typeof updateArgs?.max_output_tokens === "number") {
openaiSettings.max_output_tokens = updateArgs.max_output_tokens as number;
}
return openaiSettings;
}

View File

@@ -49,7 +49,8 @@
"updateArgs": {
"reasoning_effort": "medium",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -60,7 +61,8 @@
"updateArgs": {
"reasoning_effort": "none",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -71,7 +73,8 @@
"updateArgs": {
"reasoning_effort": "low",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -83,7 +86,8 @@
"updateArgs": {
"reasoning_effort": "medium",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -94,7 +98,8 @@
"updateArgs": {
"reasoning_effort": "high",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -105,7 +110,8 @@
"updateArgs": {
"reasoning_effort": "none",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -117,7 +123,8 @@
"updateArgs": {
"reasoning_effort": "medium",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{
@@ -128,7 +135,8 @@
"updateArgs": {
"reasoning_effort": "high",
"verbosity": "medium",
"context_window": 272000
"context_window": 272000,
"max_output_tokens": 128000
}
},
{