feat(core): add gpt-5.3-codex model support (#9628)
* feat(core): add gpt-5.3-codex model support Add OpenAI gpt-5.3-codex model: context window overrides, model pricing and capabilities, none-reasoning-effort support, and test config. 🐾 Generated with [Letta Code](https://letta.com) Co-Authored-By: Letta <noreply@letta.com> * just stage-api && just publish-api --------- Co-authored-by: Letta <noreply@letta.com>
This commit is contained in:
@@ -29104,7 +29104,7 @@
|
|||||||
"type": "integer",
|
"type": "integer",
|
||||||
"title": "Limit",
|
"title": "Limit",
|
||||||
"description": "Character limit of the block.",
|
"description": "Character limit of the block.",
|
||||||
"default": 20000
|
"default": 100000
|
||||||
},
|
},
|
||||||
"project_id": {
|
"project_id": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
@@ -29317,7 +29317,7 @@
|
|||||||
"type": "integer",
|
"type": "integer",
|
||||||
"title": "Limit",
|
"title": "Limit",
|
||||||
"description": "Character limit of the block.",
|
"description": "Character limit of the block.",
|
||||||
"default": 20000
|
"default": 100000
|
||||||
},
|
},
|
||||||
"project_id": {
|
"project_id": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
@@ -29535,7 +29535,7 @@
|
|||||||
"type": "integer",
|
"type": "integer",
|
||||||
"title": "Limit",
|
"title": "Limit",
|
||||||
"description": "Character limit of the block.",
|
"description": "Character limit of the block.",
|
||||||
"default": 20000
|
"default": 100000
|
||||||
},
|
},
|
||||||
"project_id": {
|
"project_id": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
@@ -33178,7 +33178,7 @@
|
|||||||
"type": "integer",
|
"type": "integer",
|
||||||
"title": "Limit",
|
"title": "Limit",
|
||||||
"description": "Character limit of the block.",
|
"description": "Character limit of the block.",
|
||||||
"default": 20000
|
"default": 100000
|
||||||
},
|
},
|
||||||
"project_id": {
|
"project_id": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
@@ -34592,7 +34592,7 @@
|
|||||||
"type": "integer",
|
"type": "integer",
|
||||||
"title": "Limit",
|
"title": "Limit",
|
||||||
"description": "Character limit of the block.",
|
"description": "Character limit of the block.",
|
||||||
"default": 20000
|
"default": 100000
|
||||||
},
|
},
|
||||||
"project_id": {
|
"project_id": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
@@ -37733,7 +37733,7 @@
|
|||||||
"type": "integer",
|
"type": "integer",
|
||||||
"title": "Limit",
|
"title": "Limit",
|
||||||
"description": "Character limit of the block.",
|
"description": "Character limit of the block.",
|
||||||
"default": 20000
|
"default": 100000
|
||||||
},
|
},
|
||||||
"project_id": {
|
"project_id": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
|
|||||||
@@ -278,6 +278,8 @@ LLM_MAX_CONTEXT_WINDOW = {
|
|||||||
"gpt-5.2-pro": 272000,
|
"gpt-5.2-pro": 272000,
|
||||||
"gpt-5.2-pro-2025-12-11": 272000,
|
"gpt-5.2-pro-2025-12-11": 272000,
|
||||||
"gpt-5.2-codex": 272000,
|
"gpt-5.2-codex": 272000,
|
||||||
|
# gpt-5.3
|
||||||
|
"gpt-5.3-codex": 272000,
|
||||||
# reasoners
|
# reasoners
|
||||||
"o1": 200000,
|
"o1": 200000,
|
||||||
# "o1-pro": 200000, # responses API only
|
# "o1-pro": 200000, # responses API only
|
||||||
|
|||||||
@@ -88,7 +88,7 @@ def supports_none_reasoning_effort(model: str) -> bool:
|
|||||||
|
|
||||||
Currently, GPT-5.1 and GPT-5.2 models support the 'none' reasoning effort level.
|
Currently, GPT-5.1 and GPT-5.2 models support the 'none' reasoning effort level.
|
||||||
"""
|
"""
|
||||||
return model.startswith("gpt-5.1") or model.startswith("gpt-5.2")
|
return model.startswith("gpt-5.1") or model.startswith("gpt-5.2") or model.startswith("gpt-5.3")
|
||||||
|
|
||||||
|
|
||||||
def is_openai_5_model(model: str) -> bool:
|
def is_openai_5_model(model: str) -> bool:
|
||||||
|
|||||||
@@ -17295,6 +17295,32 @@
|
|||||||
"supports_tool_choice": true,
|
"supports_tool_choice": true,
|
||||||
"supports_vision": true
|
"supports_vision": true
|
||||||
},
|
},
|
||||||
|
"gpt-5.3-codex": {
|
||||||
|
"cache_read_input_token_cost": 1.75e-7,
|
||||||
|
"cache_read_input_token_cost_priority": 3.5e-7,
|
||||||
|
"input_cost_per_token": 1.75e-6,
|
||||||
|
"input_cost_per_token_priority": 3.5e-6,
|
||||||
|
"litellm_provider": "openai",
|
||||||
|
"max_input_tokens": 272000,
|
||||||
|
"max_output_tokens": 128000,
|
||||||
|
"max_tokens": 128000,
|
||||||
|
"mode": "responses",
|
||||||
|
"output_cost_per_token": 1.4e-5,
|
||||||
|
"output_cost_per_token_priority": 2.8e-5,
|
||||||
|
"supported_endpoints": ["/v1/responses"],
|
||||||
|
"supported_modalities": ["text", "image"],
|
||||||
|
"supported_output_modalities": ["text"],
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_native_streaming": true,
|
||||||
|
"supports_parallel_function_calling": true,
|
||||||
|
"supports_pdf_input": true,
|
||||||
|
"supports_prompt_caching": true,
|
||||||
|
"supports_reasoning": true,
|
||||||
|
"supports_response_schema": true,
|
||||||
|
"supports_system_messages": false,
|
||||||
|
"supports_tool_choice": true,
|
||||||
|
"supports_vision": true
|
||||||
|
},
|
||||||
"gpt-5-mini": {
|
"gpt-5-mini": {
|
||||||
"cache_read_input_token_cost": 2.5e-8,
|
"cache_read_input_token_cost": 2.5e-8,
|
||||||
"cache_read_input_token_cost_flex": 1.25e-8,
|
"cache_read_input_token_cost_flex": 1.25e-8,
|
||||||
|
|||||||
@@ -0,0 +1,8 @@
|
|||||||
|
{
|
||||||
|
"context_window": 32000,
|
||||||
|
"model": "gpt-5.3-codex",
|
||||||
|
"model_endpoint_type": "openai",
|
||||||
|
"model_endpoint": "https://api.openai.com/v1",
|
||||||
|
"model_wrapper": null,
|
||||||
|
"reasoning_effort": "low"
|
||||||
|
}
|
||||||
Reference in New Issue
Block a user