fix: change the default context window to 30k (#2974)

Co-authored-by: Sarah Wooders <sarahwooders@gmail.com>
This commit is contained in:
Charles Packer
2025-06-24 17:42:50 -07:00
committed by GitHub
parent 19077a47f8
commit eaf7531448

View File

@@ -192,10 +192,21 @@ CORE_MEMORY_LINE_NUMBER_WARNING = (
# Constants to do with summarization / conversation length window
# The max amount of tokens supported by the underlying model (eg 8k for gpt-4 and Mistral 7B)
LLM_MAX_TOKENS = {
"DEFAULT": 8192,
"DEFAULT": 30000,
# deepseek
"deepseek-chat": 64000,
"deepseek-reasoner": 64000,
## OpenAI models: https://platform.openai.com/docs/models/overview
# reasoners
"o1": 200000,
# "o1-pro": 200000, # responses API only
"o1-2024-12-17": 200000,
"o3": 200000,
"o3-2025-04-16": 200000,
"o3-mini": 200000,
"o3-mini-2025-01-31": 200000,
# "o3-pro": 200000, # responses API only
# "o3-pro-2025-06-10": 200000,
"gpt-4.1": 1047576,
"gpt-4.1-2025-04-14": 1047576,
"gpt-4.1-mini": 1047576,
@@ -209,6 +220,7 @@ LLM_MAX_TOKENS = {
"chatgpt-4o-latest": 128000,
# "o1-preview-2024-09-12
"gpt-4o-2024-08-06": 128000,
"gpt-4o-2024-11-20": 128000,
"gpt-4-turbo-preview": 128000,
"gpt-4o": 128000,
"gpt-3.5-turbo-instruct": 16385,
@@ -218,7 +230,7 @@ LLM_MAX_TOKENS = {
# "davinci-002": 128000,
"gpt-4-turbo-2024-04-09": 128000,
# "gpt-4o-realtime-preview-2024-10-01
"gpt-4-turbo": 8192,
"gpt-4-turbo": 128000,
"gpt-4o-2024-05-13": 128000,
# "o1-mini
# "o1-mini-2024-09-12