fix: max output tokens for gemini 3 models (#7322)
This commit is contained in:
@@ -360,6 +360,9 @@ LLM_MAX_CONTEXT_WINDOW = {
|
||||
"gemini-2.5-flash-preview-09-2025": 1048576,
|
||||
"gemini-2.5-flash-lite-preview-09-2025": 1048576,
|
||||
"gemini-2.5-computer-use-preview-10-2025": 1048576,
|
||||
# gemini 3
|
||||
"gemini-3-pro-preview": 1048576,
|
||||
"gemini-3-flash-preview": 1048576,
|
||||
# gemini latest aliases
|
||||
"gemini-flash-latest": 1048576,
|
||||
"gemini-flash-lite-latest": 1048576,
|
||||
|
||||
@@ -1,4 +1,6 @@
|
||||
GOOGLE_MODEL_TO_CONTEXT_LENGTH = {
|
||||
"gemini-3-pro-preview": 1048576,
|
||||
"gemini-3-flash-preview": 1048576,
|
||||
"gemini-2.5-pro": 1048576,
|
||||
"gemini-2.5-flash": 1048576,
|
||||
"gemini-live-2.5-flash": 1048576,
|
||||
|
||||
@@ -28,7 +28,7 @@ class GoogleAIProvider(Provider):
|
||||
|
||||
def get_default_max_output_tokens(self, model_name: str) -> int:
|
||||
"""Get the default max output tokens for Google Gemini models."""
|
||||
if "2.5" in model_name or "2-5" in model_name: # gemini-2.5-* or gemini-2-5-*
|
||||
if "2.5" in model_name or "2-5" in model_name or model_name.startswith("gemini-3"):
|
||||
return 65536
|
||||
return 8192 # default for google gemini
|
||||
|
||||
|
||||
@@ -18,7 +18,7 @@ class GoogleVertexProvider(Provider):
|
||||
|
||||
def get_default_max_output_tokens(self, model_name: str) -> int:
|
||||
"""Get the default max output tokens for Google Vertex models."""
|
||||
if "2.5" in model_name or "2-5" in model_name: # gemini-2.5-* or gemini-2-5-*
|
||||
if "2.5" in model_name or "2-5" in model_name or model_name.startswith("gemini-3"):
|
||||
return 65536
|
||||
return 8192 # default for google vertex
|
||||
|
||||
|
||||
Reference in New Issue
Block a user