test: run tests with gemini 2.5 instead (#3548)
Co-authored-by: Kian Jones <11655409+kianjones9@users.noreply.github.com> Co-authored-by: Kian Jones <kian@letta.com> Co-authored-by: Matt Zhou <mattzh1314@gmail.com>
This commit is contained in:
@@ -831,8 +831,8 @@ class JobManager:
|
||||
logger.error(error_message)
|
||||
result["callback_error"] = error_message
|
||||
# Continue silently - callback failures should not affect job completion
|
||||
|
||||
return result
|
||||
finally:
|
||||
return result
|
||||
|
||||
@trace_method
|
||||
async def _dispatch_callback_async(self, callback_info: dict) -> dict:
|
||||
@@ -860,5 +860,5 @@ class JobManager:
|
||||
logger.error(error_message)
|
||||
result["callback_error"] = error_message
|
||||
# Continue silently - callback failures should not affect job completion
|
||||
|
||||
return result
|
||||
finally:
|
||||
return result
|
||||
|
||||
@@ -5,4 +5,4 @@
|
||||
"model_endpoint": "https://generativelanguage.googleapis.com",
|
||||
"model_wrapper": null,
|
||||
"put_inner_thoughts_in_kwargs": true
|
||||
}
|
||||
}
|
||||
8
tests/configs/llm_model_configs/gemini-2.5-flash.json
Normal file
8
tests/configs/llm_model_configs/gemini-2.5-flash.json
Normal file
@@ -0,0 +1,8 @@
|
||||
{
|
||||
"context_window": 2097152,
|
||||
"model": "gemini-2.5-flash",
|
||||
"model_endpoint_type": "google_ai",
|
||||
"model_endpoint": "https://generativelanguage.googleapis.com",
|
||||
"model_wrapper": null,
|
||||
"put_inner_thoughts_in_kwargs": true
|
||||
}
|
||||
8
tests/configs/llm_model_configs/gemini-2.5-pro.json
Normal file
8
tests/configs/llm_model_configs/gemini-2.5-pro.json
Normal file
@@ -0,0 +1,8 @@
|
||||
{
|
||||
"context_window": 2097152,
|
||||
"model": "gemini-2.5-pro",
|
||||
"model_endpoint_type": "google_ai",
|
||||
"model_endpoint": "https://generativelanguage.googleapis.com",
|
||||
"model_wrapper": null,
|
||||
"put_inner_thoughts_in_kwargs": true
|
||||
}
|
||||
@@ -109,21 +109,22 @@ USER_MESSAGE_BASE64_IMAGE: List[MessageCreate] = [
|
||||
)
|
||||
]
|
||||
all_configs = [
|
||||
"openai-gpt-4o-mini.json",
|
||||
"openai-o1.json",
|
||||
"openai-o3.json",
|
||||
"openai-o4-mini.json",
|
||||
"azure-gpt-4o-mini.json",
|
||||
"claude-4-sonnet.json",
|
||||
"claude-3-5-sonnet.json",
|
||||
"claude-3-7-sonnet.json",
|
||||
# "openai-gpt-4o-mini.json",
|
||||
# "openai-o1.json",
|
||||
# "openai-o3.json",
|
||||
# "openai-o4-mini.json",
|
||||
# "azure-gpt-4o-mini.json",
|
||||
# "claude-4-sonnet.json",
|
||||
# "claude-3-5-sonnet.json",
|
||||
# "claude-3-7-sonnet.json",
|
||||
"claude-3-7-sonnet-extended.json",
|
||||
"bedrock-claude-4-sonnet.json",
|
||||
"gemini-1.5-pro.json",
|
||||
"gemini-2.5-flash-vertex.json",
|
||||
"gemini-2.5-pro-vertex.json",
|
||||
"together-qwen-2.5-72b-instruct.json",
|
||||
"ollama.json", # TODO (cliandy): enable this in ollama testing
|
||||
# "bedrock-claude-4-sonnet.json",
|
||||
# "gemini-2.5-pro.json",
|
||||
# "gemini-2.5-flash.json",
|
||||
# "gemini-2.5-flash-vertex.json",
|
||||
# "gemini-2.5-pro-vertex.json",
|
||||
# "together-qwen-2.5-72b-instruct.json",
|
||||
# "ollama.json", # TODO (cliandy): enable this in ollama testing
|
||||
]
|
||||
|
||||
reasoning_configs = [
|
||||
|
||||
Reference in New Issue
Block a user