diff --git a/letta/services/job_manager.py b/letta/services/job_manager.py index b192afa8..4f8b76ff 100644 --- a/letta/services/job_manager.py +++ b/letta/services/job_manager.py @@ -831,8 +831,8 @@ class JobManager: logger.error(error_message) result["callback_error"] = error_message # Continue silently - callback failures should not affect job completion - - return result + finally: + return result @trace_method async def _dispatch_callback_async(self, callback_info: dict) -> dict: @@ -860,5 +860,5 @@ class JobManager: logger.error(error_message) result["callback_error"] = error_message # Continue silently - callback failures should not affect job completion - - return result + finally: + return result diff --git a/tests/configs/llm_model_configs/gemini-1.5-pro.json b/tests/configs/llm_model_configs/gemini-1.5-pro.json index 356d199b..cc0391a0 100644 --- a/tests/configs/llm_model_configs/gemini-1.5-pro.json +++ b/tests/configs/llm_model_configs/gemini-1.5-pro.json @@ -5,4 +5,4 @@ "model_endpoint": "https://generativelanguage.googleapis.com", "model_wrapper": null, "put_inner_thoughts_in_kwargs": true -} +} \ No newline at end of file diff --git a/tests/configs/llm_model_configs/gemini-2.5-flash.json b/tests/configs/llm_model_configs/gemini-2.5-flash.json new file mode 100644 index 00000000..d30c118f --- /dev/null +++ b/tests/configs/llm_model_configs/gemini-2.5-flash.json @@ -0,0 +1,8 @@ +{ + "context_window": 2097152, + "model": "gemini-2.5-flash", + "model_endpoint_type": "google_ai", + "model_endpoint": "https://generativelanguage.googleapis.com", + "model_wrapper": null, + "put_inner_thoughts_in_kwargs": true +} diff --git a/tests/configs/llm_model_configs/gemini-2.5-pro.json b/tests/configs/llm_model_configs/gemini-2.5-pro.json new file mode 100644 index 00000000..c291917c --- /dev/null +++ b/tests/configs/llm_model_configs/gemini-2.5-pro.json @@ -0,0 +1,8 @@ +{ + "context_window": 2097152, + "model": "gemini-2.5-pro", + "model_endpoint_type": "google_ai", + "model_endpoint": "https://generativelanguage.googleapis.com", + "model_wrapper": null, + "put_inner_thoughts_in_kwargs": true +} diff --git a/tests/integration_test_send_message.py b/tests/integration_test_send_message.py index b27f8b10..314cf702 100644 --- a/tests/integration_test_send_message.py +++ b/tests/integration_test_send_message.py @@ -109,21 +109,22 @@ USER_MESSAGE_BASE64_IMAGE: List[MessageCreate] = [ ) ] all_configs = [ - "openai-gpt-4o-mini.json", - "openai-o1.json", - "openai-o3.json", - "openai-o4-mini.json", - "azure-gpt-4o-mini.json", - "claude-4-sonnet.json", - "claude-3-5-sonnet.json", - "claude-3-7-sonnet.json", + # "openai-gpt-4o-mini.json", + # "openai-o1.json", + # "openai-o3.json", + # "openai-o4-mini.json", + # "azure-gpt-4o-mini.json", + # "claude-4-sonnet.json", + # "claude-3-5-sonnet.json", + # "claude-3-7-sonnet.json", "claude-3-7-sonnet-extended.json", - "bedrock-claude-4-sonnet.json", - "gemini-1.5-pro.json", - "gemini-2.5-flash-vertex.json", - "gemini-2.5-pro-vertex.json", - "together-qwen-2.5-72b-instruct.json", - "ollama.json", # TODO (cliandy): enable this in ollama testing + # "bedrock-claude-4-sonnet.json", + # "gemini-2.5-pro.json", + # "gemini-2.5-flash.json", + # "gemini-2.5-flash-vertex.json", + # "gemini-2.5-pro-vertex.json", + # "together-qwen-2.5-72b-instruct.json", + # "ollama.json", # TODO (cliandy): enable this in ollama testing ] reasoning_configs = [