From b53be62e7ac1c1b72edd46b2bfcfbcd96bdadba5 Mon Sep 17 00:00:00 2001 From: Matthew Zhou Date: Tue, 10 Jun 2025 13:27:00 -0700 Subject: [PATCH] fix: Turn parallel tool calling off for OpenAI (#2737) --- letta/llm_api/openai_client.py | 1 + 1 file changed, 1 insertion(+) diff --git a/letta/llm_api/openai_client.py b/letta/llm_api/openai_client.py index e0ea6187..56ca5f42 100644 --- a/letta/llm_api/openai_client.py +++ b/letta/llm_api/openai_client.py @@ -217,6 +217,7 @@ class OpenAIClient(LLMClientBase): messages=fill_image_content_in_messages(openai_message_list, messages), tools=[OpenAITool(type="function", function=f) for f in tools] if tools else None, tool_choice=tool_choice, + parallel_tool_calls=False if tools else None, # Forcibly disable parallel tool calling user=str(), max_completion_tokens=llm_config.max_tokens, # NOTE: the reasoners that don't support temperature require 1.0, not None