From 0543a60538546e90ff3dd8a2a6c710ad162d9887 Mon Sep 17 00:00:00 2001 From: Matthew Zhou Date: Tue, 14 Oct 2025 13:27:03 -0700 Subject: [PATCH] chore: Restore chunk size for openai embeddings (#5431) Restore chunk size --- letta/llm_api/openai_client.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/letta/llm_api/openai_client.py b/letta/llm_api/openai_client.py index 1e259742..0e8b8e39 100644 --- a/letta/llm_api/openai_client.py +++ b/letta/llm_api/openai_client.py @@ -727,7 +727,7 @@ class OpenAIClient(LLMClientBase): # queue of (start_idx, chunk_inputs) to process chunks_to_process = [(i, inputs[i : i + 2048]) for i in range(0, len(inputs), 2048)] - min_chunk_size = 64 + min_chunk_size = 256 while chunks_to_process: tasks = []