From f19a71dad111c19aed5e582cf47b8c34513f7d56 Mon Sep 17 00:00:00 2001 From: Ari Webb Date: Thu, 13 Nov 2025 15:44:21 -0800 Subject: [PATCH] chore: log problematic chunk (#6166) log problematic chunk Co-authored-by: Ari Webb --- letta/llm_api/openai_client.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/letta/llm_api/openai_client.py b/letta/llm_api/openai_client.py index 19e4a7f7..d9c589c0 100644 --- a/letta/llm_api/openai_client.py +++ b/letta/llm_api/openai_client.py @@ -813,9 +813,11 @@ class OpenAIClient(LLMClientBase): failed_chunks.append((start_idx, chunk_inputs[:mid], 1)) failed_chunks.append((start_idx + mid, chunk_inputs[mid:], 1)) else: + chunk_preview = str(chunk_inputs)[:500] if chunk_inputs else "None" logger.error( f"Failed to get embeddings for chunk starting at {start_idx} even with batch_size=1 " - f"and minimum chunk size {min_chunk_size}. Error: {result}" + f"and minimum chunk size {min_chunk_size}. Error: {result}. " + f"Chunk preview (first 500 chars): {chunk_preview}" ) raise result else: