From 03e7639e2be268a9f91ab659c457355dc2e0db78 Mon Sep 17 00:00:00 2001 From: Ari Webb Date: Mon, 13 Oct 2025 17:40:09 -0700 Subject: [PATCH] handle llm error on request_async [LET-5403] (#5408) handle llm error on request_async Co-authored-by: Ari Webb --- letta/adapters/simple_llm_request_adapter.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/letta/adapters/simple_llm_request_adapter.py b/letta/adapters/simple_llm_request_adapter.py index ab3f662d..4caa8d00 100644 --- a/letta/adapters/simple_llm_request_adapter.py +++ b/letta/adapters/simple_llm_request_adapter.py @@ -38,7 +38,11 @@ class SimpleLLMRequestAdapter(LettaLLMRequestAdapter): self.request_data = request_data # Make the blocking LLM request - self.response_data = await self.llm_client.request_async(request_data, self.llm_config) + try: + self.response_data = await self.llm_client.request_async(request_data, self.llm_config) + except Exception as e: + raise self.llm_client.handle_llm_error(e) + self.llm_request_finish_timestamp_ns = get_utc_timestamp_ns() # Convert response to chat completion format