feat: integrate simple adapter for non-streaming letta v1 agent (#5017)

This commit is contained in:
cthomas
2025-09-30 12:07:31 -07:00
committed by Caren Thomas
parent 74cf625f39
commit 4304a2e2ef
2 changed files with 5 additions and 2 deletions

View File

@@ -66,6 +66,10 @@ class SimpleLLMRequestAdapter(LettaLLMRequestAdapter):
else:
self.content = None
if self.reasoning_content and len(self.reasoning_content) > 0:
# Temp workaround to consolidate parts to persist reasoning content, this should be integrated better
self.content = self.reasoning_content + (self.content or [])
# Extract tool call
if self.chat_completions_response.choices[0].message.tool_calls:
self.tool_call = self.chat_completions_response.choices[0].message.tool_calls[0]

View File

@@ -4,7 +4,6 @@ from typing import AsyncGenerator, Optional
from opentelemetry.trace import Span
from letta.adapters.letta_llm_adapter import LettaLLMAdapter
from letta.adapters.letta_llm_request_adapter import LettaLLMRequestAdapter
from letta.adapters.simple_llm_request_adapter import SimpleLLMRequestAdapter
from letta.adapters.simple_llm_stream_adapter import SimpleLLMStreamAdapter
from letta.agents.helpers import (
@@ -89,7 +88,7 @@ class LettaAgentV3(LettaAgentV2):
messages=in_context_messages + self.response_messages,
input_messages_to_persist=input_messages_to_persist,
# TODO need to support non-streaming adapter too
llm_adapter=LettaLLMRequestAdapter(llm_client=self.llm_client, llm_config=self.agent_state.llm_config),
llm_adapter=SimpleLLMRequestAdapter(llm_client=self.llm_client, llm_config=self.agent_state.llm_config),
run_id=run_id,
# use_assistant_message=use_assistant_message,
include_return_message_types=include_return_message_types,