* add llm client errors to traces * update response json for telemetry * prevent silent failures and properly log errored responses in streaming path * remove double logging --------- Co-authored-by: Amy Guan <amy@letta.com> Co-authored-by: Kian Jones <kian@letta.com>
743 lines
32 KiB
Python
743 lines
32 KiB
Python
import asyncio
|
||
import json
|
||
import traceback
|
||
from typing import List, Optional, Tuple, Union
|
||
|
||
from letta.agents.ephemeral_summary_agent import EphemeralSummaryAgent
|
||
from letta.constants import (
|
||
DEFAULT_MESSAGE_TOOL,
|
||
DEFAULT_MESSAGE_TOOL_KWARG,
|
||
MESSAGE_SUMMARY_REQUEST_ACK,
|
||
TOOL_RETURN_TRUNCATION_CHARS,
|
||
)
|
||
from letta.errors import ContextWindowExceededError
|
||
from letta.helpers.message_helper import convert_message_creates_to_messages
|
||
from letta.llm_api.llm_client import LLMClient
|
||
from letta.log import get_logger
|
||
from letta.otel.tracing import trace_method
|
||
from letta.prompts import gpt_summarize
|
||
from letta.schemas.enums import AgentType, LLMCallType, MessageRole, ProviderType
|
||
from letta.schemas.letta_message_content import TextContent
|
||
from letta.schemas.llm_config import LLMConfig
|
||
from letta.schemas.message import Message, MessageCreate
|
||
from letta.schemas.user import User
|
||
from letta.services.agent_manager import AgentManager
|
||
from letta.services.message_manager import MessageManager
|
||
from letta.services.summarizer.enums import SummarizationMode
|
||
from letta.system import package_summarize_message_no_counts
|
||
from letta.utils import safe_create_task
|
||
|
||
logger = get_logger(__name__)
|
||
|
||
|
||
# NOTE: legacy, new version is functional
|
||
class Summarizer:
|
||
"""
|
||
Handles summarization or trimming of conversation messages based on
|
||
the specified SummarizationMode. For now, we demonstrate a simple
|
||
static buffer approach but leave room for more advanced strategies.
|
||
"""
|
||
|
||
def __init__(
|
||
self,
|
||
mode: SummarizationMode,
|
||
summarizer_agent: Optional[Union[EphemeralSummaryAgent, "VoiceSleeptimeAgent"]] = None,
|
||
message_buffer_limit: int = 10,
|
||
message_buffer_min: int = 3,
|
||
partial_evict_summarizer_percentage: float = 0.30,
|
||
agent_manager: Optional[AgentManager] = None,
|
||
message_manager: Optional[MessageManager] = None,
|
||
actor: Optional[User] = None,
|
||
agent_id: Optional[str] = None,
|
||
run_id: Optional[str] = None,
|
||
step_id: Optional[str] = None,
|
||
):
|
||
self.mode = mode
|
||
|
||
# Need to do validation on this
|
||
# TODO: Move this to config
|
||
self.message_buffer_limit = message_buffer_limit
|
||
self.message_buffer_min = message_buffer_min
|
||
self.summarizer_agent = summarizer_agent
|
||
self.partial_evict_summarizer_percentage = partial_evict_summarizer_percentage
|
||
|
||
# for partial buffer only
|
||
self.agent_manager = agent_manager
|
||
self.message_manager = message_manager
|
||
self.actor = actor
|
||
self.agent_id = agent_id
|
||
self.run_id = run_id
|
||
self.step_id = step_id
|
||
|
||
@trace_method
|
||
async def summarize(
|
||
self,
|
||
in_context_messages: List[Message],
|
||
new_letta_messages: List[Message],
|
||
force: bool = False,
|
||
clear: bool = False,
|
||
run_id: Optional[str] = None,
|
||
step_id: Optional[str] = None,
|
||
) -> Tuple[List[Message], bool]:
|
||
"""
|
||
Summarizes or trims in_context_messages according to the chosen mode,
|
||
and returns the updated messages plus any optional "summary message".
|
||
|
||
Args:
|
||
in_context_messages: The existing messages in the conversation's context.
|
||
new_letta_messages: The newly added Letta messages (just appended).
|
||
force: Force summarize even if the criteria is not met
|
||
run_id: Optional run ID for telemetry (overrides instance default)
|
||
step_id: Optional step ID for telemetry (overrides instance default)
|
||
|
||
Returns:
|
||
(updated_messages, summary_message)
|
||
updated_messages: The new context after trimming/summary
|
||
summary_message: Optional summarization message that was created
|
||
(could be appended to the conversation if desired)
|
||
"""
|
||
effective_run_id = run_id if run_id is not None else self.run_id
|
||
effective_step_id = step_id if step_id is not None else self.step_id
|
||
|
||
if self.mode == SummarizationMode.STATIC_MESSAGE_BUFFER:
|
||
return self._static_buffer_summarization(
|
||
in_context_messages,
|
||
new_letta_messages,
|
||
force=force,
|
||
clear=clear,
|
||
)
|
||
elif self.mode == SummarizationMode.PARTIAL_EVICT_MESSAGE_BUFFER:
|
||
return await self._partial_evict_buffer_summarization(
|
||
in_context_messages,
|
||
new_letta_messages,
|
||
force=force,
|
||
clear=clear,
|
||
run_id=effective_run_id,
|
||
step_id=effective_step_id,
|
||
)
|
||
else:
|
||
# Fallback or future logic
|
||
return in_context_messages, False
|
||
|
||
def fire_and_forget(self, coro):
|
||
task = safe_create_task(coro, label="summarizer_background_task")
|
||
|
||
def callback(t):
|
||
try:
|
||
t.result() # This re-raises exceptions from the task
|
||
except Exception:
|
||
logger.exception("Background task failed")
|
||
|
||
task.add_done_callback(callback)
|
||
return task
|
||
|
||
async def _partial_evict_buffer_summarization(
|
||
self,
|
||
in_context_messages: List[Message],
|
||
new_letta_messages: List[Message],
|
||
force: bool = False,
|
||
clear: bool = False,
|
||
run_id: Optional[str] = None,
|
||
step_id: Optional[str] = None,
|
||
) -> Tuple[List[Message], bool]:
|
||
"""Summarization as implemented in the original MemGPT loop, but using message count instead of token count.
|
||
Evict a partial amount of messages, and replace message[1] with a recursive summary.
|
||
|
||
Note that this can't be made sync, because we're waiting on the summary to inject it into the context window,
|
||
unlike the version that writes it to a block.
|
||
|
||
Unless force is True, don't summarize.
|
||
Ignore clear, we don't use it.
|
||
"""
|
||
all_in_context_messages = in_context_messages + new_letta_messages
|
||
|
||
if not force:
|
||
logger.debug("Not forcing summarization, returning in-context messages as is.")
|
||
return all_in_context_messages, False
|
||
|
||
# First step: determine how many messages to retain
|
||
total_message_count = len(all_in_context_messages)
|
||
assert self.partial_evict_summarizer_percentage >= 0.0 and self.partial_evict_summarizer_percentage <= 1.0
|
||
target_message_start = round((1.0 - self.partial_evict_summarizer_percentage) * total_message_count)
|
||
logger.info(f"Target message count: {total_message_count}->{(total_message_count - target_message_start)}")
|
||
|
||
# The summary message we'll insert is role 'user' (vs 'assistant', 'tool', or 'system')
|
||
# We are going to put it at index 1 (index 0 is the system message)
|
||
# That means that index 2 needs to be role 'assistant', so walk up the list starting at
|
||
# the target_message_count and find the first assistant message
|
||
for i in range(target_message_start, total_message_count):
|
||
if all_in_context_messages[i].role == MessageRole.assistant:
|
||
assistant_message_index = i
|
||
break
|
||
else:
|
||
raise ValueError(f"No assistant message found from indices {target_message_start} to {total_message_count}")
|
||
|
||
# The sequence to summarize is index 1 -> assistant_message_index
|
||
messages_to_summarize = all_in_context_messages[1:assistant_message_index]
|
||
logger.info(f"Eviction indices: {1}->{assistant_message_index}(/{total_message_count})")
|
||
|
||
# Dynamically get the LLMConfig from the summarizer agent
|
||
# Pretty cringe code here that we need the agent for this but we don't use it
|
||
agent_state = await self.agent_manager.get_agent_by_id_async(agent_id=self.agent_id, actor=self.actor)
|
||
|
||
# TODO if we do this via the "agent", then we can more easily allow toggling on the memory block version
|
||
from letta.settings import summarizer_settings
|
||
|
||
summary_message_str = await simple_summary(
|
||
messages=messages_to_summarize,
|
||
llm_config=agent_state.llm_config,
|
||
actor=self.actor,
|
||
include_ack=True,
|
||
agent_id=self.agent_id,
|
||
agent_tags=agent_state.tags,
|
||
run_id=run_id if run_id is not None else self.run_id,
|
||
step_id=step_id if step_id is not None else self.step_id,
|
||
compaction_settings={
|
||
"mode": str(summarizer_settings.mode.value),
|
||
"message_buffer_limit": summarizer_settings.message_buffer_limit,
|
||
"message_buffer_min": summarizer_settings.message_buffer_min,
|
||
"partial_evict_summarizer_percentage": summarizer_settings.partial_evict_summarizer_percentage,
|
||
},
|
||
)
|
||
|
||
# TODO add counts back
|
||
# Recall message count
|
||
# num_recall_messages_current = await self.message_manager.size_async(actor=self.actor, agent_id=agent_state.id)
|
||
# num_messages_evicted = len(messages_to_summarize)
|
||
# num_recall_messages_hidden = num_recall_messages_total - len()
|
||
|
||
# Create the summary message
|
||
summary_message_str_packed = package_summarize_message_no_counts(
|
||
summary=summary_message_str,
|
||
timezone=agent_state.timezone,
|
||
)
|
||
summary_message_obj = (
|
||
await convert_message_creates_to_messages(
|
||
message_creates=[
|
||
MessageCreate(
|
||
role=MessageRole.user,
|
||
content=[TextContent(text=summary_message_str_packed)],
|
||
)
|
||
],
|
||
agent_id=agent_state.id,
|
||
timezone=agent_state.timezone,
|
||
# We already packed, don't pack again
|
||
wrap_user_message=False,
|
||
wrap_system_message=False,
|
||
run_id=None, # TODO: add this
|
||
)
|
||
)[0]
|
||
|
||
# Create the message in the DB
|
||
await self.message_manager.create_many_messages_async(
|
||
pydantic_msgs=[summary_message_obj],
|
||
actor=self.actor,
|
||
project_id=agent_state.project_id,
|
||
template_id=agent_state.template_id,
|
||
)
|
||
|
||
updated_in_context_messages = all_in_context_messages[assistant_message_index:]
|
||
return [all_in_context_messages[0], summary_message_obj] + updated_in_context_messages, True
|
||
|
||
def _static_buffer_summarization(
|
||
self,
|
||
in_context_messages: List[Message],
|
||
new_letta_messages: List[Message],
|
||
force: bool = False,
|
||
clear: bool = False,
|
||
) -> Tuple[List[Message], bool]:
|
||
"""
|
||
Implements static buffer summarization by maintaining a fixed-size message buffer (< N messages).
|
||
|
||
Logic:
|
||
1. Combine existing context messages with new messages
|
||
2. If total messages <= buffer limit and not forced, return unchanged
|
||
3. Calculate how many messages to retain (0 if clear=True, otherwise message_buffer_min)
|
||
4. Find the trim index to keep the most recent messages while preserving user message boundaries
|
||
5. Evict older messages (everything between system message and trim index)
|
||
6. If summarizer agent is available, trigger background summarization of evicted messages
|
||
7. Return updated context with system message + retained recent messages
|
||
|
||
Args:
|
||
in_context_messages: Existing conversation context messages
|
||
new_letta_messages: Newly added messages to append
|
||
force: Force summarization even if buffer limit not exceeded
|
||
clear: Clear all messages except system message (retain_count = 0)
|
||
|
||
Returns:
|
||
Tuple of (updated_messages, was_summarized)
|
||
- updated_messages: New context after trimming/summarization
|
||
- was_summarized: True if messages were evicted and summarization triggered
|
||
"""
|
||
|
||
all_in_context_messages = in_context_messages + new_letta_messages
|
||
|
||
if len(all_in_context_messages) <= self.message_buffer_limit and not force:
|
||
logger.info(
|
||
f"Nothing to evict, returning in context messages as is. Current buffer length is {len(all_in_context_messages)}, limit is {self.message_buffer_limit}."
|
||
)
|
||
return all_in_context_messages, False
|
||
|
||
retain_count = 0 if clear else self.message_buffer_min
|
||
|
||
if not force:
|
||
logger.info(f"Buffer length hit {self.message_buffer_limit}, evicting until we retain only {retain_count} messages.")
|
||
else:
|
||
logger.info(f"Requested force summarization, evicting until we retain only {retain_count} messages.")
|
||
|
||
target_trim_index = max(1, len(all_in_context_messages) - retain_count)
|
||
|
||
while target_trim_index < len(all_in_context_messages) and all_in_context_messages[target_trim_index].role != MessageRole.user:
|
||
target_trim_index += 1
|
||
|
||
# If the first retained message is an approval request, also keep the assistant message before it
|
||
# (they're part of the same LLM response - assistant has reasoning/tool_calls, approval has approval-required subset)
|
||
if target_trim_index < len(all_in_context_messages):
|
||
first_retained = all_in_context_messages[target_trim_index]
|
||
if first_retained.role == MessageRole.approval and target_trim_index > 1:
|
||
# Check if the message before it is an assistant from the same step
|
||
prev_message = all_in_context_messages[target_trim_index - 1]
|
||
if prev_message.role == MessageRole.assistant and prev_message.step_id == first_retained.step_id:
|
||
# Back up to include the assistant message with reasoning
|
||
target_trim_index -= 1
|
||
|
||
evicted_messages = all_in_context_messages[1:target_trim_index] # everything except sys msg
|
||
updated_in_context_messages = all_in_context_messages[target_trim_index:] # may be empty
|
||
|
||
# If *no* messages were evicted we really have nothing to do
|
||
if not evicted_messages:
|
||
logger.info("Nothing to evict, returning in-context messages as-is.")
|
||
return all_in_context_messages, False
|
||
|
||
if self.summarizer_agent:
|
||
# Only invoke if summarizer agent is passed in
|
||
# Format
|
||
formatted_evicted_messages = format_transcript(evicted_messages)
|
||
formatted_in_context_messages = format_transcript(updated_in_context_messages)
|
||
|
||
# TODO: This is hyperspecific to voice, generalize!
|
||
# Update the message transcript of the memory agent
|
||
if not isinstance(self.summarizer_agent, EphemeralSummaryAgent):
|
||
self.summarizer_agent.update_message_transcript(
|
||
message_transcripts=formatted_evicted_messages + formatted_in_context_messages
|
||
)
|
||
|
||
# Add line numbers to the formatted messages
|
||
offset = len(formatted_evicted_messages)
|
||
formatted_evicted_messages = [f"{i}. {msg}" for (i, msg) in enumerate(formatted_evicted_messages)]
|
||
formatted_in_context_messages = [f"{i + offset}. {msg}" for (i, msg) in enumerate(formatted_in_context_messages)]
|
||
|
||
summary_request_text = build_summary_request_text(
|
||
retain_count=retain_count,
|
||
evicted_messages=formatted_evicted_messages,
|
||
in_context_messages=formatted_in_context_messages,
|
||
)
|
||
|
||
# Fire-and-forget the summarization task
|
||
self.fire_and_forget(
|
||
self.summarizer_agent.step([MessageCreate(role=MessageRole.user, content=[TextContent(text=summary_request_text)])])
|
||
)
|
||
|
||
return [all_in_context_messages[0]] + updated_in_context_messages, True
|
||
|
||
|
||
def simple_formatter(
|
||
messages: List[Message],
|
||
include_system: bool = False,
|
||
tool_return_truncation_chars: int | None = None,
|
||
) -> str:
|
||
"""Go from an OpenAI-style list of messages to a concatenated string.
|
||
|
||
Optionally clamps tool-return content to avoid ballooning the summarizer transcript.
|
||
"""
|
||
|
||
parsed_messages = Message.to_openai_dicts_from_list(
|
||
[message for message in messages if message.role != MessageRole.system or include_system],
|
||
tool_return_truncation_chars=tool_return_truncation_chars,
|
||
)
|
||
return "<start_transcript>\n" + "\n".join(json.dumps(msg) for msg in parsed_messages) + "\n<end_transcript>\n. Generate the summary."
|
||
|
||
|
||
def middle_truncate_text(
|
||
text: str,
|
||
budget_chars: int,
|
||
head_frac: float = 0.3,
|
||
tail_frac: float = 0.3,
|
||
) -> tuple[str, int]:
|
||
"""Middle-truncate a string to fit within a character budget.
|
||
|
||
Keeps the first `head_frac` and last `tail_frac` portions (by budget chars)
|
||
and drops the middle. Returns (truncated_text, dropped_char_count).
|
||
|
||
Fractions are relative to budget, not original text length.
|
||
"""
|
||
if budget_chars <= 0 or len(text) <= budget_chars:
|
||
return text, 0
|
||
|
||
head_len = max(0, int(budget_chars * head_frac))
|
||
tail_len = max(0, int(budget_chars * tail_frac))
|
||
# Ensure head + tail <= budget; allocate remainder to tail preferentially
|
||
if head_len + tail_len > budget_chars:
|
||
tail_len = max(0, budget_chars - head_len)
|
||
|
||
head = text[:head_len]
|
||
tail = text[-tail_len:] if tail_len > 0 else ""
|
||
dropped = max(0, len(text) - (len(head) + len(tail)))
|
||
|
||
marker = f"\n[TRUNCATED: dropped {dropped} middle chars due to context budget]\n"
|
||
# If marker would overflow budget, shrink tail to fit marker
|
||
available_for_marker = budget_chars - (len(head) + len(tail))
|
||
if available_for_marker < len(marker):
|
||
# reduce tail to free up space
|
||
over = len(marker) - available_for_marker
|
||
tail = tail[:-over] if over < len(tail) else ""
|
||
|
||
return head + marker + tail, dropped
|
||
|
||
|
||
def build_summary_request_text(retain_count: int, evicted_messages: List[str], in_context_messages: List[str]) -> str:
|
||
parts: List[str] = []
|
||
if retain_count == 0:
|
||
parts.append(
|
||
"You’re a memory-recall helper for an AI that is about to forget all prior messages. Scan the conversation history and write crisp notes that capture any important facts or insights about the conversation history."
|
||
)
|
||
else:
|
||
parts.append(
|
||
f"You’re a memory-recall helper for an AI that can only keep the last {retain_count} messages. Scan the conversation history, focusing on messages about to drop out of that window, and write crisp notes that capture any important facts or insights about the human so they aren’t lost."
|
||
)
|
||
|
||
if evicted_messages:
|
||
parts.append("\n(Older) Evicted Messages:")
|
||
for item in evicted_messages:
|
||
parts.append(f" {item}")
|
||
|
||
if retain_count > 0 and in_context_messages:
|
||
parts.append("\n(Newer) In-Context Messages:")
|
||
for item in in_context_messages:
|
||
parts.append(f" {item}")
|
||
|
||
return "\n".join(parts) + "\n"
|
||
|
||
|
||
def simple_message_wrapper(openai_msg: dict) -> Message:
|
||
"""Extremely simple way to map from role/content to Message object w/ throwaway dummy fields"""
|
||
|
||
if "role" not in openai_msg:
|
||
raise ValueError(f"Missing role in openai_msg: {openai_msg}")
|
||
if "content" not in openai_msg:
|
||
raise ValueError(f"Missing content in openai_msg: {openai_msg}")
|
||
|
||
if openai_msg["role"] == "user":
|
||
return Message(
|
||
role=MessageRole.user,
|
||
content=[TextContent(text=openai_msg["content"])],
|
||
)
|
||
elif openai_msg["role"] == "assistant":
|
||
return Message(
|
||
role=MessageRole.assistant,
|
||
content=[TextContent(text=openai_msg["content"])],
|
||
)
|
||
elif openai_msg["role"] == "system":
|
||
return Message(
|
||
role=MessageRole.system,
|
||
content=[TextContent(text=openai_msg["content"])],
|
||
)
|
||
else:
|
||
raise ValueError(f"Unknown role: {openai_msg['role']}")
|
||
|
||
|
||
@trace_method
|
||
async def simple_summary(
|
||
messages: List[Message],
|
||
llm_config: LLMConfig,
|
||
actor: User,
|
||
include_ack: bool = True,
|
||
prompt: str | None = None,
|
||
telemetry_manager: "TelemetryManager | None" = None,
|
||
agent_id: str | None = None,
|
||
agent_tags: List[str] | None = None,
|
||
run_id: str | None = None,
|
||
step_id: str | None = None,
|
||
compaction_settings: dict | None = None,
|
||
) -> str:
|
||
"""Generate a simple summary from a list of messages.
|
||
|
||
Intentionally kept functional due to the simplicity of the prompt.
|
||
"""
|
||
from letta.services.telemetry_manager import TelemetryManager
|
||
|
||
# Create an LLMClient from the config
|
||
llm_client = LLMClient.create(
|
||
provider_type=llm_config.model_endpoint_type,
|
||
put_inner_thoughts_first=True,
|
||
actor=actor,
|
||
)
|
||
assert llm_client is not None
|
||
|
||
# Always set telemetry context - create TelemetryManager if not provided
|
||
tm = telemetry_manager or TelemetryManager()
|
||
llm_client.set_telemetry_context(
|
||
telemetry_manager=tm,
|
||
agent_id=agent_id,
|
||
agent_tags=agent_tags,
|
||
run_id=run_id,
|
||
step_id=step_id,
|
||
call_type=LLMCallType.summarization,
|
||
org_id=actor.organization_id if actor else None,
|
||
user_id=actor.id if actor else None,
|
||
compaction_settings=compaction_settings,
|
||
)
|
||
|
||
# Prepare the messages payload to send to the LLM
|
||
system_prompt = prompt or gpt_summarize.SYSTEM
|
||
# Build the initial transcript without clamping to preserve fidelity
|
||
# TODO proactively clip here?
|
||
summary_transcript = simple_formatter(messages)
|
||
logger.info(f"Summarizing {len(messages)} messages with prompt: {system_prompt}")
|
||
|
||
if include_ack:
|
||
logger.info(f"Summarizing with ACK for model {llm_config.model}")
|
||
input_messages = [
|
||
{"role": "system", "content": system_prompt},
|
||
{"role": "assistant", "content": MESSAGE_SUMMARY_REQUEST_ACK},
|
||
{"role": "user", "content": summary_transcript},
|
||
]
|
||
else:
|
||
logger.info(f"Summarizing without ACK for model {llm_config.model}")
|
||
input_messages = [
|
||
{"role": "system", "content": system_prompt},
|
||
{"role": "user", "content": summary_transcript},
|
||
]
|
||
input_messages_obj = [simple_message_wrapper(msg) for msg in input_messages]
|
||
# Build a local LLMConfig for v1-style summarization which uses native content and must not
|
||
# include inner thoughts in kwargs to avoid conflicts in Anthropic formatting.
|
||
# We also disable enable_reasoner to avoid extended thinking requirements (Anthropic requires
|
||
# assistant messages to start with thinking blocks when extended thinking is enabled).
|
||
summarizer_llm_config = LLMConfig(**llm_config.model_dump())
|
||
summarizer_llm_config.put_inner_thoughts_in_kwargs = False
|
||
summarizer_llm_config.enable_reasoner = False
|
||
|
||
async def _run_summarizer_request(req_data: dict, req_messages_obj: list[Message]) -> str:
|
||
"""Run summarization request and return assistant text.
|
||
|
||
For Anthropic, use provider-side streaming to avoid long-request failures
|
||
(Anthropic requires streaming for requests that may exceed ~10 minutes).
|
||
"""
|
||
|
||
if summarizer_llm_config.model_endpoint_type in [ProviderType.anthropic, ProviderType.bedrock]:
|
||
logger.info(
|
||
"Summarizer: using provider streaming (%s/%s) to avoid long-request failures",
|
||
summarizer_llm_config.model_endpoint_type,
|
||
summarizer_llm_config.model,
|
||
)
|
||
# Stream from provider and accumulate the final assistant text.
|
||
from letta.interfaces.anthropic_parallel_tool_call_streaming_interface import (
|
||
SimpleAnthropicStreamingInterface,
|
||
)
|
||
|
||
interface = SimpleAnthropicStreamingInterface(
|
||
requires_approval_tools=[],
|
||
run_id=None,
|
||
step_id=None,
|
||
)
|
||
|
||
# AnthropicClient.stream_async sets request_data["stream"] = True internally.
|
||
try:
|
||
stream = await llm_client.stream_async(req_data, summarizer_llm_config)
|
||
async for _chunk in interface.process(stream):
|
||
pass
|
||
|
||
content_parts = interface.get_content()
|
||
text = "".join(part.text for part in content_parts if isinstance(part, TextContent)).strip()
|
||
|
||
await llm_client.log_provider_trace_async(
|
||
request_data=req_data,
|
||
response_json={
|
||
"content": text,
|
||
"model": summarizer_llm_config.model,
|
||
"usage": {
|
||
"input_tokens": getattr(interface, "input_tokens", None),
|
||
"output_tokens": getattr(interface, "output_tokens", None),
|
||
},
|
||
},
|
||
llm_config=summarizer_llm_config,
|
||
)
|
||
except Exception as e:
|
||
await llm_client.log_provider_trace_async(
|
||
request_data=req_data,
|
||
response_json=None,
|
||
llm_config=summarizer_llm_config,
|
||
error_msg=str(e),
|
||
error_type=type(e).__name__,
|
||
)
|
||
raise
|
||
|
||
if not text:
|
||
logger.warning("No content returned from summarizer (streaming path)")
|
||
raise Exception("Summary failed to generate")
|
||
return text
|
||
|
||
# Default: non-streaming provider request, then normalize via chat-completions conversion.
|
||
logger.debug(
|
||
"Summarizer: using non-streaming request (%s/%s)",
|
||
summarizer_llm_config.model_endpoint_type,
|
||
summarizer_llm_config.model,
|
||
)
|
||
response_data = await llm_client.request_async_with_telemetry(req_data, summarizer_llm_config)
|
||
response = await llm_client.convert_response_to_chat_completion(
|
||
response_data,
|
||
req_messages_obj,
|
||
summarizer_llm_config,
|
||
)
|
||
if response.choices[0].message.content is None:
|
||
logger.warning("No content returned from summarizer")
|
||
raise Exception("Summary failed to generate")
|
||
return response.choices[0].message.content.strip()
|
||
|
||
request_data = llm_client.build_request_data(AgentType.letta_v1_agent, input_messages_obj, summarizer_llm_config, tools=[])
|
||
try:
|
||
summary = await _run_summarizer_request(request_data, input_messages_obj)
|
||
except Exception as e:
|
||
# handle LLM error (likely a context window exceeded error)
|
||
try:
|
||
raise llm_client.handle_llm_error(e, llm_config=llm_config)
|
||
except ContextWindowExceededError as context_error:
|
||
logger.warning(f"Context window exceeded during summarization. Applying clamping fallbacks. Original error: {context_error}")
|
||
|
||
# Fallback A: rebuild transcript with clamped tool returns to shrink payload
|
||
summary_transcript = simple_formatter(
|
||
messages,
|
||
tool_return_truncation_chars=TOOL_RETURN_TRUNCATION_CHARS,
|
||
)
|
||
logger.info(f"Full summarization payload: {request_data}")
|
||
|
||
if include_ack:
|
||
logger.info(f"Fallback summarization with ACK for model {llm_config.model}")
|
||
input_messages = [
|
||
{"role": "system", "content": system_prompt},
|
||
{"role": "assistant", "content": MESSAGE_SUMMARY_REQUEST_ACK},
|
||
{"role": "user", "content": summary_transcript},
|
||
]
|
||
else:
|
||
logger.info(f"Fallback summarization without ACK for model {llm_config.model}")
|
||
input_messages = [
|
||
{"role": "system", "content": system_prompt},
|
||
{"role": "user", "content": summary_transcript},
|
||
]
|
||
input_messages_obj = [simple_message_wrapper(msg) for msg in input_messages]
|
||
|
||
request_data = llm_client.build_request_data(
|
||
AgentType.letta_v1_agent,
|
||
input_messages_obj,
|
||
summarizer_llm_config,
|
||
tools=[],
|
||
)
|
||
|
||
try:
|
||
summary = await _run_summarizer_request(request_data, input_messages_obj)
|
||
except Exception as fallback_error_a:
|
||
# Fallback B: hard-truncate the user transcript to fit a conservative char budget
|
||
logger.warning(f"Clamped tool returns still overflowed ({fallback_error_a}). Falling back to transcript truncation.")
|
||
logger.info(f"Full fallback summarization payload: {request_data}")
|
||
|
||
# Compute a conservative char budget for the transcript based on context window
|
||
try:
|
||
budget_chars = int(summarizer_llm_config.context_window * 0.6 * 4)
|
||
except Exception:
|
||
budget_chars = 48000
|
||
|
||
overhead = len(system_prompt) + (len(MESSAGE_SUMMARY_REQUEST_ACK) if include_ack else 0) + 1024
|
||
budget_chars = max(2000, budget_chars - overhead)
|
||
|
||
truncated_transcript, _ = middle_truncate_text(summary_transcript, budget_chars=budget_chars, head_frac=0.3, tail_frac=0.3)
|
||
|
||
if include_ack:
|
||
input_messages = [
|
||
{"role": "system", "content": system_prompt},
|
||
{"role": "assistant", "content": MESSAGE_SUMMARY_REQUEST_ACK},
|
||
{"role": "user", "content": truncated_transcript},
|
||
]
|
||
else:
|
||
input_messages = [
|
||
{"role": "system", "content": system_prompt},
|
||
{"role": "user", "content": truncated_transcript},
|
||
]
|
||
input_messages_obj = [simple_message_wrapper(msg) for msg in input_messages]
|
||
|
||
request_data = llm_client.build_request_data(
|
||
AgentType.letta_v1_agent,
|
||
input_messages_obj,
|
||
summarizer_llm_config,
|
||
tools=[],
|
||
)
|
||
try:
|
||
summary = await _run_summarizer_request(request_data, input_messages_obj)
|
||
except Exception as fallback_error_b:
|
||
logger.error(f"Transcript truncation fallback also failed: {fallback_error_b}. Propagating error.")
|
||
logger.info(f"Full fallback summarization payload: {request_data}")
|
||
raise llm_client.handle_llm_error(fallback_error_b, llm_config=llm_config)
|
||
|
||
logger.info(f"Summarized {len(messages)}: {summary}")
|
||
|
||
return summary
|
||
|
||
|
||
def format_transcript(messages: List[Message], include_system: bool = False) -> List[str]:
|
||
"""
|
||
Turn a list of Message objects into a human-readable transcript.
|
||
|
||
Args:
|
||
messages: List of Message instances, in chronological order.
|
||
include_system: If True, include system-role messages. Defaults to False.
|
||
|
||
Returns:
|
||
A single string, e.g.:
|
||
user: Hey, my name is Matt.
|
||
assistant: Hi Matt! It's great to meet you...
|
||
user: What's the weather like? ...
|
||
assistant: The weather in Las Vegas is sunny...
|
||
"""
|
||
lines = []
|
||
for msg in messages:
|
||
role = msg.role.value # e.g. 'user', 'assistant', 'system', 'tool'
|
||
# skip system messages by default
|
||
if role == "system" and not include_system:
|
||
continue
|
||
|
||
# 1) Try plain content
|
||
if msg.content:
|
||
# Skip tool messages where the name is "send_message"
|
||
if msg.role == MessageRole.tool and msg.name == DEFAULT_MESSAGE_TOOL:
|
||
continue
|
||
|
||
text = "".join(c.text for c in msg.content if isinstance(c, TextContent)).strip()
|
||
# Append a compact placeholder for any images
|
||
image_count = len([c for c in msg.content if isinstance(c, ImageContent)])
|
||
if image_count > 0:
|
||
placeholder = "[Image omitted]" if image_count == 1 else f"[{image_count} images omitted]"
|
||
text = (text + (" " if text else "")) + placeholder
|
||
|
||
# 2) Otherwise, try extracting from function calls
|
||
elif msg.tool_calls:
|
||
parts = []
|
||
for call in msg.tool_calls:
|
||
args_str = call.function.arguments
|
||
if call.function.name == DEFAULT_MESSAGE_TOOL:
|
||
try:
|
||
args = json.loads(args_str)
|
||
# pull out a "message" field if present
|
||
parts.append(args.get(DEFAULT_MESSAGE_TOOL_KWARG, args_str))
|
||
except json.JSONDecodeError:
|
||
parts.append(args_str)
|
||
else:
|
||
parts.append(args_str)
|
||
text = " ".join(parts).strip()
|
||
|
||
else:
|
||
# nothing to show for this message
|
||
continue
|
||
|
||
lines.append(f"{role}: {text}")
|
||
|
||
return lines
|