From efe34c9db3df526e857b36939225b50bba70f97c Mon Sep 17 00:00:00 2001 From: cthomas Date: Mon, 8 Sep 2025 23:44:19 -0700 Subject: [PATCH] fix: async generator return type (#4494) --- letta/agents/letta_agent_v2.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/letta/agents/letta_agent_v2.py b/letta/agents/letta_agent_v2.py index 79081e91..db66c541 100644 --- a/letta/agents/letta_agent_v2.py +++ b/letta/agents/letta_agent_v2.py @@ -219,7 +219,7 @@ class LettaAgentV2(BaseAgentV2): use_assistant_message: bool = True, include_return_message_types: list[MessageType] | None = None, request_start_timestamp_ns: int | None = None, - ) -> AsyncGenerator[str]: + ) -> AsyncGenerator[str, None]: """ Execute the agent loop in streaming mode, yielding chunks as they become available. If stream_tokens is True, individual tokens are streamed as they arrive from the LLM, @@ -309,7 +309,7 @@ class LettaAgentV2(BaseAgentV2): request_start_timestamp_ns: int | None = None, remaining_turns: int = -1, dry_run: bool = False, - ) -> AsyncGenerator[LettaMessage | dict]: + ) -> AsyncGenerator[LettaMessage | dict, None]: """ Execute a single agent step (one LLM call and tool execution).