61 lines
2.1 KiB
Python
61 lines
2.1 KiB
Python
from abc import ABC, abstractmethod
|
|
from typing import AsyncGenerator
|
|
|
|
from letta.constants import DEFAULT_MAX_STEPS
|
|
from letta.log import get_logger
|
|
from letta.schemas.agent import AgentState
|
|
from letta.schemas.enums import MessageStreamStatus
|
|
from letta.schemas.letta_message import LegacyLettaMessage, LettaMessage
|
|
from letta.schemas.letta_response import LettaResponse
|
|
from letta.schemas.message import MessageCreate
|
|
from letta.schemas.user import User
|
|
|
|
|
|
class BaseAgentV2(ABC):
|
|
"""
|
|
Abstract base class for the main agent execution loop for letta agents, handling
|
|
message management, llm api request, tool execution, and context tracking.
|
|
"""
|
|
|
|
def __init__(self, agent_state: AgentState, actor: User):
|
|
self.agent_state = agent_state
|
|
self.actor = actor
|
|
self.logger = get_logger(agent_state.id)
|
|
|
|
@abstractmethod
|
|
async def build_request(
|
|
self,
|
|
input_messages: list[MessageCreate],
|
|
) -> dict:
|
|
"""
|
|
Execute the agent loop in dry_run mode, returning just the generated request
|
|
payload sent to the underlying llm provider.
|
|
"""
|
|
raise NotImplementedError
|
|
|
|
@abstractmethod
|
|
async def step(
|
|
self,
|
|
input_messages: list[MessageCreate],
|
|
max_steps: int = DEFAULT_MAX_STEPS,
|
|
) -> LettaResponse:
|
|
"""
|
|
Execute the agent loop in blocking mode, returning all messages at once.
|
|
"""
|
|
raise NotImplementedError
|
|
|
|
@abstractmethod
|
|
async def stream(
|
|
self,
|
|
input_messages: list[MessageCreate],
|
|
max_steps: int = DEFAULT_MAX_STEPS,
|
|
stream_tokens: bool = True,
|
|
) -> AsyncGenerator[LettaMessage | LegacyLettaMessage | MessageStreamStatus, None]:
|
|
"""
|
|
Execute the agent loop in streaming mode, yielding chunks as they become available.
|
|
If stream_tokens is True, individual tokens are streamed as they arrive from the LLM,
|
|
providing the lowest latency experience, otherwise each complete step (reasoning +
|
|
tool call + tool return) is yielded as it completes.
|
|
"""
|
|
raise NotImplementedError
|