diff --git a/letta/server/rest_api/chat_completions_interface.py b/letta/server/rest_api/chat_completions_interface.py index ae58327f..77550a52 100644 --- a/letta/server/rest_api/chat_completions_interface.py +++ b/letta/server/rest_api/chat_completions_interface.py @@ -160,6 +160,7 @@ class ChatCompletionsStreamingInterface(AgentChunkStreamingInterface): message_id: str, message_date: datetime, expect_reasoning_content: bool = False, + name: Optional[str] = None, message_index: int = 0, ) -> None: """ diff --git a/letta/streaming_interface.py b/letta/streaming_interface.py index 6fe97286..4b2c3330 100644 --- a/letta/streaming_interface.py +++ b/letta/streaming_interface.py @@ -54,6 +54,7 @@ class AgentChunkStreamingInterface(ABC): message_id: str, message_date: datetime, expect_reasoning_content: bool = False, + name: Optional[str] = None, message_index: int = 0, ): """Process a streaming chunk from an OpenAI-compatible server""" @@ -105,6 +106,7 @@ class StreamingCLIInterface(AgentChunkStreamingInterface): message_id: str, message_date: datetime, expect_reasoning_content: bool = False, + name: Optional[str] = None, message_index: int = 0, ): assert len(chunk.choices) == 1, chunk