run black, add isort config to pyproject.toml
This commit is contained in:
@@ -14,12 +14,7 @@ from letta.agent import Agent
|
||||
from letta.config import LettaConfig
|
||||
from letta.constants import DEFAULT_HUMAN, DEFAULT_PERSONA
|
||||
from letta.embeddings import embedding_model
|
||||
from letta.errors import (
|
||||
InvalidInnerMonologueError,
|
||||
InvalidToolCallError,
|
||||
MissingInnerMonologueError,
|
||||
MissingToolCallError,
|
||||
)
|
||||
from letta.errors import InvalidInnerMonologueError, InvalidToolCallError, MissingInnerMonologueError, MissingToolCallError
|
||||
from letta.llm_api.llm_api_tools import create
|
||||
from letta.local_llm.constants import INNER_THOUGHTS_KWARG
|
||||
from letta.schemas.agent import AgentState
|
||||
@@ -28,12 +23,7 @@ from letta.schemas.letta_message import LettaMessage, ReasoningMessage, ToolCall
|
||||
from letta.schemas.letta_response import LettaResponse
|
||||
from letta.schemas.llm_config import LLMConfig
|
||||
from letta.schemas.memory import ChatMemory
|
||||
from letta.schemas.openai.chat_completion_response import (
|
||||
ChatCompletionResponse,
|
||||
Choice,
|
||||
FunctionCall,
|
||||
Message,
|
||||
)
|
||||
from letta.schemas.openai.chat_completion_response import ChatCompletionResponse, Choice, FunctionCall, Message
|
||||
from letta.utils import get_human_text, get_persona_text, json_dumps
|
||||
from tests.helpers.utils import cleanup
|
||||
|
||||
|
||||
Reference in New Issue
Block a user