from datetime import datetime, timezone from typing import Dict, List from letta.errors import AgentFileExportError, AgentFileImportError from letta.log import get_logger from letta.schemas.agent import AgentState, CreateAgent from letta.schemas.agent_file import ( AgentFileSchema, AgentSchema, BlockSchema, FileAgentSchema, FileSchema, GroupSchema, ImportResult, MessageSchema, SourceSchema, ToolSchema, ) from letta.schemas.block import Block from letta.schemas.file import FileMetadata from letta.schemas.message import Message from letta.schemas.source import Source from letta.schemas.tool import Tool from letta.schemas.user import User from letta.services.agent_manager import AgentManager from letta.services.block_manager import BlockManager from letta.services.file_manager import FileManager from letta.services.file_processor.embedder.base_embedder import BaseEmbedder from letta.services.file_processor.file_processor import FileProcessor from letta.services.file_processor.parser.mistral_parser import MistralFileParser from letta.services.files_agents_manager import FileAgentManager from letta.services.group_manager import GroupManager from letta.services.mcp_manager import MCPManager from letta.services.message_manager import MessageManager from letta.services.source_manager import SourceManager from letta.services.tool_manager import ToolManager from letta.utils import get_latest_alembic_revision logger = get_logger(__name__) class AgentFileManager: """ Manages export and import of agent files between database and AgentFileSchema format. Handles: - ID mapping between database IDs and human-readable file IDs - Coordination across multiple entity managers - Transaction safety during imports - Referential integrity validation """ def __init__( self, agent_manager: AgentManager, tool_manager: ToolManager, source_manager: SourceManager, block_manager: BlockManager, group_manager: GroupManager, mcp_manager: MCPManager, file_manager: FileManager, file_agent_manager: FileAgentManager, message_manager: MessageManager, embedder: BaseEmbedder, file_parser: MistralFileParser, using_pinecone: bool = False, ): self.agent_manager = agent_manager self.tool_manager = tool_manager self.source_manager = source_manager self.block_manager = block_manager self.group_manager = group_manager self.mcp_manager = mcp_manager self.file_manager = file_manager self.file_agent_manager = file_agent_manager self.message_manager = message_manager self.embedder = embedder self.file_parser = file_parser self.using_pinecone = using_pinecone # ID mapping state for export self._db_to_file_ids: Dict[str, str] = {} # Counters for generating Stripe-style IDs self._id_counters: Dict[str, int] = { AgentSchema.__id_prefix__: 0, GroupSchema.__id_prefix__: 0, BlockSchema.__id_prefix__: 0, FileSchema.__id_prefix__: 0, SourceSchema.__id_prefix__: 0, ToolSchema.__id_prefix__: 0, MessageSchema.__id_prefix__: 0, FileAgentSchema.__id_prefix__: 0, # MCPServerSchema.__id_prefix__: 0, } def _reset_state(self): """Reset internal state for a new operation""" self._db_to_file_ids.clear() for key in self._id_counters: self._id_counters[key] = 0 def _generate_file_id(self, entity_type: str) -> str: """Generate a Stripe-style ID for the given entity type""" counter = self._id_counters[entity_type] file_id = f"{entity_type}-{counter}" self._id_counters[entity_type] += 1 return file_id def _map_db_to_file_id(self, db_id: str, entity_type: str, allow_new: bool = True) -> str: """Map a database UUID to a file ID, creating if needed (export only)""" if db_id in self._db_to_file_ids: return self._db_to_file_ids[db_id] if not allow_new: raise AgentFileExportError( f"Unexpected new {entity_type} ID '{db_id}' encountered during conversion. " f"All IDs should have been mapped during agent processing." ) file_id = self._generate_file_id(entity_type) self._db_to_file_ids[db_id] = file_id return file_id def _extract_unique_tools(self, agent_states: List[AgentState]) -> List: """Extract unique tools across all agent states by ID""" all_tools = [] for agent_state in agent_states: if agent_state.tools: all_tools.extend(agent_state.tools) unique_tools = {} for tool in all_tools: unique_tools[tool.id] = tool return sorted(unique_tools.values(), key=lambda x: x.name) def _extract_unique_blocks(self, agent_states: List[AgentState]) -> List: """Extract unique blocks across all agent states by ID""" all_blocks = [] for agent_state in agent_states: if agent_state.memory and agent_state.memory.blocks: all_blocks.extend(agent_state.memory.blocks) unique_blocks = {} for block in all_blocks: unique_blocks[block.id] = block return sorted(unique_blocks.values(), key=lambda x: x.label) async def _extract_unique_sources_and_files_from_agents( self, agent_states: List[AgentState], actor: User, files_agents_cache: dict = None ) -> tuple[List[Source], List[FileMetadata]]: """Extract unique sources and files from agent states using bulk operations""" all_source_ids = set() all_file_ids = set() for agent_state in agent_states: files_agents = await self.file_agent_manager.list_files_for_agent( agent_id=agent_state.id, actor=actor, is_open_only=False, return_as_blocks=False ) # cache the results for reuse during conversion if files_agents_cache is not None: files_agents_cache[agent_state.id] = files_agents for file_agent in files_agents: all_source_ids.add(file_agent.source_id) all_file_ids.add(file_agent.file_id) sources = await self.source_manager.get_sources_by_ids_async(list(all_source_ids), actor) files = await self.file_manager.get_files_by_ids_async(list(all_file_ids), actor, include_content=True) return sources, files async def _convert_agent_state_to_schema(self, agent_state: AgentState, actor: User, files_agents_cache: dict = None) -> AgentSchema: """Convert AgentState to AgentSchema with ID remapping""" agent_file_id = self._map_db_to_file_id(agent_state.id, AgentSchema.__id_prefix__) # use cached file-agent data if available, otherwise fetch if files_agents_cache is not None and agent_state.id in files_agents_cache: files_agents = files_agents_cache[agent_state.id] else: files_agents = await self.file_agent_manager.list_files_for_agent( agent_id=agent_state.id, actor=actor, is_open_only=False, return_as_blocks=False ) agent_schema = await AgentSchema.from_agent_state( agent_state, message_manager=self.message_manager, files_agents=files_agents, actor=actor ) agent_schema.id = agent_file_id if agent_schema.messages: for message in agent_schema.messages: message_file_id = self._map_db_to_file_id(message.id, MessageSchema.__id_prefix__) message.id = message_file_id message.agent_id = agent_file_id if agent_schema.in_context_message_ids: agent_schema.in_context_message_ids = [ self._map_db_to_file_id(message_id, MessageSchema.__id_prefix__, allow_new=False) for message_id in agent_schema.in_context_message_ids ] if agent_schema.tool_ids: agent_schema.tool_ids = [self._map_db_to_file_id(tool_id, ToolSchema.__id_prefix__) for tool_id in agent_schema.tool_ids] if agent_schema.source_ids: agent_schema.source_ids = [ self._map_db_to_file_id(source_id, SourceSchema.__id_prefix__) for source_id in agent_schema.source_ids ] if agent_schema.block_ids: agent_schema.block_ids = [self._map_db_to_file_id(block_id, BlockSchema.__id_prefix__) for block_id in agent_schema.block_ids] if agent_schema.files_agents: for file_agent in agent_schema.files_agents: file_agent.file_id = self._map_db_to_file_id(file_agent.file_id, FileSchema.__id_prefix__) file_agent.source_id = self._map_db_to_file_id(file_agent.source_id, SourceSchema.__id_prefix__) file_agent.agent_id = agent_file_id return agent_schema def _convert_tool_to_schema(self, tool) -> ToolSchema: """Convert Tool to ToolSchema with ID remapping""" tool_file_id = self._map_db_to_file_id(tool.id, ToolSchema.__id_prefix__, allow_new=False) tool_schema = ToolSchema.from_tool(tool) tool_schema.id = tool_file_id return tool_schema def _convert_block_to_schema(self, block) -> BlockSchema: """Convert Block to BlockSchema with ID remapping""" block_file_id = self._map_db_to_file_id(block.id, BlockSchema.__id_prefix__, allow_new=False) block_schema = BlockSchema.from_block(block) block_schema.id = block_file_id return block_schema def _convert_source_to_schema(self, source) -> SourceSchema: """Convert Source to SourceSchema with ID remapping""" source_file_id = self._map_db_to_file_id(source.id, SourceSchema.__id_prefix__, allow_new=False) source_schema = SourceSchema.from_source(source) source_schema.id = source_file_id return source_schema def _convert_file_to_schema(self, file_metadata) -> FileSchema: """Convert FileMetadata to FileSchema with ID remapping""" file_file_id = self._map_db_to_file_id(file_metadata.id, FileSchema.__id_prefix__, allow_new=False) file_schema = FileSchema.from_file_metadata(file_metadata) file_schema.id = file_file_id file_schema.source_id = self._map_db_to_file_id(file_metadata.source_id, SourceSchema.__id_prefix__, allow_new=False) return file_schema async def export(self, agent_ids: List[str], actor: User) -> AgentFileSchema: """ Export agents and their related entities to AgentFileSchema format. Args: agent_ids: List of agent UUIDs to export Returns: AgentFileSchema with all related entities Raises: AgentFileExportError: If export fails """ try: self._reset_state() agent_states = await self.agent_manager.get_agents_by_ids_async(agent_ids=agent_ids, actor=actor) # Validate that all requested agents were found if len(agent_states) != len(agent_ids): found_ids = {agent.id for agent in agent_states} missing_ids = [agent_id for agent_id in agent_ids if agent_id not in found_ids] raise AgentFileExportError(f"The following agent IDs were not found: {missing_ids}") # cache for file-agent relationships to avoid duplicate queries files_agents_cache = {} # Maps agent_id to list of file_agent relationships # Extract unique entities across all agents tool_set = self._extract_unique_tools(agent_states) block_set = self._extract_unique_blocks(agent_states) # Extract sources and files from agent states BEFORE conversion (with caching) source_set, file_set = await self._extract_unique_sources_and_files_from_agents(agent_states, actor, files_agents_cache) # Convert to schemas with ID remapping (reusing cached file-agent data) agent_schemas = [ await self._convert_agent_state_to_schema(agent_state, actor=actor, files_agents_cache=files_agents_cache) for agent_state in agent_states ] tool_schemas = [self._convert_tool_to_schema(tool) for tool in tool_set] block_schemas = [self._convert_block_to_schema(block) for block in block_set] source_schemas = [self._convert_source_to_schema(source) for source in source_set] file_schemas = [self._convert_file_to_schema(file_metadata) for file_metadata in file_set] logger.info(f"Exporting {len(agent_ids)} agents to agent file format") # Return AgentFileSchema with converted entities return AgentFileSchema( agents=agent_schemas, groups=[], # TODO: Extract and convert groups blocks=block_schemas, files=file_schemas, sources=source_schemas, tools=tool_schemas, # mcp_servers=[], # TODO: Extract and convert MCP servers metadata={"revision_id": await get_latest_alembic_revision()}, created_at=datetime.now(timezone.utc), ) except Exception as e: logger.error(f"Failed to export agent file: {e}") raise AgentFileExportError(f"Export failed: {e}") from e async def import_file(self, schema: AgentFileSchema, actor: User, dry_run: bool = False) -> ImportResult: """ Import AgentFileSchema into the database. Args: schema: The agent file schema to import dry_run: If True, validate but don't commit changes Returns: ImportResult with success status and details Raises: AgentFileImportError: If import fails """ try: self._reset_state() if dry_run: logger.info("Starting dry run import validation") else: logger.info("Starting agent file import") # Validate schema first self._validate_schema(schema) if dry_run: return ImportResult( success=True, message="Dry run validation passed", imported_count=0, ) # Import in dependency order imported_count = 0 file_to_db_ids = {} # Maps file IDs to new database IDs # in-memory cache for file metadata to avoid repeated db calls file_metadata_cache = {} # Maps database file ID to FileMetadata # 1. Create tools first (no dependencies) - using bulk upsert for efficiency if schema.tools: # convert tool schemas to pydantic tools pydantic_tools = [] for tool_schema in schema.tools: pydantic_tools.append(Tool(**tool_schema.model_dump(exclude={"id"}))) # bulk upsert all tools at once created_tools = await self.tool_manager.bulk_upsert_tools_async(pydantic_tools, actor) # map file ids to database ids # note: tools are matched by name during upsert, so we need to match by name here too created_tools_by_name = {tool.name: tool for tool in created_tools} for tool_schema in schema.tools: created_tool = created_tools_by_name.get(tool_schema.name) if created_tool: file_to_db_ids[tool_schema.id] = created_tool.id imported_count += 1 else: logger.warning(f"Tool {tool_schema.name} was not created during bulk upsert") # 2. Create blocks (no dependencies) - using batch create for efficiency if schema.blocks: # convert block schemas to pydantic blocks (excluding IDs to create new blocks) pydantic_blocks = [] for block_schema in schema.blocks: pydantic_blocks.append(Block(**block_schema.model_dump(exclude={"id"}))) # batch create all blocks at once created_blocks = await self.block_manager.batch_create_blocks_async(pydantic_blocks, actor) # map file ids to database ids for block_schema, created_block in zip(schema.blocks, created_blocks): file_to_db_ids[block_schema.id] = created_block.id imported_count += 1 # 3. Create sources (no dependencies) - using bulk upsert for efficiency if schema.sources: # convert source schemas to pydantic sources pydantic_sources = [] for source_schema in schema.sources: source_data = source_schema.model_dump(exclude={"id", "embedding", "embedding_chunk_size"}) pydantic_sources.append(Source(**source_data)) # bulk upsert all sources at once created_sources = await self.source_manager.bulk_upsert_sources_async(pydantic_sources, actor) # map file ids to database ids # note: sources are matched by name during upsert, so we need to match by name here too created_sources_by_name = {source.name: source for source in created_sources} for source_schema in schema.sources: created_source = created_sources_by_name.get(source_schema.name) if created_source: file_to_db_ids[source_schema.id] = created_source.id imported_count += 1 else: logger.warning(f"Source {source_schema.name} was not created during bulk upsert") # 4. Create files (depends on sources) for file_schema in schema.files: # Convert FileSchema back to FileMetadata file_data = file_schema.model_dump(exclude={"id", "content"}) # Remap source_id from file ID to database ID file_data["source_id"] = file_to_db_ids[file_schema.source_id] file_metadata = FileMetadata(**file_data) created_file = await self.file_manager.create_file(file_metadata, actor, text=file_schema.content) file_to_db_ids[file_schema.id] = created_file.id imported_count += 1 # 5. Process files for chunking/embedding (depends on files and sources) file_processor = FileProcessor( file_parser=self.file_parser, embedder=self.embedder, actor=actor, using_pinecone=self.using_pinecone, ) for file_schema in schema.files: if file_schema.content: # Only process files with content file_db_id = file_to_db_ids[file_schema.id] source_db_id = file_to_db_ids[file_schema.source_id] # Get the created file metadata (with caching) if file_db_id not in file_metadata_cache: file_metadata_cache[file_db_id] = await self.file_manager.get_file_by_id(file_db_id, actor) file_metadata = file_metadata_cache[file_db_id] # Save the db call of fetching content again file_metadata.content = file_schema.content # Process the file for chunking/embedding passages = await file_processor.process_imported_file(file_metadata=file_metadata, source_id=source_db_id) imported_count += len(passages) # 6. Create agents with empty message history for agent_schema in schema.agents: # Convert AgentSchema back to CreateAgent, remapping tool/block IDs agent_data = agent_schema.model_dump(exclude={"id", "in_context_message_ids", "messages"}) # Remap tool_ids from file IDs to database IDs if agent_data.get("tool_ids"): agent_data["tool_ids"] = [file_to_db_ids[file_id] for file_id in agent_data["tool_ids"]] # Remap block_ids from file IDs to database IDs if agent_data.get("block_ids"): agent_data["block_ids"] = [file_to_db_ids[file_id] for file_id in agent_data["block_ids"]] agent_create = CreateAgent(**agent_data) created_agent = await self.agent_manager.create_agent_async(agent_create, actor, _init_with_no_messages=True) file_to_db_ids[agent_schema.id] = created_agent.id imported_count += 1 # 7. Create messages and update agent message_ids for agent_schema in schema.agents: agent_db_id = file_to_db_ids[agent_schema.id] message_file_to_db_ids = {} # Create messages for this agent messages = [] for message_schema in agent_schema.messages: # Convert MessageSchema back to Message, setting agent_id to new DB ID message_data = message_schema.model_dump(exclude={"id"}) message_data["agent_id"] = agent_db_id # Remap agent_id to new database ID message_obj = Message(**message_data) messages.append(message_obj) # Map file ID to the generated database ID immediately message_file_to_db_ids[message_schema.id] = message_obj.id created_messages = await self.message_manager.create_many_messages_async(pydantic_msgs=messages, actor=actor) imported_count += len(created_messages) # Remap in_context_message_ids from file IDs to database IDs in_context_db_ids = [message_file_to_db_ids[message_schema_id] for message_schema_id in agent_schema.in_context_message_ids] # Update agent with the correct message_ids await self.agent_manager.update_message_ids_async(agent_id=agent_db_id, message_ids=in_context_db_ids, actor=actor) # 8. Create file-agent relationships (depends on agents and files) for agent_schema in schema.agents: if agent_schema.files_agents: agent_db_id = file_to_db_ids[agent_schema.id] # Prepare files for bulk attachment files_for_agent = [] visible_content_map = {} for file_agent_schema in agent_schema.files_agents: file_db_id = file_to_db_ids[file_agent_schema.file_id] # Use cached file metadata if available if file_db_id not in file_metadata_cache: file_metadata_cache[file_db_id] = await self.file_manager.get_file_by_id(file_db_id, actor) file_metadata = file_metadata_cache[file_db_id] files_for_agent.append(file_metadata) if file_agent_schema.visible_content: visible_content_map[file_db_id] = file_agent_schema.visible_content # Bulk attach files to agent await self.file_agent_manager.attach_files_bulk( agent_id=agent_db_id, files_metadata=files_for_agent, visible_content_map=visible_content_map, actor=actor ) imported_count += len(files_for_agent) return ImportResult( success=True, message=f"Import completed successfully. Imported {imported_count} entities.", imported_count=imported_count, id_mappings=file_to_db_ids, ) except Exception as e: logger.exception(f"Failed to import agent file: {e}") raise AgentFileImportError(f"Import failed: {e}") from e def _validate_id_format(self, schema: AgentFileSchema) -> List[str]: """Validate that all IDs follow the expected format""" errors = [] # Define entity types and their expected prefixes entity_checks = [ (schema.agents, AgentSchema.__id_prefix__), (schema.groups, GroupSchema.__id_prefix__), (schema.blocks, BlockSchema.__id_prefix__), (schema.files, FileSchema.__id_prefix__), (schema.sources, SourceSchema.__id_prefix__), (schema.tools, ToolSchema.__id_prefix__), ] for entities, expected_prefix in entity_checks: for entity in entities: if not entity.id.startswith(f"{expected_prefix}-"): errors.append(f"Invalid ID format: {entity.id} should start with '{expected_prefix}-'") else: # Check that the suffix is a valid integer try: suffix = entity.id[len(expected_prefix) + 1 :] int(suffix) except ValueError: errors.append(f"Invalid ID format: {entity.id} should have integer suffix") # Also check message IDs within agents for agent in schema.agents: for message in agent.messages: if not message.id.startswith(f"{MessageSchema.__id_prefix__}-"): errors.append(f"Invalid message ID format: {message.id} should start with '{MessageSchema.__id_prefix__}-'") else: # Check that the suffix is a valid integer try: suffix = message.id[len(MessageSchema.__id_prefix__) + 1 :] int(suffix) except ValueError: errors.append(f"Invalid message ID format: {message.id} should have integer suffix") return errors def _validate_duplicate_ids(self, schema: AgentFileSchema) -> List[str]: """Validate that there are no duplicate IDs within or across entity types""" errors = [] all_ids = set() # Check each entity type for internal duplicates and collect all IDs entity_collections = [ ("agents", schema.agents), ("groups", schema.groups), ("blocks", schema.blocks), ("files", schema.files), ("sources", schema.sources), ("tools", schema.tools), ] for entity_type, entities in entity_collections: entity_ids = [entity.id for entity in entities] # Check for duplicates within this entity type seen = set() duplicates = set() for entity_id in entity_ids: if entity_id in seen: duplicates.add(entity_id) else: seen.add(entity_id) if duplicates: errors.append(f"Duplicate {entity_type} IDs found: {duplicates}") # Check for duplicates across all entity types for entity_id in entity_ids: if entity_id in all_ids: errors.append(f"Duplicate ID across entity types: {entity_id}") all_ids.add(entity_id) # Also check message IDs within agents for agent in schema.agents: message_ids = [msg.id for msg in agent.messages] # Check for duplicates within agent messages seen = set() duplicates = set() for message_id in message_ids: if message_id in seen: duplicates.add(message_id) else: seen.add(message_id) if duplicates: errors.append(f"Duplicate message IDs in agent {agent.id}: {duplicates}") # Check for duplicates across all entity types for message_id in message_ids: if message_id in all_ids: errors.append(f"Duplicate ID across entity types: {message_id}") all_ids.add(message_id) return errors def _validate_file_source_references(self, schema: AgentFileSchema) -> List[str]: """Validate that all file source_id references exist""" errors = [] source_ids = {source.id for source in schema.sources} for file in schema.files: if file.source_id not in source_ids: errors.append(f"File {file.id} references non-existent source {file.source_id}") return errors def _validate_file_agent_references(self, schema: AgentFileSchema) -> List[str]: """Validate that all file-agent relationships reference existing entities""" errors = [] file_ids = {file.id for file in schema.files} source_ids = {source.id for source in schema.sources} {agent.id for agent in schema.agents} for agent in schema.agents: for file_agent in agent.files_agents: if file_agent.file_id not in file_ids: errors.append(f"File-agent relationship references non-existent file {file_agent.file_id}") if file_agent.source_id not in source_ids: errors.append(f"File-agent relationship references non-existent source {file_agent.source_id}") if file_agent.agent_id != agent.id: errors.append(f"File-agent relationship has mismatched agent_id {file_agent.agent_id} vs {agent.id}") return errors def _validate_schema(self, schema: AgentFileSchema): """ Validate the agent file schema for consistency and referential integrity. Args: schema: The schema to validate Raises: AgentFileImportError: If validation fails """ errors = [] # 1. ID Format Validation errors.extend(self._validate_id_format(schema)) # 2. Duplicate ID Detection errors.extend(self._validate_duplicate_ids(schema)) # 3. File Source Reference Validation errors.extend(self._validate_file_source_references(schema)) # 4. File-Agent Reference Validation errors.extend(self._validate_file_agent_references(schema)) if errors: raise AgentFileImportError(f"Schema validation failed: {'; '.join(errors)}") logger.info("Schema validation passed")