feat: offload llama index init logic (#8387)

This commit is contained in:
cthomas
2026-01-07 11:58:51 -08:00
committed by Caren Thomas
parent 6e64b59d8a
commit a775bbfdf2

View File

@@ -50,8 +50,10 @@ class FileProcessor:
"""Chunk text and generate embeddings with fallback to default chunker if needed"""
filename = file_metadata.file_name
# Create file-type-specific chunker
text_chunker = LlamaIndexChunker(file_type=file_metadata.file_type, chunk_size=self.embedder.embedding_config.embedding_chunk_size)
# Create file-type-specific chunker in thread pool to avoid blocking event loop
text_chunker = await asyncio.to_thread(
LlamaIndexChunker, file_type=file_metadata.file_type, chunk_size=self.embedder.embedding_config.embedding_chunk_size
)
# First attempt with file-specific chunker
try: