* wait I forgot to comit locally * cp the entire core directory and then rm the .git subdir
61 lines
1.7 KiB
Python
61 lines
1.7 KiB
Python
from letta_client import Letta
|
|
|
|
client = Letta(base_url="http://localhost:8283")
|
|
|
|
# list available models
|
|
models = client.models.list_llms()
|
|
for model in models:
|
|
print(f"Provider {model.model_endpoint_type} model {model.model}: {model.handle}")
|
|
|
|
# list available embedding models
|
|
embedding_models = client.models.list_embedding_models()
|
|
for model in embedding_models:
|
|
print(f"Provider {model.handle}")
|
|
|
|
# openai
|
|
openai_agent = client.agents.create(
|
|
model="openai/gpt-4o-mini",
|
|
embedding="openai/text-embedding-3-small",
|
|
# optional configuration
|
|
context_window_limit=16000,
|
|
embedding_chunk_size=300,
|
|
)
|
|
|
|
# Azure OpenAI
|
|
azure_openai_agent = client.agents.create(
|
|
model="azure/gpt-4o-mini",
|
|
embedding="azure/text-embedding-3-small",
|
|
# optional configuration
|
|
context_window_limit=16000,
|
|
embedding_chunk_size=300,
|
|
)
|
|
|
|
# anthropic
|
|
anthropic_agent = client.agents.create(
|
|
model="anthropic/claude-3-5-sonnet-20241022",
|
|
# note: anthropic does not support embeddings so you will need another provider
|
|
embedding="openai/text-embedding-3-small",
|
|
# optional configuration
|
|
context_window_limit=16000,
|
|
embedding_chunk_size=300,
|
|
)
|
|
|
|
# Groq
|
|
groq_agent = client.agents.create(
|
|
model="groq/llama-3.3-70b-versatile",
|
|
# note: groq does not support embeddings so you will need another provider
|
|
embedding="openai/text-embedding-3-small",
|
|
# optional configuration
|
|
context_window_limit=16000,
|
|
embedding_chunk_size=300,
|
|
)
|
|
|
|
# Ollama
|
|
ollama_agent = client.agents.create(
|
|
model="ollama/thewindmom/hermes-3-llama-3.1-8b:latest",
|
|
embedding="ollama/mxbai-embed-large:latest",
|
|
# optional configuration
|
|
context_window_limit=16000,
|
|
embedding_chunk_size=300,
|
|
)
|