Files
letta-server/letta/schemas/providers/groq.py
jnjpng 350f3a751c fix: update more plaintext non async callsites (#7223)
* bae

* update

* fix

* clean up

* last
2025-12-17 17:31:02 -08:00

38 lines
1.6 KiB
Python

from typing import Literal
from pydantic import Field
from letta.schemas.enums import ProviderCategory, ProviderType
from letta.schemas.llm_config import LLMConfig
from letta.schemas.providers.openai import OpenAIProvider
class GroqProvider(OpenAIProvider):
provider_type: Literal[ProviderType.groq] = Field(ProviderType.groq, description="The type of the provider.")
provider_category: ProviderCategory = Field(ProviderCategory.base, description="The category of the provider (base or byok)")
base_url: str = "https://api.groq.com/openai/v1"
api_key: str | None = Field(None, description="API key for the Groq API.", deprecated=True)
async def list_llm_models_async(self) -> list[LLMConfig]:
from letta.llm_api.openai import openai_get_model_list_async
api_key = await self.api_key_enc.get_plaintext_async() if self.api_key_enc else None
response = await openai_get_model_list_async(self.base_url, api_key=api_key)
configs = []
for model in response["data"]:
if "context_window" not in model:
continue
configs.append(
LLMConfig(
model=model["id"],
model_endpoint_type="groq",
model_endpoint=self.base_url,
context_window=model["context_window"],
handle=self.get_handle(model["id"]),
max_tokens=self.get_default_max_output_tokens(model["id"]),
provider_name=self.name,
provider_category=self.provider_category,
)
)
return configs