From 62241e0d03b268170d8bfef4ea43a8d2ecf9fd04 Mon Sep 17 00:00:00 2001 From: Matt Zhou Date: Tue, 1 Oct 2024 17:12:09 -0700 Subject: [PATCH] Remove groq old imports --- letta/local_llm/chat_completion_proxy.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/letta/local_llm/chat_completion_proxy.py b/letta/local_llm/chat_completion_proxy.py index bdec58b6..25b91420 100644 --- a/letta/local_llm/chat_completion_proxy.py +++ b/letta/local_llm/chat_completion_proxy.py @@ -12,7 +12,6 @@ from letta.local_llm.grammars.gbnf_grammar_generator import ( create_dynamic_model_from_function, generate_gbnf_grammar_and_documentation, ) -from letta.local_llm.groq.api import get_groq_completion from letta.local_llm.koboldcpp.api import get_koboldcpp_completion from letta.local_llm.llamacpp.api import get_llamacpp_completion from letta.local_llm.llm_chat_completion_wrappers import simple_summary_wrapper @@ -170,8 +169,6 @@ def get_chat_completion( result, usage = get_ollama_completion(endpoint, auth_type, auth_key, model, prompt, context_window) elif endpoint_type == "vllm": result, usage = get_vllm_completion(endpoint, auth_type, auth_key, model, prompt, context_window, user) - elif endpoint_type == "groq": - result, usage = get_groq_completion(endpoint, auth_type, auth_key, model, prompt, context_window) else: raise LocalLLMError( f"Invalid endpoint type {endpoint_type}, please set variable depending on your backend (webui, lmstudio, llamacpp, koboldcpp)"