From 4f30288c14c8c71ff4c647a0385d873cecfbf060 Mon Sep 17 00:00:00 2001 From: Sarah Wooders Date: Sat, 9 Nov 2024 14:46:39 -0800 Subject: [PATCH] fix: fix `Unsupport authentication type` error for Ollama (#2018) --- letta/local_llm/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/letta/local_llm/utils.py b/letta/local_llm/utils.py index 2b2c153b..a5381159 100644 --- a/letta/local_llm/utils.py +++ b/letta/local_llm/utils.py @@ -18,7 +18,7 @@ def post_json_auth_request(uri, json_payload, auth_type, auth_key): """Send a POST request with a JSON payload and optional authentication""" # By default most local LLM inference servers do not have authorization enabled - if auth_type is None: + if auth_type is None or auth_type == "": response = requests.post(uri, json=json_payload) # Used by OpenAI, together.ai, Mistral AI