fix: patch errors with OllamaProvider (#1875)

This commit is contained in:
Charles Packer
2024-10-12 20:19:20 -07:00
committed by GitHub
parent 9d0da9549b
commit 1099809e49
4 changed files with 65 additions and 17 deletions

View File

@@ -85,9 +85,7 @@ def get_chat_completion(
elif wrapper is None:
# Warn the user that we're using the fallback
if not has_shown_warning:
print(
f"{CLI_WARNING_PREFIX}no wrapper specified for local LLM, using the default wrapper (you can remove this warning by specifying the wrapper with --model-wrapper)"
)
print(f"{CLI_WARNING_PREFIX}no prompt formatter specified for local LLM, using the default formatter")
has_shown_warning = True
llm_wrapper = DEFAULT_WRAPPER()