* allow entering custom model name when using openai/azure
* pull models from endpoint
* added/tested vllm and azure
* no print
* make red
* make the endpoint question give you an opportunity to enter your openai api key again in case you made a mitake / want to swap it out
* add cascading workflow for openai+azure model listings
* patched bug w/ azure listing
* Update autogen.md
* in groupchat example add an azure elif
* fixed missing azure mappings + corrected the gpt-4-turbo one
* Updated MemGPT AutoGen agent to take credentials and store them in the config (allows users to use memgpt+autogen without running memgpt configure), also patched api_base kwarg for autogen >=v0.2
* add note about 0.2 testing
* added overview to autogen integration page
* default examples to openai, sync config header between the two main examples, change speaker mode to round-robin in 2-way chat to supress warning
* sync config header on last example (not used in docs)
* refactor to make sure we use existing config when writing out extra credentials
* fixed bug in local LLM where we need to comment out api_type (for pyautogen>=0.2.0)
* mark depricated API section
* CLI bug fixes for azure
* check azure before running
* Update README.md
* Update README.md
* bug fix with persona loading
* remove print
* make errors for cli flags more clear
* format
* fix imports
* fix imports
* add prints
* update lock
* update config fields
* cleanup config loading
* commit
* remove asserts
* refactor configure
* put into different functions
* add embedding default
* pass in config
* fixes
* allow overriding openai embedding endpoint
* black
* trying to patch tests (some circular import errors)
* update flags and docs
* patched support for local llms using endpoint and endpoint type passed via configs, not env vars
* missing files
* fix naming
* fix import
* fix two runtime errors
* patch ollama typo, move ollama model question pre-wrapper, modify question phrasing to include link to readthedocs, also have a default ollama model that has a tag included
* disable debug messages
* made error message for failed load more informative
* don't print dynamic linking function warning unless --debug
* updated tests to work with new cli workflow (disabled openai config test for now)
* added skips for tests when vars are missing
* update bad arg
* revise test to soft pass on empty string too
* don't run configure twice
* extend timeout (try to pass against nltk download)
* update defaults
* typo with endpoint type default
* patch runtime errors for when model is None
* catching another case of 'x in model' when model is None (preemptively)
* allow overrides to local llm related config params
* made model wrapper selection from a list vs raw input
* update test for select instead of input
* Fixed bug in endpoint when using local->openai selection, also added validation loop to manual endpoint entry
* updated error messages to be more informative with links to readthedocs
* add back gpt3.5-turbo
---------
Co-authored-by: cpacker <packercharles@gmail.com>
* Remove AsyncAgent and async from cli
Refactor agent.py memory.py
Refactor interface.py
Refactor main.py
Refactor openai_tools.py
Refactor cli/cli.py
stray asyncs
save
make legacy embeddings not use async
Refactor presets
Remove deleted function from import
* remove stray prints
* typo
* another stray print
* patch test
---------
Co-authored-by: cpacker <packercharles@gmail.com>
* various fixes to get autogen working again
* MemGPT+Autogen+Local LLM example working
Co-Authored-By: nmx0 <nmx0@users.noreply.github.com>
* propagate model to other memgpt_agent constructor
* fix agent_groupchat
* tested on lm studio
---------
Co-authored-by: nmx0 <nmx0@users.noreply.github.com>
Co-authored-by: cpacker <packercharles@gmail.com>