* trying to patch summarize when running with local llms
* moved token magic numbers to constants, made special localllm exception class (TODO catch these for retry), fix summarize bug where it exits early if empty list
* missing file
* raise an exception on no-op summary
* changed summarization logic to walk forwards in list until fraction of tokens in buffer is reached
* added same diff to sync agent
* reverted default max tokens to 8k, cleanup + more error wrapping for better error messages that get caught on retry
* patch for web UI context limit error propogation, using best guess for what the web UI error message is
* add webui token length exception
* remove print
* make no wrapper warning only pop up once
* cleanup
* Add errors to other wrappers
---------
Co-authored-by: Vivian Fang <hi@vivi.sh>
Currently, if you run the /memory command the application breaks if the LocalArchivalMemory
has no existing archival storage and defaults to the EmptyIndex. This is caused by EmptyIndex
not having a ref_doc_info implementation and throwing an Exception when that is used to print
the memory information to the console. This hot fix simply makes sure that we do not try to
use the function if using EmptyIndex and instead prints a message to the console indicating
an EmptyIndex is used.
* make tests dummy to make sure github workflow is fine
* black test
* strip circular import
* further dummy-fy the test
* use pexpect
* need y
* Update tests.yml
* Update tests.yml
* added prints
* sleep before decode print
* updated test to match legacy flow
* revising test where it fails
* comment out enter your message check for now, pexpect seems to be stuck on only setting the bootup message
* weird now it's not showing Bootup sequence complete?
* added debug
* handle none
* allow more time
* loosen string check
* add enter after commands
* modify saved compontent snippet
* add try again check
* more sendlines
* more excepts
* test passing locally
* Update tests.yml
* dont clearline
* add EOF catch that seems to only happen on github actiosn (ubuntu) but not macos
* more eof
* try flushing
* add strip_ui flag
* fix archival_memory_search and memory print output
* Don't use questionary for input if strip_ui
* Run black
* Always strip UI if TEST is set
* Add another flush
* expect Enter your message
* more debug prints
* one more shot at printing debug info
* stray fore color in stripped ui
* tests pass locally
* cleanup
---------
Co-authored-by: Vivian Fang <hi@vivi.sh>
* various fixes to get autogen working again
* MemGPT+Autogen+Local LLM example working
Co-Authored-By: nmx0 <nmx0@users.noreply.github.com>
* propagate model to other memgpt_agent constructor
* fix agent_groupchat
* tested on lm studio
---------
Co-authored-by: nmx0 <nmx0@users.noreply.github.com>
Co-authored-by: cpacker <packercharles@gmail.com>
* Added db load ability, updated README
* Fixed spacing on multiplication
* Fixed double quotes
* Added double quotes to .db
* After reading the code, realized we can return here for .db extension, and let main handle loading the .db into archival memory.
Issue is that the final else was treating the db as a file and reading it in raw, causing issues.
* Removed .csv reference, not currently working.
* Changes to lmstudio to fix JSON decode error
* black formatting
* properly handle context overflow error (propogate exception up the stack with recognizable error message) + add backwards compat option to use completions endpoint
* set max tokens to 8k, comment out the overflow policy (use memgpt's overflow policy)
* 8k not 3k
---------
Co-authored-by: Matt Poff <mattpoff@Matts-MacBook-Pro-2.local>
Co-authored-by: cpacker <packercharles@gmail.com>