Files
letta-server/memgpt/autogen
Wes 2597ff2eb8 Add load and load_and_attach functions to memgpt autogen agent. (#430)
* Add load and load_and_attach functions to memgpt autogen agent.

* Only recompute files if dataset does not exist.
2023-11-14 22:51:21 -08:00
..
2023-10-20 20:00:36 -07:00

MemGPT + Autogen examples

examples/agent_groupchat.py contains an example of a groupchat where one of the agents is powered by MemGPT.

Local LLM support

In order to run MemGPT+Autogen on a local LLM, go to lines 32-55 in examples/agent_groupchat.py and fill in the config files with your local LLM's deployment details. For example, if you are using webui, it will look something like this:

config_list = [
    {
        "model": "dolphin-2.1-mistral-7b",  # this indicates the MODEL, not the WRAPPER (no concept of wrappers for AutoGen)
        "api_base": "http://127.0.0.1:5001/v1"
        "api_key": "NULL", # this is a placeholder
        "api_type": "open_ai",
    },
]
config_list_memgpt = [
    {
        "model": "airoboros-l2-70b-2.1",  # this specifies the WRAPPER MemGPT will use, not the MODEL
    },
]

config_list is used by non-MemGPT agents, which expect an OpenAI-compatible API.

config_list_memgpt is used by MemGPT agents. Currently, MemGPT interfaces with the LLM backend by exporting OPENAI_API_BASE and BACKEND_TYPE as described in Local LLM support. Note that MemGPT does not use the OpenAI-compatible API (it uses the direct API).

If you're using WebUI and want to run the non-MemGPT agents with a local LLM instead of OpenAI, enable the openai extension and point config_list's api_base to the appropriate URL (usually port 5001). Then, for MemGPT agents, export OPENAI_API_BASE and BACKEND_TYPE as described in Local LLM support (usually port 5000).