* Relax verify_first_message_correctness to accept any function call * Also allow missing internal monologue if request_heartbeat * Cleanup * get instead of raw dict access * Support attach in memgpt autogen agent * Add docs example * Add documentation, cleanup
89 lines
3.1 KiB
Python
89 lines
3.1 KiB
Python
"""Example of how to add MemGPT into an AutoGen groupchat and chat with docs.
|
|
|
|
See https://memgpt.readthedocs.io/en/latest/autogen/#loading-documents
|
|
|
|
Based on the official AutoGen example here: https://github.com/microsoft/autogen/blob/main/notebook/agentchat_groupchat.ipynb
|
|
|
|
Begin by doing:
|
|
pip install "pyautogen[teachable]"
|
|
pip install pymemgpt
|
|
or
|
|
pip install -e . (inside the MemGPT home directory)
|
|
"""
|
|
|
|
|
|
import os
|
|
import autogen
|
|
from memgpt.autogen.memgpt_agent import create_autogen_memgpt_agent, create_memgpt_autogen_agent_from_config
|
|
|
|
# This config is for autogen agents that are not powered by MemGPT
|
|
config_list = [
|
|
{
|
|
"model": "gpt-4",
|
|
"api_key": os.getenv("OPENAI_API_KEY"),
|
|
}
|
|
]
|
|
|
|
# This config is for autogen agents that powered by MemGPT
|
|
config_list_memgpt = [
|
|
{
|
|
"model": "gpt-4",
|
|
},
|
|
]
|
|
|
|
USE_AUTOGEN_WORKFLOW = True
|
|
|
|
# Set to True if you want to print MemGPT's inner workings.
|
|
DEBUG = False
|
|
|
|
interface_kwargs = {
|
|
"debug": DEBUG,
|
|
"show_inner_thoughts": DEBUG,
|
|
"show_function_outputs": DEBUG,
|
|
}
|
|
|
|
llm_config = {"config_list": config_list, "seed": 42}
|
|
llm_config_memgpt = {"config_list": config_list_memgpt, "seed": 42}
|
|
|
|
# The user agent
|
|
user_proxy = autogen.UserProxyAgent(
|
|
name="User_proxy",
|
|
system_message="A human admin.",
|
|
code_execution_config={"last_n_messages": 2, "work_dir": "groupchat"},
|
|
human_input_mode="TERMINATE", # needed?
|
|
default_auto_reply="...", # Set a default auto-reply message here (non-empty auto-reply is required for LM Studio)
|
|
)
|
|
|
|
# In our example, we swap this AutoGen agent with a MemGPT agent
|
|
# This MemGPT agent will have all the benefits of MemGPT, ie persistent memory, etc.
|
|
if not USE_AUTOGEN_WORKFLOW:
|
|
coder = create_autogen_memgpt_agent(
|
|
"MemGPT_coder",
|
|
persona_description="I am a 10x engineer, trained in Python. I was the first engineer at Uber "
|
|
"(which I make sure to tell everyone I work with).",
|
|
user_description=f"You are participating in a group chat with a user ({user_proxy.name}) " f"and a product manager ({pm.name}).",
|
|
model=config_list_memgpt[0]["model"],
|
|
interface_kwargs=interface_kwargs,
|
|
)
|
|
else:
|
|
coder = create_memgpt_autogen_agent_from_config(
|
|
"MemGPT_coder",
|
|
llm_config=llm_config_memgpt,
|
|
system_message=f"I am a 10x engineer, trained in Python. I was the first engineer at Uber "
|
|
f"(which I make sure to tell everyone I work with).\n"
|
|
f"You are participating in a group chat with a user ({user_proxy.name}).",
|
|
interface_kwargs=interface_kwargs,
|
|
)
|
|
coder.attach("memgpt_research_paper") # See https://memgpt.readthedocs.io/en/latest/autogen/#loading-documents
|
|
|
|
# Initialize the group chat between the user and two LLM agents (PM and coder)
|
|
groupchat = autogen.GroupChat(agents=[user_proxy, coder], messages=[], max_round=12)
|
|
manager = autogen.GroupChatManager(groupchat=groupchat, llm_config=llm_config)
|
|
|
|
|
|
# Begin the group chat with a message from the user
|
|
user_proxy.initiate_chat(
|
|
manager,
|
|
message="Tell me what a virtual context in MemGPT is. Search your archival memory.",
|
|
)
|