Files
letta-server/README.md
Charles Packer d6f6ea65e7 Update README.md
2023-10-15 16:53:44 -07:00

107 lines
3.7 KiB
Markdown

<a href="#user-content-memgpt"><img src="https://memgpt.ai/assets/img/memgpt_logo_circle.png" alt="MemGPT logo" width="75" align="right"></a>
# [MemGPT](https://memgpt.ai)
<div align="center">
[![Website](https://img.shields.io/badge/Website-Visit-008000?logo=firefox-browser&style=flat-square)](https://memgpt.ai)
[![Discord](https://img.shields.io/discord/1161736243340640419?label=Discord&logo=discord&logoColor=5865F2&style=flat-square&color=5865F2)](https://discord.gg/9GEQrxmVyE)
[![arXiv 2310.08560](https://img.shields.io/badge/arXiv-2310.08560-B31B1B?logo=arxiv&style=flat-square)](https://arxiv.org/abs/2310.08560)
Teaching LLMs memory management for unbounded context
</div>
<details open>
<summary><h2>Create perpetual chatbots with self-editing memory!</h1></summary>
<div align="center">
<img src="https://memgpt.ai/assets/img/demo.gif" alt="MemGPT demo video" width="800">
</div>
</details>
<details open>
<summary><h2>Chat with your data - try it with the <a href="https://docs.llamaindex.ai/en/stable/">LlamaIndex API docs</a>!</h1></summary>
<div align="center">
<img src="https://memgpt.ai/assets/img/docqa_demo.gif" alt="MemGPT demo video for llamaindex api docs search" width="800">
</div>
</details>
## Quick setup
Install dependencies:
```sh
pip install -r requirements.txt
```
Add your OpenAI API key to your environment:
```sh
export OPENAI_API_KEY=YOUR_API_KEY
```
## What is MemGPT?
MemoryGPT (or MemGPT in short) is a system that intelligently manages different memory tiers in LLMs in order to effectively provide extended context within the LLM's limited context window. For example, MemGPT knows when to push critical information to a vector database and when to retrieve it later in the chat, enabling perpetual conversations. Learn more about MemGPT in our [paper](https://arxiv.org/abs/2310.08560).
## Try MemGPT in your CLI
To run MemGPT for as a conversation agent in CLI mode, simply run `main.py`:
```sh
python3 main.py
```
To create a new starter user or starter persona (that MemGPT gets initialized with), create a new `.txt` file in [/memgpt/humans/examples](/memgpt/humans/examples) or [/memgpt/personas/examples](/memgpt/personas/examples), then use the `--persona` or `--human` flag when running `main.py`. For example:
```sh
# assuming you created a new file /memgpt/humans/examples/me.txt
python main.py --human me.txt
```
### `main.py` flags
```text
--persona
load a specific persona file
--human
load a specific human file
--first
allows you to send the first message in the chat (by default, MemGPT will send the first message)
--debug
enables debugging output
```
### Interactive CLI commands
While using MemGPT via the CLI you can run various commands:
```text
/exit
exit the CLI
/save
save a checkpoint of the current agent/conversation state
/load
load a saved checkpoint
/dump
view the current message log (see the contents of main context)
/memory
print the current contents of agent memory
/pop
undo the last message in the conversation
/heartbeat
send a heartbeat system message to the agent
/memorywarning
send a memory warning system message to the agent
```
### Support
* By default MemGPT will use `gpt-4`, so your API key will require `gpt-4` API access.
If you have any further questions, or have anything to share, we are excited to hear your feedback!
* For issues and feature requests, please [open a GitHub issue](https://github.com/cpacker/MemGPT/issues).
### Datasets
Datasets used in our [paper](https://arxiv.org/abs/2310.08560) can be downloaded at [HuggingFace](https://huggingface.co/MemGPT).