Empty initial commit

This commit is contained in:
cpacker
2023-10-12 16:23:55 -07:00
parent 9cd5e7866a
commit 51ae4d2cb5
7 changed files with 0 additions and 327 deletions

View File

@@ -1,51 +0,0 @@
# Sample workflow for building and deploying a Jekyll site to GitHub Pages
name: Deploy Jekyll with GitHub Pages dependencies preinstalled
on:
# Runs on pushes targeting the default branch
push:
branches: ["gh-pages"]
# Allows you to run this workflow manually from the Actions tab
workflow_dispatch:
# Sets permissions of the GITHUB_TOKEN to allow deployment to GitHub Pages
permissions:
contents: read
pages: write
id-token: write
# Allow only one concurrent deployment, skipping runs queued between the run in-progress and latest queued.
# However, do NOT cancel in-progress runs as we want to allow these production deployments to complete.
concurrency:
group: "pages"
cancel-in-progress: false
jobs:
# Build job
build:
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v3
- name: Setup Pages
uses: actions/configure-pages@v3
- name: Build with Jekyll
uses: actions/jekyll-build-pages@v1
with:
source: ./
destination: ./_site
- name: Upload artifact
uses: actions/upload-pages-artifact@v2
# Deployment job
deploy:
environment:
name: github-pages
url: ${{ steps.deployment.outputs.page_url }}
runs-on: ubuntu-latest
needs: build
steps:
- name: Deploy to GitHub Pages
id: deployment
uses: actions/deploy-pages@v2

View File

@@ -1 +0,0 @@
# MemGPT

View File

@@ -1,64 +0,0 @@
.highlight-clean {
color: #313437;
background-color: #fff;
padding: 50px 0;
}
.highlight-clean p {
color: #7d8285;
}
.highlight-clean h2 {
font-weight: bold;
margin-bottom: 25px;
line-height: 1.5;
padding-top: 0;
margin-top: 0;
color: inherit;
}
.highlight-clean .intro {
font-size: 16px;
max-width: 500px;
margin: 0 auto 25px;
}
.highlight-clean .buttons {
text-align: center;
}
.highlight-clean .buttons .btn {
padding: 16px 32px;
margin: 6px;
border: none;
background: none;
box-shadow: none;
text-shadow: none;
opacity: 0.90;
/*text-transform: uppercase;*/
font-weight: bold;
font-size: 13px;
letter-spacing: 0.4px;
line-height: 1;
outline: none;
background-color: #ddd;
}
.highlight-clean .buttons .btn:hover {
opacity: 1;
}
.highlight-clean .buttons .btn:active {
transform: translateY(1px);
}
.highlight-clean .buttons .btn-primary {
background-color: #055ada;
color: #fff;
}
body {
font-family: 'Source Sans Pro', sans-serif;
padding-bottom: 50px;
}

View File

@@ -1,84 +0,0 @@
.team-clean {
color: #313437;
background-color: #fff;
}
.team-clean p {
color: #7d8285;
}
.team-clean h2 {
font-weight: bold;
margin-bottom: 40px;
padding-top: 40px;
color: inherit;
}
@media (max-width:767px) {
.team-clean h2 {
margin-bottom: 25px;
padding-top: 25px;
font-size: 24px;
}
}
.team-clean .intro {
font-size: 16px;
max-width: 500px;
margin: 0 auto;
}
.team-clean .intro p {
margin-bottom: 0;
}
.team-clean .people {
padding-bottom: 40px;
}
.team-clean .item {
text-align: center;
padding-top: 50px;
}
.team-clean .item .name {
font-weight: bold;
margin-top: 28px;
margin-bottom: 8px;
color: inherit;
}
.team-clean .item .title {
text-transform: uppercase;
font-weight: bold;
color: #d0d0d0;
letter-spacing: 2px;
font-size: 13px;
}
.team-clean .item .description {
font-size: 15px;
margin-top: 15px;
margin-bottom: 20px;
}
.team-clean .item img {
max-width: 160px;
}
.team-clean .social {
font-size: 18px;
color: #a2a8ae;
}
.team-clean .social a {
color: inherit;
margin: 0 10px;
display: inline-block;
opacity: 0.7;
}
.team-clean .social a:hover {
opacity: 1;
}

View File

Binary file not shown.

Before

Width:  |  Height:  |  Size: 213 KiB

View File

@@ -1,127 +0,0 @@
<!DOCTYPE html>
<html>
<head>
<!-- Global site tag (gtag.js) - Google Analytics -->
<script async src="https://www.googletagmanager.com/gtag/js?id=G-BFMDYSTN1L"></script>
<script>
window.dataLayer = window.dataLayer || [];
function gtag(){dataLayer.push(arguments);}
gtag('js', new Date());
gtag('config', 'G-BFMDYSTN1L');
</script>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0, shrink-to-fit=no">
<title>MemGPT</title>
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.5.0/css/bootstrap.min.css">
<link rel="stylesheet" href="https://fonts.googleapis.com/css?family=Source+Sans+Pro">
<link rel="stylesheet" href="assets/css/Highlight-Clean.css">
<link rel="stylesheet" href="assets/css/styles.css">
<link rel="stylesheet" href="assets/css/Team-Clean.css">
</head>
<body>
<div class="highlight-clean" style="padding-bottom: 10px;">
<div class="container">
<h1 class="text-center">MemGPT: Towards LLMs as Operating Systems</h1>
</div>
<div><p></p></div>
<div class="container" style="max-width: 768px;">
<div class="row">
<div class="col-md-12">
<h4 class="text-center" style="margin: 0px;">
<a class="text-center" href="https://people.eecs.berkeley.edu/~cpacker/">Charles Packer</a>,
<a class="text-center" href="https://www.vivi.sh/">Vivian Fang</a>,
<a class="text-center" href="https://shishirpatil.github.io/">Shishir G. Patil</a>, <br>
<a class="text-center" href="https://people.eecs.berkeley.edu/~kevinlin/">Kevin Lin</a>,
<a class="text-center" href="https://www.linkedin.com/in/sarah-wooders-9243b0aa/">Sarah Wooders</a>,
<a class="text-center" href="https://people.eecs.berkeley.edu/~jegonzal/">Joseph E. Gonzalez</a>
</h4>
<h4></h4>
<h5 class="text-center">UC Berkeley</h5>
</div>
</div>
</div>
<div><p></p></div>
<div class="buttons" style="margin-bottom: 8px;">
<a class="btn btn-primary" role="button" href="https://drive.google.com/file/d/1roywY0mrsJC854LlStW4hzR0UzKOml8r/view?usp=sharing">Paper</a>
<a class="btn btn-primary" role="button" href="https://discord.gg/9GEQrxmVyE">Discord</a>
<a class="btn btn-primary" role="button" href="https://github.com/cpacker/MemGPT">GitHub</a>
</div>
<div><p></p></div>
<div>
<div class="container" style="max-width: 850px;">
<div class="row">
<div class="col-md-12">
<h3 class="text-center">Teach LLMs to manage their own memory for unbounded context!</h3>
</div>
</div>
</div>
</div>
<div><p></p></div>
<div class="container" style="max-width: 768px;">
<div class="row">
<div class="col-md-12 text-center"><img src="assets/img/memgpt-system-diagram.png"
style="width: 100%;margin-bottom: 8px;" alt="MemGPT system overview">
<em>
In MemGPT, a fixed-context LLM processor is augmented with a tiered memory system and a set of functions that allow it to manage its own memory.
Main context is the (fixed-length) LLM input.
MemGPT parses the LLM text ouputs at each processing cycle, and either yields control or executes a function call, which can be used to move data between main and external context.
When the LLM generates a function call, it can request immediate return of execution to chain together functions.
In the case of a yield, the LLM will not be run again until the next external event trigger (e.g. a user message or scheduled interrupt).
</em>
</div>
</div>
</div>
</div>
<div><p></p></div>
<hr style="max-width: 768px;">
<div class="container" style="max-width: 768px;">
<div class="row">
<div class="col-md-12">
<h2>Overview</h2>
<ul>
<li>LLMs are increasingly being used for perpetual chats </li>
<li>Limited context lengths makes perpetual chat challenging</li>
<li>MemGPT manages a <mark>virtual context</mark> (inspired by virtual memory in operating systems) to create unbounded LLM context</li>
<li><mark>With MemGPT, we demonstrate that LLMs can be taught to manage their own memory!</mark>
</li>
</ul>
</div>
</div>
<div><p></p></div>
<hr style="max-width: 768px;">
<div class="container" style="max-width: 768px;">
<div class="row">
<div class="col-md-12">
<h2>Abstract</h2>
<p>
Large language models (LLMs) have revolutionized AI, but are constrained by limited context windows, hindering their utility in tasks like extended conversations and document analysis. To enable using context beyond limited context windows, we propose virtual context management, a technique drawing inspiration from hierarchical memory systems in traditional operating systems that provide the appearance of large memory resources through data movement between fast and slow memory. Using this technique, we introduce MemGPT (Memory-GPT), a system that intelligently manages different memory tiers in order to effectively provide extended context within the LLM's limited context window, and utilizes interrupts to manage control flow between itself and the user. We evaluate our OS-inspired design in two domains where the limited context windows of modern LLMs severely handicaps their performance: document analysis, where MemGPT is able to analyze large documents that far exceed the underlying LLM's context window, and multi-session chat, where MemGPT can create conversational agents that remember, reflect, and evolve dynamically through long-term interactions with their users. We release MemGPT code and data for our experiments at <a href="https://memgpt.ai" style="color: inherit;">https://memgpt.ai</a>.
<br>
</p>
</div>
</div>
</div>
<hr style="max-width: 768px;">
<div class="container" style="max-width: 768px;">
<div class="row">
<div class="col-md-12">
<h2>Citation</h2>
<code>
@inproceedings{packer2023memgpt,<br>
&nbsp; title={{MemGPT}: Towards LLMs as Operating Systems}, <br>
&nbsp; author={Packer, Charles and Fang, Vivian and Patil, Shishir G. <br>
&nbsp; and Lin, Kevin and Wooders, Sarah and Gonzalez, Joseph E.}, <br>
&nbsp; year={2023}<br>
}
</code></div>
</div>
</div>
</body>
</html>