Co-authored-by: Charles Packer <packercharles@gmail.com> Co-authored-by: Shubham Naik <shubham.naik10@gmail.com> Co-authored-by: Shubham Naik <shub@memgpt.ai>
38 lines
2.3 KiB
Markdown
38 lines
2.3 KiB
Markdown
---
|
|
title: oobabooga web UI
|
|
excerpt: Setting up Letta with web UI
|
|
category: 6580da9a40bb410016b8b0c3
|
|
---
|
|
|
|
> 📘 web UI troubleshooting
|
|
>
|
|
> If you have problems getting web UI set up, please use the [official web UI repo for support](https://github.com/oobabooga/text-generation-webui)! There will be more answered questions about web UI there vs here on the Letta repo.
|
|
|
|
To get Letta to work with a local LLM, you need to have the LLM running on a server that takes API requests.
|
|
|
|
In this example we'll set up [oobabooga web UI](https://github.com/oobabooga/text-generation-webui#starting-the-web-ui) locally - if you're running on a remote service like Runpod, you'll want to follow Runpod specific instructions for installing web UI and determining your endpoint IP address (for example use [TheBloke's one-click UI and API](https://github.com/TheBlokeAI/dockerLLM/blob/main/README_Runpod_LocalLLMsUIandAPI.md)).
|
|
|
|
1. Install oobabooga web UI using the instructions [here](https://github.com/oobabooga/text-generation-webui#starting-the-web-ui)
|
|
2. Once installed, launch the web server with `python server.py`
|
|
3. Navigate to the web app (if local, this is probably [`http://127.0.0.1:7860`](http://localhost:7860)), select the model you want to use, adjust your GPU and CPU memory settings, and click "load"
|
|
4. If the model was loaded successfully, you should be able to access it via the API (if local, this is probably on port `5000`)
|
|
5. Assuming steps 1-4 went correctly, the LLM is now properly hosted on a port you can point Letta to!
|
|
|
|
In your terminal where you're running Letta, run `letta configure` to set the default backend for Letta to point at web UI:
|
|
|
|
```text
|
|
# if you are running web UI locally, the default IP address + port will be http://localhost:5000
|
|
? Select LLM inference provider: local
|
|
? Select LLM backend (select 'openai' if you have an OpenAI compatible proxy): webui
|
|
? Enter default endpoint: http://localhost:5000
|
|
...
|
|
```
|
|
|
|
If you have an existing agent that you want to move to the web UI backend, add extra flags to `letta run`:
|
|
|
|
```sh
|
|
letta run --agent your_agent --model-endpoint-type webui --model-endpoint http://localhost:5000
|
|
```
|
|
|
|
Text gen web UI exposes a lot of parameters that can dramatically change LLM outputs, to change these you can modify the [web UI settings file](https://github.com/cpacker/Letta/blob/main/letta/local_llm/webui/settings.py).
|