Configuring storage backends
Customizing the MemGPT storage backend
Switching storage backends
MemGPT can only use one storage backend at a time. If you switch from local to database storage, you will need to re-load data and start agents from scratch. We currently do not support migrating between storage backends.
MemGPT supports both local and database storage for archival memory. You can configure which storage backend to use via memgpt configure
. For larger datasets, we recommend using a database backend.
Local
MemGPT will default to using local storage (saved at ~/.memgpt/archival/
for loaded data sources, and ~/.memgpt/agents/
for agent storage).
Postgres
In order to use the Postgres backend, you must have a running Postgres database that MemGPT can write to. You can enable the Postgres backend by running memgpt configure
and selecting postgres
for archival storage, which will then prompt for the database URI (e.g. postgresql+pg8000://<USER>:<PASSWORD>@<IP>:5432/<DB_NAME>
). To enable the Postgres backend, make sure to install the required dependencies with:
pip install 'pymemgpt[postgres]'
Running Postgres
To run the Postgres backend, you will need a URI to a Postgres database that supports pgvector. Follow these steps to set up and run your Postgres server easily with Docker:
-
Give the
run_postgres.sh
script permissions to execute:
chmod +x db/run_postgres.sh
-
Configure the environment for
pgvector
. You can either:-
Add the following line to your shell profile (e.g.,
~/.bashrc
,~/.zshrc
):export PGVECTOR_TEST_DB_URL=postgresql+pg8000://memgpt:memgpt@localhost:8888/memgpt
-
Or create a
.env
file in the root project directory with:PGVECTOR_TEST_DB_URL=postgresql+pg8000://memgpt:memgpt@localhost:8888/memgpt
-
-
Run the script from the root project directory:
bash db/run_postgres.sh
Note: You can either use a hosted provider or install pgvector. You do not need to do this manually if you use our Docker container, however.
Chroma
You can configure Chroma with both the HTTP and persistent storage client via memgpt configure
. You will need to specify either a persistent storage path or host/port dependending on your client choice. The example below shows how to configure Chroma with local persistent storage:
? Select LLM inference provider: openai
? Override default endpoint: https://api.openai.com/v1
? Select default model (recommended: gpt-4): gpt-4
? Select embedding provider: openai
? Select default preset: memgpt_chat
? Select default persona: sam_pov
? Select default human: cs_phd
? Select storage backend for archival data: chroma
? Select chroma backend: persistent
? Enter persistent storage location: /Users/sarahwooders/.memgpt/config/chroma
LanceDB
You have to enable the LanceDB backend by running
memgpt configure
and selecting lancedb
for archival storage, and database URI (e.g. ./.lancedb
"), Empty archival uri is also handled and default uri is set at ./.lancedb
. For more checkout lancedb docs
Milvus
To enable the Milvus backend, make sure to install the required dependencies with:
pip install 'pymemgpt[milvus]'
You can configure Milvus connection via command memgpt configure
.
...
? Select storage backend for archival data: milvus
? Enter the Milvus connection URI (Default: ~/.memgpt/milvus.db): ~/.memgpt/milvus.db
You just set the URI to the local file path, e.g. ~/.memgpt/milvus.db
, which will automatically invoke the local Milvus service instance through Milvus Lite.
If you have large scale of data such as more than a million docs, we recommend setting up a more performant Milvus server on docker or kubenetes. And in this case, your URI should be the server URI, e.g. http://localhost:19530
.
Updated about 1 month ago