Configuring storage backends

Customizing the MemGPT storage backend


Switching storage backends

MemGPT can only use one storage backend at a time. If you switch from local to database storage, you will need to re-load data and start agents from scratch. We currently do not support migrating between storage backends.

MemGPT supports both local and database storage for archival memory. You can configure which storage backend to use via memgpt configure. For larger datasets, we recommend using a database backend.


MemGPT will default to using local storage (saved at ~/.memgpt/archival/ for loaded data sources, and ~/.memgpt/agents/ for agent storage).


In order to use the Postgres backend, you must have a running Postgres database that MemGPT can write to. You can enable the Postgres backend by running memgpt configure and selecting postgres for archival storage, which will then prompt for the database URI (e.g. postgresql+pg8000://<USER>:<PASSWORD>@<IP>:5432/<DB_NAME>). To enable the Postgres backend, make sure to install the required dependencies with:

pip install 'pymemgpt[postgres]'

Running Postgres

To run the Postgres backend, you will need a URI to a Postgres database that supports pgvector. Follow these steps to set up and run your Postgres server easily with Docker:

  1. Install Docker

  2. Give the script permissions to execute:

chmod +x db/
  1. Configure the environment for pgvector. You can either:

    • Add the following line to your shell profile (e.g., ~/.bashrc, ~/.zshrc):

      export PGVECTOR_TEST_DB_URL=postgresql+pg8000://memgpt:memgpt@localhost:8888/memgpt
    • Or create a .env file in the root project directory with:

  2. Run the script from the root project directory:

bash db/

Note: You can either use a hosted provider or install pgvector. You do not need to do this manually if you use our Docker container, however.


You can configure Chroma with both the HTTP and persistent storage client via memgpt configure. You will need to specify either a persistent storage path or host/port dependending on your client choice. The example below shows how to configure Chroma with local persistent storage:

? Select LLM inference provider: openai
? Override default endpoint:
? Select default model (recommended: gpt-4): gpt-4
? Select embedding provider: openai
? Select default preset: memgpt_chat
? Select default persona: sam_pov
? Select default human: cs_phd
? Select storage backend for archival data: chroma
? Select chroma backend: persistent
? Enter persistent storage location: /Users/sarahwooders/.memgpt/config/chroma


You have to enable the LanceDB backend by running

memgpt configure

and selecting lancedb for archival storage, and database URI (e.g. ./.lancedb"), Empty archival uri is also handled and default uri is set at ./.lancedb. For more checkout lancedb docs


To enable the Milvus backend, make sure to install the required dependencies with:

pip install 'pymemgpt[milvus]'

You can configure Milvus connection via command memgpt configure.

? Select storage backend for archival data: milvus
? Enter the Milvus connection URI (Default: ~/.memgpt/milvus.db): ~/.memgpt/milvus.db

You just set the URI to the local file path, e.g. ~/.memgpt/milvus.db, which will automatically invoke the local Milvus service instance through Milvus Lite.

If you have large scale of data such as more than a million docs, we recommend setting up a more performant Milvus server on docker or kubenetes. And in this case, your URI should be the server URI, e.g. http://localhost:19530.