Importing an existing Postgres database

The simplest way to import a database into BigAnimal is using logical backups taken with pg_dump and loaded using pg_restore. This approach provides a way to export and import a database across different versions of Postgres, including exporting from PostgreSQL and EDB Postgres Advanced Server versions prior to 10.

The high-level steps are:

  1. Export existing roles.
  2. Import existing roles.
  3. For each database, you are migrating:
    1. logical export using pg_dump
    2. logical import with pg_restore

In case your source PostgreSQL instance hosts multiple databases, you can segment them in multiple BigAnimal clusters for easier management, better performance, increased predictability, and finer control of resources. For example, if your host has 10 databases, you can import one database and related users on a different BigAnimal cluster, one at a time.

Downtime considerations

This approach requires suspending write operations to the database application for the duration of the export/import process. You can then resume the write operations on the new system. This is because pg_dump takes an online snapshot of the source database. As a result, the changes after the backup starts aren't included in the output.

The required downtime depends on many factors, including:

  • Size of the database
  • Speed of the network between the two systems
  • Your team's familiarity with the migration procedure

To minimize the downtime, you can test the process as many times as needed before the actual migration. You can perform the export with pg_dump online, and the process is repeatable and measurable.

Before you begin

Make sure that you:

Terminology conventions

source databasepg-sourcePostgres instance from which you want to import your data.
target databasepg-targetPostgres cluster in BigAnimal where you want to import your data.
migration hostpg-migrationTemporary Linux machine in your trusted network from which to execute the export of the database and the subsequent import into BigAnimal. The migration host needs access to both the source and target databases. Or, if your source and target databases are on the same version of Postgres, the source host can serve as your migration host.

Postgres client libraries

The following client binaries must be on the migration host:

  • pg_dumpall
  • pg_dump
  • pg_restore
  • psql

They must be the same version as the Postgres version of the target database. For example, if you want to import a PostgreSQL 10 database from your private network into a PostgreSQL 14 database in BigAnimal, use the client libraries and binaries from version 14.

Access to the source and target database

Access requirements:

  • PostgreSQL superuser access to the source database. This can be the postgres user or another user with superuser privileges.
  • Access to the target database in BigAnimal as the edb_admin user.

Verify your access

  1. Connect to the source database using psql. For example:

    psql -d "host=<pg-source> user=postgres dbname=postgres"

    Replace <pg-source> with the actual hostname or IP address of the source database and the user and dbname values as appropriate. If the connection doesn't work, contact your system and database administrators to make sure that you can access the source database. This might require changes to your pg_hba.conf and network settings. If pg_hba.conf changes, reload the configuration with either SELECT pg_reload_conf(); using a psql connection or pg_ctl reload in a shell connection to the database host.

  2. Connect to the target database using the edb_admin user. For example:

    psql -d "host=<pg-target> user=edb_admin dbname=edb_admin"

    Replace <pg-target> with the actual hostname of your BigAnimal cluster.

Export existing roles

Export the existing roles from your source Postgres instance by running the following command on the migration host:

pg_dumpall -r -d "host=pg-source user=postgres dbname=postgres" > roles.sql

The generated SQL file looks like this:

-- PostgreSQL database cluster dump

SET default_transaction_read_only = off;

SET client_encoding = 'UTF8';
SET standard_conforming_strings = on;

-- Roles

--- … Your roles are here ...

-- PostgreSQL database cluster dump complete

Import the roles

  1. Your BigAnimal cluster already contains the edb_admin user, as well as the following-system required roles:
  • postgres The superuser, needed by BigAnimal to manage the cluster.

  • streaming_replica Required to manage streaming replication.

    As a result, you need to modify the roles.sql file to:

  1. Remove the lines involving the postgres user. For example, remove lines like these:

    CREATE ROLE postgres;
  2. Remove any role with superuser or replication privileges. For example, remove lines like these:

    CREATE ROLE admin;
  3. For every role that's created, grant the new role to the edb_admin user immediately after creating the user. For example:

    CREATE ROLE my_role;
    GRANT my_role TO edb_admin;
  4. Remove the NOSUPERUSER, NOCREATEROLE, NOCREATEDB, NOREPLICATION, NOBYPASSRLS permission attributes on the other users.

    The role section in the modified file, then, looks similar to:

    CREATE ROLE my_role;
    GRANT my_role TO edb_admin;
    ALTER ROLE my_role WITH INHERIT LOGIN PASSWORD 'SCRAM-SHA-256$4096:my-Scrambled-Password';
  5. From the migration host, execute:

    psql -1 -f roles.sql -d "postgres://edb_admin@<pg-target>:5432/edb_admin?sslmode=verify-full"

    Replace <pg-target> with the fully qualified domain name (FQDN) of your BigAnimal cluster.

    This command tries to create the roles in a single transaction. In case of errors, the transaction is rolled back, leaving the database cluster in the same state as before the import attempt. Enforce this behavior using the -1 option of psql.

Export a database

From the migration host, use the pg_dump command to export the source database into the target database in BigAnimal. For example:

pg_dump -Fc -d "host=pg-source user=postgres dbname=app" -f app.dump

You can use the --verbose option to monitor the progress of the operation.

The command generates a custom .dump archive (app.dump in this example), which contains the compressed dump of the source database. How long it takes the command to execute varies depending on several variables, including size of the database, network speed, disk speed, and CPU of both the source instance and the migration host. You can inspect the table of contents of the dump with pg_restore -l <db_name>.dump.

As with any other custom format dump produced with pg_dump, you can take advantage of the features that pg_restore provides you with, including:

  • Selecting a subset of the import tasks by editing the table of contents and passing it to the -L option.
  • Running the command in parallel using the -j option with the directory format.

For more information, see the pg_restore documentation.

Import a database

Use the pg_restore command and the .dump file you created when exporting the source database to import the database into BigAnimal. For example:

pg_restore -C -d  "postgres://edb_admin@pg-target:5432/edb_admin?sslmode=verify-full" app.dump

This process might take some time depending on the size of the database and the speed of the network.

In case of error, repeat the restore operation after you delete the database using the following command:

psql -d "postgres://edb_admin@pg-target:5432/edb_admin?sslmode=verify-full" \
    -c ‘DROP DATABASE app’