Connectors
...
PostgreSQL
Logical Replication Setup

Test Decoding Replication Setup

21min

Setup Logical Replication :: test_decoding

Test decoding is the longest running change tracking solution for PostgreSQL, and it can capture deletes which enables the soft delete function in DataLakeHouse.io.

Check to see if any logical replication slots already exist by running the command: select * from pg_catalog.pg_replication_slots;

There is almost always one default one called pghoard_local with a physical slot_type. If this is the only one that exist then you have NO logical replication slots yet.

Execute and Record the Following Settings

Review Current Settings
ο»Ώ

Ensure that:

  • WAL_LEVEL = 'logical'
  • WAL_SENDER_TIMEOUT = 0 or 60000 (0 = infinity)
  • MAX_REPLICATION_SLOTS >= 8
  • MAX_WAL_SENDERS >= 16

Create a User for DataLakeHouse.io

Replace the placeholders, schema name, etc. below with your actual values and record them for later references as you'll need them to setup the user that will have access to the database you will connect with in the DataLakeHouse.io connection info, for example we usually recommend the '' as 'datalakehouse_sync_svc':

Create a User
ο»Ώ

as an example, to create a user with access to all tables:

Pgsql
ο»Ώ

ο»Ώ

Create the Logical Replication

Creating the logical replication slots require some basic configurations on the PostgreSQL server.

Adjust a few configurations: -- Set timeout to 30 mins SET statement_timeout = '1800';

Now create the actual logical replication slot, using specifically the name 'datalakehouseio_replication_slot': SELECT pg_create_logical_replication_slot('datalakehouseio_replication_slot', 'test_decoding');

Create a ROLE and Grant permissions to the user created previously,

  • CREATE ROLE datalakehouse_sync_role;
  • ALTER ROLE datalakehouse_sync_role WITH REPLICATION;
  • GRANT datalakehouse_sync_role TO datalakehouse_sync_svc;

Other Options Acceptible

Creating a logical replication slot has the ability to add options such as "include-xids" and "skip-empty-xacts". As of 12/1/2021, DataLakeHouse requires the Xids (the default), but does not concern over skip-empty-xacts, but we suggest that leaving the default settings as-is unless otherwise communicated by our support team. There are other not well-known options for test_decoding and wal2json which unless communicated by our support team should not be used.

Analyzing Logical Replication Slots

Peeking at the logical changes are the easiest way to check if the logical replication is working and if to see if there are any tracked changes flowing through without actually changing the pointerlike when you use get_changes command:

  • SELECT count(*) FROM pg_logical_slot_peek_changes('datalakehouseio_replication_slot', null, null);

Also, retrieve the change inacted as well, as this is the same logic that DataLakeHouse.io uses, SELECT * FROM pg_logical_slot_get_changes('datalakehouseio_replication_slot', null, null); which will move the pointer of the tracking forward and clear out the change records from the WAL thus removing the disk space previously consumed so the disk does not grow out of control.

[NB]The pg_current_wal_lsn and the lsn_distance continue to change as they are based on the Log Sequence Number (LSN), a 64-bit integer used to determine the position of the WRite Ahead Log (WAL) stream which is there to preserve data integrity acting as a pointer to the XLOG. The print out is two hexadecimal numbers upt to 8-digits each, separated by a slash (/), for examplte 63/B30000220. If you compare to LSNs by using basic operator like =, >, < , βˆ’, the result is the number of bytes between the two WAL positions.

ο»Ώ

Test Your Logical Replication

In order to confirm that your logical replication is working, you can artificially create, or wait for, a DML activity of INSERT, UPDATE, DELETE.

Then run the above peek_changes command,

SQL
ο»Ώ

Get the last received WAL Position

This will show the last received and last replayed WAL positions. If there is a delta (in Bytes) beteween pg_last_wal_receive_lsn and pg_last_wal_replay_lsn, then there is a lag then data is available typically.

SQL
ο»Ώ

Confirm the Logical Replication Lag/Pointer

To quickly shown any flush confirmations from using the get_changes:

Pgsql
ο»Ώ

Which will look similar to the following output

slot_name

confirmed_flush_lsn

pg_current_wal_lsn

lsn_distance

pghoard_local

ο»Ώ

ο»Ώ

63/B00008F0

datalakehouseio_replication_slot

63/390006B0

63/B00008F0

1996489280

(2 rows)

ο»Ώ

ο»Ώ

ο»Ώ

Overall Big Picture Replication View

Pgsql
ο»Ώ

Get the Time Lag In Human Readable Format

Pgsql
ο»Ώ

General Logical Replication Slot Details

This table provides the basic information about the logic replication you have created. When DataLakeHouse.io is retrieving from the slot, the active colum will typically be set to "t" instead of "f" for true and false repsectively.

Pgsql
ο»Ώ

Peeking at Logical Slot Changes

There are several parameters for working with the pg_logical_slot_peek_changes function as further described here, https://pgpedia.info/p/pg_logical_slot_peek_changes.html, for testing the flow of WAL changes as identifying the LSNs.

Pgsql
ο»Ώ

Changes to the pg_last_wal_replay_lsn

Is not leveraged by DataLakeHouse.io as this relates to a streaming replication standby replication instance, where DataLakeHouse.io is on-demand frequency sync scheduled query-based.

Disk Space Increase is Normal

Aiven.io has a good article on standard operation increase in disk space due to WAL, https://developer.aiven.io/docs/products/postgresql/concepts/pg-disk-usage.htmlο»Ώ

Run any of the main postgreSQL commands to check disk space on the instance:

  • \l or \l+
  • SELECT pg_size_pretty(pg_database_size('yourdbname'));

Logical Replication Failues

In a standard process where the DDL of your postgreSQL tables do not change much there is very little cause for failure other than not have a sync frequency that is not aligned with the volume of data your database tables produce commensurate with the disk space of the database/server. Since Logical Replication does not track DDL changes an error could occur if a DDL change is made but downstream impacts are not considered. In the case of DataLakeHouse.io a manual change may be required on your target system in order to reflect a DDL change on your source system.

When errors, if any occur, please report them immediately by opening up a support ticket. Methods used on your source database side to clear issues may include things such as using:

  • pg_replication_origin_advance, to skip the transaction that is failing

Other References

ο»Ώ