postgresql/src/backend/replication
Tomas Vondra c91f71b9dc Fix publish_as_relid with multiple publications
Commit 83fd4532a7 allowed publishing of changes via ancestors, for
publications defined with publish_via_partition_root. But the way
the ancestor was determined in get_rel_sync_entry() was incorrect,
simply updating the same variable. So with multiple publications,
replicating different ancestors, the outcome depended on the order
of publications in the list - the value from the last loop was used,
even if it wasn't the top-most ancestor.

This is a probably rare situation, as in most cases publications do
not overlap, so each partition has exactly one candidate ancestor
to replicate as and there's no ambiguity.

Fixed by tracking the "ancestor level" for each publication, and
picking the top-most ancestor. Adds a test case, verifying the
correct ancestor is used for publishing the changes and that this
does not depend on order of publications in the list.

Older releases have another bug in this loop - once all actions are
replicated, the loop is terminated, on the assumption that inspecting
additional publications is unecessary. But that misses the fact that
those additional applications may replicate different ancestors.

Fixed by removal of this break condition. We might still terminate the
loop in some cases (e.g. when replicating all actions and the ancestor
is the partition root).

Backpatch to 13, where publish_via_partition_root was introduced.

Initial report and fix by me, test added by Hou zj. Reviews and
improvements by Amit Kapila.

Author: Tomas Vondra, Hou zj, Amit Kapila
Reviewed-by: Amit Kapila, Hou zj
Discussion: https://postgr.es/m/d26d24dd-2fab-3c48-0162-2b7f84a9c893%40enterprisedb.com
2022-03-16 18:05:58 +01:00
..
libpqwalreceiver Update copyright for 2022 2022-01-07 19:04:57 -05:00
logical Fix compiler warning introduced in commit 705e20f855. 2022-03-15 08:11:17 +05:30
pgoutput Fix publish_as_relid with multiple publications 2022-03-16 18:05:58 +01:00
.gitignore Support multiple synchronous standby servers. 2016-04-06 17:18:25 +09:00
backup_manifest.c Improve error handling of cryptohash computations 2022-01-11 09:55:16 +09:00
basebackup.c Allow extensions to add new backup targets. 2022-03-15 13:22:04 -04:00
basebackup_copy.c Remove server support for the previous base backup protocol. 2022-02-10 12:12:43 -05:00
basebackup_gzip.c pg_basebackup: Add a dummy return to bbsink_gzip_new(). 2022-01-27 14:20:18 -05:00
basebackup_lz4.c Change HAVE_LIBLZ4 and HAVE_LIBZSTD tests to USE_LZ4 and USE_ZSTD. 2022-03-15 13:06:25 -04:00
basebackup_progress.c Update copyright for 2022 2022-01-07 19:04:57 -05:00
basebackup_server.c Fix server crash bug in 'server' backup target. 2022-02-02 13:50:33 -05:00
basebackup_sink.c Fix collection of typos in the code and the documentation 2022-03-15 11:29:35 +09:00
basebackup_target.c Suppress compiler warnings. 2022-03-16 09:26:48 -04:00
basebackup_throttle.c Update copyright for 2022 2022-01-07 19:04:57 -05:00
basebackup_zstd.c Change HAVE_LIBLZ4 and HAVE_LIBZSTD tests to USE_LZ4 and USE_ZSTD. 2022-03-15 13:06:25 -04:00
Makefile Allow extensions to add new backup targets. 2022-03-15 13:22:04 -04:00
README code: replace 'master' with 'primary' where appropriate. 2020-07-08 12:57:23 -07:00
repl_gram.y Remove server support for old BASE_BACKUP command syntax. 2022-02-10 10:48:33 -05:00
repl_scanner.l Remove server support for old BASE_BACKUP command syntax. 2022-02-10 10:48:33 -05:00
slot.c Fix warning on mingw due to pid_t width, introduced in fe0972ee5e. 2022-02-26 16:07:07 -08:00
slotfuncs.c Create routine able to set single-call SRFs for Materialize mode 2022-03-07 10:26:29 +09:00
syncrep.c Update copyright for 2022 2022-01-07 19:04:57 -05:00
syncrep_gram.y Update copyright for 2022 2022-01-07 19:04:57 -05:00
syncrep_scanner.l Update copyright for 2022 2022-01-07 19:04:57 -05:00
walreceiver.c Split xlog.c into xlog.c and xlogrecovery.c. 2022-02-16 09:30:38 +02:00
walreceiverfuncs.c Split xlog.c into xlog.c and xlogrecovery.c. 2022-02-16 09:30:38 +02:00
walsender.c Create routine able to set single-call SRFs for Materialize mode 2022-03-07 10:26:29 +09:00

src/backend/replication/README

Walreceiver - libpqwalreceiver API
----------------------------------

The transport-specific part of walreceiver, responsible for connecting to
the primary server, receiving WAL files and sending messages, is loaded
dynamically to avoid having to link the main server binary with libpq.
The dynamically loaded module is in libpqwalreceiver subdirectory.

The dynamically loaded module implements a set of functions with details
about each one of them provided in src/include/replication/walreceiver.h.

This API should be considered internal at the moment, but we could open it
up for 3rd party replacements of libpqwalreceiver in the future, allowing
pluggable methods for receiving WAL.

Walreceiver IPC
---------------

When the WAL replay in startup process has reached the end of archived WAL,
restorable using restore_command, it starts up the walreceiver process
to fetch more WAL (if streaming replication is configured).

Walreceiver is a postmaster subprocess, so the startup process can't fork it
directly. Instead, it sends a signal to postmaster, asking postmaster to launch
it. Before that, however, startup process fills in WalRcvData->conninfo
and WalRcvData->slotname, and initializes the starting point in
WalRcvData->receiveStart.

As walreceiver receives WAL from the primary server, and writes and flushes
it to disk (in pg_wal), it updates WalRcvData->flushedUpto and signals
the startup process to know how far WAL replay can advance.

Walreceiver sends information about replication progress to the primary server
whenever it either writes or flushes new WAL, or the specified interval elapses.
This is used for reporting purpose.

Walsender IPC
-------------

At shutdown, postmaster handles walsender processes differently from regular
backends. It waits for regular backends to die before writing the
shutdown checkpoint and terminating pgarch and other auxiliary processes, but
that's not desirable for walsenders, because we want the standby servers to
receive all the WAL, including the shutdown checkpoint, before the primary
is shut down. Therefore postmaster treats walsenders like the pgarch process,
and instructs them to terminate at PM_SHUTDOWN_2 phase, after all regular
backends have died and checkpointer has issued the shutdown checkpoint.

When postmaster accepts a connection, it immediately forks a new process
to handle the handshake and authentication, and the process initializes to
become a backend. Postmaster doesn't know if the process becomes a regular
backend or a walsender process at that time - that's indicated in the
connection handshake - so we need some extra signaling to let postmaster
identify walsender processes.

When walsender process starts up, it marks itself as a walsender process in
the PMSignal array. That way postmaster can tell it apart from regular
backends.

Note that no big harm is done if postmaster thinks that a walsender is a
regular backend; it will just terminate the walsender earlier in the shutdown
phase. A walsender will look like a regular backend until it's done with the
initialization and has marked itself in PMSignal array, and at process
termination, after unmarking the PMSignal slot.

Each walsender allocates an entry from the WalSndCtl array, and tracks
information about replication progress. User can monitor them via
statistics views.


Walsender - walreceiver protocol
--------------------------------

See manual.