postgresql/src/backend/replication
Tom Lane d7ab908fba Distinguish wait-for-connection from wait-for-write-ready on Windows.
The API for WaitLatch and friends followed the Unix convention in which
waiting for a socket connection to complete is identical to waiting for
the socket to accept a write.  While Windows provides a select(2)
emulation that agrees with that, the native WaitForMultipleObjects API
treats them as quite different --- and for some bizarre reason, it will
report a not-yet-connected socket as write-ready.  libpq itself has so
far escaped dealing with this because it waits with select(), but in
libpqwalreceiver.c we want to wait using WaitLatchOrSocket.  The semantics
mismatch resulted in replication connection failures on Windows, but only
for remote connections (apparently, localhost connections complete
immediately, or at least too fast for anyone to have noticed the problem
in single-machine testing).

To fix, introduce an additional WL_SOCKET_CONNECTED wait flag for
WaitLatchOrSocket, which is identical to WL_SOCKET_WRITEABLE on
non-Windows, but results in waiting for FD_CONNECT events on Windows.

Ideally, we would also distinguish the two conditions in the API for
PQconnectPoll(), but changing that API at this point seems infeasible.
Instead, cheat by checking for PQstatus() == CONNECTION_STARTED to
determine that we're still waiting for the connection to complete.
(This is a cheat mainly because CONNECTION_STARTED is documented as an
internal state rather than something callers should rely on.  Perhaps
we ought to change the documentation ... but this patch doesn't.)

Per reports from Jobin Augustine and Igor Neyman.  Back-patch to v10
where commit 1e8a85009 exposed this longstanding shortcoming.

Andres Freund, minor fix and some code review/beautification by me

Discussion: https://postgr.es/m/CAHBggj8g2T+ZDcACZ2FmzX9CTxkWjKBsHd6NkYB4i9Ojf6K1Fw@mail.gmail.com
2017-08-15 11:08:25 -04:00
..
libpqwalreceiver Distinguish wait-for-connection from wait-for-write-ready on Windows. 2017-08-15 11:08:25 -04:00
logical Final pgindent + perltidy run for v10. 2017-08-14 17:29:33 -04:00
pgoutput Phase 3 of pgindent updates. 2017-06-21 15:35:54 -04:00
.gitignore Support multiple synchronous standby servers. 2016-04-06 17:18:25 +09:00
basebackup.c Phase 3 of pgindent updates. 2017-06-21 15:35:54 -04:00
Makefile Rethink flex flags for syncrep_scanner.l. 2017-05-19 18:05:20 -04:00
README Rename "pg_xlog" directory to "pg_wal". 2016-10-20 11:32:18 -04:00
repl_gram.y Logical replication support for initial data copy 2017-03-23 08:55:37 -04:00
repl_scanner.l Logical replication support for initial data copy 2017-03-23 08:55:37 -04:00
slot.c Final pgindent + perltidy run for v10. 2017-08-14 17:29:33 -04:00
slotfuncs.c Fix race conditions in replication slot operations 2017-07-25 13:26:49 -04:00
syncrep.c Final pgindent + perltidy run for v10. 2017-08-14 17:29:33 -04:00
syncrep_gram.y Update copyright via script for 2017 2017-01-03 13:48:53 -05:00
syncrep_scanner.l Rethink flex flags for syncrep_scanner.l. 2017-05-19 18:05:20 -04:00
walreceiver.c Fix locking in WAL receiver/sender shmem state structs 2017-06-30 18:06:33 -04:00
walreceiverfuncs.c Spelling fixes 2017-03-14 12:58:39 -04:00
walsender.c Remove uses of "slave" in replication contexts 2017-08-10 22:55:41 -04:00

src/backend/replication/README

Walreceiver - libpqwalreceiver API
----------------------------------

The transport-specific part of walreceiver, responsible for connecting to
the primary server, receiving WAL files and sending messages, is loaded
dynamically to avoid having to link the main server binary with libpq.
The dynamically loaded module is in libpqwalreceiver subdirectory.

The dynamically loaded module implements four functions:


bool walrcv_connect(char *conninfo, XLogRecPtr startpoint)

Establish connection to the primary, and starts streaming from 'startpoint'.
Returns true on success.

int walrcv_receive(char **buffer, pgsocket *wait_fd)

Retrieve any message available without blocking through the
connection.  If a message was successfully read, returns its
length. If the connection is closed, returns -1.  Otherwise returns 0
to indicate that no data is available, and sets *wait_fd to a socket
descriptor which can be waited on before trying again.  On success, a
pointer to the message payload is stored in *buffer. The returned
buffer is valid until the next call to walrcv_* functions, and the
caller should not attempt to free it.

void walrcv_send(const char *buffer, int nbytes)

Send a message to XLOG stream.

void walrcv_disconnect(void);

Disconnect.


This API should be considered internal at the moment, but we could open it
up for 3rd party replacements of libpqwalreceiver in the future, allowing
pluggable methods for receiving WAL.

Walreceiver IPC
---------------

When the WAL replay in startup process has reached the end of archived WAL,
restorable using restore_command, it starts up the walreceiver process
to fetch more WAL (if streaming replication is configured).

Walreceiver is a postmaster subprocess, so the startup process can't fork it
directly. Instead, it sends a signal to postmaster, asking postmaster to launch
it. Before that, however, startup process fills in WalRcvData->conninfo
and WalRcvData->slotname, and initializes the starting point in
WalRcvData->receiveStart.

As walreceiver receives WAL from the master server, and writes and flushes
it to disk (in pg_wal), it updates WalRcvData->receivedUpto and signals
the startup process to know how far WAL replay can advance.

Walreceiver sends information about replication progress to the master server
whenever it either writes or flushes new WAL, or the specified interval elapses.
This is used for reporting purpose.

Walsender IPC
-------------

At shutdown, postmaster handles walsender processes differently from regular
backends. It waits for regular backends to die before writing the
shutdown checkpoint and terminating pgarch and other auxiliary processes, but
that's not desirable for walsenders, because we want the standby servers to
receive all the WAL, including the shutdown checkpoint, before the master
is shut down. Therefore postmaster treats walsenders like the pgarch process,
and instructs them to terminate at PM_SHUTDOWN_2 phase, after all regular
backends have died and checkpointer has issued the shutdown checkpoint.

When postmaster accepts a connection, it immediately forks a new process
to handle the handshake and authentication, and the process initializes to
become a backend. Postmaster doesn't know if the process becomes a regular
backend or a walsender process at that time - that's indicated in the
connection handshake - so we need some extra signaling to let postmaster
identify walsender processes.

When walsender process starts up, it marks itself as a walsender process in
the PMSignal array. That way postmaster can tell it apart from regular
backends.

Note that no big harm is done if postmaster thinks that a walsender is a
regular backend; it will just terminate the walsender earlier in the shutdown
phase. A walsender will look like a regular backend until it's done with the
initialization and has marked itself in PMSignal array, and at process
termination, after unmarking the PMSignal slot.

Each walsender allocates an entry from the WalSndCtl array, and tracks
information about replication progress. User can monitor them via
statistics views.


Walsender - walreceiver protocol
--------------------------------

See manual.