mirror of
https://github.com/postgres/postgres.git
synced 2025-05-28 05:21:27 +03:00
This commit documents that the failover option is not copied when using the pg_copy_logical_replication_slot function. In passing, we modify the comments in the function clarifying the reason for this behavior. Reported-by: <duffieldzane@gmail.com> Author: Hou Zhijie <houzj.fnst@fujitsu.com> Reviewed-by: Amit Kapila <amit.kapila16@gmail.com> Backpatch-through: 17, where it was introduced Discussion: https://postgr.es/m/173976850802.682632.11315364077431550250@wrigleys.postgresql.org
913 lines
26 KiB
C
913 lines
26 KiB
C
/*-------------------------------------------------------------------------
|
|
*
|
|
* slotfuncs.c
|
|
* Support functions for replication slots
|
|
*
|
|
* Copyright (c) 2012-2024, PostgreSQL Global Development Group
|
|
*
|
|
* IDENTIFICATION
|
|
* src/backend/replication/slotfuncs.c
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
#include "postgres.h"
|
|
|
|
#include "access/htup_details.h"
|
|
#include "access/xlog_internal.h"
|
|
#include "access/xlogrecovery.h"
|
|
#include "access/xlogutils.h"
|
|
#include "funcapi.h"
|
|
#include "miscadmin.h"
|
|
#include "replication/decode.h"
|
|
#include "replication/logical.h"
|
|
#include "replication/slot.h"
|
|
#include "replication/slotsync.h"
|
|
#include "utils/builtins.h"
|
|
#include "utils/guc.h"
|
|
#include "utils/inval.h"
|
|
#include "utils/pg_lsn.h"
|
|
#include "utils/resowner.h"
|
|
|
|
/*
|
|
* Helper function for creating a new physical replication slot with
|
|
* given arguments. Note that this function doesn't release the created
|
|
* slot.
|
|
*
|
|
* If restart_lsn is a valid value, we use it without WAL reservation
|
|
* routine. So the caller must guarantee that WAL is available.
|
|
*/
|
|
static void
|
|
create_physical_replication_slot(char *name, bool immediately_reserve,
|
|
bool temporary, XLogRecPtr restart_lsn)
|
|
{
|
|
Assert(!MyReplicationSlot);
|
|
|
|
/* acquire replication slot, this will check for conflicting names */
|
|
ReplicationSlotCreate(name, false,
|
|
temporary ? RS_TEMPORARY : RS_PERSISTENT, false,
|
|
false, false);
|
|
|
|
if (immediately_reserve)
|
|
{
|
|
/* Reserve WAL as the user asked for it */
|
|
if (XLogRecPtrIsInvalid(restart_lsn))
|
|
ReplicationSlotReserveWal();
|
|
else
|
|
MyReplicationSlot->data.restart_lsn = restart_lsn;
|
|
|
|
/* Write this slot to disk */
|
|
ReplicationSlotMarkDirty();
|
|
ReplicationSlotSave();
|
|
}
|
|
}
|
|
|
|
/*
|
|
* SQL function for creating a new physical (streaming replication)
|
|
* replication slot.
|
|
*/
|
|
Datum
|
|
pg_create_physical_replication_slot(PG_FUNCTION_ARGS)
|
|
{
|
|
Name name = PG_GETARG_NAME(0);
|
|
bool immediately_reserve = PG_GETARG_BOOL(1);
|
|
bool temporary = PG_GETARG_BOOL(2);
|
|
Datum values[2];
|
|
bool nulls[2];
|
|
TupleDesc tupdesc;
|
|
HeapTuple tuple;
|
|
Datum result;
|
|
|
|
if (get_call_result_type(fcinfo, NULL, &tupdesc) != TYPEFUNC_COMPOSITE)
|
|
elog(ERROR, "return type must be a row type");
|
|
|
|
CheckSlotPermissions();
|
|
|
|
CheckSlotRequirements();
|
|
|
|
create_physical_replication_slot(NameStr(*name),
|
|
immediately_reserve,
|
|
temporary,
|
|
InvalidXLogRecPtr);
|
|
|
|
values[0] = NameGetDatum(&MyReplicationSlot->data.name);
|
|
nulls[0] = false;
|
|
|
|
if (immediately_reserve)
|
|
{
|
|
values[1] = LSNGetDatum(MyReplicationSlot->data.restart_lsn);
|
|
nulls[1] = false;
|
|
}
|
|
else
|
|
nulls[1] = true;
|
|
|
|
tuple = heap_form_tuple(tupdesc, values, nulls);
|
|
result = HeapTupleGetDatum(tuple);
|
|
|
|
ReplicationSlotRelease();
|
|
|
|
PG_RETURN_DATUM(result);
|
|
}
|
|
|
|
|
|
/*
|
|
* Helper function for creating a new logical replication slot with
|
|
* given arguments. Note that this function doesn't release the created
|
|
* slot.
|
|
*
|
|
* When find_startpoint is false, the slot's confirmed_flush is not set; it's
|
|
* caller's responsibility to ensure it's set to something sensible.
|
|
*/
|
|
static void
|
|
create_logical_replication_slot(char *name, char *plugin,
|
|
bool temporary, bool two_phase,
|
|
bool failover,
|
|
XLogRecPtr restart_lsn,
|
|
bool find_startpoint)
|
|
{
|
|
LogicalDecodingContext *ctx = NULL;
|
|
|
|
Assert(!MyReplicationSlot);
|
|
|
|
/*
|
|
* Acquire a logical decoding slot, this will check for conflicting names.
|
|
* Initially create persistent slot as ephemeral - that allows us to
|
|
* nicely handle errors during initialization because it'll get dropped if
|
|
* this transaction fails. We'll make it persistent at the end. Temporary
|
|
* slots can be created as temporary from beginning as they get dropped on
|
|
* error as well.
|
|
*/
|
|
ReplicationSlotCreate(name, true,
|
|
temporary ? RS_TEMPORARY : RS_EPHEMERAL, two_phase,
|
|
failover, false);
|
|
|
|
/*
|
|
* Create logical decoding context to find start point or, if we don't
|
|
* need it, to 1) bump slot's restart_lsn and xmin 2) check plugin sanity.
|
|
*
|
|
* Note: when !find_startpoint this is still important, because it's at
|
|
* this point that the output plugin is validated.
|
|
*/
|
|
ctx = CreateInitDecodingContext(plugin, NIL,
|
|
false, /* just catalogs is OK */
|
|
restart_lsn,
|
|
XL_ROUTINE(.page_read = read_local_xlog_page,
|
|
.segment_open = wal_segment_open,
|
|
.segment_close = wal_segment_close),
|
|
NULL, NULL, NULL);
|
|
|
|
/*
|
|
* If caller needs us to determine the decoding start point, do so now.
|
|
* This might take a while.
|
|
*/
|
|
if (find_startpoint)
|
|
DecodingContextFindStartpoint(ctx);
|
|
|
|
/* don't need the decoding context anymore */
|
|
FreeDecodingContext(ctx);
|
|
}
|
|
|
|
/*
|
|
* SQL function for creating a new logical replication slot.
|
|
*/
|
|
Datum
|
|
pg_create_logical_replication_slot(PG_FUNCTION_ARGS)
|
|
{
|
|
Name name = PG_GETARG_NAME(0);
|
|
Name plugin = PG_GETARG_NAME(1);
|
|
bool temporary = PG_GETARG_BOOL(2);
|
|
bool two_phase = PG_GETARG_BOOL(3);
|
|
bool failover = PG_GETARG_BOOL(4);
|
|
Datum result;
|
|
TupleDesc tupdesc;
|
|
HeapTuple tuple;
|
|
Datum values[2];
|
|
bool nulls[2];
|
|
|
|
if (get_call_result_type(fcinfo, NULL, &tupdesc) != TYPEFUNC_COMPOSITE)
|
|
elog(ERROR, "return type must be a row type");
|
|
|
|
CheckSlotPermissions();
|
|
|
|
CheckLogicalDecodingRequirements();
|
|
|
|
create_logical_replication_slot(NameStr(*name),
|
|
NameStr(*plugin),
|
|
temporary,
|
|
two_phase,
|
|
failover,
|
|
InvalidXLogRecPtr,
|
|
true);
|
|
|
|
values[0] = NameGetDatum(&MyReplicationSlot->data.name);
|
|
values[1] = LSNGetDatum(MyReplicationSlot->data.confirmed_flush);
|
|
|
|
memset(nulls, 0, sizeof(nulls));
|
|
|
|
tuple = heap_form_tuple(tupdesc, values, nulls);
|
|
result = HeapTupleGetDatum(tuple);
|
|
|
|
/* ok, slot is now fully created, mark it as persistent if needed */
|
|
if (!temporary)
|
|
ReplicationSlotPersist();
|
|
ReplicationSlotRelease();
|
|
|
|
PG_RETURN_DATUM(result);
|
|
}
|
|
|
|
|
|
/*
|
|
* SQL function for dropping a replication slot.
|
|
*/
|
|
Datum
|
|
pg_drop_replication_slot(PG_FUNCTION_ARGS)
|
|
{
|
|
Name name = PG_GETARG_NAME(0);
|
|
|
|
CheckSlotPermissions();
|
|
|
|
CheckSlotRequirements();
|
|
|
|
ReplicationSlotDrop(NameStr(*name), true);
|
|
|
|
PG_RETURN_VOID();
|
|
}
|
|
|
|
/*
|
|
* pg_get_replication_slots - SQL SRF showing all replication slots
|
|
* that currently exist on the database cluster.
|
|
*/
|
|
Datum
|
|
pg_get_replication_slots(PG_FUNCTION_ARGS)
|
|
{
|
|
#define PG_GET_REPLICATION_SLOTS_COLS 19
|
|
ReturnSetInfo *rsinfo = (ReturnSetInfo *) fcinfo->resultinfo;
|
|
XLogRecPtr currlsn;
|
|
int slotno;
|
|
|
|
/*
|
|
* We don't require any special permission to see this function's data
|
|
* because nothing should be sensitive. The most critical being the slot
|
|
* name, which shouldn't contain anything particularly sensitive.
|
|
*/
|
|
|
|
InitMaterializedSRF(fcinfo, 0);
|
|
|
|
currlsn = GetXLogWriteRecPtr();
|
|
|
|
LWLockAcquire(ReplicationSlotControlLock, LW_SHARED);
|
|
for (slotno = 0; slotno < max_replication_slots; slotno++)
|
|
{
|
|
ReplicationSlot *slot = &ReplicationSlotCtl->replication_slots[slotno];
|
|
ReplicationSlot slot_contents;
|
|
Datum values[PG_GET_REPLICATION_SLOTS_COLS];
|
|
bool nulls[PG_GET_REPLICATION_SLOTS_COLS];
|
|
WALAvailability walstate;
|
|
int i;
|
|
ReplicationSlotInvalidationCause cause;
|
|
|
|
if (!slot->in_use)
|
|
continue;
|
|
|
|
/* Copy slot contents while holding spinlock, then examine at leisure */
|
|
SpinLockAcquire(&slot->mutex);
|
|
slot_contents = *slot;
|
|
SpinLockRelease(&slot->mutex);
|
|
|
|
memset(values, 0, sizeof(values));
|
|
memset(nulls, 0, sizeof(nulls));
|
|
|
|
i = 0;
|
|
values[i++] = NameGetDatum(&slot_contents.data.name);
|
|
|
|
if (slot_contents.data.database == InvalidOid)
|
|
nulls[i++] = true;
|
|
else
|
|
values[i++] = NameGetDatum(&slot_contents.data.plugin);
|
|
|
|
if (slot_contents.data.database == InvalidOid)
|
|
values[i++] = CStringGetTextDatum("physical");
|
|
else
|
|
values[i++] = CStringGetTextDatum("logical");
|
|
|
|
if (slot_contents.data.database == InvalidOid)
|
|
nulls[i++] = true;
|
|
else
|
|
values[i++] = ObjectIdGetDatum(slot_contents.data.database);
|
|
|
|
values[i++] = BoolGetDatum(slot_contents.data.persistency == RS_TEMPORARY);
|
|
values[i++] = BoolGetDatum(slot_contents.active_pid != 0);
|
|
|
|
if (slot_contents.active_pid != 0)
|
|
values[i++] = Int32GetDatum(slot_contents.active_pid);
|
|
else
|
|
nulls[i++] = true;
|
|
|
|
if (slot_contents.data.xmin != InvalidTransactionId)
|
|
values[i++] = TransactionIdGetDatum(slot_contents.data.xmin);
|
|
else
|
|
nulls[i++] = true;
|
|
|
|
if (slot_contents.data.catalog_xmin != InvalidTransactionId)
|
|
values[i++] = TransactionIdGetDatum(slot_contents.data.catalog_xmin);
|
|
else
|
|
nulls[i++] = true;
|
|
|
|
if (slot_contents.data.restart_lsn != InvalidXLogRecPtr)
|
|
values[i++] = LSNGetDatum(slot_contents.data.restart_lsn);
|
|
else
|
|
nulls[i++] = true;
|
|
|
|
if (slot_contents.data.confirmed_flush != InvalidXLogRecPtr)
|
|
values[i++] = LSNGetDatum(slot_contents.data.confirmed_flush);
|
|
else
|
|
nulls[i++] = true;
|
|
|
|
/*
|
|
* If the slot has not been invalidated, test availability from
|
|
* restart_lsn.
|
|
*/
|
|
if (slot_contents.data.invalidated != RS_INVAL_NONE)
|
|
walstate = WALAVAIL_REMOVED;
|
|
else
|
|
walstate = GetWALAvailability(slot_contents.data.restart_lsn);
|
|
|
|
switch (walstate)
|
|
{
|
|
case WALAVAIL_INVALID_LSN:
|
|
nulls[i++] = true;
|
|
break;
|
|
|
|
case WALAVAIL_RESERVED:
|
|
values[i++] = CStringGetTextDatum("reserved");
|
|
break;
|
|
|
|
case WALAVAIL_EXTENDED:
|
|
values[i++] = CStringGetTextDatum("extended");
|
|
break;
|
|
|
|
case WALAVAIL_UNRESERVED:
|
|
values[i++] = CStringGetTextDatum("unreserved");
|
|
break;
|
|
|
|
case WALAVAIL_REMOVED:
|
|
|
|
/*
|
|
* If we read the restart_lsn long enough ago, maybe that file
|
|
* has been removed by now. However, the walsender could have
|
|
* moved forward enough that it jumped to another file after
|
|
* we looked. If checkpointer signalled the process to
|
|
* termination, then it's definitely lost; but if a process is
|
|
* still alive, then "unreserved" seems more appropriate.
|
|
*
|
|
* If we do change it, save the state for safe_wal_size below.
|
|
*/
|
|
if (!XLogRecPtrIsInvalid(slot_contents.data.restart_lsn))
|
|
{
|
|
int pid;
|
|
|
|
SpinLockAcquire(&slot->mutex);
|
|
pid = slot->active_pid;
|
|
slot_contents.data.restart_lsn = slot->data.restart_lsn;
|
|
SpinLockRelease(&slot->mutex);
|
|
if (pid != 0)
|
|
{
|
|
values[i++] = CStringGetTextDatum("unreserved");
|
|
walstate = WALAVAIL_UNRESERVED;
|
|
break;
|
|
}
|
|
}
|
|
values[i++] = CStringGetTextDatum("lost");
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* safe_wal_size is only computed for slots that have not been lost,
|
|
* and only if there's a configured maximum size.
|
|
*/
|
|
if (walstate == WALAVAIL_REMOVED || max_slot_wal_keep_size_mb < 0)
|
|
nulls[i++] = true;
|
|
else
|
|
{
|
|
XLogSegNo targetSeg;
|
|
uint64 slotKeepSegs;
|
|
uint64 keepSegs;
|
|
XLogSegNo failSeg;
|
|
XLogRecPtr failLSN;
|
|
|
|
XLByteToSeg(slot_contents.data.restart_lsn, targetSeg, wal_segment_size);
|
|
|
|
/* determine how many segments can be kept by slots */
|
|
slotKeepSegs = XLogMBVarToSegs(max_slot_wal_keep_size_mb, wal_segment_size);
|
|
/* ditto for wal_keep_size */
|
|
keepSegs = XLogMBVarToSegs(wal_keep_size_mb, wal_segment_size);
|
|
|
|
/* if currpos reaches failLSN, we lose our segment */
|
|
failSeg = targetSeg + Max(slotKeepSegs, keepSegs) + 1;
|
|
XLogSegNoOffsetToRecPtr(failSeg, 0, wal_segment_size, failLSN);
|
|
|
|
values[i++] = Int64GetDatum(failLSN - currlsn);
|
|
}
|
|
|
|
values[i++] = BoolGetDatum(slot_contents.data.two_phase);
|
|
|
|
if (slot_contents.inactive_since > 0)
|
|
values[i++] = TimestampTzGetDatum(slot_contents.inactive_since);
|
|
else
|
|
nulls[i++] = true;
|
|
|
|
cause = slot_contents.data.invalidated;
|
|
|
|
if (SlotIsPhysical(&slot_contents))
|
|
nulls[i++] = true;
|
|
else
|
|
{
|
|
/*
|
|
* rows_removed and wal_level_insufficient are the only two
|
|
* reasons for the logical slot's conflict with recovery.
|
|
*/
|
|
if (cause == RS_INVAL_HORIZON ||
|
|
cause == RS_INVAL_WAL_LEVEL)
|
|
values[i++] = BoolGetDatum(true);
|
|
else
|
|
values[i++] = BoolGetDatum(false);
|
|
}
|
|
|
|
if (cause == RS_INVAL_NONE)
|
|
nulls[i++] = true;
|
|
else
|
|
values[i++] = CStringGetTextDatum(SlotInvalidationCauses[cause]);
|
|
|
|
values[i++] = BoolGetDatum(slot_contents.data.failover);
|
|
|
|
values[i++] = BoolGetDatum(slot_contents.data.synced);
|
|
|
|
Assert(i == PG_GET_REPLICATION_SLOTS_COLS);
|
|
|
|
tuplestore_putvalues(rsinfo->setResult, rsinfo->setDesc,
|
|
values, nulls);
|
|
}
|
|
|
|
LWLockRelease(ReplicationSlotControlLock);
|
|
|
|
return (Datum) 0;
|
|
}
|
|
|
|
/*
|
|
* Helper function for advancing our physical replication slot forward.
|
|
*
|
|
* The LSN position to move to is compared simply to the slot's restart_lsn,
|
|
* knowing that any position older than that would be removed by successive
|
|
* checkpoints.
|
|
*/
|
|
static XLogRecPtr
|
|
pg_physical_replication_slot_advance(XLogRecPtr moveto)
|
|
{
|
|
XLogRecPtr startlsn = MyReplicationSlot->data.restart_lsn;
|
|
XLogRecPtr retlsn = startlsn;
|
|
|
|
Assert(moveto != InvalidXLogRecPtr);
|
|
|
|
if (startlsn < moveto)
|
|
{
|
|
SpinLockAcquire(&MyReplicationSlot->mutex);
|
|
MyReplicationSlot->data.restart_lsn = moveto;
|
|
SpinLockRelease(&MyReplicationSlot->mutex);
|
|
retlsn = moveto;
|
|
|
|
/*
|
|
* Dirty the slot so as it is written out at the next checkpoint. Note
|
|
* that the LSN position advanced may still be lost in the event of a
|
|
* crash, but this makes the data consistent after a clean shutdown.
|
|
*/
|
|
ReplicationSlotMarkDirty();
|
|
|
|
/*
|
|
* Wake up logical walsenders holding logical failover slots after
|
|
* updating the restart_lsn of the physical slot.
|
|
*/
|
|
PhysicalWakeupLogicalWalSnd();
|
|
}
|
|
|
|
return retlsn;
|
|
}
|
|
|
|
/*
|
|
* Advance our logical replication slot forward. See
|
|
* LogicalSlotAdvanceAndCheckSnapState for details.
|
|
*/
|
|
static XLogRecPtr
|
|
pg_logical_replication_slot_advance(XLogRecPtr moveto)
|
|
{
|
|
return LogicalSlotAdvanceAndCheckSnapState(moveto, NULL);
|
|
}
|
|
|
|
/*
|
|
* SQL function for moving the position in a replication slot.
|
|
*/
|
|
Datum
|
|
pg_replication_slot_advance(PG_FUNCTION_ARGS)
|
|
{
|
|
Name slotname = PG_GETARG_NAME(0);
|
|
XLogRecPtr moveto = PG_GETARG_LSN(1);
|
|
XLogRecPtr endlsn;
|
|
XLogRecPtr minlsn;
|
|
TupleDesc tupdesc;
|
|
Datum values[2];
|
|
bool nulls[2];
|
|
HeapTuple tuple;
|
|
Datum result;
|
|
|
|
Assert(!MyReplicationSlot);
|
|
|
|
CheckSlotPermissions();
|
|
|
|
if (XLogRecPtrIsInvalid(moveto))
|
|
ereport(ERROR,
|
|
(errmsg("invalid target WAL LSN")));
|
|
|
|
/* Build a tuple descriptor for our result type */
|
|
if (get_call_result_type(fcinfo, NULL, &tupdesc) != TYPEFUNC_COMPOSITE)
|
|
elog(ERROR, "return type must be a row type");
|
|
|
|
/*
|
|
* We can't move slot past what's been flushed/replayed so clamp the
|
|
* target position accordingly.
|
|
*/
|
|
if (!RecoveryInProgress())
|
|
moveto = Min(moveto, GetFlushRecPtr(NULL));
|
|
else
|
|
moveto = Min(moveto, GetXLogReplayRecPtr(NULL));
|
|
|
|
/* Acquire the slot so we "own" it */
|
|
ReplicationSlotAcquire(NameStr(*slotname), true);
|
|
|
|
/* A slot whose restart_lsn has never been reserved cannot be advanced */
|
|
if (XLogRecPtrIsInvalid(MyReplicationSlot->data.restart_lsn))
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
|
|
errmsg("replication slot \"%s\" cannot be advanced",
|
|
NameStr(*slotname)),
|
|
errdetail("This slot has never previously reserved WAL, or it has been invalidated.")));
|
|
|
|
/*
|
|
* Check if the slot is not moving backwards. Physical slots rely simply
|
|
* on restart_lsn as a minimum point, while logical slots have confirmed
|
|
* consumption up to confirmed_flush, meaning that in both cases data
|
|
* older than that is not available anymore.
|
|
*/
|
|
if (OidIsValid(MyReplicationSlot->data.database))
|
|
minlsn = MyReplicationSlot->data.confirmed_flush;
|
|
else
|
|
minlsn = MyReplicationSlot->data.restart_lsn;
|
|
|
|
if (moveto < minlsn)
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
|
|
errmsg("cannot advance replication slot to %X/%X, minimum is %X/%X",
|
|
LSN_FORMAT_ARGS(moveto), LSN_FORMAT_ARGS(minlsn))));
|
|
|
|
/* Do the actual slot update, depending on the slot type */
|
|
if (OidIsValid(MyReplicationSlot->data.database))
|
|
endlsn = pg_logical_replication_slot_advance(moveto);
|
|
else
|
|
endlsn = pg_physical_replication_slot_advance(moveto);
|
|
|
|
values[0] = NameGetDatum(&MyReplicationSlot->data.name);
|
|
nulls[0] = false;
|
|
|
|
/*
|
|
* Recompute the minimum LSN and xmin across all slots to adjust with the
|
|
* advancing potentially done.
|
|
*/
|
|
ReplicationSlotsComputeRequiredXmin(false);
|
|
ReplicationSlotsComputeRequiredLSN();
|
|
|
|
ReplicationSlotRelease();
|
|
|
|
/* Return the reached position. */
|
|
values[1] = LSNGetDatum(endlsn);
|
|
nulls[1] = false;
|
|
|
|
tuple = heap_form_tuple(tupdesc, values, nulls);
|
|
result = HeapTupleGetDatum(tuple);
|
|
|
|
PG_RETURN_DATUM(result);
|
|
}
|
|
|
|
/*
|
|
* Helper function of copying a replication slot.
|
|
*/
|
|
static Datum
|
|
copy_replication_slot(FunctionCallInfo fcinfo, bool logical_slot)
|
|
{
|
|
Name src_name = PG_GETARG_NAME(0);
|
|
Name dst_name = PG_GETARG_NAME(1);
|
|
ReplicationSlot *src = NULL;
|
|
ReplicationSlot first_slot_contents;
|
|
ReplicationSlot second_slot_contents;
|
|
XLogRecPtr src_restart_lsn;
|
|
bool src_islogical;
|
|
bool temporary;
|
|
char *plugin;
|
|
Datum values[2];
|
|
bool nulls[2];
|
|
Datum result;
|
|
TupleDesc tupdesc;
|
|
HeapTuple tuple;
|
|
|
|
if (get_call_result_type(fcinfo, NULL, &tupdesc) != TYPEFUNC_COMPOSITE)
|
|
elog(ERROR, "return type must be a row type");
|
|
|
|
CheckSlotPermissions();
|
|
|
|
if (logical_slot)
|
|
CheckLogicalDecodingRequirements();
|
|
else
|
|
CheckSlotRequirements();
|
|
|
|
LWLockAcquire(ReplicationSlotControlLock, LW_SHARED);
|
|
|
|
/*
|
|
* We need to prevent the source slot's reserved WAL from being removed,
|
|
* but we don't want to lock that slot for very long, and it can advance
|
|
* in the meantime. So obtain the source slot's data, and create a new
|
|
* slot using its restart_lsn. Afterwards we lock the source slot again
|
|
* and verify that the data we copied (name, type) has not changed
|
|
* incompatibly. No inconvenient WAL removal can occur once the new slot
|
|
* is created -- but since WAL removal could have occurred before we
|
|
* managed to create the new slot, we advance the new slot's restart_lsn
|
|
* to the source slot's updated restart_lsn the second time we lock it.
|
|
*/
|
|
for (int i = 0; i < max_replication_slots; i++)
|
|
{
|
|
ReplicationSlot *s = &ReplicationSlotCtl->replication_slots[i];
|
|
|
|
if (s->in_use && strcmp(NameStr(s->data.name), NameStr(*src_name)) == 0)
|
|
{
|
|
/* Copy the slot contents while holding spinlock */
|
|
SpinLockAcquire(&s->mutex);
|
|
first_slot_contents = *s;
|
|
SpinLockRelease(&s->mutex);
|
|
src = s;
|
|
break;
|
|
}
|
|
}
|
|
|
|
LWLockRelease(ReplicationSlotControlLock);
|
|
|
|
if (src == NULL)
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_UNDEFINED_OBJECT),
|
|
errmsg("replication slot \"%s\" does not exist", NameStr(*src_name))));
|
|
|
|
src_islogical = SlotIsLogical(&first_slot_contents);
|
|
src_restart_lsn = first_slot_contents.data.restart_lsn;
|
|
temporary = (first_slot_contents.data.persistency == RS_TEMPORARY);
|
|
plugin = logical_slot ? NameStr(first_slot_contents.data.plugin) : NULL;
|
|
|
|
/* Check type of replication slot */
|
|
if (src_islogical != logical_slot)
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
|
|
src_islogical ?
|
|
errmsg("cannot copy physical replication slot \"%s\" as a logical replication slot",
|
|
NameStr(*src_name)) :
|
|
errmsg("cannot copy logical replication slot \"%s\" as a physical replication slot",
|
|
NameStr(*src_name))));
|
|
|
|
/* Copying non-reserved slot doesn't make sense */
|
|
if (XLogRecPtrIsInvalid(src_restart_lsn))
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
|
|
errmsg("cannot copy a replication slot that doesn't reserve WAL")));
|
|
|
|
/* Overwrite params from optional arguments */
|
|
if (PG_NARGS() >= 3)
|
|
temporary = PG_GETARG_BOOL(2);
|
|
if (PG_NARGS() >= 4)
|
|
{
|
|
Assert(logical_slot);
|
|
plugin = NameStr(*(PG_GETARG_NAME(3)));
|
|
}
|
|
|
|
/* Create new slot and acquire it */
|
|
if (logical_slot)
|
|
{
|
|
/*
|
|
* We must not try to read WAL, since we haven't reserved it yet --
|
|
* hence pass find_startpoint false. confirmed_flush will be set
|
|
* below, by copying from the source slot.
|
|
*
|
|
* We don't copy the failover option to prevent potential issues with
|
|
* slot synchronization. For instance, if a slot was synchronized to
|
|
* the standby, then dropped on the primary, and immediately recreated
|
|
* by copying from another existing slot with much earlier restart_lsn
|
|
* and confirmed_flush_lsn, the slot synchronization would only
|
|
* observe the LSN of the same slot moving backward. As slot
|
|
* synchronization does not copy the restart_lsn and
|
|
* confirmed_flush_lsn backward (see update_local_synced_slot() for
|
|
* details), if a failover happens before the primary's slot catches
|
|
* up, logical replication cannot continue using the synchronized slot
|
|
* on the promoted standby because the slot retains the restart_lsn
|
|
* and confirmed_flush_lsn that are much later than expected.
|
|
*/
|
|
create_logical_replication_slot(NameStr(*dst_name),
|
|
plugin,
|
|
temporary,
|
|
false,
|
|
false,
|
|
src_restart_lsn,
|
|
false);
|
|
}
|
|
else
|
|
create_physical_replication_slot(NameStr(*dst_name),
|
|
true,
|
|
temporary,
|
|
src_restart_lsn);
|
|
|
|
/*
|
|
* Update the destination slot to current values of the source slot;
|
|
* recheck that the source slot is still the one we saw previously.
|
|
*/
|
|
{
|
|
TransactionId copy_effective_xmin;
|
|
TransactionId copy_effective_catalog_xmin;
|
|
TransactionId copy_xmin;
|
|
TransactionId copy_catalog_xmin;
|
|
XLogRecPtr copy_restart_lsn;
|
|
XLogRecPtr copy_confirmed_flush;
|
|
bool copy_islogical;
|
|
char *copy_name;
|
|
|
|
/* Copy data of source slot again */
|
|
SpinLockAcquire(&src->mutex);
|
|
second_slot_contents = *src;
|
|
SpinLockRelease(&src->mutex);
|
|
|
|
copy_effective_xmin = second_slot_contents.effective_xmin;
|
|
copy_effective_catalog_xmin = second_slot_contents.effective_catalog_xmin;
|
|
|
|
copy_xmin = second_slot_contents.data.xmin;
|
|
copy_catalog_xmin = second_slot_contents.data.catalog_xmin;
|
|
copy_restart_lsn = second_slot_contents.data.restart_lsn;
|
|
copy_confirmed_flush = second_slot_contents.data.confirmed_flush;
|
|
|
|
/* for existence check */
|
|
copy_name = NameStr(second_slot_contents.data.name);
|
|
copy_islogical = SlotIsLogical(&second_slot_contents);
|
|
|
|
/*
|
|
* Check if the source slot still exists and is valid. We regard it as
|
|
* invalid if the type of replication slot or name has been changed,
|
|
* or the restart_lsn either is invalid or has gone backward. (The
|
|
* restart_lsn could go backwards if the source slot is dropped and
|
|
* copied from an older slot during installation.)
|
|
*
|
|
* Since erroring out will release and drop the destination slot we
|
|
* don't need to release it here.
|
|
*/
|
|
if (copy_restart_lsn < src_restart_lsn ||
|
|
src_islogical != copy_islogical ||
|
|
strcmp(copy_name, NameStr(*src_name)) != 0)
|
|
ereport(ERROR,
|
|
(errmsg("could not copy replication slot \"%s\"",
|
|
NameStr(*src_name)),
|
|
errdetail("The source replication slot was modified incompatibly during the copy operation.")));
|
|
|
|
/* The source slot must have a consistent snapshot */
|
|
if (src_islogical && XLogRecPtrIsInvalid(copy_confirmed_flush))
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
|
|
errmsg("cannot copy unfinished logical replication slot \"%s\"",
|
|
NameStr(*src_name)),
|
|
errhint("Retry when the source replication slot's confirmed_flush_lsn is valid.")));
|
|
|
|
/* Install copied values again */
|
|
SpinLockAcquire(&MyReplicationSlot->mutex);
|
|
MyReplicationSlot->effective_xmin = copy_effective_xmin;
|
|
MyReplicationSlot->effective_catalog_xmin = copy_effective_catalog_xmin;
|
|
|
|
MyReplicationSlot->data.xmin = copy_xmin;
|
|
MyReplicationSlot->data.catalog_xmin = copy_catalog_xmin;
|
|
MyReplicationSlot->data.restart_lsn = copy_restart_lsn;
|
|
MyReplicationSlot->data.confirmed_flush = copy_confirmed_flush;
|
|
SpinLockRelease(&MyReplicationSlot->mutex);
|
|
|
|
ReplicationSlotMarkDirty();
|
|
ReplicationSlotsComputeRequiredXmin(false);
|
|
ReplicationSlotsComputeRequiredLSN();
|
|
ReplicationSlotSave();
|
|
|
|
#ifdef USE_ASSERT_CHECKING
|
|
/* Check that the restart_lsn is available */
|
|
{
|
|
XLogSegNo segno;
|
|
|
|
XLByteToSeg(copy_restart_lsn, segno, wal_segment_size);
|
|
Assert(XLogGetLastRemovedSegno() < segno);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/* target slot fully created, mark as persistent if needed */
|
|
if (logical_slot && !temporary)
|
|
ReplicationSlotPersist();
|
|
|
|
/* All done. Set up the return values */
|
|
values[0] = NameGetDatum(dst_name);
|
|
nulls[0] = false;
|
|
if (!XLogRecPtrIsInvalid(MyReplicationSlot->data.confirmed_flush))
|
|
{
|
|
values[1] = LSNGetDatum(MyReplicationSlot->data.confirmed_flush);
|
|
nulls[1] = false;
|
|
}
|
|
else
|
|
nulls[1] = true;
|
|
|
|
tuple = heap_form_tuple(tupdesc, values, nulls);
|
|
result = HeapTupleGetDatum(tuple);
|
|
|
|
ReplicationSlotRelease();
|
|
|
|
PG_RETURN_DATUM(result);
|
|
}
|
|
|
|
/* The wrappers below are all to appease opr_sanity */
|
|
Datum
|
|
pg_copy_logical_replication_slot_a(PG_FUNCTION_ARGS)
|
|
{
|
|
return copy_replication_slot(fcinfo, true);
|
|
}
|
|
|
|
Datum
|
|
pg_copy_logical_replication_slot_b(PG_FUNCTION_ARGS)
|
|
{
|
|
return copy_replication_slot(fcinfo, true);
|
|
}
|
|
|
|
Datum
|
|
pg_copy_logical_replication_slot_c(PG_FUNCTION_ARGS)
|
|
{
|
|
return copy_replication_slot(fcinfo, true);
|
|
}
|
|
|
|
Datum
|
|
pg_copy_physical_replication_slot_a(PG_FUNCTION_ARGS)
|
|
{
|
|
return copy_replication_slot(fcinfo, false);
|
|
}
|
|
|
|
Datum
|
|
pg_copy_physical_replication_slot_b(PG_FUNCTION_ARGS)
|
|
{
|
|
return copy_replication_slot(fcinfo, false);
|
|
}
|
|
|
|
/*
|
|
* Synchronize failover enabled replication slots to a standby server
|
|
* from the primary server.
|
|
*/
|
|
Datum
|
|
pg_sync_replication_slots(PG_FUNCTION_ARGS)
|
|
{
|
|
WalReceiverConn *wrconn;
|
|
char *err;
|
|
StringInfoData app_name;
|
|
|
|
CheckSlotPermissions();
|
|
|
|
if (!RecoveryInProgress())
|
|
ereport(ERROR,
|
|
errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
|
|
errmsg("replication slots can only be synchronized to a standby server"));
|
|
|
|
ValidateSlotSyncParams(ERROR);
|
|
|
|
/* Load the libpq-specific functions */
|
|
load_file("libpqwalreceiver", false);
|
|
|
|
(void) CheckAndGetDbnameFromConninfo();
|
|
|
|
initStringInfo(&app_name);
|
|
if (cluster_name[0])
|
|
appendStringInfo(&app_name, "%s_slotsync", cluster_name);
|
|
else
|
|
appendStringInfoString(&app_name, "slotsync");
|
|
|
|
/* Connect to the primary server. */
|
|
wrconn = walrcv_connect(PrimaryConnInfo, false, false, false,
|
|
app_name.data, &err);
|
|
pfree(app_name.data);
|
|
|
|
if (!wrconn)
|
|
ereport(ERROR,
|
|
errcode(ERRCODE_CONNECTION_FAILURE),
|
|
errmsg("could not connect to the primary server: %s", err));
|
|
|
|
SyncReplicationSlots(wrconn);
|
|
|
|
walrcv_disconnect(wrconn);
|
|
|
|
PG_RETURN_VOID();
|
|
}
|