mirror of
https://github.com/postgres/postgres.git
synced 2025-09-02 04:21:28 +03:00
Pre-beta mechanical code beautification.
Run pgindent, pgperltidy, and reformat-dat-files. This set of diffs is a bit larger than typical. We've updated to pg_bsd_indent 2.1.2, which properly indents variable declarations that have multi-line initialization expressions (the continuation lines are now indented one tab stop). We've also updated to perltidy version 20230309 and changed some of its settings, which reduces its desire to add whitespace to lines to make assignments etc. line up. Going forward, that should make for fewer random-seeming changes to existing code. Discussion: https://postgr.es/m/20230428092545.qfb3y5wcu4cm75ur@alvherre.pgsql
This commit is contained in:
@@ -700,8 +700,8 @@ bringetbitmap(IndexScanDesc scan, TIDBitmap *tbm)
|
||||
}
|
||||
|
||||
/*
|
||||
* If we found a scan key eliminating the range, no need to
|
||||
* check additional ones.
|
||||
* If we found a scan key eliminating the range, no need
|
||||
* to check additional ones.
|
||||
*/
|
||||
if (!addrange)
|
||||
break;
|
||||
@@ -1223,7 +1223,7 @@ brin_build_desc(Relation rel)
|
||||
* Obtain BrinOpcInfo for each indexed column. While at it, accumulate
|
||||
* the number of columns stored, since the number is opclass-defined.
|
||||
*/
|
||||
opcinfo = palloc_array(BrinOpcInfo*, tupdesc->natts);
|
||||
opcinfo = palloc_array(BrinOpcInfo *, tupdesc->natts);
|
||||
for (keyno = 0; keyno < tupdesc->natts; keyno++)
|
||||
{
|
||||
FmgrInfo *opcInfoFn;
|
||||
@@ -1801,8 +1801,8 @@ add_values_to_range(Relation idxRel, BrinDesc *bdesc, BrinMemTuple *dtup,
|
||||
bval = &dtup->bt_columns[keyno];
|
||||
|
||||
/*
|
||||
* Does the range have actual NULL values? Either of the flags can
|
||||
* be set, but we ignore the state before adding first row.
|
||||
* Does the range have actual NULL values? Either of the flags can be
|
||||
* set, but we ignore the state before adding first row.
|
||||
*
|
||||
* We have to remember this, because we'll modify the flags and we
|
||||
* need to know if the range started as empty.
|
||||
@@ -1842,12 +1842,12 @@ add_values_to_range(Relation idxRel, BrinDesc *bdesc, BrinMemTuple *dtup,
|
||||
|
||||
/*
|
||||
* If the range was had actual NULL values (i.e. did not start empty),
|
||||
* make sure we don't forget about the NULL values. Either the allnulls
|
||||
* flag is still set to true, or (if the opclass cleared it) we need to
|
||||
* set hasnulls=true.
|
||||
* make sure we don't forget about the NULL values. Either the
|
||||
* allnulls flag is still set to true, or (if the opclass cleared it)
|
||||
* we need to set hasnulls=true.
|
||||
*
|
||||
* XXX This can only happen when the opclass modified the tuple, so the
|
||||
* modified flag should be set.
|
||||
* XXX This can only happen when the opclass modified the tuple, so
|
||||
* the modified flag should be set.
|
||||
*/
|
||||
if (has_nulls && !(bval->bv_hasnulls || bval->bv_allnulls))
|
||||
{
|
||||
@@ -1859,9 +1859,9 @@ add_values_to_range(Relation idxRel, BrinDesc *bdesc, BrinMemTuple *dtup,
|
||||
/*
|
||||
* After updating summaries for all the keys, mark it as not empty.
|
||||
*
|
||||
* If we're actually changing the flag value (i.e. tuple started as empty),
|
||||
* we should have modified the tuple. So we should not see empty range that
|
||||
* was not modified.
|
||||
* If we're actually changing the flag value (i.e. tuple started as
|
||||
* empty), we should have modified the tuple. So we should not see empty
|
||||
* range that was not modified.
|
||||
*/
|
||||
Assert(!dtup->bt_empty_range || modified);
|
||||
dtup->bt_empty_range = false;
|
||||
|
@@ -1717,7 +1717,7 @@ allocateReloptStruct(Size base, relopt_value *options, int numoptions)
|
||||
if (optstr->fill_cb)
|
||||
{
|
||||
const char *val = optval->isset ? optval->values.string_val :
|
||||
optstr->default_isnull ? NULL : optstr->default_val;
|
||||
optstr->default_isnull ? NULL : optstr->default_val;
|
||||
|
||||
size += optstr->fill_cb(val, NULL);
|
||||
}
|
||||
@@ -1796,8 +1796,8 @@ fillRelOptions(void *rdopts, Size basesize,
|
||||
if (optstring->fill_cb)
|
||||
{
|
||||
Size size =
|
||||
optstring->fill_cb(string_val,
|
||||
(char *) rdopts + offset);
|
||||
optstring->fill_cb(string_val,
|
||||
(char *) rdopts + offset);
|
||||
|
||||
if (size)
|
||||
{
|
||||
|
@@ -1117,7 +1117,7 @@ gistformdownlink(Relation rel, Buffer buf, GISTSTATE *giststate,
|
||||
for (offset = FirstOffsetNumber; offset <= maxoff; offset = OffsetNumberNext(offset))
|
||||
{
|
||||
IndexTuple ituple = (IndexTuple)
|
||||
PageGetItem(page, PageGetItemId(page, offset));
|
||||
PageGetItem(page, PageGetItemId(page, offset));
|
||||
|
||||
if (downlink == NULL)
|
||||
downlink = CopyIndexTuple(ituple);
|
||||
|
@@ -598,7 +598,7 @@ gistRelocateBuildBuffersOnSplit(GISTBuildBuffers *gfbb, GISTSTATE *giststate,
|
||||
{
|
||||
GISTPageSplitInfo *si = (GISTPageSplitInfo *) lfirst(lc);
|
||||
GISTNodeBuffer *newNodeBuffer;
|
||||
int i = foreach_current_index(lc);
|
||||
int i = foreach_current_index(lc);
|
||||
|
||||
/* Decompress parent index tuple of node buffer page. */
|
||||
gistDeCompressAtt(giststate, r,
|
||||
|
@@ -657,7 +657,7 @@ gistgettuple(IndexScanDesc scan, ScanDirection dir)
|
||||
if (so->killedItems == NULL)
|
||||
{
|
||||
MemoryContext oldCxt =
|
||||
MemoryContextSwitchTo(so->giststate->scanCxt);
|
||||
MemoryContextSwitchTo(so->giststate->scanCxt);
|
||||
|
||||
so->killedItems =
|
||||
(OffsetNumber *) palloc(MaxIndexTuplesPerPage
|
||||
@@ -694,7 +694,7 @@ gistgettuple(IndexScanDesc scan, ScanDirection dir)
|
||||
if (so->killedItems == NULL)
|
||||
{
|
||||
MemoryContext oldCxt =
|
||||
MemoryContextSwitchTo(so->giststate->scanCxt);
|
||||
MemoryContextSwitchTo(so->giststate->scanCxt);
|
||||
|
||||
so->killedItems =
|
||||
(OffsetNumber *) palloc(MaxIndexTuplesPerPage
|
||||
|
@@ -125,7 +125,7 @@ gistRedoPageUpdateRecord(XLogReaderState *record)
|
||||
if (data - begin < datalen)
|
||||
{
|
||||
OffsetNumber off = (PageIsEmpty(page)) ? FirstOffsetNumber :
|
||||
OffsetNumberNext(PageGetMaxOffsetNumber(page));
|
||||
OffsetNumberNext(PageGetMaxOffsetNumber(page));
|
||||
|
||||
while (data - begin < datalen)
|
||||
{
|
||||
|
@@ -289,7 +289,8 @@ hashtext(PG_FUNCTION_ARGS)
|
||||
}
|
||||
else
|
||||
{
|
||||
Size bsize, rsize;
|
||||
Size bsize,
|
||||
rsize;
|
||||
char *buf;
|
||||
const char *keydata = VARDATA_ANY(key);
|
||||
size_t keylen = VARSIZE_ANY_EXHDR(key);
|
||||
@@ -304,8 +305,8 @@ hashtext(PG_FUNCTION_ARGS)
|
||||
|
||||
/*
|
||||
* In principle, there's no reason to include the terminating NUL
|
||||
* character in the hash, but it was done before and the behavior
|
||||
* must be preserved.
|
||||
* character in the hash, but it was done before and the behavior must
|
||||
* be preserved.
|
||||
*/
|
||||
result = hash_any((uint8_t *) buf, bsize + 1);
|
||||
|
||||
@@ -343,7 +344,8 @@ hashtextextended(PG_FUNCTION_ARGS)
|
||||
}
|
||||
else
|
||||
{
|
||||
Size bsize, rsize;
|
||||
Size bsize,
|
||||
rsize;
|
||||
char *buf;
|
||||
const char *keydata = VARDATA_ANY(key);
|
||||
size_t keylen = VARSIZE_ANY_EXHDR(key);
|
||||
@@ -357,8 +359,8 @@ hashtextextended(PG_FUNCTION_ARGS)
|
||||
|
||||
/*
|
||||
* In principle, there's no reason to include the terminating NUL
|
||||
* character in the hash, but it was done before and the behavior
|
||||
* must be preserved.
|
||||
* character in the hash, but it was done before and the behavior must
|
||||
* be preserved.
|
||||
*/
|
||||
result = hash_any_extended((uint8_t *) buf, bsize + 1,
|
||||
PG_GETARG_INT64(1));
|
||||
|
@@ -2491,7 +2491,7 @@ static inline bool
|
||||
xmax_infomask_changed(uint16 new_infomask, uint16 old_infomask)
|
||||
{
|
||||
const uint16 interesting =
|
||||
HEAP_XMAX_IS_MULTI | HEAP_XMAX_LOCK_ONLY | HEAP_LOCK_MASK;
|
||||
HEAP_XMAX_IS_MULTI | HEAP_XMAX_LOCK_ONLY | HEAP_LOCK_MASK;
|
||||
|
||||
if ((new_infomask & interesting) != (old_infomask & interesting))
|
||||
return true;
|
||||
|
@@ -334,8 +334,8 @@ heapam_tuple_update(Relation relation, ItemPointer otid, TupleTableSlot *slot,
|
||||
* Note: heap_update returns the tid (location) of the new tuple in the
|
||||
* t_self field.
|
||||
*
|
||||
* If the update is not HOT, we must update all indexes. If the update
|
||||
* is HOT, it could be that we updated summarized columns, so we either
|
||||
* If the update is not HOT, we must update all indexes. If the update is
|
||||
* HOT, it could be that we updated summarized columns, so we either
|
||||
* update only summarized indexes, or none at all.
|
||||
*/
|
||||
if (result != TM_Ok)
|
||||
|
@@ -376,7 +376,7 @@ RelationAddBlocks(Relation relation, BulkInsertState bistate,
|
||||
if (use_fsm && i >= not_in_fsm_pages)
|
||||
{
|
||||
Size freespace = BufferGetPageSize(victim_buffers[i]) -
|
||||
SizeOfPageHeaderData;
|
||||
SizeOfPageHeaderData;
|
||||
|
||||
RecordPageWithFreeSpace(relation, curBlock, freespace);
|
||||
}
|
||||
|
@@ -532,7 +532,7 @@ heap_prune_satisfies_vacuum(PruneState *prstate, HeapTuple tup, Buffer buffer)
|
||||
if (!TransactionIdIsValid(prstate->old_snap_xmin))
|
||||
{
|
||||
TransactionId horizon =
|
||||
GlobalVisTestNonRemovableHorizon(prstate->vistest);
|
||||
GlobalVisTestNonRemovableHorizon(prstate->vistest);
|
||||
|
||||
TransactionIdLimitedForOldSnapshots(horizon, prstate->rel,
|
||||
&prstate->old_snap_xmin,
|
||||
|
@@ -389,6 +389,7 @@ heap_vacuum_rel(Relation rel, VacuumParams *params,
|
||||
Assert(params->index_cleanup != VACOPTVALUE_UNSPECIFIED);
|
||||
Assert(params->truncate != VACOPTVALUE_UNSPECIFIED &&
|
||||
params->truncate != VACOPTVALUE_AUTO);
|
||||
|
||||
/*
|
||||
* While VacuumFailSafeActive is reset to false before calling this, we
|
||||
* still need to reset it here due to recursive calls.
|
||||
@@ -1813,12 +1814,12 @@ retry:
|
||||
{
|
||||
/*
|
||||
* We have no freeze plans to execute, so there's no added cost
|
||||
* from following the freeze path. That's why it was chosen.
|
||||
* This is important in the case where the page only contains
|
||||
* totally frozen tuples at this point (perhaps only following
|
||||
* pruning). Such pages can be marked all-frozen in the VM by our
|
||||
* caller, even though none of its tuples were newly frozen here
|
||||
* (note that the "no freeze" path never sets pages all-frozen).
|
||||
* from following the freeze path. That's why it was chosen. This
|
||||
* is important in the case where the page only contains totally
|
||||
* frozen tuples at this point (perhaps only following pruning).
|
||||
* Such pages can be marked all-frozen in the VM by our caller,
|
||||
* even though none of its tuples were newly frozen here (note
|
||||
* that the "no freeze" path never sets pages all-frozen).
|
||||
*
|
||||
* We never increment the frozen_pages instrumentation counter
|
||||
* here, since it only counts pages with newly frozen tuples
|
||||
@@ -3117,8 +3118,8 @@ dead_items_max_items(LVRelState *vacrel)
|
||||
{
|
||||
int64 max_items;
|
||||
int vac_work_mem = IsAutoVacuumWorkerProcess() &&
|
||||
autovacuum_work_mem != -1 ?
|
||||
autovacuum_work_mem : maintenance_work_mem;
|
||||
autovacuum_work_mem != -1 ?
|
||||
autovacuum_work_mem : maintenance_work_mem;
|
||||
|
||||
if (vacrel->nindexes > 0)
|
||||
{
|
||||
|
@@ -626,7 +626,7 @@ vm_readbuf(Relation rel, BlockNumber blkno, bool extend)
|
||||
static Buffer
|
||||
vm_extend(Relation rel, BlockNumber vm_nblocks)
|
||||
{
|
||||
Buffer buf;
|
||||
Buffer buf;
|
||||
|
||||
buf = ExtendBufferedRelTo(EB_REL(rel), VISIBILITYMAP_FORKNUM, NULL,
|
||||
EB_CREATE_FORK_IF_NEEDED |
|
||||
|
@@ -2947,7 +2947,7 @@ void
|
||||
_bt_pendingfsm_finalize(Relation rel, BTVacState *vstate)
|
||||
{
|
||||
IndexBulkDeleteResult *stats = vstate->stats;
|
||||
Relation heaprel = vstate->info->heaprel;
|
||||
Relation heaprel = vstate->info->heaprel;
|
||||
|
||||
Assert(stats->pages_newly_deleted >= vstate->npendingpages);
|
||||
|
||||
@@ -3027,7 +3027,7 @@ _bt_pendingfsm_add(BTVacState *vstate,
|
||||
if (vstate->npendingpages > 0)
|
||||
{
|
||||
FullTransactionId lastsafexid =
|
||||
vstate->pendingpages[vstate->npendingpages - 1].safexid;
|
||||
vstate->pendingpages[vstate->npendingpages - 1].safexid;
|
||||
|
||||
Assert(FullTransactionIdFollowsOrEquals(safexid, lastsafexid));
|
||||
}
|
||||
|
@@ -27,7 +27,7 @@ dbase_desc(StringInfo buf, XLogReaderState *record)
|
||||
if (info == XLOG_DBASE_CREATE_FILE_COPY)
|
||||
{
|
||||
xl_dbase_create_file_copy_rec *xlrec =
|
||||
(xl_dbase_create_file_copy_rec *) rec;
|
||||
(xl_dbase_create_file_copy_rec *) rec;
|
||||
|
||||
appendStringInfo(buf, "copy dir %u/%u to %u/%u",
|
||||
xlrec->src_tablespace_id, xlrec->src_db_id,
|
||||
@@ -36,7 +36,7 @@ dbase_desc(StringInfo buf, XLogReaderState *record)
|
||||
else if (info == XLOG_DBASE_CREATE_WAL_LOG)
|
||||
{
|
||||
xl_dbase_create_wal_log_rec *xlrec =
|
||||
(xl_dbase_create_wal_log_rec *) rec;
|
||||
(xl_dbase_create_wal_log_rec *) rec;
|
||||
|
||||
appendStringInfo(buf, "create dir %u/%u",
|
||||
xlrec->tablespace_id, xlrec->db_id);
|
||||
|
@@ -120,7 +120,7 @@ gin_desc(StringInfo buf, XLogReaderState *record)
|
||||
else
|
||||
{
|
||||
ginxlogInsertDataInternal *insertData =
|
||||
(ginxlogInsertDataInternal *) payload;
|
||||
(ginxlogInsertDataInternal *) payload;
|
||||
|
||||
appendStringInfo(buf, " pitem: %u-%u/%u",
|
||||
PostingItemGetBlockNumber(&insertData->newitem),
|
||||
@@ -156,7 +156,7 @@ gin_desc(StringInfo buf, XLogReaderState *record)
|
||||
else
|
||||
{
|
||||
ginxlogVacuumDataLeafPage *xlrec =
|
||||
(ginxlogVacuumDataLeafPage *) XLogRecGetBlockData(record, 0, NULL);
|
||||
(ginxlogVacuumDataLeafPage *) XLogRecGetBlockData(record, 0, NULL);
|
||||
|
||||
desc_recompress_leaf(buf, &xlrec->data);
|
||||
}
|
||||
|
@@ -115,7 +115,7 @@ spgAllocSearchItem(SpGistScanOpaque so, bool isnull, double *distances)
|
||||
{
|
||||
/* allocate distance array only for non-NULL items */
|
||||
SpGistSearchItem *item =
|
||||
palloc(SizeOfSpGistSearchItem(isnull ? 0 : so->numberOfNonNullOrderBys));
|
||||
palloc(SizeOfSpGistSearchItem(isnull ? 0 : so->numberOfNonNullOrderBys));
|
||||
|
||||
item->isNull = isnull;
|
||||
|
||||
@@ -130,7 +130,7 @@ static void
|
||||
spgAddStartItem(SpGistScanOpaque so, bool isnull)
|
||||
{
|
||||
SpGistSearchItem *startEntry =
|
||||
spgAllocSearchItem(so, isnull, so->zeroDistances);
|
||||
spgAllocSearchItem(so, isnull, so->zeroDistances);
|
||||
|
||||
ItemPointerSet(&startEntry->heapPtr,
|
||||
isnull ? SPGIST_NULL_BLKNO : SPGIST_ROOT_BLKNO,
|
||||
@@ -768,7 +768,7 @@ spgTestLeafTuple(SpGistScanOpaque so,
|
||||
storeRes_func storeRes)
|
||||
{
|
||||
SpGistLeafTuple leafTuple = (SpGistLeafTuple)
|
||||
PageGetItem(page, PageGetItemId(page, offset));
|
||||
PageGetItem(page, PageGetItemId(page, offset));
|
||||
|
||||
if (leafTuple->tupstate != SPGIST_LIVE)
|
||||
{
|
||||
@@ -896,7 +896,7 @@ redirect:
|
||||
else /* page is inner */
|
||||
{
|
||||
SpGistInnerTuple innerTuple = (SpGistInnerTuple)
|
||||
PageGetItem(page, PageGetItemId(page, offset));
|
||||
PageGetItem(page, PageGetItemId(page, offset));
|
||||
|
||||
if (innerTuple->tupstate != SPGIST_LIVE)
|
||||
{
|
||||
@@ -974,7 +974,7 @@ storeGettuple(SpGistScanOpaque so, ItemPointer heapPtr,
|
||||
else
|
||||
{
|
||||
IndexOrderByDistance *distances =
|
||||
palloc(sizeof(distances[0]) * so->numberOfOrderBys);
|
||||
palloc(sizeof(distances[0]) * so->numberOfOrderBys);
|
||||
int i;
|
||||
|
||||
for (i = 0; i < so->numberOfOrderBys; i++)
|
||||
|
@@ -112,7 +112,7 @@ TableScanDesc
|
||||
table_beginscan_catalog(Relation relation, int nkeys, struct ScanKeyData *key)
|
||||
{
|
||||
uint32 flags = SO_TYPE_SEQSCAN |
|
||||
SO_ALLOW_STRAT | SO_ALLOW_SYNC | SO_ALLOW_PAGEMODE | SO_TEMP_SNAPSHOT;
|
||||
SO_ALLOW_STRAT | SO_ALLOW_SYNC | SO_ALLOW_PAGEMODE | SO_TEMP_SNAPSHOT;
|
||||
Oid relid = RelationGetRelid(relation);
|
||||
Snapshot snapshot = RegisterSnapshot(GetCatalogSnapshot(relid));
|
||||
|
||||
@@ -176,7 +176,7 @@ table_beginscan_parallel(Relation relation, ParallelTableScanDesc pscan)
|
||||
{
|
||||
Snapshot snapshot;
|
||||
uint32 flags = SO_TYPE_SEQSCAN |
|
||||
SO_ALLOW_STRAT | SO_ALLOW_SYNC | SO_ALLOW_PAGEMODE;
|
||||
SO_ALLOW_STRAT | SO_ALLOW_SYNC | SO_ALLOW_PAGEMODE;
|
||||
|
||||
Assert(RelationGetRelid(relation) == pscan->phs_relid);
|
||||
|
||||
|
@@ -3270,7 +3270,7 @@ multixact_redo(XLogReaderState *record)
|
||||
else if (info == XLOG_MULTIXACT_CREATE_ID)
|
||||
{
|
||||
xl_multixact_create *xlrec =
|
||||
(xl_multixact_create *) XLogRecGetData(record);
|
||||
(xl_multixact_create *) XLogRecGetData(record);
|
||||
TransactionId max_xid;
|
||||
int i;
|
||||
|
||||
|
@@ -375,8 +375,8 @@ InitializeParallelDSM(ParallelContext *pcxt)
|
||||
shm_toc_insert(pcxt->toc, PARALLEL_KEY_COMBO_CID, combocidspace);
|
||||
|
||||
/*
|
||||
* Serialize the transaction snapshot if the transaction
|
||||
* isolation level uses a transaction snapshot.
|
||||
* Serialize the transaction snapshot if the transaction isolation
|
||||
* level uses a transaction snapshot.
|
||||
*/
|
||||
if (IsolationUsesXactSnapshot())
|
||||
{
|
||||
@@ -1497,8 +1497,8 @@ ParallelWorkerMain(Datum main_arg)
|
||||
RestoreClientConnectionInfo(clientconninfospace);
|
||||
|
||||
/*
|
||||
* Initialize SystemUser now that MyClientConnectionInfo is restored.
|
||||
* Also ensure that auth_method is actually valid, aka authn_id is not NULL.
|
||||
* Initialize SystemUser now that MyClientConnectionInfo is restored. Also
|
||||
* ensure that auth_method is actually valid, aka authn_id is not NULL.
|
||||
*/
|
||||
if (MyClientConnectionInfo.authn_id)
|
||||
InitializeSystemUser(MyClientConnectionInfo.authn_id,
|
||||
|
@@ -3152,10 +3152,9 @@ CommitTransactionCommand(void)
|
||||
break;
|
||||
|
||||
/*
|
||||
* The user issued a SAVEPOINT inside a transaction block.
|
||||
* Start a subtransaction. (DefineSavepoint already did
|
||||
* PushTransaction, so as to have someplace to put the SUBBEGIN
|
||||
* state.)
|
||||
* The user issued a SAVEPOINT inside a transaction block. Start a
|
||||
* subtransaction. (DefineSavepoint already did PushTransaction,
|
||||
* so as to have someplace to put the SUBBEGIN state.)
|
||||
*/
|
||||
case TBLOCK_SUBBEGIN:
|
||||
StartSubTransaction();
|
||||
@@ -4696,9 +4695,9 @@ RollbackAndReleaseCurrentSubTransaction(void)
|
||||
|
||||
s = CurrentTransactionState; /* changed by pop */
|
||||
Assert(s->blockState == TBLOCK_SUBINPROGRESS ||
|
||||
s->blockState == TBLOCK_INPROGRESS ||
|
||||
s->blockState == TBLOCK_IMPLICIT_INPROGRESS ||
|
||||
s->blockState == TBLOCK_STARTED);
|
||||
s->blockState == TBLOCK_INPROGRESS ||
|
||||
s->blockState == TBLOCK_IMPLICIT_INPROGRESS ||
|
||||
s->blockState == TBLOCK_STARTED);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@@ -5460,8 +5460,8 @@ StartupXLOG(void)
|
||||
missingContrecPtr = endOfRecoveryInfo->missingContrecPtr;
|
||||
|
||||
/*
|
||||
* Reset ps status display, so as no information related to recovery
|
||||
* shows up.
|
||||
* Reset ps status display, so as no information related to recovery shows
|
||||
* up.
|
||||
*/
|
||||
set_ps_display("");
|
||||
|
||||
@@ -5596,9 +5596,9 @@ StartupXLOG(void)
|
||||
if (!XLogRecPtrIsInvalid(missingContrecPtr))
|
||||
{
|
||||
/*
|
||||
* We should only have a missingContrecPtr if we're not switching to
|
||||
* a new timeline. When a timeline switch occurs, WAL is copied from
|
||||
* the old timeline to the new only up to the end of the last complete
|
||||
* We should only have a missingContrecPtr if we're not switching to a
|
||||
* new timeline. When a timeline switch occurs, WAL is copied from the
|
||||
* old timeline to the new only up to the end of the last complete
|
||||
* record, so there can't be an incomplete WAL record that we need to
|
||||
* disregard.
|
||||
*/
|
||||
@@ -8494,7 +8494,7 @@ do_pg_backup_start(const char *backupidstr, bool fast, List **tablespaces,
|
||||
*/
|
||||
if (rllen > datadirpathlen &&
|
||||
strncmp(linkpath, DataDir, datadirpathlen) == 0 &&
|
||||
IS_DIR_SEP(linkpath[datadirpathlen]))
|
||||
IS_DIR_SEP(linkpath[datadirpathlen]))
|
||||
relpath = pstrdup(linkpath + datadirpathlen + 1);
|
||||
|
||||
/*
|
||||
|
@@ -897,8 +897,8 @@ XLogRecordAssemble(RmgrId rmid, uint8 info,
|
||||
*
|
||||
* XLogReader machinery is only able to handle records up to a certain
|
||||
* size (ignoring machine resource limitations), so make sure that we will
|
||||
* not emit records larger than the sizes advertised to be supported.
|
||||
* This cap is based on DecodeXLogRecordRequiredSpace().
|
||||
* not emit records larger than the sizes advertised to be supported. This
|
||||
* cap is based on DecodeXLogRecordRequiredSpace().
|
||||
*/
|
||||
if (total_len >= XLogRecordMaxSize)
|
||||
ereport(ERROR,
|
||||
|
@@ -569,7 +569,7 @@ XLogPrefetcherNextBlock(uintptr_t pgsr_private, XLogRecPtr *lsn)
|
||||
if (record_type == XLOG_DBASE_CREATE_FILE_COPY)
|
||||
{
|
||||
xl_dbase_create_file_copy_rec *xlrec =
|
||||
(xl_dbase_create_file_copy_rec *) record->main_data;
|
||||
(xl_dbase_create_file_copy_rec *) record->main_data;
|
||||
RelFileLocator rlocator =
|
||||
{InvalidOid, xlrec->db_id, InvalidRelFileNumber};
|
||||
|
||||
@@ -596,7 +596,7 @@ XLogPrefetcherNextBlock(uintptr_t pgsr_private, XLogRecPtr *lsn)
|
||||
if (record_type == XLOG_SMGR_CREATE)
|
||||
{
|
||||
xl_smgr_create *xlrec = (xl_smgr_create *)
|
||||
record->main_data;
|
||||
record->main_data;
|
||||
|
||||
if (xlrec->forkNum == MAIN_FORKNUM)
|
||||
{
|
||||
@@ -624,7 +624,7 @@ XLogPrefetcherNextBlock(uintptr_t pgsr_private, XLogRecPtr *lsn)
|
||||
else if (record_type == XLOG_SMGR_TRUNCATE)
|
||||
{
|
||||
xl_smgr_truncate *xlrec = (xl_smgr_truncate *)
|
||||
record->main_data;
|
||||
record->main_data;
|
||||
|
||||
/*
|
||||
* Don't consider prefetching anything in the truncated
|
||||
|
@@ -282,7 +282,7 @@ XLogRecPtr
|
||||
XLogReleasePreviousRecord(XLogReaderState *state)
|
||||
{
|
||||
DecodedXLogRecord *record;
|
||||
XLogRecPtr next_lsn;
|
||||
XLogRecPtr next_lsn;
|
||||
|
||||
if (!state->record)
|
||||
return InvalidXLogRecPtr;
|
||||
|
@@ -3215,7 +3215,7 @@ XLogPageRead(XLogReaderState *xlogreader, XLogRecPtr targetPagePtr, int reqLen,
|
||||
XLogRecPtr targetRecPtr, char *readBuf)
|
||||
{
|
||||
XLogPageReadPrivate *private =
|
||||
(XLogPageReadPrivate *) xlogreader->private_data;
|
||||
(XLogPageReadPrivate *) xlogreader->private_data;
|
||||
int emode = private->emode;
|
||||
uint32 targetPageOff;
|
||||
XLogSegNo targetSegNo PG_USED_FOR_ASSERTS_ONLY;
|
||||
|
Reference in New Issue
Block a user