mirror of
https://github.com/postgres/postgres.git
synced 2025-11-06 07:49:08 +03:00
Final pgindent + perltidy run for v10.
This commit is contained in:
@@ -1454,7 +1454,7 @@ BeginCopy(ParseState *pstate,
|
||||
*/
|
||||
if (cstate->transition_capture != NULL)
|
||||
{
|
||||
int i;
|
||||
int i;
|
||||
|
||||
cstate->transition_tupconv_maps = (TupleConversionMap **)
|
||||
palloc0(sizeof(TupleConversionMap *) *
|
||||
@@ -2651,6 +2651,7 @@ CopyFrom(CopyState cstate)
|
||||
cstate->transition_capture->tcs_map = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* We might need to convert from the parent rowtype to the
|
||||
* partition rowtype.
|
||||
|
||||
@@ -919,9 +919,10 @@ DropSubscription(DropSubscriptionStmt *stmt, bool isTopLevel)
|
||||
LWLockAcquire(LogicalRepWorkerLock, LW_SHARED);
|
||||
subworkers = logicalrep_workers_find(subid, false);
|
||||
LWLockRelease(LogicalRepWorkerLock);
|
||||
foreach (lc, subworkers)
|
||||
foreach(lc, subworkers)
|
||||
{
|
||||
LogicalRepWorker *w = (LogicalRepWorker *) lfirst(lc);
|
||||
|
||||
if (slotname)
|
||||
logicalrep_worker_stop(w->subid, w->relid);
|
||||
else
|
||||
|
||||
@@ -13509,8 +13509,8 @@ ATExecAttachPartition(List **wqueue, Relation rel, PartitionCmd *cmd)
|
||||
* having to construct this list again, so we request the strongest lock
|
||||
* on all partitions. We need the strongest lock, because we may decide
|
||||
* to scan them if we find out that the table being attached (or its leaf
|
||||
* partitions) may contain rows that violate the partition constraint.
|
||||
* If the table has a constraint that would prevent such rows, which by
|
||||
* partitions) may contain rows that violate the partition constraint. If
|
||||
* the table has a constraint that would prevent such rows, which by
|
||||
* definition is present in all the partitions, we need not scan the
|
||||
* table, nor its partitions. But we cannot risk a deadlock by taking a
|
||||
* weaker lock now and the stronger one only when needed.
|
||||
|
||||
@@ -2071,11 +2071,11 @@ FindTriggerIncompatibleWithInheritance(TriggerDesc *trigdesc)
|
||||
{
|
||||
if (trigdesc != NULL)
|
||||
{
|
||||
int i;
|
||||
int i;
|
||||
|
||||
for (i = 0; i < trigdesc->numtriggers; ++i)
|
||||
{
|
||||
Trigger *trigger = &trigdesc->triggers[i];
|
||||
Trigger *trigger = &trigdesc->triggers[i];
|
||||
|
||||
if (trigger->tgoldtable != NULL || trigger->tgnewtable != NULL)
|
||||
return trigger->tgname;
|
||||
@@ -5253,12 +5253,12 @@ AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
|
||||
*/
|
||||
if (row_trigger && transition_capture != NULL)
|
||||
{
|
||||
HeapTuple original_insert_tuple = transition_capture->tcs_original_insert_tuple;
|
||||
HeapTuple original_insert_tuple = transition_capture->tcs_original_insert_tuple;
|
||||
TupleConversionMap *map = transition_capture->tcs_map;
|
||||
bool delete_old_table = transition_capture->tcs_delete_old_table;
|
||||
bool update_old_table = transition_capture->tcs_update_old_table;
|
||||
bool update_new_table = transition_capture->tcs_update_new_table;
|
||||
bool insert_new_table = transition_capture->tcs_insert_new_table;;
|
||||
bool delete_old_table = transition_capture->tcs_delete_old_table;
|
||||
bool update_old_table = transition_capture->tcs_update_old_table;
|
||||
bool update_new_table = transition_capture->tcs_update_new_table;
|
||||
bool insert_new_table = transition_capture->tcs_insert_new_table;;
|
||||
|
||||
if ((event == TRIGGER_EVENT_DELETE && delete_old_table) ||
|
||||
(event == TRIGGER_EVENT_UPDATE && update_old_table))
|
||||
|
||||
@@ -529,11 +529,11 @@ lazy_scan_heap(Relation onerel, int options, LVRelStats *vacrelstats,
|
||||
* safely set for relfrozenxid or relminmxid.
|
||||
*
|
||||
* Before entering the main loop, establish the invariant that
|
||||
* next_unskippable_block is the next block number >= blkno that we
|
||||
* can't skip based on the visibility map, either all-visible for a
|
||||
* regular scan or all-frozen for an aggressive scan. We set it to
|
||||
* nblocks if there's no such block. We also set up the skipping_blocks
|
||||
* flag correctly at this stage.
|
||||
* next_unskippable_block is the next block number >= blkno that we can't
|
||||
* skip based on the visibility map, either all-visible for a regular scan
|
||||
* or all-frozen for an aggressive scan. We set it to nblocks if there's
|
||||
* no such block. We also set up the skipping_blocks flag correctly at
|
||||
* this stage.
|
||||
*
|
||||
* Note: The value returned by visibilitymap_get_status could be slightly
|
||||
* out-of-date, since we make this test before reading the corresponding
|
||||
|
||||
Reference in New Issue
Block a user