1
0
mirror of https://github.com/postgres/postgres.git synced 2025-07-07 00:36:50 +03:00

Fix more typos and inconsistencies in the tree

Author: Alexander Lakhin
Discussion: https://postgr.es/m/0a5419ea-1452-a4e6-72ff-545b1a5a8076@gmail.com
This commit is contained in:
Michael Paquier
2019-06-17 16:13:16 +09:00
parent 9d20b0ec8f
commit 3412030205
41 changed files with 56 additions and 56 deletions

View File

@ -2876,8 +2876,9 @@ EvalPlanQualStart(EPQState *epqstate, EState *parentestate, Plan *planTree)
/*
* Each EState must have its own es_epqScanDone state, but if we have
* nested EPQ checks they should share es_epqTuple arrays. This allows
* sub-rechecks to inherit the values being examined by an outer recheck.
* nested EPQ checks they should share es_epqTupleSlot arrays. This
* allows sub-rechecks to inherit the values being examined by an outer
* recheck.
*/
estate->es_epqScanDone = (bool *) palloc0(rtsize * sizeof(bool));
if (parentestate->es_epqTupleSlot != NULL)

View File

@ -642,7 +642,7 @@ ExecAssignScanType(ScanState *scanstate, TupleDesc tupDesc)
}
/* ----------------
* ExecCreateSlotFromOuterPlan
* ExecCreateScanSlotFromOuterPlan
* ----------------
*/
void

View File

@ -1049,8 +1049,8 @@ ExecHashIncreaseNumBatches(HashJoinTable hashtable)
/*
* ExecParallelHashIncreaseNumBatches
* Every participant attached to grow_barrier must run this function
* when it observes growth == PHJ_GROWTH_NEED_MORE_BATCHES.
* Every participant attached to grow_batches_barrier must run this
* function when it observes growth == PHJ_GROWTH_NEED_MORE_BATCHES.
*/
static void
ExecParallelHashIncreaseNumBatches(HashJoinTable hashtable)
@ -1106,7 +1106,7 @@ ExecParallelHashIncreaseNumBatches(HashJoinTable hashtable)
* The combined work_mem of all participants wasn't
* enough. Therefore one batch per participant would be
* approximately equivalent and would probably also be
* insufficient. So try two batches per particiant,
* insufficient. So try two batches per participant,
* rounded up to a power of two.
*/
new_nbatch = 1 << my_log2(pstate->nparticipants * 2);
@ -1674,7 +1674,7 @@ ExecHashTableInsert(HashJoinTable hashtable,
}
/*
* ExecHashTableParallelInsert
* ExecParallelHashTableInsert
* insert a tuple into a shared hash table or shared batch tuplestore
*/
void

View File

@ -297,11 +297,12 @@ ExecInitProjectSet(ProjectSet *node, EState *estate, int eflags)
Assert(node->plan.qual == NIL);
/*
* Create a memory context that ExecMakeFunctionResult can use to evaluate
* function arguments in. We can't use the per-tuple context for this
* because it gets reset too often; but we don't want to leak evaluation
* results into the query-lifespan context either. We use one context for
* the arguments of all tSRFs, as they have roughly equivalent lifetimes.
* Create a memory context that ExecMakeFunctionResultSet can use to
* evaluate function arguments in. We can't use the per-tuple context for
* this because it gets reset too often; but we don't want to leak
* evaluation results into the query-lifespan context either. We use one
* context for the arguments of all tSRFs, as they have roughly equivalent
* lifetimes.
*/
state->argcontext = AllocSetContextCreate(CurrentMemoryContext,
"tSRF function arguments",