mirror of
https://github.com/postgres/postgres.git
synced 2025-07-07 00:36:50 +03:00
Fix more typos and inconsistencies in the tree
Author: Alexander Lakhin Discussion: https://postgr.es/m/0a5419ea-1452-a4e6-72ff-545b1a5a8076@gmail.com
This commit is contained in:
@ -2876,8 +2876,9 @@ EvalPlanQualStart(EPQState *epqstate, EState *parentestate, Plan *planTree)
|
||||
|
||||
/*
|
||||
* Each EState must have its own es_epqScanDone state, but if we have
|
||||
* nested EPQ checks they should share es_epqTuple arrays. This allows
|
||||
* sub-rechecks to inherit the values being examined by an outer recheck.
|
||||
* nested EPQ checks they should share es_epqTupleSlot arrays. This
|
||||
* allows sub-rechecks to inherit the values being examined by an outer
|
||||
* recheck.
|
||||
*/
|
||||
estate->es_epqScanDone = (bool *) palloc0(rtsize * sizeof(bool));
|
||||
if (parentestate->es_epqTupleSlot != NULL)
|
||||
|
@ -642,7 +642,7 @@ ExecAssignScanType(ScanState *scanstate, TupleDesc tupDesc)
|
||||
}
|
||||
|
||||
/* ----------------
|
||||
* ExecCreateSlotFromOuterPlan
|
||||
* ExecCreateScanSlotFromOuterPlan
|
||||
* ----------------
|
||||
*/
|
||||
void
|
||||
|
@ -1049,8 +1049,8 @@ ExecHashIncreaseNumBatches(HashJoinTable hashtable)
|
||||
|
||||
/*
|
||||
* ExecParallelHashIncreaseNumBatches
|
||||
* Every participant attached to grow_barrier must run this function
|
||||
* when it observes growth == PHJ_GROWTH_NEED_MORE_BATCHES.
|
||||
* Every participant attached to grow_batches_barrier must run this
|
||||
* function when it observes growth == PHJ_GROWTH_NEED_MORE_BATCHES.
|
||||
*/
|
||||
static void
|
||||
ExecParallelHashIncreaseNumBatches(HashJoinTable hashtable)
|
||||
@ -1106,7 +1106,7 @@ ExecParallelHashIncreaseNumBatches(HashJoinTable hashtable)
|
||||
* The combined work_mem of all participants wasn't
|
||||
* enough. Therefore one batch per participant would be
|
||||
* approximately equivalent and would probably also be
|
||||
* insufficient. So try two batches per particiant,
|
||||
* insufficient. So try two batches per participant,
|
||||
* rounded up to a power of two.
|
||||
*/
|
||||
new_nbatch = 1 << my_log2(pstate->nparticipants * 2);
|
||||
@ -1674,7 +1674,7 @@ ExecHashTableInsert(HashJoinTable hashtable,
|
||||
}
|
||||
|
||||
/*
|
||||
* ExecHashTableParallelInsert
|
||||
* ExecParallelHashTableInsert
|
||||
* insert a tuple into a shared hash table or shared batch tuplestore
|
||||
*/
|
||||
void
|
||||
|
@ -297,11 +297,12 @@ ExecInitProjectSet(ProjectSet *node, EState *estate, int eflags)
|
||||
Assert(node->plan.qual == NIL);
|
||||
|
||||
/*
|
||||
* Create a memory context that ExecMakeFunctionResult can use to evaluate
|
||||
* function arguments in. We can't use the per-tuple context for this
|
||||
* because it gets reset too often; but we don't want to leak evaluation
|
||||
* results into the query-lifespan context either. We use one context for
|
||||
* the arguments of all tSRFs, as they have roughly equivalent lifetimes.
|
||||
* Create a memory context that ExecMakeFunctionResultSet can use to
|
||||
* evaluate function arguments in. We can't use the per-tuple context for
|
||||
* this because it gets reset too often; but we don't want to leak
|
||||
* evaluation results into the query-lifespan context either. We use one
|
||||
* context for the arguments of all tSRFs, as they have roughly equivalent
|
||||
* lifetimes.
|
||||
*/
|
||||
state->argcontext = AllocSetContextCreate(CurrentMemoryContext,
|
||||
"tSRF function arguments",
|
||||
|
Reference in New Issue
Block a user