mirror of
https://github.com/postgres/postgres.git
synced 2025-07-21 16:02:15 +03:00
Harmonize parameter names in storage and AM code.
Make sure that function declarations use names that exactly match the corresponding names from function definitions in storage, catalog, access method, executor, and logical replication code, as well as in miscellaneous utility/library code. Like other recent commits that cleaned up function parameter names, this commit was written with help from clang-tidy. Later commits will do the same for other parts of the codebase. Author: Peter Geoghegan <pg@bowt.ie> Reviewed-By: David Rowley <dgrowleyml@gmail.com> Discussion: https://postgr.es/m/CAH2-WznJt9CMM9KJTMjJh_zbL5hD9oX44qdJ4aqZtjFi-zA3Tg@mail.gmail.com
This commit is contained in:
@ -130,7 +130,7 @@ static bool check_exclusion_or_unique_constraint(Relation heap, Relation index,
|
||||
Datum *values, bool *isnull,
|
||||
EState *estate, bool newIndex,
|
||||
CEOUC_WAIT_MODE waitMode,
|
||||
bool errorOK,
|
||||
bool violationOK,
|
||||
ItemPointer conflictTid);
|
||||
|
||||
static bool index_recheck_constraint(Relation index, Oid *constr_procs,
|
||||
|
@ -126,9 +126,9 @@ typedef struct ExecParallelInitializeDSMContext
|
||||
|
||||
/* Helper functions that run in the parallel leader. */
|
||||
static char *ExecSerializePlan(Plan *plan, EState *estate);
|
||||
static bool ExecParallelEstimate(PlanState *node,
|
||||
static bool ExecParallelEstimate(PlanState *planstate,
|
||||
ExecParallelEstimateContext *e);
|
||||
static bool ExecParallelInitializeDSM(PlanState *node,
|
||||
static bool ExecParallelInitializeDSM(PlanState *planstate,
|
||||
ExecParallelInitializeDSMContext *d);
|
||||
static shm_mq_handle **ExecParallelSetupTupleQueues(ParallelContext *pcxt,
|
||||
bool reinitialize);
|
||||
|
@ -396,7 +396,7 @@ static void prepare_projection_slot(AggState *aggstate,
|
||||
TupleTableSlot *slot,
|
||||
int currentSet);
|
||||
static void finalize_aggregates(AggState *aggstate,
|
||||
AggStatePerAgg peragg,
|
||||
AggStatePerAgg peraggs,
|
||||
AggStatePerGroup pergroup);
|
||||
static TupleTableSlot *project_aggregates(AggState *aggstate);
|
||||
static void find_cols(AggState *aggstate, Bitmapset **aggregated,
|
||||
@ -407,12 +407,11 @@ static void build_hash_table(AggState *aggstate, int setno, long nbuckets);
|
||||
static void hashagg_recompile_expressions(AggState *aggstate, bool minslot,
|
||||
bool nullcheck);
|
||||
static long hash_choose_num_buckets(double hashentrysize,
|
||||
long estimated_nbuckets,
|
||||
Size memory);
|
||||
long ngroups, Size memory);
|
||||
static int hash_choose_num_partitions(double input_groups,
|
||||
double hashentrysize,
|
||||
int used_bits,
|
||||
int *log2_npartittions);
|
||||
int *log2_npartitions);
|
||||
static void initialize_hash_entry(AggState *aggstate,
|
||||
TupleHashTable hashtable,
|
||||
TupleHashEntry entry);
|
||||
@ -432,11 +431,11 @@ static HashAggBatch *hashagg_batch_new(LogicalTape *input_tape, int setno,
|
||||
int64 input_tuples, double input_card,
|
||||
int used_bits);
|
||||
static MinimalTuple hashagg_batch_read(HashAggBatch *batch, uint32 *hashp);
|
||||
static void hashagg_spill_init(HashAggSpill *spill, LogicalTapeSet *lts,
|
||||
static void hashagg_spill_init(HashAggSpill *spill, LogicalTapeSet *tapeset,
|
||||
int used_bits, double input_groups,
|
||||
double hashentrysize);
|
||||
static Size hashagg_spill_tuple(AggState *aggstate, HashAggSpill *spill,
|
||||
TupleTableSlot *slot, uint32 hash);
|
||||
TupleTableSlot *inputslot, uint32 hash);
|
||||
static void hashagg_spill_finish(AggState *aggstate, HashAggSpill *spill,
|
||||
int setno);
|
||||
static Datum GetAggInitVal(Datum textInitVal, Oid transtype);
|
||||
|
@ -62,9 +62,9 @@ static HashJoinTuple ExecParallelHashTupleAlloc(HashJoinTable hashtable,
|
||||
dsa_pointer *shared);
|
||||
static void MultiExecPrivateHash(HashState *node);
|
||||
static void MultiExecParallelHash(HashState *node);
|
||||
static inline HashJoinTuple ExecParallelHashFirstTuple(HashJoinTable table,
|
||||
static inline HashJoinTuple ExecParallelHashFirstTuple(HashJoinTable hashtable,
|
||||
int bucketno);
|
||||
static inline HashJoinTuple ExecParallelHashNextTuple(HashJoinTable table,
|
||||
static inline HashJoinTuple ExecParallelHashNextTuple(HashJoinTable hashtable,
|
||||
HashJoinTuple tuple);
|
||||
static inline void ExecParallelHashPushTuple(dsa_pointer_atomic *head,
|
||||
HashJoinTuple tuple,
|
||||
@ -73,7 +73,7 @@ static void ExecParallelHashJoinSetUpBatches(HashJoinTable hashtable, int nbatch
|
||||
static void ExecParallelHashEnsureBatchAccessors(HashJoinTable hashtable);
|
||||
static void ExecParallelHashRepartitionFirst(HashJoinTable hashtable);
|
||||
static void ExecParallelHashRepartitionRest(HashJoinTable hashtable);
|
||||
static HashMemoryChunk ExecParallelHashPopChunkQueue(HashJoinTable table,
|
||||
static HashMemoryChunk ExecParallelHashPopChunkQueue(HashJoinTable hashtable,
|
||||
dsa_pointer *shared);
|
||||
static bool ExecParallelHashTuplePrealloc(HashJoinTable hashtable,
|
||||
int batchno,
|
||||
|
@ -145,7 +145,7 @@ static TupleTableSlot *ExecHashJoinGetSavedTuple(HashJoinState *hjstate,
|
||||
TupleTableSlot *tupleSlot);
|
||||
static bool ExecHashJoinNewBatch(HashJoinState *hjstate);
|
||||
static bool ExecParallelHashJoinNewBatch(HashJoinState *hjstate);
|
||||
static void ExecParallelHashJoinPartitionOuter(HashJoinState *node);
|
||||
static void ExecParallelHashJoinPartitionOuter(HashJoinState *hjstate);
|
||||
|
||||
|
||||
/* ----------------------------------------------------------------
|
||||
@ -1502,11 +1502,11 @@ ExecHashJoinInitializeDSM(HashJoinState *state, ParallelContext *pcxt)
|
||||
* ----------------------------------------------------------------
|
||||
*/
|
||||
void
|
||||
ExecHashJoinReInitializeDSM(HashJoinState *state, ParallelContext *cxt)
|
||||
ExecHashJoinReInitializeDSM(HashJoinState *state, ParallelContext *pcxt)
|
||||
{
|
||||
int plan_node_id = state->js.ps.plan->plan_node_id;
|
||||
ParallelHashJoinState *pstate =
|
||||
shm_toc_lookup(cxt->toc, plan_node_id, false);
|
||||
shm_toc_lookup(pcxt->toc, plan_node_id, false);
|
||||
|
||||
/*
|
||||
* It would be possible to reuse the shared hash table in single-batch
|
||||
|
@ -133,8 +133,8 @@ typedef struct MemoizeEntry
|
||||
static uint32 MemoizeHash_hash(struct memoize_hash *tb,
|
||||
const MemoizeKey *key);
|
||||
static bool MemoizeHash_equal(struct memoize_hash *tb,
|
||||
const MemoizeKey *params1,
|
||||
const MemoizeKey *params2);
|
||||
const MemoizeKey *key1,
|
||||
const MemoizeKey *key2);
|
||||
|
||||
#define SH_PREFIX memoize
|
||||
#define SH_ELEMENT_TYPE MemoizeEntry
|
||||
|
Reference in New Issue
Block a user