1
0
mirror of https://github.com/postgres/postgres.git synced 2025-06-05 23:56:58 +03:00

Fix misc typos.

Oskari Saarenmaa. Backpatch to stable branches where applicable.
This commit is contained in:
Heikki Linnakangas 2015-09-05 11:35:49 +03:00
parent 9e9b310d8b
commit 658ec62640
8 changed files with 11 additions and 11 deletions

View File

@ -374,7 +374,7 @@ gbt_ts_penalty(PG_FUNCTION_ARGS)
newdbl[2]; newdbl[2];
/* /*
* We are allways using "double" timestamps here. Precision should be good * We are always using "double" timestamps here. Precision should be good
* enough. * enough.
*/ */
orgdbl[0] = ((double) origentry->lower); orgdbl[0] = ((double) origentry->lower);

View File

@ -52,7 +52,7 @@ gbt_var_decompress(PG_FUNCTION_ARGS)
PG_RETURN_POINTER(entry); PG_RETURN_POINTER(entry);
} }
/* Returns a better readable representaion of variable key ( sets pointer ) */ /* Returns a better readable representation of variable key ( sets pointer ) */
GBT_VARKEY_R GBT_VARKEY_R
gbt_var_key_readable(const GBT_VARKEY *k) gbt_var_key_readable(const GBT_VARKEY *k)
{ {

View File

@ -807,7 +807,7 @@ heap_modify_tuple(HeapTuple tuple,
* repl information, as appropriate. * repl information, as appropriate.
* *
* NOTE: it's debatable whether to use heap_deform_tuple() here or just * NOTE: it's debatable whether to use heap_deform_tuple() here or just
* heap_getattr() only the non-replaced colums. The latter could win if * heap_getattr() only the non-replaced columns. The latter could win if
* there are many replaced columns and few non-replaced ones. However, * there are many replaced columns and few non-replaced ones. However,
* heap_deform_tuple costs only O(N) while the heap_getattr way would cost * heap_deform_tuple costs only O(N) while the heap_getattr way would cost
* O(N^2) if there are many non-replaced columns, so it seems better to * O(N^2) if there are many non-replaced columns, so it seems better to

View File

@ -881,8 +881,8 @@ ginInsertCleanup(GinState *ginstate,
* locking */ * locking */
/* /*
* remove readed pages from pending list, at this point all * remove read pages from pending list, at this point all
* content of readed pages is in regular structure * content of read pages is in regular structure
*/ */
if (shiftList(index, metabuffer, blkno, stats)) if (shiftList(index, metabuffer, blkno, stats))
{ {

View File

@ -578,7 +578,7 @@ gist_box_picksplit(PG_FUNCTION_ARGS)
* We first consider splits where b is the lower bound of an entry. * We first consider splits where b is the lower bound of an entry.
* We iterate through all entries, and for each b, calculate the * We iterate through all entries, and for each b, calculate the
* smallest possible a. Then we consider splits where a is the * smallest possible a. Then we consider splits where a is the
* uppper bound of an entry, and for each a, calculate the greatest * upper bound of an entry, and for each a, calculate the greatest
* possible b. * possible b.
* *
* In the above example, the first loop would consider splits: * In the above example, the first loop would consider splits:
@ -628,7 +628,7 @@ gist_box_picksplit(PG_FUNCTION_ARGS)
} }
/* /*
* Iterate over upper bound of left group finding greates possible * Iterate over upper bound of left group finding greatest possible
* lower bound of right group. * lower bound of right group.
*/ */
i1 = nentries - 1; i1 = nentries - 1;

View File

@ -5115,7 +5115,7 @@ l4:
* *
* The initial tuple is assumed to be already locked. * The initial tuple is assumed to be already locked.
* *
* This function doesn't check visibility, it just inconditionally marks the * This function doesn't check visibility, it just unconditionally marks the
* tuple(s) as locked. If any tuple in the updated chain is being deleted * tuple(s) as locked. If any tuple in the updated chain is being deleted
* concurrently (or updated with the key being modified), sleep until the * concurrently (or updated with the key being modified), sleep until the
* transaction doing it is finished. * transaction doing it is finished.
@ -5608,7 +5608,7 @@ heap_prepare_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid,
/* /*
* NB -- some of these transformations are only valid because * NB -- some of these transformations are only valid because
* we know the return Xid is a tuple updater (i.e. not merely a * we know the return Xid is a tuple updater (i.e. not merely a
* locker.) Also note that the only reason we don't explicitely * locker.) Also note that the only reason we don't explicitly
* worry about HEAP_KEYS_UPDATED is because it lives in t_infomask2 * worry about HEAP_KEYS_UPDATED is because it lives in t_infomask2
* rather than t_infomask. * rather than t_infomask.
*/ */

View File

@ -931,7 +931,7 @@ cost_tidscan(Path *path, PlannerInfo *root,
/* /*
* The TID qual expressions will be computed once, any other baserestrict * The TID qual expressions will be computed once, any other baserestrict
* quals once per retrived tuple. * quals once per retrieved tuple.
*/ */
cost_qual_eval(&tid_qual_cost, tidquals, root); cost_qual_eval(&tid_qual_cost, tidquals, root);

View File

@ -698,7 +698,7 @@ format_operator_internal(Oid operator_oid, bool force_qualify)
/* /*
* Would this oper be found (given the right args) by regoperatorin? * Would this oper be found (given the right args) by regoperatorin?
* If not, or if caller explicitely requests it, we need to qualify * If not, or if caller explicitly requests it, we need to qualify
* it. * it.
*/ */
if (force_qualify || !OperatorIsVisible(operator_oid)) if (force_qualify || !OperatorIsVisible(operator_oid))