mirror of
https://github.com/postgres/postgres.git
synced 2025-08-30 06:01:21 +03:00
Fix typos in comments.
Backpatch to all supported versions, where applicable, to make backpatching of future fixes go more smoothly. Josh Soref Discussion: https://www.postgresql.org/message-id/CACZqfqCf+5qRztLPgmmosr-B0Ye4srWzzw_mo4c_8_B_mtjmJQ@mail.gmail.com
This commit is contained in:
@@ -28,7 +28,7 @@ The current implementation of GiST supports:
|
||||
|
||||
The support for concurrency implemented in PostgreSQL was developed based on
|
||||
the paper "Access Methods for Next-Generation Database Systems" by
|
||||
Marcel Kornaker:
|
||||
Marcel Kornacker:
|
||||
|
||||
http://www.sai.msu.su/~megera/postgres/gist/papers/concurrency/access-methods-for-next-generation.pdf.gz
|
||||
|
||||
|
@@ -671,7 +671,7 @@ createdb(const CreatedbStmt *stmt)
|
||||
|
||||
/*
|
||||
* Force synchronous commit, thus minimizing the window between
|
||||
* creation of the database files and commital of the transaction. If
|
||||
* creation of the database files and committal of the transaction. If
|
||||
* we crash before committing, we'll have a DB that's taking up disk
|
||||
* space but is not in pg_database, which is not good.
|
||||
*/
|
||||
@@ -916,7 +916,7 @@ dropdb(const char *dbname, bool missing_ok)
|
||||
|
||||
/*
|
||||
* Force synchronous commit, thus minimizing the window between removal of
|
||||
* the database files and commital of the transaction. If we crash before
|
||||
* the database files and committal of the transaction. If we crash before
|
||||
* committing, we'll have a DB that's gone on disk but still there
|
||||
* according to pg_database, which is not good.
|
||||
*/
|
||||
@@ -1266,7 +1266,7 @@ movedb(const char *dbname, const char *tblspcname)
|
||||
|
||||
/*
|
||||
* Force synchronous commit, thus minimizing the window between
|
||||
* copying the database files and commital of the transaction. If we
|
||||
* copying the database files and committal of the transaction. If we
|
||||
* crash before committing, we'll leave an orphaned set of files on
|
||||
* disk, which is not fatal but not good either.
|
||||
*/
|
||||
|
@@ -2438,7 +2438,7 @@ ExplainYAMLLineStarting(ExplainState *es)
|
||||
}
|
||||
|
||||
/*
|
||||
* YAML is a superset of JSON; unfortuantely, the YAML quoting rules are
|
||||
* YAML is a superset of JSON; unfortunately, the YAML quoting rules are
|
||||
* ridiculously complicated -- as documented in sections 5.3 and 7.3.3 of
|
||||
* http://yaml.org/spec/1.2/spec.html -- so we chose to just quote everything.
|
||||
* Empty strings, strings with leading or trailing whitespace, and strings
|
||||
|
@@ -1524,7 +1524,7 @@ CreateCast(CreateCastStmt *stmt)
|
||||
(errcode(ERRCODE_WRONG_OBJECT_TYPE),
|
||||
errmsg("cast will be ignored because the target data type is a domain")));
|
||||
|
||||
/* Detemine the cast method */
|
||||
/* Determine the cast method */
|
||||
if (stmt->func != NULL)
|
||||
castmethod = COERCION_METHOD_FUNCTION;
|
||||
else if (stmt->inout)
|
||||
|
@@ -95,7 +95,7 @@ static void RangeVarCallbackForReindexIndex(const RangeVar *relation,
|
||||
* Errors arising from the attribute list still apply.
|
||||
*
|
||||
* Most column type changes that can skip a table rewrite do not invalidate
|
||||
* indexes. We ackowledge this when all operator classes, collations and
|
||||
* indexes. We acknowledge this when all operator classes, collations and
|
||||
* exclusion operators match. Though we could further permit intra-opfamily
|
||||
* changes for btree and hash indexes, that adds subtle complexity with no
|
||||
* concrete benefit for core types.
|
||||
@@ -920,7 +920,7 @@ CheckMutability(Expr *expr)
|
||||
* indxpath.c could do something with. However, that seems overly
|
||||
* restrictive. One useful application of partial indexes is to apply
|
||||
* a UNIQUE constraint across a subset of a table, and in that scenario
|
||||
* any evaluatable predicate will work. So accept any predicate here
|
||||
* any evaluable predicate will work. So accept any predicate here
|
||||
* (except ones requiring a plan), and let indxpath.c fend for itself.
|
||||
*/
|
||||
static void
|
||||
|
@@ -5661,7 +5661,7 @@ ATAddCheckConstraint(List **wqueue, AlteredTableInfo *tab, Relation rel,
|
||||
|
||||
/*
|
||||
* Check if ONLY was specified with ALTER TABLE. If so, allow the
|
||||
* contraint creation only if there are no children currently. Error out
|
||||
* constraint creation only if there are no children currently. Error out
|
||||
* otherwise.
|
||||
*/
|
||||
if (!recurse && children != NIL)
|
||||
|
@@ -111,7 +111,7 @@ gimme_edge_table(PlannerInfo *root, Gene *tour1, Gene *tour2,
|
||||
for (index1 = 0; index1 < num_gene; index1++)
|
||||
{
|
||||
/*
|
||||
* presume the tour is circular, i.e. 1->2, 2->3, 3->1 this operaton
|
||||
* presume the tour is circular, i.e. 1->2, 2->3, 3->1 this operation
|
||||
* maps n back to 1
|
||||
*/
|
||||
|
||||
@@ -314,7 +314,7 @@ gimme_gene(PlannerInfo *root, Edge edge, Edge *edge_table)
|
||||
/*
|
||||
* give priority to candidates with fewest remaining unused edges;
|
||||
* find out what the minimum number of unused edges is
|
||||
* (minimum_edges); if there is more than one cadidate with the
|
||||
* (minimum_edges); if there is more than one candidate with the
|
||||
* minimum number of unused edges keep count of this number
|
||||
* (minimum_count);
|
||||
*/
|
||||
|
@@ -1367,7 +1367,7 @@ select_mergejoin_clauses(PlannerInfo *root,
|
||||
/*
|
||||
* Insist that each side have a non-redundant eclass. This
|
||||
* restriction is needed because various bits of the planner expect
|
||||
* that each clause in a merge be associatable with some pathkey in a
|
||||
* that each clause in a merge be associable with some pathkey in a
|
||||
* canonical pathkey list, but redundant eclasses can't appear in
|
||||
* canonical sort orderings. (XXX it might be worth relaxing this,
|
||||
* but not enough time to address it for 8.3.)
|
||||
|
@@ -212,7 +212,7 @@ query_planner(PlannerInfo *root, List *tlist,
|
||||
/*
|
||||
* Now distribute "placeholders" to base rels as needed. This has to be
|
||||
* done after join removal because removal could change whether a
|
||||
* placeholder is evaluatable at a base rel.
|
||||
* placeholder is evaluable at a base rel.
|
||||
*/
|
||||
add_placeholders_to_base_rels(root);
|
||||
|
||||
|
@@ -24,7 +24,7 @@
|
||||
* Detect whether there is a joinclause that involves
|
||||
* the two given relations.
|
||||
*
|
||||
* Note: the joinclause does not have to be evaluatable with only these two
|
||||
* Note: the joinclause does not have to be evaluable with only these two
|
||||
* relations. This is intentional. For example consider
|
||||
* SELECT * FROM a, b, c WHERE a.x = (b.y + c.z)
|
||||
* If a is much larger than the other tables, it may be worthwhile to
|
||||
|
@@ -718,7 +718,7 @@ join_clause_is_movable_into(RestrictInfo *rinfo,
|
||||
Relids currentrelids,
|
||||
Relids current_and_outer)
|
||||
{
|
||||
/* Clause must be evaluatable given available context */
|
||||
/* Clause must be evaluable given available context */
|
||||
if (!bms_is_subset(rinfo->clause_relids, current_and_outer))
|
||||
return false;
|
||||
|
||||
|
@@ -4604,7 +4604,7 @@ PostmasterRandom(void)
|
||||
}
|
||||
|
||||
/*
|
||||
* Count up number of child processes of specified types (dead_end chidren
|
||||
* Count up number of child processes of specified types (dead_end children
|
||||
* are always excluded).
|
||||
*/
|
||||
static int
|
||||
|
@@ -2687,7 +2687,7 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode)
|
||||
vxids = (VirtualTransactionId *)
|
||||
palloc0(sizeof(VirtualTransactionId) * (MaxBackends + 1));
|
||||
|
||||
/* Compute hash code and partiton lock, and look up conflicting modes. */
|
||||
/* Compute hash code and partition lock, and look up conflicting modes. */
|
||||
hashcode = LockTagHashCode(locktag);
|
||||
partitionLock = LockHashPartitionLock(hashcode);
|
||||
conflictMask = lockMethodTable->conflictTab[lockmode];
|
||||
|
@@ -3192,7 +3192,7 @@ ReleasePredicateLocks(bool isCommit)
|
||||
/*
|
||||
* We can't trust XactReadOnly here, because a transaction which started
|
||||
* as READ WRITE can show as READ ONLY later, e.g., within
|
||||
* substransactions. We want to flag a transaction as READ ONLY if it
|
||||
* subtransactions. We want to flag a transaction as READ ONLY if it
|
||||
* commits without writing so that de facto READ ONLY transactions get the
|
||||
* benefit of some RO optimizations, so we will use this local variable to
|
||||
* get some cleanup logic right which is based on whether the transaction
|
||||
|
@@ -179,7 +179,7 @@ LexizeExec(LexizeData *ld, ParsedLex **correspondLexem)
|
||||
if (ld->curDictId == InvalidOid)
|
||||
{
|
||||
/*
|
||||
* usial mode: dictionary wants only one word, but we should keep in
|
||||
* usual mode: dictionary wants only one word, but we should keep in
|
||||
* mind that we should go through all stack
|
||||
*/
|
||||
|
||||
@@ -272,7 +272,7 @@ LexizeExec(LexizeData *ld, ParsedLex **correspondLexem)
|
||||
|
||||
/*
|
||||
* We should be sure that current type of lexeme is recognized
|
||||
* by our dictinonary: we just check is it exist in list of
|
||||
* by our dictionary: we just check is it exist in list of
|
||||
* dictionaries ?
|
||||
*/
|
||||
for (i = 0; i < map->len && !dictExists; i++)
|
||||
@@ -618,7 +618,7 @@ generateHeadline(HeadlineParsedText *prs)
|
||||
/* start of a new fragment */
|
||||
infrag = 1;
|
||||
numfragments++;
|
||||
/* add a fragment delimitor if this is after the first one */
|
||||
/* add a fragment delimiter if this is after the first one */
|
||||
if (numfragments > 1)
|
||||
{
|
||||
memcpy(ptr, prs->fragdelim, prs->fragdelimlen);
|
||||
|
@@ -2419,7 +2419,7 @@ mark_hl_words(HeadlineParsedText *prs, TSQuery query, int highlight,
|
||||
break;
|
||||
}
|
||||
if (curlen < min_words && i >= prs->curwords)
|
||||
{ /* got end of text and our cover is shoter
|
||||
{ /* got end of text and our cover is shorter
|
||||
* than min_words */
|
||||
for (i = p - 1; i >= 0; i--)
|
||||
{
|
||||
|
@@ -2285,7 +2285,7 @@ seq_search(char *name, char **array, int type, int max, int *len)
|
||||
|
||||
for (last = 0, a = array; *a != NULL; a++)
|
||||
{
|
||||
/* comperate first chars */
|
||||
/* compare first chars */
|
||||
if (*name != **a)
|
||||
continue;
|
||||
|
||||
|
@@ -777,7 +777,7 @@ calc_rank_cd(float4 *arrdata, TSVector txt, TSQuery query, int method)
|
||||
|
||||
/*
|
||||
* if doc are big enough then ext.q may be equal to ext.p due to limit
|
||||
* of posional information. In this case we approximate number of
|
||||
* of positional information. In this case we approximate number of
|
||||
* noise word as half cover's length
|
||||
*/
|
||||
nNoise = (ext.q - ext.p) - (ext.end - ext.begin);
|
||||
@@ -786,7 +786,7 @@ calc_rank_cd(float4 *arrdata, TSVector txt, TSQuery query, int method)
|
||||
Wdoc += Cpos / ((double) (1 + nNoise));
|
||||
|
||||
CurExtPos = ((double) (ext.q + ext.p)) / 2.0;
|
||||
if (NExtent > 0 && CurExtPos > PrevExtPos /* prevent devision by
|
||||
if (NExtent > 0 && CurExtPos > PrevExtPos /* prevent division by
|
||||
* zero in a case of
|
||||
multiple lexize */ )
|
||||
SumDist += 1.0 / (CurExtPos - PrevExtPos);
|
||||
|
@@ -342,7 +342,7 @@ window_lag(PG_FUNCTION_ARGS)
|
||||
|
||||
/*
|
||||
* lag_with_offset
|
||||
* returns the value of VE evelulated on a row that is OFFSET
|
||||
* returns the value of VE evaluated on a row that is OFFSET
|
||||
* rows before the current row within a partition,
|
||||
* per spec.
|
||||
*/
|
||||
|
@@ -18,7 +18,7 @@ OBJS = guc.o help_config.o pg_rusage.o ps_status.o superuser.o tzparser.o \
|
||||
rbtree.o
|
||||
|
||||
# This location might depend on the installation directories. Therefore
|
||||
# we can't subsitute it into pg_config.h.
|
||||
# we can't substitute it into pg_config.h.
|
||||
ifdef krb_srvtab
|
||||
override CPPFLAGS += -DPG_KRB_SRVTAB='"$(krb_srvtab)"'
|
||||
endif
|
||||
|
Reference in New Issue
Block a user