mirror of
https://github.com/postgres/postgres.git
synced 2025-07-02 09:02:37 +03:00
Fix typos in comments, code and documentation
While on it, newlines are removed from the end of two elog() strings.
The others are simple grammar mistakes. One comment in pg_upgrade
referred incorrectly to sequences since a7e5457
.
Author: Justin Pryzby
Discussion: https://postgr.es/m/20221230231257.GI1153@telsasoft.com
Backpatch-through: 11
This commit is contained in:
@ -11,7 +11,7 @@ env:
|
||||
CONTAINER_REPO: us-docker.pkg.dev/${GCP_PROJECT}/ci
|
||||
|
||||
# The lower depth accelerates git clone. Use a bit of depth so that
|
||||
# concurrent tasks and retrying older jobs has a chance of working.
|
||||
# concurrent tasks and retrying older jobs have a chance of working.
|
||||
CIRRUS_CLONE_DEPTH: 500
|
||||
# Useful to be able to analyse what in a script takes long
|
||||
CIRRUS_LOG_TIMESTAMP: true
|
||||
@ -66,7 +66,7 @@ task:
|
||||
|
||||
ccache_cache:
|
||||
folder: $CCACHE_DIR
|
||||
# Workaround around performance issues due to 32KB block size
|
||||
# Work around performance issues due to 32KB block size
|
||||
repartition_script: src/tools/ci/gcp_freebsd_repartition.sh
|
||||
create_user_script: |
|
||||
pw useradd postgres
|
||||
|
@ -128,7 +128,7 @@ EXPLAIN SELECT * FROM pgbench_accounts WHERE filler LIKE '%x%';
|
||||
|
||||
<para>
|
||||
In addition, the system must not be running in single-user mode. Since
|
||||
the entire database system is running in single process in this situation,
|
||||
the entire database system is running as a single process in this situation,
|
||||
no background workers will be available.
|
||||
</para>
|
||||
|
||||
|
@ -1001,7 +1001,7 @@ WITH ( MODULUS <replaceable class="parameter">numeric_literal</replaceable>, REM
|
||||
constraint. This does not work, however, if any of the partition keys
|
||||
is an expression and the partition does not accept
|
||||
<literal>NULL</literal> values. If attaching a list partition that will
|
||||
not accept <literal>NULL</literal> values, also add
|
||||
not accept <literal>NULL</literal> values, also add a
|
||||
<literal>NOT NULL</literal> constraint to the partition key column,
|
||||
unless it's an expression.
|
||||
</para>
|
||||
@ -1373,7 +1373,7 @@ WITH ( MODULUS <replaceable class="parameter">numeric_literal</replaceable>, REM
|
||||
type or an unconstrained domain over the new type, a table rewrite is not
|
||||
needed. However, indexes must always be rebuilt unless the system can
|
||||
verify that the new index would be logically equivalent to the existing
|
||||
one. For example, if the collation for a column has been changed an index
|
||||
one. For example, if the collation for a column has been changed, an index
|
||||
rebuild is always required because the new sort order might be different.
|
||||
However, in the absence of a collation change, a column can be changed
|
||||
from <type>text</type> to <type>varchar</type> (or vice versa) without
|
||||
|
@ -350,7 +350,7 @@ CREATE SUBSCRIPTION <replaceable class="parameter">subscription_name</replaceabl
|
||||
<literal>FOR TABLES IN SCHEMA</literal>, rows are always published
|
||||
regardless of the definition of the other expressions.
|
||||
If the subscriber is a <productname>PostgreSQL</productname> version before
|
||||
15 then any row filtering is ignored during the initial data synchronization
|
||||
15, then any row filtering is ignored during the initial data synchronization
|
||||
phase. For this case, the user might want to consider deleting any initially
|
||||
copied data that would be incompatible with subsequent filtering.
|
||||
Because initial data synchronization does not take into account the publication
|
||||
|
@ -935,8 +935,8 @@ BETTER: unrecognized node type: 42
|
||||
<simplesect>
|
||||
<title>Function-Like Macros and Inline Functions</title>
|
||||
<para>
|
||||
Both, macros with arguments and <literal>static inline</literal>
|
||||
functions, may be used. The latter are preferable if there are
|
||||
Both macros with arguments and <literal>static inline</literal>
|
||||
functions may be used. The latter are preferable if there are
|
||||
multiple-evaluation hazards when written as a macro, as e.g., the
|
||||
case with
|
||||
<programlisting>
|
||||
|
@ -78,7 +78,7 @@ mask_unused_space(Page page)
|
||||
if (pd_lower > pd_upper || pd_special < pd_upper ||
|
||||
pd_lower < SizeOfPageHeaderData || pd_special > BLCKSZ)
|
||||
{
|
||||
elog(ERROR, "invalid page pd_lower %u pd_upper %u pd_special %u\n",
|
||||
elog(ERROR, "invalid page pd_lower %u pd_upper %u pd_special %u",
|
||||
pd_lower, pd_upper, pd_special);
|
||||
}
|
||||
|
||||
|
@ -1301,7 +1301,7 @@ spgproperty(Oid index_oid, int attno,
|
||||
/*
|
||||
* Currently, SP-GiST distance-ordered scans require that there be a
|
||||
* distance operator in the opclass with the default types. So we assume
|
||||
* that if such a operator exists, then there's a reason for it.
|
||||
* that if such an operator exists, then there's a reason for it.
|
||||
*/
|
||||
|
||||
/* First we need to know the column's opclass. */
|
||||
|
@ -816,7 +816,7 @@ llvm_session_initialize(void)
|
||||
|
||||
if (LLVMGetTargetFromTriple(llvm_triple, &llvm_targetref, &error) != 0)
|
||||
{
|
||||
elog(FATAL, "failed to query triple %s\n", error);
|
||||
elog(FATAL, "failed to query triple %s", error);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -865,7 +865,7 @@ apply_pathtarget_labeling_to_tlist(List *tlist, PathTarget *target)
|
||||
*
|
||||
* The outputs of this function are two parallel lists, one a list of
|
||||
* PathTargets and the other an integer list of bool flags indicating
|
||||
* whether the corresponding PathTarget contains any evaluatable SRFs.
|
||||
* whether the corresponding PathTarget contains any evaluable SRFs.
|
||||
* The lists are given in the order they'd need to be evaluated in, with
|
||||
* the "lowest" PathTarget first. So the last list entry is always the
|
||||
* originally given PathTarget, and any entries before it indicate evaluation
|
||||
|
@ -11416,7 +11416,7 @@ get_opclass_name(Oid opclass, Oid actual_datatype,
|
||||
|
||||
/*
|
||||
* generate_opclass_name
|
||||
* Compute the name to display for a opclass specified by OID
|
||||
* Compute the name to display for an opclass specified by OID
|
||||
*
|
||||
* The result includes all necessary quoting and schema-prefixing.
|
||||
*/
|
||||
|
@ -408,11 +408,10 @@ get_rel_infos(ClusterInfo *cluster, DbInfo *dbinfo)
|
||||
query[0] = '\0'; /* initialize query string to empty */
|
||||
|
||||
/*
|
||||
* Create a CTE that collects OIDs of regular user tables, including
|
||||
* matviews and sequences, but excluding toast tables and indexes. We
|
||||
* assume that relations with OIDs >= FirstNormalObjectId belong to the
|
||||
* user. (That's probably redundant with the namespace-name exclusions,
|
||||
* but let's be safe.)
|
||||
* Create a CTE that collects OIDs of regular user tables and matviews,
|
||||
* but excluding toast tables and indexes. We assume that relations with
|
||||
* OIDs >= FirstNormalObjectId belong to the user. (That's probably
|
||||
* redundant with the namespace-name exclusions, but let's be safe.)
|
||||
*
|
||||
* pg_largeobject contains user data that does not appear in pg_dump
|
||||
* output, so we have to copy that system table. It's easiest to do that
|
||||
|
@ -182,7 +182,7 @@ create table header_copytest (
|
||||
b int,
|
||||
c text
|
||||
);
|
||||
-- Make sure it works with with dropped columns
|
||||
-- Make sure it works with dropped columns
|
||||
alter table header_copytest drop column c;
|
||||
alter table header_copytest add column c text;
|
||||
copy header_copytest to stdout with (header match);
|
||||
|
@ -57,7 +57,7 @@ SELECT now()::timestamp::text = localtimestamp::text;
|
||||
t
|
||||
(1 row)
|
||||
|
||||
-- current_role/user/user is tested in rolnames.sql
|
||||
-- current_role/user/user is tested in rolenames.sql
|
||||
-- current database / catalog
|
||||
SELECT current_catalog = current_database();
|
||||
?column?
|
||||
|
@ -204,7 +204,7 @@ create table header_copytest (
|
||||
b int,
|
||||
c text
|
||||
);
|
||||
-- Make sure it works with with dropped columns
|
||||
-- Make sure it works with dropped columns
|
||||
alter table header_copytest drop column c;
|
||||
alter table header_copytest add column c text;
|
||||
copy header_copytest to stdout with (header match);
|
||||
|
@ -24,7 +24,7 @@ SELECT length(current_timestamp::text) >= length(current_timestamp(0)::text);
|
||||
-- localtimestamp
|
||||
SELECT now()::timestamp::text = localtimestamp::text;
|
||||
|
||||
-- current_role/user/user is tested in rolnames.sql
|
||||
-- current_role/user/user is tested in rolenames.sql
|
||||
|
||||
-- current database / catalog
|
||||
SELECT current_catalog = current_database();
|
||||
|
@ -257,7 +257,7 @@ The certificate file to use. Implementation is SSL backend specific.
|
||||
|
||||
=item keyfile => B<value>
|
||||
|
||||
The private key to to use. Implementation is SSL backend specific.
|
||||
The private key to use. Implementation is SSL backend specific.
|
||||
|
||||
=item crlfile => B<value>
|
||||
|
||||
|
Reference in New Issue
Block a user