diff --git a/contrib/pg_upgrade/TESTING b/contrib/pg_upgrade/TESTING
index d38b205f49c..2043b40ae8c 100644
--- a/contrib/pg_upgrade/TESTING
+++ b/contrib/pg_upgrade/TESTING
@@ -13,8 +13,8 @@ Here are the steps needed to create a regression database dump file:
This database can be created by running 'gmake installcheck' from
src/test/regression.
-2) Use pg_dump to dump out the regression database. Use the new
- cluster's pg_dump on the old database to minimize whitespace
+2) Use pg_dump to dump out the regression database. Use the new
+ cluster's pg_dump on the old database to minimize whitespace
differences in the diff.
3) Adjust the regression database dump file
diff --git a/doc/src/sgml/client-auth.sgml b/doc/src/sgml/client-auth.sgml
index 31ce45d4ca2..cfdb33a1fd6 100644
--- a/doc/src/sgml/client-auth.sgml
+++ b/doc/src/sgml/client-auth.sgml
@@ -188,7 +188,7 @@ hostnossl database user
obsolete but still accepted spelling of samerole>.)
Superusers are not considered to be members of a role for the
purposes of samerole> unless they are explicitly
- members of the role, directly or indirectly, and not just by
+ members of the role, directly or indirectly, and not just by
virtue of being a superuser.
The value replication> specifies that the record
matches if a replication connection is requested (note that
diff --git a/doc/src/sgml/func.sgml b/doc/src/sgml/func.sgml
index 59f5299f7f1..29538888cb5 100644
--- a/doc/src/sgml/func.sgml
+++ b/doc/src/sgml/func.sgml
@@ -9648,8 +9648,8 @@ table2-mapping
array_to_json(anyarray [, pretty_bool])
- Returns the array as JSON. A Postgres multi-dimensional array
- becomes a JSON array of arrays. Line feeds will be added between
+ Returns the array as JSON. A Postgres multi-dimensional array
+ becomes a JSON array of arrays. Line feeds will be added between
dimension 1 elements if pretty_bool is true.
array_to_json('{{1,5},{99,100}}'::int[])
@@ -9663,7 +9663,7 @@ table2-mapping
row_to_json(record [, pretty_bool])
- Returns the row as JSON. Line feeds will be added between level
+ Returns the row as JSON. Line feeds will be added between level
1 elements if pretty_bool is true.
row_to_json(row(1,'foo'))
@@ -13812,7 +13812,7 @@ SELECT pg_type_is_visible('myschema.widget'::regtype);
pg_get_viewdef(view_name , pretty_bool>)
text
- get underlying SELECT command for view,
+ get underlying SELECT command for view,
lines with fields are wrapped to 80 columns if pretty_bool is true (deprecated )
@@ -13823,13 +13823,13 @@ SELECT pg_type_is_visible('myschema.widget'::regtype);
pg_get_viewdef(view_oid , pretty_bool>)
text
- get underlying SELECT command for view,
+ get underlying SELECT command for view,
lines with fields are wrapped to 80 columns if pretty_bool is true
pg_get_viewdef(view_oid , wrap_int>)
text
- get underlying SELECT command for view,
+ get underlying SELECT command for view,
wrapping lines with fields as specified, pretty printing is implied
diff --git a/doc/src/sgml/installation.sgml b/doc/src/sgml/installation.sgml
index 609e004a336..d6fc3f0fbbd 100644
--- a/doc/src/sgml/installation.sgml
+++ b/doc/src/sgml/installation.sgml
@@ -1622,7 +1622,7 @@ PostgreSQL, contrib and HTML documentation successfully made. Ready to install.
On some systems with shared libraries
you need to tell the system how to find the newly installed
shared libraries. The systems on which this is
- not necessary include
+ not necessary include
FreeBSD>,
HP-UX>, IRIX>, Linux>,
diff --git a/doc/src/sgml/pgupgrade.sgml b/doc/src/sgml/pgupgrade.sgml
index ef99212b0fd..99d248b43a5 100644
--- a/doc/src/sgml/pgupgrade.sgml
+++ b/doc/src/sgml/pgupgrade.sgml
@@ -270,7 +270,7 @@ gmake prefix=/usr/local/pgsql.new install
Install any custom shared object files (or DLLs) used by the old cluster
- into the new cluster, e.g. pgcrypto.so ,
+ into the new cluster, e.g. pgcrypto.so ,
whether they are from contrib
or some other source. Do not install the schema definitions, e.g.
pgcrypto.sql>, because these will be upgraded from the old cluster.
diff --git a/doc/src/sgml/ref/pg_dump.sgml b/doc/src/sgml/ref/pg_dump.sgml
index a5e53043267..4f7f5109fb9 100644
--- a/doc/src/sgml/ref/pg_dump.sgml
+++ b/doc/src/sgml/ref/pg_dump.sgml
@@ -403,11 +403,11 @@ PostgreSQL documentation
Dump only the object definitions (schema), not data.
- To exclude table data for only a subset of tables in the database,
+ To exclude table data for only a subset of tables in the database,
see --exclude-table-data>.
- This option is equivalent to specifying
+ This option is equivalent to specifying
--section=pre-data --section=post-data>.
@@ -417,13 +417,13 @@ PostgreSQL documentation
--section=sectionname
- Only dump the named section. The name can be one of pre-data>, data>
- and post-data>.
+ Only dump the named section. The name can be one of pre-data>, data>
+ and post-data>.
This option can be specified more than once. The default is to dump all sections.
- Post-data items consist of definitions of indexes, triggers, rules
- and constraints other than validated check constraints.
+ Post-data items consist of definitions of indexes, triggers, rules
+ and constraints other than validated check constraints.
Pre-data items consist of all other data definition items.
@@ -640,7 +640,7 @@ PostgreSQL documentation
Do not dump data for any tables matching the table pattern. The pattern is
interpreted according to the same rules as for -t>.
- --exclude-table-data> can be given more than once to
+ --exclude-table-data> can be given more than once to
exclude tables matching any of several patterns. This option is
useful when you need the definition of a particular table even
though you do not need the data in it.
diff --git a/doc/src/sgml/ref/pg_restore.sgml b/doc/src/sgml/ref/pg_restore.sgml
index 47e972223e6..513b946e571 100644
--- a/doc/src/sgml/ref/pg_restore.sgml
+++ b/doc/src/sgml/ref/pg_restore.sgml
@@ -363,7 +363,7 @@
uses the word schema> in a different meaning.)
- This option is equivalent to specifying
+ This option is equivalent to specifying
--section=pre-data --section=post-data>.
@@ -515,13 +515,13 @@
--section=sectionname
- Only restore the named section. The name can be one of pre-data>, data>
- and post-data>.
+ Only restore the named section. The name can be one of pre-data>, data>
+ and post-data>.
This option can be specified more than once. The default is to restore all sections.
- Post-data items consist of definitions of indexes, triggers, rules
- and constraints other than validated check constraints.
+ Post-data items consist of definitions of indexes, triggers, rules
+ and constraints other than validated check constraints.
Pre-data items consist of all other data definition items.
diff --git a/doc/src/sgml/release.sgml b/doc/src/sgml/release.sgml
index 4103c5d48bd..2515329b99e 100644
--- a/doc/src/sgml/release.sgml
+++ b/doc/src/sgml/release.sgml
@@ -29,7 +29,7 @@ non-ASCII characters find using grep -P '[\x80-\xFF]'
does not support it
http://www.pemberley.com/janeinfo/latin1.html#latexta
- do not use numeric _UTF_ numeric character escapes (nnn;),
+ do not use numeric _UTF_ numeric character escapes (nnn;),
we can only use Latin1
Example: Alvaro Herrera is Álvaro Herrera
diff --git a/src/backend/utils/misc/guc-file.l b/src/backend/utils/misc/guc-file.l
index 1a5e383b150..ca7619034f4 100644
--- a/src/backend/utils/misc/guc-file.l
+++ b/src/backend/utils/misc/guc-file.l
@@ -431,8 +431,8 @@ ParseConfigFile(const char *config_file, const char *calling_file, bool strict,
return false;
}
- ereport(LOG,
- (errmsg("skipping missing configuration file \"%s\"",
+ ereport(LOG,
+ (errmsg("skipping missing configuration file \"%s\"",
config_file)));
return OK;
}
diff --git a/src/interfaces/libpq/Makefile b/src/interfaces/libpq/Makefile
index ec4fdd403e4..b71b0c76ef3 100644
--- a/src/interfaces/libpq/Makefile
+++ b/src/interfaces/libpq/Makefile
@@ -85,7 +85,7 @@ backend_src = $(top_srcdir)/src/backend
# compile with appropriate options to build a shared lib, we can't
# necessarily use the same object files as the backend uses. Instead,
# symlink the source files in here and build our own object file.
-# For some libpgport modules, this only happens if configure decides
+# For some libpgport modules, this only happens if configure decides
# the module is needed (see filter hack in OBJS, above).
chklocale.c crypt.c getaddrinfo.c getpeereid.c inet_aton.c inet_net_ntop.c noblock.c open.c pgsleep.c pgstrcasecmp.c snprintf.c strerror.c strlcpy.c thread.c win32error.c win32setlocale.c: % : $(top_srcdir)/src/port/%
@@ -141,7 +141,7 @@ clean distclean: clean-lib
rm -f pg_config_paths.h
rm -f inet_net_ntop.c noblock.c pgstrcasecmp.c thread.c
rm -f chklocale.c crypt.c getaddrinfo.c getpeereid.c inet_aton.c open.c snprintf.c strerror.c strlcpy.c win32error.c win32setlocale.c
- rm -f pgsleep.c
+ rm -f pgsleep.c
rm -f md5.c ip.c
rm -f encnames.c wchar.c
diff --git a/src/pl/plperl/expected/plperl_array.out b/src/pl/plperl/expected/plperl_array.out
index 829733fef4f..efa884b359f 100644
--- a/src/pl/plperl/expected/plperl_array.out
+++ b/src/pl/plperl/expected/plperl_array.out
@@ -131,7 +131,7 @@ CREATE OR REPLACE FUNCTION plperl_sum_row_elements(rowfoo) RETURNS TEXT AS $$
}
else {
$result = $row_ref->{bar};
- die "not an array reference".ref ($row_ref->{baz})
+ die "not an array reference".ref ($row_ref->{baz})
unless (is_array_ref($row_ref->{baz}));
# process a single-dimensional array
foreach my $elem (@{$row_ref->{baz}}) {
@@ -159,7 +159,7 @@ CREATE OR REPLACE FUNCTION plperl_sum_array_of_rows(rowbar) RETURNS TEXT AS $$
foreach my $row_ref (@{$row_array_ref}) {
if (ref $row_ref eq 'HASH') {
$result += $row_ref->{bar};
- die "not an array reference".ref ($row_ref->{baz})
+ die "not an array reference".ref ($row_ref->{baz})
unless (is_array_ref($row_ref->{baz}));
foreach my $elem (@{$row_ref->{baz}}) {
$result += $elem unless ref $elem;
@@ -177,7 +177,7 @@ CREATE OR REPLACE FUNCTION plperl_sum_array_of_rows(rowbar) RETURNS TEXT AS $$
}
return $result;
$$ LANGUAGE plperl;
-select plperl_sum_array_of_rows(ROW(ARRAY[ROW(1, ARRAY[2,3,4,5,6,7,8,9,10])::rowfoo,
+select plperl_sum_array_of_rows(ROW(ARRAY[ROW(1, ARRAY[2,3,4,5,6,7,8,9,10])::rowfoo,
ROW(11, ARRAY[12,13,14,15,16,17,18,19,20])::rowfoo])::rowbar);
plperl_sum_array_of_rows
--------------------------
diff --git a/src/pl/plperl/sql/plperl_array.sql b/src/pl/plperl/sql/plperl_array.sql
index 818a48ec200..00314726459 100644
--- a/src/pl/plperl/sql/plperl_array.sql
+++ b/src/pl/plperl/sql/plperl_array.sql
@@ -90,7 +90,7 @@ CREATE OR REPLACE FUNCTION plperl_sum_row_elements(rowfoo) RETURNS TEXT AS $$
}
else {
$result = $row_ref->{bar};
- die "not an array reference".ref ($row_ref->{baz})
+ die "not an array reference".ref ($row_ref->{baz})
unless (is_array_ref($row_ref->{baz}));
# process a single-dimensional array
foreach my $elem (@{$row_ref->{baz}}) {
@@ -116,7 +116,7 @@ CREATE OR REPLACE FUNCTION plperl_sum_array_of_rows(rowbar) RETURNS TEXT AS $$
foreach my $row_ref (@{$row_array_ref}) {
if (ref $row_ref eq 'HASH') {
$result += $row_ref->{bar};
- die "not an array reference".ref ($row_ref->{baz})
+ die "not an array reference".ref ($row_ref->{baz})
unless (is_array_ref($row_ref->{baz}));
foreach my $elem (@{$row_ref->{baz}}) {
$result += $elem unless ref $elem;
@@ -135,7 +135,7 @@ CREATE OR REPLACE FUNCTION plperl_sum_array_of_rows(rowbar) RETURNS TEXT AS $$
return $result;
$$ LANGUAGE plperl;
-select plperl_sum_array_of_rows(ROW(ARRAY[ROW(1, ARRAY[2,3,4,5,6,7,8,9,10])::rowfoo,
+select plperl_sum_array_of_rows(ROW(ARRAY[ROW(1, ARRAY[2,3,4,5,6,7,8,9,10])::rowfoo,
ROW(11, ARRAY[12,13,14,15,16,17,18,19,20])::rowfoo])::rowbar);
-- check arrays as out parameters
diff --git a/src/test/regress/expected/alter_table.out b/src/test/regress/expected/alter_table.out
index 890a51fd9e8..456a729517f 100644
--- a/src/test/regress/expected/alter_table.out
+++ b/src/test/regress/expected/alter_table.out
@@ -384,7 +384,7 @@ DROP TABLE tmp2;
-- NOT VALID with plan invalidation -- ensure we don't use a constraint for
-- exclusion until validated
set constraint_exclusion TO 'partition';
-create table nv_parent (d date);
+create table nv_parent (d date);
create table nv_child_2010 () inherits (nv_parent);
create table nv_child_2011 () inherits (nv_parent);
alter table nv_child_2010 add check (d between '2010-01-01'::date and '2010-12-31'::date) not valid;
@@ -1968,7 +1968,7 @@ create view alter1.v1 as select * from alter1.t1;
create function alter1.plus1(int) returns int as 'select $1+1' language sql;
create domain alter1.posint integer check (value > 0);
create type alter1.ctype as (f1 int, f2 text);
-create function alter1.same(alter1.ctype, alter1.ctype) returns boolean language sql
+create function alter1.same(alter1.ctype, alter1.ctype) returns boolean language sql
as 'select $1.f1 is not distinct from $2.f1 and $1.f2 is not distinct from $2.f2';
create operator alter1.=(procedure = alter1.same, leftarg = alter1.ctype, rightarg = alter1.ctype);
create operator class alter1.ctype_hash_ops default for type alter1.ctype using hash as
diff --git a/src/test/regress/expected/create_function_3.out b/src/test/regress/expected/create_function_3.out
index eac3de57ccb..e7952322dc2 100644
--- a/src/test/regress/expected/create_function_3.out
+++ b/src/test/regress/expected/create_function_3.out
@@ -8,7 +8,7 @@ CREATE SCHEMA temp_func_test;
GRANT ALL ON SCHEMA temp_func_test TO public;
SET search_path TO temp_func_test, public;
--
--- ARGUMENT and RETURN TYPES
+-- ARGUMENT and RETURN TYPES
--
CREATE FUNCTION functest_A_1(text, date) RETURNS bool LANGUAGE 'sql'
AS 'SELECT $1 = ''abcd'' AND $2 > ''2001-01-01''';
diff --git a/src/test/regress/expected/foreign_data.out b/src/test/regress/expected/foreign_data.out
index b3165d59014..af72c91d74f 100644
--- a/src/test/regress/expected/foreign_data.out
+++ b/src/test/regress/expected/foreign_data.out
@@ -1183,7 +1183,7 @@ NOTICE: drop cascades to user mapping for public
DROP SERVER t2;
DROP USER MAPPING FOR regress_test_role SERVER s6;
-- This test causes some order dependent cascade detail output,
--- so switch to terse mode for it.
+-- so switch to terse mode for it.
\set VERBOSITY terse
DROP FOREIGN DATA WRAPPER foo CASCADE;
NOTICE: drop cascades to 5 other objects
diff --git a/src/test/regress/expected/json.out b/src/test/regress/expected/json.out
index fa8415cdb77..a26c64a599d 100644
--- a/src/test/regress/expected/json.out
+++ b/src/test/regress/expected/json.out
@@ -279,10 +279,10 @@ SELECT array_to_json(array_agg(q),true) from (select x as b, x * 2 as c from gen
(1 row)
SELECT array_to_json(array_agg(q),false)
- FROM ( SELECT $$a$$ || x AS b, y AS c,
+ FROM ( SELECT $$a$$ || x AS b, y AS c,
ARRAY[ROW(x.*,ARRAY[1,2,3]),
- ROW(y.*,ARRAY[4,5,6])] AS z
- FROM generate_series(1,2) x,
+ ROW(y.*,ARRAY[4,5,6])] AS z
+ FROM generate_series(1,2) x,
generate_series(4,5) y) q;
array_to_json
-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
@@ -308,12 +308,12 @@ SELECT row_to_json(row(1,'foo'));
{"f1":1,"f2":"foo"}
(1 row)
-SELECT row_to_json(q)
-FROM (SELECT $$a$$ || x AS b,
- y AS c,
+SELECT row_to_json(q)
+FROM (SELECT $$a$$ || x AS b,
+ y AS c,
ARRAY[ROW(x.*,ARRAY[1,2,3]),
- ROW(y.*,ARRAY[4,5,6])] AS z
- FROM generate_series(1,2) x,
+ ROW(y.*,ARRAY[4,5,6])] AS z
+ FROM generate_series(1,2) x,
generate_series(4,5) y) q;
row_to_json
--------------------------------------------------------------------
@@ -323,12 +323,12 @@ FROM (SELECT $$a$$ || x AS b,
{"b":"a2","c":5,"z":[{"f1":2,"f2":[1,2,3]},{"f1":5,"f2":[4,5,6]}]}
(4 rows)
-SELECT row_to_json(q,true)
-FROM (SELECT $$a$$ || x AS b,
- y AS c,
+SELECT row_to_json(q,true)
+FROM (SELECT $$a$$ || x AS b,
+ y AS c,
ARRAY[ROW(x.*,ARRAY[1,2,3]),
- ROW(y.*,ARRAY[4,5,6])] AS z
- FROM generate_series(1,2) x,
+ ROW(y.*,ARRAY[4,5,6])] AS z
+ FROM generate_series(1,2) x,
generate_series(4,5) y) q;
row_to_json
-----------------------------------------------------
@@ -349,7 +349,7 @@ FROM (SELECT $$a$$ || x AS b,
CREATE TEMP TABLE rows AS
SELECT x, 'txt' || x as y
FROM generate_series(1,3) AS x;
-SELECT row_to_json(q,true)
+SELECT row_to_json(q,true)
FROM rows q;
row_to_json
--------------
diff --git a/src/test/regress/sql/alter_table.sql b/src/test/regress/sql/alter_table.sql
index 5c03123b4e0..5ca537e91bc 100644
--- a/src/test/regress/sql/alter_table.sql
+++ b/src/test/regress/sql/alter_table.sql
@@ -327,7 +327,7 @@ DROP TABLE tmp2;
-- NOT VALID with plan invalidation -- ensure we don't use a constraint for
-- exclusion until validated
set constraint_exclusion TO 'partition';
-create table nv_parent (d date);
+create table nv_parent (d date);
create table nv_child_2010 () inherits (nv_parent);
create table nv_child_2011 () inherits (nv_parent);
alter table nv_child_2010 add check (d between '2010-01-01'::date and '2010-12-31'::date) not valid;
@@ -1348,7 +1348,7 @@ create domain alter1.posint integer check (value > 0);
create type alter1.ctype as (f1 int, f2 text);
-create function alter1.same(alter1.ctype, alter1.ctype) returns boolean language sql
+create function alter1.same(alter1.ctype, alter1.ctype) returns boolean language sql
as 'select $1.f1 is not distinct from $2.f1 and $1.f2 is not distinct from $2.f2';
create operator alter1.=(procedure = alter1.same, leftarg = alter1.ctype, rightarg = alter1.ctype);
diff --git a/src/test/regress/sql/create_function_3.sql b/src/test/regress/sql/create_function_3.sql
index 4693db63cdc..e2dd9a35a02 100644
--- a/src/test/regress/sql/create_function_3.sql
+++ b/src/test/regress/sql/create_function_3.sql
@@ -11,7 +11,7 @@ GRANT ALL ON SCHEMA temp_func_test TO public;
SET search_path TO temp_func_test, public;
--
--- ARGUMENT and RETURN TYPES
+-- ARGUMENT and RETURN TYPES
--
CREATE FUNCTION functest_A_1(text, date) RETURNS bool LANGUAGE 'sql'
AS 'SELECT $1 = ''abcd'' AND $2 > ''2001-01-01''';
diff --git a/src/test/regress/sql/foreign_data.sql b/src/test/regress/sql/foreign_data.sql
index 4137969e7de..38057db7b44 100644
--- a/src/test/regress/sql/foreign_data.sql
+++ b/src/test/regress/sql/foreign_data.sql
@@ -484,7 +484,7 @@ DROP SERVER t1 CASCADE;
DROP SERVER t2;
DROP USER MAPPING FOR regress_test_role SERVER s6;
-- This test causes some order dependent cascade detail output,
--- so switch to terse mode for it.
+-- so switch to terse mode for it.
\set VERBOSITY terse
DROP FOREIGN DATA WRAPPER foo CASCADE;
\set VERBOSITY default
diff --git a/src/test/regress/sql/json.sql b/src/test/regress/sql/json.sql
index ab1c41c1c4c..27454ad3623 100644
--- a/src/test/regress/sql/json.sql
+++ b/src/test/regress/sql/json.sql
@@ -62,10 +62,10 @@ SELECT array_to_json(array(select 1 as a));
SELECT array_to_json(array_agg(q),false) from (select x as b, x * 2 as c from generate_series(1,3) x) q;
SELECT array_to_json(array_agg(q),true) from (select x as b, x * 2 as c from generate_series(1,3) x) q;
SELECT array_to_json(array_agg(q),false)
- FROM ( SELECT $$a$$ || x AS b, y AS c,
+ FROM ( SELECT $$a$$ || x AS b, y AS c,
ARRAY[ROW(x.*,ARRAY[1,2,3]),
- ROW(y.*,ARRAY[4,5,6])] AS z
- FROM generate_series(1,2) x,
+ ROW(y.*,ARRAY[4,5,6])] AS z
+ FROM generate_series(1,2) x,
generate_series(4,5) y) q;
SELECT array_to_json(array_agg(x),false) from generate_series(5,10) x;
SELECT array_to_json('{{1,5},{99,100}}'::int[]);
@@ -73,27 +73,27 @@ SELECT array_to_json('{{1,5},{99,100}}'::int[]);
-- row_to_json
SELECT row_to_json(row(1,'foo'));
-SELECT row_to_json(q)
-FROM (SELECT $$a$$ || x AS b,
- y AS c,
+SELECT row_to_json(q)
+FROM (SELECT $$a$$ || x AS b,
+ y AS c,
ARRAY[ROW(x.*,ARRAY[1,2,3]),
- ROW(y.*,ARRAY[4,5,6])] AS z
- FROM generate_series(1,2) x,
+ ROW(y.*,ARRAY[4,5,6])] AS z
+ FROM generate_series(1,2) x,
generate_series(4,5) y) q;
-SELECT row_to_json(q,true)
-FROM (SELECT $$a$$ || x AS b,
- y AS c,
+SELECT row_to_json(q,true)
+FROM (SELECT $$a$$ || x AS b,
+ y AS c,
ARRAY[ROW(x.*,ARRAY[1,2,3]),
- ROW(y.*,ARRAY[4,5,6])] AS z
- FROM generate_series(1,2) x,
+ ROW(y.*,ARRAY[4,5,6])] AS z
+ FROM generate_series(1,2) x,
generate_series(4,5) y) q;
CREATE TEMP TABLE rows AS
SELECT x, 'txt' || x as y
FROM generate_series(1,3) AS x;
-SELECT row_to_json(q,true)
+SELECT row_to_json(q,true)
FROM rows q;
SELECT row_to_json(row((select array_agg(x) as d from generate_series(5,10) x)),false);
@@ -111,4 +111,3 @@ FROM (SELECT '-Infinity'::float8 AS "float8field") q;
-- json input
SELECT row_to_json(q)
FROM (SELECT '{"a":1,"b": [2,3,4,"d","e","f"],"c":{"p":1,"q":2}}'::json AS "jsonfield") q;
-
diff --git a/src/tools/copyright.pl b/src/tools/copyright.pl
index db5a2f023c3..d78a8d8ce88 100755
--- a/src/tools/copyright.pl
+++ b/src/tools/copyright.pl
@@ -1,4 +1,4 @@
-#!/usr/bin/perl
+#!/usr/bin/perl
#################################################################
# copyright.pl -- update copyright notices throughout the source tree, idempotently.
#
diff --git a/src/tools/msvc/README b/src/tools/msvc/README
index bbadb34829c..3b2939ad686 100644
--- a/src/tools/msvc/README
+++ b/src/tools/msvc/README
@@ -76,7 +76,7 @@ Solution.pm module containing the code to generate the Visual
Studio solution files.
VCBuildProject.pm module containing the code to generate VCBuild based
project files (Visual Studio 2005/2008)
-VSObjectFactory.pm factory module providing the code to create the
+VSObjectFactory.pm factory module providing the code to create the
appropriate project/solution files for the current
environment
diff --git a/src/tools/msvc/install.bat b/src/tools/msvc/install.bat
index 1d2ab9c9b0a..c636cbd7f70 100644
--- a/src/tools/msvc/install.bat
+++ b/src/tools/msvc/install.bat
@@ -18,7 +18,7 @@ IF NOT EXIST buildenv.pl goto nobuildenv
perl -e "require 'buildenv.pl'; while(($k,$v) = each %%ENV) { print qq[\@SET $k=$v\n]; }" > bldenv.bat
CALL bldenv.bat
del bldenv.bat
-:nobuildenv
+:nobuildenv
perl install.pl "%1"
diff --git a/src/tools/pginclude/pgrminclude b/src/tools/pginclude/pgrminclude
index 383c5044bd7..7cbd2e7c9ca 100755
--- a/src/tools/pginclude/pgrminclude
+++ b/src/tools/pginclude/pgrminclude
@@ -43,7 +43,7 @@ verbose_output() {
process_includes_in_file() {
# loop through all includes mentioned in the file
- cat "$FILE" |
+ cat "$FILE" |
grep "^#include\>" |
grep -v '/\* *pgrminclude *ignore *\*/' |
sed 's/^#include[ ]*[<"]\([^>"]*\).*$/\1/g' |
@@ -80,7 +80,7 @@ compile_file() {
grep -v "^#else" |
grep -v "^#elif" |
grep -v "^#endif" |
- # with #if blocks gone, now undef #defines to avoid redefine
+ # with #if blocks gone, now undef #defines to avoid redefine
# warning and failure
sed 's/#define[ ][ ]*\([A-Za-z0-9_]*\).*$/#undef \1\n&/' >/tmp/$$a
@@ -121,7 +121,7 @@ compile_file() {
fi
}
-# Process include files first because they can affect the compilation
+# Process include files first because they can affect the compilation
# of *.c files.
(find . \( -name .git -a -prune \) -o -type f -name '*.h' -print | sort;
find . \( -name .git -a -prune \) -o -type f -name '*.c' -print | sort) |