Commit b5bce6c1 authored by Tom Lane's avatar Tom Lane

Final pgindent + perltidy run for 9.6.

parent 05d8dec6
......@@ -6698,6 +6698,7 @@ heap_prepare_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid,
if (tuple->t_infomask & HEAP_MOVED)
{
xid = HeapTupleHeaderGetXvac(tuple);
/*
* For Xvac, we ignore the cutoff_xid and just always perform the
* freeze operation. The oldest release in which such a value can
......
......@@ -193,10 +193,11 @@ SyncRepWaitForLSN(XLogRecPtr lsn, bool commit)
ResetLatch(MyLatch);
/*
* Acquiring the lock is not needed, the latch ensures proper barriers.
* If it looks like we're done, we must really be done, because once
* walsender changes the state to SYNC_REP_WAIT_COMPLETE, it will never
* update it again, so we can't be seeing a stale value in that case.
* Acquiring the lock is not needed, the latch ensures proper
* barriers. If it looks like we're done, we must really be done,
* because once walsender changes the state to SYNC_REP_WAIT_COMPLETE,
* it will never update it again, so we can't be seeing a stale value
* in that case.
*/
if (MyProc->syncRepState == SYNC_REP_WAIT_COMPLETE)
break;
......
......@@ -433,7 +433,8 @@ getNextFlagFromString(IspellDict *Conf, char **sflagset, char *sflag)
if (Conf->flagMode == FM_LONG && maxstep > 0)
ereport(ERROR,
(errcode(ERRCODE_CONFIG_FILE_ERROR),
errmsg("invalid affix flag \"%s\" with \"long\" flag value", sbuf)));
errmsg("invalid affix flag \"%s\" with \"long\" flag value",
sbuf)));
*sflag = '\0';
}
......@@ -1263,8 +1264,9 @@ NIImportOOAffixes(IspellDict *Conf, const char *filename)
else if (STRNCMP(s, "default") != 0)
ereport(ERROR,
(errcode(ERRCODE_CONFIG_FILE_ERROR),
errmsg("Ispell dictionary supports only \"default\", "
"\"long\", and \"num\" flag value")));
errmsg("Ispell dictionary supports only "
"\"default\", \"long\", "
"and \"num\" flag values")));
}
}
......
......@@ -1849,9 +1849,9 @@ jsonb_object_agg_transfn(PG_FUNCTION_ARGS)
single_scalar = false;
/*
* values can be anything, including structured and null, so we treat
* them as in json_agg_transfn, except that single scalars are always
* pushed as WJB_VALUE items.
* values can be anything, including structured and null, so we treat them
* as in json_agg_transfn, except that single scalars are always pushed as
* WJB_VALUE items.
*/
while ((type = JsonbIteratorNext(&it, &v, false)) != WJB_DONE)
......
......@@ -430,7 +430,8 @@ pushStop(TSQueryParserState state)
#define STACKDEPTH 32
typedef struct OperatorElement {
typedef struct OperatorElement
{
int8 op;
int16 distance;
} OperatorElement;
......@@ -453,7 +454,7 @@ cleanOpStack(TSQueryParserState state,
{
int opPriority = OP_PRIORITY(op);
while(*lenstack)
while (*lenstack)
{
/* NOT is right associative unlike to others */
if ((op != OP_NOT && opPriority > OP_PRIORITY(stack[*lenstack - 1].op)) ||
......@@ -503,7 +504,7 @@ makepol(TSQueryParserState state,
makepol(state, pushval, opaque);
break;
case PT_CLOSE:
cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */);
cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */ );
return;
case PT_ERR:
default:
......@@ -514,7 +515,7 @@ makepol(TSQueryParserState state,
}
}
cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */);
cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */ );
}
static void
......@@ -916,7 +917,7 @@ tsqueryout(PG_FUNCTION_ARGS)
nrm.cur = nrm.buf = (char *) palloc(sizeof(char) * nrm.buflen);
*(nrm.cur) = '\0';
nrm.op = GETOPERAND(query);
infix(&nrm, -1 /* lowest priority */, false);
infix(&nrm, -1 /* lowest priority */ , false);
PG_FREE_IF_COPY(query, 0);
PG_RETURN_CSTRING(nrm.buf);
......
......@@ -430,8 +430,8 @@ normalize_phrase_tree(NODE *node)
node->right = normalize_phrase_tree(node->right);
/*
* if subtree contains only nodes with higher "priority" then
* we are done. See comment near NODE_PRIORITY()
* if subtree contains only nodes with higher "priority" then we are
* done. See comment near NODE_PRIORITY()
*/
if (NODE_PRIORITY(node) <= NODE_PRIORITY(node->right) &&
NODE_PRIORITY(node) <= NODE_PRIORITY(node->left))
......
......@@ -299,9 +299,9 @@ struct Tuplesortstate
* Function to read a stored tuple from tape back into memory. 'len' is
* the already-read length of the stored tuple. Create a palloc'd copy,
* initialize tuple/datum1/isnull1 in the target SortTuple struct, and
* decrease state->availMem by the amount of memory space consumed.
* (See batchUsed notes for details on how memory is handled when
* incremental accounting is abandoned.)
* decrease state->availMem by the amount of memory space consumed. (See
* batchUsed notes for details on how memory is handled when incremental
* accounting is abandoned.)
*/
void (*readtup) (Tuplesortstate *state, SortTuple *stup,
int tapenum, unsigned int len);
......@@ -309,8 +309,8 @@ struct Tuplesortstate
/*
* Function to move a caller tuple. This is usually implemented as a
* memmove() shim, but function may also perform additional fix-up of
* caller tuple where needed. Batch memory support requires the
* movement of caller tuples from one location in memory to another.
* caller tuple where needed. Batch memory support requires the movement
* of caller tuples from one location in memory to another.
*/
void (*movetup) (void *dest, void *src, unsigned int len);
......
......@@ -6042,15 +6042,16 @@ getOwnedSeqs(Archive *fout, TableInfo tblinfo[], int numTables)
* We need to dump the components that are being dumped for the table
* and any components which the sequence is explicitly marked with.
*
* We can't simply use the set of components which are being dumped for
* the table as the table might be in an extension (and only the
* We can't simply use the set of components which are being dumped
* for the table as the table might be in an extension (and only the
* non-extension components, eg: ACLs if changed, security labels, and
* policies, are being dumped) while the sequence is not (and therefore
* the definition and other components should also be dumped).
* policies, are being dumped) while the sequence is not (and
* therefore the definition and other components should also be
* dumped).
*
* If the sequence is part of the extension then it should be properly
* marked by checkExtensionMembership() and this will be a no-op as the
* table will be equivalently marked.
* marked by checkExtensionMembership() and this will be a no-op as
* the table will be equivalently marked.
*/
seqinfo->dobj.dump = seqinfo->dobj.dump | owning_tab->dobj.dump;
......
......@@ -1098,7 +1098,8 @@ dumpTablespaces(PGconn *conn)
else if (server_version >= 90200)
res = executeQuery(conn, "SELECT oid, spcname, "
"pg_catalog.pg_get_userbyid(spcowner) AS spcowner, "
"pg_catalog.pg_tablespace_location(oid), spcacl, '' as rspcacl, "
"pg_catalog.pg_tablespace_location(oid), "
"spcacl, '' as rspcacl, "
"array_to_string(spcoptions, ', '),"
"pg_catalog.shobj_description(oid, 'pg_tablespace') "
"FROM pg_catalog.pg_tablespace "
......@@ -1329,10 +1330,10 @@ dumpCreateDB(PGconn *conn)
"datistemplate, "
"(SELECT pg_catalog.array_agg(acl) FROM (SELECT pg_catalog.unnest(coalesce(datacl,pg_catalog.acldefault('d',datdba))) AS acl "
"EXCEPT SELECT pg_catalog.unnest(pg_catalog.acldefault('d',datdba))) as foo)"
"AS datacl,"
"AS datacl, "
"(SELECT pg_catalog.array_agg(acl) FROM (SELECT pg_catalog.unnest(pg_catalog.acldefault('d',datdba)) AS acl "
"EXCEPT SELECT pg_catalog.unnest(coalesce(datacl,pg_catalog.acldefault('d',datdba)))) as foo)"
"AS rdatacl,"
"AS rdatacl, "
"datconnlimit, "
"(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
"FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
......@@ -1343,7 +1344,8 @@ dumpCreateDB(PGconn *conn)
"coalesce(rolname, (select rolname from pg_authid where oid=(select datdba from pg_database where datname='template0'))), "
"pg_encoding_to_char(d.encoding), "
"datcollate, datctype, datfrozenxid, datminmxid, "
"datistemplate, datacl, '' as rdatacl, datconnlimit, "
"datistemplate, datacl, '' as rdatacl, "
"datconnlimit, "
"(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
"FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
"WHERE datallowconn ORDER BY 1");
......@@ -1353,7 +1355,8 @@ dumpCreateDB(PGconn *conn)
"coalesce(rolname, (select rolname from pg_authid where oid=(select datdba from pg_database where datname='template0'))), "
"pg_encoding_to_char(d.encoding), "
"datcollate, datctype, datfrozenxid, 0 AS datminmxid, "
"datistemplate, datacl, '' as rdatacl, datconnlimit, "
"datistemplate, datacl, '' as rdatacl, "
"datconnlimit, "
"(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
"FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
"WHERE datallowconn ORDER BY 1");
......@@ -1363,7 +1366,8 @@ dumpCreateDB(PGconn *conn)
"coalesce(rolname, (select rolname from pg_authid where oid=(select datdba from pg_database where datname='template0'))), "
"pg_encoding_to_char(d.encoding), "
"null::text AS datcollate, null::text AS datctype, datfrozenxid, 0 AS datminmxid, "
"datistemplate, datacl, '' as rdatacl, datconnlimit, "
"datistemplate, datacl, '' as rdatacl, "
"datconnlimit, "
"(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
"FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
"WHERE datallowconn ORDER BY 1");
......@@ -1373,7 +1377,8 @@ dumpCreateDB(PGconn *conn)
"coalesce(usename, (select usename from pg_shadow where usesysid=(select datdba from pg_database where datname='template0'))), "
"pg_encoding_to_char(d.encoding), "
"null::text AS datcollate, null::text AS datctype, datfrozenxid, 0 AS datminmxid, "
"datistemplate, datacl, '' as rdatacl, -1 as datconnlimit, "
"datistemplate, datacl, '' as rdatacl, "
"-1 as datconnlimit, "
"(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
"FROM pg_database d LEFT JOIN pg_shadow u ON (datdba = usesysid) "
"WHERE datallowconn ORDER BY 1");
......@@ -1383,7 +1388,8 @@ dumpCreateDB(PGconn *conn)
"coalesce(usename, (select usename from pg_shadow where usesysid=(select datdba from pg_database where datname='template0'))), "
"pg_encoding_to_char(d.encoding), "
"null::text AS datcollate, null::text AS datctype, datfrozenxid, 0 AS datminmxid, "
"datistemplate, datacl, '' as rdatacl, -1 as datconnlimit, "
"datistemplate, datacl, '' as rdatacl, "
"-1 as datconnlimit, "
"'pg_default' AS dattablespace "
"FROM pg_database d LEFT JOIN pg_shadow u ON (datdba = usesysid) "
"WHERE datallowconn ORDER BY 1");
......@@ -1395,7 +1401,8 @@ dumpCreateDB(PGconn *conn)
"(select usename from pg_shadow where usesysid=(select datdba from pg_database where datname='template0'))), "
"pg_encoding_to_char(d.encoding), "
"null::text AS datcollate, null::text AS datctype, 0 AS datfrozenxid, 0 AS datminmxid, "
"datistemplate, '' as datacl, '' as rdatacl, -1 as datconnlimit, "
"datistemplate, '' as datacl, '' as rdatacl, "
"-1 as datconnlimit, "
"'pg_default' AS dattablespace "
"FROM pg_database d "
"WHERE datallowconn ORDER BY 1");
......@@ -1522,7 +1529,8 @@ dumpCreateDB(PGconn *conn)
}
if (!skip_acls &&
!buildACLCommands(fdbname, NULL, "DATABASE", dbacl, rdbacl, dbowner,
!buildACLCommands(fdbname, NULL, "DATABASE",
dbacl, rdbacl, dbowner,
"", server_version, buf))
{
fprintf(stderr, _("%s: could not parse ACL list (%s) for database \"%s\"\n"),
......
......@@ -161,8 +161,8 @@ my %pgdump_runs = (
'postgres', ], },
role => {
dump_cmd => [
'pg_dump', '-f',
"$tempdir/role.sql", '--role=regress_dump_test_role',
'pg_dump', '-f', "$tempdir/role.sql",
'--role=regress_dump_test_role',
'--schema=dump_test_second_schema', 'postgres', ], },
schema_only => {
dump_cmd =>
......@@ -2192,12 +2192,12 @@ my %tests = (
test_schema_plus_blobs => 1, }, },
'GRANT CREATE ON DATABASE dump_test' => {
create_order => 48,
create_sql => 'GRANT CREATE ON DATABASE dump_test TO regress_dump_test_role;',
create_sql =>
'GRANT CREATE ON DATABASE dump_test TO regress_dump_test_role;',
regexp => qr/^
\QGRANT CREATE ON DATABASE dump_test TO regress_dump_test_role;\E
/xm,
like => {
pg_dumpall_dbprivs => 1, },
like => { pg_dumpall_dbprivs => 1, },
unlike => {
binary_upgrade => 1,
clean => 1,
......@@ -2212,13 +2212,14 @@ my %tests = (
only_dump_test_table => 1,
pg_dumpall_globals => 1,
schema_only => 1,
section_pre_data => 1, ,
section_pre_data => 1,
test_schema_plus_blobs => 1, }, },
'GRANT SELECT ON TABLE test_table' => {
create_order => 5,
create_sql => 'GRANT SELECT ON TABLE dump_test.test_table
TO regress_dump_test_role;',
regexp => qr/^GRANT SELECT ON TABLE test_table TO regress_dump_test_role;/m,
regexp =>
qr/^GRANT SELECT ON TABLE test_table TO regress_dump_test_role;/m,
like => {
binary_upgrade => 1,
clean => 1,
......@@ -2242,7 +2243,8 @@ my %tests = (
create_sql => 'GRANT SELECT ON
TABLE dump_test_second_schema.test_third_table
TO regress_dump_test_role;',
regexp => qr/^GRANT SELECT ON TABLE test_third_table TO regress_dump_test_role;/m,
regexp =>
qr/^GRANT SELECT ON TABLE test_third_table TO regress_dump_test_role;/m,
like => {
binary_upgrade => 1,
clean => 1,
......@@ -2484,8 +2486,7 @@ my %tests = (
\QREVOKE CONNECT,TEMPORARY ON DATABASE dump_test FROM PUBLIC;\E\n
\QGRANT TEMPORARY ON DATABASE dump_test TO PUBLIC;\E
/xm,
like => {
pg_dumpall_dbprivs => 1, },
like => { pg_dumpall_dbprivs => 1, },
unlike => {
binary_upgrade => 1,
clean => 1,
......
......@@ -353,8 +353,8 @@ my %tests = (
no_privs => 1,
pg_dumpall_globals => 1,
section_post_data => 1, }, },
'GRANT SELECT(col2) ON regress_pg_dump_table TO regress_dump_test_role' => {
create_order => 4,
'GRANT SELECT(col2) ON regress_pg_dump_table TO regress_dump_test_role' =>
{ create_order => 4,
create_sql => 'GRANT SELECT(col2) ON regress_pg_dump_table
TO regress_dump_test_role;',
regexp => qr/^
......@@ -373,7 +373,8 @@ my %tests = (
no_privs => 1,
pg_dumpall_globals => 1,
section_post_data => 1, }, },
'GRANT USAGE ON regress_pg_dump_table_col1_seq TO regress_dump_test_role' => {
'GRANT USAGE ON regress_pg_dump_table_col1_seq TO regress_dump_test_role'
=> {
create_order => 5,
create_sql => 'GRANT USAGE ON SEQUENCE regress_pg_dump_table_col1_seq
TO regress_dump_test_role;',
......@@ -397,8 +398,7 @@ my %tests = (
regexp => qr/^
\QGRANT USAGE ON SEQUENCE regress_pg_dump_seq TO regress_dump_test_role;\E
$/xm,
like => {
binary_upgrade => 1, },
like => { binary_upgrade => 1, },
unlike => {
clean => 1,
clean_if_exists => 1,
......
......@@ -475,8 +475,8 @@ sub backup
my $name = $self->name;
print "# Taking pg_basebackup $backup_name from node \"$name\"\n";
TestLib::system_or_bail('pg_basebackup', '-D', $backup_path,
'-p', $port, '-x');
TestLib::system_or_bail('pg_basebackup', '-D', $backup_path, '-p', $port,
'-x');
print "# Backup finished\n";
}
......
......@@ -66,15 +66,15 @@ $node_master->backup('my_backup');
# target TXID.
$node_master->safe_psql('postgres',
"INSERT INTO tab_int VALUES (generate_series(1001,2000))");
my $ret =
$node_master->safe_psql('postgres', "SELECT pg_current_xlog_location(), txid_current();");
my $ret = $node_master->safe_psql('postgres',
"SELECT pg_current_xlog_location(), txid_current();");
my ($lsn2, $recovery_txid) = split /\|/, $ret;
# More data, with recovery target timestamp
$node_master->safe_psql('postgres',
"INSERT INTO tab_int VALUES (generate_series(2001,3000))");
$ret =
$node_master->safe_psql('postgres', "SELECT pg_current_xlog_location(), now();");
$ret = $node_master->safe_psql('postgres',
"SELECT pg_current_xlog_location(), now();");
my ($lsn3, $recovery_time) = split /\|/, $ret;
# Even more data, this time with a recovery target name
......
......@@ -386,13 +386,11 @@ sub GenerateTimezoneFiles
print "Generating timezone files...";
my @args = ("$conf/zic/zic",
'-d',
"$target/share/timezone");
my @args = ("$conf/zic/zic", '-d', "$target/share/timezone");
foreach (@tzfiles)
{
my $tzfile = $_;
push(@args, "src/timezone/data/$tzfile")
push(@args, "src/timezone/data/$tzfile");
}
system(@args);
......@@ -643,8 +641,8 @@ sub CopyIncludeFiles
next unless (-d "src/include/$d");
EnsureDirectories("$target/include/server/$d");
my @args = ('xcopy', '/s', '/i', '/q', '/r', '/y',
"src\\include\\$d\\*.h",
my @args = (
'xcopy', '/s', '/i', '/q', '/r', '/y', "src\\include\\$d\\*.h",
"$ctarget\\include\\server\\$d\\");
system(@args) && croak("Failed to copy include directory $d\n");
}
......@@ -699,9 +697,10 @@ sub GenerateNLSFiles
EnsureDirectories($target, "share/locale/$lang",
"share/locale/$lang/LC_MESSAGES");
my @args = ("$nlspath\\bin\\msgfmt",
my @args = (
"$nlspath\\bin\\msgfmt",
'-o',
"$target\\share\\locale\\$lang\\LC_MESSAGES\\$prgm-$majorver.mo",
"$target\\share\\locale\\$lang\\LC_MESSAGES\\$prgm-$majorver.mo",
$_);
system(@args) && croak("Could not run msgfmt on $dir\\$_");
print ".";
......
......@@ -469,8 +469,9 @@ sub upgradecheck
print "\nSetting up new cluster\n\n";
standard_initdb() or exit 1;
print "\nRunning pg_upgrade\n\n";
@args = ('pg_upgrade', '-d', "$data.old", '-D', $data, '-b', $bindir,
'-B', $bindir);
@args = (
'pg_upgrade', '-d', "$data.old", '-D', $data, '-b',
$bindir, '-B', $bindir);
system(@args) == 0 or exit 1;
print "\nStarting new cluster\n\n";
@args = ('pg_ctl', '-l', "$logdir/postmaster2.log", '-w', 'start');
......
......@@ -40,6 +40,7 @@ AggClauseCosts
AggHashEntry
AggInfo
AggPath
AggSplit
AggState
AggStatePerAgg
AggStatePerGroup
......@@ -123,6 +124,7 @@ ArrayMetaState
ArrayParseState
ArrayRef
ArrayRefExprState
ArrayRemapInfo
ArrayType
AsyncQueueControl
AsyncQueueEntry
......@@ -612,6 +614,8 @@ FmgrBuiltin
FmgrHookEventType
FmgrInfo
ForeignDataWrapper
ForeignKeyCacheInfo
ForeignKeyOptInfo
ForeignPath
ForeignScan
ForeignScanState
......@@ -917,6 +921,7 @@ ImportQual
InclusionOpaque
IncrementVarSublevelsUp_context
Index
IndexAMProperty
IndexAmRoutine
IndexArrayKeyInfo
IndexAttrBitmapKind
......@@ -1239,6 +1244,7 @@ OpFamilyMember
OpFamilyOpFuncGroup
OpclassInfo
Operator
OperatorElement
OpfamilyInfo
OprCacheEntry
OprCacheKey
......@@ -1456,7 +1462,6 @@ ParsedText
ParsedWord
ParserSetupHook
ParserState
PartialAggType
Path
PathClauseUsage
PathCostComparison
......@@ -1651,6 +1656,7 @@ RangeBox
RangeFunction
RangeIOData
RangeQueryClause
RangeRemapInfo
RangeSubselect
RangeTableSample
RangeTblEntry
......@@ -1671,7 +1677,8 @@ RecheckForeignScan_function
RecordCacheEntry
RecordCompareData
RecordIOData
RecordTypemodMap
RecordRemapInfo
RecordTypmodMap
RecoveryTargetAction
RecoveryTargetType
RectBox
......@@ -1708,8 +1715,6 @@ RelfilenodeMapEntry
RelfilenodeMapKey
Relids
RelocationBufferInfo
RemapClass
RemapInfo
RenameStmt
ReopenPtr
ReorderBuffer
......@@ -2123,6 +2128,8 @@ TupleHashEntryData
TupleHashIterator
TupleHashTable
TupleQueueReader
TupleRemapClass
TupleRemapInfo
TupleTableSlot
Tuplesortstate
Tuplestorestate
......@@ -2323,6 +2330,7 @@ amgettuple_function
aminsert_function
ammarkpos_function
amoptions_function
amproperty_function
amrescan_function
amrestrpos_function
amvacuumcleanup_function
......@@ -2353,6 +2361,7 @@ cashKEY
celt
cfp
check_agg_arguments_context
check_function_callback
check_network_data
check_object_relabel_type
check_password_hook_type
......@@ -2374,7 +2383,6 @@ core_yy_extra_type
core_yyscan_t
corrupt_items
cost_qual_eval_context
count_agg_clauses_context
create_upper_paths_hook_type
createdb_failure_params
crosstab_HashEnt
......@@ -2452,6 +2460,7 @@ generate_series_numeric_fctx
generate_series_timestamp_fctx
generate_series_timestamptz_fctx
generate_subscripts_fctx
get_agg_clause_costs_context
get_attavgwidth_hook_type
get_index_stats_hook_type
get_relation_info_hook_type
......@@ -2580,7 +2589,6 @@ pairingheap_comparator
pairingheap_node
parallel_worker_main_type
parse_error_callback_arg
partial_agg_context
pcolor
pendingPosition
pgParameterStatus
......@@ -2839,6 +2847,7 @@ vbits
walrcv_connect_type
walrcv_disconnect_type
walrcv_endstreaming_type
walrcv_get_conninfo_type
walrcv_identify_system_type
walrcv_readtimelinehistoryfile_type
walrcv_receive_type
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment