diff --git a/contrib/auto_explain/auto_explain.c b/contrib/auto_explain/auto_explain.c index e48ea489dc5bb87f48037d058a4aae9f30a54f0d..ad333b664471aee25c15ff6dd4dcde5ac702845c 100644 --- a/contrib/auto_explain/auto_explain.c +++ b/contrib/auto_explain/auto_explain.c @@ -23,7 +23,7 @@ static int auto_explain_log_min_duration = -1; /* msec or -1 */ static bool auto_explain_log_analyze = false; static bool auto_explain_log_verbose = false; static bool auto_explain_log_buffers = false; -static bool auto_explain_log_timing = false; +static bool auto_explain_log_timing = false; static int auto_explain_log_format = EXPLAIN_FORMAT_TEXT; static bool auto_explain_log_nested_statements = false; diff --git a/contrib/dblink/dblink.c b/contrib/dblink/dblink.c index 71acb35af6ff3a7d6e5a7cd3e57688cb56a687fb..1e62d8091a9d2bdf60af6745d5a01ee14ee5cf5a 100644 --- a/contrib/dblink/dblink.c +++ b/contrib/dblink/dblink.c @@ -1140,7 +1140,7 @@ storeHandler(PGresult *res, const PGdataValue *columns, * strings and add null termination. As a micro-optimization, allocate * all the strings with one palloc. */ - pbuflen = nfields; /* count the null terminators themselves */ + pbuflen = nfields; /* count the null terminators themselves */ for (i = 0; i < nfields; i++) { int len = columns[i].len; diff --git a/contrib/file_fdw/file_fdw.c b/contrib/file_fdw/file_fdw.c index 66fd0e62cc8da2b9f9b31ae4568ddf1fb3b0d3b6..e3b9223b3ee7a102efef7e1c18192eb64494864b 100644 --- a/contrib/file_fdw/file_fdw.c +++ b/contrib/file_fdw/file_fdw.c @@ -109,17 +109,17 @@ PG_FUNCTION_INFO_V1(file_fdw_validator); * FDW callback routines */ static void fileGetForeignRelSize(PlannerInfo *root, - RelOptInfo *baserel, - Oid foreigntableid); + RelOptInfo *baserel, + Oid foreigntableid); static void fileGetForeignPaths(PlannerInfo *root, - RelOptInfo *baserel, - Oid foreigntableid); + RelOptInfo *baserel, + Oid foreigntableid); static ForeignScan *fileGetForeignPlan(PlannerInfo *root, - RelOptInfo *baserel, - Oid foreigntableid, - ForeignPath *best_path, - List *tlist, - List *scan_clauses); + RelOptInfo *baserel, + Oid foreigntableid, + ForeignPath *best_path, + List *tlist, + List *scan_clauses); static void fileExplainForeignScan(ForeignScanState *node, ExplainState *es); static void fileBeginForeignScan(ForeignScanState *node, int eflags); static TupleTableSlot *fileIterateForeignScan(ForeignScanState *node); @@ -141,7 +141,7 @@ static void estimate_size(PlannerInfo *root, RelOptInfo *baserel, static void estimate_costs(PlannerInfo *root, RelOptInfo *baserel, FileFdwPlanState *fdw_private, Cost *startup_cost, Cost *total_cost); -static int file_acquire_sample_rows(Relation onerel, int elevel, +static int file_acquire_sample_rows(Relation onerel, int elevel, HeapTuple *rows, int targrows, double *totalrows, double *totaldeadrows); @@ -180,7 +180,7 @@ file_fdw_validator(PG_FUNCTION_ARGS) List *options_list = untransformRelOptions(PG_GETARG_DATUM(0)); Oid catalog = PG_GETARG_OID(1); char *filename = NULL; - DefElem *force_not_null = NULL; + DefElem *force_not_null = NULL; List *other_options = NIL; ListCell *cell; @@ -233,7 +233,7 @@ file_fdw_validator(PG_FUNCTION_ARGS) buf.len > 0 ? errhint("Valid options in this context are: %s", buf.data) - : errhint("There are no valid options in this context."))); + : errhint("There are no valid options in this context."))); } /* @@ -393,13 +393,13 @@ get_file_fdw_attribute_options(Oid relid) options = GetForeignColumnOptions(relid, attnum); foreach(lc, options) { - DefElem *def = (DefElem *) lfirst(lc); + DefElem *def = (DefElem *) lfirst(lc); if (strcmp(def->defname, "force_not_null") == 0) { if (defGetBoolean(def)) { - char *attname = pstrdup(NameStr(attr->attname)); + char *attname = pstrdup(NameStr(attr->attname)); fnncolumns = lappend(fnncolumns, makeString(attname)); } @@ -429,8 +429,8 @@ fileGetForeignRelSize(PlannerInfo *root, FileFdwPlanState *fdw_private; /* - * Fetch options. We only need filename at this point, but we might - * as well get everything and not need to re-fetch it later in planning. + * Fetch options. We only need filename at this point, but we might as + * well get everything and not need to re-fetch it later in planning. */ fdw_private = (FileFdwPlanState *) palloc(sizeof(FileFdwPlanState)); fileGetOptions(foreigntableid, @@ -468,13 +468,14 @@ fileGetForeignPaths(PlannerInfo *root, baserel->rows, startup_cost, total_cost, - NIL, /* no pathkeys */ - NULL, /* no outer rel either */ - NIL)); /* no fdw_private data */ + NIL, /* no pathkeys */ + NULL, /* no outer rel either */ + NIL)); /* no fdw_private data */ /* * If data file was sorted, and we knew it somehow, we could insert - * appropriate pathkeys into the ForeignPath node to tell the planner that. + * appropriate pathkeys into the ForeignPath node to tell the planner + * that. */ } @@ -505,8 +506,8 @@ fileGetForeignPlan(PlannerInfo *root, return make_foreignscan(tlist, scan_clauses, scan_relid, - NIL, /* no expressions to evaluate */ - NIL); /* no private state either */ + NIL, /* no expressions to evaluate */ + NIL); /* no private state either */ } /* @@ -665,14 +666,14 @@ fileAnalyzeForeignTable(Relation relation, { char *filename; List *options; - struct stat stat_buf; + struct stat stat_buf; /* Fetch options of foreign table */ fileGetOptions(RelationGetRelid(relation), &filename, &options); /* - * Get size of the file. (XXX if we fail here, would it be better to - * just return false to skip analyzing the table?) + * Get size of the file. (XXX if we fail here, would it be better to just + * return false to skip analyzing the table?) */ if (stat(filename, &stat_buf) < 0) ereport(ERROR, @@ -746,7 +747,7 @@ estimate_size(PlannerInfo *root, RelOptInfo *baserel, * planner's idea of the relation width; which is bogus if not all * columns are being read, not to mention that the text representation * of a row probably isn't the same size as its internal - * representation. Possibly we could do something better, but the + * representation. Possibly we could do something better, but the * real answer to anyone who complains is "ANALYZE" ... */ int tuple_width; @@ -811,7 +812,7 @@ estimate_costs(PlannerInfo *root, RelOptInfo *baserel, * which must have at least targrows entries. * The actual number of rows selected is returned as the function result. * We also count the total number of rows in the file and return it into - * *totalrows. Note that *totaldeadrows is always set to 0. + * *totalrows. Note that *totaldeadrows is always set to 0. * * Note that the returned list of rows is not always in order by physical * position in the file. Therefore, correlation estimates derived later @@ -824,7 +825,7 @@ file_acquire_sample_rows(Relation onerel, int elevel, double *totalrows, double *totaldeadrows) { int numrows = 0; - double rowstoskip = -1; /* -1 means not set yet */ + double rowstoskip = -1; /* -1 means not set yet */ double rstate; TupleDesc tupDesc; Datum *values; @@ -853,8 +854,8 @@ file_acquire_sample_rows(Relation onerel, int elevel, cstate = BeginCopyFrom(onerel, filename, NIL, options); /* - * Use per-tuple memory context to prevent leak of memory used to read rows - * from the file with Copy routines. + * Use per-tuple memory context to prevent leak of memory used to read + * rows from the file with Copy routines. */ tupcontext = AllocSetContextCreate(CurrentMemoryContext, "file_fdw temporary context", @@ -912,10 +913,10 @@ file_acquire_sample_rows(Relation onerel, int elevel, if (rowstoskip <= 0) { /* - * Found a suitable tuple, so save it, replacing one - * old tuple at random + * Found a suitable tuple, so save it, replacing one old tuple + * at random */ - int k = (int) (targrows * anl_random_fract()); + int k = (int) (targrows * anl_random_fract()); Assert(k >= 0 && k < targrows); heap_freetuple(rows[k]); diff --git a/contrib/pg_archivecleanup/pg_archivecleanup.c b/contrib/pg_archivecleanup/pg_archivecleanup.c index 20977805c8deab6df569263894069afa4a494ca5..a226101bbc47b36e3f21a9ae374b65adbcf5b166 100644 --- a/contrib/pg_archivecleanup/pg_archivecleanup.c +++ b/contrib/pg_archivecleanup/pg_archivecleanup.c @@ -37,7 +37,7 @@ const char *progname; /* Options and defaults */ bool debug = false; /* are we debugging? */ bool dryrun = false; /* are we performing a dry-run operation? */ -char *additional_ext = NULL; /* Extension to remove from filenames */ +char *additional_ext = NULL; /* Extension to remove from filenames */ char *archiveLocation; /* where to find the archive? */ char *restartWALFileName; /* the file from which we can restart restore */ @@ -136,12 +136,13 @@ CleanupPriorWALFiles(void) * they were originally written, in case this worries you. */ if (strlen(walfile) == XLOG_DATA_FNAME_LEN && - strspn(walfile, "0123456789ABCDEF") == XLOG_DATA_FNAME_LEN && + strspn(walfile, "0123456789ABCDEF") == XLOG_DATA_FNAME_LEN && strcmp(walfile + 8, exclusiveCleanupFileName + 8) < 0) { - /* - * Use the original file name again now, including any extension - * that might have been chopped off before testing the sequence. + /* + * Use the original file name again now, including any + * extension that might have been chopped off before testing + * the sequence. */ snprintf(WALFilePath, MAXPGPATH, "%s/%s", archiveLocation, xlde->d_name); @@ -150,7 +151,7 @@ CleanupPriorWALFiles(void) { /* * Prints the name of the file to be removed and skips the - * actual removal. The regular printout is so that the + * actual removal. The regular printout is so that the * user can pipe the output into some other program. */ printf("%s\n", WALFilePath); @@ -298,7 +299,8 @@ main(int argc, char **argv) dryrun = true; break; case 'x': - additional_ext = optarg; /* Extension to remove from xlogfile names */ + additional_ext = optarg; /* Extension to remove from + * xlogfile names */ break; default: fprintf(stderr, "Try \"%s --help\" for more information.\n", progname); diff --git a/contrib/pg_stat_statements/pg_stat_statements.c b/contrib/pg_stat_statements/pg_stat_statements.c index 06869fa344e8a5b23c613574f26787dd2e424fb8..aa11c144d68cf4c8a224d1cdcdd61a74c824ed47 100644 --- a/contrib/pg_stat_statements/pg_stat_statements.c +++ b/contrib/pg_stat_statements/pg_stat_statements.c @@ -103,19 +103,19 @@ typedef struct Counters int64 calls; /* # of times executed */ double total_time; /* total execution time, in msec */ int64 rows; /* total # of retrieved or affected rows */ - int64 shared_blks_hit; /* # of shared buffer hits */ + int64 shared_blks_hit; /* # of shared buffer hits */ int64 shared_blks_read; /* # of shared disk blocks read */ int64 shared_blks_dirtied; /* # of shared disk blocks dirtied */ int64 shared_blks_written; /* # of shared disk blocks written */ - int64 local_blks_hit; /* # of local buffer hits */ - int64 local_blks_read; /* # of local disk blocks read */ + int64 local_blks_hit; /* # of local buffer hits */ + int64 local_blks_read; /* # of local disk blocks read */ int64 local_blks_dirtied; /* # of local disk blocks dirtied */ int64 local_blks_written; /* # of local disk blocks written */ - int64 temp_blks_read; /* # of temp blocks read */ + int64 temp_blks_read; /* # of temp blocks read */ int64 temp_blks_written; /* # of temp blocks written */ - double blk_read_time; /* time spent reading, in msec */ - double blk_write_time; /* time spent writing, in msec */ - double usage; /* usage factor */ + double blk_read_time; /* time spent reading, in msec */ + double blk_write_time; /* time spent writing, in msec */ + double usage; /* usage factor */ } Counters; /* @@ -140,7 +140,7 @@ typedef struct pgssSharedState { LWLockId lock; /* protects hashtable search/modification */ int query_size; /* max query length in bytes */ - double cur_median_usage; /* current median usage in hashtable */ + double cur_median_usage; /* current median usage in hashtable */ } pgssSharedState; /* @@ -150,7 +150,7 @@ typedef struct pgssLocationLen { int location; /* start offset in query text */ int length; /* length in bytes, or -1 to ignore */ -} pgssLocationLen; +} pgssLocationLen; /* * Working state for computing a query jumble and producing a normalized @@ -172,7 +172,7 @@ typedef struct pgssJumbleState /* Current number of valid entries in clocations array */ int clocations_count; -} pgssJumbleState; +} pgssJumbleState; /*---- Local variables ----*/ @@ -248,21 +248,21 @@ static uint32 pgss_hash_string(const char *str); static void pgss_store(const char *query, uint32 queryId, double total_time, uint64 rows, const BufferUsage *bufusage, - pgssJumbleState * jstate); + pgssJumbleState *jstate); static Size pgss_memsize(void); static pgssEntry *entry_alloc(pgssHashKey *key, const char *query, - int query_len, bool sticky); + int query_len, bool sticky); static void entry_dealloc(void); static void entry_reset(void); -static void AppendJumble(pgssJumbleState * jstate, +static void AppendJumble(pgssJumbleState *jstate, const unsigned char *item, Size size); -static void JumbleQuery(pgssJumbleState * jstate, Query *query); -static void JumbleRangeTable(pgssJumbleState * jstate, List *rtable); -static void JumbleExpr(pgssJumbleState * jstate, Node *node); -static void RecordConstLocation(pgssJumbleState * jstate, int location); -static char *generate_normalized_query(pgssJumbleState * jstate, const char *query, +static void JumbleQuery(pgssJumbleState *jstate, Query *query); +static void JumbleRangeTable(pgssJumbleState *jstate, List *rtable); +static void JumbleExpr(pgssJumbleState *jstate, Node *node); +static void RecordConstLocation(pgssJumbleState *jstate, int location); +static char *generate_normalized_query(pgssJumbleState *jstate, const char *query, int *query_len_p, int encoding); -static void fill_in_constant_lengths(pgssJumbleState * jstate, const char *query); +static void fill_in_constant_lengths(pgssJumbleState *jstate, const char *query); static int comp_location(const void *a, const void *b); @@ -513,8 +513,8 @@ pgss_shmem_startup(void) FreeFile(file); /* - * Remove the file so it's not included in backups/replication - * slaves, etc. A new file will be written on next shutdown. + * Remove the file so it's not included in backups/replication slaves, + * etc. A new file will be written on next shutdown. */ unlink(PGSS_DUMP_FILE); @@ -600,7 +600,7 @@ error: ereport(LOG, (errcode_for_file_access(), errmsg("could not write pg_stat_statement file \"%s\": %m", - PGSS_DUMP_FILE ".tmp"))); + PGSS_DUMP_FILE ".tmp"))); if (file) FreeFile(file); unlink(PGSS_DUMP_FILE ".tmp"); @@ -626,8 +626,8 @@ pgss_post_parse_analyze(ParseState *pstate, Query *query) * the statement contains an optimizable statement for which a queryId * could be derived (such as EXPLAIN or DECLARE CURSOR). For such cases, * runtime control will first go through ProcessUtility and then the - * executor, and we don't want the executor hooks to do anything, since - * we are already measuring the statement's costs at the utility level. + * executor, and we don't want the executor hooks to do anything, since we + * are already measuring the statement's costs at the utility level. */ if (query->utilityStmt) { @@ -768,7 +768,7 @@ pgss_ExecutorEnd(QueryDesc *queryDesc) pgss_store(queryDesc->sourceText, queryId, - queryDesc->totaltime->total * 1000.0, /* convert to msec */ + queryDesc->totaltime->total * 1000.0, /* convert to msec */ queryDesc->estate->es_processed, &queryDesc->totaltime->bufusage, NULL); @@ -789,10 +789,9 @@ pgss_ProcessUtility(Node *parsetree, const char *queryString, DestReceiver *dest, char *completionTag) { /* - * If it's an EXECUTE statement, we don't track it and don't increment - * the nesting level. This allows the cycles to be charged to the - * underlying PREPARE instead (by the Executor hooks), which is much more - * useful. + * If it's an EXECUTE statement, we don't track it and don't increment the + * nesting level. This allows the cycles to be charged to the underlying + * PREPARE instead (by the Executor hooks), which is much more useful. * * We also don't track execution of PREPARE. If we did, we would get one * hash table entry for the PREPARE (with hash calculated from the query @@ -942,7 +941,7 @@ static void pgss_store(const char *query, uint32 queryId, double total_time, uint64 rows, const BufferUsage *bufusage, - pgssJumbleState * jstate) + pgssJumbleState *jstate) { pgssHashKey key; pgssEntry *entry; @@ -1355,7 +1354,7 @@ entry_reset(void) * the current jumble. */ static void -AppendJumble(pgssJumbleState * jstate, const unsigned char *item, Size size) +AppendJumble(pgssJumbleState *jstate, const unsigned char *item, Size size) { unsigned char *jumble = jstate->jumble; Size jumble_len = jstate->jumble_len; @@ -1404,7 +1403,7 @@ AppendJumble(pgssJumbleState * jstate, const unsigned char *item, Size size) * of information). */ static void -JumbleQuery(pgssJumbleState * jstate, Query *query) +JumbleQuery(pgssJumbleState *jstate, Query *query) { Assert(IsA(query, Query)); Assert(query->utilityStmt == NULL); @@ -1431,7 +1430,7 @@ JumbleQuery(pgssJumbleState * jstate, Query *query) * Jumble a range table */ static void -JumbleRangeTable(pgssJumbleState * jstate, List *rtable) +JumbleRangeTable(pgssJumbleState *jstate, List *rtable) { ListCell *lc; @@ -1485,11 +1484,11 @@ JumbleRangeTable(pgssJumbleState * jstate, List *rtable) * * Note: the reason we don't simply use expression_tree_walker() is that the * point of that function is to support tree walkers that don't care about - * most tree node types, but here we care about all types. We should complain + * most tree node types, but here we care about all types. We should complain * about any unrecognized node type. */ static void -JumbleExpr(pgssJumbleState * jstate, Node *node) +JumbleExpr(pgssJumbleState *jstate, Node *node) { ListCell *temp; @@ -1874,7 +1873,7 @@ JumbleExpr(pgssJumbleState * jstate, Node *node) * that is currently being walked. */ static void -RecordConstLocation(pgssJumbleState * jstate, int location) +RecordConstLocation(pgssJumbleState *jstate, int location) { /* -1 indicates unknown or undefined location */ if (location >= 0) @@ -1909,7 +1908,7 @@ RecordConstLocation(pgssJumbleState * jstate, int location) * Returns a palloc'd string, which is not necessarily null-terminated. */ static char * -generate_normalized_query(pgssJumbleState * jstate, const char *query, +generate_normalized_query(pgssJumbleState *jstate, const char *query, int *query_len_p, int encoding) { char *norm_query; @@ -2007,7 +2006,7 @@ generate_normalized_query(pgssJumbleState * jstate, const char *query, * a problem. * * Duplicate constant pointers are possible, and will have their lengths - * marked as '-1', so that they are later ignored. (Actually, we assume the + * marked as '-1', so that they are later ignored. (Actually, we assume the * lengths were initialized as -1 to start with, and don't change them here.) * * N.B. There is an assumption that a '-' character at a Const location begins @@ -2015,7 +2014,7 @@ generate_normalized_query(pgssJumbleState * jstate, const char *query, * reason for a constant to start with a '-'. */ static void -fill_in_constant_lengths(pgssJumbleState * jstate, const char *query) +fill_in_constant_lengths(pgssJumbleState *jstate, const char *query) { pgssLocationLen *locs; core_yyscan_t yyscanner; diff --git a/contrib/pg_test_fsync/pg_test_fsync.c b/contrib/pg_test_fsync/pg_test_fsync.c index 7f92bc88182beedf3bfd1881f90ec5c4d0fcf4ce..9fe2301e419a569d6f6111c606a22c01c2690c9a 100644 --- a/contrib/pg_test_fsync/pg_test_fsync.c +++ b/contrib/pg_test_fsync/pg_test_fsync.c @@ -29,7 +29,7 @@ /* These are macros to avoid timing the function call overhead. */ #ifndef WIN32 -#define START_TIMER \ +#define START_TIMER \ do { \ alarm_triggered = false; \ alarm(secs_per_test); \ @@ -37,7 +37,7 @@ do { \ } while (0) #else /* WIN32 doesn't support alarm, so we create a thread and sleep there */ -#define START_TIMER \ +#define START_TIMER \ do { \ alarm_triggered = false; \ if (CreateThread(NULL, 0, process_alarm, NULL, 0, NULL) == \ @@ -55,7 +55,7 @@ do { \ gettimeofday(&stop_t, NULL); \ print_elapse(start_t, stop_t, ops); \ } while (0) - + static const char *progname; @@ -77,6 +77,7 @@ static void test_sync(int writes_per_op); static void test_open_syncs(void); static void test_open_sync(const char *msg, int writes_size); static void test_file_descriptor_sync(void); + #ifndef WIN32 static void process_alarm(int sig); #else diff --git a/contrib/pg_test_timing/pg_test_timing.c b/contrib/pg_test_timing/pg_test_timing.c index 4e43694338bed648fe0b674952465b694e058844..b3f98abe5cb6b1bc37253d00d0ba68b6fd75e543 100644 --- a/contrib/pg_test_timing/pg_test_timing.c +++ b/contrib/pg_test_timing/pg_test_timing.c @@ -1,7 +1,7 @@ /* * pg_test_timing.c - * tests overhead of timing calls and their monotonicity: that - * they always move forward + * tests overhead of timing calls and their monotonicity: that + * they always move forward */ #include "postgres_fe.h" @@ -35,8 +35,8 @@ handle_args(int argc, char *argv[]) {"duration", required_argument, NULL, 'd'}, {NULL, 0, NULL, 0} }; - int option; /* Command line option */ - int optindex = 0; /* used by getopt_long */ + int option; /* Command line option */ + int optindex = 0; /* used by getopt_long */ if (argc > 1) { @@ -87,7 +87,7 @@ handle_args(int argc, char *argv[]) else { fprintf(stderr, - "%s: duration must be a positive integer (duration is \"%d\")\n", + "%s: duration must be a positive integer (duration is \"%d\")\n", progname, test_duration); fprintf(stderr, "Try \"%s --help\" for more information.\n", progname); @@ -98,16 +98,22 @@ handle_args(int argc, char *argv[]) static void test_timing(int32 duration) { - uint64 total_time; - int64 time_elapsed = 0; - uint64 loop_count = 0; - uint64 prev, cur; - int32 diff, i, bits, found; - - instr_time start_time, end_time, temp; + uint64 total_time; + int64 time_elapsed = 0; + uint64 loop_count = 0; + uint64 prev, + cur; + int32 diff, + i, + bits, + found; + + instr_time start_time, + end_time, + temp; static int64 histogram[32]; - char buf[100]; + char buf[100]; total_time = duration > 0 ? duration * 1000000 : 0; @@ -146,7 +152,7 @@ test_timing(int32 duration) INSTR_TIME_SUBTRACT(end_time, start_time); printf("Per loop time including overhead: %0.2f nsec\n", - INSTR_TIME_GET_DOUBLE(end_time) * 1e9 / loop_count); + INSTR_TIME_GET_DOUBLE(end_time) * 1e9 / loop_count); printf("Histogram of timing durations:\n"); printf("%9s: %10s %9s\n", "< usec", "count", "percent"); diff --git a/contrib/pg_trgm/trgm_gist.c b/contrib/pg_trgm/trgm_gist.c index 57bce012070a49bb2c5e2d45140bffcd8a307669..d59c8eb670b44c4c8b79e6b92baa0f68e6ce4e73 100644 --- a/contrib/pg_trgm/trgm_gist.c +++ b/contrib/pg_trgm/trgm_gist.c @@ -199,9 +199,9 @@ gtrgm_consistent(PG_FUNCTION_ARGS) * trigram extraction is relatively CPU-expensive. We must include * strategy number because trigram extraction depends on strategy. * - * The cached structure contains the strategy number, then the input - * query (starting at a MAXALIGN boundary), then the TRGM value (also - * starting at a MAXALIGN boundary). + * The cached structure contains the strategy number, then the input query + * (starting at a MAXALIGN boundary), then the TRGM value (also starting + * at a MAXALIGN boundary). */ if (cache == NULL || strategy != *((StrategyNumber *) cache) || @@ -341,8 +341,7 @@ gtrgm_distance(PG_FUNCTION_ARGS) char *cache = (char *) fcinfo->flinfo->fn_extra; /* - * Cache the generated trigrams across multiple calls with the same - * query. + * Cache the generated trigrams across multiple calls with the same query. */ if (cache == NULL || VARSIZE(cache) != querysize || diff --git a/contrib/pg_upgrade/check.c b/contrib/pg_upgrade/check.c index 2669c096586743aa2580696915d00f10abaeed3c..eed4a1eba7c7cebf271aa9ba22728f515a65d358 100644 --- a/contrib/pg_upgrade/check.c +++ b/contrib/pg_upgrade/check.c @@ -168,7 +168,7 @@ issue_warnings(char *sequence_script_file_name) SYSTEMQUOTE "\"%s/psql\" --echo-queries " "--set ON_ERROR_STOP=on " "--no-psqlrc --port %d --username \"%s\" " - "-f \"%s\" --dbname template1 >> \"%s\" 2>&1" SYSTEMQUOTE, + "-f \"%s\" --dbname template1 >> \"%s\" 2>&1" SYSTEMQUOTE, new_cluster.bindir, new_cluster.port, os_info.user, sequence_script_file_name, UTILITY_LOG_FILE); unlink(sequence_script_file_name); @@ -204,7 +204,7 @@ output_completion_banner(char *analyze_script_file_name, else pg_log(PG_REPORT, "Optimizer statistics and free space information are not transferred\n" - "by pg_upgrade so, once you start the new server, consider running:\n" + "by pg_upgrade so, once you start the new server, consider running:\n" " %s\n\n", analyze_script_file_name); pg_log(PG_REPORT, @@ -238,7 +238,8 @@ check_cluster_versions(void) /* * We can't allow downgrading because we use the target pg_dumpall, and - * pg_dumpall cannot operate on new database versions, only older versions. + * pg_dumpall cannot operate on new database versions, only older + * versions. */ if (old_cluster.major_version > new_cluster.major_version) pg_log(PG_FATAL, "This utility cannot be used to downgrade to older major PostgreSQL versions.\n"); @@ -402,31 +403,31 @@ create_script_for_cluster_analyze(char **analyze_script_file_name) #endif fprintf(script, "echo %sThis script will generate minimal optimizer statistics rapidly%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %sso your system is usable, and then gather statistics twice more%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %swith increasing accuracy. When it is done, your system will%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %shave the default level of optimizer statistics.%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo\n\n"); fprintf(script, "echo %sIf you have used ALTER TABLE to modify the statistics target for%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %sany tables, you might want to remove them and restore them after%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %srunning this script because they will delay fast statistics generation.%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo\n\n"); fprintf(script, "echo %sIf you would like default statistics as quickly as possible, cancel%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %sthis script and run:%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %s vacuumdb --all %s%s\n", ECHO_QUOTE, - /* Did we copy the free space files? */ - (GET_MAJOR_VERSION(old_cluster.major_version) >= 804) ? - "--analyze-only" : "--analyze", ECHO_QUOTE); + /* Did we copy the free space files? */ + (GET_MAJOR_VERSION(old_cluster.major_version) >= 804) ? + "--analyze-only" : "--analyze", ECHO_QUOTE); fprintf(script, "echo\n\n"); #ifndef WIN32 @@ -441,15 +442,15 @@ create_script_for_cluster_analyze(char **analyze_script_file_name) #endif fprintf(script, "echo %sGenerating minimal optimizer statistics (1 target)%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %s--------------------------------------------------%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "vacuumdb --all --analyze-only\n"); fprintf(script, "echo\n"); fprintf(script, "echo %sThe server is now available with minimal optimizer statistics.%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %sQuery performance will be optimal once this script completes.%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo\n\n"); #ifndef WIN32 @@ -462,9 +463,9 @@ create_script_for_cluster_analyze(char **analyze_script_file_name) #endif fprintf(script, "echo %sGenerating medium optimizer statistics (10 targets)%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %s---------------------------------------------------%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "vacuumdb --all --analyze-only\n"); fprintf(script, "echo\n\n"); @@ -475,17 +476,17 @@ create_script_for_cluster_analyze(char **analyze_script_file_name) #endif fprintf(script, "echo %sGenerating default (full) optimizer statistics (100 targets?)%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "echo %s-------------------------------------------------------------%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fprintf(script, "vacuumdb --all %s\n", - /* Did we copy the free space files? */ - (GET_MAJOR_VERSION(old_cluster.major_version) >= 804) ? - "--analyze-only" : "--analyze"); + /* Did we copy the free space files? */ + (GET_MAJOR_VERSION(old_cluster.major_version) >= 804) ? + "--analyze-only" : "--analyze"); fprintf(script, "echo\n\n"); fprintf(script, "echo %sDone%s\n", - ECHO_QUOTE, ECHO_QUOTE); + ECHO_QUOTE, ECHO_QUOTE); fclose(script); @@ -716,8 +717,8 @@ check_for_isn_and_int8_passing_mismatch(ClusterInfo *cluster) pg_log(PG_REPORT, "fatal\n"); pg_log(PG_FATAL, "Your installation contains \"contrib/isn\" functions which rely on the\n" - "bigint data type. Your old and new clusters pass bigint values\n" - "differently so this cluster cannot currently be upgraded. You can\n" + "bigint data type. Your old and new clusters pass bigint values\n" + "differently so this cluster cannot currently be upgraded. You can\n" "manually upgrade databases that use \"contrib/isn\" facilities and remove\n" "\"contrib/isn\" from the old cluster and restart the upgrade. A list of\n" "the problem functions is in the file:\n" @@ -764,9 +765,9 @@ check_for_reg_data_type_usage(ClusterInfo *cluster) PGconn *conn = connectToServer(cluster, active_db->db_name); /* - * While several relkinds don't store any data, e.g. views, they - * can be used to define data types of other columns, so we - * check all relkinds. + * While several relkinds don't store any data, e.g. views, they can + * be used to define data types of other columns, so we check all + * relkinds. */ res = executeQueryOrDie(conn, "SELECT n.nspname, c.relname, a.attname " @@ -777,16 +778,16 @@ check_for_reg_data_type_usage(ClusterInfo *cluster) " NOT a.attisdropped AND " " a.atttypid IN ( " " 'pg_catalog.regproc'::pg_catalog.regtype, " - " 'pg_catalog.regprocedure'::pg_catalog.regtype, " + " 'pg_catalog.regprocedure'::pg_catalog.regtype, " " 'pg_catalog.regoper'::pg_catalog.regtype, " - " 'pg_catalog.regoperator'::pg_catalog.regtype, " + " 'pg_catalog.regoperator'::pg_catalog.regtype, " /* regclass.oid is preserved, so 'regclass' is OK */ /* regtype.oid is preserved, so 'regtype' is OK */ - " 'pg_catalog.regconfig'::pg_catalog.regtype, " - " 'pg_catalog.regdictionary'::pg_catalog.regtype) AND " - " c.relnamespace = n.oid AND " - " n.nspname != 'pg_catalog' AND " - " n.nspname != 'information_schema'"); + " 'pg_catalog.regconfig'::pg_catalog.regtype, " + " 'pg_catalog.regdictionary'::pg_catalog.regtype) AND " + " c.relnamespace = n.oid AND " + " n.nspname != 'pg_catalog' AND " + " n.nspname != 'information_schema'"); ntups = PQntuples(res); i_nspname = PQfnumber(res, "nspname"); @@ -822,8 +823,8 @@ check_for_reg_data_type_usage(ClusterInfo *cluster) pg_log(PG_REPORT, "fatal\n"); pg_log(PG_FATAL, "Your installation contains one of the reg* data types in user tables.\n" - "These data types reference system OIDs that are not preserved by\n" - "pg_upgrade, so this cluster cannot currently be upgraded. You can\n" + "These data types reference system OIDs that are not preserved by\n" + "pg_upgrade, so this cluster cannot currently be upgraded. You can\n" "remove the problem tables and restart the upgrade. A list of the problem\n" "columns is in the file:\n" " %s\n\n", output_path); @@ -836,9 +837,11 @@ check_for_reg_data_type_usage(ClusterInfo *cluster) static void get_bin_version(ClusterInfo *cluster) { - char cmd[MAXPGPATH], cmd_output[MAX_STRING]; + char cmd[MAXPGPATH], + cmd_output[MAX_STRING]; FILE *output; - int pre_dot, post_dot; + int pre_dot, + post_dot; snprintf(cmd, sizeof(cmd), "\"%s/pg_ctl\" --version", cluster->bindir); @@ -858,4 +861,3 @@ get_bin_version(ClusterInfo *cluster) cluster->bin_version = (pre_dot * 100 + post_dot) * 100; } - diff --git a/contrib/pg_upgrade/controldata.c b/contrib/pg_upgrade/controldata.c index e01280db9e284b5527c5d7db0afc481247e1e017..6bffe549e59610846eaa391f63bb956ab05d37f3 100644 --- a/contrib/pg_upgrade/controldata.c +++ b/contrib/pg_upgrade/controldata.c @@ -129,6 +129,7 @@ get_control_data(ClusterInfo *cluster, bool live_check) pg_log(PG_VERBOSE, "%s", bufin); #ifdef WIN32 + /* * Due to an installer bug, LANG=C doesn't work for PG 8.3.3, but does * work 8.2.6 and 8.3.7, so check for non-ASCII output and suggest a @@ -506,7 +507,7 @@ check_control_data(ControlData *oldctrl, * This is a common 8.3 -> 8.4 upgrade problem, so we are more verbose */ pg_log(PG_FATAL, - "You will need to rebuild the new server with configure option\n" + "You will need to rebuild the new server with configure option\n" "--disable-integer-datetimes or get server binaries built with those\n" "options.\n"); } @@ -531,6 +532,6 @@ disable_old_cluster(void) pg_log(PG_REPORT, "\n" "If you want to start the old cluster, you will need to remove\n" "the \".old\" suffix from %s/global/pg_control.old.\n" - "Because \"link\" mode was used, the old cluster cannot be safely\n" - "started once the new cluster has been started.\n\n", old_cluster.pgdata); + "Because \"link\" mode was used, the old cluster cannot be safely\n" + "started once the new cluster has been started.\n\n", old_cluster.pgdata); } diff --git a/contrib/pg_upgrade/exec.c b/contrib/pg_upgrade/exec.c index 68cf0795aabdb7f734d2528980cf4bbad6a8a46f..9e63bd5856839b1c6f099cdb0b5e53ea34ec19d4 100644 --- a/contrib/pg_upgrade/exec.c +++ b/contrib/pg_upgrade/exec.c @@ -18,8 +18,9 @@ static void check_data_dir(const char *pg_data); static void check_bin_dir(ClusterInfo *cluster); static void validate_exec(const char *dir, const char *cmdName); + #ifdef WIN32 -static int win32_check_directory_write_permissions(void); +static int win32_check_directory_write_permissions(void); #endif @@ -64,7 +65,7 @@ exec_prog(bool throw_error, bool is_priv, pg_log(throw_error ? PG_FATAL : PG_REPORT, "Consult the last few lines of \"%s\" for\n" "the probable cause of the failure.\n", - log_file); + log_file); return 1; } @@ -142,12 +143,12 @@ verify_directories(void) static int win32_check_directory_write_permissions(void) { - int fd; + int fd; /* - * We open a file we would normally create anyway. We do this even in - * 'check' mode, which isn't ideal, but this is the best we can do. - */ + * We open a file we would normally create anyway. We do this even in + * 'check' mode, which isn't ideal, but this is the best we can do. + */ if ((fd = open(GLOBALS_DUMP_FILE, O_RDWR | O_CREAT, S_IRUSR | S_IWUSR)) < 0) return -1; close(fd); @@ -184,7 +185,7 @@ check_data_dir(const char *pg_data) struct stat statBuf; snprintf(subDirName, sizeof(subDirName), "%s%s%s", pg_data, - /* Win32 can't stat() a directory with a trailing slash. */ + /* Win32 can't stat() a directory with a trailing slash. */ *requiredSubdirs[subdirnum] ? "/" : "", requiredSubdirs[subdirnum]); diff --git a/contrib/pg_upgrade/file.c b/contrib/pg_upgrade/file.c index 0276636e03fa4208134b0d4aa91c2c037c9c432f..1dd3722142c9e83c1ec228099c3a3fd302a2179b 100644 --- a/contrib/pg_upgrade/file.c +++ b/contrib/pg_upgrade/file.c @@ -233,7 +233,7 @@ copy_file(const char *srcfile, const char *dstfile, bool force) * large number of times. */ int -load_directory(const char *dirname, struct dirent ***namelist) +load_directory(const char *dirname, struct dirent *** namelist) { DIR *dirdesc; struct dirent *direntry; @@ -251,7 +251,7 @@ load_directory(const char *dirname, struct dirent ***namelist) count++; *namelist = (struct dirent **) realloc((void *) (*namelist), - (size_t) ((name_num + 1) * sizeof(struct dirent *))); + (size_t) ((name_num + 1) * sizeof(struct dirent *))); if (*namelist == NULL) { @@ -314,7 +314,6 @@ win32_pghardlink(const char *src, const char *dst) else return 0; } - #endif @@ -322,13 +321,11 @@ win32_pghardlink(const char *src, const char *dst) FILE * fopen_priv(const char *path, const char *mode) { - mode_t old_umask = umask(S_IRWXG | S_IRWXO); - FILE *fp; + mode_t old_umask = umask(S_IRWXG | S_IRWXO); + FILE *fp; fp = fopen(path, mode); umask(old_umask); return fp; } - - diff --git a/contrib/pg_upgrade/function.c b/contrib/pg_upgrade/function.c index b4b17badb28ce8abfa2f946c33572c01f98fa021..77bd3a0359771498d7b6aaabc6467ba37a7ef55c 100644 --- a/contrib/pg_upgrade/function.c +++ b/contrib/pg_upgrade/function.c @@ -133,7 +133,7 @@ get_loadable_libraries(void) int totaltups; int dbnum; bool found_public_plpython_handler = false; - + ress = (PGresult **) pg_malloc(old_cluster.dbarr.ndbs * sizeof(PGresult *)); totaltups = 0; @@ -144,10 +144,10 @@ get_loadable_libraries(void) PGconn *conn = connectToServer(&old_cluster, active_db->db_name); /* - * Fetch all libraries referenced in this DB. We can't exclude - * the "pg_catalog" schema because, while such functions are not - * explicitly dumped by pg_dump, they do reference implicit objects - * that pg_dump does dump, e.g. CREATE LANGUAGE plperl. + * Fetch all libraries referenced in this DB. We can't exclude the + * "pg_catalog" schema because, while such functions are not + * explicitly dumped by pg_dump, they do reference implicit objects + * that pg_dump does dump, e.g. CREATE LANGUAGE plperl. */ ress[dbnum] = executeQueryOrDie(conn, "SELECT DISTINCT probin " @@ -158,26 +158,26 @@ get_loadable_libraries(void) FirstNormalObjectId); totaltups += PQntuples(ress[dbnum]); - /* - * Systems that install plpython before 8.1 have - * plpython_call_handler() defined in the "public" schema, causing - * pg_dumpall to dump it. However that function still references - * "plpython" (no "2"), so it throws an error on restore. This code - * checks for the problem function, reports affected databases to the - * user and explains how to remove them. - * 8.1 git commit: e0dedd0559f005d60c69c9772163e69c204bac69 - * http://archives.postgresql.org/pgsql-hackers/2012-03/msg01101.php - * http://archives.postgresql.org/pgsql-bugs/2012-05/msg00206.php - */ + /* + * Systems that install plpython before 8.1 have + * plpython_call_handler() defined in the "public" schema, causing + * pg_dumpall to dump it. However that function still references + * "plpython" (no "2"), so it throws an error on restore. This code + * checks for the problem function, reports affected databases to the + * user and explains how to remove them. 8.1 git commit: + * e0dedd0559f005d60c69c9772163e69c204bac69 + * http://archives.postgresql.org/pgsql-hackers/2012-03/msg01101.php + * http://archives.postgresql.org/pgsql-bugs/2012-05/msg00206.php + */ if (GET_MAJOR_VERSION(old_cluster.major_version) < 901) { - PGresult *res; + PGresult *res; res = executeQueryOrDie(conn, "SELECT 1 " - "FROM pg_catalog.pg_proc JOIN pg_namespace " - " ON pronamespace = pg_namespace.oid " - "WHERE proname = 'plpython_call_handler' AND " + "FROM pg_catalog.pg_proc JOIN pg_namespace " + " ON pronamespace = pg_namespace.oid " + "WHERE proname = 'plpython_call_handler' AND " "nspname = 'public' AND " "prolang = 13 /* C */ AND " "probin = '$libdir/plpython' AND " @@ -188,23 +188,23 @@ get_loadable_libraries(void) if (!found_public_plpython_handler) { pg_log(PG_WARNING, - "\nThe old cluster has a \"plpython_call_handler\" function defined\n" - "in the \"public\" schema which is a duplicate of the one defined\n" - "in the \"pg_catalog\" schema. You can confirm this by executing\n" - "in psql:\n" - "\n" - " \\df *.plpython_call_handler\n" - "\n" - "The \"public\" schema version of this function was created by a\n" - "pre-8.1 install of plpython, and must be removed for pg_upgrade\n" - "to complete because it references a now-obsolete \"plpython\"\n" - "shared object file. You can remove the \"public\" schema version\n" - "of this function by running the following command:\n" - "\n" - " DROP FUNCTION public.plpython_call_handler()\n" - "\n" - "in each affected database:\n" - "\n"); + "\nThe old cluster has a \"plpython_call_handler\" function defined\n" + "in the \"public\" schema which is a duplicate of the one defined\n" + "in the \"pg_catalog\" schema. You can confirm this by executing\n" + "in psql:\n" + "\n" + " \\df *.plpython_call_handler\n" + "\n" + "The \"public\" schema version of this function was created by a\n" + "pre-8.1 install of plpython, and must be removed for pg_upgrade\n" + "to complete because it references a now-obsolete \"plpython\"\n" + "shared object file. You can remove the \"public\" schema version\n" + "of this function by running the following command:\n" + "\n" + " DROP FUNCTION public.plpython_call_handler()\n" + "\n" + "in each affected database:\n" + "\n"); } pg_log(PG_WARNING, " %s\n", active_db->db_name); found_public_plpython_handler = true; @@ -217,9 +217,9 @@ get_loadable_libraries(void) if (found_public_plpython_handler) pg_log(PG_FATAL, - "Remove the problem functions from the old cluster to continue.\n"); - - totaltups++; /* reserve for pg_upgrade_support */ + "Remove the problem functions from the old cluster to continue.\n"); + + totaltups++; /* reserve for pg_upgrade_support */ /* Allocate what's certainly enough space */ os_info.libraries = (char **) pg_malloc(totaltups * sizeof(char *)); @@ -293,17 +293,17 @@ check_loadable_libraries(void) PGresult *res; /* - * In Postgres 9.0, Python 3 support was added, and to do that, a - * plpython2u language was created with library name plpython2.so - * as a symbolic link to plpython.so. In Postgres 9.1, only the - * plpython2.so library was created, and both plpythonu and - * plpython2u pointing to it. For this reason, any reference to - * library name "plpython" in an old PG <= 9.1 cluster must look - * for "plpython2" in the new cluster. + * In Postgres 9.0, Python 3 support was added, and to do that, a + * plpython2u language was created with library name plpython2.so as a + * symbolic link to plpython.so. In Postgres 9.1, only the + * plpython2.so library was created, and both plpythonu and plpython2u + * pointing to it. For this reason, any reference to library name + * "plpython" in an old PG <= 9.1 cluster must look for "plpython2" in + * the new cluster. * - * For this case, we could check pg_pltemplate, but that only works - * for languages, and does not help with function shared objects, - * so we just do a general fix. + * For this case, we could check pg_pltemplate, but that only works + * for languages, and does not help with function shared objects, so + * we just do a general fix. */ if (GET_MAJOR_VERSION(old_cluster.major_version) < 901 && strcmp(lib, "$libdir/plpython") == 0) @@ -325,7 +325,7 @@ check_loadable_libraries(void) /* exit and report missing support library with special message */ if (strcmp(lib, PG_UPGRADE_SUPPORT) == 0) pg_log(PG_FATAL, - "The pg_upgrade_support module must be created and installed in the new cluster.\n"); + "The pg_upgrade_support module must be created and installed in the new cluster.\n"); if (script == NULL && (script = fopen_priv(output_path, "w")) == NULL) pg_log(PG_FATAL, "Could not open file \"%s\": %s\n", diff --git a/contrib/pg_upgrade/info.c b/contrib/pg_upgrade/info.c index 5b2b9eb28c18a126b15d9df65d2b17d1d39c7ab6..74b13e782da28e17788f7b44061269cb6acf17fc 100644 --- a/contrib/pg_upgrade/info.c +++ b/contrib/pg_upgrade/info.c @@ -57,12 +57,12 @@ gen_db_file_maps(DbInfo *old_db, DbInfo *new_db, old_db->db_name, old_rel->reloid, new_rel->reloid); /* - * TOAST table names initially match the heap pg_class oid. - * In pre-8.4, TOAST table names change during CLUSTER; in pre-9.0, - * TOAST table names change during ALTER TABLE ALTER COLUMN SET TYPE. - * In >= 9.0, TOAST relation names always use heap table oids, hence - * we cannot check relation names when upgrading from pre-9.0. - * Clusters upgraded to 9.0 will get matching TOAST names. + * TOAST table names initially match the heap pg_class oid. In + * pre-8.4, TOAST table names change during CLUSTER; in pre-9.0, TOAST + * table names change during ALTER TABLE ALTER COLUMN SET TYPE. In >= + * 9.0, TOAST relation names always use heap table oids, hence we + * cannot check relation names when upgrading from pre-9.0. Clusters + * upgraded to 9.0 will get matching TOAST names. */ if (strcmp(old_rel->nspname, new_rel->nspname) != 0 || ((GET_MAJOR_VERSION(old_cluster.major_version) >= 900 || @@ -194,16 +194,16 @@ get_db_infos(ClusterInfo *cluster) char query[QUERY_ALLOC]; snprintf(query, sizeof(query), - "SELECT d.oid, d.datname, %s " - "FROM pg_catalog.pg_database d " - " LEFT OUTER JOIN pg_catalog.pg_tablespace t " - " ON d.dattablespace = t.oid " - "WHERE d.datallowconn = true " + "SELECT d.oid, d.datname, %s " + "FROM pg_catalog.pg_database d " + " LEFT OUTER JOIN pg_catalog.pg_tablespace t " + " ON d.dattablespace = t.oid " + "WHERE d.datallowconn = true " /* we don't preserve pg_database.oid so we sort by name */ - "ORDER BY 2", + "ORDER BY 2", /* 9.2 removed the spclocation column */ - (GET_MAJOR_VERSION(cluster->major_version) <= 901) ? - "t.spclocation" : "pg_catalog.pg_tablespace_location(t.oid) AS spclocation"); + (GET_MAJOR_VERSION(cluster->major_version) <= 901) ? + "t.spclocation" : "pg_catalog.pg_tablespace_location(t.oid) AS spclocation"); res = executeQueryOrDie(conn, "%s", query); @@ -276,7 +276,7 @@ get_rel_infos(ClusterInfo *cluster, DbInfo *dbinfo) " LEFT OUTER JOIN pg_catalog.pg_tablespace t " " ON c.reltablespace = t.oid " "WHERE relkind IN ('r','t', 'i'%s) AND " - /* exclude possible orphaned temp tables */ + /* exclude possible orphaned temp tables */ " ((n.nspname !~ '^pg_temp_' AND " " n.nspname !~ '^pg_toast_temp_' AND " " n.nspname NOT IN ('pg_catalog', 'information_schema', 'binary_upgrade') AND " diff --git a/contrib/pg_upgrade/option.c b/contrib/pg_upgrade/option.c index 66a70cac8ac983243a9d8fd6eb391e6af90326eb..ccf00434d3b610238900fa372360a8dcdcbe70b4 100644 --- a/contrib/pg_upgrade/option.c +++ b/contrib/pg_upgrade/option.c @@ -56,10 +56,10 @@ parseCommandLine(int argc, char *argv[]) int option; /* Command line option */ int optindex = 0; /* used by getopt_long */ int os_user_effective_id; - FILE *fp; - char **filename; + FILE *fp; + char **filename; time_t run_time = time(NULL); - + user_opts.transfer_mode = TRANSFER_MODE_COPY; os_info.progname = get_progname(argv[0]); @@ -138,11 +138,11 @@ parseCommandLine(int argc, char *argv[]) new_cluster.pgopts = pg_strdup(optarg); break; - /* - * Someday, the port number option could be removed and - * passed using -o/-O, but that requires postmaster -C - * to be supported on all old/new versions. - */ + /* + * Someday, the port number option could be removed and passed + * using -o/-O, but that requires postmaster -C to be + * supported on all old/new versions. + */ case 'p': if ((old_cluster.port = atoi(optarg)) <= 0) { @@ -196,21 +196,21 @@ parseCommandLine(int argc, char *argv[]) /* Start with newline because we might be appending to a file. */ fprintf(fp, "\n" "-----------------------------------------------------------------\n" - " pg_upgrade run on %s" - "-----------------------------------------------------------------\n\n", - ctime(&run_time)); + " pg_upgrade run on %s" + "-----------------------------------------------------------------\n\n", + ctime(&run_time)); fclose(fp); } /* Get values from env if not already set */ check_required_directory(&old_cluster.bindir, "PGBINOLD", "-b", - "old cluster binaries reside"); + "old cluster binaries reside"); check_required_directory(&new_cluster.bindir, "PGBINNEW", "-B", - "new cluster binaries reside"); + "new cluster binaries reside"); check_required_directory(&old_cluster.pgdata, "PGDATAOLD", "-d", - "old cluster data resides"); + "old cluster data resides"); check_required_directory(&new_cluster.pgdata, "PGDATANEW", "-D", - "new cluster data resides"); + "new cluster data resides"); } @@ -285,7 +285,7 @@ or\n"), old_cluster.port, new_cluster.port, os_info.user); */ static void check_required_directory(char **dirpath, char *envVarName, - char *cmdLineOption, char *description) + char *cmdLineOption, char *description) { if (*dirpath == NULL || strlen(*dirpath) == 0) { @@ -322,8 +322,10 @@ void adjust_data_dir(ClusterInfo *cluster) { char filename[MAXPGPATH]; - char cmd[MAXPGPATH], cmd_output[MAX_STRING]; - FILE *fp, *output; + char cmd[MAXPGPATH], + cmd_output[MAX_STRING]; + FILE *fp, + *output; /* If there is no postgresql.conf, it can't be a config-only dir */ snprintf(filename, sizeof(filename), "%s/postgresql.conf", cluster->pgconfig); @@ -345,10 +347,9 @@ adjust_data_dir(ClusterInfo *cluster) CLUSTER_NAME(cluster)); /* - * We don't have a data directory yet, so we can't check the PG - * version, so this might fail --- only works for PG 9.2+. If this - * fails, pg_upgrade will fail anyway because the data files will not - * be found. + * We don't have a data directory yet, so we can't check the PG version, + * so this might fail --- only works for PG 9.2+. If this fails, + * pg_upgrade will fail anyway because the data files will not be found. */ snprintf(cmd, sizeof(cmd), "\"%s/postmaster\" -D \"%s\" -C data_directory", cluster->bindir, cluster->pgconfig); @@ -356,7 +357,7 @@ adjust_data_dir(ClusterInfo *cluster) if ((output = popen(cmd, "r")) == NULL || fgets(cmd_output, sizeof(cmd_output), output) == NULL) pg_log(PG_FATAL, "Could not get data directory using %s: %s\n", - cmd, getErrorText(errno)); + cmd, getErrorText(errno)); pclose(output); diff --git a/contrib/pg_upgrade/pg_upgrade.c b/contrib/pg_upgrade/pg_upgrade.c index 465ecdd6b324275293044646b9fcf49a6d180600..3537fc2bd05bbd9b9c7dfda6028826eb376485a9 100644 --- a/contrib/pg_upgrade/pg_upgrade.c +++ b/contrib/pg_upgrade/pg_upgrade.c @@ -55,7 +55,7 @@ ClusterInfo old_cluster, new_cluster; OSInfo os_info; -char *output_files[] = { +char *output_files[] = { SERVER_LOG_FILE, #ifdef WIN32 /* unique file for pg_ctl start */ @@ -122,11 +122,10 @@ main(int argc, char **argv) stop_postmaster(false); /* - * Most failures happen in create_new_objects(), which has - * completed at this point. We do this here because it is just - * before linking, which will link the old and new cluster data - * files, preventing the old cluster from being safely started - * once the new cluster is started. + * Most failures happen in create_new_objects(), which has completed at + * this point. We do this here because it is just before linking, which + * will link the old and new cluster data files, preventing the old + * cluster from being safely started once the new cluster is started. */ if (user_opts.transfer_mode == TRANSFER_MODE_LINK) disable_old_cluster(); @@ -215,8 +214,8 @@ prepare_new_cluster(void) exec_prog(true, true, UTILITY_LOG_FILE, SYSTEMQUOTE "\"%s/vacuumdb\" --port %d --username \"%s\" " "--all --analyze %s >> \"%s\" 2>&1" SYSTEMQUOTE, - new_cluster.bindir, new_cluster.port, os_info.user, - log_opts.verbose ? "--verbose" : "", UTILITY_LOG_FILE); + new_cluster.bindir, new_cluster.port, os_info.user, + log_opts.verbose ? "--verbose" : "", UTILITY_LOG_FILE); check_ok(); /* @@ -229,8 +228,8 @@ prepare_new_cluster(void) exec_prog(true, true, UTILITY_LOG_FILE, SYSTEMQUOTE "\"%s/vacuumdb\" --port %d --username \"%s\" " "--all --freeze %s >> \"%s\" 2>&1" SYSTEMQUOTE, - new_cluster.bindir, new_cluster.port, os_info.user, - log_opts.verbose ? "--verbose" : "", UTILITY_LOG_FILE); + new_cluster.bindir, new_cluster.port, os_info.user, + log_opts.verbose ? "--verbose" : "", UTILITY_LOG_FILE); check_ok(); get_pg_database_relfilenode(&new_cluster); @@ -252,8 +251,8 @@ prepare_new_databases(void) /* * Install support functions in the global-object restore database to - * preserve pg_authid.oid. pg_dumpall uses 'template0' as its template - * database so objects we add into 'template1' are not propogated. They + * preserve pg_authid.oid. pg_dumpall uses 'template0' as its template + * database so objects we add into 'template1' are not propogated. They * are removed on pg_upgrade exit. */ install_support_functions_in_new_db("template1"); @@ -267,7 +266,7 @@ prepare_new_databases(void) exec_prog(true, true, RESTORE_LOG_FILE, SYSTEMQUOTE "\"%s/psql\" --echo-queries " "--set ON_ERROR_STOP=on " - /* --no-psqlrc prevents AUTOCOMMIT=off */ + /* --no-psqlrc prevents AUTOCOMMIT=off */ "--no-psqlrc --port %d --username \"%s\" " "-f \"%s\" --dbname template1 >> \"%s\" 2>&1" SYSTEMQUOTE, new_cluster.bindir, new_cluster.port, os_info.user, @@ -453,13 +452,13 @@ set_frozenxids(void) static void cleanup(void) { - + fclose(log_opts.internal); /* Remove dump and log files? */ if (!log_opts.retain) { - char **filename; + char **filename; for (filename = output_files; *filename != NULL; filename++) unlink(*filename); diff --git a/contrib/pg_upgrade/pg_upgrade.h b/contrib/pg_upgrade/pg_upgrade.h index 26aa7bb1d24732f6a1528c662dd4de71dd566e1d..d12590ac6ba294d4d76551b6d5f7726fbcde17ef 100644 --- a/contrib/pg_upgrade/pg_upgrade.h +++ b/contrib/pg_upgrade/pg_upgrade.h @@ -75,7 +75,7 @@ extern char *output_files[]; #define RM_CMD "rm -f" #define RMDIR_CMD "rm -rf" #define SCRIPT_EXT "sh" -#define ECHO_QUOTE "'" +#define ECHO_QUOTE "'" #else #define pg_copy_file CopyFile #define pg_mv_file pgrename @@ -85,7 +85,7 @@ extern char *output_files[]; #define RMDIR_CMD "RMDIR /s/q" #define SCRIPT_EXT "bat" #define EXE_EXT ".exe" -#define ECHO_QUOTE "" +#define ECHO_QUOTE "" #endif #define CLUSTER_NAME(cluster) ((cluster) == &old_cluster ? "old" : \ @@ -98,7 +98,7 @@ extern char *output_files[]; /* postmaster/postgres -b (binary_upgrade) flag added during PG 9.1 development */ #define BINARY_UPGRADE_SERVER_FLAG_CAT_VER 201104251 /* - * Visibility map changed with this 9.2 commit, + * Visibility map changed with this 9.2 commit, * 8f9fe6edce358f7904e0db119416b4d1080a83aa; pick later catalog version. */ #define VISIBILITY_MAP_CRASHSAFE_CAT_VER 201107031 @@ -114,7 +114,7 @@ typedef struct Oid reloid; /* relation oid */ Oid relfilenode; /* relation relfile node */ /* relation tablespace path, or "" for the cluster default */ - char tablespace[MAXPGPATH]; + char tablespace[MAXPGPATH]; } RelInfo; typedef struct @@ -222,9 +222,11 @@ typedef struct ControlData controldata; /* pg_control information */ DbInfoArr dbarr; /* dbinfos array */ char *pgdata; /* pathname for cluster's $PGDATA directory */ - char *pgconfig; /* pathname for cluster's config file directory */ + char *pgconfig; /* pathname for cluster's config file + * directory */ char *bindir; /* pathname for cluster's executable directory */ - char *pgopts; /* options to pass to the server, like pg_ctl -o */ + char *pgopts; /* options to pass to the server, like pg_ctl + * -o */ unsigned short port; /* port number where postmaster is waiting */ uint32 major_version; /* PG_VERSION of cluster */ char major_version_str[64]; /* string PG_VERSION of cluster */ @@ -291,8 +293,8 @@ void check_old_cluster(bool live_check, void check_new_cluster(void); void report_clusters_compatible(void); void issue_warnings(char *sequence_script_file_name); -void output_completion_banner(char *analyze_script_file_name, - char *deletion_script_file_name); +void output_completion_banner(char *analyze_script_file_name, + char *deletion_script_file_name); void check_cluster_versions(void); void check_cluster_compatibility(bool live_check); void create_script_for_old_cluster_deletion(char **deletion_script_file_name); @@ -314,9 +316,10 @@ void split_old_dump(void); /* exec.c */ -int exec_prog(bool throw_error, bool is_priv, - const char *log_file, const char *cmd, ...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 4, 5))); +int +exec_prog(bool throw_error, bool is_priv, + const char *log_file, const char *cmd,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 4, 5))); void verify_directories(void); bool is_server_running(const char *datadir); @@ -353,14 +356,14 @@ const char *setupPageConverter(pageCnvCtx **result); typedef void *pageCnvCtx; #endif -int load_directory(const char *dirname, struct dirent ***namelist); +int load_directory(const char *dirname, struct dirent *** namelist); const char *copyAndUpdateFile(pageCnvCtx *pageConverter, const char *src, const char *dst, bool force); const char *linkAndUpdateFile(pageCnvCtx *pageConverter, const char *src, const char *dst); void check_hard_link(void); -FILE *fopen_priv(const char *path, const char *mode); +FILE *fopen_priv(const char *path, const char *mode); /* function.c */ @@ -399,8 +402,9 @@ void init_tablespaces(void); /* server.c */ PGconn *connectToServer(ClusterInfo *cluster, const char *db_name); -PGresult *executeQueryOrDie(PGconn *conn, const char *fmt, ...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); +PGresult * +executeQueryOrDie(PGconn *conn, const char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); void start_postmaster(ClusterInfo *cluster); void stop_postmaster(bool fast); @@ -413,12 +417,15 @@ void check_pghost_envvar(void); char *quote_identifier(const char *s); int get_user_info(char **user_name); void check_ok(void); -void report_status(eLogType type, const char *fmt, ...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); -void pg_log(eLogType type, char *fmt, ...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); -void prep_status(const char *fmt, ...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); +void +report_status(eLogType type, const char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); +void +pg_log(eLogType type, char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); +void +prep_status(const char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); void check_ok(void); char *pg_strdup(const char *s); void *pg_malloc(int size); diff --git a/contrib/pg_upgrade/relfilenode.c b/contrib/pg_upgrade/relfilenode.c index 45d6c5415bf15a92a16f9ae9d3938408a7c40b47..3509585de7b568af921d09443d077d7ed3b3bf47 100644 --- a/contrib/pg_upgrade/relfilenode.c +++ b/contrib/pg_upgrade/relfilenode.c @@ -34,26 +34,28 @@ const char * transfer_all_new_dbs(DbInfoArr *old_db_arr, DbInfoArr *new_db_arr, char *old_pgdata, char *new_pgdata) { - int old_dbnum, new_dbnum; + int old_dbnum, + new_dbnum; const char *msg = NULL; prep_status("%s user relation files\n", - user_opts.transfer_mode == TRANSFER_MODE_LINK ? "Linking" : "Copying"); + user_opts.transfer_mode == TRANSFER_MODE_LINK ? "Linking" : "Copying"); /* Scan the old cluster databases and transfer their files */ for (old_dbnum = new_dbnum = 0; old_dbnum < old_db_arr->ndbs; old_dbnum++, new_dbnum++) { - DbInfo *old_db = &old_db_arr->dbs[old_dbnum], *new_db = NULL; + DbInfo *old_db = &old_db_arr->dbs[old_dbnum], + *new_db = NULL; FileNameMap *mappings; int n_maps; pageCnvCtx *pageConverter = NULL; /* - * Advance past any databases that exist in the new cluster - * but not in the old, e.g. "postgres". (The user might - * have removed the 'postgres' database from the old cluster.) + * Advance past any databases that exist in the new cluster but not in + * the old, e.g. "postgres". (The user might have removed the + * 'postgres' database from the old cluster.) */ for (; new_dbnum < new_db_arr->ndbs; new_dbnum++) { @@ -83,8 +85,8 @@ transfer_all_new_dbs(DbInfoArr *old_db_arr, } } - prep_status(" "); /* in case nothing printed; pass a space so gcc - * doesn't complain about empty format + prep_status(" "); /* in case nothing printed; pass a space so + * gcc doesn't complain about empty format * string */ check_ok(); @@ -137,14 +139,14 @@ transfer_single_new_db(pageCnvCtx *pageConverter, int mapnum; int fileno; bool vm_crashsafe_change = false; - + old_dir[0] = '\0'; /* Do not copy non-crashsafe vm files for binaries that assume crashsafety */ if (old_cluster.controldata.cat_ver < VISIBILITY_MAP_CRASHSAFE_CAT_VER && new_cluster.controldata.cat_ver >= VISIBILITY_MAP_CRASHSAFE_CAT_VER) vm_crashsafe_change = true; - + for (mapnum = 0; mapnum < size; mapnum++) { char old_file[MAXPGPATH]; @@ -190,8 +192,8 @@ transfer_single_new_db(pageCnvCtx *pageConverter, for (fileno = 0; fileno < numFiles; fileno++) { - char *vm_offset = strstr(namelist[fileno]->d_name, "_vm"); - bool is_vm_file = false; + char *vm_offset = strstr(namelist[fileno]->d_name, "_vm"); + bool is_vm_file = false; /* Is a visibility map file? (name ends with _vm) */ if (vm_offset && strlen(vm_offset) == strlen("_vm")) diff --git a/contrib/pg_upgrade/server.c b/contrib/pg_upgrade/server.c index f557453df2d92e0ff946588eacaa1fae5962ca1e..f83d6fa866cc87e142a11a2289ff938c8c2f1ab6 100644 --- a/contrib/pg_upgrade/server.c +++ b/contrib/pg_upgrade/server.c @@ -161,7 +161,7 @@ start_postmaster(ClusterInfo *cluster) snprintf(cmd, sizeof(cmd), SYSTEMQUOTE "\"%s/pg_ctl\" -w -l \"%s\" -D \"%s\" " "-o \"-p %d %s %s\" start >> \"%s\" 2>&1" SYSTEMQUOTE, - cluster->bindir, SERVER_LOG_FILE, cluster->pgconfig, cluster->port, + cluster->bindir, SERVER_LOG_FILE, cluster->pgconfig, cluster->port, (cluster->controldata.cat_ver >= BINARY_UPGRADE_SERVER_FLAG_CAT_VER) ? "-b" : "-c autovacuum=off -c autovacuum_freeze_max_age=2000000000", @@ -172,11 +172,11 @@ start_postmaster(ClusterInfo *cluster) * it might supply a reason for the failure. */ pg_ctl_return = exec_prog(false, true, - /* pass both file names if the differ */ - (strcmp(SERVER_LOG_FILE, SERVER_START_LOG_FILE) == 0) ? - SERVER_LOG_FILE : - SERVER_LOG_FILE " or " SERVER_START_LOG_FILE, - "%s", cmd); + /* pass both file names if the differ */ + (strcmp(SERVER_LOG_FILE, SERVER_START_LOG_FILE) == 0) ? + SERVER_LOG_FILE : + SERVER_LOG_FILE " or " SERVER_START_LOG_FILE, + "%s", cmd); /* Check to see if we can connect to the server; if not, report it. */ if ((conn = get_db_conn(cluster, "template1")) == NULL || @@ -211,14 +211,14 @@ stop_postmaster(bool fast) else if (os_info.running_cluster == &new_cluster) cluster = &new_cluster; else - return; /* no cluster running */ + return; /* no cluster running */ snprintf(cmd, sizeof(cmd), SYSTEMQUOTE "\"%s/pg_ctl\" -w -D \"%s\" -o \"%s\" " "%s stop >> \"%s\" 2>&1" SYSTEMQUOTE, cluster->bindir, cluster->pgconfig, cluster->pgopts ? cluster->pgopts : "", - fast ? "-m fast" : "", SERVER_STOP_LOG_FILE); + fast ? "-m fast" : "", SERVER_STOP_LOG_FILE); exec_prog(fast ? false : true, true, SERVER_STOP_LOG_FILE, "%s", cmd); diff --git a/contrib/pg_upgrade/tablespace.c b/contrib/pg_upgrade/tablespace.c index 6b61f4bac11cc320b106ea65adc7152016be4820..b783b6251e6e248f01953fcd0ade964044505363 100644 --- a/contrib/pg_upgrade/tablespace.c +++ b/contrib/pg_upgrade/tablespace.c @@ -52,8 +52,8 @@ get_tablespace_paths(void) "WHERE spcname != 'pg_default' AND " " spcname != 'pg_global'", /* 9.2 removed the spclocation column */ - (GET_MAJOR_VERSION(old_cluster.major_version) <= 901) ? - "spclocation" : "pg_catalog.pg_tablespace_location(oid) AS spclocation"); + (GET_MAJOR_VERSION(old_cluster.major_version) <= 901) ? + "spclocation" : "pg_catalog.pg_tablespace_location(oid) AS spclocation"); res = executeQueryOrDie(conn, "%s", query); diff --git a/contrib/pg_upgrade/version_old_8_3.c b/contrib/pg_upgrade/version_old_8_3.c index 542425c7c9eff3d761f7839904fcfab7ac90478f..b681c0984e6e9d0bee6c0a561907d6efbbc3a1a4 100644 --- a/contrib/pg_upgrade/version_old_8_3.c +++ b/contrib/pg_upgrade/version_old_8_3.c @@ -60,10 +60,10 @@ old_8_3_check_for_name_data_type_usage(ClusterInfo *cluster) " NOT a.attisdropped AND " " a.atttypid = 'pg_catalog.name'::pg_catalog.regtype AND " " c.relnamespace = n.oid AND " - /* exclude possible orphaned temp tables */ + /* exclude possible orphaned temp tables */ " n.nspname !~ '^pg_temp_' AND " - " n.nspname !~ '^pg_toast_temp_' AND " - " n.nspname NOT IN ('pg_catalog', 'information_schema')"); + " n.nspname !~ '^pg_toast_temp_' AND " + " n.nspname NOT IN ('pg_catalog', 'information_schema')"); ntups = PQntuples(res); i_nspname = PQfnumber(res, "nspname"); @@ -98,9 +98,9 @@ old_8_3_check_for_name_data_type_usage(ClusterInfo *cluster) pg_log(PG_REPORT, "fatal\n"); pg_log(PG_FATAL, "Your installation contains the \"name\" data type in user tables. This\n" - "data type changed its internal alignment between your old and new\n" + "data type changed its internal alignment between your old and new\n" "clusters so this cluster cannot currently be upgraded. You can remove\n" - "the problem tables and restart the upgrade. A list of the problem\n" + "the problem tables and restart the upgrade. A list of the problem\n" "columns is in the file:\n" " %s\n\n", output_path); } @@ -150,10 +150,10 @@ old_8_3_check_for_tsquery_usage(ClusterInfo *cluster) " NOT a.attisdropped AND " " a.atttypid = 'pg_catalog.tsquery'::pg_catalog.regtype AND " " c.relnamespace = n.oid AND " - /* exclude possible orphaned temp tables */ + /* exclude possible orphaned temp tables */ " n.nspname !~ '^pg_temp_' AND " - " n.nspname !~ '^pg_toast_temp_' AND " - " n.nspname NOT IN ('pg_catalog', 'information_schema')"); + " n.nspname !~ '^pg_toast_temp_' AND " + " n.nspname NOT IN ('pg_catalog', 'information_schema')"); ntups = PQntuples(res); i_nspname = PQfnumber(res, "nspname"); @@ -189,7 +189,7 @@ old_8_3_check_for_tsquery_usage(ClusterInfo *cluster) pg_log(PG_FATAL, "Your installation contains the \"tsquery\" data type. This data type\n" "added a new internal field between your old and new clusters so this\n" - "cluster cannot currently be upgraded. You can remove the problem\n" + "cluster cannot currently be upgraded. You can remove the problem\n" "columns and restart the upgrade. A list of the problem columns is in the\n" "file:\n" " %s\n\n", output_path); @@ -328,10 +328,10 @@ old_8_3_rebuild_tsvector_tables(ClusterInfo *cluster, bool check_mode) " NOT a.attisdropped AND " " a.atttypid = 'pg_catalog.tsvector'::pg_catalog.regtype AND " " c.relnamespace = n.oid AND " - /* exclude possible orphaned temp tables */ + /* exclude possible orphaned temp tables */ " n.nspname !~ '^pg_temp_' AND " - " n.nspname !~ '^pg_toast_temp_' AND " - " n.nspname NOT IN ('pg_catalog', 'information_schema')"); + " n.nspname !~ '^pg_toast_temp_' AND " + " n.nspname NOT IN ('pg_catalog', 'information_schema')"); /* * This macro is used below to avoid reindexing indexes already rebuilt @@ -527,7 +527,7 @@ old_8_3_invalidate_hash_gin_indexes(ClusterInfo *cluster, bool check_mode) "must be reindexed with the REINDEX command. The file:\n" " %s\n" "when executed by psql by the database superuser will recreate all invalid\n" - "indexes; until then, none of these indexes will be used.\n\n", + "indexes; until then, none of these indexes will be used.\n\n", output_path); } else @@ -648,10 +648,10 @@ old_8_3_invalidate_bpchar_pattern_ops_indexes(ClusterInfo *cluster, pg_log(PG_WARNING, "\n" "Your installation contains indexes using \"bpchar_pattern_ops\". These\n" "indexes have different internal formats between your old and new clusters\n" - "so they must be reindexed with the REINDEX command. The file:\n" + "so they must be reindexed with the REINDEX command. The file:\n" " %s\n" "when executed by psql by the database superuser will recreate all invalid\n" - "indexes; until then, none of these indexes will be used.\n\n", + "indexes; until then, none of these indexes will be used.\n\n", output_path); } else @@ -699,10 +699,10 @@ old_8_3_create_sequence_script(ClusterInfo *cluster) " pg_catalog.pg_namespace n " "WHERE c.relkind = 'S' AND " " c.relnamespace = n.oid AND " - /* exclude possible orphaned temp tables */ + /* exclude possible orphaned temp tables */ " n.nspname !~ '^pg_temp_' AND " - " n.nspname !~ '^pg_toast_temp_' AND " - " n.nspname NOT IN ('pg_catalog', 'information_schema')"); + " n.nspname !~ '^pg_toast_temp_' AND " + " n.nspname NOT IN ('pg_catalog', 'information_schema')"); ntups = PQntuples(res); i_nspname = PQfnumber(res, "nspname"); diff --git a/contrib/pgbench/pgbench.c b/contrib/pgbench/pgbench.c index b0e699187b5c99fd331f3d34348b663173718ab4..25fb15a847155db6f852ae82343bc98b1d5d0b64 100644 --- a/contrib/pgbench/pgbench.c +++ b/contrib/pgbench/pgbench.c @@ -66,7 +66,7 @@ typedef struct win32_pthread *pthread_t; typedef int pthread_attr_t; -static int pthread_create(pthread_t *thread, pthread_attr_t * attr, void *(*start_routine) (void *), void *arg); +static int pthread_create(pthread_t *thread, pthread_attr_t *attr, void *(*start_routine) (void *), void *arg); static int pthread_join(pthread_t th, void **thread_return); #elif defined(ENABLE_THREAD_SAFETY) /* Use platform-dependent pthread capability */ @@ -84,7 +84,7 @@ static int pthread_join(pthread_t th, void **thread_return); typedef struct fork_pthread *pthread_t; typedef int pthread_attr_t; -static int pthread_create(pthread_t *thread, pthread_attr_t * attr, void *(*start_routine) (void *), void *arg); +static int pthread_create(pthread_t *thread, pthread_attr_t *attr, void *(*start_routine) (void *), void *arg); static int pthread_join(pthread_t th, void **thread_return); #endif @@ -198,7 +198,7 @@ typedef struct instr_time start_time; /* thread start time */ instr_time *exec_elapsed; /* time spent executing cmds (per Command) */ int *exec_count; /* number of cmd executions (per Command) */ - unsigned short random_state[3]; /* separate randomness for each thread */ + unsigned short random_state[3]; /* separate randomness for each thread */ } TState; #define INVALID_THREAD ((pthread_t) 0) @@ -1075,7 +1075,7 @@ top: /* * getrand() neeeds to be able to subtract max from min and add - * one the result without overflowing. Since we know max > min, + * one the result without overflowing. Since we know max > min, * we can detect overflow just by checking for a negative result. * But we must check both that the subtraction doesn't overflow, * and that adding one to the result doesn't overflow either. @@ -1267,10 +1267,11 @@ init(void) * versions. Since pgbench has never pretended to be fully TPC-B * compliant anyway, we stick with the historical behavior. */ - struct ddlinfo { - char *table; - char *cols; - int declare_fillfactor; + struct ddlinfo + { + char *table; + char *cols; + int declare_fillfactor; }; struct ddlinfo DDLs[] = { { @@ -1321,15 +1322,16 @@ init(void) /* Construct new create table statement. */ opts[0] = '\0'; if (ddl->declare_fillfactor) - snprintf(opts+strlen(opts), 256-strlen(opts), - " with (fillfactor=%d)", fillfactor); + snprintf(opts + strlen(opts), 256 - strlen(opts), + " with (fillfactor=%d)", fillfactor); if (tablespace != NULL) { - char *escape_tablespace; + char *escape_tablespace; + escape_tablespace = PQescapeIdentifier(con, tablespace, strlen(tablespace)); - snprintf(opts+strlen(opts), 256-strlen(opts), - " tablespace %s", escape_tablespace); + snprintf(opts + strlen(opts), 256 - strlen(opts), + " tablespace %s", escape_tablespace); PQfreemem(escape_tablespace); } snprintf(buffer, 256, "create%s table %s(%s)%s", @@ -1404,17 +1406,18 @@ init(void) fprintf(stderr, "set primary key...\n"); for (i = 0; i < lengthof(DDLAFTERs); i++) { - char buffer[256]; + char buffer[256]; strncpy(buffer, DDLAFTERs[i], 256); if (index_tablespace != NULL) { - char *escape_tablespace; + char *escape_tablespace; + escape_tablespace = PQescapeIdentifier(con, index_tablespace, strlen(index_tablespace)); - snprintf(buffer+strlen(buffer), 256-strlen(buffer), - " using index tablespace %s", escape_tablespace); + snprintf(buffer + strlen(buffer), 256 - strlen(buffer), + " using index tablespace %s", escape_tablespace); PQfreemem(escape_tablespace); } @@ -1861,10 +1864,10 @@ main(int argc, char **argv) int i; static struct option long_options[] = { - {"index-tablespace", required_argument, NULL, 3}, - {"tablespace", required_argument, NULL, 2}, - {"unlogged-tables", no_argument, &unlogged_tables, 1}, - {NULL, 0, NULL, 0} + {"index-tablespace", required_argument, NULL, 3}, + {"tablespace", required_argument, NULL, 2}, + {"unlogged-tables", no_argument, &unlogged_tables, 1}, + {NULL, 0, NULL, 0} }; #ifdef HAVE_GETRLIMIT @@ -2065,10 +2068,10 @@ main(int argc, char **argv) case 0: /* This covers long options which take no argument. */ break; - case 2: /* tablespace */ + case 2: /* tablespace */ tablespace = optarg; break; - case 3: /* index-tablespace */ + case 3: /* index-tablespace */ index_tablespace = optarg; break; default: @@ -2571,7 +2574,7 @@ typedef struct fork_pthread static int pthread_create(pthread_t *thread, - pthread_attr_t * attr, + pthread_attr_t *attr, void *(*start_routine) (void *), void *arg) { @@ -2687,7 +2690,7 @@ win32_pthread_run(void *arg) static int pthread_create(pthread_t *thread, - pthread_attr_t * attr, + pthread_attr_t *attr, void *(*start_routine) (void *), void *arg) { diff --git a/contrib/pgcrypto/crypt-md5.c b/contrib/pgcrypto/crypt-md5.c index 6c7a2b329e487cd9d5c164f6ca5302ad8c558524..2a5cd70208a90a8cbd047a77aca7e30ed6bd5ea2 100644 --- a/contrib/pgcrypto/crypt-md5.c +++ b/contrib/pgcrypto/crypt-md5.c @@ -34,8 +34,8 @@ char * px_crypt_md5(const char *pw, const char *salt, char *passwd, unsigned dstlen) { static char *magic = "$1$"; /* This string is magic for this algorithm. - * Having it this way, we can get better - * later on */ + * Having it this way, we can get better later + * on */ static char *p; static const char *sp, *ep; diff --git a/contrib/pgcrypto/px.h b/contrib/pgcrypto/px.h index 610b7fad789a83d655c38e21431d5c1d49c57e3c..80e8624460ddea1549fd868645d3a221137f2990 100644 --- a/contrib/pgcrypto/px.h +++ b/contrib/pgcrypto/px.h @@ -204,8 +204,9 @@ const char *px_resolve_alias(const PX_Alias *aliases, const char *name); void px_set_debug_handler(void (*handler) (const char *)); #ifdef PX_DEBUG -void px_debug(const char *fmt, ...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); +void +px_debug(const char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); #else #define px_debug(...) #endif diff --git a/contrib/pgstattuple/pgstatindex.c b/contrib/pgstattuple/pgstatindex.c index 9f2ec1f210840a45852a8b3dbd7d9892c3b1c4c8..d4fc8a0fd609f359c1cfba928f140519c99a17da 100644 --- a/contrib/pgstattuple/pgstatindex.c +++ b/contrib/pgstattuple/pgstatindex.c @@ -95,7 +95,7 @@ pgstatindex(PG_FUNCTION_ARGS) BlockNumber nblocks; BlockNumber blkno; BTIndexStat indexStat; - BufferAccessStrategy bstrategy = GetAccessStrategy(BAS_BULKREAD); + BufferAccessStrategy bstrategy = GetAccessStrategy(BAS_BULKREAD); if (!superuser()) ereport(ERROR, @@ -160,7 +160,7 @@ pgstatindex(PG_FUNCTION_ARGS) CHECK_FOR_INTERRUPTS(); /* Read and lock buffer */ - buffer = ReadBufferExtended(rel, MAIN_FORKNUM, blkno, RBM_NORMAL, bstrategy); + buffer = ReadBufferExtended(rel, MAIN_FORKNUM, blkno, RBM_NORMAL, bstrategy); LockBuffer(buffer, BUFFER_LOCK_SHARE); page = BufferGetPage(buffer); diff --git a/contrib/pgstattuple/pgstattuple.c b/contrib/pgstattuple/pgstattuple.c index c9be8c92e4bf2aecdd53e2e60683cb9a8cd5b524..2b62b7850639d4bbaff7ddb4473ea0d671ba06e5 100644 --- a/contrib/pgstattuple/pgstattuple.c +++ b/contrib/pgstattuple/pgstattuple.c @@ -62,7 +62,7 @@ typedef struct pgstattuple_type } pgstattuple_type; typedef void (*pgstat_page) (pgstattuple_type *, Relation, BlockNumber, - BufferAccessStrategy); + BufferAccessStrategy); static Datum build_pgstattuple_type(pgstattuple_type *stat, FunctionCallInfo fcinfo); diff --git a/contrib/sepgsql/database.c b/contrib/sepgsql/database.c index 0c395c42a3cb0643c6d5f99808c84a625468e316..5a4246752a3d358a0483f5f1f1f5615bad221ce7 100644 --- a/contrib/sepgsql/database.c +++ b/contrib/sepgsql/database.c @@ -32,19 +32,19 @@ void sepgsql_database_post_create(Oid databaseId, const char *dtemplate) { Relation rel; - ScanKeyData skey; - SysScanDesc sscan; + ScanKeyData skey; + SysScanDesc sscan; HeapTuple tuple; char *tcontext; char *ncontext; char audit_name[NAMEDATALEN + 20]; - ObjectAddress object; - Form_pg_database datForm; + ObjectAddress object; + Form_pg_database datForm; /* - * Oid of the source database is not saved in pg_database catalog, - * so we collect its identifier using contextual information. - * If NULL, its default is "template1" according to createdb(). + * Oid of the source database is not saved in pg_database catalog, so we + * collect its identifier using contextual information. If NULL, its + * default is "template1" according to createdb(). */ if (!dtemplate) dtemplate = "template1"; @@ -56,6 +56,7 @@ sepgsql_database_post_create(Oid databaseId, const char *dtemplate) tcontext = sepgsql_get_label(object.classId, object.objectId, object.objectSubId); + /* * check db_database:{getattr} permission */ @@ -67,11 +68,11 @@ sepgsql_database_post_create(Oid databaseId, const char *dtemplate) true); /* - * Compute a default security label of the newly created database - * based on a pair of security label of client and source database. + * Compute a default security label of the newly created database based on + * a pair of security label of client and source database. * - * XXX - uncoming version of libselinux supports to take object - * name to handle special treatment on default security label. + * XXX - uncoming version of libselinux supports to take object name to + * handle special treatment on default security label. */ rel = heap_open(DatabaseRelationId, AccessShareLock); @@ -91,6 +92,7 @@ sepgsql_database_post_create(Oid databaseId, const char *dtemplate) ncontext = sepgsql_compute_create(sepgsql_get_client_label(), tcontext, SEPG_CLASS_DB_DATABASE); + /* * check db_database:{create} permission */ @@ -126,8 +128,8 @@ sepgsql_database_post_create(Oid databaseId, const char *dtemplate) void sepgsql_database_drop(Oid databaseId) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; /* * check db_database:{drop} permission @@ -153,8 +155,8 @@ sepgsql_database_drop(Oid databaseId) void sepgsql_database_relabel(Oid databaseId, const char *seclabel) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; object.classId = DatabaseRelationId; object.objectId = databaseId; @@ -170,6 +172,7 @@ sepgsql_database_relabel(Oid databaseId, const char *seclabel) SEPG_DB_DATABASE__RELABELFROM, audit_name, true); + /* * check db_database:{relabelto} permission */ diff --git a/contrib/sepgsql/dml.c b/contrib/sepgsql/dml.c index 17aa41cf4e7eea43ccf033aa49d21a00d9fc153f..47a108741746f99f9a542051f13f68812a43e585 100644 --- a/contrib/sepgsql/dml.c +++ b/contrib/sepgsql/dml.c @@ -150,7 +150,7 @@ check_relation_privileges(Oid relOid, uint32 required, bool abort) { - ObjectAddress object; + ObjectAddress object; char *audit_name; Bitmapset *columns; int index; diff --git a/contrib/sepgsql/hooks.c b/contrib/sepgsql/hooks.c index ffa078677c8490ab0973991726be6afc79c288ba..914519109c84bb1053bb3eef9fde06d92a79904d 100644 --- a/contrib/sepgsql/hooks.c +++ b/contrib/sepgsql/hooks.c @@ -52,9 +52,9 @@ typedef struct * command. Elsewhere (including the case of default) NULL. */ const char *createdb_dtemplate; -} sepgsql_context_info_t; +} sepgsql_context_info_t; -static sepgsql_context_info_t sepgsql_context_info; +static sepgsql_context_info_t sepgsql_context_info; /* * GUC: sepgsql.permissive = (on|off) @@ -101,7 +101,7 @@ sepgsql_object_access(ObjectAccessType access, { case DatabaseRelationId: sepgsql_database_post_create(objectId, - sepgsql_context_info.createdb_dtemplate); + sepgsql_context_info.createdb_dtemplate); break; case NamespaceRelationId: @@ -115,9 +115,8 @@ sepgsql_object_access(ObjectAccessType access, * All cases we want to apply permission checks on * creation of a new relation are invocation of the * heap_create_with_catalog via DefineRelation or - * OpenIntoRel. - * Elsewhere, we need neither assignment of security - * label nor permission checks. + * OpenIntoRel. Elsewhere, we need neither assignment + * of security label nor permission checks. */ switch (sepgsql_context_info.cmdtype) { @@ -150,12 +149,12 @@ sepgsql_object_access(ObjectAccessType access, case OAT_DROP: { - ObjectAccessDrop *drop_arg = (ObjectAccessDrop *)arg; + ObjectAccessDrop *drop_arg = (ObjectAccessDrop *) arg; /* - * No need to apply permission checks on object deletion - * due to internal cleanups; such as removal of temporary - * database object on session closed. + * No need to apply permission checks on object deletion due + * to internal cleanups; such as removal of temporary database + * object on session closed. */ if ((drop_arg->dropflags & PERFORM_DELETION_INTERNAL) != 0) break; @@ -219,13 +218,13 @@ sepgsql_exec_check_perms(List *rangeTabls, bool abort) /* * sepgsql_executor_start * - * It saves contextual information during ExecutorStart to distinguish + * It saves contextual information during ExecutorStart to distinguish * a case with/without permission checks later. */ static void sepgsql_executor_start(QueryDesc *queryDesc, int eflags) { - sepgsql_context_info_t saved_context_info = sepgsql_context_info; + sepgsql_context_info_t saved_context_info = sepgsql_context_info; PG_TRY(); { @@ -270,28 +269,29 @@ sepgsql_utility_command(Node *parsetree, DestReceiver *dest, char *completionTag) { - sepgsql_context_info_t saved_context_info = sepgsql_context_info; - ListCell *cell; + sepgsql_context_info_t saved_context_info = sepgsql_context_info; + ListCell *cell; PG_TRY(); { /* * Check command tag to avoid nefarious operations, and save the - * current contextual information to determine whether we should - * apply permission checks here, or not. + * current contextual information to determine whether we should apply + * permission checks here, or not. */ sepgsql_context_info.cmdtype = nodeTag(parsetree); switch (nodeTag(parsetree)) { case T_CreatedbStmt: + /* * We hope to reference name of the source database, but it * does not appear in system catalog. So, we save it here. */ - foreach (cell, ((CreatedbStmt *) parsetree)->options) + foreach(cell, ((CreatedbStmt *) parsetree)->options) { - DefElem *defel = (DefElem *) lfirst(cell); + DefElem *defel = (DefElem *) lfirst(cell); if (strcmp(defel->defname, "template") == 0) { @@ -303,6 +303,7 @@ sepgsql_utility_command(Node *parsetree, break; case T_LoadStmt: + /* * We reject LOAD command across the board on enforcing mode, * because a binary module can arbitrarily override hooks. @@ -315,6 +316,7 @@ sepgsql_utility_command(Node *parsetree, } break; default: + /* * Right now we don't check any other utility commands, * because it needs more detailed information to make access diff --git a/contrib/sepgsql/label.c b/contrib/sepgsql/label.c index 85f4efe072ad552e6f93e0b3da6139bc6cf3502d..23577b5844db3d37e3a6aaddbaf39cd6b980b140 100644 --- a/contrib/sepgsql/label.c +++ b/contrib/sepgsql/label.c @@ -58,17 +58,18 @@ static fmgr_hook_type next_fmgr_hook = NULL; * we use the list client_label_pending of pending_label to keep track of which * labels were set during the (sub-)transactions. */ -static char *client_label_peer = NULL; /* set by getpeercon(3) */ -static List *client_label_pending = NIL; /* pending list being set by - * sepgsql_setcon() */ -static char *client_label_committed = NULL; /* set by sepgsql_setcon(), - * and already committed */ -static char *client_label_func = NULL; /* set by trusted procedure */ - -typedef struct { - SubTransactionId subid; - char *label; -} pending_label; +static char *client_label_peer = NULL; /* set by getpeercon(3) */ +static List *client_label_pending = NIL; /* pending list being set by + * sepgsql_setcon() */ +static char *client_label_committed = NULL; /* set by sepgsql_setcon(), + * and already committed */ +static char *client_label_func = NULL; /* set by trusted procedure */ + +typedef struct +{ + SubTransactionId subid; + char *label; +} pending_label; /* * sepgsql_get_client_label @@ -87,7 +88,7 @@ sepgsql_get_client_label(void) /* uncommitted sepgsql_setcon() value */ if (client_label_pending) { - pending_label *plabel = llast(client_label_pending); + pending_label *plabel = llast(client_label_pending); if (plabel->label) return plabel->label; @@ -104,16 +105,16 @@ sepgsql_get_client_label(void) * sepgsql_set_client_label * * This routine tries to switch the current security label of the client, and - * checks related permissions. The supplied new label shall be added to the + * checks related permissions. The supplied new label shall be added to the * client_label_pending list, then saved at transaction-commit time to ensure * transaction-awareness. */ static void sepgsql_set_client_label(const char *new_label) { - const char *tcontext; - MemoryContext oldcxt; - pending_label *plabel; + const char *tcontext; + MemoryContext oldcxt; + pending_label *plabel; /* Reset to the initial client label, if NULL */ if (!new_label) @@ -140,9 +141,10 @@ sepgsql_set_client_label(const char *new_label) SEPG_PROCESS__DYNTRANSITION, NULL, true); + /* - * Append the supplied new_label on the pending list until - * the current transaction is committed. + * Append the supplied new_label on the pending list until the current + * transaction is committed. */ oldcxt = MemoryContextSwitchTo(CurTransactionContext); @@ -158,7 +160,7 @@ sepgsql_set_client_label(const char *new_label) /* * sepgsql_xact_callback * - * A callback routine of transaction commit/abort/prepare. Commmit or abort + * A callback routine of transaction commit/abort/prepare. Commmit or abort * changes in the client_label_pending list. */ static void @@ -168,8 +170,8 @@ sepgsql_xact_callback(XactEvent event, void *arg) { if (client_label_pending != NIL) { - pending_label *plabel = llast(client_label_pending); - char *new_label; + pending_label *plabel = llast(client_label_pending); + char *new_label; if (plabel->label) new_label = MemoryContextStrdup(TopMemoryContext, @@ -181,10 +183,11 @@ sepgsql_xact_callback(XactEvent event, void *arg) pfree(client_label_committed); client_label_committed = new_label; + /* - * XXX - Note that items of client_label_pending are allocated - * on CurTransactionContext, thus, all acquired memory region - * shall be released implicitly. + * XXX - Note that items of client_label_pending are allocated on + * CurTransactionContext, thus, all acquired memory region shall + * be released implicitly. */ client_label_pending = NIL; } @@ -212,7 +215,8 @@ sepgsql_subxact_callback(SubXactEvent event, SubTransactionId mySubid, prev = NULL; for (cell = list_head(client_label_pending); cell; cell = next) { - pending_label *plabel = lfirst(cell); + pending_label *plabel = lfirst(cell); + next = lnext(cell); if (plabel->subid == mySubid) @@ -272,7 +276,7 @@ sepgsql_client_auth(Port *port, int status) static bool sepgsql_needs_fmgr_hook(Oid functionId) { - ObjectAddress object; + ObjectAddress object; if (next_needs_fmgr_hook && (*next_needs_fmgr_hook) (functionId)) @@ -340,8 +344,8 @@ sepgsql_fmgr_hook(FmgrHookEventType event, /* * process:transition permission between old and new label, - * when user tries to switch security label of the client - * on execution of trusted procedure. + * when user tries to switch security label of the client on + * execution of trusted procedure. */ if (stack->new_label) sepgsql_avc_check_perms_label(stack->new_label, diff --git a/contrib/sepgsql/proc.c b/contrib/sepgsql/proc.c index 1efbc906c6b63ef6ec837ca292e120b06a374118..b68314d87851de0dea8cab5f501e9c33e9dcdbb1 100644 --- a/contrib/sepgsql/proc.c +++ b/contrib/sepgsql/proc.c @@ -42,9 +42,9 @@ sepgsql_proc_post_create(Oid functionId) char *tcontext; char *ncontext; int i; - StringInfoData audit_name; - ObjectAddress object; - Form_pg_proc proForm; + StringInfoData audit_name; + ObjectAddress object; + Form_pg_proc proForm; /* * Fetch namespace of the new procedure. Because pg_proc entry is not @@ -77,6 +77,7 @@ sepgsql_proc_post_create(Oid functionId) SEPG_DB_SCHEMA__ADD_NAME, getObjectDescription(&object), true); + /* * XXX - db_language:{implement} also should be checked here */ @@ -97,9 +98,10 @@ sepgsql_proc_post_create(Oid functionId) */ initStringInfo(&audit_name); appendStringInfo(&audit_name, "function %s(", NameStr(proForm->proname)); - for (i=0; i < proForm->pronargs; i++) + for (i = 0; i < proForm->pronargs; i++) { - Oid typeoid = proForm->proargtypes.values[i]; + Oid typeoid = proForm->proargtypes.values[i]; + if (i > 0) appendStringInfoChar(&audit_name, ','); appendStringInfoString(&audit_name, format_type_be(typeoid)); @@ -111,6 +113,7 @@ sepgsql_proc_post_create(Oid functionId) SEPG_DB_PROCEDURE__CREATE, audit_name.data, true); + /* * Assign the default security label on a new procedure */ @@ -138,8 +141,8 @@ sepgsql_proc_post_create(Oid functionId) void sepgsql_proc_drop(Oid functionId) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; /* * check db_schema:{remove_name} permission @@ -156,19 +159,19 @@ sepgsql_proc_drop(Oid functionId) true); pfree(audit_name); - /* - * check db_procedure:{drop} permission - */ + /* + * check db_procedure:{drop} permission + */ object.classId = ProcedureRelationId; object.objectId = functionId; object.objectSubId = 0; audit_name = getObjectDescription(&object); - sepgsql_avc_check_perms(&object, - SEPG_CLASS_DB_PROCEDURE, - SEPG_DB_PROCEDURE__DROP, - audit_name, - true); + sepgsql_avc_check_perms(&object, + SEPG_CLASS_DB_PROCEDURE, + SEPG_DB_PROCEDURE__DROP, + audit_name, + true); pfree(audit_name); } @@ -181,8 +184,8 @@ sepgsql_proc_drop(Oid functionId) void sepgsql_proc_relabel(Oid functionId, const char *seclabel) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; object.classId = ProcedureRelationId; object.objectId = functionId; @@ -198,6 +201,7 @@ sepgsql_proc_relabel(Oid functionId, const char *seclabel) SEPG_DB_PROCEDURE__RELABELFROM, audit_name, true); + /* * check db_procedure:{relabelto} permission */ diff --git a/contrib/sepgsql/relation.c b/contrib/sepgsql/relation.c index 259be492684ffb14ae127861b333bd5f790ad936..e759a7d98ee686ded63385ce4475a127817bd44a 100644 --- a/contrib/sepgsql/relation.c +++ b/contrib/sepgsql/relation.c @@ -44,9 +44,9 @@ sepgsql_attribute_post_create(Oid relOid, AttrNumber attnum) char *scontext; char *tcontext; char *ncontext; - char audit_name[2*NAMEDATALEN + 20]; + char audit_name[2 * NAMEDATALEN + 20]; ObjectAddress object; - Form_pg_attribute attForm; + Form_pg_attribute attForm; /* * Only attributes within regular relation have individual security @@ -84,6 +84,7 @@ sepgsql_attribute_post_create(Oid relOid, AttrNumber attnum) tcontext = sepgsql_get_label(RelationRelationId, relOid, 0); ncontext = sepgsql_compute_create(scontext, tcontext, SEPG_CLASS_DB_COLUMN); + /* * check db_column:{create} permission */ @@ -118,8 +119,8 @@ sepgsql_attribute_post_create(Oid relOid, AttrNumber attnum) void sepgsql_attribute_drop(Oid relOid, AttrNumber attnum) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; if (get_rel_relkind(relOid) != RELKIND_RELATION) return; @@ -151,7 +152,7 @@ sepgsql_attribute_relabel(Oid relOid, AttrNumber attnum, const char *seclabel) { ObjectAddress object; - char *audit_name; + char *audit_name; if (get_rel_relkind(relOid) != RELKIND_RELATION) ereport(ERROR, @@ -172,6 +173,7 @@ sepgsql_attribute_relabel(Oid relOid, AttrNumber attnum, SEPG_DB_COLUMN__RELABELFROM, audit_name, true); + /* * check db_column:{relabelto} permission */ @@ -203,7 +205,7 @@ sepgsql_relation_post_create(Oid relOid) char *tcontext; /* schema */ char *rcontext; /* relation */ char *ccontext; /* column */ - char audit_name[2*NAMEDATALEN + 20]; + char audit_name[2 * NAMEDATALEN + 20]; /* * Fetch catalog record of the new relation. Because pg_class entry is not @@ -254,6 +256,7 @@ sepgsql_relation_post_create(Oid relOid) SEPG_DB_SCHEMA__ADD_NAME, getObjectDescription(&object), true); + /* * Compute a default security label when we create a new relation object * under the specified namespace. @@ -273,6 +276,7 @@ sepgsql_relation_post_create(Oid relOid) SEPG_DB_DATABASE__CREATE, audit_name, true); + /* * Assign the default security label on the new relation */ @@ -288,10 +292,10 @@ sepgsql_relation_post_create(Oid relOid) if (classForm->relkind == RELKIND_RELATION) { Relation arel; - ScanKeyData akey; - SysScanDesc ascan; + ScanKeyData akey; + SysScanDesc ascan; HeapTuple atup; - Form_pg_attribute attForm; + Form_pg_attribute attForm; arel = heap_open(AttributeRelationId, AccessShareLock); @@ -315,6 +319,7 @@ sepgsql_relation_post_create(Oid relOid) ccontext = sepgsql_compute_create(scontext, rcontext, SEPG_CLASS_DB_COLUMN); + /* * check db_column:{create} permission */ @@ -348,10 +353,10 @@ out: void sepgsql_relation_drop(Oid relOid) { - ObjectAddress object; - char *audit_name; - uint16_t tclass = 0; - char relkind; + ObjectAddress object; + char *audit_name; + uint16_t tclass = 0; + char relkind; relkind = get_rel_relkind(relOid); if (relkind == RELKIND_RELATION) @@ -398,13 +403,13 @@ sepgsql_relation_drop(Oid relOid) */ if (relkind == RELKIND_RELATION) { - Form_pg_attribute attForm; + Form_pg_attribute attForm; CatCList *attrList; HeapTuple atttup; int i; attrList = SearchSysCacheList1(ATTNUM, ObjectIdGetDatum(relOid)); - for (i=0; i < attrList->n_members; i++) + for (i = 0; i < attrList->n_members; i++) { atttup = &attrList->members[i]->tuple; attForm = (Form_pg_attribute) GETSTRUCT(atttup); @@ -436,7 +441,7 @@ sepgsql_relation_drop(Oid relOid) void sepgsql_relation_relabel(Oid relOid, const char *seclabel) { - ObjectAddress object; + ObjectAddress object; char *audit_name; char relkind; uint16_t tclass = 0; @@ -468,6 +473,7 @@ sepgsql_relation_relabel(Oid relOid, const char *seclabel) SEPG_DB_TABLE__RELABELFROM, audit_name, true); + /* * check db_xxx:{relabelto} permission */ diff --git a/contrib/sepgsql/schema.c b/contrib/sepgsql/schema.c index 31d60efe18514daa722af0080240f0210360e781..230449dc4b3558d9c1d67cd2111afa0e846e7496 100644 --- a/contrib/sepgsql/schema.c +++ b/contrib/sepgsql/schema.c @@ -35,22 +35,22 @@ void sepgsql_schema_post_create(Oid namespaceId) { Relation rel; - ScanKeyData skey; - SysScanDesc sscan; + ScanKeyData skey; + SysScanDesc sscan; HeapTuple tuple; char *tcontext; char *ncontext; char audit_name[NAMEDATALEN + 20]; - ObjectAddress object; - Form_pg_namespace nspForm; + ObjectAddress object; + Form_pg_namespace nspForm; /* * Compute a default security label when we create a new schema object * under the working database. * - * XXX - uncoming version of libselinux supports to take object - * name to handle special treatment on default security label; - * such as special label on "pg_temp" schema. + * XXX - uncoming version of libselinux supports to take object name to + * handle special treatment on default security label; such as special + * label on "pg_temp" schema. */ rel = heap_open(NamespaceRelationId, AccessShareLock); @@ -71,6 +71,7 @@ sepgsql_schema_post_create(Oid namespaceId) ncontext = sepgsql_compute_create(sepgsql_get_client_label(), tcontext, SEPG_CLASS_DB_SCHEMA); + /* * check db_schema:{create} */ @@ -104,8 +105,8 @@ sepgsql_schema_post_create(Oid namespaceId) void sepgsql_schema_drop(Oid namespaceId) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; /* * check db_schema:{drop} permission @@ -116,7 +117,7 @@ sepgsql_schema_drop(Oid namespaceId) audit_name = getObjectDescription(&object); sepgsql_avc_check_perms(&object, - SEPG_CLASS_DB_SCHEMA, + SEPG_CLASS_DB_SCHEMA, SEPG_DB_SCHEMA__DROP, audit_name, true); @@ -132,8 +133,8 @@ sepgsql_schema_drop(Oid namespaceId) void sepgsql_schema_relabel(Oid namespaceId, const char *seclabel) { - ObjectAddress object; - char *audit_name; + ObjectAddress object; + char *audit_name; object.classId = NamespaceRelationId; object.objectId = namespaceId; @@ -149,6 +150,7 @@ sepgsql_schema_relabel(Oid namespaceId, const char *seclabel) SEPG_DB_SCHEMA__RELABELFROM, audit_name, true); + /* * check db_schema:{relabelto} permission */ diff --git a/contrib/sepgsql/sepgsql.h b/contrib/sepgsql/sepgsql.h index 708d4ee6563c82dd90d8da8b0e67b3f82baa49f0..479b136909bca0b49c8a8e67dbc4ae87867e56be 100644 --- a/contrib/sepgsql/sepgsql.h +++ b/contrib/sepgsql/sepgsql.h @@ -248,20 +248,21 @@ extern bool sepgsql_check_perms(const char *scontext, uint32 required, const char *audit_name, bool abort); + /* * uavc.c */ #define SEPGSQL_AVC_NOAUDIT ((void *)(-1)) extern bool sepgsql_avc_check_perms_label(const char *tcontext, - uint16 tclass, - uint32 required, - const char *audit_name, - bool abort); + uint16 tclass, + uint32 required, + const char *audit_name, + bool abort); extern bool sepgsql_avc_check_perms(const ObjectAddress *tobject, - uint16 tclass, - uint32 required, - const char *audit_name, - bool abort); + uint16 tclass, + uint32 required, + const char *audit_name, + bool abort); extern char *sepgsql_avc_trusted_proc(Oid functionId); extern void sepgsql_avc_init(void); @@ -269,7 +270,7 @@ extern void sepgsql_avc_init(void); * label.c */ extern char *sepgsql_get_client_label(void); -extern void sepgsql_init_client_label(void); +extern void sepgsql_init_client_label(void); extern char *sepgsql_get_label(Oid relOid, Oid objOid, int32 subId); extern void sepgsql_object_relabel(const ObjectAddress *object, @@ -290,7 +291,7 @@ extern bool sepgsql_dml_privileges(List *rangeTabls, bool abort); * database.c */ extern void sepgsql_database_post_create(Oid databaseId, - const char *dtemplate); + const char *dtemplate); extern void sepgsql_database_drop(Oid databaseId); extern void sepgsql_database_relabel(Oid databaseId, const char *seclabel); diff --git a/contrib/sepgsql/uavc.c b/contrib/sepgsql/uavc.c index 905f87dfc8dabc5f5eb40a66f619ae2982ea6c24..9641a17d79e84aec4fe8c521516e3d2d7662a595 100644 --- a/contrib/sepgsql/uavc.c +++ b/contrib/sepgsql/uavc.c @@ -30,22 +30,22 @@ */ typedef struct { - uint32 hash; /* hash value of this cache entry */ - char *scontext; /* security context of the subject */ - char *tcontext; /* security context of the target */ - uint16 tclass; /* object class of the target */ + uint32 hash; /* hash value of this cache entry */ + char *scontext; /* security context of the subject */ + char *tcontext; /* security context of the target */ + uint16 tclass; /* object class of the target */ - uint32 allowed; /* permissions to be allowed */ - uint32 auditallow; /* permissions to be audited on allowed */ - uint32 auditdeny; /* permissions to be audited on denied */ + uint32 allowed; /* permissions to be allowed */ + uint32 auditallow; /* permissions to be audited on allowed */ + uint32 auditdeny; /* permissions to be audited on denied */ - bool permissive; /* true, if permissive rule */ - bool hot_cache; /* true, if recently referenced */ + bool permissive; /* true, if permissive rule */ + bool hot_cache; /* true, if recently referenced */ bool tcontext_is_valid; - /* true, if tcontext is valid */ - char *ncontext; /* temporary scontext on execution of trusted - * procedure, or NULL elsewhere */ -} avc_cache; + /* true, if tcontext is valid */ + char *ncontext; /* temporary scontext on execution of trusted + * procedure, or NULL elsewhere */ +} avc_cache; /* * Declaration of static variables @@ -54,12 +54,12 @@ typedef struct #define AVC_NUM_RECLAIM 16 #define AVC_DEF_THRESHOLD 384 -static MemoryContext avc_mem_cxt; -static List *avc_slots[AVC_NUM_SLOTS]; /* avc's hash buckets */ -static int avc_num_caches; /* number of caches currently used */ -static int avc_lru_hint; /* index of the buckets to be reclaimed next */ -static int avc_threshold; /* threshold to launch cache-reclaiming */ -static char *avc_unlabeled; /* system 'unlabeled' label */ +static MemoryContext avc_mem_cxt; +static List *avc_slots[AVC_NUM_SLOTS]; /* avc's hash buckets */ +static int avc_num_caches; /* number of caches currently used */ +static int avc_lru_hint; /* index of the buckets to be reclaimed next */ +static int avc_threshold; /* threshold to launch cache-reclaiming */ +static char *avc_unlabeled; /* system 'unlabeled' label */ /* * Hash function @@ -67,8 +67,8 @@ static char *avc_unlabeled; /* system 'unlabeled' label */ static uint32 sepgsql_avc_hash(const char *scontext, const char *tcontext, uint16 tclass) { - return hash_any((const unsigned char *)scontext, strlen(scontext)) - ^ hash_any((const unsigned char *)tcontext, strlen(tcontext)) + return hash_any((const unsigned char *) scontext, strlen(scontext)) + ^ hash_any((const unsigned char *) tcontext, strlen(tcontext)) ^ tclass; } @@ -88,7 +88,7 @@ sepgsql_avc_reset(void) /* * Reclaim caches recently unreferenced - */ + */ static void sepgsql_avc_reclaim(void) { @@ -142,15 +142,15 @@ sepgsql_avc_reclaim(void) * Access control decisions must be atomic, but multiple system calls may * be required to make a decision; thus, when referencing the access vector * cache, we must loop until we complete without an intervening cache flush - * event. In practice, looping even once should be very rare. Callers should + * event. In practice, looping even once should be very rare. Callers should * do something like this: * - * sepgsql_avc_check_valid(); - * do { - * : - * <reference to uavc> - * : - * } while (!sepgsql_avc_check_valid()) + * sepgsql_avc_check_valid(); + * do { + * : + * <reference to uavc> + * : + * } while (!sepgsql_avc_check_valid()) * * ------------------------------------------------------------------------- */ @@ -169,7 +169,7 @@ sepgsql_avc_check_valid(void) /* * sepgsql_avc_unlabeled * - * Returns an alternative label to be applied when no label or an invalid + * Returns an alternative label to be applied when no label or an invalid * label would otherwise be assigned. */ static char * @@ -177,12 +177,12 @@ sepgsql_avc_unlabeled(void) { if (!avc_unlabeled) { - security_context_t unlabeled; + security_context_t unlabeled; if (security_get_initial_context_raw("unlabeled", &unlabeled) < 0) ereport(ERROR, - (errcode(ERRCODE_INTERNAL_ERROR), - errmsg("SELinux: failed to get initial security label: %m"))); + (errcode(ERRCODE_INTERNAL_ERROR), + errmsg("SELinux: failed to get initial security label: %m"))); PG_TRY(); { avc_unlabeled = MemoryContextStrdup(avc_mem_cxt, unlabeled); @@ -200,7 +200,7 @@ sepgsql_avc_unlabeled(void) } /* - * sepgsql_avc_compute + * sepgsql_avc_compute * * A fallback path, when cache mishit. It asks SELinux its access control * decision for the supplied pair of security context and object class. @@ -208,24 +208,24 @@ sepgsql_avc_unlabeled(void) static avc_cache * sepgsql_avc_compute(const char *scontext, const char *tcontext, uint16 tclass) { - char *ucontext = NULL; - char *ncontext = NULL; - MemoryContext oldctx; - avc_cache *cache; - uint32 hash; - int index; - struct av_decision avd; + char *ucontext = NULL; + char *ncontext = NULL; + MemoryContext oldctx; + avc_cache *cache; + uint32 hash; + int index; + struct av_decision avd; hash = sepgsql_avc_hash(scontext, tcontext, tclass); index = hash % AVC_NUM_SLOTS; /* - * Validation check of the supplied security context. - * Because it always invoke system-call, frequent check should be avoided. - * Unless security policy is reloaded, validation status shall be kept, so - * we also cache whether the supplied security context was valid, or not. + * Validation check of the supplied security context. Because it always + * invoke system-call, frequent check should be avoided. Unless security + * policy is reloaded, validation status shall be kept, so we also cache + * whether the supplied security context was valid, or not. */ - if (security_check_context_raw((security_context_t)tcontext) != 0) + if (security_check_context_raw((security_context_t) tcontext) != 0) ucontext = sepgsql_avc_unlabeled(); /* @@ -237,15 +237,14 @@ sepgsql_avc_compute(const char *scontext, const char *tcontext, uint16 tclass) sepgsql_compute_avd(scontext, ucontext, tclass, &avd); /* - * It also caches a security label to be switched when a client - * labeled as 'scontext' executes a procedure labeled as 'tcontext', - * not only access control decision on the procedure. - * The security label to be switched shall be computed uniquely on - * a pair of 'scontext' and 'tcontext', thus, it is reasonable to - * cache the new label on avc, and enables to reduce unnecessary - * system calls. - * It shall be referenced at sepgsql_needs_fmgr_hook to check whether - * the supplied function is a trusted procedure, or not. + * It also caches a security label to be switched when a client labeled as + * 'scontext' executes a procedure labeled as 'tcontext', not only access + * control decision on the procedure. The security label to be switched + * shall be computed uniquely on a pair of 'scontext' and 'tcontext', + * thus, it is reasonable to cache the new label on avc, and enables to + * reduce unnecessary system calls. It shall be referenced at + * sepgsql_needs_fmgr_hook to check whether the supplied function is a + * trusted procedure, or not. */ if (tclass == SEPG_CLASS_DB_PROCEDURE) { @@ -269,7 +268,7 @@ sepgsql_avc_compute(const char *scontext, const char *tcontext, uint16 tclass) cache = palloc0(sizeof(avc_cache)); - cache->hash = hash; + cache->hash = hash; cache->scontext = pstrdup(scontext); cache->tcontext = pstrdup(tcontext); cache->tclass = tclass; @@ -314,7 +313,7 @@ sepgsql_avc_lookup(const char *scontext, const char *tcontext, uint16 tclass) hash = sepgsql_avc_hash(scontext, tcontext, tclass); index = hash % AVC_NUM_SLOTS; - foreach (cell, avc_slots[index]) + foreach(cell, avc_slots[index]) { cache = lfirst(cell); @@ -348,14 +347,15 @@ sepgsql_avc_check_perms_label(const char *tcontext, uint16 tclass, uint32 required, const char *audit_name, bool abort) { - char *scontext = sepgsql_get_client_label(); + char *scontext = sepgsql_get_client_label(); avc_cache *cache; uint32 denied; uint32 audited; bool result; sepgsql_avc_check_valid(); - do { + do + { result = true; /* @@ -377,16 +377,16 @@ sepgsql_avc_check_perms_label(const char *tcontext, audited = (denied ? (denied & ~0) : (required & ~0)); else audited = denied ? (denied & cache->auditdeny) - : (required & cache->auditallow); + : (required & cache->auditallow); if (denied) { /* * In permissive mode or permissive domain, violated permissions * shall be audited to the log files at once, and then implicitly - * allowed to avoid a flood of access denied logs, because - * the purpose of permissive mode/domain is to collect a violation - * log that will make it possible to fix up the security policy. + * allowed to avoid a flood of access denied logs, because the + * purpose of permissive mode/domain is to collect a violation log + * that will make it possible to fix up the security policy. */ if (!sepgsql_getenforce() || cache->permissive) cache->allowed |= required; @@ -397,10 +397,10 @@ sepgsql_avc_check_perms_label(const char *tcontext, /* * In the case when we have something auditable actions here, - * sepgsql_audit_log shall be called with text representation of - * security labels for both of subject and object. - * It records this access violation, so DBA will be able to find - * out unexpected security problems later. + * sepgsql_audit_log shall be called with text representation of security + * labels for both of subject and object. It records this access + * violation, so DBA will be able to find out unexpected security problems + * later. */ if (audited != 0 && audit_name != SEPGSQL_AVC_NOAUDIT && @@ -428,8 +428,8 @@ sepgsql_avc_check_perms(const ObjectAddress *tobject, uint16 tclass, uint32 required, const char *audit_name, bool abort) { - char *tcontext = GetSecurityLabel(tobject, SEPGSQL_LABEL_TAG); - bool rc; + char *tcontext = GetSecurityLabel(tobject, SEPGSQL_LABEL_TAG); + bool rc; rc = sepgsql_avc_check_perms_label(tcontext, tclass, required, @@ -450,10 +450,10 @@ sepgsql_avc_check_perms(const ObjectAddress *tobject, char * sepgsql_avc_trusted_proc(Oid functionId) { - char *scontext = sepgsql_get_client_label(); - char *tcontext; - ObjectAddress tobject; - avc_cache *cache; + char *scontext = sepgsql_get_client_label(); + char *tcontext; + ObjectAddress tobject; + avc_cache *cache; tobject.classId = ProcedureRelationId; tobject.objectId = functionId; @@ -461,7 +461,8 @@ sepgsql_avc_trusted_proc(Oid functionId) tcontext = GetSecurityLabel(&tobject, SEPGSQL_LABEL_TAG); sepgsql_avc_check_valid(); - do { + do + { if (tcontext) cache = sepgsql_avc_lookup(scontext, tcontext, SEPG_CLASS_DB_PROCEDURE); @@ -492,7 +493,7 @@ sepgsql_avc_exit(int code, Datum arg) void sepgsql_avc_init(void) { - int rc; + int rc; /* * All the avc stuff shall be allocated on avc_mem_cxt @@ -508,12 +509,11 @@ sepgsql_avc_init(void) avc_threshold = AVC_DEF_THRESHOLD; /* - * SELinux allows to mmap(2) its kernel status page in read-only mode - * to inform userspace applications its status updating (such as - * policy reloading) without system-call invocations. - * This feature is only supported in Linux-2.6.38 or later, however, - * libselinux provides a fallback mode to know its status using - * netlink sockets. + * SELinux allows to mmap(2) its kernel status page in read-only mode to + * inform userspace applications its status updating (such as policy + * reloading) without system-call invocations. This feature is only + * supported in Linux-2.6.38 or later, however, libselinux provides a + * fallback mode to know its status using netlink sockets. */ rc = selinux_status_open(1); if (rc < 0) diff --git a/contrib/spi/refint.c b/contrib/spi/refint.c index 39a0160587461a31f0ba01c5e7e26fc9ad14fbbf..8dc565a190775f5ba94ae11e684de4da027debfe 100644 --- a/contrib/spi/refint.c +++ b/contrib/spi/refint.c @@ -536,8 +536,7 @@ check_foreign_key(PG_FUNCTION_ARGS) /* * Remember that SPI_prepare places plan in current memory context - * - so, we have to save plan in Top memory context for later - * use. + * - so, we have to save plan in Top memory context for later use. */ if (SPI_keepplan(pplan)) /* internal error */ diff --git a/contrib/vacuumlo/vacuumlo.c b/contrib/vacuumlo/vacuumlo.c index 641a8c3425d97614a14a547f91651f0bbd83eae8..958a496b24abb85a82cb1f738d02db1ac9ff22b6 100644 --- a/contrib/vacuumlo/vacuumlo.c +++ b/contrib/vacuumlo/vacuumlo.c @@ -69,7 +69,7 @@ vacuumlo(const char *database, const struct _param * param) int i; static char *password = NULL; bool new_pass; - bool success = true; + bool success = true; /* Note: password can be carried over from a previous call */ if (param->pg_prompt == TRI_YES && password == NULL) @@ -261,8 +261,8 @@ vacuumlo(const char *database, const struct _param * param) * We don't want to run each delete as an individual transaction, because * the commit overhead would be high. However, since 9.0 the backend will * acquire a lock per deleted LO, so deleting too many LOs per transaction - * risks running out of room in the shared-memory lock table. - * Accordingly, we delete up to transaction_limit LOs per transaction. + * risks running out of room in the shared-memory lock table. Accordingly, + * we delete up to transaction_limit LOs per transaction. */ res = PQexec(conn, "begin"); if (PQresultStatus(res) != PGRES_COMMAND_OK) @@ -459,8 +459,8 @@ main(int argc, char **argv) if (param.transaction_limit < 0) { fprintf(stderr, - "%s: transaction limit must not be negative (0 disables)\n", - progname); + "%s: transaction limit must not be negative (0 disables)\n", + progname); exit(1); } break; diff --git a/contrib/xml2/xpath.c b/contrib/xml2/xpath.c index 2ddee59fcb779ba3677affdbe16eb8e7fc3ff204..660d25c3490bc6bfda8298e3d474186fe7e14492 100644 --- a/contrib/xml2/xpath.c +++ b/contrib/xml2/xpath.c @@ -702,126 +702,126 @@ xpath_table(PG_FUNCTION_ARGS) PG_TRY(); { - /* For each row i.e. document returned from SPI */ - for (i = 0; i < proc; i++) - { - char *pkey; - char *xmldoc; - xmlXPathContextPtr ctxt; - xmlXPathObjectPtr res; - xmlChar *resstr; - xmlXPathCompExprPtr comppath; - - /* Extract the row data as C Strings */ - spi_tuple = tuptable->vals[i]; - pkey = SPI_getvalue(spi_tuple, spi_tupdesc, 1); - xmldoc = SPI_getvalue(spi_tuple, spi_tupdesc, 2); - - /* - * Clear the values array, so that not-well-formed documents return - * NULL in all columns. Note that this also means that spare columns - * will be NULL. - */ - for (j = 0; j < ret_tupdesc->natts; j++) - values[j] = NULL; - - /* Insert primary key */ - values[0] = pkey; - - /* Parse the document */ - if (xmldoc) - doctree = xmlParseMemory(xmldoc, strlen(xmldoc)); - else /* treat NULL as not well-formed */ - doctree = NULL; - - if (doctree == NULL) + /* For each row i.e. document returned from SPI */ + for (i = 0; i < proc; i++) { - /* not well-formed, so output all-NULL tuple */ - ret_tuple = BuildTupleFromCStrings(attinmeta, values); - tuplestore_puttuple(tupstore, ret_tuple); - heap_freetuple(ret_tuple); - } - else - { - /* New loop here - we have to deal with nodeset results */ - rownr = 0; - - do + char *pkey; + char *xmldoc; + xmlXPathContextPtr ctxt; + xmlXPathObjectPtr res; + xmlChar *resstr; + xmlXPathCompExprPtr comppath; + + /* Extract the row data as C Strings */ + spi_tuple = tuptable->vals[i]; + pkey = SPI_getvalue(spi_tuple, spi_tupdesc, 1); + xmldoc = SPI_getvalue(spi_tuple, spi_tupdesc, 2); + + /* + * Clear the values array, so that not-well-formed documents + * return NULL in all columns. Note that this also means that + * spare columns will be NULL. + */ + for (j = 0; j < ret_tupdesc->natts; j++) + values[j] = NULL; + + /* Insert primary key */ + values[0] = pkey; + + /* Parse the document */ + if (xmldoc) + doctree = xmlParseMemory(xmldoc, strlen(xmldoc)); + else /* treat NULL as not well-formed */ + doctree = NULL; + + if (doctree == NULL) { - /* Now evaluate the set of xpaths. */ - had_values = false; - for (j = 0; j < numpaths; j++) + /* not well-formed, so output all-NULL tuple */ + ret_tuple = BuildTupleFromCStrings(attinmeta, values); + tuplestore_puttuple(tupstore, ret_tuple); + heap_freetuple(ret_tuple); + } + else + { + /* New loop here - we have to deal with nodeset results */ + rownr = 0; + + do { - ctxt = xmlXPathNewContext(doctree); - ctxt->node = xmlDocGetRootElement(doctree); + /* Now evaluate the set of xpaths. */ + had_values = false; + for (j = 0; j < numpaths; j++) + { + ctxt = xmlXPathNewContext(doctree); + ctxt->node = xmlDocGetRootElement(doctree); - /* compile the path */ - comppath = xmlXPathCompile(xpaths[j]); - if (comppath == NULL) - xml_ereport(xmlerrcxt, ERROR, - ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, - "XPath Syntax Error"); + /* compile the path */ + comppath = xmlXPathCompile(xpaths[j]); + if (comppath == NULL) + xml_ereport(xmlerrcxt, ERROR, + ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, + "XPath Syntax Error"); - /* Now evaluate the path expression. */ - res = xmlXPathCompiledEval(comppath, ctxt); - xmlXPathFreeCompExpr(comppath); + /* Now evaluate the path expression. */ + res = xmlXPathCompiledEval(comppath, ctxt); + xmlXPathFreeCompExpr(comppath); - if (res != NULL) - { - switch (res->type) + if (res != NULL) { - case XPATH_NODESET: - /* We see if this nodeset has enough nodes */ - if (res->nodesetval != NULL && - rownr < res->nodesetval->nodeNr) - { - resstr = xmlXPathCastNodeToString(res->nodesetval->nodeTab[rownr]); - had_values = true; - } - else - resstr = NULL; - - break; - - case XPATH_STRING: - resstr = xmlStrdup(res->stringval); - break; - - default: - elog(NOTICE, "unsupported XQuery result: %d", res->type); - resstr = xmlStrdup((const xmlChar *) "<unsupported/>"); + switch (res->type) + { + case XPATH_NODESET: + /* We see if this nodeset has enough nodes */ + if (res->nodesetval != NULL && + rownr < res->nodesetval->nodeNr) + { + resstr = xmlXPathCastNodeToString(res->nodesetval->nodeTab[rownr]); + had_values = true; + } + else + resstr = NULL; + + break; + + case XPATH_STRING: + resstr = xmlStrdup(res->stringval); + break; + + default: + elog(NOTICE, "unsupported XQuery result: %d", res->type); + resstr = xmlStrdup((const xmlChar *) "<unsupported/>"); + } + + /* + * Insert this into the appropriate column in the + * result tuple. + */ + values[j + 1] = (char *) resstr; } - - /* - * Insert this into the appropriate column in the - * result tuple. - */ - values[j + 1] = (char *) resstr; + xmlXPathFreeContext(ctxt); } - xmlXPathFreeContext(ctxt); - } - /* Now add the tuple to the output, if there is one. */ - if (had_values) - { - ret_tuple = BuildTupleFromCStrings(attinmeta, values); - tuplestore_puttuple(tupstore, ret_tuple); - heap_freetuple(ret_tuple); - } + /* Now add the tuple to the output, if there is one. */ + if (had_values) + { + ret_tuple = BuildTupleFromCStrings(attinmeta, values); + tuplestore_puttuple(tupstore, ret_tuple); + heap_freetuple(ret_tuple); + } - rownr++; - } while (had_values); - } + rownr++; + } while (had_values); + } - if (doctree != NULL) - xmlFreeDoc(doctree); - doctree = NULL; + if (doctree != NULL) + xmlFreeDoc(doctree); + doctree = NULL; - if (pkey) - pfree(pkey); - if (xmldoc) - pfree(xmldoc); - } + if (pkey) + pfree(pkey); + if (xmldoc) + pfree(xmldoc); + } } PG_CATCH(); { diff --git a/contrib/xml2/xslt_proc.c b/contrib/xml2/xslt_proc.c index ba1171a041fadb9d605ef78b879bf43946c10b02..a93931d2618ccecaefa3a71fe752fcd52132302c 100644 --- a/contrib/xml2/xslt_proc.c +++ b/contrib/xml2/xslt_proc.c @@ -85,40 +85,40 @@ xslt_process(PG_FUNCTION_ARGS) { /* Check to see if document is a file or a literal */ - if (VARDATA(doct)[0] == '<') - doctree = xmlParseMemory((char *) VARDATA(doct), VARSIZE(doct) - VARHDRSZ); - else - doctree = xmlParseFile(text_to_cstring(doct)); - - if (doctree == NULL) - xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, - "error parsing XML document"); + if (VARDATA(doct)[0] == '<') + doctree = xmlParseMemory((char *) VARDATA(doct), VARSIZE(doct) - VARHDRSZ); + else + doctree = xmlParseFile(text_to_cstring(doct)); - /* Same for stylesheet */ - if (VARDATA(ssheet)[0] == '<') - { - ssdoc = xmlParseMemory((char *) VARDATA(ssheet), - VARSIZE(ssheet) - VARHDRSZ); - if (ssdoc == NULL) + if (doctree == NULL) xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, - "error parsing stylesheet as XML document"); + "error parsing XML document"); - stylesheet = xsltParseStylesheetDoc(ssdoc); - } - else - stylesheet = xsltParseStylesheetFile((xmlChar *) text_to_cstring(ssheet)); + /* Same for stylesheet */ + if (VARDATA(ssheet)[0] == '<') + { + ssdoc = xmlParseMemory((char *) VARDATA(ssheet), + VARSIZE(ssheet) - VARHDRSZ); + if (ssdoc == NULL) + xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, + "error parsing stylesheet as XML document"); + + stylesheet = xsltParseStylesheetDoc(ssdoc); + } + else + stylesheet = xsltParseStylesheetFile((xmlChar *) text_to_cstring(ssheet)); - if (stylesheet == NULL) - xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, - "failed to parse stylesheet"); + if (stylesheet == NULL) + xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, + "failed to parse stylesheet"); - restree = xsltApplyStylesheet(stylesheet, doctree, params); + restree = xsltApplyStylesheet(stylesheet, doctree, params); - if (restree == NULL) - xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, - "failed to apply stylesheet"); + if (restree == NULL) + xml_ereport(xmlerrcxt, ERROR, ERRCODE_EXTERNAL_ROUTINE_EXCEPTION, + "failed to apply stylesheet"); - resstat = xsltSaveResultToString(&resstr, &reslen, restree, stylesheet); + resstat = xsltSaveResultToString(&resstr, &reslen, restree, stylesheet); } PG_CATCH(); { diff --git a/src/backend/access/gist/gist.c b/src/backend/access/gist/gist.c index 1efaaee1a8636745f6b55ab375dd027bd5b72abf..783590ea55e59134bbd22cbf17bd4712e4e543a4 100644 --- a/src/backend/access/gist/gist.c +++ b/src/backend/access/gist/gist.c @@ -27,7 +27,7 @@ /* non-export function prototypes */ static void gistfixsplit(GISTInsertState *state, GISTSTATE *giststate); static bool gistinserttuple(GISTInsertState *state, GISTInsertStack *stack, - GISTSTATE *giststate, IndexTuple tuple, OffsetNumber oldoffnum); + GISTSTATE *giststate, IndexTuple tuple, OffsetNumber oldoffnum); static bool gistinserttuples(GISTInsertState *state, GISTInsertStack *stack, GISTSTATE *giststate, IndexTuple *tuples, int ntup, OffsetNumber oldoffnum, @@ -781,8 +781,8 @@ gistFindPath(Relation r, BlockNumber child, OffsetNumber *downlinkoffnum) { /* * Page was split while we looked elsewhere. We didn't see the - * downlink to the right page when we scanned the parent, so - * add it to the queue now. + * downlink to the right page when we scanned the parent, so add + * it to the queue now. * * Put the right page ahead of the queue, so that we visit it * next. That's important, because if this is the lowest internal @@ -829,7 +829,7 @@ gistFindPath(Relation r, BlockNumber child, OffsetNumber *downlinkoffnum) elog(ERROR, "failed to re-find parent of a page in index \"%s\", block %u", RelationGetRelationName(r), child); - return NULL; /* keep compiler quiet */ + return NULL; /* keep compiler quiet */ } /* @@ -1046,7 +1046,7 @@ gistfixsplit(GISTInsertState *state, GISTSTATE *giststate) */ static bool gistinserttuple(GISTInsertState *state, GISTInsertStack *stack, - GISTSTATE *giststate, IndexTuple tuple, OffsetNumber oldoffnum) + GISTSTATE *giststate, IndexTuple tuple, OffsetNumber oldoffnum) { return gistinserttuples(state, stack, giststate, &tuple, 1, oldoffnum, InvalidBuffer, InvalidBuffer, false, false); @@ -1308,7 +1308,7 @@ initGISTstate(Relation index) giststate = (GISTSTATE *) palloc(sizeof(GISTSTATE)); giststate->scanCxt = scanCxt; - giststate->tempCxt = scanCxt; /* caller must change this if needed */ + giststate->tempCxt = scanCxt; /* caller must change this if needed */ giststate->tupdesc = index->rd_att; for (i = 0; i < index->rd_att->natts; i++) diff --git a/src/backend/access/gist/gistbuild.c b/src/backend/access/gist/gistbuild.c index 712e59ac908c9c7fc5e1702eab27cea2bc1170d8..8caf4856763b68828c6f23d9c1df6c9a60f8e7b5 100644 --- a/src/backend/access/gist/gistbuild.c +++ b/src/backend/access/gist/gistbuild.c @@ -48,7 +48,7 @@ typedef enum * before switching to the buffering build * mode */ GIST_BUFFERING_ACTIVE /* in buffering build mode */ -} GistBufferingMode; +} GistBufferingMode; /* Working state for gistbuild and its callback */ typedef struct @@ -263,7 +263,7 @@ gistValidateBufferingOption(char *value) ereport(ERROR, (errcode(ERRCODE_INVALID_PARAMETER_VALUE), errmsg("invalid value for \"buffering\" option"), - errdetail("Valid values are \"on\", \"off\", and \"auto\"."))); + errdetail("Valid values are \"on\", \"off\", and \"auto\"."))); } } @@ -567,7 +567,7 @@ gistProcessItup(GISTBuildState *buildstate, IndexTuple itup, BlockNumber childblkno; Buffer buffer; bool result = false; - BlockNumber blkno; + BlockNumber blkno; int level; OffsetNumber downlinkoffnum = InvalidOffsetNumber; BlockNumber parentblkno = InvalidBlockNumber; @@ -623,7 +623,7 @@ gistProcessItup(GISTBuildState *buildstate, IndexTuple itup, { gistbufferinginserttuples(buildstate, buffer, level, &newtup, 1, childoffnum, - InvalidBlockNumber, InvalidOffsetNumber); + InvalidBlockNumber, InvalidOffsetNumber); /* gistbufferinginserttuples() released the buffer */ } else @@ -716,26 +716,26 @@ gistbufferinginserttuples(GISTBuildState *buildstate, Buffer buffer, int level, /* * All the downlinks on the old root page are now on one of the child - * pages. Visit all the new child pages to memorize the parents of - * the grandchildren. + * pages. Visit all the new child pages to memorize the parents of the + * grandchildren. */ if (gfbb->rootlevel > 1) { maxoff = PageGetMaxOffsetNumber(page); for (off = FirstOffsetNumber; off <= maxoff; off++) { - ItemId iid = PageGetItemId(page, off); - IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); + ItemId iid = PageGetItemId(page, off); + IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); BlockNumber childblkno = ItemPointerGetBlockNumber(&(idxtuple->t_tid)); - Buffer childbuf = ReadBuffer(buildstate->indexrel, childblkno); + Buffer childbuf = ReadBuffer(buildstate->indexrel, childblkno); LockBuffer(childbuf, GIST_SHARE); gistMemorizeAllDownlinks(buildstate, childbuf); UnlockReleaseBuffer(childbuf); /* - * Also remember that the parent of the new child page is - * the root block. + * Also remember that the parent of the new child page is the + * root block. */ gistMemorizeParent(buildstate, childblkno, GIST_ROOT_BLKNO); } @@ -789,8 +789,8 @@ gistbufferinginserttuples(GISTBuildState *buildstate, Buffer buffer, int level, * Remember the parent of each new child page in our parent map. * This assumes that the downlinks fit on the parent page. If the * parent page is split, too, when we recurse up to insert the - * downlinks, the recursive gistbufferinginserttuples() call - * will update the map again. + * downlinks, the recursive gistbufferinginserttuples() call will + * update the map again. */ if (level > 0) gistMemorizeParent(buildstate, @@ -879,8 +879,9 @@ gistBufferingFindCorrectParent(GISTBuildState *buildstate, if (parent == *parentblkno && *parentblkno != InvalidBlockNumber && *downlinkoffnum != InvalidOffsetNumber && *downlinkoffnum <= maxoff) { - ItemId iid = PageGetItemId(page, *downlinkoffnum); - IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); + ItemId iid = PageGetItemId(page, *downlinkoffnum); + IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); + if (ItemPointerGetBlockNumber(&(idxtuple->t_tid)) == childblkno) { /* Still there */ @@ -889,16 +890,17 @@ gistBufferingFindCorrectParent(GISTBuildState *buildstate, } /* - * Downlink was not at the offset where it used to be. Scan the page - * to find it. During normal gist insertions, it might've moved to another - * page, to the right, but during a buffering build, we keep track of - * the parent of each page in the lookup table so we should always know - * what page it's on. + * Downlink was not at the offset where it used to be. Scan the page to + * find it. During normal gist insertions, it might've moved to another + * page, to the right, but during a buffering build, we keep track of the + * parent of each page in the lookup table so we should always know what + * page it's on. */ for (off = FirstOffsetNumber; off <= maxoff; off = OffsetNumberNext(off)) { - ItemId iid = PageGetItemId(page, off); - IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); + ItemId iid = PageGetItemId(page, off); + IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); + if (ItemPointerGetBlockNumber(&(idxtuple->t_tid)) == childblkno) { /* yes!!, found it */ @@ -908,7 +910,7 @@ gistBufferingFindCorrectParent(GISTBuildState *buildstate, } elog(ERROR, "failed to re-find parent for block %u", childblkno); - return InvalidBuffer; /* keep compiler quiet */ + return InvalidBuffer; /* keep compiler quiet */ } /* @@ -1129,7 +1131,7 @@ gistGetMaxLevel(Relation index) typedef struct { - BlockNumber childblkno; /* hash key */ + BlockNumber childblkno; /* hash key */ BlockNumber parentblkno; } ParentMapEntry; @@ -1156,9 +1158,9 @@ gistMemorizeParent(GISTBuildState *buildstate, BlockNumber child, BlockNumber pa bool found; entry = (ParentMapEntry *) hash_search(buildstate->parentMap, - (const void *) &child, - HASH_ENTER, - &found); + (const void *) &child, + HASH_ENTER, + &found); entry->parentblkno = parent; } @@ -1171,16 +1173,17 @@ gistMemorizeAllDownlinks(GISTBuildState *buildstate, Buffer parentbuf) OffsetNumber maxoff; OffsetNumber off; BlockNumber parentblkno = BufferGetBlockNumber(parentbuf); - Page page = BufferGetPage(parentbuf); + Page page = BufferGetPage(parentbuf); Assert(!GistPageIsLeaf(page)); maxoff = PageGetMaxOffsetNumber(page); for (off = FirstOffsetNumber; off <= maxoff; off++) { - ItemId iid = PageGetItemId(page, off); - IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); + ItemId iid = PageGetItemId(page, off); + IndexTuple idxtuple = (IndexTuple) PageGetItem(page, iid); BlockNumber childblkno = ItemPointerGetBlockNumber(&(idxtuple->t_tid)); + gistMemorizeParent(buildstate, childblkno, parentblkno); } } @@ -1193,9 +1196,9 @@ gistGetParent(GISTBuildState *buildstate, BlockNumber child) /* Find node buffer in hash table */ entry = (ParentMapEntry *) hash_search(buildstate->parentMap, - (const void *) &child, - HASH_FIND, - &found); + (const void *) &child, + HASH_FIND, + &found); if (!found) elog(ERROR, "could not find parent of block %d in lookup table", child); diff --git a/src/backend/access/gist/gistbuildbuffers.c b/src/backend/access/gist/gistbuildbuffers.c index 3feca263a7621ded24ba818c315067832d355bee..39aec856f9270a4751ef3e97ee469e5c14efbf5f 100644 --- a/src/backend/access/gist/gistbuildbuffers.c +++ b/src/backend/access/gist/gistbuildbuffers.c @@ -528,7 +528,7 @@ typedef struct bool isnull[INDEX_MAX_KEYS]; GISTPageSplitInfo *splitinfo; GISTNodeBuffer *nodeBuffer; -} RelocationBufferInfo; +} RelocationBufferInfo; /* * At page split, distribute tuples from the buffer of the split page to diff --git a/src/backend/access/gist/gistproc.c b/src/backend/access/gist/gistproc.c index d97c64ede3f9326fb507824556c05686070f3e57..09e911d0981ef335baf99812b56fc3b468958a14 100644 --- a/src/backend/access/gist/gistproc.c +++ b/src/backend/access/gist/gistproc.c @@ -244,7 +244,7 @@ typedef struct int index; /* Delta between penalties of entry insertion into different groups */ double delta; -} CommonEntry; +} CommonEntry; /* * Context for g_box_consider_split. Contains information about currently @@ -267,7 +267,7 @@ typedef struct int dim; /* axis of this split */ double range; /* width of general MBR projection to the * selected axis */ -} ConsiderSplitContext; +} ConsiderSplitContext; /* * Interval represents projection of box to axis. @@ -276,7 +276,7 @@ typedef struct { double lower, upper; -} SplitInterval; +} SplitInterval; /* * Interval comparison function by lower bound of the interval; diff --git a/src/backend/access/gist/gistscan.c b/src/backend/access/gist/gistscan.c index bf139de8246f40b6373bcc49ad184573522dbb2c..c9fc9ba97f9bd3c78f6fe748e1c0a044474f5265 100644 --- a/src/backend/access/gist/gistscan.c +++ b/src/backend/access/gist/gistscan.c @@ -124,7 +124,7 @@ gistbeginscan(PG_FUNCTION_ARGS) so->giststate = giststate; giststate->tempCxt = createTempGistContext(); so->queue = NULL; - so->queueCxt = giststate->scanCxt; /* see gistrescan */ + so->queueCxt = giststate->scanCxt; /* see gistrescan */ /* workspaces with size dependent on numberOfOrderBys: */ so->tmpTreeItem = palloc(GSTIHDRSZ + sizeof(double) * scan->numberOfOrderBys); diff --git a/src/backend/access/gist/gistsplit.c b/src/backend/access/gist/gistsplit.c index 2ec69a60d454818d0cf005d9fde66a8f88145d22..739fc597ceb70613fc72790ec1eae7ec6c54cf99 100644 --- a/src/backend/access/gist/gistsplit.c +++ b/src/backend/access/gist/gistsplit.c @@ -581,8 +581,7 @@ gistSplitByKey(Relation r, Page page, IndexTuple *itup, int len, GISTSTATE *gist if (v->spl_equiv == NULL) { /* - * simple case: left and right keys for attno column are - * equal + * simple case: left and right keys for attno column are equal */ gistSplitByKey(r, page, itup, len, giststate, v, entryvec, attno + 1); } diff --git a/src/backend/access/hash/hashovfl.c b/src/backend/access/hash/hashovfl.c index 96dabdb48ae689d3fc551cb6a3d854f3c5e7e8a1..bbea5e4eaca275bd37be2a056921e050ff4f6901 100644 --- a/src/backend/access/hash/hashovfl.c +++ b/src/backend/access/hash/hashovfl.c @@ -391,7 +391,7 @@ _hash_freeovflpage(Relation rel, Buffer ovflbuf, uint32 ovflbitno; int32 bitmappage, bitmapbit; - Bucket bucket PG_USED_FOR_ASSERTS_ONLY; + Bucket bucket PG_USED_FOR_ASSERTS_ONLY; /* Get information from the doomed page */ _hash_checkpage(rel, ovflbuf, LH_OVERFLOW_PAGE); diff --git a/src/backend/access/heap/heapam.c b/src/backend/access/heap/heapam.c index 2d81383ae8ad467427c6161694517856df26d055..9519e73e54c6c6315015a648ef0f958527315c80 100644 --- a/src/backend/access/heap/heapam.c +++ b/src/backend/access/heap/heapam.c @@ -223,9 +223,9 @@ heapgetpage(HeapScanDesc scan, BlockNumber page) } /* - * Be sure to check for interrupts at least once per page. Checks at - * higher code levels won't be able to stop a seqscan that encounters - * many pages' worth of consecutive dead tuples. + * Be sure to check for interrupts at least once per page. Checks at + * higher code levels won't be able to stop a seqscan that encounters many + * pages' worth of consecutive dead tuples. */ CHECK_FOR_INTERRUPTS(); @@ -997,8 +997,8 @@ relation_openrv(const RangeVar *relation, LOCKMODE lockmode) * * Same as relation_openrv, but with an additional missing_ok argument * allowing a NULL return rather than an error if the relation is not - * found. (Note that some other causes, such as permissions problems, - * will still result in an ereport.) + * found. (Note that some other causes, such as permissions problems, + * will still result in an ereport.) * ---------------- */ Relation @@ -1105,7 +1105,7 @@ heap_openrv(const RangeVar *relation, LOCKMODE lockmode) * by a RangeVar node * * As above, but optionally return NULL instead of failing for - * relation-not-found. + * relation-not-found. * ---------------- */ Relation @@ -1588,10 +1588,10 @@ heap_hot_search_buffer(ItemPointer tid, Relation relation, Buffer buffer, /* * When first_call is true (and thus, skip is initially false) we'll - * return the first tuple we find. But on later passes, heapTuple + * return the first tuple we find. But on later passes, heapTuple * will initially be pointing to the tuple we returned last time. - * Returning it again would be incorrect (and would loop forever), - * so we skip it and return the next match we find. + * Returning it again would be incorrect (and would loop forever), so + * we skip it and return the next match we find. */ if (!skip) { @@ -1651,7 +1651,7 @@ heap_hot_search(ItemPointer tid, Relation relation, Snapshot snapshot, { bool result; Buffer buffer; - HeapTupleData heapTuple; + HeapTupleData heapTuple; buffer = ReadBuffer(relation, ItemPointerGetBlockNumber(tid)); LockBuffer(buffer, BUFFER_LOCK_SHARE); @@ -1885,14 +1885,14 @@ heap_insert(Relation relation, HeapTuple tup, CommandId cid, heaptup = heap_prepare_insert(relation, tup, xid, cid, options); /* - * We're about to do the actual insert -- but check for conflict first, - * to avoid possibly having to roll back work we've just done. + * We're about to do the actual insert -- but check for conflict first, to + * avoid possibly having to roll back work we've just done. * - * For a heap insert, we only need to check for table-level SSI locks. - * Our new tuple can't possibly conflict with existing tuple locks, and - * heap page locks are only consolidated versions of tuple locks; they do - * not lock "gaps" as index page locks do. So we don't need to identify - * a buffer before making the call. + * For a heap insert, we only need to check for table-level SSI locks. Our + * new tuple can't possibly conflict with existing tuple locks, and heap + * page locks are only consolidated versions of tuple locks; they do not + * lock "gaps" as index page locks do. So we don't need to identify a + * buffer before making the call. */ CheckForSerializableConflictIn(relation, NULL, InvalidBuffer); @@ -2123,11 +2123,11 @@ heap_multi_insert(Relation relation, HeapTuple *tuples, int ntuples, * We're about to do the actual inserts -- but check for conflict first, * to avoid possibly having to roll back work we've just done. * - * For a heap insert, we only need to check for table-level SSI locks. - * Our new tuple can't possibly conflict with existing tuple locks, and - * heap page locks are only consolidated versions of tuple locks; they do - * not lock "gaps" as index page locks do. So we don't need to identify - * a buffer before making the call. + * For a heap insert, we only need to check for table-level SSI locks. Our + * new tuple can't possibly conflict with existing tuple locks, and heap + * page locks are only consolidated versions of tuple locks; they do not + * lock "gaps" as index page locks do. So we don't need to identify a + * buffer before making the call. */ CheckForSerializableConflictIn(relation, NULL, InvalidBuffer); @@ -2137,12 +2137,11 @@ heap_multi_insert(Relation relation, HeapTuple *tuples, int ntuples, Buffer buffer; Buffer vmbuffer = InvalidBuffer; bool all_visible_cleared = false; - int nthispage; + int nthispage; /* - * Find buffer where at least the next tuple will fit. If the page - * is all-visible, this will also pin the requisite visibility map - * page. + * Find buffer where at least the next tuple will fit. If the page is + * all-visible, this will also pin the requisite visibility map page. */ buffer = RelationGetBufferForTuple(relation, heaptuples[ndone]->t_len, InvalidBuffer, options, bistate, @@ -2358,7 +2357,7 @@ heap_delete(Relation relation, ItemPointer tid, ItemId lp; HeapTupleData tp; Page page; - BlockNumber block; + BlockNumber block; Buffer buffer; Buffer vmbuffer = InvalidBuffer; bool have_tuple_lock = false; @@ -2372,10 +2371,10 @@ heap_delete(Relation relation, ItemPointer tid, page = BufferGetPage(buffer); /* - * Before locking the buffer, pin the visibility map page if it appears - * to be necessary. Since we haven't got the lock yet, someone else might - * be in the middle of changing this, so we'll need to recheck after - * we have the lock. + * Before locking the buffer, pin the visibility map page if it appears to + * be necessary. Since we haven't got the lock yet, someone else might be + * in the middle of changing this, so we'll need to recheck after we have + * the lock. */ if (PageIsAllVisible(page)) visibilitymap_pin(relation, block, &vmbuffer); @@ -2717,7 +2716,7 @@ heap_update(Relation relation, ItemPointer otid, HeapTuple newtup, HeapTupleData oldtup; HeapTuple heaptup; Page page; - BlockNumber block; + BlockNumber block; Buffer buffer, newbuf, vmbuffer = InvalidBuffer, @@ -2753,10 +2752,10 @@ heap_update(Relation relation, ItemPointer otid, HeapTuple newtup, page = BufferGetPage(buffer); /* - * Before locking the buffer, pin the visibility map page if it appears - * to be necessary. Since we haven't got the lock yet, someone else might - * be in the middle of changing this, so we'll need to recheck after - * we have the lock. + * Before locking the buffer, pin the visibility map page if it appears to + * be necessary. Since we haven't got the lock yet, someone else might be + * in the middle of changing this, so we'll need to recheck after we have + * the lock. */ if (PageIsAllVisible(page)) visibilitymap_pin(relation, block, &vmbuffer); @@ -2900,11 +2899,11 @@ l2: /* * If we didn't pin the visibility map page and the page has become all - * visible while we were busy locking the buffer, or during some subsequent - * window during which we had it unlocked, we'll have to unlock and - * re-lock, to avoid holding the buffer lock across an I/O. That's a bit - * unfortunate, esepecially since we'll now have to recheck whether the - * tuple has been locked or updated under us, but hopefully it won't + * visible while we were busy locking the buffer, or during some + * subsequent window during which we had it unlocked, we'll have to unlock + * and re-lock, to avoid holding the buffer lock across an I/O. That's a + * bit unfortunate, esepecially since we'll now have to recheck whether + * the tuple has been locked or updated under us, but hopefully it won't * happen very often. */ if (vmbuffer == InvalidBuffer && PageIsAllVisible(page)) @@ -3196,11 +3195,11 @@ l2: /* * Mark old tuple for invalidation from system caches at next command - * boundary, and mark the new tuple for invalidation in case we abort. - * We have to do this before releasing the buffer because oldtup is in - * the buffer. (heaptup is all in local memory, but it's necessary to - * process both tuple versions in one call to inval.c so we can avoid - * redundant sinval messages.) + * boundary, and mark the new tuple for invalidation in case we abort. We + * have to do this before releasing the buffer because oldtup is in the + * buffer. (heaptup is all in local memory, but it's necessary to process + * both tuple versions in one call to inval.c so we can avoid redundant + * sinval messages.) */ CacheInvalidateHeapTuple(relation, &oldtup, heaptup); @@ -4069,7 +4068,7 @@ heap_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid) */ bool heap_tuple_needs_freeze(HeapTupleHeader tuple, TransactionId cutoff_xid, - Buffer buf) + Buffer buf) { TransactionId xid; @@ -4368,9 +4367,9 @@ log_heap_freeze(Relation reln, Buffer buffer, } /* - * Perform XLogInsert for a heap-visible operation. 'block' is the block + * Perform XLogInsert for a heap-visible operation. 'block' is the block * being marked all-visible, and vm_buffer is the buffer containing the - * corresponding visibility map block. Both should have already been modified + * corresponding visibility map block. Both should have already been modified * and dirtied. */ XLogRecPtr @@ -4705,7 +4704,7 @@ heap_xlog_visible(XLogRecPtr lsn, XLogRecord *record) Page page; /* - * Read the heap page, if it still exists. If the heap file has been + * Read the heap page, if it still exists. If the heap file has been * dropped or truncated later in recovery, this might fail. In that case, * there's no point in doing anything further, since the visibility map * will have to be cleared out at the same time. @@ -4731,17 +4730,16 @@ heap_xlog_visible(XLogRecPtr lsn, XLogRecord *record) LockBuffer(buffer, BUFFER_LOCK_EXCLUSIVE); /* - * We don't bump the LSN of the heap page when setting the visibility - * map bit, because that would generate an unworkable volume of - * full-page writes. This exposes us to torn page hazards, but since - * we're not inspecting the existing page contents in any way, we - * don't care. + * We don't bump the LSN of the heap page when setting the visibility map + * bit, because that would generate an unworkable volume of full-page + * writes. This exposes us to torn page hazards, but since we're not + * inspecting the existing page contents in any way, we don't care. * - * However, all operations that clear the visibility map bit *do* bump - * the LSN, and those operations will only be replayed if the XLOG LSN - * follows the page LSN. Thus, if the page LSN has advanced past our - * XLOG record's LSN, we mustn't mark the page all-visible, because - * the subsequent update won't be replayed to clear the flag. + * However, all operations that clear the visibility map bit *do* bump the + * LSN, and those operations will only be replayed if the XLOG LSN follows + * the page LSN. Thus, if the page LSN has advanced past our XLOG + * record's LSN, we mustn't mark the page all-visible, because the + * subsequent update won't be replayed to clear the flag. */ if (!XLByteLE(lsn, PageGetLSN(page))) { @@ -4772,10 +4770,10 @@ heap_xlog_visible(XLogRecPtr lsn, XLogRecord *record) * Don't set the bit if replay has already passed this point. * * It might be safe to do this unconditionally; if replay has past - * this point, we'll replay at least as far this time as we did before, - * and if this bit needs to be cleared, the record responsible for - * doing so should be again replayed, and clear it. For right now, - * out of an abundance of conservatism, we use the same test here + * this point, we'll replay at least as far this time as we did + * before, and if this bit needs to be cleared, the record responsible + * for doing so should be again replayed, and clear it. For right + * now, out of an abundance of conservatism, we use the same test here * we did for the heap page; if this results in a dropped bit, no real * harm is done; and the next VACUUM will fix it. */ @@ -5183,7 +5181,7 @@ heap_xlog_update(XLogRecPtr lsn, XLogRecord *record, bool hot_update) if (xlrec->all_visible_cleared) { Relation reln = CreateFakeRelcacheEntry(xlrec->target.node); - BlockNumber block = ItemPointerGetBlockNumber(&xlrec->target.tid); + BlockNumber block = ItemPointerGetBlockNumber(&xlrec->target.tid); Buffer vmbuffer = InvalidBuffer; visibilitymap_pin(reln, block, &vmbuffer); @@ -5267,7 +5265,7 @@ newt:; if (xlrec->new_all_visible_cleared) { Relation reln = CreateFakeRelcacheEntry(xlrec->target.node); - BlockNumber block = ItemPointerGetBlockNumber(&xlrec->newtid); + BlockNumber block = ItemPointerGetBlockNumber(&xlrec->newtid); Buffer vmbuffer = InvalidBuffer; visibilitymap_pin(reln, block, &vmbuffer); @@ -5690,7 +5688,7 @@ heap2_desc(StringInfo buf, uint8 xl_info, char *rec) else appendStringInfo(buf, "multi-insert: "); appendStringInfo(buf, "rel %u/%u/%u; blk %u; %d tuples", - xlrec->node.spcNode, xlrec->node.dbNode, xlrec->node.relNode, + xlrec->node.spcNode, xlrec->node.dbNode, xlrec->node.relNode, xlrec->blkno, xlrec->ntuples); } else diff --git a/src/backend/access/heap/hio.c b/src/backend/access/heap/hio.c index 30ef1bf7e04efc8571a618515e234ccfdf4766bf..19a34923c7a74ddbf09d073ffe09f7cb5893cf80 100644 --- a/src/backend/access/heap/hio.c +++ b/src/backend/access/heap/hio.c @@ -109,8 +109,8 @@ GetVisibilityMapPins(Relation relation, Buffer buffer1, Buffer buffer2, BlockNumber block1, BlockNumber block2, Buffer *vmbuffer1, Buffer *vmbuffer2) { - bool need_to_pin_buffer1; - bool need_to_pin_buffer2; + bool need_to_pin_buffer1; + bool need_to_pin_buffer2; Assert(BufferIsValid(buffer1)); Assert(buffer2 == InvalidBuffer || buffer1 <= buffer2); @@ -145,7 +145,7 @@ GetVisibilityMapPins(Relation relation, Buffer buffer1, Buffer buffer2, /* * If there are two buffers involved and we pinned just one of them, * it's possible that the second one became all-visible while we were - * busy pinning the first one. If it looks like that's a possible + * busy pinning the first one. If it looks like that's a possible * scenario, we'll need to make a second pass through this loop. */ if (buffer2 == InvalidBuffer || buffer1 == buffer2 @@ -302,11 +302,11 @@ RelationGetBufferForTuple(Relation relation, Size len, * block if one was given, taking suitable care with lock ordering and * the possibility they are the same block. * - * If the page-level all-visible flag is set, caller will need to clear - * both that and the corresponding visibility map bit. However, by the - * time we return, we'll have x-locked the buffer, and we don't want to - * do any I/O while in that state. So we check the bit here before - * taking the lock, and pin the page if it appears necessary. + * If the page-level all-visible flag is set, caller will need to + * clear both that and the corresponding visibility map bit. However, + * by the time we return, we'll have x-locked the buffer, and we don't + * want to do any I/O while in that state. So we check the bit here + * before taking the lock, and pin the page if it appears necessary. * Checking without the lock creates a risk of getting the wrong * answer, so we'll have to recheck after acquiring the lock. */ @@ -347,23 +347,24 @@ RelationGetBufferForTuple(Relation relation, Size len, /* * We now have the target page (and the other buffer, if any) pinned - * and locked. However, since our initial PageIsAllVisible checks - * were performed before acquiring the lock, the results might now - * be out of date, either for the selected victim buffer, or for the - * other buffer passed by the caller. In that case, we'll need to give - * up our locks, go get the pin(s) we failed to get earlier, and + * and locked. However, since our initial PageIsAllVisible checks + * were performed before acquiring the lock, the results might now be + * out of date, either for the selected victim buffer, or for the + * other buffer passed by the caller. In that case, we'll need to + * give up our locks, go get the pin(s) we failed to get earlier, and * re-lock. That's pretty painful, but hopefully shouldn't happen * often. * - * Note that there's a small possibility that we didn't pin the - * page above but still have the correct page pinned anyway, either - * because we've already made a previous pass through this loop, or - * because caller passed us the right page anyway. + * Note that there's a small possibility that we didn't pin the page + * above but still have the correct page pinned anyway, either because + * we've already made a previous pass through this loop, or because + * caller passed us the right page anyway. * * Note also that it's possible that by the time we get the pin and * retake the buffer locks, the visibility map bit will have been - * cleared by some other backend anyway. In that case, we'll have done - * a bit of extra work for no gain, but there's no real harm done. + * cleared by some other backend anyway. In that case, we'll have + * done a bit of extra work for no gain, but there's no real harm + * done. */ if (otherBuffer == InvalidBuffer || buffer <= otherBuffer) GetVisibilityMapPins(relation, buffer, otherBuffer, diff --git a/src/backend/access/heap/tuptoaster.c b/src/backend/access/heap/tuptoaster.c index 28b5a20ae706156eb80a898fd106f58d0e6d2c27..050f048a9b0458d9c7469d4ec30210c90a090abe 100644 --- a/src/backend/access/heap/tuptoaster.c +++ b/src/backend/access/heap/tuptoaster.c @@ -75,7 +75,7 @@ do { \ static void toast_delete_datum(Relation rel, Datum value); static Datum toast_save_datum(Relation rel, Datum value, - struct varlena *oldexternal, int options); + struct varlena * oldexternal, int options); static bool toastrel_valueid_exists(Relation toastrel, Oid valueid); static bool toastid_valueid_exists(Oid toastrelid, Oid valueid); static struct varlena *toast_fetch_datum(struct varlena * attr); @@ -1233,7 +1233,7 @@ toast_compress_datum(Datum value) */ static Datum toast_save_datum(Relation rel, Datum value, - struct varlena *oldexternal, int options) + struct varlena * oldexternal, int options) { Relation toastrel; Relation toastidx; @@ -1353,7 +1353,7 @@ toast_save_datum(Relation rel, Datum value, * those versions could easily reference the same toast value. * When we copy the second or later version of such a row, * reusing the OID will mean we select an OID that's already - * in the new toast table. Check for that, and if so, just + * in the new toast table. Check for that, and if so, just * fall through without writing the data again. * * While annoying and ugly-looking, this is a good thing diff --git a/src/backend/access/heap/visibilitymap.c b/src/backend/access/heap/visibilitymap.c index 9152c7d151150573fbdbd5bb8d7a8125da18adf6..eb5625906f12904b3248b9b0364998f03a625318 100644 --- a/src/backend/access/heap/visibilitymap.c +++ b/src/backend/access/heap/visibilitymap.c @@ -16,7 +16,7 @@ * visibilitymap_pin_ok - check whether correct map page is already pinned * visibilitymap_set - set a bit in a previously pinned page * visibilitymap_test - test if a bit is set - * visibilitymap_count - count number of bits set in visibility map + * visibilitymap_count - count number of bits set in visibility map * visibilitymap_truncate - truncate the visibility map * * NOTES @@ -27,7 +27,7 @@ * the sense that we make sure that whenever a bit is set, we know the * condition is true, but if a bit is not set, it might or might not be true. * - * Clearing a visibility map bit is not separately WAL-logged. The callers + * Clearing a visibility map bit is not separately WAL-logged. The callers * must make sure that whenever a bit is cleared, the bit is cleared on WAL * replay of the updating operation as well. * @@ -36,9 +36,9 @@ * it may still be the case that every tuple on the page is visible to all * transactions; we just don't know that for certain. The difficulty is that * there are two bits which are typically set together: the PD_ALL_VISIBLE bit - * on the page itself, and the visibility map bit. If a crash occurs after the + * on the page itself, and the visibility map bit. If a crash occurs after the * visibility map page makes it to disk and before the updated heap page makes - * it to disk, redo must set the bit on the heap page. Otherwise, the next + * it to disk, redo must set the bit on the heap page. Otherwise, the next * insert, update, or delete on the heap page will fail to realize that the * visibility map bit must be cleared, possibly causing index-only scans to * return wrong answers. @@ -59,10 +59,10 @@ * the buffer lock over any I/O that may be required to read in the visibility * map page. To avoid this, we examine the heap page before locking it; * if the page-level PD_ALL_VISIBLE bit is set, we pin the visibility map - * bit. Then, we lock the buffer. But this creates a race condition: there + * bit. Then, we lock the buffer. But this creates a race condition: there * is a possibility that in the time it takes to lock the buffer, the * PD_ALL_VISIBLE bit gets set. If that happens, we have to unlock the - * buffer, pin the visibility map page, and relock the buffer. This shouldn't + * buffer, pin the visibility map page, and relock the buffer. This shouldn't * happen often, because only VACUUM currently sets visibility map bits, * and the race will only occur if VACUUM processes a given page at almost * exactly the same time that someone tries to further modify it. @@ -227,9 +227,9 @@ visibilitymap_pin_ok(BlockNumber heapBlk, Buffer buf) * visibilitymap_set - set a bit on a previously pinned page * * recptr is the LSN of the XLOG record we're replaying, if we're in recovery, - * or InvalidXLogRecPtr in normal running. The page LSN is advanced to the + * or InvalidXLogRecPtr in normal running. The page LSN is advanced to the * one provided; in normal running, we generate a new XLOG record and set the - * page LSN to that value. cutoff_xid is the largest xmin on the page being + * page LSN to that value. cutoff_xid is the largest xmin on the page being * marked all-visible; it is needed for Hot Standby, and can be * InvalidTransactionId if the page contains no tuples. * @@ -295,10 +295,10 @@ visibilitymap_set(Relation rel, BlockNumber heapBlk, XLogRecPtr recptr, * releasing *buf after it's done testing and setting bits. * * NOTE: This function is typically called without a lock on the heap page, - * so somebody else could change the bit just after we look at it. In fact, + * so somebody else could change the bit just after we look at it. In fact, * since we don't lock the visibility map page either, it's even possible that * someone else could have changed the bit just before we look at it, but yet - * we might see the old value. It is the caller's responsibility to deal with + * we might see the old value. It is the caller's responsibility to deal with * all concurrency issues! */ bool @@ -344,7 +344,7 @@ visibilitymap_test(Relation rel, BlockNumber heapBlk, Buffer *buf) } /* - * visibilitymap_count - count number of bits set in visibility map + * visibilitymap_count - count number of bits set in visibility map * * Note: we ignore the possibility of race conditions when the table is being * extended concurrently with the call. New pages added to the table aren't @@ -356,16 +356,16 @@ visibilitymap_count(Relation rel) BlockNumber result = 0; BlockNumber mapBlock; - for (mapBlock = 0; ; mapBlock++) + for (mapBlock = 0;; mapBlock++) { Buffer mapBuffer; unsigned char *map; int i; /* - * Read till we fall off the end of the map. We assume that any - * extra bytes in the last page are zeroed, so we don't bother - * excluding them from the count. + * Read till we fall off the end of the map. We assume that any extra + * bytes in the last page are zeroed, so we don't bother excluding + * them from the count. */ mapBuffer = vm_readbuf(rel, mapBlock, false); if (!BufferIsValid(mapBuffer)) @@ -496,11 +496,11 @@ vm_readbuf(Relation rel, BlockNumber blkno, bool extend) Buffer buf; /* - * We might not have opened the relation at the smgr level yet, or we might - * have been forced to close it by a sinval message. The code below won't - * necessarily notice relation extension immediately when extend = false, - * so we rely on sinval messages to ensure that our ideas about the size of - * the map aren't too far out of date. + * We might not have opened the relation at the smgr level yet, or we + * might have been forced to close it by a sinval message. The code below + * won't necessarily notice relation extension immediately when extend = + * false, so we rely on sinval messages to ensure that our ideas about the + * size of the map aren't too far out of date. */ RelationOpenSmgr(rel); diff --git a/src/backend/access/index/genam.c b/src/backend/access/index/genam.c index d54b669bf359d5f005d8b5131fe8e91ffa04dd82..26fd9b6e11438eb67b1cb234a9cb2ebbb540cc97 100644 --- a/src/backend/access/index/genam.c +++ b/src/backend/access/index/genam.c @@ -93,7 +93,7 @@ RelationGetIndexScan(Relation indexRelation, int nkeys, int norderbys) else scan->orderByData = NULL; - scan->xs_want_itup = false; /* may be set later */ + scan->xs_want_itup = false; /* may be set later */ /* * During recovery we ignore killed tuples and don't bother to kill them diff --git a/src/backend/access/index/indexam.c b/src/backend/access/index/indexam.c index 16ac4e1b9ffcae0a7d902297a1288a82f02295fb..d64df319c5cac37ac71fd3cf436c6f635cd4bb1e 100644 --- a/src/backend/access/index/indexam.c +++ b/src/backend/access/index/indexam.c @@ -435,7 +435,7 @@ index_restrpos(IndexScanDesc scan) ItemPointer index_getnext_tid(IndexScanDesc scan, ScanDirection direction) { - FmgrInfo *procedure; + FmgrInfo *procedure; bool found; SCAN_CHECKS; @@ -495,7 +495,7 @@ index_getnext_tid(IndexScanDesc scan, ScanDirection direction) HeapTuple index_fetch_heap(IndexScanDesc scan) { - ItemPointer tid = &scan->xs_ctup.t_self; + ItemPointer tid = &scan->xs_ctup.t_self; bool all_dead = false; bool got_heap_tuple; @@ -530,8 +530,8 @@ index_fetch_heap(IndexScanDesc scan) if (got_heap_tuple) { /* - * Only in a non-MVCC snapshot can more than one member of the - * HOT chain be visible. + * Only in a non-MVCC snapshot can more than one member of the HOT + * chain be visible. */ scan->xs_continue_hot = !IsMVCCSnapshot(scan->xs_snapshot); pgstat_count_heap_fetch(scan->indexRelation); @@ -544,7 +544,7 @@ index_fetch_heap(IndexScanDesc scan) /* * If we scanned a whole HOT chain and found only dead tuples, tell index * AM to kill its entry for that TID (this will take effect in the next - * amgettuple call, in index_getnext_tid). We do not do this when in + * amgettuple call, in index_getnext_tid). We do not do this when in * recovery because it may violate MVCC to do so. See comments in * RelationGetIndexScan(). */ diff --git a/src/backend/access/nbtree/nbtcompare.c b/src/backend/access/nbtree/nbtcompare.c index fedde934a389921e5c01deb672de79f22d9dd05b..d610bef79830cb35bd5cc89c48eabd6aa7433d4c 100644 --- a/src/backend/access/nbtree/nbtcompare.c +++ b/src/backend/access/nbtree/nbtcompare.c @@ -82,7 +82,7 @@ btint2fastcmp(Datum x, Datum y, SortSupport ssup) Datum btint2sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btint2fastcmp; PG_RETURN_VOID(); @@ -119,7 +119,7 @@ btint4fastcmp(Datum x, Datum y, SortSupport ssup) Datum btint4sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btint4fastcmp; PG_RETURN_VOID(); @@ -156,7 +156,7 @@ btint8fastcmp(Datum x, Datum y, SortSupport ssup) Datum btint8sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btint8fastcmp; PG_RETURN_VOID(); @@ -277,7 +277,7 @@ btoidfastcmp(Datum x, Datum y, SortSupport ssup) Datum btoidsortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btoidfastcmp; PG_RETURN_VOID(); @@ -338,7 +338,7 @@ btnamefastcmp(Datum x, Datum y, SortSupport ssup) Datum btnamesortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btnamefastcmp; PG_RETURN_VOID(); diff --git a/src/backend/access/nbtree/nbtpage.c b/src/backend/access/nbtree/nbtpage.c index e6dec618c7781b7a454b899954be345593864f79..016ce2283c2655ec3e25178ebe6d40ad7bf9c02a 100644 --- a/src/backend/access/nbtree/nbtpage.c +++ b/src/backend/access/nbtree/nbtpage.c @@ -1362,7 +1362,7 @@ _bt_pagedel(Relation rel, Buffer buf, BTStack stack) * we're in VACUUM and would not otherwise have an XID. Having already * updated links to the target, ReadNewTransactionId() suffices as an * upper bound. Any scan having retained a now-stale link is advertising - * in its PGXACT an xmin less than or equal to the value we read here. It + * in its PGXACT an xmin less than or equal to the value we read here. It * will continue to do so, holding back RecentGlobalXmin, for the duration * of that scan. */ diff --git a/src/backend/access/nbtree/nbtree.c b/src/backend/access/nbtree/nbtree.c index 184fc3bb79b234a5423a06af253cbe836ab4c14d..41d06edb15487130104a19457b96dacd7ccc0331 100644 --- a/src/backend/access/nbtree/nbtree.c +++ b/src/backend/access/nbtree/nbtree.c @@ -433,7 +433,7 @@ btbeginscan(PG_FUNCTION_ARGS) /* * We don't know yet whether the scan will be index-only, so we do not - * allocate the tuple workspace arrays until btrescan. However, we set up + * allocate the tuple workspace arrays until btrescan. However, we set up * scan->xs_itupdesc whether we'll need it or not, since that's so cheap. */ so->currTuples = so->markTuples = NULL; @@ -478,7 +478,7 @@ btrescan(PG_FUNCTION_ARGS) /* * Allocate tuple workspace arrays, if needed for an index-only scan and - * not already done in a previous rescan call. To save on palloc + * not already done in a previous rescan call. To save on palloc * overhead, both workspaces are allocated as one palloc block; only this * function and btendscan know that. * diff --git a/src/backend/access/nbtree/nbtsearch.c b/src/backend/access/nbtree/nbtsearch.c index b701c3f819cb0dd76b9ef75fadd6086796198b37..e0c952368b9c1c72712f04334326a9c9de40ec68 100644 --- a/src/backend/access/nbtree/nbtsearch.c +++ b/src/backend/access/nbtree/nbtsearch.c @@ -564,11 +564,11 @@ _bt_first(IndexScanDesc scan, ScanDirection dir) ScanKeyEntryInitialize(chosen, (SK_SEARCHNOTNULL | SK_ISNULL | (impliesNN->sk_flags & - (SK_BT_DESC | SK_BT_NULLS_FIRST))), + (SK_BT_DESC | SK_BT_NULLS_FIRST))), curattr, - ((impliesNN->sk_flags & SK_BT_NULLS_FIRST) ? - BTGreaterStrategyNumber : - BTLessStrategyNumber), + ((impliesNN->sk_flags & SK_BT_NULLS_FIRST) ? + BTGreaterStrategyNumber : + BTLessStrategyNumber), InvalidOid, InvalidOid, InvalidOid, diff --git a/src/backend/access/nbtree/nbtutils.c b/src/backend/access/nbtree/nbtutils.c index f79ce552b6206005b4a24ab92ad06183d603c146..33ad8915f5a7982e6a7cef3216f5bffeb1e90894 100644 --- a/src/backend/access/nbtree/nbtutils.c +++ b/src/backend/access/nbtree/nbtutils.c @@ -37,10 +37,10 @@ typedef struct BTSortArrayContext static Datum _bt_find_extreme_element(IndexScanDesc scan, ScanKey skey, StrategyNumber strat, Datum *elems, int nelems); -static int _bt_sort_array_elements(IndexScanDesc scan, ScanKey skey, +static int _bt_sort_array_elements(IndexScanDesc scan, ScanKey skey, bool reverse, Datum *elems, int nelems); -static int _bt_compare_array_elements(const void *a, const void *b, void *arg); +static int _bt_compare_array_elements(const void *a, const void *b, void *arg); static bool _bt_compare_scankey_args(IndexScanDesc scan, ScanKey op, ScanKey leftarg, ScanKey rightarg, bool *result); @@ -227,8 +227,8 @@ _bt_preprocess_array_keys(IndexScanDesc scan) } /* - * Make a scan-lifespan context to hold array-associated data, or reset - * it if we already have one from a previous rescan cycle. + * Make a scan-lifespan context to hold array-associated data, or reset it + * if we already have one from a previous rescan cycle. */ if (so->arrayContext == NULL) so->arrayContext = AllocSetContextCreate(CurrentMemoryContext, @@ -269,7 +269,7 @@ _bt_preprocess_array_keys(IndexScanDesc scan) continue; /* - * First, deconstruct the array into elements. Anything allocated + * First, deconstruct the array into elements. Anything allocated * here (including a possibly detoasted array value) is in the * workspace context. */ @@ -283,7 +283,7 @@ _bt_preprocess_array_keys(IndexScanDesc scan) &elem_values, &elem_nulls, &num_elems); /* - * Compress out any null elements. We can ignore them since we assume + * Compress out any null elements. We can ignore them since we assume * all btree operators are strict. */ num_nonnulls = 0; @@ -338,7 +338,7 @@ _bt_preprocess_array_keys(IndexScanDesc scan) * successive primitive indexscans produce data in index order. */ num_elems = _bt_sort_array_elements(scan, cur, - (indoption[cur->sk_attno - 1] & INDOPTION_DESC) != 0, + (indoption[cur->sk_attno - 1] & INDOPTION_DESC) != 0, elem_values, num_nonnulls); /* @@ -387,9 +387,10 @@ _bt_find_extreme_element(IndexScanDesc scan, ScanKey skey, /* * Look up the appropriate comparison operator in the opfamily. * - * Note: it's possible that this would fail, if the opfamily is incomplete, - * but it seems quite unlikely that an opfamily would omit non-cross-type - * comparison operators for any datatype that it supports at all. + * Note: it's possible that this would fail, if the opfamily is + * incomplete, but it seems quite unlikely that an opfamily would omit + * non-cross-type comparison operators for any datatype that it supports + * at all. */ cmp_op = get_opfamily_member(rel->rd_opfamily[skey->sk_attno - 1], elemtype, @@ -455,9 +456,10 @@ _bt_sort_array_elements(IndexScanDesc scan, ScanKey skey, /* * Look up the appropriate comparison function in the opfamily. * - * Note: it's possible that this would fail, if the opfamily is incomplete, - * but it seems quite unlikely that an opfamily would omit non-cross-type - * support functions for any datatype that it supports at all. + * Note: it's possible that this would fail, if the opfamily is + * incomplete, but it seems quite unlikely that an opfamily would omit + * non-cross-type support functions for any datatype that it supports at + * all. */ cmp_proc = get_opfamily_proc(rel->rd_opfamily[skey->sk_attno - 1], elemtype, @@ -515,7 +517,7 @@ _bt_compare_array_elements(const void *a, const void *b, void *arg) * _bt_start_array_keys() -- Initialize array keys at start of a scan * * Set up the cur_elem counters and fill in the first sk_argument value for - * each array scankey. We can't do this until we know the scan direction. + * each array scankey. We can't do this until we know the scan direction. */ void _bt_start_array_keys(IndexScanDesc scan, ScanDirection dir) @@ -609,8 +611,8 @@ _bt_advance_array_keys(IndexScanDesc scan, ScanDirection dir) * so that the index sorts in the desired direction. * * One key purpose of this routine is to discover which scan keys must be - * satisfied to continue the scan. It also attempts to eliminate redundant - * keys and detect contradictory keys. (If the index opfamily provides + * satisfied to continue the scan. It also attempts to eliminate redundant + * keys and detect contradictory keys. (If the index opfamily provides * incomplete sets of cross-type operators, we may fail to detect redundant * or contradictory keys, but we can survive that.) * @@ -676,7 +678,7 @@ _bt_advance_array_keys(IndexScanDesc scan, ScanDirection dir) * Note: the reason we have to copy the preprocessed scan keys into private * storage is that we are modifying the array based on comparisons of the * key argument values, which could change on a rescan or after moving to - * new elements of array keys. Therefore we can't overwrite the source data. + * new elements of array keys. Therefore we can't overwrite the source data. */ void _bt_preprocess_keys(IndexScanDesc scan) @@ -781,8 +783,8 @@ _bt_preprocess_keys(IndexScanDesc scan) * set qual_ok to false and abandon further processing. * * We also have to deal with the case of "key IS NULL", which is - * unsatisfiable in combination with any other index condition. - * By the time we get here, that's been classified as an equality + * unsatisfiable in combination with any other index condition. By + * the time we get here, that's been classified as an equality * check, and we've rejected any combination of it with a regular * equality condition; but not with other types of conditions. */ @@ -1421,12 +1423,12 @@ _bt_checkkeys(IndexScanDesc scan, /* * Since NULLs are sorted before non-NULLs, we know we have * reached the lower limit of the range of values for this - * index attr. On a backward scan, we can stop if this qual + * index attr. On a backward scan, we can stop if this qual * is one of the "must match" subset. We can stop regardless * of whether the qual is > or <, so long as it's required, - * because it's not possible for any future tuples to pass. - * On a forward scan, however, we must keep going, because we - * may have initially positioned to the start of the index. + * because it's not possible for any future tuples to pass. On + * a forward scan, however, we must keep going, because we may + * have initially positioned to the start of the index. */ if ((key->sk_flags & (SK_BT_REQFWD | SK_BT_REQBKWD)) && ScanDirectionIsBackward(dir)) @@ -1437,11 +1439,11 @@ _bt_checkkeys(IndexScanDesc scan, /* * Since NULLs are sorted after non-NULLs, we know we have * reached the upper limit of the range of values for this - * index attr. On a forward scan, we can stop if this qual is - * one of the "must match" subset. We can stop regardless of + * index attr. On a forward scan, we can stop if this qual is + * one of the "must match" subset. We can stop regardless of * whether the qual is > or <, so long as it's required, - * because it's not possible for any future tuples to pass. - * On a backward scan, however, we must keep going, because we + * because it's not possible for any future tuples to pass. On + * a backward scan, however, we must keep going, because we * may have initially positioned to the end of the index. */ if ((key->sk_flags & (SK_BT_REQFWD | SK_BT_REQBKWD)) && @@ -1532,12 +1534,12 @@ _bt_check_rowcompare(ScanKey skey, IndexTuple tuple, TupleDesc tupdesc, /* * Since NULLs are sorted before non-NULLs, we know we have * reached the lower limit of the range of values for this - * index attr. On a backward scan, we can stop if this qual + * index attr. On a backward scan, we can stop if this qual * is one of the "must match" subset. We can stop regardless * of whether the qual is > or <, so long as it's required, - * because it's not possible for any future tuples to pass. - * On a forward scan, however, we must keep going, because we - * may have initially positioned to the start of the index. + * because it's not possible for any future tuples to pass. On + * a forward scan, however, we must keep going, because we may + * have initially positioned to the start of the index. */ if ((subkey->sk_flags & (SK_BT_REQFWD | SK_BT_REQBKWD)) && ScanDirectionIsBackward(dir)) @@ -1548,11 +1550,11 @@ _bt_check_rowcompare(ScanKey skey, IndexTuple tuple, TupleDesc tupdesc, /* * Since NULLs are sorted after non-NULLs, we know we have * reached the upper limit of the range of values for this - * index attr. On a forward scan, we can stop if this qual is - * one of the "must match" subset. We can stop regardless of + * index attr. On a forward scan, we can stop if this qual is + * one of the "must match" subset. We can stop regardless of * whether the qual is > or <, so long as it's required, - * because it's not possible for any future tuples to pass. - * On a backward scan, however, we must keep going, because we + * because it's not possible for any future tuples to pass. On + * a backward scan, however, we must keep going, because we * may have initially positioned to the end of the index. */ if ((subkey->sk_flags & (SK_BT_REQFWD | SK_BT_REQBKWD)) && diff --git a/src/backend/access/spgist/spgdoinsert.c b/src/backend/access/spgist/spgdoinsert.c index 98a7bea742ea87bb81611586df035a8d6c169ac2..b3f8f6a231372fb1f4b6e2c1d42f124b2ca215cb 100644 --- a/src/backend/access/spgist/spgdoinsert.c +++ b/src/backend/access/spgist/spgdoinsert.c @@ -24,7 +24,7 @@ /* * SPPageDesc tracks all info about a page we are inserting into. In some * situations it actually identifies a tuple, or even a specific node within - * an inner tuple. But any of the fields can be invalid. If the buffer + * an inner tuple. But any of the fields can be invalid. If the buffer * field is valid, it implies we hold pin and exclusive lock on that buffer. * page pointer should be valid exactly when buffer is. */ @@ -129,8 +129,8 @@ spgPageIndexMultiDelete(SpGistState *state, Page page, int firststate, int reststate, BlockNumber blkno, OffsetNumber offnum) { - OffsetNumber firstItem; - OffsetNumber *sortednos; + OffsetNumber firstItem; + OffsetNumber *sortednos; SpGistDeadTuple tuple = NULL; int i; @@ -155,8 +155,8 @@ spgPageIndexMultiDelete(SpGistState *state, Page page, for (i = 0; i < nitems; i++) { - OffsetNumber itemno = sortednos[i]; - int tupstate; + OffsetNumber itemno = sortednos[i]; + int tupstate; tupstate = (itemno == firstItem) ? firststate : reststate; if (tuple == NULL || tuple->tupstate != tupstate) @@ -200,7 +200,7 @@ saveNodeLink(Relation index, SPPageDesc *parent, */ static void addLeafTuple(Relation index, SpGistState *state, SpGistLeafTuple leafTuple, - SPPageDesc *current, SPPageDesc *parent, bool isNulls, bool isNew) + SPPageDesc *current, SPPageDesc *parent, bool isNulls, bool isNew) { XLogRecData rdata[4]; spgxlogAddLeaf xlrec; @@ -230,7 +230,7 @@ addLeafTuple(Relation index, SpGistState *state, SpGistLeafTuple leafTuple, /* Tuple is not part of a chain */ leafTuple->nextOffset = InvalidOffsetNumber; current->offnum = SpGistPageAddNewItem(state, current->page, - (Item) leafTuple, leafTuple->size, + (Item) leafTuple, leafTuple->size, NULL, false); xlrec.offnumLeaf = current->offnum; @@ -250,9 +250,9 @@ addLeafTuple(Relation index, SpGistState *state, SpGistLeafTuple leafTuple, else { /* - * Tuple must be inserted into existing chain. We mustn't change - * the chain's head address, but we don't need to chase the entire - * chain to put the tuple at the end; we can insert it second. + * Tuple must be inserted into existing chain. We mustn't change the + * chain's head address, but we don't need to chase the entire chain + * to put the tuple at the end; we can insert it second. * * Also, it's possible that the "chain" consists only of a DEAD tuple, * in which case we should replace the DEAD tuple in-place. @@ -261,7 +261,7 @@ addLeafTuple(Relation index, SpGistState *state, SpGistLeafTuple leafTuple, OffsetNumber offnum; head = (SpGistLeafTuple) PageGetItem(current->page, - PageGetItemId(current->page, current->offnum)); + PageGetItemId(current->page, current->offnum)); if (head->tupstate == SPGIST_LIVE) { leafTuple->nextOffset = head->nextOffset; @@ -274,7 +274,7 @@ addLeafTuple(Relation index, SpGistState *state, SpGistLeafTuple leafTuple, * and set new second element */ head = (SpGistLeafTuple) PageGetItem(current->page, - PageGetItemId(current->page, current->offnum)); + PageGetItemId(current->page, current->offnum)); head->nextOffset = offnum; xlrec.offnumLeaf = offnum; @@ -483,7 +483,7 @@ moveLeafs(Relation index, SpGistState *state, for (i = 0; i < nDelete; i++) { it = (SpGistLeafTuple) PageGetItem(current->page, - PageGetItemId(current->page, toDelete[i])); + PageGetItemId(current->page, toDelete[i])); Assert(it->tupstate == SPGIST_LIVE); /* @@ -516,12 +516,12 @@ moveLeafs(Relation index, SpGistState *state, leafptr += newLeafTuple->size; /* - * Now delete the old tuples, leaving a redirection pointer behind for - * the first one, unless we're doing an index build; in which case there - * can't be any concurrent scan so we need not provide a redirect. + * Now delete the old tuples, leaving a redirection pointer behind for the + * first one, unless we're doing an index build; in which case there can't + * be any concurrent scan so we need not provide a redirect. */ spgPageIndexMultiDelete(state, current->page, toDelete, nDelete, - state->isBuild ? SPGIST_PLACEHOLDER : SPGIST_REDIRECT, + state->isBuild ? SPGIST_PLACEHOLDER : SPGIST_REDIRECT, SPGIST_PLACEHOLDER, nblkno, r); @@ -575,7 +575,7 @@ setRedirectionTuple(SPPageDesc *current, OffsetNumber position, SpGistDeadTuple dt; dt = (SpGistDeadTuple) PageGetItem(current->page, - PageGetItemId(current->page, position)); + PageGetItemId(current->page, position)); Assert(dt->tupstate == SPGIST_REDIRECT); Assert(ItemPointerGetBlockNumber(&dt->pointer) == SPGIST_METAPAGE_BLKNO); ItemPointerSet(&dt->pointer, blkno, offnum); @@ -640,7 +640,7 @@ checkAllTheSame(spgPickSplitIn *in, spgPickSplitOut *out, bool tooBig, /* The opclass may not use node labels, but if it does, duplicate 'em */ if (out->nodeLabels) { - Datum theLabel = out->nodeLabels[theNode]; + Datum theLabel = out->nodeLabels[theNode]; out->nodeLabels = (Datum *) palloc(sizeof(Datum) * out->nNodes); for (i = 0; i < out->nNodes; i++) @@ -754,8 +754,8 @@ doPickSplit(Relation index, SpGistState *state, { /* * We are splitting the root (which up to now is also a leaf page). - * Its tuples are not linked, so scan sequentially to get them all. - * We ignore the original value of current->offnum. + * Its tuples are not linked, so scan sequentially to get them all. We + * ignore the original value of current->offnum. */ for (i = FirstOffsetNumber; i <= max; i++) { @@ -773,7 +773,7 @@ doPickSplit(Relation index, SpGistState *state, /* we will delete the tuple altogether, so count full space */ spaceToDelete += it->size + sizeof(ItemIdData); } - else /* tuples on root should be live */ + else /* tuples on root should be live */ elog(ERROR, "unexpected SPGiST tuple state: %d", it->tupstate); } } @@ -820,7 +820,7 @@ doPickSplit(Relation index, SpGistState *state, * We may not actually insert new tuple because another picksplit may be * necessary due to too large value, but we will try to allocate enough * space to include it; and in any case it has to be included in the input - * for the picksplit function. So don't increment nToInsert yet. + * for the picksplit function. So don't increment nToInsert yet. */ in.datums[in.nTuples] = SGLTDATUM(newLeafTuple, state); heapPtrs[in.nTuples] = newLeafTuple->heapPtr; @@ -878,7 +878,7 @@ doPickSplit(Relation index, SpGistState *state, /* * Check to see if the picksplit function failed to separate the values, * ie, it put them all into the same child node. If so, select allTheSame - * mode and create a random split instead. See comments for + * mode and create a random split instead. See comments for * checkAllTheSame as to why we need to know if the new leaf tuples could * fit on one page. */ @@ -924,8 +924,8 @@ doPickSplit(Relation index, SpGistState *state, innerTuple->allTheSame = allTheSame; /* - * Update nodes[] array to point into the newly formed innerTuple, so - * that we can adjust their downlinks below. + * Update nodes[] array to point into the newly formed innerTuple, so that + * we can adjust their downlinks below. */ SGITITERATE(innerTuple, i, node) { @@ -944,13 +944,13 @@ doPickSplit(Relation index, SpGistState *state, } /* - * To perform the split, we must insert a new inner tuple, which can't - * go on a leaf page; and unless we are splitting the root page, we - * must then update the parent tuple's downlink to point to the inner - * tuple. If there is room, we'll put the new inner tuple on the same - * page as the parent tuple, otherwise we need another non-leaf buffer. - * But if the parent page is the root, we can't add the new inner tuple - * there, because the root page must have only one inner tuple. + * To perform the split, we must insert a new inner tuple, which can't go + * on a leaf page; and unless we are splitting the root page, we must then + * update the parent tuple's downlink to point to the inner tuple. If + * there is room, we'll put the new inner tuple on the same page as the + * parent tuple, otherwise we need another non-leaf buffer. But if the + * parent page is the root, we can't add the new inner tuple there, + * because the root page must have only one inner tuple. */ xlrec.initInner = false; if (parent->buffer != InvalidBuffer && @@ -965,9 +965,9 @@ doPickSplit(Relation index, SpGistState *state, { /* Send tuple to page with next triple parity (see README) */ newInnerBuffer = SpGistGetBuffer(index, - GBUF_INNER_PARITY(parent->blkno + 1) | + GBUF_INNER_PARITY(parent->blkno + 1) | (isNulls ? GBUF_NULLS : 0), - innerTuple->size + sizeof(ItemIdData), + innerTuple->size + sizeof(ItemIdData), &xlrec.initInner); } else @@ -977,22 +977,22 @@ doPickSplit(Relation index, SpGistState *state, } /* - * Because a WAL record can't involve more than four buffers, we can - * only afford to deal with two leaf pages in each picksplit action, - * ie the current page and at most one other. + * Because a WAL record can't involve more than four buffers, we can only + * afford to deal with two leaf pages in each picksplit action, ie the + * current page and at most one other. * - * The new leaf tuples converted from the existing ones should require - * the same or less space, and therefore should all fit onto one page + * The new leaf tuples converted from the existing ones should require the + * same or less space, and therefore should all fit onto one page * (although that's not necessarily the current page, since we can't * delete the old tuples but only replace them with placeholders). - * However, the incoming new tuple might not also fit, in which case - * we might need another picksplit cycle to reduce it some more. + * However, the incoming new tuple might not also fit, in which case we + * might need another picksplit cycle to reduce it some more. * - * If there's not room to put everything back onto the current page, - * then we decide on a per-node basis which tuples go to the new page. - * (We do it like that because leaf tuple chains can't cross pages, - * so we must place all leaf tuples belonging to the same parent node - * on the same page.) + * If there's not room to put everything back onto the current page, then + * we decide on a per-node basis which tuples go to the new page. (We do + * it like that because leaf tuple chains can't cross pages, so we must + * place all leaf tuples belonging to the same parent node on the same + * page.) * * If we are splitting the root page (turning it from a leaf page into an * inner page), then no leaf tuples can go back to the current page; they @@ -1037,12 +1037,13 @@ doPickSplit(Relation index, SpGistState *state, int newspace; newLeafBuffer = SpGistGetBuffer(index, - GBUF_LEAF | (isNulls ? GBUF_NULLS : 0), + GBUF_LEAF | (isNulls ? GBUF_NULLS : 0), Min(totalLeafSizes, SPGIST_PAGE_CAPACITY), &xlrec.initDest); + /* - * Attempt to assign node groups to the two pages. We might fail to + * Attempt to assign node groups to the two pages. We might fail to * do so, even if totalLeafSizes is less than the available space, * because we can't split a group across pages. */ @@ -1054,12 +1055,12 @@ doPickSplit(Relation index, SpGistState *state, { if (leafSizes[i] <= curspace) { - nodePageSelect[i] = 0; /* signifies current page */ + nodePageSelect[i] = 0; /* signifies current page */ curspace -= leafSizes[i]; } else { - nodePageSelect[i] = 1; /* signifies new leaf page */ + nodePageSelect[i] = 1; /* signifies new leaf page */ newspace -= leafSizes[i]; } } @@ -1075,7 +1076,7 @@ doPickSplit(Relation index, SpGistState *state, else if (includeNew) { /* We must exclude the new leaf tuple from the split */ - int nodeOfNewTuple = out.mapTuplesToNodes[in.nTuples - 1]; + int nodeOfNewTuple = out.mapTuplesToNodes[in.nTuples - 1]; leafSizes[nodeOfNewTuple] -= newLeafs[in.nTuples - 1]->size + sizeof(ItemIdData); @@ -1087,12 +1088,12 @@ doPickSplit(Relation index, SpGistState *state, { if (leafSizes[i] <= curspace) { - nodePageSelect[i] = 0; /* signifies current page */ + nodePageSelect[i] = 0; /* signifies current page */ curspace -= leafSizes[i]; } else { - nodePageSelect[i] = 1; /* signifies new leaf page */ + nodePageSelect[i] = 1; /* signifies new leaf page */ newspace -= leafSizes[i]; } } @@ -1204,7 +1205,7 @@ doPickSplit(Relation index, SpGistState *state, for (i = 0; i < nToInsert; i++) { SpGistLeafTuple it = newLeafs[i]; - Buffer leafBuffer; + Buffer leafBuffer; BlockNumber leafBlock; OffsetNumber newoffset; @@ -1584,12 +1585,12 @@ spgAddNodeAction(Relation index, SpGistState *state, xlrec.nodeI = parent->node; /* - * obtain new buffer with the same parity as current, since it will - * be a child of same parent tuple + * obtain new buffer with the same parity as current, since it will be + * a child of same parent tuple */ current->buffer = SpGistGetBuffer(index, GBUF_INNER_PARITY(current->blkno), - newInnerTuple->size + sizeof(ItemIdData), + newInnerTuple->size + sizeof(ItemIdData), &xlrec.newPage); current->blkno = BufferGetBlockNumber(current->buffer); current->page = BufferGetPage(current->buffer); @@ -1597,15 +1598,15 @@ spgAddNodeAction(Relation index, SpGistState *state, xlrec.blknoNew = current->blkno; /* - * Let's just make real sure new current isn't same as old. Right - * now that's impossible, but if SpGistGetBuffer ever got smart enough - * to delete placeholder tuples before checking space, maybe it - * wouldn't be impossible. The case would appear to work except that - * WAL replay would be subtly wrong, so I think a mere assert isn't - * enough here. + * Let's just make real sure new current isn't same as old. Right now + * that's impossible, but if SpGistGetBuffer ever got smart enough to + * delete placeholder tuples before checking space, maybe it wouldn't + * be impossible. The case would appear to work except that WAL + * replay would be subtly wrong, so I think a mere assert isn't enough + * here. */ - if (xlrec.blknoNew == xlrec.blkno) - elog(ERROR, "SPGiST new buffer shouldn't be same as old buffer"); + if (xlrec.blknoNew == xlrec.blkno) + elog(ERROR, "SPGiST new buffer shouldn't be same as old buffer"); /* * New current and parent buffer will both be modified; but note that @@ -1707,9 +1708,9 @@ spgSplitNodeAction(Relation index, SpGistState *state, Assert(!SpGistPageStoresNulls(current->page)); /* - * Construct new prefix tuple, containing a single node with the - * specified label. (We'll update the node's downlink to point to the - * new postfix tuple, below.) + * Construct new prefix tuple, containing a single node with the specified + * label. (We'll update the node's downlink to point to the new postfix + * tuple, below.) */ node = spgFormNodeTuple(state, out->result.splitTuple.nodeLabel, false); @@ -1888,9 +1889,9 @@ spgdoinsert(Relation index, SpGistState *state, (errcode(ERRCODE_PROGRAM_LIMIT_EXCEEDED), errmsg("index row size %lu exceeds maximum %lu for index \"%s\"", (unsigned long) (leafSize - sizeof(ItemIdData)), - (unsigned long) (SPGIST_PAGE_CAPACITY - sizeof(ItemIdData)), + (unsigned long) (SPGIST_PAGE_CAPACITY - sizeof(ItemIdData)), RelationGetRelationName(index)), - errhint("Values larger than a buffer page cannot be indexed."))); + errhint("Values larger than a buffer page cannot be indexed."))); /* Initialize "current" to the appropriate root page */ current.blkno = isnull ? SPGIST_NULL_BLKNO : SPGIST_ROOT_BLKNO; @@ -1920,7 +1921,7 @@ spgdoinsert(Relation index, SpGistState *state, if (current.blkno == InvalidBlockNumber) { /* - * Create a leaf page. If leafSize is too large to fit on a page, + * Create a leaf page. If leafSize is too large to fit on a page, * we won't actually use the page yet, but it simplifies the API * for doPickSplit to always have a leaf page at hand; so just * quietly limit our request to a page size. @@ -1968,7 +1969,7 @@ spgdoinsert(Relation index, SpGistState *state, } else if ((sizeToSplit = checkSplitConditions(index, state, ¤t, - &nToSplit)) < SPGIST_PAGE_CAPACITY / 2 && + &nToSplit)) < SPGIST_PAGE_CAPACITY / 2 && nToSplit < 64 && leafTuple->size + sizeof(ItemIdData) + sizeToSplit <= SPGIST_PAGE_CAPACITY) { @@ -2077,8 +2078,8 @@ spgdoinsert(Relation index, SpGistState *state, } /* - * Loop around and attempt to insert the new leafDatum - * at "current" (which might reference an existing child + * Loop around and attempt to insert the new leafDatum at + * "current" (which might reference an existing child * tuple, or might be invalid to force us to find a new * page for the tuple). * @@ -2102,8 +2103,8 @@ spgdoinsert(Relation index, SpGistState *state, out.result.addNode.nodeLabel); /* - * Retry insertion into the enlarged node. We assume - * that we'll get a MatchNode result this time. + * Retry insertion into the enlarged node. We assume that + * we'll get a MatchNode result this time. */ goto process_inner_tuple; break; diff --git a/src/backend/access/spgist/spginsert.c b/src/backend/access/spgist/spginsert.c index 8ff9245e179ac71063804b79db08cb9ca35a2c9e..456a71fbba5dabb557510167baaccccea4fa6c9b 100644 --- a/src/backend/access/spgist/spginsert.c +++ b/src/backend/access/spgist/spginsert.c @@ -123,7 +123,7 @@ spgbuild(PG_FUNCTION_ARGS) buildstate.spgstate.isBuild = true; buildstate.tmpCtx = AllocSetContextCreate(CurrentMemoryContext, - "SP-GiST build temporary context", + "SP-GiST build temporary context", ALLOCSET_DEFAULT_MINSIZE, ALLOCSET_DEFAULT_INITSIZE, ALLOCSET_DEFAULT_MAXSIZE); diff --git a/src/backend/access/spgist/spgkdtreeproc.c b/src/backend/access/spgist/spgkdtreeproc.c index adfe287581b8c101af79a461f176222b338dfecf..db472db9d6310cecef0cd6cb1c91977c1f1cb1d9 100644 --- a/src/backend/access/spgist/spgkdtreeproc.c +++ b/src/backend/access/spgist/spgkdtreeproc.c @@ -135,12 +135,12 @@ spg_kd_picksplit(PG_FUNCTION_ARGS) /* * Note: points that have coordinates exactly equal to coord may get - * classified into either node, depending on where they happen to fall - * in the sorted list. This is okay as long as the inner_consistent - * function descends into both sides for such cases. This is better - * than the alternative of trying to have an exact boundary, because - * it keeps the tree balanced even when we have many instances of the - * same point value. So we should never trigger the allTheSame logic. + * classified into either node, depending on where they happen to fall in + * the sorted list. This is okay as long as the inner_consistent function + * descends into both sides for such cases. This is better than the + * alternative of trying to have an exact boundary, because it keeps the + * tree balanced even when we have many instances of the same point value. + * So we should never trigger the allTheSame logic. */ for (i = 0; i < in->nTuples; i++) { diff --git a/src/backend/access/spgist/spgquadtreeproc.c b/src/backend/access/spgist/spgquadtreeproc.c index 10fafe58643b249ae6a32fee229c71b394e0e4eb..5da265025e6192cfc7ea397595c5dac820b42dd5 100644 --- a/src/backend/access/spgist/spgquadtreeproc.c +++ b/src/backend/access/spgist/spgquadtreeproc.c @@ -253,8 +253,8 @@ spg_quad_inner_consistent(PG_FUNCTION_ARGS) boxQuery = DatumGetBoxP(in->scankeys[i].sk_argument); if (DatumGetBool(DirectFunctionCall2(box_contain_pt, - PointerGetDatum(boxQuery), - PointerGetDatum(centroid)))) + PointerGetDatum(boxQuery), + PointerGetDatum(centroid)))) { /* centroid is in box, so all quadrants are OK */ } diff --git a/src/backend/access/spgist/spgscan.c b/src/backend/access/spgist/spgscan.c index 7a3a96230d176b6d13e78a771373a55c04802f4f..2a083b7c388beb3f00de872bec34252e04391f62 100644 --- a/src/backend/access/spgist/spgscan.c +++ b/src/backend/access/spgist/spgscan.c @@ -24,7 +24,7 @@ typedef void (*storeRes_func) (SpGistScanOpaque so, ItemPointer heapPtr, - Datum leafValue, bool isnull, bool recheck); + Datum leafValue, bool isnull, bool recheck); typedef struct ScanStackEntry { @@ -88,7 +88,7 @@ resetSpGistScanOpaque(SpGistScanOpaque so) if (so->want_itup) { /* Must pfree IndexTuples to avoid memory leak */ - int i; + int i; for (i = 0; i < so->nPtrs; i++) pfree(so->indexTups[i]); @@ -102,7 +102,7 @@ resetSpGistScanOpaque(SpGistScanOpaque so) * Sets searchNulls, searchNonNulls, numberOfKeys, keyData fields of *so. * * The point here is to eliminate null-related considerations from what the - * opclass consistent functions need to deal with. We assume all SPGiST- + * opclass consistent functions need to deal with. We assume all SPGiST- * indexable operators are strict, so any null RHS value makes the scan * condition unsatisfiable. We also pull out any IS NULL/IS NOT NULL * conditions; their effect is reflected into searchNulls/searchNonNulls. @@ -177,6 +177,7 @@ spgbeginscan(PG_FUNCTION_ARGS) { Relation rel = (Relation) PG_GETARG_POINTER(0); int keysz = PG_GETARG_INT32(1); + /* ScanKey scankey = (ScanKey) PG_GETARG_POINTER(2); */ IndexScanDesc scan; SpGistScanOpaque so; @@ -457,7 +458,7 @@ redirect: MemoryContext oldCtx; innerTuple = (SpGistInnerTuple) PageGetItem(page, - PageGetItemId(page, offset)); + PageGetItemId(page, offset)); if (innerTuple->tupstate != SPGIST_LIVE) { @@ -522,7 +523,7 @@ redirect: for (i = 0; i < out.nNodes; i++) { - int nodeN = out.nodeNumbers[i]; + int nodeN = out.nodeNumbers[i]; Assert(nodeN >= 0 && nodeN < in.nNodes); if (ItemPointerIsValid(&nodes[nodeN]->t_tid)) @@ -598,7 +599,7 @@ storeGettuple(SpGistScanOpaque so, ItemPointer heapPtr, if (so->want_itup) { /* - * Reconstruct desired IndexTuple. We have to copy the datum out of + * Reconstruct desired IndexTuple. We have to copy the datum out of * the temp context anyway, so we may as well create the tuple here. */ so->indexTups[so->nPtrs] = index_form_tuple(so->indexTupDesc, @@ -636,7 +637,7 @@ spggettuple(PG_FUNCTION_ARGS) if (so->want_itup) { /* Must pfree IndexTuples to avoid memory leak */ - int i; + int i; for (i = 0; i < so->nPtrs; i++) pfree(so->indexTups[i]); diff --git a/src/backend/access/spgist/spgtextproc.c b/src/backend/access/spgist/spgtextproc.c index 656015ea7e6598f1fa194c2676747944d3f7bcca..520d7b24c5597d3059066dd440fce02179af6c35 100644 --- a/src/backend/access/spgist/spgtextproc.c +++ b/src/backend/access/spgist/spgtextproc.c @@ -26,7 +26,7 @@ * In the worst case, a inner tuple in a text suffix tree could have as many * as 256 nodes (one for each possible byte value). Each node can take 16 * bytes on MAXALIGN=8 machines. The inner tuple must fit on an index page - * of size BLCKSZ. Rather than assuming we know the exact amount of overhead + * of size BLCKSZ. Rather than assuming we know the exact amount of overhead * imposed by page headers, tuple headers, etc, we leave 100 bytes for that * (the actual overhead should be no more than 56 bytes at this writing, so * there is slop in this number). The upshot is that the maximum safe prefix @@ -209,9 +209,9 @@ spg_text_choose(PG_FUNCTION_ARGS) { /* * Descend to existing node. (If in->allTheSame, the core code will - * ignore our nodeN specification here, but that's OK. We still - * have to provide the correct levelAdd and restDatum values, and - * those are the same regardless of which node gets chosen by core.) + * ignore our nodeN specification here, but that's OK. We still have + * to provide the correct levelAdd and restDatum values, and those are + * the same regardless of which node gets chosen by core.) */ out->resultType = spgMatchNode; out->result.matchNode.nodeN = i; @@ -227,10 +227,10 @@ spg_text_choose(PG_FUNCTION_ARGS) else if (in->allTheSame) { /* - * Can't use AddNode action, so split the tuple. The upper tuple - * has the same prefix as before and uses an empty node label for - * the lower tuple. The lower tuple has no prefix and the same - * node labels as the original tuple. + * Can't use AddNode action, so split the tuple. The upper tuple has + * the same prefix as before and uses an empty node label for the + * lower tuple. The lower tuple has no prefix and the same node + * labels as the original tuple. */ out->resultType = spgSplitTuple; out->result.splitTuple.prefixHasPrefix = in->hasPrefix; @@ -315,13 +315,13 @@ spg_text_picksplit(PG_FUNCTION_ARGS) if (commonLen < VARSIZE_ANY_EXHDR(texti)) nodes[i].c = *(uint8 *) (VARDATA_ANY(texti) + commonLen); else - nodes[i].c = '\0'; /* use \0 if string is all common */ + nodes[i].c = '\0'; /* use \0 if string is all common */ nodes[i].i = i; nodes[i].d = in->datums[i]; } /* - * Sort by label bytes so that we can group the values into nodes. This + * Sort by label bytes so that we can group the values into nodes. This * also ensures that the nodes are ordered by label value, allowing the * use of binary search in searchChar. */ @@ -371,7 +371,7 @@ spg_text_inner_consistent(PG_FUNCTION_ARGS) /* * Reconstruct values represented at this tuple, including parent data, - * prefix of this tuple if any, and the node label if any. in->level + * prefix of this tuple if any, and the node label if any. in->level * should be the length of the previously reconstructed value, and the * number of bytes added here is prefixSize or prefixSize + 1. * @@ -381,7 +381,7 @@ spg_text_inner_consistent(PG_FUNCTION_ARGS) * long-format reconstructed values. */ Assert(in->level == 0 ? DatumGetPointer(in->reconstructedValue) == NULL : - VARSIZE_ANY_EXHDR(DatumGetPointer(in->reconstructedValue)) == in->level); + VARSIZE_ANY_EXHDR(DatumGetPointer(in->reconstructedValue)) == in->level); maxReconstrLen = in->level + 1; if (in->hasPrefix) @@ -530,7 +530,7 @@ spg_text_leaf_consistent(PG_FUNCTION_ARGS) } else { - text *fullText = palloc(VARHDRSZ + fullLen); + text *fullText = palloc(VARHDRSZ + fullLen); SET_VARSIZE(fullText, VARHDRSZ + fullLen); fullValue = VARDATA(fullText); diff --git a/src/backend/access/spgist/spgutils.c b/src/backend/access/spgist/spgutils.c index 46a10f6a20617e165d32e0a79e46f632e650612f..d56c2325fe57d1e3d846ff6ddec42705d298b15a 100644 --- a/src/backend/access/spgist/spgutils.c +++ b/src/backend/access/spgist/spgutils.c @@ -235,7 +235,7 @@ SpGistUpdateMetaPage(Relation index) * * When requesting an inner page, if we get one with the wrong parity, * we just release the buffer and try again. We will get a different page - * because GetFreeIndexPage will have marked the page used in FSM. The page + * because GetFreeIndexPage will have marked the page used in FSM. The page * is entered in our local lastUsedPages cache, so there's some hope of * making use of it later in this session, but otherwise we rely on VACUUM * to eventually re-enter the page in FSM, making it available for recycling. @@ -245,7 +245,7 @@ SpGistUpdateMetaPage(Relation index) * * When we return a buffer to the caller, the page is *not* entered into * the lastUsedPages cache; we expect the caller will do so after it's taken - * whatever space it will use. This is because after the caller has used up + * whatever space it will use. This is because after the caller has used up * some space, the page might have less space than whatever was cached already * so we'd rather not trash the old cache entry. */ @@ -275,7 +275,7 @@ allocNewBuffer(Relation index, int flags) else { BlockNumber blkno = BufferGetBlockNumber(buffer); - int blkFlags = GBUF_INNER_PARITY(blkno); + int blkFlags = GBUF_INNER_PARITY(blkno); if ((flags & GBUF_PARITY_MASK) == blkFlags) { @@ -317,7 +317,7 @@ SpGistGetBuffer(Relation index, int flags, int needSpace, bool *isNew) /* * If possible, increase the space request to include relation's - * fillfactor. This ensures that when we add unrelated tuples to a page, + * fillfactor. This ensures that when we add unrelated tuples to a page, * we try to keep 100-fillfactor% available for adding tuples that are * related to the ones already on it. But fillfactor mustn't cause an * error for requests that would otherwise be legal. @@ -664,7 +664,7 @@ spgFormInnerTuple(SpGistState *state, bool hasPrefix, Datum prefix, errmsg("SPGiST inner tuple size %lu exceeds maximum %lu", (unsigned long) size, (unsigned long) (SPGIST_PAGE_CAPACITY - sizeof(ItemIdData))), - errhint("Values larger than a buffer page cannot be indexed."))); + errhint("Values larger than a buffer page cannot be indexed."))); /* * Check for overflow of header fields --- probably can't fail if the @@ -801,7 +801,7 @@ SpGistPageAddNewItem(SpGistState *state, Page page, Item item, Size size, for (; i <= maxoff; i++) { SpGistDeadTuple it = (SpGistDeadTuple) PageGetItem(page, - PageGetItemId(page, i)); + PageGetItemId(page, i)); if (it->tupstate == SPGIST_PLACEHOLDER) { diff --git a/src/backend/access/spgist/spgvacuum.c b/src/backend/access/spgist/spgvacuum.c index 856790ee2aa41cac1728021a079338a10b21e429..27b55170cb4b3bfa083ce0d7ffa7cbbb1d1f2733 100644 --- a/src/backend/access/spgist/spgvacuum.c +++ b/src/backend/access/spgist/spgvacuum.c @@ -31,8 +31,8 @@ /* Entry in pending-list of TIDs we need to revisit */ typedef struct spgVacPendingItem { - ItemPointerData tid; /* redirection target to visit */ - bool done; /* have we dealt with this? */ + ItemPointerData tid; /* redirection target to visit */ + bool done; /* have we dealt with this? */ struct spgVacPendingItem *next; /* list link */ } spgVacPendingItem; @@ -46,10 +46,10 @@ typedef struct spgBulkDeleteState void *callback_state; /* Additional working state */ - SpGistState spgstate; /* for SPGiST operations that need one */ - spgVacPendingItem *pendingList; /* TIDs we need to (re)visit */ - TransactionId myXmin; /* for detecting newly-added redirects */ - TransactionId OldestXmin; /* for deciding a redirect is obsolete */ + SpGistState spgstate; /* for SPGiST operations that need one */ + spgVacPendingItem *pendingList; /* TIDs we need to (re)visit */ + TransactionId myXmin; /* for detecting newly-added redirects */ + TransactionId OldestXmin; /* for deciding a redirect is obsolete */ BlockNumber lastFilledBlock; /* last non-deletable block */ } spgBulkDeleteState; @@ -213,7 +213,7 @@ vacuumLeafPage(spgBulkDeleteState *bds, Relation index, Buffer buffer, * Figure out exactly what we have to do. We do this separately from * actually modifying the page, mainly so that we have a representation * that can be dumped into WAL and then the replay code can do exactly - * the same thing. The output of this step consists of six arrays + * the same thing. The output of this step consists of six arrays * describing four kinds of operations, to be performed in this order: * * toDead[]: tuple numbers to be replaced with DEAD tuples @@ -276,8 +276,8 @@ vacuumLeafPage(spgBulkDeleteState *bds, Relation index, Buffer buffer, else if (prevLive == InvalidOffsetNumber) { /* - * This is the first live tuple in the chain. It has - * to move to the head position. + * This is the first live tuple in the chain. It has to move + * to the head position. */ moveSrc[xlrec.nMove] = j; moveDest[xlrec.nMove] = i; @@ -289,7 +289,7 @@ vacuumLeafPage(spgBulkDeleteState *bds, Relation index, Buffer buffer, else { /* - * Second or later live tuple. Arrange to re-chain it to the + * Second or later live tuple. Arrange to re-chain it to the * previous live one, if there was a gap. */ if (interveningDeletable) @@ -353,11 +353,11 @@ vacuumLeafPage(spgBulkDeleteState *bds, Relation index, Buffer buffer, InvalidBlockNumber, InvalidOffsetNumber); /* - * We implement the move step by swapping the item pointers of the - * source and target tuples, then replacing the newly-source tuples - * with placeholders. This is perhaps unduly friendly with the page - * data representation, but it's fast and doesn't risk page overflow - * when a tuple to be relocated is large. + * We implement the move step by swapping the item pointers of the source + * and target tuples, then replacing the newly-source tuples with + * placeholders. This is perhaps unduly friendly with the page data + * representation, but it's fast and doesn't risk page overflow when a + * tuple to be relocated is large. */ for (i = 0; i < xlrec.nMove; i++) { @@ -518,7 +518,7 @@ vacuumRedirectAndPlaceholder(Relation index, Buffer buffer, */ for (i = max; i >= FirstOffsetNumber && - (opaque->nRedirection > 0 || !hasNonPlaceholder); + (opaque->nRedirection > 0 || !hasNonPlaceholder); i--) { SpGistDeadTuple dt; @@ -651,9 +651,9 @@ spgvacuumpage(spgBulkDeleteState *bds, BlockNumber blkno) /* * The root pages must never be deleted, nor marked as available in FSM, - * because we don't want them ever returned by a search for a place to - * put a new tuple. Otherwise, check for empty/deletable page, and - * make sure FSM knows about it. + * because we don't want them ever returned by a search for a place to put + * a new tuple. Otherwise, check for empty/deletable page, and make sure + * FSM knows about it. */ if (!SpGistBlockIsRoot(blkno)) { @@ -688,7 +688,7 @@ spgprocesspending(spgBulkDeleteState *bds) Relation index = bds->info->index; spgVacPendingItem *pitem; spgVacPendingItem *nitem; - BlockNumber blkno; + BlockNumber blkno; Buffer buffer; Page page; @@ -741,11 +741,11 @@ spgprocesspending(spgBulkDeleteState *bds) else { /* - * On an inner page, visit the referenced inner tuple and add - * all its downlinks to the pending list. We might have pending - * items for more than one inner tuple on the same page (in fact - * this is pretty likely given the way space allocation works), - * so get them all while we are here. + * On an inner page, visit the referenced inner tuple and add all + * its downlinks to the pending list. We might have pending items + * for more than one inner tuple on the same page (in fact this is + * pretty likely given the way space allocation works), so get + * them all while we are here. */ for (nitem = pitem; nitem != NULL; nitem = nitem->next) { @@ -774,7 +774,7 @@ spgprocesspending(spgBulkDeleteState *bds) { /* transfer attention to redirect point */ spgAddPendingTID(bds, - &((SpGistDeadTuple) innerTuple)->pointer); + &((SpGistDeadTuple) innerTuple)->pointer); } else elog(ERROR, "unexpected SPGiST tuple state: %d", @@ -825,8 +825,8 @@ spgvacuumscan(spgBulkDeleteState *bds) * physical order (we hope the kernel will cooperate in providing * read-ahead for speed). It is critical that we visit all leaf pages, * including ones added after we start the scan, else we might fail to - * delete some deletable tuples. See more extensive comments about - * this in btvacuumscan(). + * delete some deletable tuples. See more extensive comments about this + * in btvacuumscan(). */ blkno = SPGIST_METAPAGE_BLKNO + 1; for (;;) diff --git a/src/backend/access/spgist/spgxlog.c b/src/backend/access/spgist/spgxlog.c index 8e87e2adc9060ffee2a5dcc4b76f2cd2b7726256..82f8c8b978a01a426d4019aa5f7fdbfc463a5b36 100644 --- a/src/backend/access/spgist/spgxlog.c +++ b/src/backend/access/spgist/spgxlog.c @@ -40,7 +40,7 @@ fillFakeState(SpGistState *state, spgxlogState stateSrc) } /* - * Add a leaf tuple, or replace an existing placeholder tuple. This is used + * Add a leaf tuple, or replace an existing placeholder tuple. This is used * to replay SpGistPageAddNewItem() operations. If the offset points at an * existing tuple, it had better be a placeholder tuple. */ @@ -50,7 +50,7 @@ addOrReplaceTuple(Page page, Item tuple, int size, OffsetNumber offset) if (offset <= PageGetMaxOffsetNumber(page)) { SpGistDeadTuple dt = (SpGistDeadTuple) PageGetItem(page, - PageGetItemId(page, offset)); + PageGetItemId(page, offset)); if (dt->tupstate != SPGIST_PLACEHOLDER) elog(ERROR, "SPGiST tuple to be replaced is not a placeholder"); @@ -126,7 +126,7 @@ spgRedoAddLeaf(XLogRecPtr lsn, XLogRecord *record) if (xldata->newPage) SpGistInitBuffer(buffer, - SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); + SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); if (!XLByteLE(lsn, PageGetLSN(page))) { @@ -143,7 +143,7 @@ spgRedoAddLeaf(XLogRecPtr lsn, XLogRecord *record) SpGistLeafTuple head; head = (SpGistLeafTuple) PageGetItem(page, - PageGetItemId(page, xldata->offnumHeadLeaf)); + PageGetItemId(page, xldata->offnumHeadLeaf)); Assert(head->nextOffset == leafTuple->nextOffset); head->nextOffset = xldata->offnumLeaf; } @@ -154,7 +154,7 @@ spgRedoAddLeaf(XLogRecPtr lsn, XLogRecord *record) PageIndexTupleDelete(page, xldata->offnumLeaf); if (PageAddItem(page, (Item) leafTuple, leafTuple->size, - xldata->offnumLeaf, false, false) != xldata->offnumLeaf) + xldata->offnumLeaf, false, false) != xldata->offnumLeaf) elog(ERROR, "failed to add item of size %u to SPGiST index page", leafTuple->size); } @@ -180,7 +180,7 @@ spgRedoAddLeaf(XLogRecPtr lsn, XLogRecord *record) SpGistInnerTuple tuple; tuple = (SpGistInnerTuple) PageGetItem(page, - PageGetItemId(page, xldata->offnumParent)); + PageGetItemId(page, xldata->offnumParent)); spgUpdateNodeLink(tuple, xldata->nodeI, xldata->blknoLeaf, xldata->offnumLeaf); @@ -229,7 +229,7 @@ spgRedoMoveLeafs(XLogRecPtr lsn, XLogRecord *record) if (xldata->newPage) SpGistInitBuffer(buffer, - SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); + SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); if (!XLByteLE(lsn, PageGetLSN(page))) { @@ -261,7 +261,7 @@ spgRedoMoveLeafs(XLogRecPtr lsn, XLogRecord *record) if (!XLByteLE(lsn, PageGetLSN(page))) { spgPageIndexMultiDelete(&state, page, toDelete, xldata->nMoves, - state.isBuild ? SPGIST_PLACEHOLDER : SPGIST_REDIRECT, + state.isBuild ? SPGIST_PLACEHOLDER : SPGIST_REDIRECT, SPGIST_PLACEHOLDER, xldata->blknoDst, toInsert[nInsert - 1]); @@ -286,7 +286,7 @@ spgRedoMoveLeafs(XLogRecPtr lsn, XLogRecord *record) SpGistInnerTuple tuple; tuple = (SpGistInnerTuple) PageGetItem(page, - PageGetItemId(page, xldata->offnumParent)); + PageGetItemId(page, xldata->offnumParent)); spgUpdateNodeLink(tuple, xldata->nodeI, xldata->blknoDst, toInsert[nInsert - 1]); @@ -413,7 +413,7 @@ spgRedoAddNode(XLogRecPtr lsn, XLogRecord *record) } /* - * Update parent downlink. Since parent could be in either of the + * Update parent downlink. Since parent could be in either of the * previous two buffers, it's a bit tricky to determine which BKP bit * applies. */ @@ -435,7 +435,7 @@ spgRedoAddNode(XLogRecPtr lsn, XLogRecord *record) SpGistInnerTuple innerTuple; innerTuple = (SpGistInnerTuple) PageGetItem(page, - PageGetItemId(page, xldata->offnumParent)); + PageGetItemId(page, xldata->offnumParent)); spgUpdateNodeLink(innerTuple, xldata->nodeI, xldata->blknoNew, xldata->offnumNew); @@ -504,7 +504,7 @@ spgRedoSplitTuple(XLogRecPtr lsn, XLogRecord *record) { PageIndexTupleDelete(page, xldata->offnumPrefix); if (PageAddItem(page, (Item) prefixTuple, prefixTuple->size, - xldata->offnumPrefix, false, false) != xldata->offnumPrefix) + xldata->offnumPrefix, false, false) != xldata->offnumPrefix) elog(ERROR, "failed to add item of size %u to SPGiST index page", prefixTuple->size); @@ -571,7 +571,7 @@ spgRedoPickSplit(XLogRecPtr lsn, XLogRecord *record) page = (Page) BufferGetPage(srcBuffer); SpGistInitBuffer(srcBuffer, - SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); + SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); /* don't update LSN etc till we're done with it */ } else @@ -587,8 +587,8 @@ spgRedoPickSplit(XLogRecPtr lsn, XLogRecord *record) { /* * We have it a bit easier here than in doPickSplit(), - * because we know the inner tuple's location already, - * so we can inject the correct redirection tuple now. + * because we know the inner tuple's location already, so + * we can inject the correct redirection tuple now. */ if (!state.isBuild) spgPageIndexMultiDelete(&state, page, @@ -627,7 +627,7 @@ spgRedoPickSplit(XLogRecPtr lsn, XLogRecord *record) page = (Page) BufferGetPage(destBuffer); SpGistInitBuffer(destBuffer, - SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); + SPGIST_LEAF | (xldata->storesNulls ? SPGIST_NULLS : 0)); /* don't update LSN etc till we're done with it */ } else @@ -707,9 +707,9 @@ spgRedoPickSplit(XLogRecPtr lsn, XLogRecord *record) SpGistInnerTuple parent; parent = (SpGistInnerTuple) PageGetItem(page, - PageGetItemId(page, xldata->offnumParent)); + PageGetItemId(page, xldata->offnumParent)); spgUpdateNodeLink(parent, xldata->nodeI, - xldata->blknoInner, xldata->offnumInner); + xldata->blknoInner, xldata->offnumInner); } PageSetLSN(page, lsn); @@ -742,9 +742,9 @@ spgRedoPickSplit(XLogRecPtr lsn, XLogRecord *record) SpGistInnerTuple parent; parent = (SpGistInnerTuple) PageGetItem(page, - PageGetItemId(page, xldata->offnumParent)); + PageGetItemId(page, xldata->offnumParent)); spgUpdateNodeLink(parent, xldata->nodeI, - xldata->blknoInner, xldata->offnumInner); + xldata->blknoInner, xldata->offnumInner); PageSetLSN(page, lsn); PageSetTLI(page, ThisTimeLineID); @@ -803,7 +803,7 @@ spgRedoVacuumLeaf(XLogRecPtr lsn, XLogRecord *record) spgPageIndexMultiDelete(&state, page, toPlaceholder, xldata->nPlaceholder, - SPGIST_PLACEHOLDER, SPGIST_PLACEHOLDER, + SPGIST_PLACEHOLDER, SPGIST_PLACEHOLDER, InvalidBlockNumber, InvalidOffsetNumber); @@ -821,7 +821,7 @@ spgRedoVacuumLeaf(XLogRecPtr lsn, XLogRecord *record) spgPageIndexMultiDelete(&state, page, moveSrc, xldata->nMove, - SPGIST_PLACEHOLDER, SPGIST_PLACEHOLDER, + SPGIST_PLACEHOLDER, SPGIST_PLACEHOLDER, InvalidBlockNumber, InvalidOffsetNumber); @@ -906,7 +906,7 @@ spgRedoVacuumRedirect(XLogRecPtr lsn, XLogRecord *record) SpGistDeadTuple dt; dt = (SpGistDeadTuple) PageGetItem(page, - PageGetItemId(page, itemToPlaceholder[i])); + PageGetItemId(page, itemToPlaceholder[i])); Assert(dt->tupstate == SPGIST_REDIRECT); dt->tupstate = SPGIST_PLACEHOLDER; ItemPointerSetInvalid(&dt->pointer); diff --git a/src/backend/access/transam/clog.c b/src/backend/access/transam/clog.c index 33b5ca2d36f3391a4884c2173b55a149bf096a59..7f2f6921d5db4de494633788aa407a1f6f85e72d 100644 --- a/src/backend/access/transam/clog.c +++ b/src/backend/access/transam/clog.c @@ -417,7 +417,7 @@ TransactionIdGetStatus(TransactionId xid, XLogRecPtr *lsn) * Testing during the PostgreSQL 9.2 development cycle revealed that on a * large multi-processor system, it was possible to have more CLOG page * requests in flight at one time than the numebr of CLOG buffers which existed - * at that time, which was hardcoded to 8. Further testing revealed that + * at that time, which was hardcoded to 8. Further testing revealed that * performance dropped off with more than 32 CLOG buffers, possibly because * the linear buffer search algorithm doesn't scale well. * diff --git a/src/backend/access/transam/slru.c b/src/backend/access/transam/slru.c index a8e3f19119a568858e8107bf53196bffa5cfa2c5..dd69c232eb4f18dbc76bd0901a4f2be28289f649 100644 --- a/src/backend/access/transam/slru.c +++ b/src/backend/access/transam/slru.c @@ -903,12 +903,12 @@ SlruSelectLRUPage(SlruCtl ctl, int pageno) { int slotno; int cur_count; - int bestvalidslot = 0; /* keep compiler quiet */ + int bestvalidslot = 0; /* keep compiler quiet */ int best_valid_delta = -1; - int best_valid_page_number = 0; /* keep compiler quiet */ - int bestinvalidslot = 0; /* keep compiler quiet */ + int best_valid_page_number = 0; /* keep compiler quiet */ + int bestinvalidslot = 0; /* keep compiler quiet */ int best_invalid_delta = -1; - int best_invalid_page_number = 0; /* keep compiler quiet */ + int best_invalid_page_number = 0; /* keep compiler quiet */ /* See if page already has a buffer assigned */ for (slotno = 0; slotno < shared->num_slots; slotno++) @@ -920,15 +920,15 @@ SlruSelectLRUPage(SlruCtl ctl, int pageno) /* * If we find any EMPTY slot, just select that one. Else choose a - * victim page to replace. We normally take the least recently used + * victim page to replace. We normally take the least recently used * valid page, but we will never take the slot containing - * latest_page_number, even if it appears least recently used. We + * latest_page_number, even if it appears least recently used. We * will select a slot that is already I/O busy only if there is no * other choice: a read-busy slot will not be least recently used once * the read finishes, and waiting for an I/O on a write-busy slot is * inferior to just picking some other slot. Testing shows the slot - * we pick instead will often be clean, allowing us to begin a read - * at once. + * we pick instead will often be clean, allowing us to begin a read at + * once. * * Normally the page_lru_count values will all be different and so * there will be a well-defined LRU page. But since we allow @@ -997,10 +997,10 @@ SlruSelectLRUPage(SlruCtl ctl, int pageno) /* * If all pages (except possibly the latest one) are I/O busy, we'll - * have to wait for an I/O to complete and then retry. In that unhappy - * case, we choose to wait for the I/O on the least recently used slot, - * on the assumption that it was likely initiated first of all the I/Os - * in progress and may therefore finish first. + * have to wait for an I/O to complete and then retry. In that + * unhappy case, we choose to wait for the I/O on the least recently + * used slot, on the assumption that it was likely initiated first of + * all the I/Os in progress and may therefore finish first. */ if (best_valid_delta < 0) { @@ -1168,20 +1168,20 @@ restart:; /* * SlruScanDirectory callback - * This callback reports true if there's any segment prior to the one - * containing the page passed as "data". + * This callback reports true if there's any segment prior to the one + * containing the page passed as "data". */ bool SlruScanDirCbReportPresence(SlruCtl ctl, char *filename, int segpage, void *data) { - int cutoffPage = *(int *) data; + int cutoffPage = *(int *) data; cutoffPage -= cutoffPage % SLRU_PAGES_PER_SEGMENT; if (ctl->PagePrecedes(segpage, cutoffPage)) - return true; /* found one; don't iterate any more */ + return true; /* found one; don't iterate any more */ - return false; /* keep going */ + return false; /* keep going */ } /* @@ -1191,8 +1191,8 @@ SlruScanDirCbReportPresence(SlruCtl ctl, char *filename, int segpage, void *data static bool SlruScanDirCbDeleteCutoff(SlruCtl ctl, char *filename, int segpage, void *data) { - char path[MAXPGPATH]; - int cutoffPage = *(int *) data; + char path[MAXPGPATH]; + int cutoffPage = *(int *) data; if (ctl->PagePrecedes(segpage, cutoffPage)) { @@ -1202,7 +1202,7 @@ SlruScanDirCbDeleteCutoff(SlruCtl ctl, char *filename, int segpage, void *data) unlink(path); } - return false; /* keep going */ + return false; /* keep going */ } /* @@ -1212,14 +1212,14 @@ SlruScanDirCbDeleteCutoff(SlruCtl ctl, char *filename, int segpage, void *data) bool SlruScanDirCbDeleteAll(SlruCtl ctl, char *filename, int segpage, void *data) { - char path[MAXPGPATH]; + char path[MAXPGPATH]; snprintf(path, MAXPGPATH, "%s/%s", ctl->Dir, filename); ereport(DEBUG2, (errmsg("removing file \"%s\"", path))); unlink(path); - return false; /* keep going */ + return false; /* keep going */ } /* diff --git a/src/backend/access/transam/twophase.c b/src/backend/access/transam/twophase.c index 0b41a76a3251461e3c53ed0d33ae8789a7c0d22f..b94fae3740251426b80dff346fc03424e7e6a76f 100644 --- a/src/backend/access/transam/twophase.c +++ b/src/backend/access/transam/twophase.c @@ -360,8 +360,9 @@ static void GXactLoadSubxactData(GlobalTransaction gxact, int nsubxacts, TransactionId *children) { - PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; - PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; + PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + /* We need no extra lock since the GXACT isn't valid yet */ if (nsubxacts > PGPROC_MAX_CACHED_SUBXIDS) { @@ -410,7 +411,7 @@ LockGXact(const char *gid, Oid user) for (i = 0; i < TwoPhaseState->numPrepXacts; i++) { GlobalTransaction gxact = TwoPhaseState->prepXacts[i]; - PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; + PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; /* Ignore not-yet-valid GIDs */ if (!gxact->valid) @@ -523,7 +524,7 @@ TransactionIdIsPrepared(TransactionId xid) for (i = 0; i < TwoPhaseState->numPrepXacts; i++) { GlobalTransaction gxact = TwoPhaseState->prepXacts[i]; - PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; if (gxact->valid && pgxact->xid == xid) { @@ -648,8 +649,8 @@ pg_prepared_xact(PG_FUNCTION_ARGS) while (status->array != NULL && status->currIdx < status->ngxacts) { GlobalTransaction gxact = &status->array[status->currIdx++]; - PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; - PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; + PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; Datum values[5]; bool nulls[5]; HeapTuple tuple; @@ -719,7 +720,7 @@ TwoPhaseGetDummyProc(TransactionId xid) for (i = 0; i < TwoPhaseState->numPrepXacts; i++) { GlobalTransaction gxact = TwoPhaseState->prepXacts[i]; - PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; if (pgxact->xid == xid) { @@ -850,8 +851,8 @@ save_state_data(const void *data, uint32 len) void StartPrepare(GlobalTransaction gxact) { - PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; - PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + PGPROC *proc = &ProcGlobal->allProcs[gxact->pgprocno]; + PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; TransactionId xid = pgxact->xid; TwoPhaseFileHeader hdr; TransactionId *children; @@ -1063,9 +1064,9 @@ EndPrepare(GlobalTransaction gxact) errmsg("could not close two-phase state file: %m"))); /* - * Mark the prepared transaction as valid. As soon as xact.c marks MyPgXact - * as not running our XID (which it will do immediately after this - * function returns), others can commit/rollback the xact. + * Mark the prepared transaction as valid. As soon as xact.c marks + * MyPgXact as not running our XID (which it will do immediately after + * this function returns), others can commit/rollback the xact. * * NB: a side effect of this is to make a dummy ProcArray entry for the * prepared XID. This must happen before we clear the XID from MyPgXact, @@ -1551,7 +1552,7 @@ CheckPointTwoPhase(XLogRecPtr redo_horizon) for (i = 0; i < TwoPhaseState->numPrepXacts; i++) { GlobalTransaction gxact = TwoPhaseState->prepXacts[i]; - PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; + PGXACT *pgxact = &ProcGlobal->allPgXact[gxact->pgprocno]; if (gxact->valid && XLByteLE(gxact->prepare_lsn, redo_horizon)) @@ -1707,7 +1708,7 @@ PrescanPreparedTransactions(TransactionId **xids_p, int *nxids_p) * XID, and they may force us to advance nextXid. * * We don't expect anyone else to modify nextXid, hence we don't - * need to hold a lock while examining it. We still acquire the + * need to hold a lock while examining it. We still acquire the * lock to modify it, though. */ subxids = (TransactionId *) diff --git a/src/backend/access/transam/varsup.c b/src/backend/access/transam/varsup.c index 892a46abc3fa34dd16627e6747b2279387b65b30..7abf9343beeff774da70e1debb3ee6174ec5cafd 100644 --- a/src/backend/access/transam/varsup.c +++ b/src/backend/access/transam/varsup.c @@ -174,8 +174,8 @@ GetNewTransactionId(bool isSubXact) * latestCompletedXid is present in the ProcArray, which is essential for * correct OldestXmin tracking; see src/backend/access/transam/README. * - * XXX by storing xid into MyPgXact without acquiring ProcArrayLock, we are - * relying on fetch/store of an xid to be atomic, else other backends + * XXX by storing xid into MyPgXact without acquiring ProcArrayLock, we + * are relying on fetch/store of an xid to be atomic, else other backends * might see a partially-set xid here. But holding both locks at once * would be a nasty concurrency hit. So for now, assume atomicity. * diff --git a/src/backend/access/transam/xact.c b/src/backend/access/transam/xact.c index 659b53524cdf71dd8d36e1723f09838c49c82e1c..8f00186dd741b565c2769da496b28b40fbddc881 100644 --- a/src/backend/access/transam/xact.c +++ b/src/backend/access/transam/xact.c @@ -1019,6 +1019,7 @@ RecordTransactionCommit(void) XLogRecData rdata[4]; int lastrdata = 0; xl_xact_commit xlrec; + /* * Set flags required for recovery processing of commits. */ @@ -1073,7 +1074,8 @@ RecordTransactionCommit(void) { XLogRecData rdata[2]; int lastrdata = 0; - xl_xact_commit_compact xlrec; + xl_xact_commit_compact xlrec; + xlrec.xact_time = xactStopTimestamp; xlrec.nsubxacts = nchildren; rdata[0].data = (char *) (&xlrec); @@ -2102,7 +2104,7 @@ PrepareTransaction(void) if (XactHasExportedSnapshots()) ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("cannot PREPARE a transaction that has exported snapshots"))); + errmsg("cannot PREPARE a transaction that has exported snapshots"))); /* Prevent cancel/die interrupt while cleaning up */ HOLD_INTERRUPTS(); @@ -2602,10 +2604,10 @@ CommitTransactionCommand(void) break; /* - * We were issued a RELEASE command, so we end the - * current subtransaction and return to the parent transaction. - * The parent might be ended too, so repeat till we find an - * INPROGRESS transaction or subtransaction. + * We were issued a RELEASE command, so we end the current + * subtransaction and return to the parent transaction. The parent + * might be ended too, so repeat till we find an INPROGRESS + * transaction or subtransaction. */ case TBLOCK_SUBRELEASE: do @@ -2623,9 +2625,9 @@ CommitTransactionCommand(void) * hierarchy and perform final commit. We do this by rolling up * any subtransactions into their parent, which leads to O(N^2) * operations with respect to resource owners - this isn't that - * bad until we approach a thousands of savepoints but is necessary - * for correctness should after triggers create new resource - * owners. + * bad until we approach a thousands of savepoints but is + * necessary for correctness should after triggers create new + * resource owners. */ case TBLOCK_SUBCOMMIT: do @@ -4551,11 +4553,11 @@ xactGetCommittedChildren(TransactionId **ptr) */ static void xact_redo_commit_internal(TransactionId xid, XLogRecPtr lsn, - TransactionId *sub_xids, int nsubxacts, - SharedInvalidationMessage *inval_msgs, int nmsgs, - RelFileNode *xnodes, int nrels, - Oid dbId, Oid tsId, - uint32 xinfo) + TransactionId *sub_xids, int nsubxacts, + SharedInvalidationMessage *inval_msgs, int nmsgs, + RelFileNode *xnodes, int nrels, + Oid dbId, Oid tsId, + uint32 xinfo) { TransactionId max_xid; int i; @@ -4659,12 +4661,13 @@ xact_redo_commit_internal(TransactionId xid, XLogRecPtr lsn, XLogFlush(lsn); } + /* * Utility function to call xact_redo_commit_internal after breaking down xlrec */ static void xact_redo_commit(xl_xact_commit *xlrec, - TransactionId xid, XLogRecPtr lsn) + TransactionId xid, XLogRecPtr lsn) { TransactionId *subxacts; SharedInvalidationMessage *inval_msgs; @@ -4675,11 +4678,11 @@ xact_redo_commit(xl_xact_commit *xlrec, inval_msgs = (SharedInvalidationMessage *) &(subxacts[xlrec->nsubxacts]); xact_redo_commit_internal(xid, lsn, subxacts, xlrec->nsubxacts, - inval_msgs, xlrec->nmsgs, - xlrec->xnodes, xlrec->nrels, - xlrec->dbId, - xlrec->tsId, - xlrec->xinfo); + inval_msgs, xlrec->nmsgs, + xlrec->xnodes, xlrec->nrels, + xlrec->dbId, + xlrec->tsId, + xlrec->xinfo); } /* @@ -4687,14 +4690,14 @@ xact_redo_commit(xl_xact_commit *xlrec, */ static void xact_redo_commit_compact(xl_xact_commit_compact *xlrec, - TransactionId xid, XLogRecPtr lsn) + TransactionId xid, XLogRecPtr lsn) { xact_redo_commit_internal(xid, lsn, xlrec->subxacts, xlrec->nsubxacts, - NULL, 0, /* inval msgs */ - NULL, 0, /* relfilenodes */ - InvalidOid, /* dbId */ - InvalidOid, /* tsId */ - 0); /* xinfo */ + NULL, 0, /* inval msgs */ + NULL, 0, /* relfilenodes */ + InvalidOid, /* dbId */ + InvalidOid, /* tsId */ + 0); /* xinfo */ } /* diff --git a/src/backend/access/transam/xlog.c b/src/backend/access/transam/xlog.c index 0f2678cfda04a750d436d0eb1a2d7632ba2d5b57..bcb71c45b2a1b677d81c4634b14b6af96faffe97 100644 --- a/src/backend/access/transam/xlog.c +++ b/src/backend/access/transam/xlog.c @@ -344,10 +344,10 @@ typedef struct XLogCtlInsert /* * fullPageWrites is the master copy used by all backends to determine - * whether to write full-page to WAL, instead of using process-local - * one. This is required because, when full_page_writes is changed - * by SIGHUP, we must WAL-log it before it actually affects - * WAL-logging by backends. Checkpointer sets at startup or after SIGHUP. + * whether to write full-page to WAL, instead of using process-local one. + * This is required because, when full_page_writes is changed by SIGHUP, + * we must WAL-log it before it actually affects WAL-logging by backends. + * Checkpointer sets at startup or after SIGHUP. */ bool fullPageWrites; @@ -455,8 +455,11 @@ typedef struct XLogCtlData XLogRecPtr recoveryLastRecPtr; /* timestamp of last COMMIT/ABORT record replayed (or being replayed) */ TimestampTz recoveryLastXTime; - /* timestamp of when we started replaying the current chunk of WAL data, - * only relevant for replication or archive recovery */ + + /* + * timestamp of when we started replaying the current chunk of WAL data, + * only relevant for replication or archive recovery + */ TimestampTz currentChunkStartTime; /* end of the last record restored from the archive */ XLogRecPtr restoreLastRecPtr; @@ -580,7 +583,7 @@ static bool updateMinRecoveryPoint = true; * to replay all the WAL, so reachedConsistency is never set. During archive * recovery, the database is consistent once minRecoveryPoint is reached. */ -bool reachedConsistency = false; +bool reachedConsistency = false; static bool InRedo = false; @@ -750,8 +753,8 @@ XLogInsert(RmgrId rmid, uint8 info, XLogRecData *rdata) * insert lock, but it seems better to avoid doing CRC calculations while * holding the lock. * - * We add entries for backup blocks to the chain, so that they don't - * need any special treatment in the critical section where the chunks are + * We add entries for backup blocks to the chain, so that they don't need + * any special treatment in the critical section where the chunks are * copied into the WAL buffers. Those entries have to be unlinked from the * chain if we have to loop back here. */ @@ -896,10 +899,10 @@ begin:; /* * Calculate CRC of the data, including all the backup blocks * - * Note that the record header isn't added into the CRC initially since - * we don't know the prev-link yet. Thus, the CRC will represent the CRC - * of the whole record in the order: rdata, then backup blocks, then - * record header. + * Note that the record header isn't added into the CRC initially since we + * don't know the prev-link yet. Thus, the CRC will represent the CRC of + * the whole record in the order: rdata, then backup blocks, then record + * header. */ INIT_CRC32(rdata_crc); for (rdt = rdata; rdt != NULL; rdt = rdt->next) @@ -948,10 +951,10 @@ begin:; } /* - * Also check to see if fullPageWrites or forcePageWrites was just turned on; - * if we weren't already doing full-page writes then go back and recompute. - * (If it was just turned off, we could recompute the record without full pages, - * but we choose not to bother.) + * Also check to see if fullPageWrites or forcePageWrites was just turned + * on; if we weren't already doing full-page writes then go back and + * recompute. (If it was just turned off, we could recompute the record + * without full pages, but we choose not to bother.) */ if ((Insert->fullPageWrites || Insert->forcePageWrites) && !doPageWrites) { @@ -1575,15 +1578,15 @@ AdvanceXLInsertBuffer(bool new_segment) * WAL records beginning in this page have removable backup blocks. This * allows the WAL archiver to know whether it is safe to compress archived * WAL data by transforming full-block records into the non-full-block - * format. It is sufficient to record this at the page level because we + * format. It is sufficient to record this at the page level because we * force a page switch (in fact a segment switch) when starting a backup, * so the flag will be off before any records can be written during the - * backup. At the end of a backup, the last page will be marked as all + * backup. At the end of a backup, the last page will be marked as all * unsafe when perhaps only part is unsafe, but at worst the archiver * would miss the opportunity to compress a few records. */ if (!Insert->forcePageWrites) - NewPage->xlp_info |= XLP_BKP_REMOVABLE; + NewPage ->xlp_info |= XLP_BKP_REMOVABLE; /* * If first page of an XLOG segment file, make it a long header. @@ -1827,11 +1830,11 @@ XLogWrite(XLogwrtRqst WriteRqst, bool flexible, bool xlog_switch) Write->lastSegSwitchTime = (pg_time_t) time(NULL); /* - * Request a checkpoint if we've consumed too - * much xlog since the last one. For speed, we first check - * using the local copy of RedoRecPtr, which might be out of - * date; if it looks like a checkpoint is needed, forcibly - * update RedoRecPtr and recheck. + * Request a checkpoint if we've consumed too much xlog since + * the last one. For speed, we first check using the local + * copy of RedoRecPtr, which might be out of date; if it looks + * like a checkpoint is needed, forcibly update RedoRecPtr and + * recheck. */ if (IsUnderPostmaster && XLogCheckpointNeeded(openLogId, openLogSeg)) @@ -1931,7 +1934,7 @@ XLogSetAsyncXactLSN(XLogRecPtr asyncXactLSN) /* * If the WALWriter is sleeping, we should kick it to make it come out of - * low-power mode. Otherwise, determine whether there's a full page of + * low-power mode. Otherwise, determine whether there's a full page of * WAL available to write. */ if (!sleeping) @@ -1945,9 +1948,9 @@ XLogSetAsyncXactLSN(XLogRecPtr asyncXactLSN) } /* - * Nudge the WALWriter: it has a full page of WAL to write, or we want - * it to come out of low-power mode so that this async commit will reach - * disk within the expected amount of time. + * Nudge the WALWriter: it has a full page of WAL to write, or we want it + * to come out of low-power mode so that this async commit will reach disk + * within the expected amount of time. */ if (ProcGlobal->walwriterLatch) SetLatch(ProcGlobal->walwriterLatch); @@ -2076,8 +2079,8 @@ XLogFlush(XLogRecPtr record) WriteRqstPtr = record; /* - * Now wait until we get the write lock, or someone else does the - * flush for us. + * Now wait until we get the write lock, or someone else does the flush + * for us. */ for (;;) { @@ -2182,7 +2185,7 @@ XLogFlush(XLogRecPtr record) * block, and flush through the latest one of those. Thus, if async commits * are not being used, we will flush complete blocks only. We can guarantee * that async commits reach disk after at most three cycles; normally only - * one or two. (When flushing complete blocks, we allow XLogWrite to write + * one or two. (When flushing complete blocks, we allow XLogWrite to write * "flexibly", meaning it can stop at the end of the buffer ring; this makes a * difference only with very high load or long wal_writer_delay, but imposes * one extra cycle for the worst case for async commits.) @@ -2273,7 +2276,8 @@ XLogBackgroundFlush(void) /* * If we wrote something then we have something to send to standbys also, - * otherwise the replication delay become around 7s with just async commit. + * otherwise the replication delay become around 7s with just async + * commit. */ if (wrote_something) WalSndWakeup(); @@ -2776,17 +2780,17 @@ XLogFileRead(uint32 log, uint32 seg, int emode, TimeLineID tli, } /* - * If the segment was fetched from archival storage, replace - * the existing xlog segment (if any) with the archival version. + * If the segment was fetched from archival storage, replace the existing + * xlog segment (if any) with the archival version. */ if (source == XLOG_FROM_ARCHIVE) { /* use volatile pointer to prevent code rearrangement */ volatile XLogCtlData *xlogctl = XLogCtl; - XLogRecPtr endptr; - char xlogfpath[MAXPGPATH]; - bool reload = false; - struct stat statbuf; + XLogRecPtr endptr; + char xlogfpath[MAXPGPATH]; + bool reload = false; + struct stat statbuf; XLogFilePath(xlogfpath, tli, log, seg); if (stat(xlogfpath, &statbuf) == 0) @@ -2801,9 +2805,9 @@ XLogFileRead(uint32 log, uint32 seg, int emode, TimeLineID tli, if (rename(path, xlogfpath) < 0) ereport(ERROR, - (errcode_for_file_access(), - errmsg("could not rename file \"%s\" to \"%s\": %m", - path, xlogfpath))); + (errcode_for_file_access(), + errmsg("could not rename file \"%s\" to \"%s\": %m", + path, xlogfpath))); /* * If the existing segment was replaced, since walsenders might have @@ -3812,7 +3816,7 @@ ReadRecord(XLogRecPtr *RecPtr, int emode, bool fetching_ckpt) RecPtr = &tmpRecPtr; /* - * RecPtr is pointing to end+1 of the previous WAL record. We must + * RecPtr is pointing to end+1 of the previous WAL record. We must * advance it if necessary to where the next record starts. First, * align to next page if no more records can fit on the current page. */ @@ -5389,10 +5393,10 @@ readRecoveryCommandFile(void) } if (rtli) ereport(DEBUG2, - (errmsg_internal("recovery_target_timeline = %u", rtli))); + (errmsg_internal("recovery_target_timeline = %u", rtli))); else ereport(DEBUG2, - (errmsg_internal("recovery_target_timeline = latest"))); + (errmsg_internal("recovery_target_timeline = latest"))); } else if (strcmp(item->name, "recovery_target_xid") == 0) { @@ -5404,7 +5408,7 @@ readRecoveryCommandFile(void) item->value))); ereport(DEBUG2, (errmsg_internal("recovery_target_xid = %u", - recoveryTargetXid))); + recoveryTargetXid))); recoveryTarget = RECOVERY_TARGET_XID; } else if (strcmp(item->name, "recovery_target_time") == 0) @@ -5428,7 +5432,7 @@ readRecoveryCommandFile(void) Int32GetDatum(-1))); ereport(DEBUG2, (errmsg_internal("recovery_target_time = '%s'", - timestamptz_to_str(recoveryTargetTime)))); + timestamptz_to_str(recoveryTargetTime)))); } else if (strcmp(item->name, "recovery_target_name") == 0) { @@ -5576,13 +5580,13 @@ exitArchiveRecovery(TimeLineID endTLI, uint32 endLogId, uint32 endLogSeg) } /* - * If we are establishing a new timeline, we have to copy data from - * the last WAL segment of the old timeline to create a starting WAL - * segment for the new timeline. + * If we are establishing a new timeline, we have to copy data from the + * last WAL segment of the old timeline to create a starting WAL segment + * for the new timeline. * - * Notify the archiver that the last WAL segment of the old timeline - * is ready to copy to archival storage. Otherwise, it is not archived - * for a while. + * Notify the archiver that the last WAL segment of the old timeline is + * ready to copy to archival storage. Otherwise, it is not archived for a + * while. */ if (endTLI != ThisTimeLineID) { @@ -5604,8 +5608,8 @@ exitArchiveRecovery(TimeLineID endTLI, uint32 endLogId, uint32 endLogSeg) XLogArchiveCleanup(xlogpath); /* - * Since there might be a partial WAL segment named RECOVERYXLOG, - * get rid of it. + * Since there might be a partial WAL segment named RECOVERYXLOG, get rid + * of it. */ snprintf(recoveryPath, MAXPGPATH, XLOGDIR "/RECOVERYXLOG"); unlink(recoveryPath); /* ignore any error */ @@ -6323,11 +6327,11 @@ StartupXLOG(void) /* * Set backupStartPoint if we're starting recovery from a base backup. * - * Set backupEndPoint and use minRecoveryPoint as the backup end location - * if we're starting recovery from a base backup which was taken from - * the standby. In this case, the database system status in pg_control must - * indicate DB_IN_ARCHIVE_RECOVERY. If not, which means that backup - * is corrupted, so we cancel recovery. + * Set backupEndPoint and use minRecoveryPoint as the backup end + * location if we're starting recovery from a base backup which was + * taken from the standby. In this case, the database system status in + * pg_control must indicate DB_IN_ARCHIVE_RECOVERY. If not, which + * means that backup is corrupted, so we cancel recovery. */ if (haveBackupLabel) { @@ -6340,7 +6344,7 @@ StartupXLOG(void) ereport(FATAL, (errmsg("backup_label contains inconsistent data with control file"), errhint("This means that the backup is corrupted and you will " - "have to use another backup for recovery."))); + "have to use another backup for recovery."))); ControlFile->backupEndPoint = ControlFile->minRecoveryPoint; } } @@ -6383,15 +6387,15 @@ StartupXLOG(void) /* * We're in recovery, so unlogged relations may be trashed and must be - * reset. This should be done BEFORE allowing Hot Standby connections, - * so that read-only backends don't try to read whatever garbage is - * left over from before. + * reset. This should be done BEFORE allowing Hot Standby + * connections, so that read-only backends don't try to read whatever + * garbage is left over from before. */ ResetUnloggedRelations(UNLOGGED_RELATION_CLEANUP); /* - * Likewise, delete any saved transaction snapshot files that got - * left behind by crashed backends. + * Likewise, delete any saved transaction snapshot files that got left + * behind by crashed backends. */ DeleteAllExportedSnapshotFiles(); @@ -6489,10 +6493,11 @@ StartupXLOG(void) /* * Let postmaster know we've started redo now, so that it can launch - * checkpointer to perform restartpoints. We don't bother during crash - * recovery as restartpoints can only be performed during archive - * recovery. And we'd like to keep crash recovery simple, to avoid - * introducing bugs that could affect you when recovering after crash. + * checkpointer to perform restartpoints. We don't bother during + * crash recovery as restartpoints can only be performed during + * archive recovery. And we'd like to keep crash recovery simple, to + * avoid introducing bugs that could affect you when recovering after + * crash. * * After this point, we can no longer assume that we're the only * process in addition to postmaster! Also, fsync requests are @@ -6649,8 +6654,8 @@ StartupXLOG(void) { /* * We have reached the end of base backup, the point where - * the minimum recovery point in pg_control indicates. - * The data on disk is now consistent. Reset backupStartPoint + * the minimum recovery point in pg_control indicates. The + * data on disk is now consistent. Reset backupStartPoint * and backupEndPoint. */ elog(DEBUG1, "end of backup reached"); @@ -6863,9 +6868,9 @@ StartupXLOG(void) oldestActiveXID = PrescanPreparedTransactions(NULL, NULL); /* - * Update full_page_writes in shared memory and write an - * XLOG_FPW_CHANGE record before resource manager writes cleanup - * WAL records or checkpoint record is written. + * Update full_page_writes in shared memory and write an XLOG_FPW_CHANGE + * record before resource manager writes cleanup WAL records or checkpoint + * record is written. */ Insert->fullPageWrites = lastFullPageWrites; LocalSetXLogInsertAllowed(); @@ -6954,8 +6959,8 @@ StartupXLOG(void) LWLockRelease(ProcArrayLock); /* - * Start up the commit log and subtrans, if not already done for - * hot standby. + * Start up the commit log and subtrans, if not already done for hot + * standby. */ if (standbyState == STANDBY_DISABLED) { @@ -7705,9 +7710,9 @@ CreateCheckPoint(int flags) checkPoint.time = (pg_time_t) time(NULL); /* - * For Hot Standby, derive the oldestActiveXid before we fix the redo pointer. - * This allows us to begin accumulating changes to assemble our starting - * snapshot of locks and transactions. + * For Hot Standby, derive the oldestActiveXid before we fix the redo + * pointer. This allows us to begin accumulating changes to assemble our + * starting snapshot of locks and transactions. */ if (!shutdown && XLogStandbyInfoActive()) checkPoint.oldestActiveXid = GetOldestActiveTransactionId(); @@ -8062,7 +8067,7 @@ RecoveryRestartPoint(const CheckPoint *checkPoint) volatile XLogCtlData *xlogctl = XLogCtl; /* - * Is it safe to restartpoint? We must ask each of the resource managers + * Is it safe to restartpoint? We must ask each of the resource managers * whether they have any partial state information that might prevent a * correct restart from this point. If so, we skip this opportunity, but * return at the next checkpoint record for another try. @@ -8082,10 +8087,11 @@ RecoveryRestartPoint(const CheckPoint *checkPoint) } /* - * Also refrain from creating a restartpoint if we have seen any references - * to non-existent pages. Restarting recovery from the restartpoint would - * not see the references, so we would lose the cross-check that the pages - * belonged to a relation that was dropped later. + * Also refrain from creating a restartpoint if we have seen any + * references to non-existent pages. Restarting recovery from the + * restartpoint would not see the references, so we would lose the + * cross-check that the pages belonged to a relation that was dropped + * later. */ if (XLogHaveInvalidPages()) { @@ -8098,8 +8104,8 @@ RecoveryRestartPoint(const CheckPoint *checkPoint) } /* - * Copy the checkpoint record to shared memory, so that checkpointer - * can work out the next time it wants to perform a restartpoint. + * Copy the checkpoint record to shared memory, so that checkpointer can + * work out the next time it wants to perform a restartpoint. */ SpinLockAcquire(&xlogctl->info_lck); XLogCtl->lastCheckPointRecPtr = ReadRecPtr; @@ -8493,8 +8499,8 @@ UpdateFullPageWrites(void) * Do nothing if full_page_writes has not been changed. * * It's safe to check the shared full_page_writes without the lock, - * because we assume that there is no concurrently running process - * which can update it. + * because we assume that there is no concurrently running process which + * can update it. */ if (fullPageWrites == Insert->fullPageWrites) return; @@ -8505,8 +8511,8 @@ UpdateFullPageWrites(void) * It's always safe to take full page images, even when not strictly * required, but not the other round. So if we're setting full_page_writes * to true, first set it true and then write the WAL record. If we're - * setting it to false, first write the WAL record and then set the - * global flag. + * setting it to false, first write the WAL record and then set the global + * flag. */ if (fullPageWrites) { @@ -8516,12 +8522,12 @@ UpdateFullPageWrites(void) } /* - * Write an XLOG_FPW_CHANGE record. This allows us to keep - * track of full_page_writes during archive recovery, if required. + * Write an XLOG_FPW_CHANGE record. This allows us to keep track of + * full_page_writes during archive recovery, if required. */ if (XLogStandbyInfoActive() && !RecoveryInProgress()) { - XLogRecData rdata; + XLogRecData rdata; rdata.data = (char *) (&fullPageWrites); rdata.len = sizeof(bool); @@ -8561,7 +8567,7 @@ xlog_redo(XLogRecPtr lsn, XLogRecord *record) /* * We used to try to take the maximum of ShmemVariableCache->nextOid * and the recorded nextOid, but that fails if the OID counter wraps - * around. Since no OID allocation should be happening during replay + * around. Since no OID allocation should be happening during replay * anyway, better to just believe the record exactly. We still take * OidGenLock while setting the variable, just in case. */ @@ -8597,7 +8603,7 @@ xlog_redo(XLogRecPtr lsn, XLogRecord *record) !XLogRecPtrIsInvalid(ControlFile->backupStartPoint) && XLogRecPtrIsInvalid(ControlFile->backupEndPoint)) ereport(PANIC, - (errmsg("online backup was canceled, recovery cannot continue"))); + (errmsg("online backup was canceled, recovery cannot continue"))); /* * If we see a shutdown checkpoint, we know that nothing was running @@ -8797,9 +8803,9 @@ xlog_redo(XLogRecPtr lsn, XLogRecord *record) memcpy(&fpw, XLogRecGetData(record), sizeof(bool)); /* - * Update the LSN of the last replayed XLOG_FPW_CHANGE record - * so that do_pg_start_backup() and do_pg_stop_backup() can check - * whether full_page_writes has been disabled during online backup. + * Update the LSN of the last replayed XLOG_FPW_CHANGE record so that + * do_pg_start_backup() and do_pg_stop_backup() can check whether + * full_page_writes has been disabled during online backup. */ if (!fpw) { @@ -8825,7 +8831,7 @@ xlog_desc(StringInfo buf, uint8 xl_info, char *rec) CheckPoint *checkpoint = (CheckPoint *) rec; appendStringInfo(buf, "checkpoint: redo %X/%X; " - "tli %u; fpw %s; xid %u/%u; oid %u; multi %u; offset %u; " + "tli %u; fpw %s; xid %u/%u; oid %u; multi %u; offset %u; " "oldest xid %u in DB %u; oldest running xid %u; %s", checkpoint->redo.xlogid, checkpoint->redo.xrecoff, checkpoint->ThisTimeLineID, @@ -9115,8 +9121,8 @@ do_pg_start_backup(const char *backupidstr, bool fast, char **labelfile) errhint("WAL control functions cannot be executed during recovery."))); /* - * During recovery, we don't need to check WAL level. Because, if WAL level - * is not sufficient, it's impossible to get here during recovery. + * During recovery, we don't need to check WAL level. Because, if WAL + * level is not sufficient, it's impossible to get here during recovery. */ if (!backup_started_in_recovery && !XLogIsNeeded()) ereport(ERROR, @@ -9179,7 +9185,7 @@ do_pg_start_backup(const char *backupidstr, bool fast, char **labelfile) * old timeline IDs. That would otherwise happen if you called * pg_start_backup() right after restoring from a PITR archive: the * first WAL segment containing the startup checkpoint has pages in - * the beginning with the old timeline ID. That can cause trouble at + * the beginning with the old timeline ID. That can cause trouble at * recovery: we won't have a history file covering the old timeline if * pg_xlog directory was not included in the base backup and the WAL * archive was cleared too before starting the backup. @@ -9202,17 +9208,18 @@ do_pg_start_backup(const char *backupidstr, bool fast, char **labelfile) bool checkpointfpw; /* - * Force a CHECKPOINT. Aside from being necessary to prevent torn + * Force a CHECKPOINT. Aside from being necessary to prevent torn * page problems, this guarantees that two successive backup runs * will have different checkpoint positions and hence different * history file names, even if nothing happened in between. * - * During recovery, establish a restartpoint if possible. We use the last - * restartpoint as the backup starting checkpoint. This means that two - * successive backup runs can have same checkpoint positions. + * During recovery, establish a restartpoint if possible. We use + * the last restartpoint as the backup starting checkpoint. This + * means that two successive backup runs can have same checkpoint + * positions. * - * Since the fact that we are executing do_pg_start_backup() during - * recovery means that checkpointer is running, we can use + * Since the fact that we are executing do_pg_start_backup() + * during recovery means that checkpointer is running, we can use * RequestCheckpoint() to establish a restartpoint. * * We use CHECKPOINT_IMMEDIATE only if requested by user (via @@ -9237,12 +9244,12 @@ do_pg_start_backup(const char *backupidstr, bool fast, char **labelfile) { /* use volatile pointer to prevent code rearrangement */ volatile XLogCtlData *xlogctl = XLogCtl; - XLogRecPtr recptr; + XLogRecPtr recptr; /* - * Check to see if all WAL replayed during online backup (i.e., - * since last restartpoint used as backup starting checkpoint) - * contain full-page writes. + * Check to see if all WAL replayed during online backup + * (i.e., since last restartpoint used as backup starting + * checkpoint) contain full-page writes. */ SpinLockAcquire(&xlogctl->info_lck); recptr = xlogctl->lastFpwDisableRecPtr; @@ -9250,20 +9257,20 @@ do_pg_start_backup(const char *backupidstr, bool fast, char **labelfile) if (!checkpointfpw || XLByteLE(startpoint, recptr)) ereport(ERROR, - (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), - errmsg("WAL generated with full_page_writes=off was replayed " - "since last restartpoint"), - errhint("This means that the backup being taken on standby " - "is corrupt and should not be used. " - "Enable full_page_writes and run CHECKPOINT on the master, " - "and then try an online backup again."))); + (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), + errmsg("WAL generated with full_page_writes=off was replayed " + "since last restartpoint"), + errhint("This means that the backup being taken on standby " + "is corrupt and should not be used. " + "Enable full_page_writes and run CHECKPOINT on the master, " + "and then try an online backup again."))); /* * During recovery, since we don't use the end-of-backup WAL - * record and don't write the backup history file, the starting WAL - * location doesn't need to be unique. This means that two base - * backups started at the same time might use the same checkpoint - * as starting locations. + * record and don't write the backup history file, the + * starting WAL location doesn't need to be unique. This means + * that two base backups started at the same time might use + * the same checkpoint as starting locations. */ gotUniqueStartpoint = true; } @@ -9443,8 +9450,8 @@ do_pg_stop_backup(char *labelfile, bool waitforarchive) errhint("WAL control functions cannot be executed during recovery."))); /* - * During recovery, we don't need to check WAL level. Because, if WAL level - * is not sufficient, it's impossible to get here during recovery. + * During recovery, we don't need to check WAL level. Because, if WAL + * level is not sufficient, it's impossible to get here during recovery. */ if (!backup_started_in_recovery && !XLogIsNeeded()) ereport(ERROR, @@ -9537,9 +9544,9 @@ do_pg_stop_backup(char *labelfile, bool waitforarchive) remaining = strchr(labelfile, '\n') + 1; /* %n is not portable enough */ /* - * Parse the BACKUP FROM line. If we are taking an online backup from - * the standby, we confirm that the standby has not been promoted - * during the backup. + * Parse the BACKUP FROM line. If we are taking an online backup from the + * standby, we confirm that the standby has not been promoted during the + * backup. */ ptr = strstr(remaining, "BACKUP FROM:"); if (!ptr || sscanf(ptr, "BACKUP FROM: %19s\n", backupfrom) != 1) @@ -9555,30 +9562,30 @@ do_pg_stop_backup(char *labelfile, bool waitforarchive) "Try taking another online backup."))); /* - * During recovery, we don't write an end-of-backup record. We assume - * that pg_control was backed up last and its minimum recovery - * point can be available as the backup end location. Since we don't - * have an end-of-backup record, we use the pg_control value to check - * whether we've reached the end of backup when starting recovery from - * this backup. We have no way of checking if pg_control wasn't backed - * up last however. + * During recovery, we don't write an end-of-backup record. We assume that + * pg_control was backed up last and its minimum recovery point can be + * available as the backup end location. Since we don't have an + * end-of-backup record, we use the pg_control value to check whether + * we've reached the end of backup when starting recovery from this + * backup. We have no way of checking if pg_control wasn't backed up last + * however. * * We don't force a switch to new WAL file and wait for all the required - * files to be archived. This is okay if we use the backup to start - * the standby. But, if it's for an archive recovery, to ensure all the - * required files are available, a user should wait for them to be archived, - * or include them into the backup. + * files to be archived. This is okay if we use the backup to start the + * standby. But, if it's for an archive recovery, to ensure all the + * required files are available, a user should wait for them to be + * archived, or include them into the backup. * * We return the current minimum recovery point as the backup end * location. Note that it's would be bigger than the exact backup end - * location if the minimum recovery point is updated since the backup - * of pg_control. This is harmless for current uses. + * location if the minimum recovery point is updated since the backup of + * pg_control. This is harmless for current uses. * * XXX currently a backup history file is for informational and debug * purposes only. It's not essential for an online backup. Furthermore, * even if it's created, it will not be archived during recovery because - * an archiver is not invoked. So it doesn't seem worthwhile to write - * a backup history file during recovery. + * an archiver is not invoked. So it doesn't seem worthwhile to write a + * backup history file during recovery. */ if (backup_started_in_recovery) { @@ -9597,12 +9604,12 @@ do_pg_stop_backup(char *labelfile, bool waitforarchive) if (XLByteLE(startpoint, recptr)) ereport(ERROR, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), - errmsg("WAL generated with full_page_writes=off was replayed " - "during online backup"), - errhint("This means that the backup being taken on standby " - "is corrupt and should not be used. " - "Enable full_page_writes and run CHECKPOINT on the master, " - "and then try an online backup again."))); + errmsg("WAL generated with full_page_writes=off was replayed " + "during online backup"), + errhint("This means that the backup being taken on standby " + "is corrupt and should not be used. " + "Enable full_page_writes and run CHECKPOINT on the master, " + "and then try an online backup again."))); LWLockAcquire(ControlFileLock, LW_SHARED); @@ -9905,10 +9912,11 @@ read_backup_label(XLogRecPtr *checkPointLoc, bool *backupEndRequired, ereport(FATAL, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), errmsg("invalid data in file \"%s\"", BACKUP_LABEL_FILE))); + /* - * BACKUP METHOD and BACKUP FROM lines are new in 9.2. We can't - * restore from an older backup anyway, but since the information on it - * is not strictly required, don't error out if it's missing for some reason. + * BACKUP METHOD and BACKUP FROM lines are new in 9.2. We can't restore + * from an older backup anyway, but since the information on it is not + * strictly required, don't error out if it's missing for some reason. */ if (fscanf(lfp, "BACKUP METHOD: %19s\n", backuptype) == 1) { @@ -10050,8 +10058,8 @@ XLogPageRead(XLogRecPtr *RecPtr, int emode, bool fetching_ckpt, if (readFile >= 0 && !XLByteInSeg(*RecPtr, readId, readSeg)) { /* - * Request a restartpoint if we've replayed too much - * xlog since the last one. + * Request a restartpoint if we've replayed too much xlog since the + * last one. */ if (StandbyMode && bgwriterLaunched) { diff --git a/src/backend/access/transam/xlogutils.c b/src/backend/access/transam/xlogutils.c index f286cdfc073b39e1f352e101f40be27710e84136..6ddcc59b37a45941fd17f8e0b7d3ff0976b346e0 100644 --- a/src/backend/access/transam/xlogutils.c +++ b/src/backend/access/transam/xlogutils.c @@ -80,10 +80,10 @@ log_invalid_page(RelFileNode node, ForkNumber forkno, BlockNumber blkno, /* * Once recovery has reached a consistent state, the invalid-page table * should be empty and remain so. If a reference to an invalid page is - * found after consistency is reached, PANIC immediately. This might - * seem aggressive, but it's better than letting the invalid reference - * linger in the hash table until the end of recovery and PANIC there, - * which might come only much later if this is a standby server. + * found after consistency is reached, PANIC immediately. This might seem + * aggressive, but it's better than letting the invalid reference linger + * in the hash table until the end of recovery and PANIC there, which + * might come only much later if this is a standby server. */ if (reachedConsistency) { diff --git a/src/backend/catalog/aclchk.c b/src/backend/catalog/aclchk.c index 9315e79c991d81dcc35881d4ae24fcb71ea30804..45cd0808ce80c609373de189bb330659c54b9895 100644 --- a/src/backend/catalog/aclchk.c +++ b/src/backend/catalog/aclchk.c @@ -186,10 +186,10 @@ merge_acl_with_grant(Acl *old_acl, bool is_grant, foreach(j, grantees) { - AclItem aclitem; + AclItem aclitem; Acl *newer_acl; - aclitem. ai_grantee = lfirst_oid(j); + aclitem.ai_grantee = lfirst_oid(j); /* * Grant options can only be granted to individual roles, not PUBLIC. @@ -202,7 +202,7 @@ merge_acl_with_grant(Acl *old_acl, bool is_grant, (errcode(ERRCODE_INVALID_GRANT_OPERATION), errmsg("grant options can only be granted to roles"))); - aclitem. ai_grantor = grantorId; + aclitem.ai_grantor = grantorId; /* * The asymmetry in the conditions here comes from the spec. In @@ -3073,7 +3073,7 @@ ExecGrant_Type(InternalGrant *istmt) ereport(ERROR, (errcode(ERRCODE_INVALID_GRANT_OPERATION), errmsg("cannot set privileges of array types"), - errhint("Set the privileges of the element type instead."))); + errhint("Set the privileges of the element type instead."))); /* Used GRANT DOMAIN on a non-domain? */ if (istmt->objtype == ACL_OBJECT_DOMAIN && @@ -4184,7 +4184,7 @@ pg_type_aclmask(Oid type_oid, Oid roleid, AclMode mask, AclMaskHow how) /* "True" array types don't manage permissions of their own */ if (typeForm->typelem != 0 && typeForm->typlen == -1) { - Oid elttype_oid = typeForm->typelem; + Oid elttype_oid = typeForm->typelem; ReleaseSysCache(tuple); diff --git a/src/backend/catalog/dependency.c b/src/backend/catalog/dependency.c index db6769cb90b699313ef12dbb31d6176d6f3e7d7d..d4e1f76f315bcb58c48f386428956108bf10c766 100644 --- a/src/backend/catalog/dependency.c +++ b/src/backend/catalog/dependency.c @@ -173,7 +173,7 @@ static void reportDependentObjects(const ObjectAddresses *targetObjects, int msglevel, const ObjectAddress *origObject); static void deleteOneObject(const ObjectAddress *object, - Relation depRel, int32 flags); + Relation depRel, int32 flags); static void doDeletion(const ObjectAddress *object, int flags); static void AcquireDeletionLock(const ObjectAddress *object, int flags); static void ReleaseDeletionLock(const ObjectAddress *object); @@ -352,7 +352,8 @@ performMultipleDeletions(const ObjectAddresses *objects, free_object_addresses(targetObjects); /* - * We closed depRel earlier in deleteOneObject if doing a drop concurrently + * We closed depRel earlier in deleteOneObject if doing a drop + * concurrently */ if ((flags & PERFORM_DELETION_CONCURRENTLY) != PERFORM_DELETION_CONCURRENTLY) heap_close(depRel, RowExclusiveLock); @@ -424,7 +425,7 @@ deleteWhatDependsOn(const ObjectAddress *object, * Since this function is currently only used to clean out temporary * schemas, we pass PERFORM_DELETION_INTERNAL here, indicating that * the operation is an automatic system operation rather than a user - * action. If, in the future, this function is used for other + * action. If, in the future, this function is used for other * purposes, we might need to revisit this. */ deleteOneObject(thisobj, depRel, PERFORM_DELETION_INTERNAL); @@ -514,12 +515,12 @@ findDependentObjects(const ObjectAddress *object, /* * The target object might be internally dependent on some other object * (its "owner"), and/or be a member of an extension (also considered its - * owner). If so, and if we aren't recursing from the owning object, we + * owner). If so, and if we aren't recursing from the owning object, we * have to transform this deletion request into a deletion request of the * owning object. (We'll eventually recurse back to this object, but the - * owning object has to be visited first so it will be deleted after.) - * The way to find out about this is to scan the pg_depend entries that - * show what this object depends on. + * owning object has to be visited first so it will be deleted after.) The + * way to find out about this is to scan the pg_depend entries that show + * what this object depends on. */ ScanKeyInit(&key[0], Anum_pg_depend_classid, @@ -577,7 +578,7 @@ findDependentObjects(const ObjectAddress *object, /* * Exception 1a: if the owning object is listed in * pendingObjects, just release the caller's lock and - * return. We'll eventually complete the DROP when we + * return. We'll eventually complete the DROP when we * reach that entry in the pending list. */ if (pendingObjects && @@ -593,8 +594,8 @@ findDependentObjects(const ObjectAddress *object, * Exception 1b: if the owning object is the extension * currently being created/altered, it's okay to continue * with the deletion. This allows dropping of an - * extension's objects within the extension's scripts, - * as well as corner cases such as dropping a transient + * extension's objects within the extension's scripts, as + * well as corner cases such as dropping a transient * object created within such a script. */ if (creating_extension && @@ -618,8 +619,8 @@ findDependentObjects(const ObjectAddress *object, * it's okay to continue with the deletion. This holds when * recursing from a whole object that includes the nominal * other end as a component, too. Since there can be more - * than one "owning" object, we have to allow matches that - * are more than one level down in the stack. + * than one "owning" object, we have to allow matches that are + * more than one level down in the stack. */ if (stack_address_present_add_flags(&otherObject, 0, stack)) break; @@ -630,7 +631,7 @@ findDependentObjects(const ObjectAddress *object, * owning object. * * First, release caller's lock on this object and get - * deletion lock on the owning object. (We must release + * deletion lock on the owning object. (We must release * caller's lock to avoid deadlock against a concurrent * deletion of the owning object.) */ @@ -999,7 +1000,8 @@ deleteOneObject(const ObjectAddress *object, Relation depRel, int flags) /* DROP hook of the objects being removed */ if (object_access_hook) { - ObjectAccessDrop drop_arg; + ObjectAccessDrop drop_arg; + drop_arg.dropflags = flags; InvokeObjectAccessHook(OAT_DROP, object->classId, object->objectId, object->objectSubId, &drop_arg); @@ -1049,8 +1051,8 @@ deleteOneObject(const ObjectAddress *object, Relation depRel, int flags) object->objectSubId); /* - * Close depRel if we are doing a drop concurrently because it - * commits the transaction, so we don't want dangling references. + * Close depRel if we are doing a drop concurrently because it commits the + * transaction, so we don't want dangling references. */ if ((flags & PERFORM_DELETION_CONCURRENTLY) == PERFORM_DELETION_CONCURRENTLY) heap_close(depRel, RowExclusiveLock); @@ -1093,8 +1095,8 @@ doDeletion(const ObjectAddress *object, int flags) if (relKind == RELKIND_INDEX) { - bool concurrent = ((flags & PERFORM_DELETION_CONCURRENTLY) - == PERFORM_DELETION_CONCURRENTLY); + bool concurrent = ((flags & PERFORM_DELETION_CONCURRENTLY) + == PERFORM_DELETION_CONCURRENTLY); Assert(object->objectSubId == 0); index_drop(object->objectId, concurrent); diff --git a/src/backend/catalog/heap.c b/src/backend/catalog/heap.c index 8fc69ae72022a057ae35dc668dce60c37357596e..49e7644699383755529b93461e91558705d17699 100644 --- a/src/backend/catalog/heap.c +++ b/src/backend/catalog/heap.c @@ -1957,7 +1957,7 @@ StoreRelCheck(Relation rel, char *ccname, Node *expr, ccsrc, /* Source form of check constraint */ is_local, /* conislocal */ inhcount, /* coninhcount */ - is_no_inherit); /* connoinherit */ + is_no_inherit); /* connoinherit */ pfree(ccbin); pfree(ccsrc); @@ -1998,7 +1998,7 @@ StoreConstraints(Relation rel, List *cooked_constraints) break; case CONSTR_CHECK: StoreRelCheck(rel, con->name, con->expr, !con->skip_validation, - con->is_local, con->inhcount, con->is_no_inherit); + con->is_local, con->inhcount, con->is_no_inherit); numchecks++; break; default: @@ -2345,8 +2345,8 @@ MergeWithExistingConstraint(Relation rel, char *ccname, Node *expr, } /* OK to update the tuple */ ereport(NOTICE, - (errmsg("merging constraint \"%s\" with inherited definition", - ccname))); + (errmsg("merging constraint \"%s\" with inherited definition", + ccname))); simple_heap_update(conDesc, &tup->t_self, tup); CatalogUpdateIndexes(conDesc, tup); break; diff --git a/src/backend/catalog/index.c b/src/backend/catalog/index.c index 998379c8af3e4dee11a8bd55367c0606f8259cc7..9e8b1cc49bc0cbd4d94e3cc4fa802ca59395a68b 100644 --- a/src/backend/catalog/index.c +++ b/src/backend/catalog/index.c @@ -1155,7 +1155,7 @@ index_constraint_create(Relation heapRelation, NULL, NULL, true, /* islocal */ - 0, /* inhcount */ + 0, /* inhcount */ false); /* noinherit */ /* @@ -1324,8 +1324,8 @@ index_drop(Oid indexId, bool concurrent) CheckTableNotInUse(userIndexRelation, "DROP INDEX"); /* - * Drop Index concurrently is similar in many ways to creating an - * index concurrently, so some actions are similar to DefineIndex() + * Drop Index concurrently is similar in many ways to creating an index + * concurrently, so some actions are similar to DefineIndex() */ if (concurrent) { @@ -1339,7 +1339,7 @@ index_drop(Oid indexId, bool concurrent) indexRelation = heap_open(IndexRelationId, RowExclusiveLock); tuple = SearchSysCacheCopy1(INDEXRELID, - ObjectIdGetDatum(indexId)); + ObjectIdGetDatum(indexId)); if (!HeapTupleIsValid(tuple)) elog(ERROR, "cache lookup failed for index %u", indexId); indexForm = (Form_pg_index) GETSTRUCT(tuple); @@ -1373,15 +1373,15 @@ index_drop(Oid indexId, bool concurrent) * will be marked not indisvalid, so that no one else tries to either * insert into it or use it for queries. * - * We must commit our current transaction so that the index update becomes - * visible; then start another. Note that all the data structures we just - * built are lost in the commit. The only data we keep past here are the - * relation IDs. + * We must commit our current transaction so that the index update + * becomes visible; then start another. Note that all the data + * structures we just built are lost in the commit. The only data we + * keep past here are the relation IDs. * * Before committing, get a session-level lock on the table, to ensure * that neither it nor the index can be dropped before we finish. This - * cannot block, even if someone else is waiting for access, because we - * already have the same lock within our transaction. + * cannot block, even if someone else is waiting for access, because + * we already have the same lock within our transaction. */ LockRelationIdForSession(&heaprelid, ShareUpdateExclusiveLock); LockRelationIdForSession(&indexrelid, ShareUpdateExclusiveLock); @@ -1391,23 +1391,23 @@ index_drop(Oid indexId, bool concurrent) StartTransactionCommand(); /* - * Now we must wait until no running transaction could have the table open - * with the old list of indexes. To do this, inquire which xacts - * currently would conflict with AccessExclusiveLock on the table -- ie, - * which ones have a lock of any kind on the table. Then wait for each of - * these xacts to commit or abort. Note we do not need to worry about - * xacts that open the table for writing after this point; they will see - * the index as invalid when they open the relation. + * Now we must wait until no running transaction could have the table + * open with the old list of indexes. To do this, inquire which xacts + * currently would conflict with AccessExclusiveLock on the table -- + * ie, which ones have a lock of any kind on the table. Then wait for + * each of these xacts to commit or abort. Note we do not need to + * worry about xacts that open the table for writing after this point; + * they will see the index as invalid when they open the relation. * - * Note: the reason we use actual lock acquisition here, rather than just - * checking the ProcArray and sleeping, is that deadlock is possible if - * one of the transactions in question is blocked trying to acquire an - * exclusive lock on our table. The lock code will detect deadlock and - * error out properly. + * Note: the reason we use actual lock acquisition here, rather than + * just checking the ProcArray and sleeping, is that deadlock is + * possible if one of the transactions in question is blocked trying + * to acquire an exclusive lock on our table. The lock code will + * detect deadlock and error out properly. * - * Note: GetLockConflicts() never reports our own xid, hence we need not - * check for that. Also, prepared xacts are not reported, which is fine - * since they certainly aren't going to do anything more. + * Note: GetLockConflicts() never reports our own xid, hence we need + * not check for that. Also, prepared xacts are not reported, which + * is fine since they certainly aren't going to do anything more. */ old_lockholders = GetLockConflicts(&heaplocktag, AccessExclusiveLock); @@ -1786,7 +1786,7 @@ index_update_stats(Relation rel, if (rd_rel->relkind != RELKIND_INDEX) relallvisible = visibilitymap_count(rel); - else /* don't bother for indexes */ + else /* don't bother for indexes */ relallvisible = 0; if (rd_rel->relpages != (int32) relpages) diff --git a/src/backend/catalog/namespace.c b/src/backend/catalog/namespace.c index e92efd863ed74fb77425333d772c194c3d36851b..1b6bb3bb6d7e02c065b3950bf83a715341aa8110 100644 --- a/src/backend/catalog/namespace.c +++ b/src/backend/catalog/namespace.c @@ -226,7 +226,7 @@ Datum pg_is_other_temp_schema(PG_FUNCTION_ARGS); Oid RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, bool missing_ok, bool nowait, - RangeVarGetRelidCallback callback, void *callback_arg) + RangeVarGetRelidCallback callback, void *callback_arg) { uint64 inval_count; Oid relId; @@ -247,20 +247,20 @@ RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, } /* - * DDL operations can change the results of a name lookup. Since all - * such operations will generate invalidation messages, we keep track - * of whether any such messages show up while we're performing the - * operation, and retry until either (1) no more invalidation messages - * show up or (2) the answer doesn't change. + * DDL operations can change the results of a name lookup. Since all such + * operations will generate invalidation messages, we keep track of + * whether any such messages show up while we're performing the operation, + * and retry until either (1) no more invalidation messages show up or (2) + * the answer doesn't change. * * But if lockmode = NoLock, then we assume that either the caller is OK * with the answer changing under them, or that they already hold some * appropriate lock, and therefore return the first answer we get without - * checking for invalidation messages. Also, if the requested lock is + * checking for invalidation messages. Also, if the requested lock is * already held, no LockRelationOid will not AcceptInvalidationMessages, * so we may fail to notice a change. We could protect against that case - * by calling AcceptInvalidationMessages() before beginning this loop, - * but that would add a significant amount overhead, so for now we don't. + * by calling AcceptInvalidationMessages() before beginning this loop, but + * that would add a significant amount overhead, so for now we don't. */ for (;;) { @@ -282,17 +282,18 @@ RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, if (relation->relpersistence == RELPERSISTENCE_TEMP) { if (!OidIsValid(myTempNamespace)) - relId = InvalidOid; /* this probably can't happen? */ + relId = InvalidOid; /* this probably can't happen? */ else { if (relation->schemaname) { - Oid namespaceId; + Oid namespaceId; + namespaceId = LookupExplicitNamespace(relation->schemaname); if (namespaceId != myTempNamespace) ereport(ERROR, (errcode(ERRCODE_INVALID_TABLE_DEFINITION), - errmsg("temporary tables cannot specify a schema name"))); + errmsg("temporary tables cannot specify a schema name"))); } relId = get_relname_relid(relation->relname, myTempNamespace); @@ -315,12 +316,12 @@ RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, /* * Invoke caller-supplied callback, if any. * - * This callback is a good place to check permissions: we haven't taken - * the table lock yet (and it's really best to check permissions before - * locking anything!), but we've gotten far enough to know what OID we - * think we should lock. Of course, concurrent DDL might change things - * while we're waiting for the lock, but in that case the callback will - * be invoked again for the new OID. + * This callback is a good place to check permissions: we haven't + * taken the table lock yet (and it's really best to check permissions + * before locking anything!), but we've gotten far enough to know what + * OID we think we should lock. Of course, concurrent DDL might + * change things while we're waiting for the lock, but in that case + * the callback will be invoked again for the new OID. */ if (callback) callback(relation, relId, oldRelId, callback_arg); @@ -328,21 +329,21 @@ RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, /* * If no lock requested, we assume the caller knows what they're * doing. They should have already acquired a heavyweight lock on - * this relation earlier in the processing of this same statement, - * so it wouldn't be appropriate to AcceptInvalidationMessages() - * here, as that might pull the rug out from under them. + * this relation earlier in the processing of this same statement, so + * it wouldn't be appropriate to AcceptInvalidationMessages() here, as + * that might pull the rug out from under them. */ if (lockmode == NoLock) break; /* - * If, upon retry, we get back the same OID we did last time, then - * the invalidation messages we processed did not change the final - * answer. So we're done. + * If, upon retry, we get back the same OID we did last time, then the + * invalidation messages we processed did not change the final answer. + * So we're done. * * If we got a different OID, we've locked the relation that used to - * have this name rather than the one that does now. So release - * the lock. + * have this name rather than the one that does now. So release the + * lock. */ if (retry) { @@ -384,8 +385,8 @@ RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, break; /* - * Something may have changed. Let's repeat the name lookup, to - * make sure this name still references the same relation it did + * Something may have changed. Let's repeat the name lookup, to make + * sure this name still references the same relation it did * previously. */ retry = true; @@ -550,8 +551,8 @@ RangeVarGetAndCheckCreationNamespace(RangeVar *relation, relid = InvalidOid; /* - * In bootstrap processing mode, we don't bother with permissions - * or locking. Permissions might not be working yet, and locking is + * In bootstrap processing mode, we don't bother with permissions or + * locking. Permissions might not be working yet, and locking is * unnecessary. */ if (IsBootstrapProcessingMode()) diff --git a/src/backend/catalog/objectaddress.c b/src/backend/catalog/objectaddress.c index d133f64776ee59a9611d3cf094a4a3832314781d..5a06fcbf41d5184adb538768b8f24747c2b687e2 100644 --- a/src/backend/catalog/objectaddress.c +++ b/src/backend/catalog/objectaddress.c @@ -75,10 +75,10 @@ */ typedef struct { - Oid class_oid; /* oid of catalog */ - Oid oid_index_oid; /* oid of index on system oid column */ - int oid_catcache_id; /* id of catcache on system oid column */ - AttrNumber attnum_namespace; /* attnum of namespace field */ + Oid class_oid; /* oid of catalog */ + Oid oid_index_oid; /* oid of index on system oid column */ + int oid_catcache_id; /* id of catcache on system oid column */ + AttrNumber attnum_namespace; /* attnum of namespace field */ } ObjectPropertyType; static ObjectPropertyType ObjectProperty[] = @@ -286,13 +286,13 @@ get_object_address(ObjectType objtype, List *objname, List *objargs, for (;;) { /* - * Remember this value, so that, after looking up the object name - * and locking it, we can check whether any invalidation messages - * have been processed that might require a do-over. + * Remember this value, so that, after looking up the object name and + * locking it, we can check whether any invalidation messages have + * been processed that might require a do-over. */ inval_count = SharedInvalidMessageCounter; - /* Look up object address. */ + /* Look up object address. */ switch (objtype) { case OBJECT_INDEX: @@ -367,7 +367,7 @@ get_object_address(ObjectType objtype, List *objname, List *objargs, case OBJECT_OPCLASS: case OBJECT_OPFAMILY: address = get_object_address_opcf(objtype, - objname, objargs, missing_ok); + objname, objargs, missing_ok); break; case OBJECT_LARGEOBJECT: Assert(list_length(objname) == 1); @@ -377,10 +377,10 @@ get_object_address(ObjectType objtype, List *objname, List *objargs, if (!LargeObjectExists(address.objectId)) { if (!missing_ok) - ereport(ERROR, - (errcode(ERRCODE_UNDEFINED_OBJECT), - errmsg("large object %u does not exist", - address.objectId))); + ereport(ERROR, + (errcode(ERRCODE_UNDEFINED_OBJECT), + errmsg("large object %u does not exist", + address.objectId))); } break; case OBJECT_CAST: @@ -475,8 +475,8 @@ get_object_address(ObjectType objtype, List *objname, List *objargs, * At this point, we've resolved the name to an OID and locked the * corresponding database object. However, it's possible that by the * time we acquire the lock on the object, concurrent DDL has modified - * the database in such a way that the name we originally looked up - * no longer resolves to that OID. + * the database in such a way that the name we originally looked up no + * longer resolves to that OID. * * We can be certain that this isn't an issue if (a) no shared * invalidation messages have been processed or (b) we've locked a @@ -488,12 +488,12 @@ get_object_address(ObjectType objtype, List *objname, List *objargs, * the relation, which is enough to freeze out any concurrent DDL. * * In all other cases, however, it's possible that the name we looked - * up no longer refers to the object we locked, so we retry the - * lookup and see whether we get the same answer. + * up no longer refers to the object we locked, so we retry the lookup + * and see whether we get the same answer. */ - if (inval_count == SharedInvalidMessageCounter || relation != NULL) - break; - old_address = address; + if (inval_count == SharedInvalidMessageCounter || relation != NULL) + break; + old_address = address; } /* Return the object address and the relation. */ @@ -621,7 +621,7 @@ get_relation_by_qualified_name(ObjectType objtype, List *objname, bool missing_ok) { Relation relation; - ObjectAddress address; + ObjectAddress address; address.classId = RelationRelationId; address.objectId = InvalidOid; @@ -721,8 +721,8 @@ get_object_address_relobject(ObjectType objtype, List *objname, address.objectSubId = 0; /* - * Caller is expecting to get back the relation, even though we - * didn't end up using it to find the rule. + * Caller is expecting to get back the relation, even though we didn't + * end up using it to find the rule. */ if (OidIsValid(address.objectId)) relation = heap_open(reloid, AccessShareLock); @@ -768,7 +768,7 @@ get_object_address_relobject(ObjectType objtype, List *objname, if (!OidIsValid(address.objectId)) { heap_close(relation, AccessShareLock); - relation = NULL; /* department of accident prevention */ + relation = NULL; /* department of accident prevention */ return address; } } @@ -834,9 +834,10 @@ static ObjectAddress get_object_address_type(ObjectType objtype, List *objname, bool missing_ok) { - ObjectAddress address; + ObjectAddress address; TypeName *typename; - Type tup; + Type tup; + typename = makeTypeNameFromNameList(objname); address.classId = TypeRelationId; @@ -1083,7 +1084,7 @@ get_object_namespace(const ObjectAddress *address) HeapTuple tuple; bool isnull; Oid oid; - ObjectPropertyType *property; + ObjectPropertyType *property; /* If not owned by a namespace, just return InvalidOid. */ property = get_object_property_data(address->classId); @@ -1122,5 +1123,5 @@ get_object_property_data(Oid class_id) return &ObjectProperty[index]; elog(ERROR, "unrecognized class id: %u", class_id); - return NULL; /* not reached */ + return NULL; /* not reached */ } diff --git a/src/backend/catalog/pg_constraint.c b/src/backend/catalog/pg_constraint.c index dca5d09ee69b832739287934b70355ecfbfd8db3..224859d76e76e6b577c4fca786033660698953ca 100644 --- a/src/backend/catalog/pg_constraint.c +++ b/src/backend/catalog/pg_constraint.c @@ -831,8 +831,8 @@ get_domain_constraint_oid(Oid typid, const char *conname, bool missing_ok) if (OidIsValid(conOid)) ereport(ERROR, (errcode(ERRCODE_DUPLICATE_OBJECT), - errmsg("domain \"%s\" has multiple constraints named \"%s\"", - format_type_be(typid), conname))); + errmsg("domain \"%s\" has multiple constraints named \"%s\"", + format_type_be(typid), conname))); conOid = HeapTupleGetOid(tuple); } } diff --git a/src/backend/catalog/pg_depend.c b/src/backend/catalog/pg_depend.c index 843f03d2c3877fe63c34a91b04eb825523f21a7d..8e584356064f1a309754914c273b80505dd94137 100644 --- a/src/backend/catalog/pg_depend.c +++ b/src/backend/catalog/pg_depend.c @@ -150,7 +150,7 @@ recordDependencyOnCurrentExtension(const ObjectAddress *object, /* Only need to check for existing membership if isReplace */ if (isReplace) { - Oid oldext; + Oid oldext; oldext = getExtensionOfObject(object->classId, object->objectId); if (OidIsValid(oldext)) diff --git a/src/backend/catalog/pg_proc.c b/src/backend/catalog/pg_proc.c index ae71b939175cd3682c503f4b44c9eeda580d57c4..599f04242f3eabd95ddec8522842c853148d9ac4 100644 --- a/src/backend/catalog/pg_proc.c +++ b/src/backend/catalog/pg_proc.c @@ -228,7 +228,7 @@ ProcedureCreate(const char *procedureName, /* * Do not allow polymorphic return type unless at least one input argument - * is polymorphic. ANYRANGE return type is even stricter: must have an + * is polymorphic. ANYRANGE return type is even stricter: must have an * ANYRANGE input (since we can't deduce the specific range type from * ANYELEMENT). Also, do not allow return type INTERNAL unless at least * one input argument is INTERNAL. diff --git a/src/backend/catalog/pg_shdepend.c b/src/backend/catalog/pg_shdepend.c index a67aebbdb6a1def6e1337e83729904c9e9100dd2..1edf950c560290b3aad2442f2c10acfec8644a06 100644 --- a/src/backend/catalog/pg_shdepend.c +++ b/src/backend/catalog/pg_shdepend.c @@ -1287,7 +1287,7 @@ shdepReassignOwned(List *roleids, Oid newrole) ereport(ERROR, (errcode(ERRCODE_DEPENDENT_OBJECTS_STILL_EXIST), errmsg("cannot reassign ownership of objects owned by %s because they are required by the database system", - getObjectDescription(&obj)))); + getObjectDescription(&obj)))); /* * There's no need to tell the whole truth, which is that we diff --git a/src/backend/catalog/storage.c b/src/backend/catalog/storage.c index 97ca95b6c8d37b35cc3d4fa1085480a3375940de..993bc49c2a2d6b73dbde6fb9682adc9afbbe32f9 100644 --- a/src/backend/catalog/storage.c +++ b/src/backend/catalog/storage.c @@ -500,8 +500,8 @@ smgr_redo(XLogRecPtr lsn, XLogRecord *record) /* * Forcibly create relation if it doesn't exist (which suggests that * it was dropped somewhere later in the WAL sequence). As in - * XLogReadBuffer, we prefer to recreate the rel and replay the log - * as best we can until the drop is seen. + * XLogReadBuffer, we prefer to recreate the rel and replay the log as + * best we can until the drop is seen. */ smgrcreate(reln, MAIN_FORKNUM, true); diff --git a/src/backend/commands/analyze.c b/src/backend/commands/analyze.c index 225ea866bf52d9c419b62bb14aeb6908d5ce1c87..9612a276f35f780bed1bd0b39b8231db5410dd5d 100644 --- a/src/backend/commands/analyze.c +++ b/src/backend/commands/analyze.c @@ -96,11 +96,11 @@ static void compute_index_stats(Relation onerel, double totalrows, MemoryContext col_context); static VacAttrStats *examine_attribute(Relation onerel, int attnum, Node *index_expr); -static int acquire_sample_rows(Relation onerel, int elevel, +static int acquire_sample_rows(Relation onerel, int elevel, HeapTuple *rows, int targrows, double *totalrows, double *totaldeadrows); static int compare_rows(const void *a, const void *b); -static int acquire_inherited_sample_rows(Relation onerel, int elevel, +static int acquire_inherited_sample_rows(Relation onerel, int elevel, HeapTuple *rows, int targrows, double *totalrows, double *totaldeadrows); static void update_attstats(Oid relid, bool inh, @@ -118,7 +118,7 @@ analyze_rel(Oid relid, VacuumStmt *vacstmt, BufferAccessStrategy bstrategy) Relation onerel; int elevel; AcquireSampleRowsFunc acquirefunc = NULL; - BlockNumber relpages = 0; + BlockNumber relpages = 0; /* Select logging level */ if (vacstmt->options & VACOPT_VERBOSE) @@ -205,8 +205,8 @@ analyze_rel(Oid relid, VacuumStmt *vacstmt, BufferAccessStrategy bstrategy) } /* - * Check that it's a plain table or foreign table; we used to do this - * in get_rel_oids() but seems safer to check after we've locked the + * Check that it's a plain table or foreign table; we used to do this in + * get_rel_oids() but seems safer to check after we've locked the * relation. */ if (onerel->rd_rel->relkind == RELKIND_RELATION) @@ -235,8 +235,8 @@ analyze_rel(Oid relid, VacuumStmt *vacstmt, BufferAccessStrategy bstrategy) if (!ok) { ereport(WARNING, - (errmsg("skipping \"%s\" --- cannot analyze this foreign table", - RelationGetRelationName(onerel)))); + (errmsg("skipping \"%s\" --- cannot analyze this foreign table", + RelationGetRelationName(onerel)))); relation_close(onerel, ShareUpdateExclusiveLock); return; } @@ -464,8 +464,8 @@ do_analyze_rel(Relation onerel, VacuumStmt *vacstmt, /* * Determine how many rows we need to sample, using the worst case from * all analyzable columns. We use a lower bound of 100 rows to avoid - * possible overflow in Vitter's algorithm. (Note: that will also be - * the target in the corner case where there are no analyzable columns.) + * possible overflow in Vitter's algorithm. (Note: that will also be the + * target in the corner case where there are no analyzable columns.) */ targrows = 100; for (i = 0; i < attr_cnt; i++) @@ -1337,7 +1337,7 @@ anl_get_next_S(double t, int n, double *stateptr) double V, quot; - V = anl_random_fract(); /* Generate V */ + V = anl_random_fract(); /* Generate V */ S = 0; t += 1; /* Note: "num" in Vitter's code is always equal to t - n */ @@ -1398,7 +1398,7 @@ anl_get_next_S(double t, int n, double *stateptr) y *= numer / denom; denom -= 1; } - W = exp(-log(anl_random_fract()) / n); /* Generate W in advance */ + W = exp(-log(anl_random_fract()) / n); /* Generate W in advance */ if (exp(log(y) / n) <= (t + X) / t) break; } diff --git a/src/backend/commands/cluster.c b/src/backend/commands/cluster.c index 349d13034e32975356c755e76533ac76d704abd2..a72b0ad5ff24bd50b04bb576cb3e1ba0b897ae67 100644 --- a/src/backend/commands/cluster.c +++ b/src/backend/commands/cluster.c @@ -594,10 +594,10 @@ make_new_heap(Oid OIDOldHeap, Oid NewTableSpace) OldHeapDesc = RelationGetDescr(OldHeap); /* - * Note that the NewHeap will not - * receive any of the defaults or constraints associated with the OldHeap; - * we don't need 'em, and there's no reason to spend cycles inserting them - * into the catalogs only to delete them. + * Note that the NewHeap will not receive any of the defaults or + * constraints associated with the OldHeap; we don't need 'em, and there's + * no reason to spend cycles inserting them into the catalogs only to + * delete them. */ /* diff --git a/src/backend/commands/copy.c b/src/backend/commands/copy.c index 1d1eacd3fbd53cfa1564ee2830007cf969631e4f..98bcb2fcf3370c72b0f0a7c0df76ebe4512e9ab0 100644 --- a/src/backend/commands/copy.c +++ b/src/backend/commands/copy.c @@ -150,7 +150,7 @@ typedef struct CopyStateData Oid *typioparams; /* array of element types for in_functions */ int *defmap; /* array of default att numbers */ ExprState **defexprs; /* array of default att expressions */ - bool volatile_defexprs; /* is any of defexprs volatile? */ + bool volatile_defexprs; /* is any of defexprs volatile? */ /* * These variables are used to reduce overhead in textual COPY FROM. @@ -566,11 +566,11 @@ CopyGetData(CopyState cstate, void *databuf, int minread, int maxread) if (mtype == EOF) ereport(ERROR, (errcode(ERRCODE_CONNECTION_FAILURE), - errmsg("unexpected EOF on client connection with an open transaction"))); + errmsg("unexpected EOF on client connection with an open transaction"))); if (pq_getmessage(cstate->fe_msgbuf, 0)) ereport(ERROR, (errcode(ERRCODE_CONNECTION_FAILURE), - errmsg("unexpected EOF on client connection with an open transaction"))); + errmsg("unexpected EOF on client connection with an open transaction"))); switch (mtype) { case 'd': /* CopyData */ @@ -1861,6 +1861,7 @@ CopyFrom(CopyState cstate) uint64 processed = 0; bool useHeapMultiInsert; int nBufferedTuples = 0; + #define MAX_BUFFERED_TUPLES 1000 HeapTuple *bufferedTuples = NULL; /* initialize to silence warning */ Size bufferedTuplesSize = 0; @@ -1968,8 +1969,8 @@ CopyFrom(CopyState cstate) * processed and prepared for insertion are not there. */ if ((resultRelInfo->ri_TrigDesc != NULL && - (resultRelInfo->ri_TrigDesc->trig_insert_before_row || - resultRelInfo->ri_TrigDesc->trig_insert_instead_row)) || + (resultRelInfo->ri_TrigDesc->trig_insert_before_row || + resultRelInfo->ri_TrigDesc->trig_insert_instead_row)) || cstate->volatile_defexprs) { useHeapMultiInsert = false; @@ -2162,8 +2163,8 @@ CopyFromInsertBatch(CopyState cstate, EState *estate, CommandId mycid, int i; /* - * heap_multi_insert leaks memory, so switch to short-lived memory - * context before calling it. + * heap_multi_insert leaks memory, so switch to short-lived memory context + * before calling it. */ oldcontext = MemoryContextSwitchTo(GetPerTupleMemoryContext(estate)); heap_multi_insert(cstate->rel, @@ -2175,14 +2176,14 @@ CopyFromInsertBatch(CopyState cstate, EState *estate, CommandId mycid, MemoryContextSwitchTo(oldcontext); /* - * If there are any indexes, update them for all the inserted tuples, - * and run AFTER ROW INSERT triggers. + * If there are any indexes, update them for all the inserted tuples, and + * run AFTER ROW INSERT triggers. */ if (resultRelInfo->ri_NumIndices > 0) { for (i = 0; i < nBufferedTuples; i++) { - List *recheckIndexes; + List *recheckIndexes; ExecStoreTuple(bufferedTuples[i], myslot, InvalidBuffer, false); recheckIndexes = @@ -2194,6 +2195,7 @@ CopyFromInsertBatch(CopyState cstate, EState *estate, CommandId mycid, list_free(recheckIndexes); } } + /* * There's no indexes, but see if we need to run AFTER ROW INSERT triggers * anyway. diff --git a/src/backend/commands/createas.c b/src/backend/commands/createas.c index 5173f5a3081e6c6a161ec3801f86e8fa2545780d..dc0665e2a4137c9e76ded4158949c2ac36893580 100644 --- a/src/backend/commands/createas.c +++ b/src/backend/commands/createas.c @@ -62,12 +62,12 @@ void ExecCreateTableAs(CreateTableAsStmt *stmt, const char *queryString, ParamListInfo params, char *completionTag) { - Query *query = (Query *) stmt->query; + Query *query = (Query *) stmt->query; IntoClause *into = stmt->into; DestReceiver *dest; - List *rewritten; + List *rewritten; PlannedStmt *plan; - QueryDesc *queryDesc; + QueryDesc *queryDesc; ScanDirection dir; /* @@ -98,9 +98,9 @@ ExecCreateTableAs(CreateTableAsStmt *stmt, const char *queryString, * plancache.c. * * Because the rewriter and planner tend to scribble on the input, we make - * a preliminary copy of the source querytree. This prevents problems in + * a preliminary copy of the source querytree. This prevents problems in * the case that CTAS is in a portal or plpgsql function and is executed - * repeatedly. (See also the same hack in EXPLAIN and PREPARE.) + * repeatedly. (See also the same hack in EXPLAIN and PREPARE.) */ rewritten = QueryRewrite((Query *) copyObject(stmt->query)); @@ -115,10 +115,10 @@ ExecCreateTableAs(CreateTableAsStmt *stmt, const char *queryString, /* * Use a snapshot with an updated command ID to ensure this query sees - * results of any previously executed queries. (This could only matter - * if the planner executed an allegedly-stable function that changed - * the database contents, but let's do it anyway to be parallel to the - * EXPLAIN code path.) + * results of any previously executed queries. (This could only matter if + * the planner executed an allegedly-stable function that changed the + * database contents, but let's do it anyway to be parallel to the EXPLAIN + * code path.) */ PushCopiedSnapshot(GetActiveSnapshot()); UpdateActiveSnapshotCommandId(); @@ -211,12 +211,12 @@ intorel_startup(DestReceiver *self, int operation, TupleDesc typeinfo) DR_intorel *myState = (DR_intorel *) self; IntoClause *into = myState->into; CreateStmt *create; - Oid intoRelationId; - Relation intoRelationDesc; + Oid intoRelationId; + Relation intoRelationDesc; RangeTblEntry *rte; Datum toast_options; - ListCell *lc; - int attnum; + ListCell *lc; + int attnum; static char *validnsps[] = HEAP_RELOPT_NAMESPACES; Assert(into != NULL); /* else somebody forgot to set it */ @@ -237,8 +237,8 @@ intorel_startup(DestReceiver *self, int operation, TupleDesc typeinfo) create->if_not_exists = false; /* - * Build column definitions using "pre-cooked" type and collation info. - * If a column name list was specified in CREATE TABLE AS, override the + * Build column definitions using "pre-cooked" type and collation info. If + * a column name list was specified in CREATE TABLE AS, override the * column names derived from the query. (Too few column names are OK, too * many are not.) */ @@ -246,8 +246,8 @@ intorel_startup(DestReceiver *self, int operation, TupleDesc typeinfo) for (attnum = 0; attnum < typeinfo->natts; attnum++) { Form_pg_attribute attribute = typeinfo->attrs[attnum]; - ColumnDef *col = makeNode(ColumnDef); - TypeName *coltype = makeNode(TypeName); + ColumnDef *col = makeNode(ColumnDef); + TypeName *coltype = makeNode(TypeName); if (lc) { @@ -280,9 +280,9 @@ intorel_startup(DestReceiver *self, int operation, TupleDesc typeinfo) /* * It's possible that the column is of a collatable type but the - * collation could not be resolved, so double-check. (We must - * check this here because DefineRelation would adopt the type's - * default collation rather than complaining.) + * collation could not be resolved, so double-check. (We must check + * this here because DefineRelation would adopt the type's default + * collation rather than complaining.) */ if (!OidIsValid(col->collOid) && type_is_collatable(coltype->typeOid)) @@ -297,8 +297,8 @@ intorel_startup(DestReceiver *self, int operation, TupleDesc typeinfo) if (lc != NULL) ereport(ERROR, - (errcode(ERRCODE_SYNTAX_ERROR), - errmsg("CREATE TABLE AS specifies too many column names"))); + (errcode(ERRCODE_SYNTAX_ERROR), + errmsg("CREATE TABLE AS specifies too many column names"))); /* * Actually create the target table @@ -342,7 +342,7 @@ intorel_startup(DestReceiver *self, int operation, TupleDesc typeinfo) for (attnum = 1; attnum <= intoRelationDesc->rd_att->natts; attnum++) rte->modifiedCols = bms_add_member(rte->modifiedCols, - attnum - FirstLowInvalidHeapAttributeNumber); + attnum - FirstLowInvalidHeapAttributeNumber); ExecCheckRTPerms(list_make1(rte), true); diff --git a/src/backend/commands/dbcommands.c b/src/backend/commands/dbcommands.c index 90155b9c14b78f54d125c9704932b9b1a434e39e..b7224bde8707f9cb567afbe2c01a2dcdf4598316 100644 --- a/src/backend/commands/dbcommands.c +++ b/src/backend/commands/dbcommands.c @@ -695,8 +695,8 @@ check_encoding_locale_matches(int encoding, const char *collate, const char *cty errmsg("encoding \"%s\" does not match locale \"%s\"", pg_encoding_to_char(encoding), ctype), - errdetail("The chosen LC_CTYPE setting requires encoding \"%s\".", - pg_encoding_to_char(ctype_encoding)))); + errdetail("The chosen LC_CTYPE setting requires encoding \"%s\".", + pg_encoding_to_char(ctype_encoding)))); if (!(collate_encoding == encoding || collate_encoding == PG_SQL_ASCII || @@ -710,8 +710,8 @@ check_encoding_locale_matches(int encoding, const char *collate, const char *cty errmsg("encoding \"%s\" does not match locale \"%s\"", pg_encoding_to_char(encoding), collate), - errdetail("The chosen LC_COLLATE setting requires encoding \"%s\".", - pg_encoding_to_char(collate_encoding)))); + errdetail("The chosen LC_COLLATE setting requires encoding \"%s\".", + pg_encoding_to_char(collate_encoding)))); } /* Error cleanup callback for createdb */ @@ -784,7 +784,8 @@ dropdb(const char *dbname, bool missing_ok) /* DROP hook for the database being removed */ if (object_access_hook) { - ObjectAccessDrop drop_arg; + ObjectAccessDrop drop_arg; + memset(&drop_arg, 0, sizeof(ObjectAccessDrop)); InvokeObjectAccessHook(OAT_DROP, DatabaseRelationId, db_id, 0, &drop_arg); @@ -831,8 +832,7 @@ dropdb(const char *dbname, bool missing_ok) ReleaseSysCache(tup); /* - * Delete any comments or security labels associated with - * the database. + * Delete any comments or security labels associated with the database. */ DeleteSharedComments(db_id, DatabaseRelationId); DeleteSharedSecurityLabel(db_id, DatabaseRelationId); @@ -860,18 +860,18 @@ dropdb(const char *dbname, bool missing_ok) pgstat_drop_database(db_id); /* - * Tell checkpointer to forget any pending fsync and unlink requests for files - * in the database; else the fsyncs will fail at next checkpoint, or + * Tell checkpointer to forget any pending fsync and unlink requests for + * files in the database; else the fsyncs will fail at next checkpoint, or * worse, it will delete files that belong to a newly created database * with the same OID. */ ForgetDatabaseFsyncRequests(db_id); /* - * Force a checkpoint to make sure the checkpointer has received the message - * sent by ForgetDatabaseFsyncRequests. On Windows, this also ensures that - * background procs don't hold any open files, which would cause rmdir() to - * fail. + * Force a checkpoint to make sure the checkpointer has received the + * message sent by ForgetDatabaseFsyncRequests. On Windows, this also + * ensures that background procs don't hold any open files, which would + * cause rmdir() to fail. */ RequestCheckpoint(CHECKPOINT_IMMEDIATE | CHECKPOINT_FORCE | CHECKPOINT_WAIT); diff --git a/src/backend/commands/dropcmds.c b/src/backend/commands/dropcmds.c index 298940c7c42ba15f966b285c5b4ed86526ac90e7..1b8529ed84312620ba6cc174574ccc59a1aec17c 100644 --- a/src/backend/commands/dropcmds.c +++ b/src/backend/commands/dropcmds.c @@ -30,7 +30,7 @@ #include "utils/syscache.h" static void does_not_exist_skipping(ObjectType objtype, - List *objname, List *objargs); + List *objname, List *objargs); /* * Drop one or more objects. @@ -54,7 +54,7 @@ RemoveObjects(DropStmt *stmt) foreach(cell1, stmt->objects) { - ObjectAddress address; + ObjectAddress address; List *objname = lfirst(cell1); List *objargs = NIL; Relation relation = NULL; @@ -97,8 +97,8 @@ RemoveObjects(DropStmt *stmt) if (((Form_pg_proc) GETSTRUCT(tup))->proisagg) ereport(ERROR, (errcode(ERRCODE_WRONG_OBJECT_TYPE), - errmsg("\"%s\" is an aggregate function", - NameListToString(objname)), + errmsg("\"%s\" is an aggregate function", + NameListToString(objname)), errhint("Use DROP AGGREGATE to drop aggregate functions."))); ReleaseSysCache(tup); @@ -149,7 +149,7 @@ does_not_exist_skipping(ObjectType objtype, List *objname, List *objargs) break; case OBJECT_CONVERSION: msg = gettext_noop("conversion \"%s\" does not exist, skipping"); - name = NameListToString(objname); + name = NameListToString(objname); break; case OBJECT_SCHEMA: msg = gettext_noop("schema \"%s\" does not exist, skipping"); @@ -196,9 +196,9 @@ does_not_exist_skipping(ObjectType objtype, List *objname, List *objargs) case OBJECT_CAST: msg = gettext_noop("cast from type %s to type %s does not exist, skipping"); name = format_type_be(typenameTypeId(NULL, - (TypeName *) linitial(objname))); + (TypeName *) linitial(objname))); args = format_type_be(typenameTypeId(NULL, - (TypeName *) linitial(objargs))); + (TypeName *) linitial(objargs))); break; case OBJECT_TRIGGER: msg = gettext_noop("trigger \"%s\" for table \"%s\" does not exist, skipping"); @@ -231,7 +231,7 @@ does_not_exist_skipping(ObjectType objtype, List *objname, List *objargs) args = strVal(linitial(objargs)); break; default: - elog(ERROR, "unexpected object type (%d)", (int)objtype); + elog(ERROR, "unexpected object type (%d)", (int) objtype); break; } diff --git a/src/backend/commands/explain.c b/src/backend/commands/explain.c index e2b4b994b477a90af07759034eee49f3ddf16d9f..1e8f618a3476b9e171658900f92532b87b3cde97 100644 --- a/src/backend/commands/explain.c +++ b/src/backend/commands/explain.c @@ -117,7 +117,7 @@ ExplainQuery(ExplainStmt *stmt, const char *queryString, TupOutputState *tstate; List *rewritten; ListCell *lc; - bool timing_set = false; + bool timing_set = false; /* Initialize ExplainState. */ ExplainInitState(&es); @@ -169,7 +169,7 @@ ExplainQuery(ExplainStmt *stmt, const char *queryString, ereport(ERROR, (errcode(ERRCODE_INVALID_PARAMETER_VALUE), errmsg("EXPLAIN option BUFFERS requires ANALYZE"))); - + /* if the timing was not set explicitly, set default value */ es.timing = (timing_set) ? es.timing : es.analyze; @@ -340,9 +340,9 @@ ExplainOneUtility(Node *utilityStmt, IntoClause *into, ExplainState *es, if (IsA(utilityStmt, CreateTableAsStmt)) { /* - * We have to rewrite the contained SELECT and then pass it back - * to ExplainOneQuery. It's probably not really necessary to copy - * the contained parsetree another time, but let's be safe. + * We have to rewrite the contained SELECT and then pass it back to + * ExplainOneQuery. It's probably not really necessary to copy the + * contained parsetree another time, but let's be safe. */ CreateTableAsStmt *ctas = (CreateTableAsStmt *) utilityStmt; List *rewritten; @@ -1021,7 +1021,7 @@ ExplainNode(PlanState *planstate, List *ancestors, { if (planstate->instrument->need_timer) appendStringInfo(es->str, - " (actual time=%.3f..%.3f rows=%.0f loops=%.0f)", + " (actual time=%.3f..%.3f rows=%.0f loops=%.0f)", startup_sec, total_sec, rows, nloops); else appendStringInfo(es->str, @@ -1095,7 +1095,7 @@ ExplainNode(PlanState *planstate, List *ancestors, planstate, es); if (es->analyze) ExplainPropertyLong("Heap Fetches", - ((IndexOnlyScanState *) planstate)->ioss_HeapFetches, es); + ((IndexOnlyScanState *) planstate)->ioss_HeapFetches, es); break; case T_BitmapIndexScan: show_scan_qual(((BitmapIndexScan *) plan)->indexqualorig, @@ -1237,7 +1237,7 @@ ExplainNode(PlanState *planstate, List *ancestors, bool has_temp = (usage->temp_blks_read > 0 || usage->temp_blks_written > 0); bool has_timing = (!INSTR_TIME_IS_ZERO(usage->blk_read_time) || - !INSTR_TIME_IS_ZERO(usage->blk_write_time)); + !INSTR_TIME_IS_ZERO(usage->blk_write_time)); /* Show only positive counter values. */ if (has_shared || has_local || has_temp) @@ -1301,10 +1301,10 @@ ExplainNode(PlanState *planstate, List *ancestors, appendStringInfoString(es->str, "I/O Timings:"); if (!INSTR_TIME_IS_ZERO(usage->blk_read_time)) appendStringInfo(es->str, " read=%0.3f", - INSTR_TIME_GET_MILLISEC(usage->blk_read_time)); + INSTR_TIME_GET_MILLISEC(usage->blk_read_time)); if (!INSTR_TIME_IS_ZERO(usage->blk_write_time)) appendStringInfo(es->str, " write=%0.3f", - INSTR_TIME_GET_MILLISEC(usage->blk_write_time)); + INSTR_TIME_GET_MILLISEC(usage->blk_write_time)); appendStringInfoChar(es->str, '\n'); } } diff --git a/src/backend/commands/extension.c b/src/backend/commands/extension.c index 732791cc41366fa9770765d8b24f2b8b74203a97..cde3d60ee8896c72e4215c8e299e3b32541581ac 100644 --- a/src/backend/commands/extension.c +++ b/src/backend/commands/extension.c @@ -899,8 +899,8 @@ execute_extension_script(Oid extensionOid, ExtensionControlFile *control, { t_sql = DirectFunctionCall3(replace_text, t_sql, - CStringGetTextDatum("MODULE_PATHNAME"), - CStringGetTextDatum(control->module_pathname)); + CStringGetTextDatum("MODULE_PATHNAME"), + CStringGetTextDatum(control->module_pathname)); } /* And now back to C string */ @@ -1585,14 +1585,14 @@ RemoveExtensionById(Oid extId) * might write "DROP EXTENSION foo" in foo's own script files, as because * errors in dependency management in extension script files could give * rise to cases where an extension is dropped as a result of recursing - * from some contained object. Because of that, we must test for the case + * from some contained object. Because of that, we must test for the case * here, not at some higher level of the DROP EXTENSION command. */ if (extId == CurrentExtensionObject) ereport(ERROR, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), - errmsg("cannot drop extension \"%s\" because it is being modified", - get_extension_name(extId)))); + errmsg("cannot drop extension \"%s\" because it is being modified", + get_extension_name(extId)))); rel = heap_open(ExtensionRelationId, RowExclusiveLock); diff --git a/src/backend/commands/foreigncmds.c b/src/backend/commands/foreigncmds.c index 30135e6de8b027b7b932c188f146f545e9af5d89..342ecc293185b13dd0b73003ba49981a9f758f1c 100644 --- a/src/backend/commands/foreigncmds.c +++ b/src/backend/commands/foreigncmds.c @@ -166,7 +166,7 @@ transformGenericOptions(Oid catalogId, if (OidIsValid(fdwvalidator)) { - Datum valarg = result; + Datum valarg = result; /* * Pass a null options list as an empty array, so that validators @@ -215,13 +215,13 @@ RenameForeignDataWrapper(const char *oldname, const char *newname) if (!HeapTupleIsValid(tup)) ereport(ERROR, (errcode(ERRCODE_UNDEFINED_OBJECT), - errmsg("foreign-data wrapper \"%s\" does not exist", oldname))); + errmsg("foreign-data wrapper \"%s\" does not exist", oldname))); /* make sure the new name doesn't exist */ if (SearchSysCacheExists1(FOREIGNDATAWRAPPERNAME, CStringGetDatum(newname))) ereport(ERROR, (errcode(ERRCODE_DUPLICATE_OBJECT), - errmsg("foreign-data wrapper \"%s\" already exists", newname))); + errmsg("foreign-data wrapper \"%s\" already exists", newname))); /* must be owner of FDW */ if (!pg_foreign_data_wrapper_ownercheck(HeapTupleGetOid(tup), GetUserId())) @@ -364,7 +364,7 @@ AlterForeignDataWrapperOwner_oid(Oid fwdId, Oid newOwnerId) if (!HeapTupleIsValid(tup)) ereport(ERROR, (errcode(ERRCODE_UNDEFINED_OBJECT), - errmsg("foreign-data wrapper with OID %u does not exist", fwdId))); + errmsg("foreign-data wrapper with OID %u does not exist", fwdId))); AlterForeignDataWrapperOwner_internal(rel, tup, newOwnerId); diff --git a/src/backend/commands/functioncmds.c b/src/backend/commands/functioncmds.c index ff0836c141e25a91aba9308092cd1f0f15ec8768..13e30f4a556b993de247498f5a5e412acb68b77c 100644 --- a/src/backend/commands/functioncmds.c +++ b/src/backend/commands/functioncmds.c @@ -890,9 +890,9 @@ CreateFunction(CreateFunctionStmt *stmt, const char *queryString) ReleaseSysCache(languageTuple); /* - * Only superuser is allowed to create leakproof functions because - * it possibly allows unprivileged users to reference invisible tuples - * to be filtered out using views for row-level security. + * Only superuser is allowed to create leakproof functions because it + * possibly allows unprivileged users to reference invisible tuples to be + * filtered out using views for row-level security. */ if (isLeakProof && !superuser()) ereport(ERROR, @@ -1320,7 +1320,7 @@ AlterFunction(AlterFunctionStmt *stmt) if (intVal(leakproof_item->arg) && !superuser()) ereport(ERROR, (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE), - errmsg("only superuser can define a leakproof function"))); + errmsg("only superuser can define a leakproof function"))); procForm->proleakproof = intVal(leakproof_item->arg); } if (cost_item) diff --git a/src/backend/commands/indexcmds.c b/src/backend/commands/indexcmds.c index 6c909298b7d72fdfb08dcdbf2516b71a10a5ccbe..a68d500e5b42c374fa87c6e7e410b52a550118a0 100644 --- a/src/backend/commands/indexcmds.c +++ b/src/backend/commands/indexcmds.c @@ -96,7 +96,7 @@ static void RangeVarCallbackForReindexIndex(const RangeVar *relation, * concrete benefit for core types. * When a comparison or exclusion operator has a polymorphic input type, the - * actual input types must also match. This defends against the possibility + * actual input types must also match. This defends against the possibility * that operators could vary behavior in response to get_fn_expr_argtype(). * At present, this hazard is theoretical: check_exclusion_constraint() and * all core index access methods decline to set fn_expr for such calls. @@ -134,6 +134,7 @@ CheckIndexCompatible(Oid oldId, /* Caller should already have the relation locked in some way. */ relationId = RangeVarGetRelid(heapRelation, NoLock, false); + /* * We can pretend isconstraint = false unconditionally. It only serves to * decide the text of an error message that should never happen for us. @@ -157,10 +158,10 @@ CheckIndexCompatible(Oid oldId, ReleaseSysCache(tuple); /* - * Compute the operator classes, collations, and exclusion operators - * for the new index, so we can test whether it's compatible with the - * existing one. Note that ComputeIndexAttrs might fail here, but that's - * OK: DefineIndex would have called this function with the same arguments + * Compute the operator classes, collations, and exclusion operators for + * the new index, so we can test whether it's compatible with the existing + * one. Note that ComputeIndexAttrs might fail here, but that's OK: + * DefineIndex would have called this function with the same arguments * later on, and it would have failed then anyway. */ indexInfo = makeNode(IndexInfo); @@ -218,11 +219,11 @@ CheckIndexCompatible(Oid oldId, return false; /* For polymorphic opcintype, column type changes break compatibility. */ - irel = index_open(oldId, AccessShareLock); /* caller probably has a lock */ + irel = index_open(oldId, AccessShareLock); /* caller probably has a lock */ for (i = 0; i < old_natts; i++) { if (IsPolymorphicType(get_opclass_input_type(classObjectId[i])) && - irel->rd_att->attrs[i]->atttypid != typeObjectId[i]) + irel->rd_att->attrs[i]->atttypid != typeObjectId[i]) { ret = false; break; @@ -232,7 +233,8 @@ CheckIndexCompatible(Oid oldId, /* Any change in exclusion operator selections breaks compatibility. */ if (ret && indexInfo->ii_ExclusionOps != NULL) { - Oid *old_operators, *old_procs; + Oid *old_operators, + *old_procs; uint16 *old_strats; RelationGetExclusionInfo(irel, &old_operators, &old_procs, &old_strats); @@ -249,7 +251,7 @@ CheckIndexCompatible(Oid oldId, op_input_types(indexInfo->ii_ExclusionOps[i], &left, &right); if ((IsPolymorphicType(left) || IsPolymorphicType(right)) && - irel->rd_att->attrs[i]->atttypid != typeObjectId[i]) + irel->rd_att->attrs[i]->atttypid != typeObjectId[i]) { ret = false; break; @@ -1778,9 +1780,9 @@ RangeVarCallbackForReindexIndex(const RangeVar *relation, return; /* - * If the relation does exist, check whether it's an index. But note - * that the relation might have been dropped between the time we did the - * name lookup and now. In that case, there's nothing to do. + * If the relation does exist, check whether it's an index. But note that + * the relation might have been dropped between the time we did the name + * lookup and now. In that case, there's nothing to do. */ relkind = get_rel_relkind(relId); if (!relkind) @@ -1798,9 +1800,9 @@ RangeVarCallbackForReindexIndex(const RangeVar *relation, if (relId != oldRelId) { /* - * Lock level here should match reindex_index() heap lock. - * If the OID isn't valid, it means the index as concurrently dropped, - * which is not a problem for us; just return normally. + * Lock level here should match reindex_index() heap lock. If the OID + * isn't valid, it means the index as concurrently dropped, which is + * not a problem for us; just return normally. */ *heapOid = IndexGetRelation(relId, true); if (OidIsValid(*heapOid)) diff --git a/src/backend/commands/lockcmds.c b/src/backend/commands/lockcmds.c index fd3dcc3643007d40f7c0f9609898c520a98ced93..ab13a459000bc3e494f53042402ea1a59bd10f65 100644 --- a/src/backend/commands/lockcmds.c +++ b/src/backend/commands/lockcmds.c @@ -40,9 +40,9 @@ LockTableCommand(LockStmt *lockstmt) /* * During recovery we only accept these variations: LOCK TABLE foo IN - * ACCESS SHARE MODE LOCK TABLE foo IN ROW SHARE MODE LOCK TABLE foo - * IN ROW EXCLUSIVE MODE This test must match the restrictions defined - * in LockAcquire() + * ACCESS SHARE MODE LOCK TABLE foo IN ROW SHARE MODE LOCK TABLE foo IN + * ROW EXCLUSIVE MODE This test must match the restrictions defined in + * LockAcquire() */ if (lockstmt->mode > RowExclusiveLock) PreventCommandDuringRecovery("LOCK TABLE"); @@ -74,15 +74,16 @@ static void RangeVarCallbackForLockTable(const RangeVar *rv, Oid relid, Oid oldrelid, void *arg) { - LOCKMODE lockmode = * (LOCKMODE *) arg; + LOCKMODE lockmode = *(LOCKMODE *) arg; char relkind; AclResult aclresult; if (!OidIsValid(relid)) - return; /* doesn't exist, so no permissions check */ + return; /* doesn't exist, so no permissions check */ relkind = get_rel_relkind(relid); if (!relkind) - return; /* woops, concurrently dropped; no permissions check */ + return; /* woops, concurrently dropped; no permissions + * check */ /* Currently, we only allow plain tables to be locked */ if (relkind != RELKIND_RELATION) @@ -122,9 +123,10 @@ LockTableRecurse(Oid reloid, LOCKMODE lockmode, bool nowait) if (aclresult != ACLCHECK_OK) { char *relname = get_rel_name(childreloid); + if (!relname) - continue; /* child concurrently dropped, just skip it */ - aclcheck_error(aclresult, ACL_KIND_CLASS, relname); + continue; /* child concurrently dropped, just skip it */ + aclcheck_error(aclresult, ACL_KIND_CLASS, relname); } /* We have enough rights to lock the relation; do so. */ @@ -134,17 +136,18 @@ LockTableRecurse(Oid reloid, LOCKMODE lockmode, bool nowait) { /* try to throw error by name; relation could be deleted... */ char *relname = get_rel_name(childreloid); + if (!relname) - continue; /* child concurrently dropped, just skip it */ + continue; /* child concurrently dropped, just skip it */ ereport(ERROR, (errcode(ERRCODE_LOCK_NOT_AVAILABLE), errmsg("could not obtain lock on relation \"%s\"", - relname))); + relname))); } /* - * Even if we got the lock, child might have been concurrently dropped. - * If so, we can skip it. + * Even if we got the lock, child might have been concurrently + * dropped. If so, we can skip it. */ if (!SearchSysCacheExists1(RELOID, ObjectIdGetDatum(childreloid))) { diff --git a/src/backend/commands/opclasscmds.c b/src/backend/commands/opclasscmds.c index 87c889604e2ad0d40da4a550438c7d203de642b9..460b1d9ae288a73937106216f55dd32675a67668 100644 --- a/src/backend/commands/opclasscmds.c +++ b/src/backend/commands/opclasscmds.c @@ -1167,7 +1167,7 @@ assignProcTypes(OpFamilyMember *member, Oid amoid, Oid typeoid) if (procform->prorettype != INT4OID) ereport(ERROR, (errcode(ERRCODE_INVALID_OBJECT_DEFINITION), - errmsg("btree comparison procedures must return integer"))); + errmsg("btree comparison procedures must return integer"))); /* * If lefttype/righttype isn't specified, use the proc's input @@ -1188,7 +1188,7 @@ assignProcTypes(OpFamilyMember *member, Oid amoid, Oid typeoid) if (procform->prorettype != VOIDOID) ereport(ERROR, (errcode(ERRCODE_INVALID_OBJECT_DEFINITION), - errmsg("btree sort support procedures must return void"))); + errmsg("btree sort support procedures must return void"))); /* * Can't infer lefttype/righttype from proc, so use default rule @@ -1217,7 +1217,7 @@ assignProcTypes(OpFamilyMember *member, Oid amoid, Oid typeoid) /* * The default in CREATE OPERATOR CLASS is to use the class' opcintype as - * lefttype and righttype. In CREATE or ALTER OPERATOR FAMILY, opcintype + * lefttype and righttype. In CREATE or ALTER OPERATOR FAMILY, opcintype * isn't available, so make the user specify the types. */ if (!OidIsValid(member->lefttype)) diff --git a/src/backend/commands/prepare.c b/src/backend/commands/prepare.c index edd646e7c348f01ff4304c57aaa3873af2b65f09..2d87b1c69078edcf7ee47316ee4d391d3c0f40dc 100644 --- a/src/backend/commands/prepare.c +++ b/src/backend/commands/prepare.c @@ -174,7 +174,7 @@ PrepareQuery(PrepareStmt *stmt, const char *queryString) * ExecuteQuery --- implement the 'EXECUTE' utility statement. * * This code also supports CREATE TABLE ... AS EXECUTE. That case is - * indicated by passing a non-null intoClause. The DestReceiver is already + * indicated by passing a non-null intoClause. The DestReceiver is already * set up correctly for CREATE TABLE AS, but we still have to make a few * other adjustments here. * @@ -211,7 +211,7 @@ ExecuteQuery(ExecuteStmt *stmt, IntoClause *intoClause, { /* * Need an EState to evaluate parameters; must not delete it till end - * of query, in case parameters are pass-by-reference. Note that the + * of query, in case parameters are pass-by-reference. Note that the * passed-in "params" could possibly be referenced in the parameter * expressions. */ @@ -237,15 +237,15 @@ ExecuteQuery(ExecuteStmt *stmt, IntoClause *intoClause, /* * For CREATE TABLE ... AS EXECUTE, we must verify that the prepared * statement is one that produces tuples. Currently we insist that it be - * a plain old SELECT. In future we might consider supporting other + * a plain old SELECT. In future we might consider supporting other * things such as INSERT ... RETURNING, but there are a couple of issues * to be settled first, notably how WITH NO DATA should be handled in such * a case (do we really want to suppress execution?) and how to pass down * the OID-determining eflags (PortalStart won't handle them in such a * case, and for that matter it's not clear the executor will either). * - * For CREATE TABLE ... AS EXECUTE, we also have to ensure that the - * proper eflags and fetch count are passed to PortalStart/PortalRun. + * For CREATE TABLE ... AS EXECUTE, we also have to ensure that the proper + * eflags and fetch count are passed to PortalStart/PortalRun. */ if (intoClause) { @@ -658,7 +658,7 @@ ExplainExecuteQuery(ExecuteStmt *execstmt, IntoClause *into, ExplainState *es, { /* * Need an EState to evaluate parameters; must not delete it till end - * of query, in case parameters are pass-by-reference. Note that the + * of query, in case parameters are pass-by-reference. Note that the * passed-in "params" could possibly be referenced in the parameter * expressions. */ diff --git a/src/backend/commands/proclang.c b/src/backend/commands/proclang.c index 5d2e7dc1951ac26f3c7f2c9809e76e7dcc75fc57..354389c617f921affa6e5e383c2caebde7e292f9 100644 --- a/src/backend/commands/proclang.c +++ b/src/backend/commands/proclang.c @@ -133,7 +133,7 @@ CreateProceduralLanguage(CreatePLangStmt *stmt) false, /* isAgg */ false, /* isWindowFunc */ false, /* security_definer */ - false, /* isLeakProof */ + false, /* isLeakProof */ false, /* isStrict */ PROVOLATILE_VOLATILE, buildoidvector(funcargtypes, 0), @@ -210,7 +210,7 @@ CreateProceduralLanguage(CreatePLangStmt *stmt) false, /* isAgg */ false, /* isWindowFunc */ false, /* security_definer */ - false, /* isLeakProof */ + false, /* isLeakProof */ true, /* isStrict */ PROVOLATILE_VOLATILE, buildoidvector(funcargtypes, 1), diff --git a/src/backend/commands/seclabel.c b/src/backend/commands/seclabel.c index 2129f62e511f93d28c22b884c15edcd3e97cccea..c09a96e9f6cb813228babfc7b7ce78f111851f21 100644 --- a/src/backend/commands/seclabel.c +++ b/src/backend/commands/seclabel.c @@ -237,7 +237,7 @@ GetSecurityLabel(const ObjectAddress *object, const char *provider) return seclabel; } -/* +/* * SetSharedSecurityLabel is a helper function of SetSecurityLabel to * handle shared database objects. */ @@ -246,8 +246,8 @@ SetSharedSecurityLabel(const ObjectAddress *object, const char *provider, const char *label) { Relation pg_shseclabel; - ScanKeyData keys[4]; - SysScanDesc scan; + ScanKeyData keys[4]; + SysScanDesc scan; HeapTuple oldtup; HeapTuple newtup = NULL; Datum values[Natts_pg_shseclabel]; @@ -414,8 +414,8 @@ void DeleteSharedSecurityLabel(Oid objectId, Oid classId) { Relation pg_shseclabel; - ScanKeyData skey[2]; - SysScanDesc scan; + ScanKeyData skey[2]; + SysScanDesc scan; HeapTuple oldtup; ScanKeyInit(&skey[0], diff --git a/src/backend/commands/sequence.c b/src/backend/commands/sequence.c index 718658995e2f430495bbb0c88e81fa7290929c96..34b74f6c3844a3a172cd660357f8ccbc48f57f3c 100644 --- a/src/backend/commands/sequence.c +++ b/src/backend/commands/sequence.c @@ -430,7 +430,7 @@ AlterSequence(AlterSeqStmt *stmt) { ereport(NOTICE, (errmsg("relation \"%s\" does not exist, skipping", - stmt->sequence->relname))); + stmt->sequence->relname))); return; } @@ -514,12 +514,12 @@ nextval(PG_FUNCTION_ARGS) sequence = makeRangeVarFromNameList(textToQualifiedNameList(seqin)); /* - * XXX: This is not safe in the presence of concurrent DDL, but - * acquiring a lock here is more expensive than letting nextval_internal - * do it, since the latter maintains a cache that keeps us from hitting - * the lock manager more than once per transaction. It's not clear - * whether the performance penalty is material in practice, but for now, - * we do it this way. + * XXX: This is not safe in the presence of concurrent DDL, but acquiring + * a lock here is more expensive than letting nextval_internal do it, + * since the latter maintains a cache that keeps us from hitting the lock + * manager more than once per transaction. It's not clear whether the + * performance penalty is material in practice, but for now, we do it this + * way. */ relid = RangeVarGetRelid(sequence, NoLock, false); @@ -1543,9 +1543,9 @@ seq_redo(XLogRecPtr lsn, XLogRecord *record) * is also used for updating sequences, it's possible that a hot-standby * backend is examining the page concurrently; so we mustn't transiently * trash the buffer. The solution is to build the correct new page - * contents in local workspace and then memcpy into the buffer. Then - * only bytes that are supposed to change will change, even transiently. - * We must palloc the local page for alignment reasons. + * contents in local workspace and then memcpy into the buffer. Then only + * bytes that are supposed to change will change, even transiently. We + * must palloc the local page for alignment reasons. */ localpage = (Page) palloc(BufferGetPageSize(buffer)); diff --git a/src/backend/commands/tablecmds.c b/src/backend/commands/tablecmds.c index 6148bd62da8d1261396c159963dcad8e6b6a81cb..5c69cfb85a2e1b5b961a252780b2024f33c6a1dd 100644 --- a/src/backend/commands/tablecmds.c +++ b/src/backend/commands/tablecmds.c @@ -237,9 +237,9 @@ static const struct dropmsgstrings dropmsgstringarray[] = { struct DropRelationCallbackState { - char relkind; - Oid heapOid; - bool concurrent; + char relkind; + Oid heapOid; + bool concurrent; }; /* Alter table target-type flags for ATSimplePermissions */ @@ -372,8 +372,8 @@ static void ATPrepSetTableSpace(AlteredTableInfo *tab, Relation rel, char *tablespacename, LOCKMODE lockmode); static void ATExecSetTableSpace(Oid tableOid, Oid newTableSpace, LOCKMODE lockmode); static void ATExecSetRelOptions(Relation rel, List *defList, - AlterTableType operation, - LOCKMODE lockmode); + AlterTableType operation, + LOCKMODE lockmode); static void ATExecEnableDisableTrigger(Relation rel, char *trigname, char fires_when, bool skip_system, LOCKMODE lockmode); static void ATExecEnableDisableRule(Relation rel, char *rulename, @@ -752,7 +752,7 @@ RemoveRelations(DropStmt *drop) if (drop->behavior == DROP_CASCADE) ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("DROP INDEX CONCURRENTLY does not support CASCADE"))); + errmsg("DROP INDEX CONCURRENTLY does not support CASCADE"))); } /* @@ -799,7 +799,7 @@ RemoveRelations(DropStmt *drop) RangeVar *rel = makeRangeVarFromNameList((List *) lfirst(cell)); Oid relOid; ObjectAddress obj; - struct DropRelationCallbackState state; + struct DropRelationCallbackState state; /* * These next few steps are a great deal like relation_openrv, but we @@ -914,9 +914,9 @@ RangeVarCallbackForDropRelation(const RangeVar *rel, Oid relOid, Oid oldRelOid, /* * In DROP INDEX, attempt to acquire lock on the parent table before * locking the index. index_drop() will need this anyway, and since - * regular queries lock tables before their indexes, we risk deadlock - * if we do it the other way around. No error if we don't find a - * pg_index entry, though --- the relation may have been droppd. + * regular queries lock tables before their indexes, we risk deadlock if + * we do it the other way around. No error if we don't find a pg_index + * entry, though --- the relation may have been droppd. */ if (relkind == RELKIND_INDEX && relOid != oldRelOid) { @@ -2322,12 +2322,12 @@ static void RangeVarCallbackForRenameAttribute(const RangeVar *rv, Oid relid, Oid oldrelid, void *arg) { - HeapTuple tuple; - Form_pg_class form; + HeapTuple tuple; + Form_pg_class form; tuple = SearchSysCache1(RELOID, ObjectIdGetDatum(relid)); if (!HeapTupleIsValid(tuple)) - return; /* concurrently dropped */ + return; /* concurrently dropped */ form = (Form_pg_class) GETSTRUCT(tuple); renameatt_check(relid, form, false); ReleaseSysCache(tuple); @@ -2351,7 +2351,7 @@ renameatt(RenameStmt *stmt) { ereport(NOTICE, (errmsg("relation \"%s\" does not exist, skipping", - stmt->relation->relname))); + stmt->relation->relname))); return; } @@ -2379,7 +2379,7 @@ rename_constraint_internal(Oid myrelid, { Relation targetrelation = NULL; Oid constraintOid; - HeapTuple tuple; + HeapTuple tuple; Form_pg_constraint con; AssertArg(!myrelid || !mytypid); @@ -2391,7 +2391,11 @@ rename_constraint_internal(Oid myrelid, else { targetrelation = relation_open(myrelid, AccessExclusiveLock); - /* don't tell it whether we're recursing; we allow changing typed tables here */ + + /* + * don't tell it whether we're recursing; we allow changing typed + * tables here + */ renameatt_check(myrelid, RelationGetForm(targetrelation), false); constraintOid = get_relation_constraint_oid(myrelid, oldconname, false); @@ -2408,9 +2412,9 @@ rename_constraint_internal(Oid myrelid, if (recurse) { List *child_oids, - *child_numparents; + *child_numparents; ListCell *lo, - *li; + *li; child_oids = find_all_inheritors(myrelid, AccessExclusiveLock, &child_numparents); @@ -2455,7 +2459,7 @@ rename_constraint_internal(Oid myrelid, ReleaseSysCache(tuple); if (targetrelation) - relation_close(targetrelation, NoLock); /* close rel but keep lock */ + relation_close(targetrelation, NoLock); /* close rel but keep lock */ } void @@ -2469,7 +2473,7 @@ RenameConstraint(RenameStmt *stmt) Relation rel; HeapTuple tup; - typid = typenameTypeId(NULL, makeTypeNameFromNameList(stmt->object)); + typid = typenameTypeId(NULL, makeTypeNameFromNameList(stmt->object)); rel = heap_open(TypeRelationId, RowExclusiveLock); tup = SearchSysCache1(TYPEOID, ObjectIdGetDatum(typid)); if (!HeapTupleIsValid(tup)) @@ -2490,9 +2494,9 @@ RenameConstraint(RenameStmt *stmt) rename_constraint_internal(relid, typid, stmt->subname, stmt->newname, - stmt->relation ? interpretInhOption(stmt->relation->inhOpt) : false, /* recursive? */ + stmt->relation ? interpretInhOption(stmt->relation->inhOpt) : false, /* recursive? */ false, /* recursing? */ - 0 /* expected inhcount */); + 0 /* expected inhcount */ ); } /* @@ -2507,8 +2511,8 @@ RenameRelation(RenameStmt *stmt) * Grab an exclusive lock on the target table, index, sequence or view, * which we will NOT release until end of transaction. * - * Lock level used here should match RenameRelationInternal, to avoid - * lock escalation. + * Lock level used here should match RenameRelationInternal, to avoid lock + * escalation. */ relid = RangeVarGetRelidExtended(stmt->relation, AccessExclusiveLock, stmt->missing_ok, false, @@ -2519,7 +2523,7 @@ RenameRelation(RenameStmt *stmt) { ereport(NOTICE, (errmsg("relation \"%s\" does not exist, skipping", - stmt->relation->relname))); + stmt->relation->relname))); return; } @@ -2702,11 +2706,11 @@ AlterTableLookupRelation(AlterTableStmt *stmt, LOCKMODE lockmode) * Thanks to the magic of MVCC, an error anywhere along the way rolls back * the whole operation; we don't have to do anything special to clean up. * - * The caller must lock the relation, with an appropriate lock level + * The caller must lock the relation, with an appropriate lock level * for the subcommands requested. Any subcommand that needs to rewrite * tuples in the table forces the whole command to be executed with * AccessExclusiveLock (actually, that is currently required always, but - * we hope to relax it at some point). We pass the lock level down + * we hope to relax it at some point). We pass the lock level down * so that we can apply it recursively to inherited tables. Note that the * lock level we want as we recurse might well be higher than required for * that specific subcommand. So we pass down the overall lock requirement, @@ -2773,22 +2777,22 @@ LOCKMODE AlterTableGetLockLevel(List *cmds) { /* - * Late in 9.1 dev cycle a number of issues were uncovered with access - * to catalog relations, leading to the decision to re-enforce all DDL - * at AccessExclusiveLock level by default. + * Late in 9.1 dev cycle a number of issues were uncovered with access to + * catalog relations, leading to the decision to re-enforce all DDL at + * AccessExclusiveLock level by default. * * The issues are that there is a pervasive assumption in the code that - * the catalogs will not be read unless an AccessExclusiveLock is held. - * If that rule is relaxed, we must protect against a number of potential + * the catalogs will not be read unless an AccessExclusiveLock is held. If + * that rule is relaxed, we must protect against a number of potential * effects - infrequent, but proven possible with test cases where * multiple DDL operations occur in a stream against frequently accessed * tables. * - * 1. Catalog tables are read using SnapshotNow, which has a race bug - * that allows a scan to return no valid rows even when one is present - * in the case of a commit of a concurrent update of the catalog table. - * SnapshotNow also ignores transactions in progress, so takes the - * latest committed version without waiting for the latest changes. + * 1. Catalog tables are read using SnapshotNow, which has a race bug that + * allows a scan to return no valid rows even when one is present in the + * case of a commit of a concurrent update of the catalog table. + * SnapshotNow also ignores transactions in progress, so takes the latest + * committed version without waiting for the latest changes. * * 2. Relcache needs to be internally consistent, so unless we lock the * definition during reads we have no way to guarantee that. @@ -3156,8 +3160,8 @@ ATPrepCmd(List **wqueue, Relation rel, AlterTableCmd *cmd, pass = AT_PASS_MISC; /* doesn't actually matter */ break; case AT_SetRelOptions: /* SET (...) */ - case AT_ResetRelOptions: /* RESET (...) */ - case AT_ReplaceRelOptions: /* reset them all, then set just these */ + case AT_ResetRelOptions: /* RESET (...) */ + case AT_ReplaceRelOptions: /* reset them all, then set just these */ ATSimplePermissions(rel, ATT_TABLE | ATT_INDEX | ATT_VIEW); /* This command never recurses */ /* No command-specific prep needed */ @@ -3344,8 +3348,8 @@ ATExecCmd(List **wqueue, AlteredTableInfo *tab, Relation rel, case AT_ValidateConstraint: /* VALIDATE CONSTRAINT */ ATExecValidateConstraint(rel, cmd->name, false, false, lockmode); break; - case AT_ValidateConstraintRecurse: /* VALIDATE CONSTRAINT with - * recursion */ + case AT_ValidateConstraintRecurse: /* VALIDATE CONSTRAINT with + * recursion */ ATExecValidateConstraint(rel, cmd->name, true, false, lockmode); break; case AT_DropConstraint: /* DROP CONSTRAINT */ @@ -3361,7 +3365,7 @@ ATExecCmd(List **wqueue, AlteredTableInfo *tab, Relation rel, case AT_AlterColumnType: /* ALTER COLUMN TYPE */ ATExecAlterColumnType(tab, rel, cmd, lockmode); break; - case AT_AlterColumnGenericOptions: /* ALTER COLUMN OPTIONS */ + case AT_AlterColumnGenericOptions: /* ALTER COLUMN OPTIONS */ ATExecAlterColumnGenericOptions(rel, cmd->name, (List *) cmd->def, lockmode); break; case AT_ChangeOwner: /* ALTER OWNER */ @@ -4725,7 +4729,7 @@ static void check_for_column_name_collision(Relation rel, const char *colname) { HeapTuple attTuple; - int attnum; + int attnum; /* * this test is deliberately not attisdropped-aware, since if one tries to @@ -4737,7 +4741,7 @@ check_for_column_name_collision(Relation rel, const char *colname) if (!HeapTupleIsValid(attTuple)) return; - attnum = ((Form_pg_attribute) GETSTRUCT(attTuple))->attnum; + attnum = ((Form_pg_attribute) GETSTRUCT(attTuple))->attnum; ReleaseSysCache(attTuple); /* @@ -4745,16 +4749,16 @@ check_for_column_name_collision(Relation rel, const char *colname) * names, since they are normally not shown and the user might otherwise * be confused about the reason for the conflict. */ - if (attnum <= 0) - ereport(ERROR, - (errcode(ERRCODE_DUPLICATE_COLUMN), - errmsg("column name \"%s\" conflicts with a system column name", - colname))); - else - ereport(ERROR, - (errcode(ERRCODE_DUPLICATE_COLUMN), - errmsg("column \"%s\" of relation \"%s\" already exists", - colname, RelationGetRelationName(rel)))); + if (attnum <= 0) + ereport(ERROR, + (errcode(ERRCODE_DUPLICATE_COLUMN), + errmsg("column name \"%s\" conflicts with a system column name", + colname))); + else + ereport(ERROR, + (errcode(ERRCODE_DUPLICATE_COLUMN), + errmsg("column \"%s\" of relation \"%s\" already exists", + colname, RelationGetRelationName(rel)))); } /* @@ -4999,8 +5003,8 @@ ATExecColumnDefault(Relation rel, const char *colName, * safety, but at present we do not expect anything to depend on the * default. * - * We treat removing the existing default as an internal operation when - * it is preparatory to adding a new default, but as a user-initiated + * We treat removing the existing default as an internal operation when it + * is preparatory to adding a new default, but as a user-initiated * operation when the user asked for a drop. */ RemoveAttrDefault(RelationGetRelid(rel), attnum, DROP_RESTRICT, false, @@ -5507,13 +5511,14 @@ ATExecAddIndex(AlteredTableInfo *tab, Relation rel, /* * If TryReuseIndex() stashed a relfilenode for us, we used it for the new - * index instead of building from scratch. The DROP of the old edition of + * index instead of building from scratch. The DROP of the old edition of * this index will have scheduled the storage for deletion at commit, so * cancel that pending deletion. */ if (OidIsValid(stmt->oldNode)) { Relation irel = index_open(new_index, NoLock); + RelationPreserveStorage(irel->rd_node, true); index_close(irel, NoLock); } @@ -5687,8 +5692,8 @@ ATAddCheckConstraint(List **wqueue, AlteredTableInfo *tab, Relation rel, */ newcons = AddRelationNewConstraints(rel, NIL, list_make1(copyObject(constr)), - recursing, /* allow_merge */ - !recursing); /* is_local */ + recursing, /* allow_merge */ + !recursing); /* is_local */ /* Add each to-be-validated constraint to Phase 3's queue */ foreach(lcon, newcons) @@ -5743,7 +5748,7 @@ ATAddCheckConstraint(List **wqueue, AlteredTableInfo *tab, Relation rel, /* * Check if ONLY was specified with ALTER TABLE. If so, allow the - * contraint creation only if there are no children currently. Error out + * contraint creation only if there are no children currently. Error out * otherwise. */ if (!recurse && children != NIL) @@ -6064,11 +6069,11 @@ ATAddForeignKeyConstraint(AlteredTableInfo *tab, Relation rel, /* * Upon a change to the cast from the FK column to its pfeqop - * operand, revalidate the constraint. For this evaluation, a + * operand, revalidate the constraint. For this evaluation, a * binary coercion cast is equivalent to no cast at all. While * type implementors should design implicit casts with an eye - * toward consistency of operations like equality, we cannot assume - * here that they have done so. + * toward consistency of operations like equality, we cannot + * assume here that they have done so. * * A function with a polymorphic argument could change behavior * arbitrarily in response to get_fn_expr_argtype(). Therefore, @@ -6082,7 +6087,7 @@ ATAddForeignKeyConstraint(AlteredTableInfo *tab, Relation rel, * Necessarily, the primary key column must then be of the domain * type. Since the constraint was previously valid, all values on * the foreign side necessarily exist on the primary side and in - * turn conform to the domain. Consequently, we need not treat + * turn conform to the domain. Consequently, we need not treat * domains specially here. * * Since we require that all collations share the same notion of @@ -6091,8 +6096,8 @@ ATAddForeignKeyConstraint(AlteredTableInfo *tab, Relation rel, * * We need not directly consider the PK type. It's necessarily * binary coercible to the opcintype of the unique index column, - * and ri_triggers.c will only deal with PK datums in terms of that - * opcintype. Changing the opcintype also changes pfeqop. + * and ri_triggers.c will only deal with PK datums in terms of + * that opcintype. Changing the opcintype also changes pfeqop. */ old_check_ok = (new_pathtype == old_pathtype && new_castfunc == old_castfunc && @@ -6144,11 +6149,11 @@ ATAddForeignKeyConstraint(AlteredTableInfo *tab, Relation rel, createForeignKeyTriggers(rel, fkconstraint, constrOid, indexOid); /* - * Tell Phase 3 to check that the constraint is satisfied by existing rows. - * We can skip this during table creation, when requested explicitly by - * specifying NOT VALID in an ADD FOREIGN KEY command, and when we're - * recreating a constraint following a SET DATA TYPE operation that did not - * impugn its validity. + * Tell Phase 3 to check that the constraint is satisfied by existing + * rows. We can skip this during table creation, when requested explicitly + * by specifying NOT VALID in an ADD FOREIGN KEY command, and when we're + * recreating a constraint following a SET DATA TYPE operation that did + * not impugn its validity. */ if (!old_check_ok && !fkconstraint->skip_validation) { @@ -6236,12 +6241,12 @@ ATExecValidateConstraint(Relation rel, char *constrName, bool recurse, Relation refrel; /* - * Triggers are already in place on both tables, so a concurrent write - * that alters the result here is not possible. Normally we can run a - * query here to do the validation, which would only require - * AccessShareLock. In some cases, it is possible that we might need - * to fire triggers to perform the check, so we take a lock at - * RowShareLock level just in case. + * Triggers are already in place on both tables, so a concurrent + * write that alters the result here is not possible. Normally we + * can run a query here to do the validation, which would only + * require AccessShareLock. In some cases, it is possible that we + * might need to fire triggers to perform the check, so we take a + * lock at RowShareLock level just in case. */ refrel = heap_open(con->confrelid, RowShareLock); @@ -6278,7 +6283,7 @@ ATExecValidateConstraint(Relation rel, char *constrName, bool recurse, */ foreach(child, children) { - Oid childoid = lfirst_oid(child); + Oid childoid = lfirst_oid(child); Relation childrel; if (childoid == RelationGetRelid(rel)) @@ -6662,27 +6667,28 @@ checkFkeyPermissions(Relation rel, int16 *attnums, int natts) static void validateCheckConstraint(Relation rel, HeapTuple constrtup) { - EState *estate; - Datum val; - char *conbin; - Expr *origexpr; - List *exprstate; - TupleDesc tupdesc; - HeapScanDesc scan; - HeapTuple tuple; - ExprContext *econtext; - MemoryContext oldcxt; + EState *estate; + Datum val; + char *conbin; + Expr *origexpr; + List *exprstate; + TupleDesc tupdesc; + HeapScanDesc scan; + HeapTuple tuple; + ExprContext *econtext; + MemoryContext oldcxt; TupleTableSlot *slot; Form_pg_constraint constrForm; - bool isnull; + bool isnull; constrForm = (Form_pg_constraint) GETSTRUCT(constrtup); estate = CreateExecutorState(); + /* * XXX this tuple doesn't really come from a syscache, but this doesn't - * matter to SysCacheGetAttr, because it only wants to be able to fetch the - * tupdesc + * matter to SysCacheGetAttr, because it only wants to be able to fetch + * the tupdesc */ val = SysCacheGetAttr(CONSTROID, constrtup, Anum_pg_constraint_conbin, &isnull); @@ -7132,7 +7138,7 @@ ATExecDropConstraint(Relation rel, const char *constrName, con = (Form_pg_constraint) GETSTRUCT(copy_tuple); - if (con->coninhcount <= 0) /* shouldn't happen */ + if (con->coninhcount <= 0) /* shouldn't happen */ elog(ERROR, "relation %u has non-inherited constraint \"%s\"", childrelid, constrName); @@ -7140,8 +7146,7 @@ ATExecDropConstraint(Relation rel, const char *constrName, { /* * If the child constraint has other definition sources, just - * decrement its inheritance count; if not, recurse to delete - * it. + * decrement its inheritance count; if not, recurse to delete it. */ if (con->coninhcount == 1 && !con->conislocal) { @@ -7164,9 +7169,9 @@ ATExecDropConstraint(Relation rel, const char *constrName, else { /* - * If we were told to drop ONLY in this table (no recursion), - * we need to mark the inheritors' constraints as locally - * defined rather than inherited. + * If we were told to drop ONLY in this table (no recursion), we + * need to mark the inheritors' constraints as locally defined + * rather than inherited. */ con->coninhcount--; con->conislocal = true; @@ -7315,8 +7320,8 @@ ATPrepAlterColumnType(List **wqueue, if (transform == NULL) ereport(ERROR, (errcode(ERRCODE_DATATYPE_MISMATCH), - errmsg("column \"%s\" cannot be cast automatically to type %s", - colName, format_type_be(targettype)), + errmsg("column \"%s\" cannot be cast automatically to type %s", + colName, format_type_be(targettype)), errhint("Specify a USING expression to perform the conversion."))); /* Fix collations after all else */ @@ -7483,8 +7488,8 @@ ATExecAlterColumnType(AlteredTableInfo *tab, Relation rel, if (defaultexpr == NULL) ereport(ERROR, (errcode(ERRCODE_DATATYPE_MISMATCH), - errmsg("default for column \"%s\" cannot be cast automatically to type %s", - colName, format_type_be(targettype)))); + errmsg("default for column \"%s\" cannot be cast automatically to type %s", + colName, format_type_be(targettype)))); } else defaultexpr = NULL; @@ -8060,7 +8065,8 @@ TryReuseIndex(Oid oldId, IndexStmt *stmt) stmt->indexParams, stmt->excludeOpNames)) { - Relation irel = index_open(oldId, NoLock); + Relation irel = index_open(oldId, NoLock); + stmt->oldNode = irel->rd_node.relNode; index_close(irel, NoLock); } @@ -8085,7 +8091,7 @@ TryReuseForeignKey(Oid oldId, Constraint *con) int i; Assert(con->contype == CONSTR_FOREIGN); - Assert(con->old_conpfeqop == NIL); /* already prepared this node */ + Assert(con->old_conpfeqop == NIL); /* already prepared this node */ tup = SearchSysCache1(CONSTROID, ObjectIdGetDatum(oldId)); if (!HeapTupleIsValid(tup)) /* should not happen */ @@ -8587,8 +8593,8 @@ ATExecSetRelOptions(Relation rel, List *defList, AlterTableType operation, /* Generate new proposed reloptions (text array) */ newOptions = transformRelOptions(isnull ? (Datum) 0 : datum, - defList, NULL, validnsps, false, - operation == AT_ResetRelOptions); + defList, NULL, validnsps, false, + operation == AT_ResetRelOptions); /* Validate */ switch (rel->rd_rel->relkind) @@ -8665,8 +8671,8 @@ ATExecSetRelOptions(Relation rel, List *defList, AlterTableType operation, } newOptions = transformRelOptions(isnull ? (Datum) 0 : datum, - defList, "toast", validnsps, false, - operation == AT_ResetRelOptions); + defList, "toast", validnsps, false, + operation == AT_ResetRelOptions); (void) heap_reloptions(RELKIND_TOASTVALUE, newOptions, true); @@ -9831,7 +9837,7 @@ AlterTableNamespace(AlterObjectSchemaStmt *stmt) { ereport(NOTICE, (errmsg("relation \"%s\" does not exist, skipping", - stmt->relation->relname))); + stmt->relation->relname))); return; } @@ -9848,10 +9854,10 @@ AlterTableNamespace(AlterObjectSchemaStmt *stmt) if (sequenceIsOwned(relid, &tableId, &colId)) ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("cannot move an owned sequence into another schema"), - errdetail("Sequence \"%s\" is linked to table \"%s\".", - RelationGetRelationName(rel), - get_rel_name(tableId)))); + errmsg("cannot move an owned sequence into another schema"), + errdetail("Sequence \"%s\" is linked to table \"%s\".", + RelationGetRelationName(rel), + get_rel_name(tableId)))); } /* Get and lock schema OID and check its permissions. */ @@ -10267,9 +10273,9 @@ RangeVarCallbackOwnsTable(const RangeVar *relation, return; /* - * If the relation does exist, check whether it's an index. But note - * that the relation might have been dropped between the time we did the - * name lookup and now. In that case, there's nothing to do. + * If the relation does exist, check whether it's an index. But note that + * the relation might have been dropped between the time we did the name + * lookup and now. In that case, there's nothing to do. */ relkind = get_rel_relkind(relId); if (!relkind) @@ -10292,16 +10298,16 @@ static void RangeVarCallbackForAlterRelation(const RangeVar *rv, Oid relid, Oid oldrelid, void *arg) { - Node *stmt = (Node *) arg; - ObjectType reltype; - HeapTuple tuple; - Form_pg_class classform; - AclResult aclresult; - char relkind; + Node *stmt = (Node *) arg; + ObjectType reltype; + HeapTuple tuple; + Form_pg_class classform; + AclResult aclresult; + char relkind; tuple = SearchSysCache1(RELOID, ObjectIdGetDatum(relid)); if (!HeapTupleIsValid(tuple)) - return; /* concurrently dropped */ + return; /* concurrently dropped */ classform = (Form_pg_class) GETSTRUCT(tuple); relkind = classform->relkind; @@ -10324,7 +10330,7 @@ RangeVarCallbackForAlterRelation(const RangeVar *rv, Oid relid, Oid oldrelid, */ if (IsA(stmt, RenameStmt)) { - aclresult = pg_namespace_aclcheck(classform->relnamespace, + aclresult = pg_namespace_aclcheck(classform->relnamespace, GetUserId(), ACL_CREATE); if (aclresult != ACLCHECK_OK) aclcheck_error(aclresult, ACL_KIND_NAMESPACE, @@ -10333,20 +10339,21 @@ RangeVarCallbackForAlterRelation(const RangeVar *rv, Oid relid, Oid oldrelid, } else if (IsA(stmt, AlterObjectSchemaStmt)) reltype = ((AlterObjectSchemaStmt *) stmt)->objectType; + else if (IsA(stmt, AlterTableStmt)) reltype = ((AlterTableStmt *) stmt)->relkind; else { - reltype = OBJECT_TABLE; /* placate compiler */ + reltype = OBJECT_TABLE; /* placate compiler */ elog(ERROR, "unrecognized node type: %d", (int) nodeTag(stmt)); } /* - * For compatibility with prior releases, we allow ALTER TABLE to be - * used with most other types of relations (but not composite types). - * We allow similar flexibility for ALTER INDEX in the case of RENAME, - * but not otherwise. Otherwise, the user must select the correct form - * of the command for the relation at issue. + * For compatibility with prior releases, we allow ALTER TABLE to be used + * with most other types of relations (but not composite types). We allow + * similar flexibility for ALTER INDEX in the case of RENAME, but not + * otherwise. Otherwise, the user must select the correct form of the + * command for the relation at issue. */ if (reltype == OBJECT_SEQUENCE && relkind != RELKIND_SEQUENCE) ereport(ERROR, @@ -10391,10 +10398,10 @@ RangeVarCallbackForAlterRelation(const RangeVar *rv, Oid relid, Oid oldrelid, errhint("Use ALTER FOREIGN TABLE instead."))); /* - * Don't allow ALTER TABLE .. SET SCHEMA on relations that can't be - * moved to a different schema, such as indexes and TOAST tables. + * Don't allow ALTER TABLE .. SET SCHEMA on relations that can't be moved + * to a different schema, such as indexes and TOAST tables. */ - if (IsA(stmt, AlterObjectSchemaStmt) && relkind != RELKIND_RELATION + if (IsA(stmt, AlterObjectSchemaStmt) &&relkind != RELKIND_RELATION && relkind != RELKIND_VIEW && relkind != RELKIND_SEQUENCE && relkind != RELKIND_FOREIGN_TABLE) ereport(ERROR, diff --git a/src/backend/commands/tablespace.c b/src/backend/commands/tablespace.c index 708bebb54d0493f15beab6e6a898dd63010baa27..da9cb2f30e955fa4c99b30c119867df45db0b3cd 100644 --- a/src/backend/commands/tablespace.c +++ b/src/backend/commands/tablespace.c @@ -437,7 +437,8 @@ DropTableSpace(DropTableSpaceStmt *stmt) /* DROP hook for the tablespace being removed */ if (object_access_hook) { - ObjectAccessDrop drop_arg; + ObjectAccessDrop drop_arg; + memset(&drop_arg, 0, sizeof(ObjectAccessDrop)); InvokeObjectAccessHook(OAT_DROP, TableSpaceRelationId, tablespaceoid, 0, &drop_arg); @@ -638,7 +639,7 @@ create_tablespace_directories(const char *location, const Oid tablespaceoid) * Attempt to remove filesystem infrastructure for the tablespace. * * 'redo' indicates we are redoing a drop from XLOG; in that case we should - * not throw an ERROR for problems, just LOG them. The worst consequence of + * not throw an ERROR for problems, just LOG them. The worst consequence of * not removing files here would be failure to release some disk space, which * does not justify throwing an error that would require manual intervention * to get the database running again. @@ -678,7 +679,7 @@ destroy_tablespace_directories(Oid tablespaceoid, bool redo) * with a warning. This is because even though ProcessUtility disallows * DROP TABLESPACE in a transaction block, it's possible that a previous * DROP failed and rolled back after removing the tablespace directories - * and/or symlink. We want to allow a new DROP attempt to succeed at + * and/or symlink. We want to allow a new DROP attempt to succeed at * removing the catalog entries (and symlink if still present), so we * should not give a hard error here. */ @@ -1199,14 +1200,14 @@ check_temp_tablespaces(char **newval, void **extra, GucSource source) } /* - * In an interactive SET command, we ereport for bad info. When + * In an interactive SET command, we ereport for bad info. When * source == PGC_S_TEST, we are checking the argument of an ALTER - * DATABASE SET or ALTER USER SET command. pg_dumpall dumps all + * DATABASE SET or ALTER USER SET command. pg_dumpall dumps all * roles before tablespaces, so if we're restoring a pg_dumpall * script the tablespace might not yet exist, but will be created - * later. Because of that, issue a NOTICE if source == PGC_S_TEST, - * but accept the value anyway. Otherwise, silently ignore any - * bad list elements. + * later. Because of that, issue a NOTICE if source == + * PGC_S_TEST, but accept the value anyway. Otherwise, silently + * ignore any bad list elements. */ curoid = get_tablespace_oid(curname, source <= PGC_S_TEST); if (curoid == InvalidOid) @@ -1493,10 +1494,10 @@ tblspc_redo(XLogRecPtr lsn, XLogRecord *record) * files then do conflict processing and try again, if currently * enabled. * - * Other possible reasons for failure include bollixed file permissions - * on a standby server when they were okay on the primary, etc etc. - * There's not much we can do about that, so just remove what we can - * and press on. + * Other possible reasons for failure include bollixed file + * permissions on a standby server when they were okay on the primary, + * etc etc. There's not much we can do about that, so just remove what + * we can and press on. */ if (!destroy_tablespace_directories(xlrec->ts_id, true)) { @@ -1513,8 +1514,8 @@ tblspc_redo(XLogRecPtr lsn, XLogRecord *record) if (!destroy_tablespace_directories(xlrec->ts_id, true)) ereport(LOG, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), - errmsg("directories for tablespace %u could not be removed", - xlrec->ts_id), + errmsg("directories for tablespace %u could not be removed", + xlrec->ts_id), errhint("You can remove the directories manually if necessary."))); } } diff --git a/src/backend/commands/trigger.c b/src/backend/commands/trigger.c index 1218d033d1fbc761e7778bcc8fbfd0b3caa29ee6..4399a274465cf4a5e653604bdd8ea92b83ee1937 100644 --- a/src/backend/commands/trigger.c +++ b/src/backend/commands/trigger.c @@ -199,8 +199,8 @@ CreateTrigger(CreateTrigStmt *stmt, const char *queryString, /* * We must take a lock on the target relation to protect against * concurrent drop. It's not clear that AccessShareLock is strong - * enough, but we certainly need at least that much... otherwise, - * we might end up creating a pg_constraint entry referencing a + * enough, but we certainly need at least that much... otherwise, we + * might end up creating a pg_constraint entry referencing a * nonexistent table. */ constrrelid = RangeVarGetRelid(stmt->constrrel, AccessShareLock, false); @@ -494,8 +494,8 @@ CreateTrigger(CreateTrigStmt *stmt, const char *queryString, * can skip this for internally generated triggers, since the name * modification above should be sufficient. * - * NOTE that this is cool only because we have AccessExclusiveLock on - * the relation, so the trigger set won't be changing underneath us. + * NOTE that this is cool only because we have AccessExclusiveLock on the + * relation, so the trigger set won't be changing underneath us. */ if (!isInternal) { @@ -1168,27 +1168,27 @@ static void RangeVarCallbackForRenameTrigger(const RangeVar *rv, Oid relid, Oid oldrelid, void *arg) { - HeapTuple tuple; - Form_pg_class form; + HeapTuple tuple; + Form_pg_class form; tuple = SearchSysCache1(RELOID, ObjectIdGetDatum(relid)); if (!HeapTupleIsValid(tuple)) - return; /* concurrently dropped */ + return; /* concurrently dropped */ form = (Form_pg_class) GETSTRUCT(tuple); /* only tables and views can have triggers */ - if (form->relkind != RELKIND_RELATION && form->relkind != RELKIND_VIEW) - ereport(ERROR, - (errcode(ERRCODE_WRONG_OBJECT_TYPE), - errmsg("\"%s\" is not a table or view", rv->relname))); + if (form->relkind != RELKIND_RELATION && form->relkind != RELKIND_VIEW) + ereport(ERROR, + (errcode(ERRCODE_WRONG_OBJECT_TYPE), + errmsg("\"%s\" is not a table or view", rv->relname))); /* you must own the table to rename one of its triggers */ - if (!pg_class_ownercheck(relid, GetUserId())) - aclcheck_error(ACLCHECK_NOT_OWNER, ACL_KIND_CLASS, rv->relname); - if (!allowSystemTableMods && IsSystemClass(form)) - ereport(ERROR, - (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE), - errmsg("permission denied: \"%s\" is a system catalog", + if (!pg_class_ownercheck(relid, GetUserId())) + aclcheck_error(ACLCHECK_NOT_OWNER, ACL_KIND_CLASS, rv->relname); + if (!allowSystemTableMods && IsSystemClass(form)) + ereport(ERROR, + (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE), + errmsg("permission denied: \"%s\" is a system catalog", rv->relname))); ReleaseSysCache(tuple); diff --git a/src/backend/commands/typecmds.c b/src/backend/commands/typecmds.c index 77559842e5301f2b31596242304ee7f78ee17afd..fdb5bdbc11ba65da7044959658965b8955215ea4 100644 --- a/src/backend/commands/typecmds.c +++ b/src/backend/commands/typecmds.c @@ -609,7 +609,7 @@ DefineType(List *names, List *parameters) F_ARRAY_SEND, /* send procedure */ typmodinOid, /* typmodin procedure */ typmodoutOid, /* typmodout procedure */ - F_ARRAY_TYPANALYZE, /* analyze procedure */ + F_ARRAY_TYPANALYZE, /* analyze procedure */ typoid, /* element type ID */ true, /* yes this is an array type */ InvalidOid, /* no further array type */ @@ -1140,7 +1140,7 @@ DefineEnum(CreateEnumStmt *stmt) F_ARRAY_SEND, /* send procedure */ InvalidOid, /* typmodin procedure - none */ InvalidOid, /* typmodout procedure - none */ - F_ARRAY_TYPANALYZE, /* analyze procedure */ + F_ARRAY_TYPANALYZE, /* analyze procedure */ enumTypeOid, /* element type ID */ true, /* yes this is an array type */ InvalidOid, /* no further array type */ @@ -1450,7 +1450,7 @@ DefineRange(CreateRangeStmt *stmt) F_ARRAY_SEND, /* send procedure */ InvalidOid, /* typmodin procedure - none */ InvalidOid, /* typmodout procedure - none */ - F_ARRAY_TYPANALYZE, /* analyze procedure */ + F_ARRAY_TYPANALYZE, /* analyze procedure */ typoid, /* element type ID */ true, /* yes this is an array type */ InvalidOid, /* no further array type */ @@ -1477,15 +1477,15 @@ DefineRange(CreateRangeStmt *stmt) * impossible to define a polymorphic constructor; we have to generate new * constructor functions explicitly for each range type. * - * We actually define 4 functions, with 0 through 3 arguments. This is just + * We actually define 4 functions, with 0 through 3 arguments. This is just * to offer more convenience for the user. */ static void makeRangeConstructors(const char *name, Oid namespace, Oid rangeOid, Oid subtype) { - static const char * const prosrc[2] = {"range_constructor2", - "range_constructor3"}; + static const char *const prosrc[2] = {"range_constructor2", + "range_constructor3"}; static const int pronargs[2] = {2, 3}; Oid constructorArgTypes[3]; @@ -1509,7 +1509,7 @@ makeRangeConstructors(const char *name, Oid namespace, constructorArgTypesVector = buildoidvector(constructorArgTypes, pronargs[i]); - procOid = ProcedureCreate(name, /* name: same as range type */ + procOid = ProcedureCreate(name, /* name: same as range type */ namespace, /* namespace */ false, /* replace */ false, /* returns set */ @@ -1518,7 +1518,7 @@ makeRangeConstructors(const char *name, Oid namespace, INTERNALlanguageId, /* language */ F_FMGR_INTERNAL_VALIDATOR, /* language validator */ prosrc[i], /* prosrc */ - NULL, /* probin */ + NULL, /* probin */ false, /* isAgg */ false, /* isWindowFunc */ false, /* security_definer */ @@ -1834,9 +1834,9 @@ findRangeSubOpclass(List *opcname, Oid subtype) if (!IsBinaryCoercible(subtype, opInputType)) ereport(ERROR, (errcode(ERRCODE_DATATYPE_MISMATCH), - errmsg("operator class \"%s\" does not accept data type %s", - NameListToString(opcname), - format_type_be(subtype)))); + errmsg("operator class \"%s\" does not accept data type %s", + NameListToString(opcname), + format_type_be(subtype)))); } else { @@ -2335,8 +2335,8 @@ AlterDomainDropConstraint(List *names, const char *constrName, if (!missing_ok) ereport(ERROR, (errcode(ERRCODE_UNDEFINED_OBJECT), - errmsg("constraint \"%s\" of domain \"%s\" does not exist", - constrName, TypeNameToString(typename)))); + errmsg("constraint \"%s\" of domain \"%s\" does not exist", + constrName, TypeNameToString(typename)))); else ereport(NOTICE, (errmsg("constraint \"%s\" of domain \"%s\" does not exist, skipping", @@ -2958,7 +2958,7 @@ domainAddConstraint(Oid domainOid, Oid domainNamespace, Oid baseTypeOid, ccsrc, /* Source form of check constraint */ true, /* is local */ 0, /* inhcount */ - false); /* is only */ + false); /* is only */ /* * Return the compiled constraint expression so the calling routine can diff --git a/src/backend/commands/user.c b/src/backend/commands/user.c index 2edbabe7549ba15fad63d85b0f913638c2c82da1..a22092c202c4a11b653edac9417afa6e36bfe4f5 100644 --- a/src/backend/commands/user.c +++ b/src/backend/commands/user.c @@ -936,7 +936,8 @@ DropRole(DropRoleStmt *stmt) /* DROP hook for the role being removed */ if (object_access_hook) { - ObjectAccessDrop drop_arg; + ObjectAccessDrop drop_arg; + memset(&drop_arg, 0, sizeof(ObjectAccessDrop)); InvokeObjectAccessHook(OAT_DROP, AuthIdRelationId, roleid, 0, &drop_arg); diff --git a/src/backend/commands/vacuum.c b/src/backend/commands/vacuum.c index c43cd8e017502604d6a2933b683f67c53aaf2608..710c2afc9f3c3b14beb526b82db72edea1ddb890 100644 --- a/src/backend/commands/vacuum.c +++ b/src/backend/commands/vacuum.c @@ -322,13 +322,13 @@ get_rel_oids(Oid relid, const RangeVar *vacrel) Oid relid; /* - * Since we don't take a lock here, the relation might be gone, - * or the RangeVar might no longer refer to the OID we look up - * here. In the former case, VACUUM will do nothing; in the - * latter case, it will process the OID we looked up here, rather - * than the new one. Neither is ideal, but there's little practical - * alternative, since we're going to commit this transaction and - * begin a new one between now and then. + * Since we don't take a lock here, the relation might be gone, or the + * RangeVar might no longer refer to the OID we look up here. In the + * former case, VACUUM will do nothing; in the latter case, it will + * process the OID we looked up here, rather than the new one. + * Neither is ideal, but there's little practical alternative, since + * we're going to commit this transaction and begin a new one between + * now and then. */ relid = RangeVarGetRelid(vacrel, NoLock, false); diff --git a/src/backend/commands/vacuumlazy.c b/src/backend/commands/vacuumlazy.c index 3ff56a736648c5ab45533b376892b21e9ec569e3..5e902211649e2d0526664470054e015557098dab 100644 --- a/src/backend/commands/vacuumlazy.c +++ b/src/backend/commands/vacuumlazy.c @@ -155,9 +155,9 @@ lazy_vacuum_rel(Relation onerel, VacuumStmt *vacstmt, BlockNumber possibly_freeable; PGRUsage ru0; TimestampTz starttime = 0; - long secs; - int usecs; - double read_rate, + long secs; + int usecs; + double read_rate, write_rate; bool scan_all; TransactionId freezeTableLimit; @@ -222,17 +222,17 @@ lazy_vacuum_rel(Relation onerel, VacuumStmt *vacstmt, * * A corner case here is that if we scanned no pages at all because every * page is all-visible, we should not update relpages/reltuples, because - * we have no new information to contribute. In particular this keeps - * us from replacing relpages=reltuples=0 (which means "unknown tuple + * we have no new information to contribute. In particular this keeps us + * from replacing relpages=reltuples=0 (which means "unknown tuple * density") with nonzero relpages and reltuples=0 (which means "zero * tuple density") unless there's some actual evidence for the latter. * - * We do update relallvisible even in the corner case, since if the - * table is all-visible we'd definitely like to know that. But clamp - * the value to be not more than what we're setting relpages to. + * We do update relallvisible even in the corner case, since if the table + * is all-visible we'd definitely like to know that. But clamp the value + * to be not more than what we're setting relpages to. * - * Also, don't change relfrozenxid if we skipped any pages, since then - * we don't know for certain that all tuples have a newer xmin. + * Also, don't change relfrozenxid if we skipped any pages, since then we + * don't know for certain that all tuples have a newer xmin. */ new_rel_pages = vacrelstats->rel_pages; new_rel_tuples = vacrelstats->new_rel_tuples; @@ -265,7 +265,7 @@ lazy_vacuum_rel(Relation onerel, VacuumStmt *vacstmt, /* and log the action if appropriate */ if (IsAutoVacuumWorkerProcess() && Log_autovacuum_min_duration >= 0) { - TimestampTz endtime = GetCurrentTimestamp(); + TimestampTz endtime = GetCurrentTimestamp(); if (Log_autovacuum_min_duration == 0 || TimestampDifferenceExceeds(starttime, endtime, @@ -277,17 +277,17 @@ lazy_vacuum_rel(Relation onerel, VacuumStmt *vacstmt, write_rate = 0; if ((secs > 0) || (usecs > 0)) { - read_rate = (double) BLCKSZ * VacuumPageMiss / (1024 * 1024) / - (secs + usecs / 1000000.0); - write_rate = (double) BLCKSZ * VacuumPageDirty / (1024 * 1024) / - (secs + usecs / 1000000.0); + read_rate = (double) BLCKSZ *VacuumPageMiss / (1024 * 1024) / + (secs + usecs / 1000000.0); + write_rate = (double) BLCKSZ *VacuumPageDirty / (1024 * 1024) / + (secs + usecs / 1000000.0); } ereport(LOG, (errmsg("automatic vacuum of table \"%s.%s.%s\": index scans: %d\n" "pages: %d removed, %d remain\n" "tuples: %.0f removed, %.0f remain\n" "buffer usage: %d hits, %d misses, %d dirtied\n" - "avg read rate: %.3f MiB/s, avg write rate: %.3f MiB/s\n" + "avg read rate: %.3f MiB/s, avg write rate: %.3f MiB/s\n" "system usage: %s", get_database_name(MyDatabaseId), get_namespace_name(RelationGetNamespace(onerel)), @@ -300,7 +300,7 @@ lazy_vacuum_rel(Relation onerel, VacuumStmt *vacstmt, VacuumPageHit, VacuumPageMiss, VacuumPageDirty, - read_rate,write_rate, + read_rate, write_rate, pg_rusage_show(&ru0)))); } } @@ -501,10 +501,10 @@ lazy_scan_heap(Relation onerel, LVRelStats *vacrelstats, vacrelstats->num_dead_tuples > 0) { /* - * Before beginning index vacuuming, we release any pin we may hold - * on the visibility map page. This isn't necessary for correctness, - * but we do it anyway to avoid holding the pin across a lengthy, - * unrelated operation. + * Before beginning index vacuuming, we release any pin we may + * hold on the visibility map page. This isn't necessary for + * correctness, but we do it anyway to avoid holding the pin + * across a lengthy, unrelated operation. */ if (BufferIsValid(vmbuffer)) { @@ -535,10 +535,10 @@ lazy_scan_heap(Relation onerel, LVRelStats *vacrelstats, /* * Pin the visibility map page in case we need to mark the page * all-visible. In most cases this will be very cheap, because we'll - * already have the correct page pinned anyway. However, it's possible - * that (a) next_not_all_visible_block is covered by a different VM page - * than the current block or (b) we released our pin and did a cycle of - * index vacuuming. + * already have the correct page pinned anyway. However, it's + * possible that (a) next_not_all_visible_block is covered by a + * different VM page than the current block or (b) we released our pin + * and did a cycle of index vacuuming. */ visibilitymap_pin(onerel, blkno, &vmbuffer); @@ -873,10 +873,10 @@ lazy_scan_heap(Relation onerel, LVRelStats *vacrelstats, else if (!all_visible_according_to_vm) { /* - * It should never be the case that the visibility map page - * is set while the page-level bit is clear, but the reverse - * is allowed. Set the visibility map bit as well so that - * we get back in sync. + * It should never be the case that the visibility map page is + * set while the page-level bit is clear, but the reverse is + * allowed. Set the visibility map bit as well so that we get + * back in sync. */ visibilitymap_set(onerel, blkno, InvalidXLogRecPtr, vmbuffer, visibility_cutoff_xid); @@ -1152,7 +1152,7 @@ lazy_check_needs_freeze(Buffer buf) if (heap_tuple_needs_freeze(tupleheader, FreezeLimit, buf)) return true; - } /* scan along page */ + } /* scan along page */ return false; } diff --git a/src/backend/commands/view.c b/src/backend/commands/view.c index c887961bc9751c4ec3296573fbcec8fef37f5c16..3e7e39d8ecc9356d26d79a5a2f4b7ea81c9c1678 100644 --- a/src/backend/commands/view.c +++ b/src/backend/commands/view.c @@ -204,8 +204,8 @@ DefineVirtualRelation(RangeVar *relation, List *tlist, bool replace, checkViewTupleDesc(descriptor, rel->rd_att); /* - * The new options list replaces the existing options list, even - * if it's empty. + * The new options list replaces the existing options list, even if + * it's empty. */ atcmd = makeNode(AlterTableCmd); atcmd->subtype = AT_ReplaceRelOptions; @@ -504,7 +504,7 @@ DefineView(ViewStmt *stmt, const char *queryString) * long as the CREATE command is consistent with that --- no explicit * schema name. */ - view = copyObject(stmt->view); /* don't corrupt original command */ + view = copyObject(stmt->view); /* don't corrupt original command */ if (view->relpersistence == RELPERSISTENCE_PERMANENT && isViewOnTempTable(viewParse)) { diff --git a/src/backend/executor/execCurrent.c b/src/backend/executor/execCurrent.c index 03790bbe066fb26c8107ef64bcfbf05b13de5690..2c8929b588fdcd74cd0d3ed506bbd377165ef069 100644 --- a/src/backend/executor/execCurrent.c +++ b/src/backend/executor/execCurrent.c @@ -151,7 +151,7 @@ execCurrentOf(CurrentOfExpr *cexpr, { ScanState *scanstate; bool lisnull; - Oid tuple_tableoid PG_USED_FOR_ASSERTS_ONLY; + Oid tuple_tableoid PG_USED_FOR_ASSERTS_ONLY; ItemPointer tuple_tid; /* diff --git a/src/backend/executor/execMain.c b/src/backend/executor/execMain.c index fbb36fa6dc460654a715213d55dbcbdddefec893..440438b1807ef684ca56f094108f465bda1e55c5 100644 --- a/src/backend/executor/execMain.c +++ b/src/backend/executor/execMain.c @@ -80,7 +80,7 @@ static void ExecutePlan(EState *estate, PlanState *planstate, static bool ExecCheckRTEPerms(RangeTblEntry *rte); static void ExecCheckXactReadOnly(PlannedStmt *plannedstmt); static char *ExecBuildSlotValueDescription(TupleTableSlot *slot, - int maxfieldlen); + int maxfieldlen); static void EvalPlanQualStart(EPQState *epqstate, EState *parentestate, Plan *planTree); @@ -1520,7 +1520,7 @@ ExecConstraints(ResultRelInfo *resultRelInfo, ereport(ERROR, (errcode(ERRCODE_NOT_NULL_VIOLATION), errmsg("null value in column \"%s\" violates not-null constraint", - NameStr(rel->rd_att->attrs[attrChk - 1]->attname)), + NameStr(rel->rd_att->attrs[attrChk - 1]->attname)), errdetail("Failing row contains %s.", ExecBuildSlotValueDescription(slot, 64)))); } diff --git a/src/backend/executor/execQual.c b/src/backend/executor/execQual.c index a1193a8dc34d8f5eab2d643898b7a905066d84de..0ea21ca5f91a704ccde2c765bd92d0c5af7ead8b 100644 --- a/src/backend/executor/execQual.c +++ b/src/backend/executor/execQual.c @@ -578,15 +578,15 @@ ExecEvalVar(ExprState *exprstate, ExprContext *econtext, /* Get the input slot and attribute number we want */ switch (variable->varno) { - case INNER_VAR: /* get the tuple from the inner node */ + case INNER_VAR: /* get the tuple from the inner node */ slot = econtext->ecxt_innertuple; break; - case OUTER_VAR: /* get the tuple from the outer node */ + case OUTER_VAR: /* get the tuple from the outer node */ slot = econtext->ecxt_outertuple; break; - /* INDEX_VAR is handled by default case */ + /* INDEX_VAR is handled by default case */ default: /* get the tuple from the relation being * scanned */ @@ -763,15 +763,15 @@ ExecEvalScalarVar(ExprState *exprstate, ExprContext *econtext, /* Get the input slot and attribute number we want */ switch (variable->varno) { - case INNER_VAR: /* get the tuple from the inner node */ + case INNER_VAR: /* get the tuple from the inner node */ slot = econtext->ecxt_innertuple; break; - case OUTER_VAR: /* get the tuple from the outer node */ + case OUTER_VAR: /* get the tuple from the outer node */ slot = econtext->ecxt_outertuple; break; - /* INDEX_VAR is handled by default case */ + /* INDEX_VAR is handled by default case */ default: /* get the tuple from the relation being * scanned */ @@ -808,15 +808,15 @@ ExecEvalWholeRowVar(ExprState *exprstate, ExprContext *econtext, /* Get the input slot we want */ switch (variable->varno) { - case INNER_VAR: /* get the tuple from the inner node */ + case INNER_VAR: /* get the tuple from the inner node */ slot = econtext->ecxt_innertuple; break; - case OUTER_VAR: /* get the tuple from the outer node */ + case OUTER_VAR: /* get the tuple from the outer node */ slot = econtext->ecxt_outertuple; break; - /* INDEX_VAR is handled by default case */ + /* INDEX_VAR is handled by default case */ default: /* get the tuple from the relation being * scanned */ @@ -879,15 +879,15 @@ ExecEvalWholeRowSlow(ExprState *exprstate, ExprContext *econtext, /* Get the input slot we want */ switch (variable->varno) { - case INNER_VAR: /* get the tuple from the inner node */ + case INNER_VAR: /* get the tuple from the inner node */ slot = econtext->ecxt_innertuple; break; - case OUTER_VAR: /* get the tuple from the outer node */ + case OUTER_VAR: /* get the tuple from the outer node */ slot = econtext->ecxt_outertuple; break; - /* INDEX_VAR is handled by default case */ + /* INDEX_VAR is handled by default case */ default: /* get the tuple from the relation being * scanned */ diff --git a/src/backend/executor/execUtils.c b/src/backend/executor/execUtils.c index 40cd5ce5d19d627141549769233b2c6215fda7bb..2bd8b4283598ab8fe474a8ee3745205921638a52 100644 --- a/src/backend/executor/execUtils.c +++ b/src/backend/executor/execUtils.c @@ -578,7 +578,7 @@ ExecBuildProjectionInfo(List *targetList, projInfo->pi_lastOuterVar = attnum; break; - /* INDEX_VAR is handled by default case */ + /* INDEX_VAR is handled by default case */ default: varSlotOffsets[numSimpleVars] = offsetof(ExprContext, @@ -638,7 +638,7 @@ get_last_attnums(Node *node, ProjectionInfo *projInfo) projInfo->pi_lastOuterVar = attnum; break; - /* INDEX_VAR is handled by default case */ + /* INDEX_VAR is handled by default case */ default: if (projInfo->pi_lastScanVar < attnum) diff --git a/src/backend/executor/functions.c b/src/backend/executor/functions.c index ae8d374db216328f6db9eebe3a0f92ea18f93864..bf2f5c68829da0e22c04cc896b123a358dee82fd 100644 --- a/src/backend/executor/functions.c +++ b/src/backend/executor/functions.c @@ -128,11 +128,11 @@ typedef struct SQLFunctionParseInfo /* non-export function prototypes */ static Node *sql_fn_param_ref(ParseState *pstate, ParamRef *pref); static Node *sql_fn_post_column_ref(ParseState *pstate, - ColumnRef *cref, Node *var); + ColumnRef *cref, Node *var); static Node *sql_fn_make_param(SQLFunctionParseInfoPtr pinfo, - int paramno, int location); + int paramno, int location); static Node *sql_fn_resolve_param_name(SQLFunctionParseInfoPtr pinfo, - const char *paramname, int location); + const char *paramname, int location); static List *init_execution_state(List *queryTree_list, SQLFunctionCachePtr fcache, bool lazyEvalOK); @@ -227,13 +227,13 @@ prepare_sql_fn_parse_info(HeapTuple procedureTuple, Anum_pg_proc_proargnames, &isNull); if (isNull) - proargnames = PointerGetDatum(NULL); /* just to be sure */ + proargnames = PointerGetDatum(NULL); /* just to be sure */ proargmodes = SysCacheGetAttr(PROCNAMEARGSNSP, procedureTuple, Anum_pg_proc_proargmodes, &isNull); if (isNull) - proargmodes = PointerGetDatum(NULL); /* just to be sure */ + proargmodes = PointerGetDatum(NULL); /* just to be sure */ n_arg_names = get_func_input_arg_names(proargnames, proargmodes, &pinfo->argnames); @@ -422,7 +422,7 @@ static Node * sql_fn_resolve_param_name(SQLFunctionParseInfoPtr pinfo, const char *paramname, int location) { - int i; + int i; if (pinfo->argnames == NULL) return NULL; diff --git a/src/backend/executor/nodeBitmapHeapscan.c b/src/backend/executor/nodeBitmapHeapscan.c index 849665d4e2e6dda87986c6774086afc0705e5502..702e704098d461112230608383a7235f966836a4 100644 --- a/src/backend/executor/nodeBitmapHeapscan.c +++ b/src/backend/executor/nodeBitmapHeapscan.c @@ -66,6 +66,7 @@ BitmapHeapNext(BitmapHeapScanState *node) TIDBitmap *tbm; TBMIterator *tbmiterator; TBMIterateResult *tbmres; + #ifdef USE_PREFETCH TBMIterator *prefetch_iterator; #endif @@ -355,7 +356,7 @@ bitgetpage(HeapScanDesc scan, TBMIterateResult *tbmres) { OffsetNumber offnum = tbmres->offsets[curslot]; ItemPointerData tid; - HeapTupleData heapTuple; + HeapTupleData heapTuple; ItemPointerSet(&tid, page, offnum); if (heap_hot_search_buffer(&tid, scan->rs_rd, buffer, snapshot, diff --git a/src/backend/executor/nodeIndexonlyscan.c b/src/backend/executor/nodeIndexonlyscan.c index af31671b3eb60979fdfbcacd4c362cecc231c32c..38078763f57e8bdbcb1141bb819f8389a4a5963d 100644 --- a/src/backend/executor/nodeIndexonlyscan.c +++ b/src/backend/executor/nodeIndexonlyscan.c @@ -86,7 +86,7 @@ IndexOnlyNext(IndexOnlyScanState *node) * Note on Memory Ordering Effects: visibilitymap_test does not lock * the visibility map buffer, and therefore the result we read here * could be slightly stale. However, it can't be stale enough to - * matter. It suffices to show that (1) there is a read barrier + * matter. It suffices to show that (1) there is a read barrier * between the time we read the index TID and the time we test the * visibility map; and (2) there is a write barrier between the time * some other concurrent process clears the visibility map bit and the @@ -106,12 +106,12 @@ IndexOnlyNext(IndexOnlyScanState *node) node->ioss_HeapFetches++; tuple = index_fetch_heap(scandesc); if (tuple == NULL) - continue; /* no visible tuple, try next index entry */ + continue; /* no visible tuple, try next index entry */ /* * Only MVCC snapshots are supported here, so there should be no * need to keep following the HOT chain once a visible entry has - * been found. If we did want to allow that, we'd need to keep + * been found. If we did want to allow that, we'd need to keep * more state to remember not to call index_getnext_tid next time. */ if (scandesc->xs_continue_hot) @@ -120,7 +120,7 @@ IndexOnlyNext(IndexOnlyScanState *node) /* * Note: at this point we are holding a pin on the heap page, as * recorded in scandesc->xs_cbuf. We could release that pin now, - * but it's not clear whether it's a win to do so. The next index + * but it's not clear whether it's a win to do so. The next index * entry might require a visit to the same heap page. */ } @@ -176,8 +176,8 @@ StoreIndexTuple(TupleTableSlot *slot, IndexTuple itup, TupleDesc itupdesc) * Note: we must use the tupdesc supplied by the AM in index_getattr, not * the slot's tupdesc, in case the latter has different datatypes (this * happens for btree name_ops in particular). They'd better have the same - * number of columns though, as well as being datatype-compatible which - * is something we can't so easily check. + * number of columns though, as well as being datatype-compatible which is + * something we can't so easily check. */ Assert(slot->tts_tupleDescriptor->natts == nindexatts); @@ -494,10 +494,10 @@ ExecInitIndexOnlyScan(IndexOnlyScan *node, EState *estate, int eflags) * Initialize scan descriptor. */ indexstate->ioss_ScanDesc = index_beginscan(currentRelation, - indexstate->ioss_RelationDesc, - estate->es_snapshot, - indexstate->ioss_NumScanKeys, - indexstate->ioss_NumOrderByKeys); + indexstate->ioss_RelationDesc, + estate->es_snapshot, + indexstate->ioss_NumScanKeys, + indexstate->ioss_NumOrderByKeys); /* Set it up for index-only scan */ indexstate->ioss_ScanDesc->xs_want_itup = true; diff --git a/src/backend/executor/nodeMaterial.c b/src/backend/executor/nodeMaterial.c index 06137c6ba80bd0e8235bc352f8df22d60c8be60f..3a6bfec0dbb2ae894cf3456786cead4ec8b8508c 100644 --- a/src/backend/executor/nodeMaterial.c +++ b/src/backend/executor/nodeMaterial.c @@ -66,7 +66,7 @@ ExecMaterial(MaterialState *node) * Allocate a second read pointer to serve as the mark. We know it * must have index 1, so needn't store that. */ - int ptrno PG_USED_FOR_ASSERTS_ONLY; + int ptrno PG_USED_FOR_ASSERTS_ONLY; ptrno = tuplestore_alloc_read_pointer(tuplestorestate, node->eflags); diff --git a/src/backend/executor/nodeMergeAppend.c b/src/backend/executor/nodeMergeAppend.c index d755109a3377658cebd64a182f62e472763bb56e..d5141ba54e2433b82c0881e1e74f5900aaad085e 100644 --- a/src/backend/executor/nodeMergeAppend.c +++ b/src/backend/executor/nodeMergeAppend.c @@ -130,7 +130,7 @@ ExecInitMergeAppend(MergeAppend *node, EState *estate, int eflags) for (i = 0; i < node->numCols; i++) { - SortSupport sortKey = mergestate->ms_sortkeys + i; + SortSupport sortKey = mergestate->ms_sortkeys + i; sortKey->ssup_cxt = CurrentMemoryContext; sortKey->ssup_collation = node->collations[i]; @@ -276,7 +276,7 @@ heap_compare_slots(MergeAppendState *node, SlotNumber slot1, SlotNumber slot2) for (nkey = 0; nkey < node->ms_nkeys; nkey++) { - SortSupport sortKey = node->ms_sortkeys + nkey; + SortSupport sortKey = node->ms_sortkeys + nkey; AttrNumber attno = sortKey->ssup_attno; Datum datum1, datum2; diff --git a/src/backend/executor/nodeMergejoin.c b/src/backend/executor/nodeMergejoin.c index a1e55646c6cb0768b56f4c4f29753faf9ea9e27b..bc0b20bf825f016ba002b843e2ff1a5fe96b5cf2 100644 --- a/src/backend/executor/nodeMergejoin.c +++ b/src/backend/executor/nodeMergejoin.c @@ -247,7 +247,7 @@ MJExamineQuals(List *mergeclauses, op_lefttype, op_righttype, BTORDER_PROC); - if (!OidIsValid(sortfunc)) /* should not happen */ + if (!OidIsValid(sortfunc)) /* should not happen */ elog(ERROR, "missing support function %d(%u,%u) in opfamily %u", BTORDER_PROC, op_lefttype, op_righttype, opfamily); /* We'll use a shim to call the old-style btree comparator */ @@ -405,7 +405,7 @@ MJCompare(MergeJoinState *mergestate) */ if (clause->lisnull && clause->risnull) { - nulleqnull = true; /* NULL "=" NULL */ + nulleqnull = true; /* NULL "=" NULL */ continue; } @@ -419,8 +419,8 @@ MJCompare(MergeJoinState *mergestate) /* * If we had any NULL-vs-NULL inputs, we do not want to report that the - * tuples are equal. Instead, if result is still 0, change it to +1. - * This will result in advancing the inner side of the join. + * tuples are equal. Instead, if result is still 0, change it to +1. This + * will result in advancing the inner side of the join. * * Likewise, if there was a constant-false joinqual, do not report * equality. We have to check this as part of the mergequals, else the diff --git a/src/backend/executor/nodeModifyTable.c b/src/backend/executor/nodeModifyTable.c index dfdcb20b1d1e06cc154f74275ae93c3539aac2a2..a7bce75f0cb63c78d59306261bae5dd093a120fc 100644 --- a/src/backend/executor/nodeModifyTable.c +++ b/src/backend/executor/nodeModifyTable.c @@ -950,8 +950,8 @@ ExecInitModifyTable(ModifyTable *node, EState *estate, int eflags) * If there are indices on the result relation, open them and save * descriptors in the result relation info, so that we can add new * index entries for the tuples we add/update. We need not do this - * for a DELETE, however, since deletion doesn't affect indexes. - * Also, inside an EvalPlanQual operation, the indexes might be open + * for a DELETE, however, since deletion doesn't affect indexes. Also, + * inside an EvalPlanQual operation, the indexes might be open * already, since we share the resultrel state with the original * query. */ diff --git a/src/backend/executor/nodeSetOp.c b/src/backend/executor/nodeSetOp.c index 85590445ccc9faf5231562a152fed5d35843da39..362f4466e4494b2ec3365394b72c06d8ecfb7d92 100644 --- a/src/backend/executor/nodeSetOp.c +++ b/src/backend/executor/nodeSetOp.c @@ -344,7 +344,7 @@ setop_fill_hash_table(SetOpState *setopstate) SetOp *node = (SetOp *) setopstate->ps.plan; PlanState *outerPlan; int firstFlag; - bool in_first_rel PG_USED_FOR_ASSERTS_ONLY; + bool in_first_rel PG_USED_FOR_ASSERTS_ONLY; /* * get state info from node diff --git a/src/backend/executor/spi.c b/src/backend/executor/spi.c index 5e4ae426b1ba03ef97317dd83271bc8e757180ff..e222365d111c75dd0100c520bd6edb5e2d928643 100644 --- a/src/backend/executor/spi.c +++ b/src/backend/executor/spi.c @@ -1674,8 +1674,8 @@ _SPI_prepare_plan(const char *src, SPIPlanPtr plan, ParamListInfo boundParams) raw_parsetree_list = pg_parse_query(src); /* - * Do parse analysis and rule rewrite for each raw parsetree, storing - * the results into unsaved plancache entries. + * Do parse analysis and rule rewrite for each raw parsetree, storing the + * results into unsaved plancache entries. */ plancache_list = NIL; @@ -1686,8 +1686,8 @@ _SPI_prepare_plan(const char *src, SPIPlanPtr plan, ParamListInfo boundParams) CachedPlanSource *plansource; /* - * Create the CachedPlanSource before we do parse analysis, since - * it needs to see the unmodified raw parse tree. + * Create the CachedPlanSource before we do parse analysis, since it + * needs to see the unmodified raw parse tree. */ plansource = CreateCachedPlan(parsetree, src, @@ -1722,7 +1722,7 @@ _SPI_prepare_plan(const char *src, SPIPlanPtr plan, ParamListInfo boundParams) plan->parserSetup, plan->parserSetupArg, cursor_options, - false); /* not fixed result */ + false); /* not fixed result */ plancache_list = lappend(plancache_list, plansource); } @@ -1907,7 +1907,7 @@ _SPI_execute_plan(SPIPlanPtr plan, ParamListInfo paramLI, } else { - char completionTag[COMPLETION_TAG_BUFSIZE]; + char completionTag[COMPLETION_TAG_BUFSIZE]; ProcessUtility(stmt, plansource->query_string, @@ -2335,9 +2335,9 @@ _SPI_make_plan_non_temp(SPIPlanPtr plan) /* * Reparent all the CachedPlanSources into the procedure context. In - * theory this could fail partway through due to the pallocs, but we - * don't care too much since both the procedure context and the executor - * context would go away on error. + * theory this could fail partway through due to the pallocs, but we don't + * care too much since both the procedure context and the executor context + * would go away on error. */ foreach(lc, plan->plancache_list) { diff --git a/src/backend/libpq/auth.c b/src/backend/libpq/auth.c index 5853b068daa6715de84291955ba84ae657db2dec..9cdee2bb3e7bd2e42a6de3e37b006786066fe543 100644 --- a/src/backend/libpq/auth.c +++ b/src/backend/libpq/auth.c @@ -316,8 +316,8 @@ ClientAuthentication(Port *port) /* * Get the authentication method to use for this frontend/database * combination. Note: we do not parse the file at this point; this has - * already been done elsewhere. hba.c dropped an error message - * into the server logfile if parsing the hba config file failed. + * already been done elsewhere. hba.c dropped an error message into the + * server logfile if parsing the hba config file failed. */ hba_getauthmethod(port); @@ -1365,10 +1365,10 @@ pg_SSPI_recvauth(Port *port) } /* - * Overwrite the current context with the one we just received. - * If sspictx is NULL it was the first loop and we need to allocate - * a buffer for it. On subsequent runs, we can just overwrite the - * buffer contents since the size does not change. + * Overwrite the current context with the one we just received. If + * sspictx is NULL it was the first loop and we need to allocate a + * buffer for it. On subsequent runs, we can just overwrite the buffer + * contents since the size does not change. */ if (sspictx == NULL) { @@ -1437,8 +1437,8 @@ pg_SSPI_recvauth(Port *port) if (!GetTokenInformation(token, TokenUser, NULL, 0, &retlen) && GetLastError() != 122) ereport(ERROR, - (errmsg_internal("could not get token user size: error code %lu", - GetLastError()))); + (errmsg_internal("could not get token user size: error code %lu", + GetLastError()))); tokenuser = malloc(retlen); if (tokenuser == NULL) @@ -1453,8 +1453,8 @@ pg_SSPI_recvauth(Port *port) if (!LookupAccountSid(NULL, tokenuser->User.Sid, accountname, &accountnamesize, domainname, &domainnamesize, &accountnameuse)) ereport(ERROR, - (errmsg_internal("could not look up account SID: error code %lu", - GetLastError()))); + (errmsg_internal("could not look up account SID: error code %lu", + GetLastError()))); free(tokenuser); diff --git a/src/backend/libpq/be-secure.c b/src/backend/libpq/be-secure.c index dce0eaa20e208314c14fda855994f071c5831c34..e0ab5997fbd9337d07a38e6fe79b7f144c6d6d04 100644 --- a/src/backend/libpq/be-secure.c +++ b/src/backend/libpq/be-secure.c @@ -89,10 +89,10 @@ static void close_SSL(Port *); static const char *SSLerrmessage(void); #endif -char *ssl_cert_file; -char *ssl_key_file; -char *ssl_ca_file; -char *ssl_crl_file; +char *ssl_cert_file; +char *ssl_key_file; +char *ssl_ca_file; +char *ssl_crl_file; /* * How much data can be sent across a secure connection @@ -845,8 +845,8 @@ initialize_SSL(void) { /* * Always ask for SSL client cert, but don't fail if it's not - * presented. We might fail such connections later, depending on - * what we find in pg_hba.conf. + * presented. We might fail such connections later, depending on what + * we find in pg_hba.conf. */ SSL_CTX_set_verify(SSL_context, (SSL_VERIFY_PEER | @@ -953,7 +953,7 @@ aloop: port->peer_cn = NULL; if (port->peer != NULL) { - int len; + int len; len = X509_NAME_get_text_by_NID(X509_get_subject_name(port->peer), NID_commonName, NULL, 0); diff --git a/src/backend/libpq/hba.c b/src/backend/libpq/hba.c index 56229cb4df8b29c19a174049a75dbe24c9cfc60f..828f6dcc8e1653ad25f206f9d10b55006466b2e1 100644 --- a/src/backend/libpq/hba.c +++ b/src/backend/libpq/hba.c @@ -61,8 +61,8 @@ typedef struct check_network_data */ typedef struct HbaToken { - char *string; - bool quoted; + char *string; + bool quoted; } HbaToken; /* @@ -76,9 +76,9 @@ static MemoryContext parsed_hba_context = NULL; * These variables hold the pre-parsed contents of the ident usermap * configuration file. ident_lines is a triple-nested list of lines, fields * and tokens, as returned by tokenize_file. There will be one line in - * ident_lines for each (non-empty, non-comment) line of the file. Note there + * ident_lines for each (non-empty, non-comment) line of the file. Note there * will always be at least one field, since blank lines are not entered in the - * data structure. ident_line_nums is an integer list containing the actual + * data structure. ident_line_nums is an integer list containing the actual * line number for each line represented in ident_lines. ident_context is * the memory context holding all this. */ @@ -246,7 +246,7 @@ make_hba_token(char *token, bool quoted) static HbaToken * copy_hba_token(HbaToken *in) { - HbaToken *out = make_hba_token(in->string, in->quoted); + HbaToken *out = make_hba_token(in->string, in->quoted); return out; } @@ -283,12 +283,12 @@ next_field_expand(const char *filename, FILE *file) /* * tokenize_inc_file - * Expand a file included from another file into an hba "field" + * Expand a file included from another file into an hba "field" * * Opens and tokenises a file included from another HBA config file with @, * and returns all values found therein as a flat list of HbaTokens. If a * @-token is found, recursively expand it. The given token list is used as - * initial contents of list (so foo,bar,@baz does what you expect). + * initial contents of list (so foo,bar,@baz does what you expect). */ static List * tokenize_inc_file(List *tokens, @@ -377,8 +377,8 @@ tokenize_file(const char *filename, FILE *file, List *current_line = NIL; List *current_field = NIL; int line_number = 1; - MemoryContext linecxt; - MemoryContext oldcxt; + MemoryContext linecxt; + MemoryContext oldcxt; linecxt = AllocSetContextCreate(TopMemoryContext, "tokenize file cxt", @@ -442,11 +442,10 @@ is_member(Oid userid, const char *role) if (!OidIsValid(roleid)) return false; /* if target role not exist, say "no" */ - /* - * See if user is directly or indirectly a member of role. - * For this purpose, a superuser is not considered to be automatically - * a member of the role, so group auth only applies to explicit - * membership. + /* + * See if user is directly or indirectly a member of role. For this + * purpose, a superuser is not considered to be automatically a member of + * the role, so group auth only applies to explicit membership. */ return is_member_of_role_nosuper(userid, roleid); } @@ -457,8 +456,8 @@ is_member(Oid userid, const char *role) static bool check_role(const char *role, Oid roleid, List *tokens) { - ListCell *cell; - HbaToken *tok; + ListCell *cell; + HbaToken *tok; foreach(cell, tokens) { @@ -481,8 +480,8 @@ check_role(const char *role, Oid roleid, List *tokens) static bool check_db(const char *dbname, const char *role, Oid roleid, List *tokens) { - ListCell *cell; - HbaToken *tok; + ListCell *cell; + HbaToken *tok; foreach(cell, tokens) { @@ -825,7 +824,7 @@ parse_hba_line(List *line, int line_num) List *tokens; ListCell *tokencell; HbaToken *token; - HbaLine *parsedline; + HbaLine *parsedline; parsedline = palloc0(sizeof(HbaLine)); parsedline->linenumber = line_num; @@ -1042,8 +1041,8 @@ parse_hba_line(List *line, int line_num) (errcode(ERRCODE_CONFIG_FILE_ERROR), errmsg("specifying both host name and CIDR mask is invalid: \"%s\"", token->string), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return NULL; } @@ -1080,9 +1079,9 @@ parse_hba_line(List *line, int line_num) { ereport(LOG, (errcode(ERRCODE_CONFIG_FILE_ERROR), - errmsg("multiple values specified for netmask"), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errmsg("multiple values specified for netmask"), + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return NULL; } token = linitial(tokens); @@ -1293,6 +1292,7 @@ parse_hba_line(List *line, int line_num) foreach(tokencell, tokens) { char *val; + token = lfirst(tokencell); str = pstrdup(token->string); @@ -1310,7 +1310,7 @@ parse_hba_line(List *line, int line_num) return NULL; } - *val++ = '\0'; /* str now holds "name", val holds "value" */ + *val++ = '\0'; /* str now holds "name", val holds "value" */ if (!parse_hba_auth_opt(str, val, parsedline, line_num)) /* parse_hba_auth_opt already logged the error message */ return NULL; @@ -1397,17 +1397,16 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) else if (strcmp(name, "clientcert") == 0) { /* - * Since we require ctHostSSL, this really can never happen - * on non-SSL-enabled builds, so don't bother checking for - * USE_SSL. + * Since we require ctHostSSL, this really can never happen on + * non-SSL-enabled builds, so don't bother checking for USE_SSL. */ if (hbaline->conntype != ctHostSSL) { ereport(LOG, (errcode(ERRCODE_CONFIG_FILE_ERROR), - errmsg("clientcert can only be configured for \"hostssl\" rows"), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errmsg("clientcert can only be configured for \"hostssl\" rows"), + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return false; } if (strcmp(val, "1") == 0) @@ -1418,8 +1417,8 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) (errcode(ERRCODE_CONFIG_FILE_ERROR), errmsg("client certificates can only be checked if a root certificate store is available"), errhint("Make sure the configuration parameter \"ssl_ca_file\" is set."), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return false; } hbaline->clientcert = true; @@ -1431,8 +1430,8 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) ereport(LOG, (errcode(ERRCODE_CONFIG_FILE_ERROR), errmsg("clientcert can not be set to 0 when using \"cert\" authentication"), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return false; } hbaline->clientcert = false; @@ -1465,8 +1464,8 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) ereport(LOG, (errcode(ERRCODE_CONFIG_FILE_ERROR), errmsg("invalid LDAP port number: \"%s\"", val), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return false; } } @@ -1528,7 +1527,7 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) { struct addrinfo *gai_result; struct addrinfo hints; - int ret; + int ret; REQUIRE_AUTH_OPTION(uaRADIUS, "radiusserver", "radius"); @@ -1543,8 +1542,8 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) (errcode(ERRCODE_CONFIG_FILE_ERROR), errmsg("could not translate RADIUS server name \"%s\" to address: %s", val, gai_strerror(ret)), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); if (gai_result) pg_freeaddrinfo_all(hints.ai_family, gai_result); return false; @@ -1561,8 +1560,8 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) ereport(LOG, (errcode(ERRCODE_CONFIG_FILE_ERROR), errmsg("invalid RADIUS port number: \"%s\"", val), - errcontext("line %d of configuration file \"%s\"", - line_num, HbaFileName))); + errcontext("line %d of configuration file \"%s\"", + line_num, HbaFileName))); return false; } } @@ -1580,8 +1579,8 @@ parse_hba_auth_opt(char *name, char *val, HbaLine *hbaline, int line_num) { ereport(LOG, (errcode(ERRCODE_CONFIG_FILE_ERROR), - errmsg("unrecognized authentication option name: \"%s\"", - name), + errmsg("unrecognized authentication option name: \"%s\"", + name), errcontext("line %d of configuration file \"%s\"", line_num, HbaFileName))); return false; @@ -1693,7 +1692,7 @@ check_hba(hbaPort *port) * Read the config file and create a List of HbaLine records for the contents. * * The configuration is read into a temporary list, and if any parse error - * occurs the old list is kept in place and false is returned. Only if the + * occurs the old list is kept in place and false is returned. Only if the * whole file parses OK is the list replaced, and the function returns true. * * On a false result, caller will take care of reporting a FATAL error in case @@ -1710,9 +1709,9 @@ load_hba(void) *line_num; List *new_parsed_lines = NIL; bool ok = true; - MemoryContext linecxt; - MemoryContext oldcxt; - MemoryContext hbacxt; + MemoryContext linecxt; + MemoryContext oldcxt; + MemoryContext hbacxt; file = AllocateFile(HbaFileName, "r"); if (file == NULL) @@ -1742,8 +1741,8 @@ load_hba(void) { /* * Parse error in the file, so indicate there's a problem. NB: a - * problem in a line will free the memory for all previous lines as - * well! + * problem in a line will free the memory for all previous lines + * as well! */ MemoryContextReset(hbacxt); new_parsed_lines = NIL; @@ -1761,9 +1760,9 @@ load_hba(void) } /* - * A valid HBA file must have at least one entry; else there's no way - * to connect to the postmaster. But only complain about this if we - * didn't already have parsing errors. + * A valid HBA file must have at least one entry; else there's no way to + * connect to the postmaster. But only complain about this if we didn't + * already have parsing errors. */ if (ok && new_parsed_lines == NIL) { diff --git a/src/backend/libpq/pqcomm.c b/src/backend/libpq/pqcomm.c index 2082e3d4f6f97314c35cdca4b8f62c08083c13e4..5272811cc0af190803666cd32dc7ac7167c12c77 100644 --- a/src/backend/libpq/pqcomm.c +++ b/src/backend/libpq/pqcomm.c @@ -1247,9 +1247,9 @@ internal_flush(void) /* * We drop the buffered data anyway so that processing can - * continue, even though we'll probably quit soon. We also - * set a flag that'll cause the next CHECK_FOR_INTERRUPTS - * to terminate the connection. + * continue, even though we'll probably quit soon. We also set a + * flag that'll cause the next CHECK_FOR_INTERRUPTS to terminate + * the connection. */ PqSendStart = PqSendPointer = 0; ClientConnectionLost = 1; @@ -1373,7 +1373,7 @@ fail: void pq_putmessage_noblock(char msgtype, const char *s, size_t len) { - int res PG_USED_FOR_ASSERTS_ONLY; + int res PG_USED_FOR_ASSERTS_ONLY; int required; /* diff --git a/src/backend/nodes/bitmapset.c b/src/backend/nodes/bitmapset.c index 4c904e03296051cf2c95fdcd3d4a029aea9dc142..ba108401663f213c929491ca185df7813f52dea8 100644 --- a/src/backend/nodes/bitmapset.c +++ b/src/backend/nodes/bitmapset.c @@ -362,8 +362,8 @@ bms_subset_compare(const Bitmapset *a, const Bitmapset *b) shortlen = Min(a->nwords, b->nwords); for (i = 0; i < shortlen; i++) { - bitmapword aword = a->words[i]; - bitmapword bword = b->words[i]; + bitmapword aword = a->words[i]; + bitmapword bword = b->words[i]; if ((aword & ~bword) != 0) { diff --git a/src/backend/nodes/copyfuncs.c b/src/backend/nodes/copyfuncs.c index 0db60b161bc0b821381553fc68b109467db45887..1743b8fdc89735f0f619c290e7286be23f3e72e6 100644 --- a/src/backend/nodes/copyfuncs.c +++ b/src/backend/nodes/copyfuncs.c @@ -381,7 +381,7 @@ _copyIndexScan(const IndexScan *from) static IndexOnlyScan * _copyIndexOnlyScan(const IndexOnlyScan *from) { - IndexOnlyScan *newnode = makeNode(IndexOnlyScan); + IndexOnlyScan *newnode = makeNode(IndexOnlyScan); /* * copy node superclass fields @@ -4473,7 +4473,7 @@ copyObject(const void *from) default: elog(ERROR, "unrecognized node type: %d", (int) nodeTag(from)); - retval = 0; /* keep compiler quiet */ + retval = 0; /* keep compiler quiet */ break; } diff --git a/src/backend/nodes/equalfuncs.c b/src/backend/nodes/equalfuncs.c index 9d588feac20c76d88a0b5a2ebaff5a6ec4b00798..f19ad7702618dcfb7f3be416dc535e6ac11b1b5e 100644 --- a/src/backend/nodes/equalfuncs.c +++ b/src/backend/nodes/equalfuncs.c @@ -2360,8 +2360,8 @@ _equalXmlSerialize(const XmlSerialize *a, const XmlSerialize *b) static bool _equalList(const List *a, const List *b) { - const ListCell *item_a; - const ListCell *item_b; + const ListCell *item_a; + const ListCell *item_b; /* * Try to reject by simple scalar checks before grovelling through all the diff --git a/src/backend/nodes/list.c b/src/backend/nodes/list.c index 209b72222efa0afaab414b75e4860cdfcb7844a2..4d19aed8f4b8b7dbffdd9dd47fa757d684d4b38f 100644 --- a/src/backend/nodes/list.c +++ b/src/backend/nodes/list.c @@ -443,7 +443,7 @@ list_nth_oid(const List *list, int n) bool list_member(const List *list, const void *datum) { - const ListCell *cell; + const ListCell *cell; Assert(IsPointerList(list)); check_list_invariants(list); @@ -464,7 +464,7 @@ list_member(const List *list, const void *datum) bool list_member_ptr(const List *list, const void *datum) { - const ListCell *cell; + const ListCell *cell; Assert(IsPointerList(list)); check_list_invariants(list); @@ -484,7 +484,7 @@ list_member_ptr(const List *list, const void *datum) bool list_member_int(const List *list, int datum) { - const ListCell *cell; + const ListCell *cell; Assert(IsIntegerList(list)); check_list_invariants(list); @@ -504,7 +504,7 @@ list_member_int(const List *list, int datum) bool list_member_oid(const List *list, Oid datum) { - const ListCell *cell; + const ListCell *cell; Assert(IsOidList(list)); check_list_invariants(list); @@ -697,7 +697,7 @@ List * list_union(const List *list1, const List *list2) { List *result; - const ListCell *cell; + const ListCell *cell; Assert(IsPointerList(list1)); Assert(IsPointerList(list2)); @@ -721,7 +721,7 @@ List * list_union_ptr(const List *list1, const List *list2) { List *result; - const ListCell *cell; + const ListCell *cell; Assert(IsPointerList(list1)); Assert(IsPointerList(list2)); @@ -744,7 +744,7 @@ List * list_union_int(const List *list1, const List *list2) { List *result; - const ListCell *cell; + const ListCell *cell; Assert(IsIntegerList(list1)); Assert(IsIntegerList(list2)); @@ -767,7 +767,7 @@ List * list_union_oid(const List *list1, const List *list2) { List *result; - const ListCell *cell; + const ListCell *cell; Assert(IsOidList(list1)); Assert(IsOidList(list2)); @@ -800,7 +800,7 @@ List * list_intersection(const List *list1, const List *list2) { List *result; - const ListCell *cell; + const ListCell *cell; if (list1 == NIL || list2 == NIL) return NIL; @@ -831,7 +831,7 @@ list_intersection(const List *list1, const List *list2) List * list_difference(const List *list1, const List *list2) { - const ListCell *cell; + const ListCell *cell; List *result = NIL; Assert(IsPointerList(list1)); @@ -857,7 +857,7 @@ list_difference(const List *list1, const List *list2) List * list_difference_ptr(const List *list1, const List *list2) { - const ListCell *cell; + const ListCell *cell; List *result = NIL; Assert(IsPointerList(list1)); @@ -882,7 +882,7 @@ list_difference_ptr(const List *list1, const List *list2) List * list_difference_int(const List *list1, const List *list2) { - const ListCell *cell; + const ListCell *cell; List *result = NIL; Assert(IsIntegerList(list1)); @@ -907,7 +907,7 @@ list_difference_int(const List *list1, const List *list2) List * list_difference_oid(const List *list1, const List *list2) { - const ListCell *cell; + const ListCell *cell; List *result = NIL; Assert(IsOidList(list1)); diff --git a/src/backend/nodes/nodeFuncs.c b/src/backend/nodes/nodeFuncs.c index 6f9e053669453c6c4dab283807406546c4b5826d..813d1da1a2816b3a9a6df159ce5f3f616683be58 100644 --- a/src/backend/nodes/nodeFuncs.c +++ b/src/backend/nodes/nodeFuncs.c @@ -59,7 +59,7 @@ exprType(const Node *expr) break; case T_ArrayRef: { - const ArrayRef *arrayref = (const ArrayRef *) expr; + const ArrayRef *arrayref = (const ArrayRef *) expr; /* slice and/or store operations yield the array type */ if (arrayref->reflowerindexpr || arrayref->refassgnexpr) @@ -91,7 +91,7 @@ exprType(const Node *expr) break; case T_SubLink: { - const SubLink *sublink = (const SubLink *) expr; + const SubLink *sublink = (const SubLink *) expr; if (sublink->subLinkType == EXPR_SUBLINK || sublink->subLinkType == ARRAY_SUBLINK) @@ -125,7 +125,7 @@ exprType(const Node *expr) break; case T_SubPlan: { - const SubPlan *subplan = (const SubPlan *) expr; + const SubPlan *subplan = (const SubPlan *) expr; if (subplan->subLinkType == EXPR_SUBLINK || subplan->subLinkType == ARRAY_SUBLINK) @@ -282,7 +282,7 @@ exprTypmod(const Node *expr) break; case T_SubLink: { - const SubLink *sublink = (const SubLink *) expr; + const SubLink *sublink = (const SubLink *) expr; if (sublink->subLinkType == EXPR_SUBLINK || sublink->subLinkType == ARRAY_SUBLINK) @@ -303,7 +303,7 @@ exprTypmod(const Node *expr) break; case T_SubPlan: { - const SubPlan *subplan = (const SubPlan *) expr; + const SubPlan *subplan = (const SubPlan *) expr; if (subplan->subLinkType == EXPR_SUBLINK || subplan->subLinkType == ARRAY_SUBLINK) @@ -341,7 +341,7 @@ exprTypmod(const Node *expr) * If all the alternatives agree on type/typmod, return that * typmod, else use -1 */ - const CaseExpr *cexpr = (const CaseExpr *) expr; + const CaseExpr *cexpr = (const CaseExpr *) expr; Oid casetype = cexpr->casetype; int32 typmod; ListCell *arg; @@ -374,7 +374,7 @@ exprTypmod(const Node *expr) * If all the elements agree on type/typmod, return that * typmod, else use -1 */ - const ArrayExpr *arrayexpr = (const ArrayExpr *) expr; + const ArrayExpr *arrayexpr = (const ArrayExpr *) expr; Oid commontype; int32 typmod; ListCell *elem; @@ -493,7 +493,7 @@ exprIsLengthCoercion(const Node *expr, int32 *coercedTypmod) */ if (expr && IsA(expr, FuncExpr)) { - const FuncExpr *func = (const FuncExpr *) expr; + const FuncExpr *func = (const FuncExpr *) expr; int nargs; Const *second_arg; @@ -707,7 +707,7 @@ exprCollation(const Node *expr) break; case T_SubLink: { - const SubLink *sublink = (const SubLink *) expr; + const SubLink *sublink = (const SubLink *) expr; if (sublink->subLinkType == EXPR_SUBLINK || sublink->subLinkType == ARRAY_SUBLINK) @@ -733,7 +733,7 @@ exprCollation(const Node *expr) break; case T_SubPlan: { - const SubPlan *subplan = (const SubPlan *) expr; + const SubPlan *subplan = (const SubPlan *) expr; if (subplan->subLinkType == EXPR_SUBLINK || subplan->subLinkType == ARRAY_SUBLINK) @@ -1137,7 +1137,7 @@ exprLocation(const Node *expr) break; case T_FuncExpr: { - const FuncExpr *fexpr = (const FuncExpr *) expr; + const FuncExpr *fexpr = (const FuncExpr *) expr; /* consider both function name and leftmost arg */ loc = leftmostLoc(fexpr->location, @@ -1157,7 +1157,7 @@ exprLocation(const Node *expr) case T_DistinctExpr: /* struct-equivalent to OpExpr */ case T_NullIfExpr: /* struct-equivalent to OpExpr */ { - const OpExpr *opexpr = (const OpExpr *) expr; + const OpExpr *opexpr = (const OpExpr *) expr; /* consider both operator name and leftmost arg */ loc = leftmostLoc(opexpr->location, @@ -1175,7 +1175,7 @@ exprLocation(const Node *expr) break; case T_BoolExpr: { - const BoolExpr *bexpr = (const BoolExpr *) expr; + const BoolExpr *bexpr = (const BoolExpr *) expr; /* * Same as above, to handle either NOT or AND/OR. We can't @@ -1188,7 +1188,7 @@ exprLocation(const Node *expr) break; case T_SubLink: { - const SubLink *sublink = (const SubLink *) expr; + const SubLink *sublink = (const SubLink *) expr; /* check the testexpr, if any, and the operator/keyword */ loc = leftmostLoc(exprLocation(sublink->testexpr), @@ -1273,7 +1273,7 @@ exprLocation(const Node *expr) break; case T_XmlExpr: { - const XmlExpr *xexpr = (const XmlExpr *) expr; + const XmlExpr *xexpr = (const XmlExpr *) expr; /* consider both function name and leftmost arg */ loc = leftmostLoc(xexpr->location, @@ -1327,7 +1327,7 @@ exprLocation(const Node *expr) break; case T_A_Expr: { - const A_Expr *aexpr = (const A_Expr *) expr; + const A_Expr *aexpr = (const A_Expr *) expr; /* use leftmost of operator or left operand (if any) */ /* we assume right operand can't be to left of operator */ @@ -1346,7 +1346,7 @@ exprLocation(const Node *expr) break; case T_FuncCall: { - const FuncCall *fc = (const FuncCall *) expr; + const FuncCall *fc = (const FuncCall *) expr; /* consider both function name and leftmost arg */ /* (we assume any ORDER BY nodes must be to right of name) */ @@ -1364,7 +1364,7 @@ exprLocation(const Node *expr) break; case T_TypeCast: { - const TypeCast *tc = (const TypeCast *) expr; + const TypeCast *tc = (const TypeCast *) expr; /* * This could represent CAST(), ::, or TypeName 'literal', so diff --git a/src/backend/nodes/outfuncs.c b/src/backend/nodes/outfuncs.c index e690194b740bbd382c5409e8befb0e06b22caf37..d6dff9de47a982d6e41ea191af2f0415be181d7f 100644 --- a/src/backend/nodes/outfuncs.c +++ b/src/backend/nodes/outfuncs.c @@ -141,7 +141,7 @@ _outToken(StringInfo str, const char *s) static void _outList(StringInfo str, const List *node) { - const ListCell *lc; + const ListCell *lc; appendStringInfoChar(str, '('); diff --git a/src/backend/nodes/print.c b/src/backend/nodes/print.c index 20aeb085d8a8c5c5d9364c06c55d00e32b1f1734..8a5e59526d24ac8110156e98c3994878ae4d920d 100644 --- a/src/backend/nodes/print.c +++ b/src/backend/nodes/print.c @@ -251,7 +251,7 @@ pretty_format_node_dump(const char *dump) void print_rt(const List *rtable) { - const ListCell *l; + const ListCell *l; int i = 1; printf("resno\trefname \trelid\tinFromCl\n"); @@ -314,7 +314,7 @@ print_expr(const Node *expr, const List *rtable) if (IsA(expr, Var)) { - const Var *var = (const Var *) expr; + const Var *var = (const Var *) expr; char *relname, *attname; @@ -348,7 +348,7 @@ print_expr(const Node *expr, const List *rtable) } else if (IsA(expr, Const)) { - const Const *c = (const Const *) expr; + const Const *c = (const Const *) expr; Oid typoutput; bool typIsVarlena; char *outputstr; @@ -368,7 +368,7 @@ print_expr(const Node *expr, const List *rtable) } else if (IsA(expr, OpExpr)) { - const OpExpr *e = (const OpExpr *) expr; + const OpExpr *e = (const OpExpr *) expr; char *opname; opname = get_opname(e->opno); @@ -387,7 +387,7 @@ print_expr(const Node *expr, const List *rtable) } else if (IsA(expr, FuncExpr)) { - const FuncExpr *e = (const FuncExpr *) expr; + const FuncExpr *e = (const FuncExpr *) expr; char *funcname; ListCell *l; @@ -412,7 +412,7 @@ print_expr(const Node *expr, const List *rtable) void print_pathkeys(const List *pathkeys, const List *rtable) { - const ListCell *i; + const ListCell *i; printf("("); foreach(i, pathkeys) @@ -452,7 +452,7 @@ print_pathkeys(const List *pathkeys, const List *rtable) void print_tl(const List *tlist, const List *rtable) { - const ListCell *tl; + const ListCell *tl; printf("(\n"); foreach(tl, tlist) diff --git a/src/backend/nodes/readfuncs.c b/src/backend/nodes/readfuncs.c index 7960793641c71b569783b1f1abe87308231c9b1a..89ddf62d4dc3ca2b2b5b8d984f4db8aa4aae51e5 100644 --- a/src/backend/nodes/readfuncs.c +++ b/src/backend/nodes/readfuncs.c @@ -49,7 +49,7 @@ #define READ_TEMP_LOCALS() \ char *token; \ int length; \ - (void) token /* possibly unused */ + (void) token /* possibly unused */ /* ... but most need both */ #define READ_LOCALS(nodeTypeName) \ @@ -195,7 +195,7 @@ _readQuery(void) READ_ENUM_FIELD(commandType, CmdType); READ_ENUM_FIELD(querySource, QuerySource); - local_node->queryId = 0; /* not saved in output format */ + local_node->queryId = 0; /* not saved in output format */ READ_BOOL_FIELD(canSetTag); READ_NODE_FIELD(utilityStmt); READ_INT_FIELD(resultRelation); diff --git a/src/backend/nodes/tidbitmap.c b/src/backend/nodes/tidbitmap.c index 17dae0d1b93eec67041a203014cb7b5ff3381ab1..728619e75dada1f9effd54f5fe85e392b1d051d9 100644 --- a/src/backend/nodes/tidbitmap.c +++ b/src/backend/nodes/tidbitmap.c @@ -956,7 +956,7 @@ tbm_lossify(TIDBitmap *tbm) * * Since we are called as soon as nentries exceeds maxentries, we should * push nentries down to significantly less than maxentries, or else we'll - * just end up doing this again very soon. We shoot for maxentries/2. + * just end up doing this again very soon. We shoot for maxentries/2. */ Assert(!tbm->iterating); Assert(tbm->status == TBM_HASH); @@ -992,14 +992,14 @@ tbm_lossify(TIDBitmap *tbm) } /* - * With a big bitmap and small work_mem, it's possible that we cannot - * get under maxentries. Again, if that happens, we'd end up uselessly + * With a big bitmap and small work_mem, it's possible that we cannot get + * under maxentries. Again, if that happens, we'd end up uselessly * calling tbm_lossify over and over. To prevent this from becoming a * performance sink, force maxentries up to at least double the current * number of entries. (In essence, we're admitting inability to fit - * within work_mem when we do this.) Note that this test will not fire - * if we broke out of the loop early; and if we didn't, the current - * number of entries is simply not reducible any further. + * within work_mem when we do this.) Note that this test will not fire if + * we broke out of the loop early; and if we didn't, the current number of + * entries is simply not reducible any further. */ if (tbm->nentries > tbm->maxentries / 2) tbm->maxentries = Min(tbm->nentries, (INT_MAX - 1) / 2) * 2; @@ -1011,8 +1011,8 @@ tbm_lossify(TIDBitmap *tbm) static int tbm_comparator(const void *left, const void *right) { - BlockNumber l = (*((PagetableEntry * const *) left))->blockno; - BlockNumber r = (*((PagetableEntry * const *) right))->blockno; + BlockNumber l = (*((PagetableEntry *const *) left))->blockno; + BlockNumber r = (*((PagetableEntry *const *) right))->blockno; if (l < r) return -1; diff --git a/src/backend/optimizer/geqo/geqo_selection.c b/src/backend/optimizer/geqo/geqo_selection.c index be64576c2fd1440041d4c12ce1efcee6a3458b3f..fbdcc5ff0c997651052c5b2191236aa4e0c2eec3 100644 --- a/src/backend/optimizer/geqo/geqo_selection.c +++ b/src/backend/optimizer/geqo/geqo_selection.c @@ -65,8 +65,8 @@ geqo_selection(PlannerInfo *root, Chromosome *momma, Chromosome *daddy, * one, when we can't. * * This code was observed to hang up in an infinite loop when the - * platform's implementation of erand48() was broken. We now always - * use our own version. + * platform's implementation of erand48() was broken. We now always use + * our own version. */ if (pool->size > 1) { diff --git a/src/backend/optimizer/path/allpaths.c b/src/backend/optimizer/path/allpaths.c index 0563cae1d7ec222df38a4cc5d88110d23950e5aa..f02954982a7e41619a0282f2ce095282b30d9cd8 100644 --- a/src/backend/optimizer/path/allpaths.c +++ b/src/backend/optimizer/path/allpaths.c @@ -50,19 +50,19 @@ join_search_hook_type join_search_hook = NULL; static void set_base_rel_sizes(PlannerInfo *root); static void set_base_rel_pathlists(PlannerInfo *root); static void set_rel_size(PlannerInfo *root, RelOptInfo *rel, - Index rti, RangeTblEntry *rte); + Index rti, RangeTblEntry *rte); static void set_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, Index rti, RangeTblEntry *rte); static void set_plain_rel_size(PlannerInfo *root, RelOptInfo *rel, - RangeTblEntry *rte); + RangeTblEntry *rte); static void set_plain_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, RangeTblEntry *rte); static void set_foreign_size(PlannerInfo *root, RelOptInfo *rel, - RangeTblEntry *rte); + RangeTblEntry *rte); static void set_foreign_pathlist(PlannerInfo *root, RelOptInfo *rel, RangeTblEntry *rte); static void set_append_rel_size(PlannerInfo *root, RelOptInfo *rel, - Index rti, RangeTblEntry *rte); + Index rti, RangeTblEntry *rte); static void set_append_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, Index rti, RangeTblEntry *rte); static void generate_mergeappend_paths(PlannerInfo *root, RelOptInfo *rel, @@ -118,7 +118,7 @@ make_one_rel(PlannerInfo *root, List *joinlist) if (brel == NULL) continue; - Assert(brel->relid == rti); /* sanity check on array */ + Assert(brel->relid == rti); /* sanity check on array */ /* ignore RTEs that are "other rels" */ if (brel->reloptkind != RELOPT_BASEREL) @@ -211,7 +211,7 @@ set_base_rel_pathlists(PlannerInfo *root) */ static void set_rel_size(PlannerInfo *root, RelOptInfo *rel, - Index rti, RangeTblEntry *rte) + Index rti, RangeTblEntry *rte) { if (rel->reloptkind == RELOPT_BASEREL && relation_excluded_by_constraints(root, rel, rte)) @@ -251,6 +251,7 @@ set_rel_size(PlannerInfo *root, RelOptInfo *rel, } break; case RTE_SUBQUERY: + /* * Subqueries don't support parameterized paths, so just go * ahead and build their paths immediately. @@ -264,6 +265,7 @@ set_rel_size(PlannerInfo *root, RelOptInfo *rel, set_values_size_estimates(root, rel); break; case RTE_CTE: + /* * CTEs don't support parameterized paths, so just go ahead * and build their paths immediately. @@ -574,8 +576,8 @@ set_append_rel_size(PlannerInfo *root, RelOptInfo *rel, /* * It is possible that constraint exclusion detected a contradiction - * within a child subquery, even though we didn't prove one above. - * If so, we can skip this child. + * within a child subquery, even though we didn't prove one above. If + * so, we can skip this child. */ if (IS_DUMMY_REL(childrel)) continue; @@ -590,7 +592,7 @@ set_append_rel_size(PlannerInfo *root, RelOptInfo *rel, /* * Accumulate per-column estimates too. We need not do anything - * for PlaceHolderVars in the parent list. If child expression + * for PlaceHolderVars in the parent list. If child expression * isn't a Var, or we didn't record a width estimate for it, we * have to fall back on a datatype-based estimate. * @@ -609,7 +611,7 @@ set_append_rel_size(PlannerInfo *root, RelOptInfo *rel, if (IsA(childvar, Var)) { - int cndx = ((Var *) childvar)->varattno - childrel->min_attr; + int cndx = ((Var *) childvar)->varattno - childrel->min_attr; child_width = childrel->attr_widths[cndx]; } @@ -664,7 +666,7 @@ set_append_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, /* * Generate access paths for each member relation, and remember the - * cheapest path for each one. Also, identify all pathkeys (orderings) + * cheapest path for each one. Also, identify all pathkeys (orderings) * and parameterizations (required_outer sets) available for the member * relations. */ @@ -708,7 +710,7 @@ set_append_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, /* * Collect lists of all the available path orderings and - * parameterizations for all the children. We use these as a + * parameterizations for all the children. We use these as a * heuristic to indicate which sort orderings and parameterizations we * should build Append and MergeAppend paths for. */ @@ -753,7 +755,7 @@ set_append_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, /* Have we already seen this param set? */ foreach(lco, all_child_outers) { - Relids existing_outers = (Relids) lfirst(lco); + Relids existing_outers = (Relids) lfirst(lco); if (bms_equal(existing_outers, childouter)) { @@ -791,7 +793,7 @@ set_append_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, * so that not that many cases actually get considered here.) * * The Append node itself cannot enforce quals, so all qual checking must - * be done in the child paths. This means that to have a parameterized + * be done in the child paths. This means that to have a parameterized * Append path, we must have the exact same parameterization for each * child path; otherwise some children might be failing to check the * moved-down quals. To make them match up, we can try to increase the @@ -799,7 +801,7 @@ set_append_rel_pathlist(PlannerInfo *root, RelOptInfo *rel, */ foreach(l, all_child_outers) { - Relids required_outer = (Relids) lfirst(l); + Relids required_outer = (Relids) lfirst(l); bool ok = true; ListCell *lcr; @@ -1115,9 +1117,9 @@ set_subquery_pathlist(PlannerInfo *root, RelOptInfo *rel, rel->subroot = subroot; /* - * It's possible that constraint exclusion proved the subquery empty. - * If so, it's convenient to turn it back into a dummy path so that we - * will recognize appropriate optimizations at this level. + * It's possible that constraint exclusion proved the subquery empty. If + * so, it's convenient to turn it back into a dummy path so that we will + * recognize appropriate optimizations at this level. */ if (is_dummy_plan(rel->subplan)) { @@ -1639,7 +1641,7 @@ qual_is_pushdown_safe(Query *subquery, Index rti, Node *qual, /* * It would be unsafe to push down window function calls, but at least for - * the moment we could never see any in a qual anyhow. (The same applies + * the moment we could never see any in a qual anyhow. (The same applies * to aggregates, which we check for in pull_var_clause below.) */ Assert(!contain_window_function(qual)); diff --git a/src/backend/optimizer/path/costsize.c b/src/backend/optimizer/path/costsize.c index e45bc121e4e6307b8efbfb55a3a0672a3e258664..480c1b7425ce66a02e2e10711359143ea7ffe59c 100644 --- a/src/backend/optimizer/path/costsize.c +++ b/src/backend/optimizer/path/costsize.c @@ -432,7 +432,7 @@ cost_index(IndexPath *path, PlannerInfo *root, double loop_count) * qual clauses that we have to evaluate as qpquals. We approximate that * list as allclauses minus any clauses appearing in indexquals. (We * assume that pointer equality is enough to recognize duplicate - * RestrictInfos.) This method neglects some considerations such as + * RestrictInfos.) This method neglects some considerations such as * clauses that needn't be checked because they are implied by a partial * index's predicate. It does not seem worth the cycles to try to factor * those things in at this stage, even though createplan.c will take pains @@ -3135,7 +3135,7 @@ get_restriction_qual_cost(PlannerInfo *root, RelOptInfo *baserel, * innerrel: inner relation under consideration * jointype: must be JOIN_SEMI or JOIN_ANTI * sjinfo: SpecialJoinInfo relevant to this join - * restrictlist: join quals + * restrictlist: join quals * Output parameters: * *semifactors is filled in (see relation.h for field definitions) */ @@ -3221,8 +3221,8 @@ compute_semi_anti_join_factors(PlannerInfo *root, * * Note: it is correct to use the inner rel's "rows" count here, even * though we might later be considering a parameterized inner path with - * fewer rows. This is because we have included all the join clauses - * in the selectivity estimate. + * fewer rows. This is because we have included all the join clauses in + * the selectivity estimate. */ if (jselec > 0) /* protect against zero divide */ { @@ -3271,17 +3271,18 @@ has_indexed_join_quals(NestPath *joinpath) indexclauses = ((IndexPath *) innerpath)->indexclauses; break; case T_BitmapHeapScan: - { - /* Accept only a simple bitmap scan, not AND/OR cases */ - Path *bmqual = ((BitmapHeapPath *) innerpath)->bitmapqual; - - if (IsA(bmqual, IndexPath)) - indexclauses = ((IndexPath *) bmqual)->indexclauses; - else - return false; - break; - } + { + /* Accept only a simple bitmap scan, not AND/OR cases */ + Path *bmqual = ((BitmapHeapPath *) innerpath)->bitmapqual; + + if (IsA(bmqual, IndexPath)) + indexclauses = ((IndexPath *) bmqual)->indexclauses; + else + return false; + break; + } default: + /* * If it's not a simple indexscan, it probably doesn't run quickly * for zero rows out, even if it's a parameterized path using all @@ -3293,8 +3294,8 @@ has_indexed_join_quals(NestPath *joinpath) /* * Examine the inner path's param clauses. Any that are from the outer * path must be found in the indexclauses list, either exactly or in an - * equivalent form generated by equivclass.c. Also, we must find at - * least one such clause, else it's a clauseless join which isn't fast. + * equivalent form generated by equivclass.c. Also, we must find at least + * one such clause, else it's a clauseless join which isn't fast. */ found_one = false; foreach(lc, innerpath->param_info->ppi_clauses) diff --git a/src/backend/optimizer/path/equivclass.c b/src/backend/optimizer/path/equivclass.c index bb196b8f2a452d7be97eb419388b1fb47db6827c..e34b9553bd4a0560e2226ab8894d4b3560a54320 100644 --- a/src/backend/optimizer/path/equivclass.c +++ b/src/backend/optimizer/path/equivclass.c @@ -494,11 +494,11 @@ add_eq_member(EquivalenceClass *ec, Expr *expr, Relids relids, * * If rel is not NULL, it identifies a specific relation we're considering * a path for, and indicates that child EC members for that relation can be - * considered. Otherwise child members are ignored. (Note: since child EC + * considered. Otherwise child members are ignored. (Note: since child EC * members aren't guaranteed unique, a non-NULL value means that there could * be more than one EC that matches the expression; if so it's order-dependent * which one you get. This is annoying but it only happens in corner cases, - * so for now we live with just reporting the first match. See also + * so for now we live with just reporting the first match. See also * generate_implied_equalities_for_indexcol and match_pathkeys_to_index.) * * If create_it is TRUE, we'll build a new EquivalenceClass when there is no @@ -922,8 +922,8 @@ generate_base_implied_equalities_broken(PlannerInfo *root, * built any join RelOptInfos. * * An annoying special case for parameterized scans is that the inner rel can - * be an appendrel child (an "other rel"). In this case we must generate - * appropriate clauses using child EC members. add_child_rel_equivalences + * be an appendrel child (an "other rel"). In this case we must generate + * appropriate clauses using child EC members. add_child_rel_equivalences * must already have been done for the child rel. * * The results are sufficient for use in merge, hash, and plain nestloop join @@ -1002,9 +1002,9 @@ generate_join_implied_equalities(PlannerInfo *root, if (ec->ec_broken) sublist = generate_join_implied_equalities_broken(root, ec, - nominal_join_relids, + nominal_join_relids, outer_relids, - nominal_inner_relids, + nominal_inner_relids, inner_appinfo); result = list_concat(result, sublist); @@ -1217,9 +1217,9 @@ generate_join_implied_equalities_broken(PlannerInfo *root, /* * If we have to translate, just brute-force apply adjust_appendrel_attrs * to all the RestrictInfos at once. This will result in returning - * RestrictInfos that are not listed in ec_derives, but there shouldn't - * be any duplication, and it's a sufficiently narrow corner case that - * we shouldn't sweat too much over it anyway. + * RestrictInfos that are not listed in ec_derives, but there shouldn't be + * any duplication, and it's a sufficiently narrow corner case that we + * shouldn't sweat too much over it anyway. */ if (inner_appinfo) result = (List *) adjust_appendrel_attrs(root, (Node *) result, @@ -1966,7 +1966,7 @@ mutate_eclass_expressions(PlannerInfo *root, * is a redundant list of clauses equating the index column to each of * the other-relation values it is known to be equal to. Any one of * these clauses can be used to create a parameterized indexscan, and there - * is no value in using more than one. (But it *is* worthwhile to create + * is no value in using more than one. (But it *is* worthwhile to create * a separate parameterized path for each one, since that leads to different * join orders.) */ @@ -2014,7 +2014,7 @@ generate_implied_equalities_for_indexcol(PlannerInfo *root, * the target relation. (Unlike regular members, the same expression * could be a child member of more than one EC. Therefore, it's * potentially order-dependent which EC a child relation's index - * column gets matched to. This is annoying but it only happens in + * column gets matched to. This is annoying but it only happens in * corner cases, so for now we live with just reporting the first * match. See also get_eclass_for_sort_expr.) */ diff --git a/src/backend/optimizer/path/indxpath.c b/src/backend/optimizer/path/indxpath.c index 05530054e1388652e4c23381d037129067460a61..2e8ccd057853ef2aa138cab47d5282157a2620c3 100644 --- a/src/backend/optimizer/path/indxpath.c +++ b/src/backend/optimizer/path/indxpath.c @@ -103,12 +103,12 @@ static List *build_paths_for_OR(PlannerInfo *root, RelOptInfo *rel, List *clauses, List *other_clauses); static List *drop_indexable_join_clauses(RelOptInfo *rel, List *clauses); static Path *choose_bitmap_and(PlannerInfo *root, RelOptInfo *rel, - List *paths); + List *paths); static int path_usage_comparator(const void *a, const void *b); static Cost bitmap_scan_cost_est(PlannerInfo *root, RelOptInfo *rel, - Path *ipath); + Path *ipath); static Cost bitmap_and_cost_est(PlannerInfo *root, RelOptInfo *rel, - List *paths); + List *paths); static PathClauseUsage *classify_index_clause_usage(Path *path, List **clauselist); static Relids get_bitmap_tree_required_outer(Path *bitmapqual); @@ -117,15 +117,15 @@ static int find_list_position(Node *node, List **nodelist); static bool check_index_only(RelOptInfo *rel, IndexOptInfo *index); static double get_loop_count(PlannerInfo *root, Relids outer_relids); static void match_restriction_clauses_to_index(RelOptInfo *rel, - IndexOptInfo *index, - IndexClauseSet *clauseset); + IndexOptInfo *index, + IndexClauseSet *clauseset); static void match_join_clauses_to_index(PlannerInfo *root, RelOptInfo *rel, IndexOptInfo *index, IndexClauseSet *clauseset, List **joinorclauses); static void match_eclass_clauses_to_index(PlannerInfo *root, - IndexOptInfo *index, - IndexClauseSet *clauseset); + IndexOptInfo *index, + IndexClauseSet *clauseset); static void match_clauses_to_index(IndexOptInfo *index, List *clauses, IndexClauseSet *clauseset); @@ -237,7 +237,7 @@ create_index_paths(PlannerInfo *root, RelOptInfo *rel) match_restriction_clauses_to_index(rel, index, &rclauseset); /* - * Build index paths from the restriction clauses. These will be + * Build index paths from the restriction clauses. These will be * non-parameterized paths. Plain paths go directly to add_path(), * bitmap paths are added to bitindexpaths to be handled below. */ @@ -245,25 +245,25 @@ create_index_paths(PlannerInfo *root, RelOptInfo *rel) &bitindexpaths); /* - * Identify the join clauses that can match the index. For the moment - * we keep them separate from the restriction clauses. Note that - * this finds only "loose" join clauses that have not been merged - * into EquivalenceClasses. Also, collect join OR clauses for later. + * Identify the join clauses that can match the index. For the moment + * we keep them separate from the restriction clauses. Note that this + * finds only "loose" join clauses that have not been merged into + * EquivalenceClasses. Also, collect join OR clauses for later. */ MemSet(&jclauseset, 0, sizeof(jclauseset)); match_join_clauses_to_index(root, rel, index, &jclauseset, &joinorclauses); /* - * Look for EquivalenceClasses that can generate joinclauses - * matching the index. + * Look for EquivalenceClasses that can generate joinclauses matching + * the index. */ MemSet(&eclauseset, 0, sizeof(eclauseset)); match_eclass_clauses_to_index(root, index, &eclauseset); /* - * If we found any plain or eclass join clauses, decide what to - * do with 'em. + * If we found any plain or eclass join clauses, decide what to do + * with 'em. */ if (jclauseset.nonempty || eclauseset.nonempty) consider_index_join_clauses(root, rel, index, @@ -287,7 +287,7 @@ create_index_paths(PlannerInfo *root, RelOptInfo *rel) * the joinclause list. Add these to bitjoinpaths. */ indexpaths = generate_bitmap_or_paths(root, rel, - joinorclauses, rel->baserestrictinfo, + joinorclauses, rel->baserestrictinfo, false); bitjoinpaths = list_concat(bitjoinpaths, indexpaths); @@ -313,7 +313,7 @@ create_index_paths(PlannerInfo *root, RelOptInfo *rel) * the most promising combination of join bitmap index paths. Note there * will be only one such path no matter how many join clauses are * available. (XXX is that good enough, or do we need to consider even - * more paths for different subsets of possible join partners? Also, + * more paths for different subsets of possible join partners? Also, * should we add in restriction bitmap paths as well?) */ if (bitjoinpaths != NIL) @@ -366,19 +366,19 @@ consider_index_join_clauses(PlannerInfo *root, RelOptInfo *rel, * We can always include any restriction clauses in the index clauses. * However, it's not obvious which subsets of the join clauses are worth * generating paths from, and it's unlikely that considering every - * possible subset is worth the cycles. Our current heuristic is based - * on the index columns, with the idea that later index columns are less + * possible subset is worth the cycles. Our current heuristic is based on + * the index columns, with the idea that later index columns are less * useful than earlier ones; therefore it's unlikely to be worth trying * combinations that would remove a clause from an earlier index column - * while adding one to a later column. Also, we know that all the - * eclass clauses for a particular column are redundant, so we should - * use only one of them. However, eclass clauses will always represent - * equality which is the strongest type of index constraint, so those - * are high-value and we should try every available combination when we - * have eclass clauses for more than one column. Furthermore, it's - * unlikely to be useful to combine an eclass clause with non-eclass - * clauses for the same index column. These considerations lead to the - * following heuristics: + * while adding one to a later column. Also, we know that all the eclass + * clauses for a particular column are redundant, so we should use only + * one of them. However, eclass clauses will always represent equality + * which is the strongest type of index constraint, so those are + * high-value and we should try every available combination when we have + * eclass clauses for more than one column. Furthermore, it's unlikely to + * be useful to combine an eclass clause with non-eclass clauses for the + * same index column. These considerations lead to the following + * heuristics: * * First, start with the restriction clauses, and add on all simple join * clauses for column 1. If there are any such join clauses, generate @@ -387,7 +387,7 @@ consider_index_join_clauses(PlannerInfo *root, RelOptInfo *rel, * any other clauses we have for column 1. * * Next, add on all simple join clauses for column 2. If there are any - * such join clauses, generate paths with this collection. If there are + * such join clauses, generate paths with this collection. If there are * eclass clauses for columns 1 or 2, generate paths with each such clause * replacing other clauses for its index column, including cases where we * use restriction or simple join clauses for one column and an eclass @@ -519,7 +519,7 @@ expand_eclass_clause_combinations(PlannerInfo *root, RelOptInfo *rel, * bitmap indexpaths are added to *bitindexpaths for later processing. * * This is a fairly simple frontend to build_index_paths(). Its reason for - * existence is mainly to handle ScalarArrayOpExpr quals properly. If the + * existence is mainly to handle ScalarArrayOpExpr quals properly. If the * index AM supports them natively, we should just include them in simple * index paths. If not, we should exclude them while building simple index * paths, and then make a separate attempt to include them in bitmap paths. @@ -533,7 +533,7 @@ get_index_paths(PlannerInfo *root, RelOptInfo *rel, ListCell *lc; /* - * Build simple index paths using the clauses. Allow ScalarArrayOpExpr + * Build simple index paths using the clauses. Allow ScalarArrayOpExpr * clauses only if the index AM supports them natively. */ indexpaths = build_index_paths(root, rel, @@ -542,17 +542,16 @@ get_index_paths(PlannerInfo *root, RelOptInfo *rel, SAOP_PER_AM, ST_ANYSCAN); /* - * Submit all the ones that can form plain IndexScan plans to add_path. - * (A plain IndexPath can represent either a plain IndexScan or an + * Submit all the ones that can form plain IndexScan plans to add_path. (A + * plain IndexPath can represent either a plain IndexScan or an * IndexOnlyScan, but for our purposes here that distinction does not - * matter. However, some of the indexes might support only bitmap scans, + * matter. However, some of the indexes might support only bitmap scans, * and those we mustn't submit to add_path here.) * - * Also, pick out the ones that are usable as bitmap scans. For that, - * we must discard indexes that don't support bitmap scans, and we - * also are only interested in paths that have some selectivity; we - * should discard anything that was generated solely for ordering - * purposes. + * Also, pick out the ones that are usable as bitmap scans. For that, we + * must discard indexes that don't support bitmap scans, and we also are + * only interested in paths that have some selectivity; we should discard + * anything that was generated solely for ordering purposes. */ foreach(lc, indexpaths) { @@ -568,9 +567,9 @@ get_index_paths(PlannerInfo *root, RelOptInfo *rel, } /* - * If the index doesn't handle ScalarArrayOpExpr clauses natively, - * check to see if there are any such clauses, and if so generate - * bitmap scan paths relying on executor-managed ScalarArrayOpExpr. + * If the index doesn't handle ScalarArrayOpExpr clauses natively, check + * to see if there are any such clauses, and if so generate bitmap scan + * paths relying on executor-managed ScalarArrayOpExpr. */ if (!index->amsearcharray) { @@ -590,7 +589,7 @@ get_index_paths(PlannerInfo *root, RelOptInfo *rel, * We return a list of paths because (1) this routine checks some cases * that should cause us to not generate any IndexPath, and (2) in some * cases we want to consider both a forward and a backward scan, so as - * to obtain both sort orders. Note that the paths are just returned + * to obtain both sort orders. Note that the paths are just returned * to the caller and not immediately fed to add_path(). * * At top level, useful_predicate should be exactly the index's predOK flag @@ -658,19 +657,19 @@ build_index_paths(PlannerInfo *root, RelOptInfo *rel, /* * 1. Collect the index clauses into a single list. * - * We build a list of RestrictInfo nodes for clauses to be used with - * this index, along with an integer list of the index column numbers - * (zero based) that each clause should be used with. The clauses are - * ordered by index key, so that the column numbers form a nondecreasing - * sequence. (This order is depended on by btree and possibly other - * places.) The lists can be empty, if the index AM allows that. + * We build a list of RestrictInfo nodes for clauses to be used with this + * index, along with an integer list of the index column numbers (zero + * based) that each clause should be used with. The clauses are ordered + * by index key, so that the column numbers form a nondecreasing sequence. + * (This order is depended on by btree and possibly other places.) The + * lists can be empty, if the index AM allows that. * - * found_clause is set true only if there's at least one index clause; - * and if saop_control is SAOP_REQUIRE, it has to be a ScalarArrayOpExpr + * found_clause is set true only if there's at least one index clause; and + * if saop_control is SAOP_REQUIRE, it has to be a ScalarArrayOpExpr * clause. * - * We also build a Relids set showing which outer rels are required - * by the selected clauses. + * We also build a Relids set showing which outer rels are required by the + * selected clauses. */ index_clauses = NIL; clause_columns = NIL; @@ -706,8 +705,8 @@ build_index_paths(PlannerInfo *root, RelOptInfo *rel, * If no clauses match the first index column, check for amoptionalkey * restriction. We can't generate a scan over an index with * amoptionalkey = false unless there's at least one index clause. - * (When working on columns after the first, this test cannot fail. - * It is always okay for columns after the first to not have any + * (When working on columns after the first, this test cannot fail. It + * is always okay for columns after the first to not have any * clauses.) */ if (index_clauses == NIL && !index->amoptionalkey) @@ -759,7 +758,7 @@ build_index_paths(PlannerInfo *root, RelOptInfo *rel, } /* - * 3. Check if an index-only scan is possible. If we're not building + * 3. Check if an index-only scan is possible. If we're not building * plain indexscans, this isn't relevant since bitmap scans don't support * index data retrieval anyway. */ @@ -865,8 +864,8 @@ build_paths_for_OR(PlannerInfo *root, RelOptInfo *rel, /* * Ignore partial indexes that do not match the query. If a partial - * index is marked predOK then we know it's OK. Otherwise, we have - * to test whether the added clauses are sufficient to imply the + * index is marked predOK then we know it's OK. Otherwise, we have to + * test whether the added clauses are sufficient to imply the * predicate. If so, we can use the index in the current context. * * We set useful_predicate to true iff the predicate was proven using @@ -904,8 +903,8 @@ build_paths_for_OR(PlannerInfo *root, RelOptInfo *rel, match_clauses_to_index(index, clauses, &clauseset); /* - * If no matches so far, and the index predicate isn't useful, - * we don't want it. + * If no matches so far, and the index predicate isn't useful, we + * don't want it. */ if (!clauseset.nonempty && !useful_predicate) continue; @@ -997,7 +996,7 @@ generate_bitmap_or_paths(PlannerInfo *root, RelOptInfo *rel, generate_bitmap_or_paths(root, rel, andargs, all_clauses, - restriction_only)); + restriction_only)); } else { @@ -1053,7 +1052,7 @@ generate_bitmap_or_paths(PlannerInfo *root, RelOptInfo *rel, * * This is a helper for generate_bitmap_or_paths(). We leave OR clauses * in the list whether they are joins or not, since we might be able to - * extract a restriction item from an OR list. It's safe to leave such + * extract a restriction item from an OR list. It's safe to leave such * clauses in the list because match_clauses_to_index() will ignore them, * so there's no harm in passing such clauses to build_paths_for_OR(). */ @@ -1361,7 +1360,7 @@ bitmap_and_cost_est(PlannerInfo *root, RelOptInfo *rel, List *paths) apath.path.type = T_BitmapAndPath; apath.path.pathtype = T_BitmapAnd; apath.path.parent = rel; - apath.path.param_info = NULL; /* not used in bitmap trees */ + apath.path.param_info = NULL; /* not used in bitmap trees */ apath.path.pathkeys = NIL; apath.bitmapquals = paths; cost_bitmap_and_node(&apath, root); @@ -1464,7 +1463,7 @@ get_bitmap_tree_required_outer(Path *bitmapqual) foreach(lc, ((BitmapAndPath *) bitmapqual)->bitmapquals) { result = bms_join(result, - get_bitmap_tree_required_outer((Path *) lfirst(lc))); + get_bitmap_tree_required_outer((Path *) lfirst(lc))); } } else if (IsA(bitmapqual, BitmapOrPath)) @@ -1472,7 +1471,7 @@ get_bitmap_tree_required_outer(Path *bitmapqual) foreach(lc, ((BitmapOrPath *) bitmapqual)->bitmapquals) { result = bms_join(result, - get_bitmap_tree_required_outer((Path *) lfirst(lc))); + get_bitmap_tree_required_outer((Path *) lfirst(lc))); } } else @@ -1581,16 +1580,16 @@ check_index_only(RelOptInfo *rel, IndexOptInfo *index) return false; /* - * Check that all needed attributes of the relation are available from - * the index. + * Check that all needed attributes of the relation are available from the + * index. * * XXX this is overly conservative for partial indexes, since we will * consider attributes involved in the index predicate as required even - * though the predicate won't need to be checked at runtime. (The same - * is true for attributes used only in index quals, if we are certain - * that the index is not lossy.) However, it would be quite expensive - * to determine that accurately at this point, so for now we take the - * easy way out. + * though the predicate won't need to be checked at runtime. (The same is + * true for attributes used only in index quals, if we are certain that + * the index is not lossy.) However, it would be quite expensive to + * determine that accurately at this point, so for now we take the easy + * way out. */ /* @@ -1603,7 +1602,7 @@ check_index_only(RelOptInfo *rel, IndexOptInfo *index) /* Add all the attributes used by restriction clauses. */ foreach(lc, rel->baserestrictinfo) { - RestrictInfo *rinfo = (RestrictInfo *) lfirst(lc); + RestrictInfo *rinfo = (RestrictInfo *) lfirst(lc); pull_varattnos((Node *) rinfo->clause, rel->relid, &attrs_used); } @@ -1611,7 +1610,7 @@ check_index_only(RelOptInfo *rel, IndexOptInfo *index) /* Construct a bitmapset of columns stored in the index. */ for (i = 0; i < index->ncolumns; i++) { - int attno = index->indexkeys[i]; + int attno = index->indexkeys[i]; /* * For the moment, we just ignore index expressions. It might be nice @@ -1642,7 +1641,7 @@ check_index_only(RelOptInfo *rel, IndexOptInfo *index) * Since we produce parameterized paths before we've begun to generate join * relations, it's impossible to predict exactly how many times a parameterized * path will be iterated; we don't know the size of the relation that will be - * on the outside of the nestloop. However, we should try to account for + * on the outside of the nestloop. However, we should try to account for * multiple iterations somehow in costing the path. The heuristic embodied * here is to use the rowcount of the smallest other base relation needed in * the join clauses used by the path. (We could alternatively consider the @@ -1676,7 +1675,7 @@ get_loop_count(PlannerInfo *root, Relids outer_relids) outer_rel = root->simple_rel_array[relid]; if (outer_rel == NULL) continue; - Assert(outer_rel->relid == relid); /* sanity check on array */ + Assert(outer_rel->relid == relid); /* sanity check on array */ /* Other relation could be proven empty, if so ignore */ if (IS_DUMMY_REL(outer_rel)) @@ -1851,7 +1850,7 @@ match_clause_to_index(IndexOptInfo *index, * doesn't involve a volatile function or a Var of the index's relation. * In particular, Vars belonging to other relations of the query are * accepted here, since a clause of that form can be used in a - * parameterized indexscan. It's the responsibility of higher code levels + * parameterized indexscan. It's the responsibility of higher code levels * to manage restriction and join clauses appropriately. * * Note: we do need to check for Vars of the index's relation on the @@ -2149,7 +2148,7 @@ match_pathkeys_to_index(IndexOptInfo *index, List *pathkeys, List *clause_columns = NIL; ListCell *lc1; - *orderby_clauses_p = NIL; /* set default results */ + *orderby_clauses_p = NIL; /* set default results */ *clause_columns_p = NIL; /* Only indexes with the amcanorderbyop property are interesting here */ @@ -2195,9 +2194,9 @@ match_pathkeys_to_index(IndexOptInfo *index, List *pathkeys, /* * We allow any column of the index to match each pathkey; they - * don't have to match left-to-right as you might expect. This - * is correct for GiST, which is the sole existing AM supporting - * amcanorderbyop. We might need different logic in future for + * don't have to match left-to-right as you might expect. This is + * correct for GiST, which is the sole existing AM supporting + * amcanorderbyop. We might need different logic in future for * other implementations. */ for (indexcol = 0; indexcol < index->ncolumns; indexcol++) @@ -2393,8 +2392,8 @@ eclass_member_matches_indexcol(EquivalenceClass *ec, EquivalenceMember *em, * If it's a btree index, we can reject it if its opfamily isn't * compatible with the EC, since no clause generated from the EC could be * used with the index. For non-btree indexes, we can't easily tell - * whether clauses generated from the EC could be used with the index, - * so don't check the opfamily. This might mean we return "true" for a + * whether clauses generated from the EC could be used with the index, so + * don't check the opfamily. This might mean we return "true" for a * useless EC, so we have to recheck the results of * generate_implied_equalities_for_indexcol; see * match_eclass_clauses_to_index. @@ -2425,7 +2424,7 @@ eclass_member_matches_indexcol(EquivalenceClass *ec, EquivalenceMember *em, * if it is true. * 2. A list of expressions in this relation, and a corresponding list of * equality operators. The caller must have already checked that the operators - * represent equality. (Note: the operators could be cross-type; the + * represent equality. (Note: the operators could be cross-type; the * expressions should correspond to their RHS inputs.) * * The caller need only supply equality conditions arising from joins; @@ -2571,7 +2570,7 @@ relation_has_unique_index_for(PlannerInfo *root, RelOptInfo *rel, * notion of equality. */ - matched = true; /* column is unique */ + matched = true; /* column is unique */ break; } @@ -3300,9 +3299,9 @@ adjust_rowcompare_for_index(RowCompareExpr *clause, /* * See how many of the remaining columns match some index column in the - * same way. As in match_clause_to_indexcol(), the "other" side of - * any potential index condition is OK as long as it doesn't use Vars from - * the indexed relation. + * same way. As in match_clause_to_indexcol(), the "other" side of any + * potential index condition is OK as long as it doesn't use Vars from the + * indexed relation. */ matching_cols = 1; largs_cell = lnext(list_head(clause->largs)); diff --git a/src/backend/optimizer/path/joinpath.c b/src/backend/optimizer/path/joinpath.c index 446319d13560a6c7289c77a3b0d97f110d59589c..65f86194e155508baa2e7c690a08be7d0380299a 100644 --- a/src/backend/optimizer/path/joinpath.c +++ b/src/backend/optimizer/path/joinpath.c @@ -103,7 +103,7 @@ add_paths_to_joinrel(PlannerInfo *root, /* * If it's SEMI or ANTI join, compute correction factors for cost - * estimation. These will be the same for all paths. + * estimation. These will be the same for all paths. */ if (jointype == JOIN_SEMI || jointype == JOIN_ANTI) compute_semi_anti_join_factors(root, outerrel, innerrel, @@ -118,7 +118,7 @@ add_paths_to_joinrel(PlannerInfo *root, * to the parameter source rel instead of joining to the other input rel. * This restriction reduces the number of parameterized paths we have to * deal with at higher join levels, without compromising the quality of - * the resulting plan. We express the restriction as a Relids set that + * the resulting plan. We express the restriction as a Relids set that * must overlap the parameterization of any proposed join path. */ foreach(lc, root->join_info_list) @@ -136,7 +136,7 @@ add_paths_to_joinrel(PlannerInfo *root, !bms_overlap(joinrel->relids, sjinfo->min_lefthand)) param_source_rels = bms_join(param_source_rels, bms_difference(root->all_baserels, - sjinfo->min_righthand)); + sjinfo->min_righthand)); /* full joins constrain both sides symmetrically */ if (sjinfo->jointype == JOIN_FULL && @@ -144,7 +144,7 @@ add_paths_to_joinrel(PlannerInfo *root, !bms_overlap(joinrel->relids, sjinfo->min_righthand)) param_source_rels = bms_join(param_source_rels, bms_difference(root->all_baserels, - sjinfo->min_lefthand)); + sjinfo->min_lefthand)); } /* @@ -216,11 +216,11 @@ try_nestloop_path(PlannerInfo *root, List *pathkeys) { Relids required_outer; - JoinCostWorkspace workspace; + JoinCostWorkspace workspace; /* - * Check to see if proposed path is still parameterized, and reject if - * the parameterization wouldn't be sensible. + * Check to see if proposed path is still parameterized, and reject if the + * parameterization wouldn't be sensible. */ required_outer = calc_nestloop_required_outer(outer_path, inner_path); @@ -289,14 +289,14 @@ try_mergejoin_path(PlannerInfo *root, List *innersortkeys) { Relids required_outer; - JoinCostWorkspace workspace; + JoinCostWorkspace workspace; /* - * Check to see if proposed path is still parameterized, and reject if - * the parameterization wouldn't be sensible. + * Check to see if proposed path is still parameterized, and reject if the + * parameterization wouldn't be sensible. */ required_outer = calc_non_nestloop_required_outer(outer_path, - inner_path); + inner_path); if (required_outer && !bms_overlap(required_outer, param_source_rels)) { @@ -368,14 +368,14 @@ try_hashjoin_path(PlannerInfo *root, List *hashclauses) { Relids required_outer; - JoinCostWorkspace workspace; + JoinCostWorkspace workspace; /* - * Check to see if proposed path is still parameterized, and reject if - * the parameterization wouldn't be sensible. + * Check to see if proposed path is still parameterized, and reject if the + * parameterization wouldn't be sensible. */ required_outer = calc_non_nestloop_required_outer(outer_path, - inner_path); + inner_path); if (required_outer && !bms_overlap(required_outer, param_source_rels)) { @@ -487,7 +487,7 @@ sort_inner_and_outer(PlannerInfo *root, * * This function intentionally does not consider parameterized input paths * (implicit in the fact that it only looks at cheapest_total_path, which - * is always unparameterized). If we did so, we'd have a combinatorial + * is always unparameterized). If we did so, we'd have a combinatorial * explosion of mergejoin paths of dubious value. This interacts with * decisions elsewhere that also discriminate against mergejoins with * parameterized inputs; see comments in src/backend/optimizer/README. @@ -582,8 +582,8 @@ sort_inner_and_outer(PlannerInfo *root, * And now we can make the path. * * Note: it's possible that the cheapest paths will already be sorted - * properly. try_mergejoin_path will detect that case and suppress - * an explicit sort step, so we needn't do so here. + * properly. try_mergejoin_path will detect that case and suppress an + * explicit sort step, so we needn't do so here. */ try_mergejoin_path(root, joinrel, @@ -733,8 +733,8 @@ match_unsorted_outer(PlannerInfo *root, /* * If we need to unique-ify the outer path, it's pointless to consider - * any but the cheapest outer. (XXX we don't consider parameterized - * outers, nor inners, for unique-ified cases. Should we?) + * any but the cheapest outer. (XXX we don't consider parameterized + * outers, nor inners, for unique-ified cases. Should we?) */ if (save_jointype == JOIN_UNIQUE_OUTER) { @@ -774,9 +774,9 @@ match_unsorted_outer(PlannerInfo *root, { /* * Consider nestloop joins using this outer path and various - * available paths for the inner relation. We consider the - * cheapest-total paths for each available parameterization of - * the inner relation, including the unparameterized case. + * available paths for the inner relation. We consider the + * cheapest-total paths for each available parameterization of the + * inner relation, including the unparameterized case. */ ListCell *lc2; @@ -847,8 +847,8 @@ match_unsorted_outer(PlannerInfo *root, /* * Generate a mergejoin on the basis of sorting the cheapest inner. * Since a sort will be needed, only cheapest total cost matters. (But - * try_mergejoin_path will do the right thing if - * inner_cheapest_total is already correctly sorted.) + * try_mergejoin_path will do the right thing if inner_cheapest_total + * is already correctly sorted.) */ try_mergejoin_path(root, joinrel, @@ -873,9 +873,9 @@ match_unsorted_outer(PlannerInfo *root, * mergejoin using a subset of the merge clauses. Here, we consider * both cheap startup cost and cheap total cost. * - * Currently we do not consider parameterized inner paths here. - * This interacts with decisions elsewhere that also discriminate - * against mergejoins with parameterized inputs; see comments in + * Currently we do not consider parameterized inner paths here. This + * interacts with decisions elsewhere that also discriminate against + * mergejoins with parameterized inputs; see comments in * src/backend/optimizer/README. * * As we shorten the sortkey list, we should consider only paths that @@ -1189,7 +1189,7 @@ hash_inner_and_outer(PlannerInfo *root, if (outerpath == cheapest_startup_outer && innerpath == cheapest_total_inner) - continue; /* already tried it */ + continue; /* already tried it */ try_hashjoin_path(root, joinrel, diff --git a/src/backend/optimizer/path/joinrels.c b/src/backend/optimizer/path/joinrels.c index 24d46515070e8602d31db2fc9dc264c9ec6543b5..e6a0f8dab6db311b816b2c139eb5dbc343e181ee 100644 --- a/src/backend/optimizer/path/joinrels.c +++ b/src/backend/optimizer/path/joinrels.c @@ -90,7 +90,7 @@ join_search_one_level(PlannerInfo *root, int level) if (level == 2) /* consider remaining initial rels */ other_rels = lnext(r); - else /* consider all initial rels */ + else /* consider all initial rels */ other_rels = list_head(joinrels[1]); make_rels_by_clause_joins(root, @@ -180,7 +180,7 @@ join_search_one_level(PlannerInfo *root, int level) /*---------- * Normally, we should always have made at least one join of the current * level. However, when special joins are involved, there may be no legal - * way to make an N-way join for some values of N. For example consider + * way to make an N-way join for some values of N. For example consider * * SELECT ... FROM t1 WHERE * x IN (SELECT ... FROM t2,t3 WHERE ...) AND diff --git a/src/backend/optimizer/path/orindxpath.c b/src/backend/optimizer/path/orindxpath.c index cab795189199d57efc8302804c02aa56475eeb52..c918c4e8da92f57a0d75c7b82579bdf9890d4f8b 100644 --- a/src/backend/optimizer/path/orindxpath.c +++ b/src/backend/optimizer/path/orindxpath.c @@ -95,8 +95,8 @@ create_or_index_quals(PlannerInfo *root, RelOptInfo *rel) /* * Find potentially interesting OR joinclauses. We can use any joinclause * that is considered safe to move to this rel by the parameterized-path - * machinery, even though what we are going to do with it is not exactly - * a parameterized path. + * machinery, even though what we are going to do with it is not exactly a + * parameterized path. */ foreach(i, rel->joininfo) { @@ -109,7 +109,7 @@ create_or_index_quals(PlannerInfo *root, RelOptInfo *rel) * Use the generate_bitmap_or_paths() machinery to estimate the * value of each OR clause. We can use regular restriction * clauses along with the OR clause contents to generate - * indexquals. We pass restriction_only = true so that any + * indexquals. We pass restriction_only = true so that any * sub-clauses that are actually joins will be ignored. */ List *orpaths; diff --git a/src/backend/optimizer/path/pathkeys.c b/src/backend/optimizer/path/pathkeys.c index 0603a94e482e52f679e779aae1f6171a5726eb49..20a5644edd83ecc17782922a1232f50ca0accb5e 100644 --- a/src/backend/optimizer/path/pathkeys.c +++ b/src/backend/optimizer/path/pathkeys.c @@ -223,7 +223,7 @@ canonicalize_pathkeys(PlannerInfo *root, List *pathkeys) * * If rel is not NULL, it identifies a specific relation we're considering * a path for, and indicates that child EC members for that relation can be - * considered. Otherwise child members are ignored. (See the comments for + * considered. Otherwise child members are ignored. (See the comments for * get_eclass_for_sort_expr.) * * create_it is TRUE if we should create any missing EquivalenceClass diff --git a/src/backend/optimizer/plan/createplan.c b/src/backend/optimizer/plan/createplan.c index c34b9b8c38ea34b285e34d6688bbb78e3399c4b6..65ad1694b07bd6f46978fea43dedf83965ee7b49 100644 --- a/src/backend/optimizer/plan/createplan.c +++ b/src/backend/optimizer/plan/createplan.c @@ -1138,10 +1138,10 @@ create_indexscan_plan(PlannerInfo *root, /* * The qpqual list must contain all restrictions not automatically handled * by the index, other than pseudoconstant clauses which will be handled - * by a separate gating plan node. All the predicates in the indexquals + * by a separate gating plan node. All the predicates in the indexquals * will be checked (either by the index itself, or by nodeIndexscan.c), * but if there are any "special" operators involved then they must be - * included in qpqual. The upshot is that qpqual must contain + * included in qpqual. The upshot is that qpqual must contain * scan_clauses minus whatever appears in indexquals. * * In normal cases simple pointer equality checks will be enough to spot @@ -1189,7 +1189,7 @@ create_indexscan_plan(PlannerInfo *root, get_parse_rowmark(root->parse, baserelid) == NULL) if (predicate_implied_by(clausel, best_path->indexinfo->indpred)) - continue; /* implied by index predicate */ + continue; /* implied by index predicate */ } } qpqual = lappend(qpqual, rinfo); @@ -1228,7 +1228,7 @@ create_indexscan_plan(PlannerInfo *root, indexoid, fixed_indexquals, fixed_indexorderbys, - best_path->indexinfo->indextlist, + best_path->indexinfo->indextlist, best_path->indexscandir); else scan_plan = (Scan *) make_indexscan(tlist, @@ -1278,15 +1278,15 @@ create_bitmap_scan_plan(PlannerInfo *root, /* * The qpqual list must contain all restrictions not automatically handled * by the index, other than pseudoconstant clauses which will be handled - * by a separate gating plan node. All the predicates in the indexquals + * by a separate gating plan node. All the predicates in the indexquals * will be checked (either by the index itself, or by * nodeBitmapHeapscan.c), but if there are any "special" operators - * involved then they must be added to qpqual. The upshot is that qpqual + * involved then they must be added to qpqual. The upshot is that qpqual * must contain scan_clauses minus whatever appears in indexquals. * * This loop is similar to the comparable code in create_indexscan_plan(), * but with some differences because it has to compare the scan clauses to - * stripped (no RestrictInfos) indexquals. See comments there for more + * stripped (no RestrictInfos) indexquals. See comments there for more * info. * * In normal cases simple equal() checks will be enough to spot duplicate @@ -1880,14 +1880,14 @@ create_foreignscan_plan(PlannerInfo *root, ForeignPath *best_path, Assert(rte->rtekind == RTE_RELATION); /* - * Sort clauses into best execution order. We do this first since the - * FDW might have more info than we do and wish to adjust the ordering. + * Sort clauses into best execution order. We do this first since the FDW + * might have more info than we do and wish to adjust the ordering. */ scan_clauses = order_qual_clauses(root, scan_clauses); /* * Let the FDW perform its processing on the restriction clauses and - * generate the plan node. Note that the FDW might remove restriction + * generate the plan node. Note that the FDW might remove restriction * clauses that it intends to execute remotely, or even add more (if it * has selected some join clauses for remote use but also wants them * rechecked locally). @@ -2005,7 +2005,7 @@ create_nestloop_plan(PlannerInfo *root, bms_overlap(((PlaceHolderVar *) nlp->paramval)->phrels, outerrelids) && bms_is_subset(find_placeholder_info(root, - (PlaceHolderVar *) nlp->paramval, + (PlaceHolderVar *) nlp->paramval, false)->ph_eval_at, outerrelids)) { @@ -2523,9 +2523,9 @@ replace_nestloop_params_mutator(Node *node, PlannerInfo *root) /* * If not to be replaced, just return the PlaceHolderVar unmodified. - * We use bms_overlap as a cheap/quick test to see if the PHV might - * be evaluated in the outer rels, and then grab its PlaceHolderInfo - * to tell for sure. + * We use bms_overlap as a cheap/quick test to see if the PHV might be + * evaluated in the outer rels, and then grab its PlaceHolderInfo to + * tell for sure. */ if (!bms_overlap(phv->phrels, root->curOuterRels)) return node; @@ -2612,7 +2612,7 @@ fix_indexqual_references(PlannerInfo *root, IndexPath *index_path) /* * Check to see if the indexkey is on the right; if so, commute - * the clause. The indexkey should be the side that refers to + * the clause. The indexkey should be the side that refers to * (only) the base relation. */ if (!bms_equal(rinfo->left_relids, index->rel->relids)) @@ -3690,13 +3690,12 @@ prepare_sort_from_pathkeys(PlannerInfo *root, Plan *lefttree, List *pathkeys, { /* * If we are given a sort column number to match, only consider - * the single TLE at that position. It's possible that there - * is no such TLE, in which case fall through and generate a - * resjunk targetentry (we assume this must have happened in the - * parent plan as well). If there is a TLE but it doesn't match - * the pathkey's EC, we do the same, which is probably the wrong - * thing but we'll leave it to caller to complain about the - * mismatch. + * the single TLE at that position. It's possible that there is + * no such TLE, in which case fall through and generate a resjunk + * targetentry (we assume this must have happened in the parent + * plan as well). If there is a TLE but it doesn't match the + * pathkey's EC, we do the same, which is probably the wrong thing + * but we'll leave it to caller to complain about the mismatch. */ tle = get_tle_by_resno(tlist, reqColIdx[numsortkeys]); if (tle) @@ -3746,11 +3745,11 @@ prepare_sort_from_pathkeys(PlannerInfo *root, Plan *lefttree, List *pathkeys, if (!tle) { /* - * No matching tlist item; look for a computable expression. - * Note that we treat Aggrefs as if they were variables; this - * is necessary when attempting to sort the output from an Agg - * node for use in a WindowFunc (since grouping_planner will - * have treated the Aggrefs as variables, too). + * No matching tlist item; look for a computable expression. Note + * that we treat Aggrefs as if they were variables; this is + * necessary when attempting to sort the output from an Agg node + * for use in a WindowFunc (since grouping_planner will have + * treated the Aggrefs as variables, too). */ Expr *sortexpr = NULL; @@ -3769,7 +3768,8 @@ prepare_sort_from_pathkeys(PlannerInfo *root, Plan *lefttree, List *pathkeys, continue; /* - * Ignore child members unless they match the rel being sorted. + * Ignore child members unless they match the rel being + * sorted. */ if (em->em_is_child && !bms_equal(em->em_relids, relids)) @@ -3817,7 +3817,7 @@ prepare_sort_from_pathkeys(PlannerInfo *root, Plan *lefttree, List *pathkeys, NULL, true); tlist = lappend(tlist, tle); - lefttree->targetlist = tlist; /* just in case NIL before */ + lefttree->targetlist = tlist; /* just in case NIL before */ } /* @@ -3877,8 +3877,7 @@ find_ec_member_for_tle(EquivalenceClass *ec, /* * We shouldn't be trying to sort by an equivalence class that - * contains a constant, so no need to consider such cases any - * further. + * contains a constant, so no need to consider such cases any further. */ if (em->em_is_const) continue; diff --git a/src/backend/optimizer/plan/initsplan.c b/src/backend/optimizer/plan/initsplan.c index 231e8353733d5a356369c0a037d82af1f6b43710..3c7fa632b8ebe26c1d91e83e0ba1fa9c03b9e919 100644 --- a/src/backend/optimizer/plan/initsplan.c +++ b/src/backend/optimizer/plan/initsplan.c @@ -192,9 +192,9 @@ add_vars_to_targetlist(PlannerInfo *root, List *vars, where_needed); /* - * If we are creating PlaceHolderInfos, mark them with the - * correct maybe-needed locations. Otherwise, it's too late to - * change that. + * If we are creating PlaceHolderInfos, mark them with the correct + * maybe-needed locations. Otherwise, it's too late to change + * that. */ if (create_new_ph) mark_placeholder_maybe_needed(root, phinfo, where_needed); diff --git a/src/backend/optimizer/plan/planagg.c b/src/backend/optimizer/plan/planagg.c index 7e2c6d2c31f269a1deaaabe1bf0d0930150ebe18..be52d16ff062cfe4ccf4ff486f350743d3e0a68e 100644 --- a/src/backend/optimizer/plan/planagg.c +++ b/src/backend/optimizer/plan/planagg.c @@ -116,9 +116,9 @@ preprocess_minmax_aggregates(PlannerInfo *root, List *tlist) rtr = (RangeTblRef *) jtnode; rte = planner_rt_fetch(rtr->rtindex, root); if (rte->rtekind == RTE_RELATION) - /* ordinary relation, ok */ ; + /* ordinary relation, ok */ ; else if (rte->rtekind == RTE_SUBQUERY && rte->inh) - /* flattened UNION ALL subquery, ok */ ; + /* flattened UNION ALL subquery, ok */ ; else return; diff --git a/src/backend/optimizer/plan/planmain.c b/src/backend/optimizer/plan/planmain.c index c439e9652c1b2be8456eac2a382cc8b02b4699f4..9838dc45d5ea5ef9aa55a3cacdb7356fc182944a 100644 --- a/src/backend/optimizer/plan/planmain.c +++ b/src/backend/optimizer/plan/planmain.c @@ -146,8 +146,8 @@ query_planner(PlannerInfo *root, List *tlist, /* * Make a flattened version of the rangetable for faster access (this is - * OK because the rangetable won't change any more), and set up an - * empty array for indexing base relations. + * OK because the rangetable won't change any more), and set up an empty + * array for indexing base relations. */ setup_simple_rel_arrays(root); diff --git a/src/backend/optimizer/plan/planner.c b/src/backend/optimizer/plan/planner.c index 0b1ee971df15137f697e5ab8e18b08051237b009..df76341c0a3e6963c41373aa0f83783533194d63 100644 --- a/src/backend/optimizer/plan/planner.c +++ b/src/backend/optimizer/plan/planner.c @@ -766,9 +766,9 @@ inheritance_planner(PlannerInfo *root) /* * The rowMarks list might contain references to subquery RTEs, so - * make a copy that we can apply ChangeVarNodes to. (Fortunately, - * the executor doesn't need to see the modified copies --- we can - * just pass it the original rowMarks list.) + * make a copy that we can apply ChangeVarNodes to. (Fortunately, the + * executor doesn't need to see the modified copies --- we can just + * pass it the original rowMarks list.) */ subroot.rowMarks = (List *) copyObject(root->rowMarks); @@ -784,10 +784,11 @@ inheritance_planner(PlannerInfo *root) /* * If this isn't the first child Query, generate duplicates of all - * subquery RTEs, and adjust Var numbering to reference the duplicates. - * To simplify the loop logic, we scan the original rtable not the - * copy just made by adjust_appendrel_attrs; that should be OK since - * subquery RTEs couldn't contain any references to the target rel. + * subquery RTEs, and adjust Var numbering to reference the + * duplicates. To simplify the loop logic, we scan the original rtable + * not the copy just made by adjust_appendrel_attrs; that should be OK + * since subquery RTEs couldn't contain any references to the target + * rel. */ if (final_rtable != NIL) { @@ -800,7 +801,7 @@ inheritance_planner(PlannerInfo *root) if (rte->rtekind == RTE_SUBQUERY) { - Index newrti; + Index newrti; /* * The RTE can't contain any references to its own RT @@ -849,7 +850,7 @@ inheritance_planner(PlannerInfo *root) else final_rtable = list_concat(final_rtable, list_copy_tail(subroot.parse->rtable, - list_length(final_rtable))); + list_length(final_rtable))); /* * We need to collect all the RelOptInfos from all child plans into @@ -1317,18 +1318,17 @@ grouping_planner(PlannerInfo *root, double tuple_fraction) need_sort_for_grouping = true; /* - * Always override create_plan's tlist, so that we don't - * sort useless data from a "physical" tlist. + * Always override create_plan's tlist, so that we don't sort + * useless data from a "physical" tlist. */ need_tlist_eval = true; } /* - * create_plan returns a plan with just a "flat" tlist of - * required Vars. Usually we need to insert the sub_tlist as the - * tlist of the top plan node. However, we can skip that if we - * determined that whatever create_plan chose to return will be - * good enough. + * create_plan returns a plan with just a "flat" tlist of required + * Vars. Usually we need to insert the sub_tlist as the tlist of + * the top plan node. However, we can skip that if we determined + * that whatever create_plan chose to return will be good enough. */ if (need_tlist_eval) { @@ -1546,7 +1546,7 @@ grouping_planner(PlannerInfo *root, double tuple_fraction) * * Note: it's essential here to use PVC_INCLUDE_AGGREGATES so that * Vars mentioned only in aggregate expressions aren't pulled out - * as separate targetlist entries. Otherwise we could be putting + * as separate targetlist entries. Otherwise we could be putting * ungrouped Vars directly into an Agg node's tlist, resulting in * undefined behavior. */ @@ -2653,8 +2653,8 @@ make_subplanTargetList(PlannerInfo *root, } /* - * Otherwise, we must build a tlist containing all grouping columns, - * plus any other Vars mentioned in the targetlist and HAVING qual. + * Otherwise, we must build a tlist containing all grouping columns, plus + * any other Vars mentioned in the targetlist and HAVING qual. */ sub_tlist = NIL; non_group_cols = NIL; @@ -2705,8 +2705,8 @@ make_subplanTargetList(PlannerInfo *root, else { /* - * Non-grouping column, so just remember the expression - * for later call to pull_var_clause. There's no need for + * Non-grouping column, so just remember the expression for + * later call to pull_var_clause. There's no need for * pull_var_clause to examine the TargetEntry node itself. */ non_group_cols = lappend(non_group_cols, tle->expr); @@ -2733,7 +2733,7 @@ make_subplanTargetList(PlannerInfo *root, * add them to the result tlist if not already present. (A Var used * directly as a GROUP BY item will be present already.) Note this * includes Vars used in resjunk items, so we are covering the needs of - * ORDER BY and window specifications. Vars used within Aggrefs will be + * ORDER BY and window specifications. Vars used within Aggrefs will be * pulled out here, too. */ non_group_vars = pull_var_clause((Node *) non_group_cols, diff --git a/src/backend/optimizer/plan/setrefs.c b/src/backend/optimizer/plan/setrefs.c index db301e6c595f81cdb89c1bbf555455310bdebf2d..f375b5f76d4ded16f5045587d86f80cedeef82b3 100644 --- a/src/backend/optimizer/plan/setrefs.c +++ b/src/backend/optimizer/plan/setrefs.c @@ -327,7 +327,7 @@ set_plan_refs(PlannerInfo *root, Plan *plan, int rtoffset) break; case T_IndexOnlyScan: { - IndexOnlyScan *splan = (IndexOnlyScan *) plan; + IndexOnlyScan *splan = (IndexOnlyScan *) plan; return set_indexonlyscan_references(root, splan, rtoffset); } @@ -573,9 +573,9 @@ set_plan_refs(PlannerInfo *root, Plan *plan, int rtoffset) lcrr, splan->resultRelations, lcp, splan->plans) { - List *rlist = (List *) lfirst(lcrl); - Index resultrel = lfirst_int(lcrr); - Plan *subplan = (Plan *) lfirst(lcp); + List *rlist = (List *) lfirst(lcrl); + Index resultrel = lfirst_int(lcrr); + Plan *subplan = (Plan *) lfirst(lcp); rlist = set_returning_clause_references(root, rlist, @@ -590,7 +590,7 @@ set_plan_refs(PlannerInfo *root, Plan *plan, int rtoffset) * Set up the visible plan targetlist as being the same as * the first RETURNING list. This is for the use of * EXPLAIN; the executor won't pay any attention to the - * targetlist. We postpone this step until here so that + * targetlist. We postpone this step until here so that * we don't have to do set_returning_clause_references() * twice on identical targetlists. */ @@ -1885,7 +1885,7 @@ record_plan_function_dependency(PlannerInfo *root, Oid funcid) */ inval_item->cacheId = PROCOID; inval_item->hashValue = GetSysCacheHashValue1(PROCOID, - ObjectIdGetDatum(funcid)); + ObjectIdGetDatum(funcid)); root->glob->invalItems = lappend(root->glob->invalItems, inval_item); } diff --git a/src/backend/optimizer/plan/subselect.c b/src/backend/optimizer/plan/subselect.c index fb6c7045484e9ed68c10f44bb0adaaecd6c599d4..8ce6bee8561074e164fd42f4ea8ca8fd62861051 100644 --- a/src/backend/optimizer/plan/subselect.c +++ b/src/backend/optimizer/plan/subselect.c @@ -1822,8 +1822,8 @@ process_sublinks_mutator(Node *node, process_sublinks_context *context) } /* - * Don't recurse into the arguments of an outer PHV or aggregate here. - * Any SubLinks in the arguments have to be dealt with at the outer query + * Don't recurse into the arguments of an outer PHV or aggregate here. Any + * SubLinks in the arguments have to be dealt with at the outer query * level; they'll be handled when build_subplan collects the PHV or Aggref * into the arguments to be passed down to the current subplan. */ diff --git a/src/backend/optimizer/prep/prepjointree.c b/src/backend/optimizer/prep/prepjointree.c index 47ddae6992f7cf33431afe55aacd0a6a4e449f95..be1219eb3d12373cf166ab56697951839f437b95 100644 --- a/src/backend/optimizer/prep/prepjointree.c +++ b/src/backend/optimizer/prep/prepjointree.c @@ -332,6 +332,7 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, j->rarg = pull_up_sublinks_jointree_recurse(root, j->rarg, &child_rels); + /* * Now recursively process the pulled-up quals. Any inserted * joins can get stacked onto either j->larg or j->rarg, @@ -357,6 +358,7 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, j->rarg = pull_up_sublinks_jointree_recurse(root, j->rarg, &child_rels); + /* * Now recursively process the pulled-up quals. Any inserted * joins can get stacked onto either j->larg or j->rarg, @@ -384,6 +386,7 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, j->rarg = pull_up_sublinks_jointree_recurse(root, j->rarg, &child_rels); + /* * Now recursively process the pulled-up quals. Any inserted * joins can get stacked onto either j->larg or j->rarg, @@ -409,6 +412,7 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, j->rarg = pull_up_sublinks_jointree_recurse(root, j->rarg, &child_rels); + /* * Now recursively process the pulled-up quals. Any inserted * joins can get stacked onto either j->larg or j->rarg, @@ -439,7 +443,7 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, if (sublink->subLinkType == EXISTS_SUBLINK) { if ((j = convert_EXISTS_sublink_to_join(root, sublink, true, - available_rels1)) != NULL) + available_rels1)) != NULL) { /* Yes; insert the new join node into the join tree */ j->larg = *jtlink1; @@ -448,11 +452,12 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, j->rarg = pull_up_sublinks_jointree_recurse(root, j->rarg, &child_rels); + /* * Now recursively process the pulled-up quals. Because - * we are underneath a NOT, we can't pull up sublinks - * that reference the left-hand stuff, but it's still - * okay to pull up sublinks referencing j->rarg. + * we are underneath a NOT, we can't pull up sublinks that + * reference the left-hand stuff, but it's still okay to + * pull up sublinks referencing j->rarg. */ j->quals = pull_up_sublinks_qual_recurse(root, j->quals, @@ -464,7 +469,7 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, } if (available_rels2 != NULL && (j = convert_EXISTS_sublink_to_join(root, sublink, true, - available_rels2)) != NULL) + available_rels2)) != NULL) { /* Yes; insert the new join node into the join tree */ j->larg = *jtlink2; @@ -473,11 +478,12 @@ pull_up_sublinks_qual_recurse(PlannerInfo *root, Node *node, j->rarg = pull_up_sublinks_jointree_recurse(root, j->rarg, &child_rels); + /* * Now recursively process the pulled-up quals. Because - * we are underneath a NOT, we can't pull up sublinks - * that reference the left-hand stuff, but it's still - * okay to pull up sublinks referencing j->rarg. + * we are underneath a NOT, we can't pull up sublinks that + * reference the left-hand stuff, but it's still okay to + * pull up sublinks referencing j->rarg. */ j->quals = pull_up_sublinks_qual_recurse(root, j->quals, diff --git a/src/backend/optimizer/prep/prepunion.c b/src/backend/optimizer/prep/prepunion.c index 7b6d12de38a622135a8e24aedbb4271d5211bace..6475633ae7d06756ab9528a5b6323dd9cd3debb7 100644 --- a/src/backend/optimizer/prep/prepunion.c +++ b/src/backend/optimizer/prep/prepunion.c @@ -150,9 +150,9 @@ plan_set_operations(PlannerInfo *root, double tuple_fraction, Assert(parse->distinctClause == NIL); /* - * We'll need to build RelOptInfos for each of the leaf subqueries, - * which are RTE_SUBQUERY rangetable entries in this Query. Prepare the - * index arrays for that. + * We'll need to build RelOptInfos for each of the leaf subqueries, which + * are RTE_SUBQUERY rangetable entries in this Query. Prepare the index + * arrays for that. */ setup_simple_rel_arrays(root); diff --git a/src/backend/optimizer/util/clauses.c b/src/backend/optimizer/util/clauses.c index 344ebb79891e685bc589ee9bcd91ae0d0e153491..73f5e11abef320ef3480ffacef3102006b1ac81d 100644 --- a/src/backend/optimizer/util/clauses.c +++ b/src/backend/optimizer/util/clauses.c @@ -113,7 +113,7 @@ static Expr *simplify_function(Oid funcid, bool process_args, bool allow_non_const, eval_const_expressions_context *context); static List *expand_function_arguments(List *args, Oid result_type, - HeapTuple func_tuple); + HeapTuple func_tuple); static List *reorder_function_arguments(List *args, HeapTuple func_tuple); static List *add_function_defaults(List *args, HeapTuple func_tuple); static List *fetch_function_defaults(HeapTuple func_tuple); @@ -181,7 +181,7 @@ make_opclause(Oid opno, Oid opresulttype, bool opretset, Node * get_leftop(const Expr *clause) { - const OpExpr *expr = (const OpExpr *) clause; + const OpExpr *expr = (const OpExpr *) clause; if (expr->args != NIL) return linitial(expr->args); @@ -198,7 +198,7 @@ get_leftop(const Expr *clause) Node * get_rightop(const Expr *clause) { - const OpExpr *expr = (const OpExpr *) clause; + const OpExpr *expr = (const OpExpr *) clause; if (list_length(expr->args) >= 2) return lsecond(expr->args); @@ -1128,15 +1128,15 @@ contain_nonstrict_functions_walker(Node *node, void *context) } /***************************************************************************** - * Check clauses for non-leakproof functions + * Check clauses for non-leakproof functions *****************************************************************************/ /* * contain_leaky_functions - * Recursively search for leaky functions within a clause. + * Recursively search for leaky functions within a clause. * * Returns true if any function call with side-effect may be present in the - * clause. Qualifiers from outside the a security_barrier view should not + * clause. Qualifiers from outside the a security_barrier view should not * be pushed down into the view, lest the contents of tuples intended to be * filtered out be revealed via side effects. */ @@ -1155,8 +1155,8 @@ contain_leaky_functions_walker(Node *node, void *context) switch (nodeTag(node)) { case T_Var: - case T_Const: - case T_Param: + case T_Const: + case T_Param: case T_ArrayExpr: case T_NamedArgExpr: case T_BoolExpr: @@ -1168,6 +1168,7 @@ contain_leaky_functions_walker(Node *node, void *context) case T_NullTest: case T_BooleanTest: case T_List: + /* * We know these node types don't contain function calls; but * something further down in the node tree might. @@ -1176,7 +1177,7 @@ contain_leaky_functions_walker(Node *node, void *context) case T_FuncExpr: { - FuncExpr *expr = (FuncExpr *) node; + FuncExpr *expr = (FuncExpr *) node; if (!get_func_leakproof(expr->funcid)) return true; @@ -1187,7 +1188,7 @@ contain_leaky_functions_walker(Node *node, void *context) case T_DistinctExpr: /* struct-equivalent to OpExpr */ case T_NullIfExpr: /* struct-equivalent to OpExpr */ { - OpExpr *expr = (OpExpr *) node; + OpExpr *expr = (OpExpr *) node; set_opfuncid(expr); if (!get_func_leakproof(expr->opfuncid)) @@ -1208,11 +1209,11 @@ contain_leaky_functions_walker(Node *node, void *context) case T_CoerceViaIO: { CoerceViaIO *expr = (CoerceViaIO *) node; - Oid funcid; - Oid ioparam; - bool varlena; + Oid funcid; + Oid ioparam; + bool varlena; - getTypeInputInfo(exprType((Node *)expr->arg), + getTypeInputInfo(exprType((Node *) expr->arg), &funcid, &ioparam); if (!get_func_leakproof(funcid)) return true; @@ -1226,11 +1227,11 @@ contain_leaky_functions_walker(Node *node, void *context) case T_ArrayCoerceExpr: { ArrayCoerceExpr *expr = (ArrayCoerceExpr *) node; - Oid funcid; - Oid ioparam; - bool varlena; + Oid funcid; + Oid ioparam; + bool varlena; - getTypeInputInfo(exprType((Node *)expr->arg), + getTypeInputInfo(exprType((Node *) expr->arg), &funcid, &ioparam); if (!get_func_leakproof(funcid)) return true; @@ -1247,7 +1248,7 @@ contain_leaky_functions_walker(Node *node, void *context) foreach(opid, rcexpr->opnos) { - Oid funcid = get_opcode(lfirst_oid(opid)); + Oid funcid = get_opcode(lfirst_oid(opid)); if (!get_func_leakproof(funcid)) return true; @@ -1256,6 +1257,7 @@ contain_leaky_functions_walker(Node *node, void *context) break; default: + /* * If we don't recognize the node tag, assume it might be leaky. * This prevents an unexpected security hole if someone adds a new @@ -2683,7 +2685,7 @@ eval_const_expressions_mutator(Node *node, -1, InvalidOid, sizeof(Oid), - ObjectIdGetDatum(intypioparam), + ObjectIdGetDatum(intypioparam), false, true), makeConst(INT4OID, @@ -2812,13 +2814,13 @@ eval_const_expressions_mutator(Node *node, * TRUE: drop all remaining alternatives * If the first non-FALSE alternative is a constant TRUE, * we can simplify the entire CASE to that alternative's - * expression. If there are no non-FALSE alternatives, + * expression. If there are no non-FALSE alternatives, * we simplify the entire CASE to the default result (ELSE). * * If we have a simple-form CASE with constant test * expression, we substitute the constant value for contained * CaseTestExpr placeholder nodes, so that we have the - * opportunity to reduce constant test conditions. For + * opportunity to reduce constant test conditions. For * example this allows * CASE 0 WHEN 0 THEN 1 ELSE 1/0 END * to reduce to 1 rather than drawing a divide-by-0 error. @@ -3581,12 +3583,12 @@ simplify_function(Oid funcid, Oid result_type, int32 result_typmod, * deliver a constant result, use a transform function to generate a * substitute node tree, or expand in-line the body of the function * definition (which only works for simple SQL-language functions, but - * that is a common case). Each case needs access to the function's + * that is a common case). Each case needs access to the function's * pg_proc tuple, so fetch it just once. * * Note: the allow_non_const flag suppresses both the second and third - * strategies; so if !allow_non_const, simplify_function can only return - * a Const or NULL. Argument-list rewriting happens anyway, though. + * strategies; so if !allow_non_const, simplify_function can only return a + * Const or NULL. Argument-list rewriting happens anyway, though. */ func_tuple = SearchSysCache1(PROCOID, ObjectIdGetDatum(funcid)); if (!HeapTupleIsValid(func_tuple)) @@ -3603,7 +3605,7 @@ simplify_function(Oid funcid, Oid result_type, int32 result_typmod, { args = expand_function_arguments(args, result_type, func_tuple); args = (List *) expression_tree_mutator((Node *) args, - eval_const_expressions_mutator, + eval_const_expressions_mutator, (void *) context); /* Argument processing done, give it back to the caller */ *args_p = args; @@ -3618,7 +3620,7 @@ simplify_function(Oid funcid, Oid result_type, int32 result_typmod, if (!newexpr && allow_non_const && OidIsValid(func_form->protransform)) { /* - * Build a dummy FuncExpr node containing the simplified arg list. We + * Build a dummy FuncExpr node containing the simplified arg list. We * use this approach to present a uniform interface to the transform * function regardless of how the function is actually being invoked. */ diff --git a/src/backend/optimizer/util/pathnode.c b/src/backend/optimizer/util/pathnode.c index 61502aa64256d7fd4ab8a1e5d64606fef10a1769..00052f5c846ebe87ade0c19da26655b47c64cf73 100644 --- a/src/backend/optimizer/util/pathnode.c +++ b/src/backend/optimizer/util/pathnode.c @@ -128,11 +128,11 @@ compare_fractional_path_costs(Path *path1, Path *path2, * * The fuzz_factor argument must be 1.0 plus delta, where delta is the * fraction of the smaller cost that is considered to be a significant - * difference. For example, fuzz_factor = 1.01 makes the fuzziness limit + * difference. For example, fuzz_factor = 1.01 makes the fuzziness limit * be 1% of the smaller cost. * * The two paths are said to have "equal" costs if both startup and total - * costs are fuzzily the same. Path1 is said to be better than path2 if + * costs are fuzzily the same. Path1 is said to be better than path2 if * it has fuzzily better startup cost and fuzzily no worse total cost, * or if it has fuzzily better total cost and fuzzily no worse startup cost. * Path2 is better than path1 if the reverse holds. Finally, if one path @@ -190,9 +190,9 @@ compare_path_costs_fuzzily(Path *path1, Path *path2, double fuzz_factor) * and save them in the rel's cheapest-path fields. * * Only unparameterized paths are considered candidates for cheapest_startup - * and cheapest_total. The cheapest_parameterized_paths list collects paths + * and cheapest_total. The cheapest_parameterized_paths list collects paths * that are cheapest-total for their parameterization (i.e., there is no - * cheaper path with the same or weaker parameterization). This list always + * cheaper path with the same or weaker parameterization). This list always * includes the unparameterized cheapest-total path, too. * * This is normally called only after we've finished constructing the path @@ -294,8 +294,8 @@ set_cheapest(RelOptInfo *parent_rel) * * There is one policy decision embedded in this function, along with its * sibling add_path_precheck: we treat all parameterized paths as having - * NIL pathkeys, so that they compete only on cost. This is to reduce - * the number of parameterized paths that are kept. See discussion in + * NIL pathkeys, so that they compete only on cost. This is to reduce + * the number of parameterized paths that are kept. See discussion in * src/backend/optimizer/README. * * The pathlist is kept sorted by total_cost, with cheaper paths @@ -358,7 +358,7 @@ add_path(RelOptInfo *parent_rel, Path *new_path) p1_next = lnext(p1); /* - * Do a fuzzy cost comparison with 1% fuzziness limit. (XXX does this + * Do a fuzzy cost comparison with 1% fuzziness limit. (XXX does this * percentage need to be user-configurable?) */ costcmp = compare_path_costs_fuzzily(new_path, old_path, 1.01); @@ -388,20 +388,20 @@ add_path(RelOptInfo *parent_rel, Path *new_path) { case COSTS_EQUAL: outercmp = bms_subset_compare(PATH_REQ_OUTER(new_path), - PATH_REQ_OUTER(old_path)); + PATH_REQ_OUTER(old_path)); if (keyscmp == PATHKEYS_BETTER1) { if ((outercmp == BMS_EQUAL || outercmp == BMS_SUBSET1) && new_path->rows <= old_path->rows) - remove_old = true; /* new dominates old */ + remove_old = true; /* new dominates old */ } else if (keyscmp == PATHKEYS_BETTER2) { if ((outercmp == BMS_EQUAL || outercmp == BMS_SUBSET2) && new_path->rows >= old_path->rows) - accept_new = false; /* old dominates new */ + accept_new = false; /* old dominates new */ } else /* keyscmp == PATHKEYS_EQUAL */ { @@ -425,19 +425,20 @@ add_path(RelOptInfo *parent_rel, Path *new_path) if (new_path->rows < old_path->rows) remove_old = true; /* new dominates old */ else if (new_path->rows > old_path->rows) - accept_new = false; /* old dominates new */ + accept_new = false; /* old dominates new */ else if (compare_path_costs_fuzzily(new_path, old_path, - 1.0000000001) == COSTS_BETTER1) + 1.0000000001) == COSTS_BETTER1) remove_old = true; /* new dominates old */ else - accept_new = false; /* old equals or dominates new */ + accept_new = false; /* old equals or + * dominates new */ } else if (outercmp == BMS_SUBSET1 && new_path->rows <= old_path->rows) - remove_old = true; /* new dominates old */ + remove_old = true; /* new dominates old */ else if (outercmp == BMS_SUBSET2 && new_path->rows >= old_path->rows) - accept_new = false; /* old dominates new */ + accept_new = false; /* old dominates new */ /* else different parameterizations, keep both */ } break; @@ -445,25 +446,26 @@ add_path(RelOptInfo *parent_rel, Path *new_path) if (keyscmp != PATHKEYS_BETTER2) { outercmp = bms_subset_compare(PATH_REQ_OUTER(new_path), - PATH_REQ_OUTER(old_path)); + PATH_REQ_OUTER(old_path)); if ((outercmp == BMS_EQUAL || outercmp == BMS_SUBSET1) && new_path->rows <= old_path->rows) - remove_old = true; /* new dominates old */ + remove_old = true; /* new dominates old */ } break; case COSTS_BETTER2: if (keyscmp != PATHKEYS_BETTER1) { outercmp = bms_subset_compare(PATH_REQ_OUTER(new_path), - PATH_REQ_OUTER(old_path)); + PATH_REQ_OUTER(old_path)); if ((outercmp == BMS_EQUAL || outercmp == BMS_SUBSET2) && new_path->rows >= old_path->rows) - accept_new = false; /* old dominates new */ + accept_new = false; /* old dominates new */ } break; case COSTS_DIFFERENT: + /* * can't get here, but keep this case to keep compiler * quiet @@ -529,7 +531,7 @@ add_path(RelOptInfo *parent_rel, Path *new_path) * and have lower bounds for its costs. * * Note that we do not know the path's rowcount, since getting an estimate for - * that is too expensive to do before prechecking. We assume here that paths + * that is too expensive to do before prechecking. We assume here that paths * of a superset parameterization will generate fewer rows; if that holds, * then paths with different parameterizations cannot dominate each other * and so we can simply ignore existing paths of another parameterization. @@ -561,9 +563,9 @@ add_path_precheck(RelOptInfo *parent_rel, * pathkeys as well as both cost metrics. If we find one, we can * reject the new path. * - * For speed, we make exact rather than fuzzy cost comparisons. - * If an old path dominates the new path exactly on both costs, it - * will surely do so fuzzily. + * For speed, we make exact rather than fuzzy cost comparisons. If an + * old path dominates the new path exactly on both costs, it will + * surely do so fuzzily. */ if (total_cost >= old_path->total_cost) { @@ -588,9 +590,9 @@ add_path_precheck(RelOptInfo *parent_rel, else { /* - * Since the pathlist is sorted by total_cost, we can stop - * looking once we reach a path with a total_cost larger - * than the new path's. + * Since the pathlist is sorted by total_cost, we can stop looking + * once we reach a path with a total_cost larger than the new + * path's. */ break; } @@ -652,26 +654,26 @@ add_parameterized_path(RelOptInfo *parent_rel, Path *new_path) { if (outercmp != BMS_SUBSET2 && new_path->rows <= old_path->rows) - remove_old = true; /* new dominates old */ + remove_old = true; /* new dominates old */ } else if (costcmp > 0) { if (outercmp != BMS_SUBSET1 && new_path->rows >= old_path->rows) - accept_new = false; /* old dominates new */ + accept_new = false; /* old dominates new */ } else if (outercmp == BMS_SUBSET1 && new_path->rows <= old_path->rows) - remove_old = true; /* new dominates old */ + remove_old = true; /* new dominates old */ else if (outercmp == BMS_SUBSET2 && new_path->rows >= old_path->rows) - accept_new = false; /* old dominates new */ + accept_new = false; /* old dominates new */ else if (new_path->rows < old_path->rows) - remove_old = true; /* new dominates old */ + remove_old = true; /* new dominates old */ else { /* Same cost, rows, and param rels; arbitrarily keep old */ - accept_new = false; /* old equals or dominates new */ + accept_new = false; /* old equals or dominates new */ } } @@ -697,8 +699,8 @@ add_parameterized_path(RelOptInfo *parent_rel, Path *new_path) /* * If we found an old path that dominates new_path, we can quit - * scanning the list; we will not add new_path, and we assume - * new_path cannot dominate any other elements of the list. + * scanning the list; we will not add new_path, and we assume new_path + * cannot dominate any other elements of the list. */ if (!accept_new) break; @@ -940,7 +942,7 @@ create_append_path(RelOptInfo *rel, List *subpaths, Relids required_outer) * Compute rows and costs as sums of subplan rows and costs. We charge * nothing extra for the Append itself, which perhaps is too optimistic, * but since it doesn't do any selection or projection, it is a pretty - * cheap node. If you change this, see also make_append(). + * cheap node. If you change this, see also make_append(). */ pathnode->path.rows = 0; pathnode->path.startup_cost = 0; @@ -1772,9 +1774,9 @@ create_foreignscan_path(PlannerInfo *root, RelOptInfo *rel, Relids calc_nestloop_required_outer(Path *outer_path, Path *inner_path) { - Relids outer_paramrels = PATH_REQ_OUTER(outer_path); - Relids inner_paramrels = PATH_REQ_OUTER(inner_path); - Relids required_outer; + Relids outer_paramrels = PATH_REQ_OUTER(outer_path); + Relids inner_paramrels = PATH_REQ_OUTER(inner_path); + Relids required_outer; /* inner_path can require rels from outer path, but not vice versa */ Assert(!bms_overlap(outer_paramrels, inner_path->parent->relids)); @@ -1804,9 +1806,9 @@ calc_nestloop_required_outer(Path *outer_path, Path *inner_path) Relids calc_non_nestloop_required_outer(Path *outer_path, Path *inner_path) { - Relids outer_paramrels = PATH_REQ_OUTER(outer_path); - Relids inner_paramrels = PATH_REQ_OUTER(inner_path); - Relids required_outer; + Relids outer_paramrels = PATH_REQ_OUTER(outer_path); + Relids inner_paramrels = PATH_REQ_OUTER(inner_path); + Relids required_outer; /* neither path can require rels from the other */ Assert(!bms_overlap(outer_paramrels, inner_path->parent->relids)); @@ -1853,9 +1855,9 @@ create_nestloop_path(PlannerInfo *root, /* * If the inner path is parameterized by the outer, we must drop any - * restrict_clauses that are due to be moved into the inner path. We - * have to do this now, rather than postpone the work till createplan - * time, because the restrict_clauses list can affect the size and cost + * restrict_clauses that are due to be moved into the inner path. We have + * to do this now, rather than postpone the work till createplan time, + * because the restrict_clauses list can affect the size and cost * estimates for this path. */ if (bms_overlap(inner_req_outer, outer_path->parent->relids)) @@ -2033,7 +2035,7 @@ create_hashjoin_path(PlannerInfo *root, * same parameterization level, ensuring that they all enforce the same set * of join quals (and thus that that parameterization can be attributed to * an append path built from such paths). Currently, only a few path types - * are supported here, though more could be added at need. We return NULL + * are supported here, though more could be added at need. We return NULL * if we can't reparameterize the given path. * * Note: we intentionally do not pass created paths to add_path(); it would @@ -2058,32 +2060,33 @@ reparameterize_path(PlannerInfo *root, Path *path, return create_seqscan_path(root, rel, required_outer); case T_IndexScan: case T_IndexOnlyScan: - { - IndexPath *ipath = (IndexPath *) path; - IndexPath *newpath = makeNode(IndexPath); + { + IndexPath *ipath = (IndexPath *) path; + IndexPath *newpath = makeNode(IndexPath); - /* - * We can't use create_index_path directly, and would not want to - * because it would re-compute the indexqual conditions which is - * wasted effort. Instead we hack things a bit: flat-copy the - * path node, revise its param_info, and redo the cost estimate. - */ - memcpy(newpath, ipath, sizeof(IndexPath)); - newpath->path.param_info = - get_baserel_parampathinfo(root, rel, required_outer); - cost_index(newpath, root, loop_count); - return (Path *) newpath; - } + /* + * We can't use create_index_path directly, and would not want + * to because it would re-compute the indexqual conditions + * which is wasted effort. Instead we hack things a bit: + * flat-copy the path node, revise its param_info, and redo + * the cost estimate. + */ + memcpy(newpath, ipath, sizeof(IndexPath)); + newpath->path.param_info = + get_baserel_parampathinfo(root, rel, required_outer); + cost_index(newpath, root, loop_count); + return (Path *) newpath; + } case T_BitmapHeapScan: - { - BitmapHeapPath *bpath = (BitmapHeapPath *) path; + { + BitmapHeapPath *bpath = (BitmapHeapPath *) path; - return (Path *) create_bitmap_heap_path(root, - rel, - bpath->bitmapqual, - required_outer, - loop_count); - } + return (Path *) create_bitmap_heap_path(root, + rel, + bpath->bitmapqual, + required_outer, + loop_count); + } case T_SubqueryScan: return create_subqueryscan_path(root, rel, path->pathkeys, required_outer); diff --git a/src/backend/optimizer/util/placeholder.c b/src/backend/optimizer/util/placeholder.c index 93f1c2cdfa4e1aadec221fa7ef2ccc833709daf8..e05c8ddef1b8c3582992111c10750b2d98b1d038 100644 --- a/src/backend/optimizer/util/placeholder.c +++ b/src/backend/optimizer/util/placeholder.c @@ -61,7 +61,7 @@ make_placeholder_expr(PlannerInfo *root, Expr *expr, Relids phrels) * We build PlaceHolderInfos only for PHVs that are still present in the * simplified query passed to query_planner(). * - * Note: this should only be called after query_planner() has started. Also, + * Note: this should only be called after query_planner() has started. Also, * create_new_ph must not be TRUE after deconstruct_jointree begins, because * make_outerjoininfo assumes that we already know about all placeholders. */ @@ -259,7 +259,7 @@ mark_placeholder_maybe_needed(PlannerInfo *root, PlaceHolderInfo *phinfo, * but they aren't going to be needed where the outer PHV is referenced. * Rather, they'll be needed where the outer PHV is evaluated. We can * estimate that (conservatively) as the syntactic location of the PHV's - * expression. Recurse to take care of any such PHVs. + * expression. Recurse to take care of any such PHVs. */ mark_placeholders_in_expr(root, (Node *) phinfo->ph_var->phexpr, phinfo->ph_var->phrels); diff --git a/src/backend/optimizer/util/plancat.c b/src/backend/optimizer/util/plancat.c index aaf288a50e9e0db084e2d600236b891d88837169..38b81a05ff712d527370af834fbbc8298d980f9e 100644 --- a/src/backend/optimizer/util/plancat.c +++ b/src/backend/optimizer/util/plancat.c @@ -341,7 +341,7 @@ get_relation_info(PlannerInfo *root, Oid relationObjectId, bool inhparent, } else { - double allvisfrac; /* dummy */ + double allvisfrac; /* dummy */ estimate_rel_size(indexRelation, NULL, &info->pages, &info->tuples, &allvisfrac); @@ -403,12 +403,12 @@ estimate_rel_size(Relation rel, int32 *attr_widths, * minimum size estimate of 10 pages. The idea here is to avoid * assuming a newly-created table is really small, even if it * currently is, because that may not be true once some data gets - * loaded into it. Once a vacuum or analyze cycle has been done + * loaded into it. Once a vacuum or analyze cycle has been done * on it, it's more reasonable to believe the size is somewhat * stable. * * (Note that this is only an issue if the plan gets cached and - * used again after the table has been filled. What we're trying + * used again after the table has been filled. What we're trying * to avoid is using a nestloop-type plan on a table that has * grown substantially since the plan was made. Normally, * autovacuum/autoanalyze will occur once enough inserts have @@ -965,7 +965,7 @@ build_index_tlist(PlannerInfo *root, IndexOptInfo *index, if (indexkey < 0) att_tup = SystemAttributeDefinition(indexkey, - heapRelation->rd_rel->relhasoids); + heapRelation->rd_rel->relhasoids); else att_tup = heapRelation->rd_att->attrs[indexkey - 1]; diff --git a/src/backend/optimizer/util/predtest.c b/src/backend/optimizer/util/predtest.c index c3161c5293b5411de92f002dfcdaf6794b45f777..65d191e5d3c6332da193aeccc147b69938949c9c 100644 --- a/src/backend/optimizer/util/predtest.c +++ b/src/backend/optimizer/util/predtest.c @@ -1624,7 +1624,7 @@ get_btree_test_op(Oid pred_op, Oid clause_op, bool refute_it) clause_op_infos = get_op_btree_interpretation(clause_op); if (clause_op_infos) pred_op_infos = get_op_btree_interpretation(pred_op); - else /* no point in looking */ + else /* no point in looking */ pred_op_infos = NIL; foreach(lcp, pred_op_infos) diff --git a/src/backend/optimizer/util/relnode.c b/src/backend/optimizer/util/relnode.c index bfdd9ff222cdf50a7561a39898b29e1685ad4731..8d4ab03d20eb79de5f6b0d48c4793fa33c84be4c 100644 --- a/src/backend/optimizer/util/relnode.c +++ b/src/backend/optimizer/util/relnode.c @@ -840,12 +840,12 @@ get_joinrel_parampathinfo(PlannerInfo *root, RelOptInfo *joinrel, outer_and_req = bms_union(outer_path->parent->relids, PATH_REQ_OUTER(outer_path)); else - outer_and_req = NULL; /* outer path does not accept parameters */ + outer_and_req = NULL; /* outer path does not accept parameters */ if (inner_path->param_info) inner_and_req = bms_union(inner_path->parent->relids, PATH_REQ_OUTER(inner_path)); else - inner_and_req = NULL; /* inner path does not accept parameters */ + inner_and_req = NULL; /* inner path does not accept parameters */ pclauses = NIL; foreach(lc, joinrel->joininfo) @@ -909,7 +909,7 @@ get_joinrel_parampathinfo(PlannerInfo *root, RelOptInfo *joinrel, *restrict_clauses); /* - * And now we can build the ParamPathInfo. No point in saving the + * And now we can build the ParamPathInfo. No point in saving the * input-pair-dependent clause list, though. * * Note: in GEQO mode, we'll be called in a temporary memory context, but @@ -929,8 +929,8 @@ get_joinrel_parampathinfo(PlannerInfo *root, RelOptInfo *joinrel, * Get the ParamPathInfo for a parameterized path for an append relation. * * For an append relation, the rowcount estimate will just be the sum of - * the estimates for its children. However, we still need a ParamPathInfo - * to flag the fact that the path requires parameters. So this just creates + * the estimates for its children. However, we still need a ParamPathInfo + * to flag the fact that the path requires parameters. So this just creates * a suitable struct with zero ppi_rows (and no ppi_clauses either, since * the Append node isn't responsible for checking quals). */ diff --git a/src/backend/optimizer/util/var.c b/src/backend/optimizer/util/var.c index 2bffb0a651efed0f13695bb40d95243b31349ba4..9bc90c253139deb43415683800681f48fe0c8436 100644 --- a/src/backend/optimizer/util/var.c +++ b/src/backend/optimizer/util/var.c @@ -218,7 +218,7 @@ pull_varattnos_walker(Node *node, pull_varattnos_context *context) if (var->varno == context->varno && var->varlevelsup == 0) context->varattnos = bms_add_member(context->varattnos, - var->varattno - FirstLowInvalidHeapAttributeNumber); + var->varattno - FirstLowInvalidHeapAttributeNumber); return false; } diff --git a/src/backend/parser/analyze.c b/src/backend/parser/analyze.c index 15d848ff4fe262fe9cd2c1f113af4dd9e2e18da3..bfd3ab941a771ac8a994afc72967fb61c3321120 100644 --- a/src/backend/parser/analyze.c +++ b/src/backend/parser/analyze.c @@ -181,7 +181,7 @@ transformTopLevelStmt(ParseState *pstate, Node *parseTree) /* If it's a set-operation tree, drill down to leftmost SelectStmt */ while (stmt && stmt->op != SETOP_NONE) stmt = stmt->larg; - Assert(stmt && IsA(stmt, SelectStmt) && stmt->larg == NULL); + Assert(stmt && IsA(stmt, SelectStmt) &&stmt->larg == NULL); if (stmt->intoClause) { @@ -950,7 +950,7 @@ transformSelectStmt(ParseState *pstate, SelectStmt *stmt) (errcode(ERRCODE_SYNTAX_ERROR), errmsg("SELECT ... INTO is not allowed here"), parser_errposition(pstate, - exprLocation((Node *) stmt->intoClause)))); + exprLocation((Node *) stmt->intoClause)))); /* make FOR UPDATE/FOR SHARE info available to addRangeTableEntry */ pstate->p_locking_clause = stmt->lockingClause; @@ -1364,12 +1364,12 @@ transformSetOperationStmt(ParseState *pstate, SelectStmt *stmt) (errcode(ERRCODE_SYNTAX_ERROR), errmsg("SELECT ... INTO is not allowed here"), parser_errposition(pstate, - exprLocation((Node *) leftmostSelect->intoClause)))); + exprLocation((Node *) leftmostSelect->intoClause)))); /* - * We need to extract ORDER BY and other top-level clauses here and - * not let transformSetOperationTree() see them --- else it'll just - * recurse right back here! + * We need to extract ORDER BY and other top-level clauses here and not + * let transformSetOperationTree() see them --- else it'll just recurse + * right back here! */ sortClause = stmt->sortClause; limitOffset = stmt->limitOffset; diff --git a/src/backend/parser/parse_coerce.c b/src/backend/parser/parse_coerce.c index 3102f2089ec9bc017446b44d923ffecaf935fc3f..a1caf74fc53406baeb0fa996b02f6485c8d691ed 100644 --- a/src/backend/parser/parse_coerce.c +++ b/src/backend/parser/parse_coerce.c @@ -1302,26 +1302,26 @@ coerce_to_common_type(ParseState *pstate, Node *node, * * 1) All arguments declared ANYELEMENT must have the same datatype. * 2) All arguments declared ANYARRAY must have the same datatype, - * which must be a varlena array type. + * which must be a varlena array type. * 3) All arguments declared ANYRANGE must have the same datatype, - * which must be a range type. + * which must be a range type. * 4) If there are arguments of both ANYELEMENT and ANYARRAY, make sure the - * actual ANYELEMENT datatype is in fact the element type for the actual - * ANYARRAY datatype. + * actual ANYELEMENT datatype is in fact the element type for the actual + * ANYARRAY datatype. * 5) Similarly, if there are arguments of both ANYELEMENT and ANYRANGE, - * make sure the actual ANYELEMENT datatype is in fact the subtype for - * the actual ANYRANGE type. + * make sure the actual ANYELEMENT datatype is in fact the subtype for + * the actual ANYRANGE type. * 6) ANYENUM is treated the same as ANYELEMENT except that if it is used - * (alone or in combination with plain ANYELEMENT), we add the extra - * condition that the ANYELEMENT type must be an enum. + * (alone or in combination with plain ANYELEMENT), we add the extra + * condition that the ANYELEMENT type must be an enum. * 7) ANYNONARRAY is treated the same as ANYELEMENT except that if it is used, - * we add the extra condition that the ANYELEMENT type must not be an array. - * (This is a no-op if used in combination with ANYARRAY or ANYENUM, but - * is an extra restriction if not.) + * we add the extra condition that the ANYELEMENT type must not be an array. + * (This is a no-op if used in combination with ANYARRAY or ANYENUM, but + * is an extra restriction if not.) * * Domains over arrays match ANYARRAY, and are immediately flattened to their * base type. (Thus, for example, we will consider it a match if one ANYARRAY - * argument is a domain over int4[] while another one is just int4[].) Also + * argument is a domain over int4[] while another one is just int4[].) Also * notice that such a domain does *not* match ANYNONARRAY. * * Similarly, domains over ranges match ANYRANGE, and are immediately @@ -1475,7 +1475,7 @@ check_generic_type_consistency(Oid *actual_arg_types, * * If any polymorphic pseudotype is used in a function's arguments or * return type, we make sure the actual data types are consistent with - * each other. The argument consistency rules are shown above for + * each other. The argument consistency rules are shown above for * check_generic_type_consistency(). * * If we have UNKNOWN input (ie, an untyped literal) for any polymorphic @@ -1487,35 +1487,35 @@ check_generic_type_consistency(Oid *actual_arg_types, * if it is declared as a polymorphic type: * * 1) If return type is ANYARRAY, and any argument is ANYARRAY, use the - * argument's actual type as the function's return type. + * argument's actual type as the function's return type. * 2) Similarly, if return type is ANYRANGE, and any argument is ANYRANGE, - * use the argument's actual type as the function's return type. + * use the argument's actual type as the function's return type. * 3) If return type is ANYARRAY, no argument is ANYARRAY, but any argument is - * ANYELEMENT, use the actual type of the argument to determine the - * function's return type, i.e. the element type's corresponding array - * type. (Note: similar behavior does not exist for ANYRANGE, because it's - * impossible to determine the range type from the subtype alone.) + * ANYELEMENT, use the actual type of the argument to determine the + * function's return type, i.e. the element type's corresponding array + * type. (Note: similar behavior does not exist for ANYRANGE, because it's + * impossible to determine the range type from the subtype alone.) * 4) If return type is ANYARRAY, but no argument is ANYARRAY or ANYELEMENT, - * generate an error. Similarly, if return type is ANYRANGE, but no - * argument is ANYRANGE, generate an error. (These conditions are - * prevented by CREATE FUNCTION and therefore are not expected here.) + * generate an error. Similarly, if return type is ANYRANGE, but no + * argument is ANYRANGE, generate an error. (These conditions are + * prevented by CREATE FUNCTION and therefore are not expected here.) * 5) If return type is ANYELEMENT, and any argument is ANYELEMENT, use the - * argument's actual type as the function's return type. + * argument's actual type as the function's return type. * 6) If return type is ANYELEMENT, no argument is ANYELEMENT, but any argument - * is ANYARRAY or ANYRANGE, use the actual type of the argument to determine - * the function's return type, i.e. the array type's corresponding element - * type or the range type's corresponding subtype (or both, in which case - * they must match). + * is ANYARRAY or ANYRANGE, use the actual type of the argument to determine + * the function's return type, i.e. the array type's corresponding element + * type or the range type's corresponding subtype (or both, in which case + * they must match). * 7) If return type is ANYELEMENT, no argument is ANYELEMENT, ANYARRAY, or - * ANYRANGE, generate an error. (This condition is prevented by CREATE - * FUNCTION and therefore is not expected here.) + * ANYRANGE, generate an error. (This condition is prevented by CREATE + * FUNCTION and therefore is not expected here.) * 8) ANYENUM is treated the same as ANYELEMENT except that if it is used - * (alone or in combination with plain ANYELEMENT), we add the extra - * condition that the ANYELEMENT type must be an enum. + * (alone or in combination with plain ANYELEMENT), we add the extra + * condition that the ANYELEMENT type must be an enum. * 9) ANYNONARRAY is treated the same as ANYELEMENT except that if it is used, - * we add the extra condition that the ANYELEMENT type must not be an array. - * (This is a no-op if used in combination with ANYARRAY or ANYENUM, but - * is an extra restriction if not.) + * we add the extra condition that the ANYELEMENT type must not be an array. + * (This is a no-op if used in combination with ANYARRAY or ANYENUM, but + * is an extra restriction if not.) * * Domains over arrays or ranges match ANYARRAY or ANYRANGE arguments, * respectively, and are immediately flattened to their base type. (In @@ -1524,14 +1524,14 @@ check_generic_type_consistency(Oid *actual_arg_types, * * When allow_poly is false, we are not expecting any of the actual_arg_types * to be polymorphic, and we should not return a polymorphic result type - * either. When allow_poly is true, it is okay to have polymorphic "actual" + * either. When allow_poly is true, it is okay to have polymorphic "actual" * arg types, and we can return ANYARRAY, ANYRANGE, or ANYELEMENT as the - * result. (This case is currently used only to check compatibility of an + * result. (This case is currently used only to check compatibility of an * aggregate's declaration with the underlying transfn.) * * A special case is that we could see ANYARRAY as an actual_arg_type even * when allow_poly is false (this is possible only because pg_statistic has - * columns shown as anyarray in the catalogs). We allow this to match a + * columns shown as anyarray in the catalogs). We allow this to match a * declared ANYARRAY argument, but only if there is no ANYELEMENT argument * or result (since we can't determine a specific element type to match to * ANYELEMENT). Note this means that functions taking ANYARRAY had better diff --git a/src/backend/parser/parse_expr.c b/src/backend/parser/parse_expr.c index 973265bcb0b1af7b3411b3f1a54aee27d790297c..bb1ad9af96bb5df02b9bf2a06222dd7dca8ea926 100644 --- a/src/backend/parser/parse_expr.c +++ b/src/backend/parser/parse_expr.c @@ -844,7 +844,7 @@ transformAExprOp(ParseState *pstate, A_Expr *a) list_length(a->name) == 1 && strcmp(strVal(linitial(a->name)), "=") == 0 && (exprIsNullConstant(lexpr) || exprIsNullConstant(rexpr)) && - (!IsA(lexpr, CaseTestExpr) && !IsA(rexpr, CaseTestExpr))) + (!IsA(lexpr, CaseTestExpr) &&!IsA(rexpr, CaseTestExpr))) { NullTest *n = makeNode(NullTest); @@ -2066,9 +2066,9 @@ transformWholeRowRef(ParseState *pstate, RangeTblEntry *rte, int location) vnum = RTERangeTablePosn(pstate, rte, &sublevels_up); /* - * Build the appropriate referencing node. Note that if the RTE is a + * Build the appropriate referencing node. Note that if the RTE is a * function returning scalar, we create just a plain reference to the - * function value, not a composite containing a single column. This is + * function value, not a composite containing a single column. This is * pretty inconsistent at first sight, but it's what we've done * historically. One argument for it is that "rel" and "rel.*" mean the * same thing for composite relations, so why not for scalar functions... @@ -2268,8 +2268,8 @@ make_row_comparison_op(ParseState *pstate, List *opname, opinfo_lists[i] = get_op_btree_interpretation(opno); /* - * convert strategy numbers into a Bitmapset to make the - * intersection calculation easy. + * convert strategy numbers into a Bitmapset to make the intersection + * calculation easy. */ this_strats = NULL; foreach(j, opinfo_lists[i]) diff --git a/src/backend/parser/parse_func.c b/src/backend/parser/parse_func.c index e583fae84997e45bd3c0b28caf1f580fad2935b5..b051707d7e1b0ccd98542c3d1fb1e08fc29ff128 100644 --- a/src/backend/parser/parse_func.c +++ b/src/backend/parser/parse_func.c @@ -787,9 +787,9 @@ func_select_candidate(int nargs, * Having completed this examination, remove candidates that accept the * wrong category at any unknown position. Also, if at least one * candidate accepted a preferred type at a position, remove candidates - * that accept non-preferred types. If just one candidate remains, - * return that one. However, if this rule turns out to reject all - * candidates, keep them all instead. + * that accept non-preferred types. If just one candidate remains, return + * that one. However, if this rule turns out to reject all candidates, + * keep them all instead. */ resolved_unknowns = false; for (i = 0; i < nargs; i++) @@ -914,7 +914,7 @@ func_select_candidate(int nargs, * type, and see if that gives us a unique match. If so, use that match. * * NOTE: for a binary operator with one unknown and one non-unknown input, - * we already tried this heuristic in binary_oper_exact(). However, that + * we already tried this heuristic in binary_oper_exact(). However, that * code only finds exact matches, whereas here we will handle matches that * involve coercion, polymorphic type resolution, etc. */ diff --git a/src/backend/parser/parse_relation.c b/src/backend/parser/parse_relation.c index 2a26b0af1d8bd80b91149b1fde0e827affeebe2a..30b307b191cc4e5304dbc99e0aea95046720bd46 100644 --- a/src/backend/parser/parse_relation.c +++ b/src/backend/parser/parse_relation.c @@ -274,10 +274,11 @@ searchRangeTable(ParseState *pstate, RangeVar *relation) * relation. * * NB: It's not critical that RangeVarGetRelid return the correct answer - * here in the face of concurrent DDL. If it doesn't, the worst case - * scenario is a less-clear error message. Also, the tables involved in + * here in the face of concurrent DDL. If it doesn't, the worst case + * scenario is a less-clear error message. Also, the tables involved in * the query are already locked, which reduces the number of cases in - * which surprising behavior can occur. So we do the name lookup unlocked. + * which surprising behavior can occur. So we do the name lookup + * unlocked. */ if (!relation->schemaname) cte = scanNameSpaceForCTE(pstate, refname, &ctelevelsup); diff --git a/src/backend/parser/parse_target.c b/src/backend/parser/parse_target.c index a512d18d01ef1dbee43e6f536f970e2350273a25..3850a3bc646f3e32dc293240a78ef8ed72aa8500 100644 --- a/src/backend/parser/parse_target.c +++ b/src/backend/parser/parse_target.c @@ -1622,7 +1622,7 @@ FigureColnameInternal(Node *node, char **name) case EXPR_SUBLINK: { /* Get column name of the subquery's single target */ - SubLink *sublink = (SubLink *) node; + SubLink *sublink = (SubLink *) node; Query *query = (Query *) sublink->subselect; /* @@ -1644,7 +1644,7 @@ FigureColnameInternal(Node *node, char **name) } } break; - /* As with other operator-like nodes, these have no names */ + /* As with other operator-like nodes, these have no names */ case ALL_SUBLINK: case ANY_SUBLINK: case ROWCOMPARE_SUBLINK: diff --git a/src/backend/parser/parse_type.c b/src/backend/parser/parse_type.c index 531495433dee4137532e086cb9e83cb664e9fffd..871a7d1ce303267154731911ea2d6bbb0a2f9db8 100644 --- a/src/backend/parser/parse_type.c +++ b/src/backend/parser/parse_type.c @@ -111,8 +111,8 @@ LookupTypeName(ParseState *pstate, const TypeName *typeName, /* * Look up the field. * - * XXX: As no lock is taken here, this might fail in the presence - * of concurrent DDL. But taking a lock would carry a performance + * XXX: As no lock is taken here, this might fail in the presence of + * concurrent DDL. But taking a lock would carry a performance * penalty and would also require a permissions check. */ relid = RangeVarGetRelid(rel, NoLock, false); diff --git a/src/backend/parser/parse_utilcmd.c b/src/backend/parser/parse_utilcmd.c index 7c315f6c8718d71edbe1b1665cb20e9e0b4be7f1..8810d0dbad5d942855449ea2da8b47a4b317692e 100644 --- a/src/backend/parser/parse_utilcmd.c +++ b/src/backend/parser/parse_utilcmd.c @@ -103,7 +103,7 @@ static void transformColumnDefinition(CreateStmtContext *cxt, static void transformTableConstraint(CreateStmtContext *cxt, Constraint *constraint); static void transformTableLikeClause(CreateStmtContext *cxt, - TableLikeClause *table_like_clause); + TableLikeClause *table_like_clause); static void transformOfType(CreateStmtContext *cxt, TypeName *ofTypename); static char *chooseIndexName(const RangeVar *relation, IndexStmt *index_stmt); @@ -309,7 +309,7 @@ transformColumnDefinition(CreateStmtContext *cxt, ColumnDef *column) column->typeName->typeOid = INT2OID; } else if (strcmp(typname, "serial") == 0 || - strcmp(typname, "serial4") == 0) + strcmp(typname, "serial4") == 0) { is_serial = true; column->typeName->names = NIL; @@ -554,13 +554,13 @@ transformColumnDefinition(CreateStmtContext *cxt, ColumnDef *column) } /* - * Generate ALTER FOREIGN TABLE ALTER COLUMN statement which adds + * Generate ALTER FOREIGN TABLE ALTER COLUMN statement which adds * per-column foreign data wrapper options for this column. */ if (column->fdwoptions != NIL) { AlterTableStmt *stmt; - AlterTableCmd *cmd; + AlterTableCmd *cmd; cmd = makeNode(AlterTableCmd); cmd->subtype = AT_AlterColumnGenericOptions; @@ -667,7 +667,7 @@ transformTableLikeClause(CreateStmtContext *cxt, TableLikeClause *table_like_cla else { aclresult = pg_class_aclcheck(RelationGetRelid(relation), GetUserId(), - ACL_SELECT); + ACL_SELECT); if (aclresult != ACLCHECK_OK) aclcheck_error(aclresult, ACL_KIND_CLASS, RelationGetRelationName(relation)); @@ -803,7 +803,7 @@ transformTableLikeClause(CreateStmtContext *cxt, TableLikeClause *table_like_cla /* Copy comment on constraint */ if ((table_like_clause->options & CREATE_TABLE_LIKE_COMMENTS) && (comment = GetComment(get_relation_constraint_oid(RelationGetRelid(relation), - n->conname, false), + n->conname, false), ConstraintRelationId, 0)) != NULL) { @@ -2305,7 +2305,7 @@ transformAlterTableStmt(AlterTableStmt *stmt, const char *queryString) /* this message is consistent with relation_openrv */ ereport(NOTICE, (errmsg("relation \"%s\" does not exist, skipping", - stmt->relation->relname))); + stmt->relation->relname))); return NIL; } diff --git a/src/backend/port/darwin/system.c b/src/backend/port/darwin/system.c index e9cd136bec222201c2f4fdf3df90c561ecda3dea..d571f26ef8c94d9d9a37928424d9f7d2eaacc918 100644 --- a/src/backend/port/darwin/system.c +++ b/src/backend/port/darwin/system.c @@ -2,7 +2,7 @@ * src/backend/port/darwin/system.c * * only needed in OS X 10.1 and possibly early 10.2 releases */ -#include <AvailabilityMacros.h> /* pgrminclude ignore */ +#include <AvailabilityMacros.h> /* pgrminclude ignore */ #if MAC_OS_X_VERSION_MAX_ALLOWED <= MAC_OS_X_VERSION_10_2 || !defined(MAC_OS_X_VERSION_10_2) /* diff --git a/src/backend/port/dynloader/aix.h b/src/backend/port/dynloader/aix.h index e5afaead8e20875fdb20efbf063ec90e9f9991c7..ba5590f598a7672dc9ff4061bf3017024cfcf9e6 100644 --- a/src/backend/port/dynloader/aix.h +++ b/src/backend/port/dynloader/aix.h @@ -16,7 +16,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * In some older systems, the RTLD_NOW flag isn't defined and the mode diff --git a/src/backend/port/dynloader/cygwin.h b/src/backend/port/dynloader/cygwin.h index 81ef27f17c6178a98e68aeaba662d7b096156114..06f75d4c42b71bc193a93d131a07f6ef1c5a7a22 100644 --- a/src/backend/port/dynloader/cygwin.h +++ b/src/backend/port/dynloader/cygwin.h @@ -13,7 +13,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * In some older systems, the RTLD_NOW flag isn't defined and the mode diff --git a/src/backend/port/dynloader/freebsd.h b/src/backend/port/dynloader/freebsd.h index dbd466d9f0fc859ac15c28cc5ef77884587a9016..bfad4deb013f86bcb6f04d0cec06b9a002d8054d 100644 --- a/src/backend/port/dynloader/freebsd.h +++ b/src/backend/port/dynloader/freebsd.h @@ -18,7 +18,7 @@ #include <link.h> #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on NetBSD 1.0. diff --git a/src/backend/port/dynloader/irix.h b/src/backend/port/dynloader/irix.h index ea8fc7ca15186cb4237c50fd6fe59910d70779f9..2db03133a0cc965d960b7c48a80924dda81df477 100644 --- a/src/backend/port/dynloader/irix.h +++ b/src/backend/port/dynloader/irix.h @@ -15,7 +15,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on Irix. diff --git a/src/backend/port/dynloader/linux.h b/src/backend/port/dynloader/linux.h index f66012d37eeeec10a7cfda5f11e5e84b23162478..a6a42a7c742cd98325c45e59c7888a95c74818b9 100644 --- a/src/backend/port/dynloader/linux.h +++ b/src/backend/port/dynloader/linux.h @@ -14,7 +14,7 @@ #ifndef PORT_PROTOS_H #define PORT_PROTOS_H -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ #ifdef HAVE_DLOPEN #include <dlfcn.h> #endif diff --git a/src/backend/port/dynloader/netbsd.h b/src/backend/port/dynloader/netbsd.h index fa5bb9a28336596f7a68aa754aaf86643e1a547c..eb55194a2f100f5b62b5c5f006aa9360f3270f7c 100644 --- a/src/backend/port/dynloader/netbsd.h +++ b/src/backend/port/dynloader/netbsd.h @@ -19,7 +19,7 @@ #include <link.h> #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on NetBSD 1.0. diff --git a/src/backend/port/dynloader/openbsd.h b/src/backend/port/dynloader/openbsd.h index 15fbe20fe4e25c0704cb441ac6da840b3cb965ed..0e475c65e299eeabea94a789270756317c50ab75 100644 --- a/src/backend/port/dynloader/openbsd.h +++ b/src/backend/port/dynloader/openbsd.h @@ -18,7 +18,7 @@ #include <link.h> #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on NetBSD 1.0. diff --git a/src/backend/port/dynloader/osf.h b/src/backend/port/dynloader/osf.h index 469d751510b8099ffca9275c9f9792ead188105e..87feddf8ad2ca862116094b2221f180d5d852113 100644 --- a/src/backend/port/dynloader/osf.h +++ b/src/backend/port/dynloader/osf.h @@ -16,7 +16,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on Alpha OSF/1.x diff --git a/src/backend/port/dynloader/sco.h b/src/backend/port/dynloader/sco.h index c15d17c32e9a2096b759b802ca113ae1b71acb14..644a0c954c98199ae5427e0b9d6eda0a0df681c7 100644 --- a/src/backend/port/dynloader/sco.h +++ b/src/backend/port/dynloader/sco.h @@ -15,7 +15,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on SCO 3.2v5.0.2 diff --git a/src/backend/port/dynloader/solaris.h b/src/backend/port/dynloader/solaris.h index 518b6b93ef447dd0f5826a0f34bdc6b39f47ef0c..8b874f200f531b24aeb8642095c652f7dc92e3c6 100644 --- a/src/backend/port/dynloader/solaris.h +++ b/src/backend/port/dynloader/solaris.h @@ -15,7 +15,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * In some older systems, the RTLD_NOW flag isn't defined and the mode diff --git a/src/backend/port/dynloader/unixware.h b/src/backend/port/dynloader/unixware.h index 19141ca8d02c49a5acda57082ad4f0a8ba544c4b..9d87a7c614513eb9df6acd2e4253b08a880a35e0 100644 --- a/src/backend/port/dynloader/unixware.h +++ b/src/backend/port/dynloader/unixware.h @@ -18,7 +18,7 @@ #define PORT_PROTOS_H #include <dlfcn.h> -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ /* * Dynamic Loader on UnixWare. diff --git a/src/backend/port/dynloader/win32.h b/src/backend/port/dynloader/win32.h index 850c07bc3c0f56924b20be88e2f3ab7b2fe49b6e..f689dc8ff99f9691644b86142434f369dcfa55e1 100644 --- a/src/backend/port/dynloader/win32.h +++ b/src/backend/port/dynloader/win32.h @@ -4,7 +4,7 @@ #ifndef PORT_PROTOS_H #define PORT_PROTOS_H -#include "utils/dynamic_loader.h" /* pgrminclude ignore */ +#include "utils/dynamic_loader.h" /* pgrminclude ignore */ #define pg_dlopen(f) dlopen((f), 1) #define pg_dlsym dlsym diff --git a/src/backend/port/unix_latch.c b/src/backend/port/unix_latch.c index e64282c210532cde8301cd7aec25e42c3f96aca1..65b2fc56e0361cf76c0e8d4f7b8e3ed602cee5a0 100644 --- a/src/backend/port/unix_latch.c +++ b/src/backend/port/unix_latch.c @@ -183,6 +183,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, { int result = 0; int rc; + #ifdef HAVE_POLL struct pollfd pfds[3]; int nfds; @@ -235,14 +236,15 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, * * Note: we assume that the kernel calls involved in drainSelfPipe() * and SetLatch() will provide adequate synchronization on machines - * with weak memory ordering, so that we cannot miss seeing is_set - * if the signal byte is already in the pipe when we drain it. + * with weak memory ordering, so that we cannot miss seeing is_set if + * the signal byte is already in the pipe when we drain it. */ drainSelfPipe(); if ((wakeEvents & WL_LATCH_SET) && latch->is_set) { result |= WL_LATCH_SET; + /* * Leave loop immediately, avoid blocking again. We don't attempt * to report any other events that might also be satisfied. @@ -309,13 +311,14 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, { result |= WL_SOCKET_WRITEABLE; } + /* * We expect a POLLHUP when the remote end is closed, but because we * don't expect the pipe to become readable or to have any errors * either, treat those as postmaster death, too. */ if ((wakeEvents & WL_POSTMASTER_DEATH) && - (pfds[nfds - 1].revents & (POLLHUP | POLLIN | POLLERR | POLLNVAL))) + (pfds[nfds - 1].revents & (POLLHUP | POLLIN | POLLERR | POLLNVAL))) { /* * According to the select(2) man page on Linux, select(2) may @@ -329,8 +332,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, if (!PostmasterIsAlive()) result |= WL_POSTMASTER_DEATH; } - -#else /* !HAVE_POLL */ +#else /* !HAVE_POLL */ FD_ZERO(&input_mask); FD_ZERO(&output_mask); @@ -387,7 +389,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, result |= WL_SOCKET_WRITEABLE; } if ((wakeEvents & WL_POSTMASTER_DEATH) && - FD_ISSET(postmaster_alive_fds[POSTMASTER_FD_WATCH], &input_mask)) + FD_ISSET(postmaster_alive_fds[POSTMASTER_FD_WATCH], &input_mask)) { /* * According to the select(2) man page on Linux, select(2) may @@ -401,7 +403,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, if (!PostmasterIsAlive()) result |= WL_POSTMASTER_DEATH; } -#endif /* HAVE_POLL */ +#endif /* HAVE_POLL */ } while (result == 0); waiting = false; @@ -423,9 +425,9 @@ SetLatch(volatile Latch *latch) pid_t owner_pid; /* - * XXX there really ought to be a memory barrier operation right here, - * to ensure that any flag variables we might have changed get flushed - * to main memory before we check/set is_set. Without that, we have to + * XXX there really ought to be a memory barrier operation right here, to + * ensure that any flag variables we might have changed get flushed to + * main memory before we check/set is_set. Without that, we have to * require that callers provide their own synchronization for machines * with weak memory ordering (see latch.h). */ @@ -450,12 +452,12 @@ SetLatch(volatile Latch *latch) * Postgres; and PG database processes should handle excess SIGUSR1 * interrupts without a problem anyhow. * - * Another sort of race condition that's possible here is for a new process - * to own the latch immediately after we look, so we don't signal it. - * This is okay so long as all callers of ResetLatch/WaitLatch follow the - * standard coding convention of waiting at the bottom of their loops, - * not the top, so that they'll correctly process latch-setting events that - * happen before they enter the loop. + * Another sort of race condition that's possible here is for a new + * process to own the latch immediately after we look, so we don't signal + * it. This is okay so long as all callers of ResetLatch/WaitLatch follow + * the standard coding convention of waiting at the bottom of their loops, + * not the top, so that they'll correctly process latch-setting events + * that happen before they enter the loop. */ owner_pid = latch->owner_pid; if (owner_pid == 0) @@ -484,7 +486,7 @@ ResetLatch(volatile Latch *latch) /* * XXX there really ought to be a memory barrier operation right here, to * ensure that the write to is_set gets flushed to main memory before we - * examine any flag variables. Otherwise a concurrent SetLatch might + * examine any flag variables. Otherwise a concurrent SetLatch might * falsely conclude that it needn't signal us, even though we have missed * seeing some flag updates that SetLatch was supposed to inform us of. * For the moment, callers must supply their own synchronization of flag diff --git a/src/backend/port/win32/mingwcompat.c b/src/backend/port/win32/mingwcompat.c index 5d91c62bc1b3bf4bb0900ba8ee1ebc274baea331..0978e8cf4677e8c5f215e0a7be48999eb1597408 100644 --- a/src/backend/port/win32/mingwcompat.c +++ b/src/backend/port/win32/mingwcompat.c @@ -42,8 +42,8 @@ LoadKernel32() kernel32 = LoadLibraryEx("kernel32.dll", NULL, 0); if (kernel32 == NULL) ereport(FATAL, - (errmsg_internal("could not load kernel32.dll: error code %lu", - GetLastError()))); + (errmsg_internal("could not load kernel32.dll: error code %lu", + GetLastError()))); } diff --git a/src/backend/port/win32/socket.c b/src/backend/port/win32/socket.c index a7215cad6ec108ece1bd6f8947d29d572f77a00c..d9222231a1e3044dfe8c13b6ac14be881d120c8f 100644 --- a/src/backend/port/win32/socket.c +++ b/src/backend/port/win32/socket.c @@ -151,7 +151,7 @@ pgwin32_waitforsinglesocket(SOCKET s, int what, int timeout) (errmsg_internal("could not reset socket waiting event: error code %lu", GetLastError()))); /* - * Track whether socket is UDP or not. (NB: most likely, this is both + * Track whether socket is UDP or not. (NB: most likely, this is both * useless and wrong; there is no reason to think that the behavior of * WSAEventSelect is different for TCP and UDP.) */ @@ -160,8 +160,9 @@ pgwin32_waitforsinglesocket(SOCKET s, int what, int timeout) current_socket = s; /* - * Attach event to socket. NOTE: we must detach it again before returning, - * since other bits of code may try to attach other events to the socket. + * Attach event to socket. NOTE: we must detach it again before + * returning, since other bits of code may try to attach other events to + * the socket. */ if (WSAEventSelect(s, waitevent, what) != 0) { diff --git a/src/backend/port/win32/timer.c b/src/backend/port/win32/timer.c index 770275acefa26e3f96f4bbfca185093fc0366819..232317369d17368fb5681f3352be02638cb5699f 100644 --- a/src/backend/port/win32/timer.c +++ b/src/backend/port/win32/timer.c @@ -97,8 +97,8 @@ setitimer(int which, const struct itimerval * value, struct itimerval * ovalue) timerCommArea.event = CreateEvent(NULL, TRUE, FALSE, NULL); if (timerCommArea.event == NULL) ereport(FATAL, - (errmsg_internal("could not create timer event: error code %lu", - GetLastError()))); + (errmsg_internal("could not create timer event: error code %lu", + GetLastError()))); MemSet(&timerCommArea.value, 0, sizeof(struct itimerval)); @@ -107,8 +107,8 @@ setitimer(int which, const struct itimerval * value, struct itimerval * ovalue) timerThreadHandle = CreateThread(NULL, 0, pg_timer_thread, NULL, 0, NULL); if (timerThreadHandle == INVALID_HANDLE_VALUE) ereport(FATAL, - (errmsg_internal("could not create timer thread: error code %lu", - GetLastError()))); + (errmsg_internal("could not create timer thread: error code %lu", + GetLastError()))); } /* Request the timer thread to change settings */ diff --git a/src/backend/port/win32_latch.c b/src/backend/port/win32_latch.c index 05b34269b5a1b870e35d7dc7d882f9e8dc116c79..eb46dcad1ba360d84ba2ba614b920965e2f32524 100644 --- a/src/backend/port/win32_latch.c +++ b/src/backend/port/win32_latch.c @@ -173,6 +173,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, if ((wakeEvents & WL_LATCH_SET) && latch->is_set) { result |= WL_LATCH_SET; + /* * Leave loop immediately, avoid blocking again. We don't attempt * to report any other events that might also be satisfied. @@ -199,7 +200,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, /* Latch is set, we'll handle that on next iteration of loop */ } else if ((wakeEvents & (WL_SOCKET_READABLE | WL_SOCKET_WRITEABLE)) && - rc == WAIT_OBJECT_0 + 2) /* socket is at event slot 2 */ + rc == WAIT_OBJECT_0 + 2) /* socket is at event slot 2 */ { WSANETWORKEVENTS resEvents; @@ -222,7 +223,7 @@ WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, rc == WAIT_OBJECT_0 + pmdeath_eventno) { /* - * Postmaster apparently died. Since the consequences of falsely + * Postmaster apparently died. Since the consequences of falsely * returning WL_POSTMASTER_DEATH could be pretty unpleasant, we * take the trouble to positively verify this with * PostmasterIsAlive(), even though there is no known reason to diff --git a/src/backend/port/win32_sema.c b/src/backend/port/win32_sema.c index a093e9f163935135965a4ea5afec2285531f00ac..ef1a4c31bda94288c2308d7dabb86d8bb024c1af 100644 --- a/src/backend/port/win32_sema.c +++ b/src/backend/port/win32_sema.c @@ -162,7 +162,7 @@ PGSemaphoreLock(PGSemaphore sema, bool interruptOK) if (errno != 0) ereport(FATAL, - (errmsg("could not lock semaphore: error code %lu", GetLastError()))); + (errmsg("could not lock semaphore: error code %lu", GetLastError()))); } /* @@ -204,7 +204,7 @@ PGSemaphoreTryLock(PGSemaphore sema) /* Otherwise we are in trouble */ ereport(FATAL, - (errmsg("could not try-lock semaphore: error code %lu", GetLastError()))); + (errmsg("could not try-lock semaphore: error code %lu", GetLastError()))); /* keep compiler quiet */ return false; diff --git a/src/backend/postmaster/autovacuum.c b/src/backend/postmaster/autovacuum.c index 9ff19b7a482de2256177a8e2eded2fd6cde04f23..1cfac9e80babacd607a86238bb94846da76ecc84 100644 --- a/src/backend/postmaster/autovacuum.c +++ b/src/backend/postmaster/autovacuum.c @@ -543,9 +543,9 @@ AutoVacLauncherMain(int argc, char *argv[]) SetConfigOption("statement_timeout", "0", PGC_SUSET, PGC_S_OVERRIDE); /* - * Force default_transaction_isolation to READ COMMITTED. We don't - * want to pay the overhead of serializable mode, nor add any risk - * of causing deadlocks or delaying other transactions. + * Force default_transaction_isolation to READ COMMITTED. We don't want + * to pay the overhead of serializable mode, nor add any risk of causing + * deadlocks or delaying other transactions. */ SetConfigOption("default_transaction_isolation", "read committed", PGC_SUSET, PGC_S_OVERRIDE); @@ -1553,9 +1553,9 @@ AutoVacWorkerMain(int argc, char *argv[]) SetConfigOption("statement_timeout", "0", PGC_SUSET, PGC_S_OVERRIDE); /* - * Force default_transaction_isolation to READ COMMITTED. We don't - * want to pay the overhead of serializable mode, nor add any risk - * of causing deadlocks or delaying other transactions. + * Force default_transaction_isolation to READ COMMITTED. We don't want + * to pay the overhead of serializable mode, nor add any risk of causing + * deadlocks or delaying other transactions. */ SetConfigOption("default_transaction_isolation", "read committed", PGC_SUSET, PGC_S_OVERRIDE); diff --git a/src/backend/postmaster/bgwriter.c b/src/backend/postmaster/bgwriter.c index 32c254c8420469973023563c6dc4ee9d7099f72e..5f93fccbfab1bbb8306f5de4ad228f3cb48b0862 100644 --- a/src/backend/postmaster/bgwriter.c +++ b/src/backend/postmaster/bgwriter.c @@ -121,7 +121,7 @@ BackgroundWriterMain(void) */ pqsignal(SIGHUP, BgSigHupHandler); /* set flag to read config file */ pqsignal(SIGINT, SIG_IGN); - pqsignal(SIGTERM, ReqShutdownHandler); /* shutdown */ + pqsignal(SIGTERM, ReqShutdownHandler); /* shutdown */ pqsignal(SIGQUIT, bg_quickdie); /* hard crash time */ pqsignal(SIGALRM, SIG_IGN); pqsignal(SIGPIPE, SIG_IGN); @@ -244,8 +244,8 @@ BackgroundWriterMain(void) */ for (;;) { - bool can_hibernate; - int rc; + bool can_hibernate; + int rc; /* Clear any already-pending wakeups */ ResetLatch(&MyProc->procLatch); @@ -297,7 +297,7 @@ BackgroundWriterMain(void) */ rc = WaitLatch(&MyProc->procLatch, WL_LATCH_SET | WL_TIMEOUT | WL_POSTMASTER_DEATH, - BgWriterDelay /* ms */); + BgWriterDelay /* ms */ ); /* * If no latch event and BgBufferSync says nothing's happening, extend @@ -314,7 +314,7 @@ BackgroundWriterMain(void) * and the time we call StrategyNotifyBgWriter. While it's not * critical that we not hibernate anyway, we try to reduce the odds of * that by only hibernating when BgBufferSync says nothing's happening - * for two consecutive cycles. Also, we mitigate any possible + * for two consecutive cycles. Also, we mitigate any possible * consequences of a missed wakeup by not hibernating forever. */ if (rc == WL_TIMEOUT && can_hibernate && prev_hibernate) diff --git a/src/backend/postmaster/checkpointer.c b/src/backend/postmaster/checkpointer.c index 6aeade92e6a92e328006e74d3f19346d91a8dd20..87aa4e1ba0a2082eb47175f16bc321d058cbee99 100644 --- a/src/backend/postmaster/checkpointer.c +++ b/src/backend/postmaster/checkpointer.c @@ -2,7 +2,7 @@ * * checkpointer.c * - * The checkpointer is new as of Postgres 9.2. It handles all checkpoints. + * The checkpointer is new as of Postgres 9.2. It handles all checkpoints. * Checkpoints are automatically dispatched after a certain amount of time has * elapsed since the last one, and it can be signaled to perform requested * checkpoints as well. (The GUC parameter that mandates a checkpoint every @@ -14,7 +14,7 @@ * subprocess finishes, or as soon as recovery begins if we are doing archive * recovery. It remains alive until the postmaster commands it to terminate. * Normal termination is by SIGUSR2, which instructs the checkpointer to - * execute a shutdown checkpoint and then exit(0). (All backends must be + * execute a shutdown checkpoint and then exit(0). (All backends must be * stopped before SIGUSR2 is issued!) Emergency termination is by SIGQUIT; * like any backend, the checkpointer will simply abort and exit on SIGQUIT. * @@ -113,7 +113,7 @@ typedef struct typedef struct { - pid_t checkpointer_pid; /* PID (0 if not started) */ + pid_t checkpointer_pid; /* PID (0 if not started) */ slock_t ckpt_lck; /* protects all the ckpt_* fields */ @@ -199,7 +199,7 @@ CheckpointerMain(void) /* * If possible, make this process a group leader, so that the postmaster - * can signal any child processes too. (checkpointer probably never has + * can signal any child processes too. (checkpointer probably never has * any child processes, but for consistency we make all postmaster child * processes do this.) */ @@ -216,10 +216,11 @@ CheckpointerMain(void) * want to wait for the backends to exit, whereupon the postmaster will * tell us it's okay to shut down (via SIGUSR2). */ - pqsignal(SIGHUP, ChkptSigHupHandler); /* set flag to read config file */ - pqsignal(SIGINT, ReqCheckpointHandler); /* request checkpoint */ - pqsignal(SIGTERM, SIG_IGN); /* ignore SIGTERM */ - pqsignal(SIGQUIT, chkpt_quickdie); /* hard crash time */ + pqsignal(SIGHUP, ChkptSigHupHandler); /* set flag to read config + * file */ + pqsignal(SIGINT, ReqCheckpointHandler); /* request checkpoint */ + pqsignal(SIGTERM, SIG_IGN); /* ignore SIGTERM */ + pqsignal(SIGQUIT, chkpt_quickdie); /* hard crash time */ pqsignal(SIGALRM, SIG_IGN); pqsignal(SIGPIPE, SIG_IGN); pqsignal(SIGUSR1, chkpt_sigusr1_handler); @@ -255,10 +256,10 @@ CheckpointerMain(void) * TopMemoryContext, but resetting that would be a really bad idea. */ checkpointer_context = AllocSetContextCreate(TopMemoryContext, - "Checkpointer", - ALLOCSET_DEFAULT_MINSIZE, - ALLOCSET_DEFAULT_INITSIZE, - ALLOCSET_DEFAULT_MAXSIZE); + "Checkpointer", + ALLOCSET_DEFAULT_MINSIZE, + ALLOCSET_DEFAULT_INITSIZE, + ALLOCSET_DEFAULT_MAXSIZE); MemoryContextSwitchTo(checkpointer_context); /* @@ -280,7 +281,8 @@ CheckpointerMain(void) /* * These operations are really just a minimal subset of * AbortTransaction(). We don't have very many resources to worry - * about in checkpointer, but we do have LWLocks, buffers, and temp files. + * about in checkpointer, but we do have LWLocks, buffers, and temp + * files. */ LWLockReleaseAll(); AbortBufferIO(); @@ -351,9 +353,8 @@ CheckpointerMain(void) ThisTimeLineID = GetRecoveryTargetTLI(); /* - * Ensure all shared memory values are set correctly for the config. - * Doing this here ensures no race conditions from other concurrent - * updaters. + * Ensure all shared memory values are set correctly for the config. Doing + * this here ensures no race conditions from other concurrent updaters. */ UpdateSharedMemoryConfig(); @@ -389,9 +390,9 @@ CheckpointerMain(void) ProcessConfigFile(PGC_SIGHUP); /* - * Checkpointer is the last process to shut down, so we ask - * it to hold the keys for a range of other tasks required - * most of which have nothing to do with checkpointing at all. + * Checkpointer is the last process to shut down, so we ask it to + * hold the keys for a range of other tasks required most of which + * have nothing to do with checkpointing at all. * * For various reasons, some config values can change dynamically * so the primary copy of them is held in shared memory to make @@ -490,7 +491,8 @@ CheckpointerMain(void) errhint("Consider increasing the configuration parameter \"checkpoint_segments\"."))); /* - * Initialize checkpointer-private variables used during checkpoint + * Initialize checkpointer-private variables used during + * checkpoint */ ckpt_active = true; if (!do_restartpoint) @@ -558,8 +560,8 @@ CheckpointerMain(void) pgstat_send_bgwriter(); /* - * Sleep until we are signaled or it's time for another checkpoint - * or xlog file switch. + * Sleep until we are signaled or it's time for another checkpoint or + * xlog file switch. */ now = (pg_time_t) time(NULL); elapsed_secs = now - last_checkpoint_time; @@ -576,7 +578,7 @@ CheckpointerMain(void) rc = WaitLatch(&MyProc->procLatch, WL_LATCH_SET | WL_TIMEOUT | WL_POSTMASTER_DEATH, - cur_timeout * 1000L /* convert to ms */); + cur_timeout * 1000L /* convert to ms */ ); /* * Emergency bailout if postmaster has died. This is to avoid the @@ -687,8 +689,8 @@ CheckpointWriteDelay(int flags, double progress) return; /* - * Perform the usual duties and take a nap, unless we're behind - * schedule, in which case we just try to catch up as quickly as possible. + * Perform the usual duties and take a nap, unless we're behind schedule, + * in which case we just try to catch up as quickly as possible. */ if (!(flags & CHECKPOINT_IMMEDIATE) && !shutdown_requested && @@ -716,7 +718,8 @@ CheckpointWriteDelay(int flags, double progress) /* * This sleep used to be connected to bgwriter_delay, typically 200ms. * That resulted in more frequent wakeups if not much work to do. - * Checkpointer and bgwriter are no longer related so take the Big Sleep. + * Checkpointer and bgwriter are no longer related so take the Big + * Sleep. */ pg_usleep(100000L); } @@ -1017,7 +1020,7 @@ RequestCheckpoint(int flags) if (ntries >= 20) /* max wait 2.0 sec */ { elog((flags & CHECKPOINT_WAIT) ? ERROR : LOG, - "could not request checkpoint because checkpointer not running"); + "could not request checkpoint because checkpointer not running"); break; } } @@ -1130,9 +1133,9 @@ ForwardFsyncRequest(RelFileNodeBackend rnode, ForkNumber forknum, CheckpointerShmem->num_backend_writes++; /* - * If the checkpointer isn't running or the request queue is full, - * the backend will have to perform its own fsync request. But before - * forcing that to happen, we can try to compact the request queue. + * If the checkpointer isn't running or the request queue is full, the + * backend will have to perform its own fsync request. But before forcing + * that to happen, we can try to compact the request queue. */ if (CheckpointerShmem->checkpointer_pid == 0 || (CheckpointerShmem->num_requests >= CheckpointerShmem->max_requests && @@ -1339,8 +1342,8 @@ UpdateSharedMemoryConfig(void) SyncRepUpdateSyncStandbysDefined(); /* - * If full_page_writes has been changed by SIGHUP, we update it - * in shared memory and write an XLOG_FPW_CHANGE record. + * If full_page_writes has been changed by SIGHUP, we update it in shared + * memory and write an XLOG_FPW_CHANGE record. */ UpdateFullPageWrites(); @@ -1356,9 +1359,9 @@ FirstCallSinceLastCheckpoint(void) { /* use volatile pointer to prevent code rearrangement */ volatile CheckpointerShmemStruct *cps = CheckpointerShmem; - static int ckpt_done = 0; - int new_done; - bool FirstCall = false; + static int ckpt_done = 0; + int new_done; + bool FirstCall = false; SpinLockAcquire(&cps->ckpt_lck); new_done = cps->ckpt_done; diff --git a/src/backend/postmaster/pgarch.c b/src/backend/postmaster/pgarch.c index 37fc73592cea503649b687de81061acb561629a1..5c43cdde653b7e04f506e20738b1ee9e6293a531 100644 --- a/src/backend/postmaster/pgarch.c +++ b/src/backend/postmaster/pgarch.c @@ -361,9 +361,9 @@ pgarch_MainLoop(void) wakened = true; /* - * There shouldn't be anything for the archiver to do except to wait - * for a signal ... however, the archiver exists to protect our data, - * so she wakes up occasionally to allow herself to be proactive. + * There shouldn't be anything for the archiver to do except to wait for a + * signal ... however, the archiver exists to protect our data, so she + * wakes up occasionally to allow herself to be proactive. */ do { @@ -410,18 +410,18 @@ pgarch_MainLoop(void) * PGARCH_AUTOWAKE_INTERVAL having passed since last_copy_time, or * until postmaster dies. */ - if (!time_to_stop) /* Don't wait during last iteration */ + if (!time_to_stop) /* Don't wait during last iteration */ { - pg_time_t curtime = (pg_time_t) time(NULL); - int timeout; + pg_time_t curtime = (pg_time_t) time(NULL); + int timeout; timeout = PGARCH_AUTOWAKE_INTERVAL - (curtime - last_copy_time); if (timeout > 0) { - int rc; + int rc; rc = WaitLatch(&mainloop_latch, - WL_LATCH_SET | WL_TIMEOUT | WL_POSTMASTER_DEATH, + WL_LATCH_SET | WL_TIMEOUT | WL_POSTMASTER_DEATH, timeout * 1000L); if (rc & WL_TIMEOUT) wakened = true; diff --git a/src/backend/postmaster/pgstat.c b/src/backend/postmaster/pgstat.c index 424cfcaf4f71672307f0eb6208ed0ee42f2cb8b2..73d5b2e39c88ed0f29f4927e3de709930b406595 100644 --- a/src/backend/postmaster/pgstat.c +++ b/src/backend/postmaster/pgstat.c @@ -1522,7 +1522,7 @@ pgstat_end_function_usage(PgStat_FunctionCallUsage *fcu, bool finalize) /* * Compute the new f_total_time as the total elapsed time added to the - * pre-call value of f_total_time. This is necessary to avoid + * pre-call value of f_total_time. This is necessary to avoid * double-counting any time taken by recursive calls of myself. (We do * not need any similar kluge for self time, since that already excludes * any recursive calls.) @@ -2836,7 +2836,7 @@ pgstat_get_backend_current_activity(int pid, bool checkUser) * pgstat_get_crashed_backend_activity() - * * Return a string representing the current activity of the backend with - * the specified PID. Like the function above, but reads shared memory with + * the specified PID. Like the function above, but reads shared memory with * the expectation that it may be corrupt. On success, copy the string * into the "buffer" argument and return that pointer. On failure, * return NULL. @@ -2845,7 +2845,7 @@ pgstat_get_backend_current_activity(int pid, bool checkUser) * query that crashed a backend. In particular, no attempt is made to * follow the correct concurrency protocol when accessing the * BackendStatusArray. But that's OK, in the worst case we'll return a - * corrupted message. We also must take care not to trip on ereport(ERROR). + * corrupted message. We also must take care not to trip on ereport(ERROR). * ---------- */ const char * @@ -2890,8 +2890,8 @@ pgstat_get_crashed_backend_activity(int pid, char *buffer, int buflen) /* * Copy only ASCII-safe characters so we don't run into encoding - * problems when reporting the message; and be sure not to run - * off the end of memory. + * problems when reporting the message; and be sure not to run off + * the end of memory. */ ascii_safe_strlcpy(buffer, activity, Min(buflen, pgstat_track_activity_query_size)); @@ -3070,7 +3070,7 @@ PgstatCollectorMain(int argc, char *argv[]) * every message; instead, do that only after a recv() fails to obtain a * message. (This effectively means that if backends are sending us stuff * like mad, we won't notice postmaster death until things slack off a - * bit; which seems fine.) To do that, we have an inner loop that + * bit; which seems fine.) To do that, we have an inner loop that * iterates as long as recv() succeeds. We do recognize got_SIGHUP inside * the inner loop, which means that such interrupts will get serviced but * the latch won't get cleared until next time there is a break in the @@ -3234,13 +3234,14 @@ PgstatCollectorMain(int argc, char *argv[]) /* Sleep until there's something to do */ #ifndef WIN32 wr = WaitLatchOrSocket(&pgStatLatch, - WL_LATCH_SET | WL_POSTMASTER_DEATH | WL_SOCKET_READABLE, + WL_LATCH_SET | WL_POSTMASTER_DEATH | WL_SOCKET_READABLE, pgStatSock, -1L); #else + /* * Windows, at least in its Windows Server 2003 R2 incarnation, - * sometimes loses FD_READ events. Waking up and retrying the recv() + * sometimes loses FD_READ events. Waking up and retrying the recv() * fixes that, so don't sleep indefinitely. This is a crock of the * first water, but until somebody wants to debug exactly what's * happening there, this is the best we can do. The two-second @@ -3249,9 +3250,9 @@ PgstatCollectorMain(int argc, char *argv[]) * backend_read_statsfile. */ wr = WaitLatchOrSocket(&pgStatLatch, - WL_LATCH_SET | WL_POSTMASTER_DEATH | WL_SOCKET_READABLE | WL_TIMEOUT, + WL_LATCH_SET | WL_POSTMASTER_DEATH | WL_SOCKET_READABLE | WL_TIMEOUT, pgStatSock, - 2 * 1000L /* msec */); + 2 * 1000L /* msec */ ); #endif /* diff --git a/src/backend/postmaster/postmaster.c b/src/backend/postmaster/postmaster.c index 830a83f60e00d452b9076342a70b105b2925f80c..eeea933b1923dc06a6ab2ebcd39cd6d5359c4696 100644 --- a/src/backend/postmaster/postmaster.c +++ b/src/backend/postmaster/postmaster.c @@ -203,7 +203,7 @@ bool enable_bonjour = false; char *bonjour_name; bool restart_after_crash = true; -char *output_config_variable = NULL; +char *output_config_variable = NULL; /* PIDs of special child processes; 0 when not running */ static pid_t StartupPID = 0, @@ -243,7 +243,7 @@ static bool RecoveryError = false; /* T if WAL recovery failed */ * checkpointer are launched, while the startup process continues applying WAL. * If Hot Standby is enabled, then, after reaching a consistent point in WAL * redo, startup process signals us again, and we switch to PM_HOT_STANDBY - * state and begin accepting connections to perform read-only queries. When + * state and begin accepting connections to perform read-only queries. When * archive recovery is finished, the startup process exits with exit code 0 * and we switch to PM_RUN state. * @@ -280,7 +280,8 @@ typedef enum PM_WAIT_BACKUP, /* waiting for online backup mode to end */ PM_WAIT_READONLY, /* waiting for read only backends to exit */ PM_WAIT_BACKENDS, /* waiting for live backends to exit */ - PM_SHUTDOWN, /* waiting for checkpointer to do shutdown ckpt */ + PM_SHUTDOWN, /* waiting for checkpointer to do shutdown + * ckpt */ PM_SHUTDOWN_2, /* waiting for archiver and walsenders to * finish */ PM_WAIT_DEAD_END, /* waiting for dead_end children to exit */ @@ -481,7 +482,7 @@ static void ShmemBackendArrayRemove(Backend *bn); * File descriptors for pipe used to monitor if postmaster is alive. * First is POSTMASTER_FD_WATCH, second is POSTMASTER_FD_OWN. */ -int postmaster_alive_fds[2] = { -1, -1 }; +int postmaster_alive_fds[2] = {-1, -1}; #else /* Process handle of postmaster used for the same purpose on Windows */ HANDLE PostmasterHandle; @@ -740,11 +741,14 @@ PostmasterMain(int argc, char *argv[]) if (output_config_variable != NULL) { - /* permission is handled because the user is reading inside the data dir */ + /* + * permission is handled because the user is reading inside the data + * dir + */ puts(GetConfigOption(output_config_variable, false, false)); ExitPostmaster(0); } - + /* Verify that DataDir looks reasonable */ checkDataDir(); @@ -791,8 +795,8 @@ PostmasterMain(int argc, char *argv[]) char **p; ereport(DEBUG3, - (errmsg_internal("%s: PostmasterMain: initial environment dump:", - progname))); + (errmsg_internal("%s: PostmasterMain: initial environment dump:", + progname))); ereport(DEBUG3, (errmsg_internal("-----------------------------------------"))); for (p = environ; *p; ++p) @@ -981,6 +985,7 @@ PostmasterMain(int argc, char *argv[]) InitPostmasterDeathWatchHandle(); #ifdef WIN32 + /* * Initialize I/O completion port used to deliver list of dead children. */ @@ -1979,6 +1984,7 @@ ClosePostmasterPorts(bool am_syslogger) int i; #ifndef WIN32 + /* * Close the write end of postmaster death watch pipe. It's important to * do this as early as possible, so that if postmaster dies, others won't @@ -1986,8 +1992,8 @@ ClosePostmasterPorts(bool am_syslogger) */ if (close(postmaster_alive_fds[POSTMASTER_FD_OWN])) ereport(FATAL, - (errcode_for_file_access(), - errmsg_internal("could not close postmaster death monitoring pipe in child process: %m"))); + (errcode_for_file_access(), + errmsg_internal("could not close postmaster death monitoring pipe in child process: %m"))); postmaster_alive_fds[POSTMASTER_FD_OWN] = -1; #endif @@ -2357,13 +2363,14 @@ reaper(SIGNAL_ARGS) * disconnection. * * XXX should avoid the need for disconnection. When we do, - * am_cascading_walsender should be replaced with RecoveryInProgress() + * am_cascading_walsender should be replaced with + * RecoveryInProgress() */ if (max_wal_senders > 0 && CountChildren(BACKEND_TYPE_WALSND) > 0) { ereport(LOG, (errmsg("terminating all walsender processes to force cascaded " - "standby(s) to update timeline and reconnect"))); + "standby(s) to update timeline and reconnect"))); SignalSomeChildren(SIGUSR2, BACKEND_TYPE_WALSND); } @@ -2398,8 +2405,8 @@ reaper(SIGNAL_ARGS) } /* - * Was it the bgwriter? Normal exit can be ignored; we'll start a - * new one at the next iteration of the postmaster's main loop, if + * Was it the bgwriter? Normal exit can be ignored; we'll start a new + * one at the next iteration of the postmaster's main loop, if * necessary. Any other exit condition is treated as a crash. */ if (pid == BgWriterPID) @@ -2420,8 +2427,8 @@ reaper(SIGNAL_ARGS) if (EXIT_STATUS_0(exitstatus) && pmState == PM_SHUTDOWN) { /* - * OK, we saw normal exit of the checkpointer after it's been told - * to shut down. We expect that it wrote a shutdown + * OK, we saw normal exit of the checkpointer after it's been + * told to shut down. We expect that it wrote a shutdown * checkpoint. (If for some reason it didn't, recovery will * occur on next postmaster start.) * @@ -2457,8 +2464,8 @@ reaper(SIGNAL_ARGS) else { /* - * Any unexpected exit of the checkpointer (including FATAL exit) - * is treated as a crash. + * Any unexpected exit of the checkpointer (including FATAL + * exit) is treated as a crash. */ HandleChildCrash(pid, exitstatus, _("checkpointer process")); @@ -2847,7 +2854,7 @@ LogChildExit(int lev, const char *procname, int pid, int exitstatus) if (!EXIT_STATUS_0(exitstatus)) activity = pgstat_get_crashed_backend_activity(pid, activity_buffer, - sizeof(activity_buffer)); + sizeof(activity_buffer)); if (WIFEXITED(exitstatus)) ereport(lev, @@ -2879,7 +2886,7 @@ LogChildExit(int lev, const char *procname, int pid, int exitstatus) procname, pid, WTERMSIG(exitstatus), WTERMSIG(exitstatus) < NSIG ? sys_siglist[WTERMSIG(exitstatus)] : "(unknown)"), - activity ? errdetail("Failed process was running: %s", activity) : 0)); + activity ? errdetail("Failed process was running: %s", activity) : 0)); #else ereport(lev, @@ -2947,14 +2954,14 @@ PostmasterStateMachine(void) { /* * PM_WAIT_BACKENDS state ends when we have no regular backends - * (including autovac workers) and no walwriter, autovac launcher - * or bgwriter. If we are doing crash recovery then we expect the - * checkpointer to exit as well, otherwise not. - * The archiver, stats, and syslogger processes - * are disregarded since they are not connected to shared memory; we - * also disregard dead_end children here. Walsenders are also - * disregarded, they will be terminated later after writing the - * checkpoint record, like the archiver process. + * (including autovac workers) and no walwriter, autovac launcher or + * bgwriter. If we are doing crash recovery then we expect the + * checkpointer to exit as well, otherwise not. The archiver, stats, + * and syslogger processes are disregarded since they are not + * connected to shared memory; we also disregard dead_end children + * here. Walsenders are also disregarded, they will be terminated + * later after writing the checkpoint record, like the archiver + * process. */ if (CountChildren(BACKEND_TYPE_NORMAL | BACKEND_TYPE_AUTOVAC) == 0 && StartupPID == 0 && @@ -2997,10 +3004,10 @@ PostmasterStateMachine(void) else { /* - * If we failed to fork a checkpointer, just shut down. Any - * required cleanup will happen at next restart. We set - * FatalError so that an "abnormal shutdown" message gets - * logged when we exit. + * If we failed to fork a checkpointer, just shut down. + * Any required cleanup will happen at next restart. We + * set FatalError so that an "abnormal shutdown" message + * gets logged when we exit. */ FatalError = true; pmState = PM_WAIT_DEAD_END; @@ -3086,13 +3093,13 @@ PostmasterStateMachine(void) else { /* - * Terminate exclusive backup mode to avoid recovery after a clean fast - * shutdown. Since an exclusive backup can only be taken during normal - * running (and not, for example, while running under Hot Standby) - * it only makes sense to do this if we reached normal running. If - * we're still in recovery, the backup file is one we're - * recovering *from*, and we must keep it around so that recovery - * restarts from the right place. + * Terminate exclusive backup mode to avoid recovery after a clean + * fast shutdown. Since an exclusive backup can only be taken + * during normal running (and not, for example, while running + * under Hot Standby) it only makes sense to do this if we reached + * normal running. If we're still in recovery, the backup file is + * one we're recovering *from*, and we must keep it around so that + * recovery restarts from the right place. */ if (ReachedNormalRunning) CancelBackup(); @@ -3437,7 +3444,7 @@ BackendInitialize(Port *port) if (pg_getnameinfo_all(&port->raddr.addr, port->raddr.salen, remote_host, sizeof(remote_host), remote_port, sizeof(remote_port), - (log_hostname ? 0 : NI_NUMERICHOST) | NI_NUMERICSERV) != 0) + (log_hostname ? 0 : NI_NUMERICHOST) | NI_NUMERICSERV) != 0) { int ret = pg_getnameinfo_all(&port->raddr.addr, port->raddr.salen, remote_host, sizeof(remote_host), @@ -3930,8 +3937,8 @@ internal_forkexec(int argc, char *argv[], Port *port) INFINITE, WT_EXECUTEONLYONCE | WT_EXECUTEINWAITTHREAD)) ereport(FATAL, - (errmsg_internal("could not register process for wait: error code %lu", - GetLastError()))); + (errmsg_internal("could not register process for wait: error code %lu", + GetLastError()))); /* Don't close pi.hProcess here - the wait thread needs access to it */ @@ -4531,7 +4538,7 @@ StartChildProcess(AuxProcType type) break; case CheckpointerProcess: ereport(LOG, - (errmsg("could not fork checkpointer process: %m"))); + (errmsg("could not fork checkpointer process: %m"))); break; case WalWriterProcess: ereport(LOG, @@ -5111,7 +5118,6 @@ pgwin32_deadchild_callback(PVOID lpParameter, BOOLEAN TimerOrWaitFired) /* Queue SIGCHLD signal */ pg_queue_signal(SIGCHLD); } - #endif /* WIN32 */ /* @@ -5124,10 +5130,11 @@ static void InitPostmasterDeathWatchHandle(void) { #ifndef WIN32 + /* * Create a pipe. Postmaster holds the write end of the pipe open - * (POSTMASTER_FD_OWN), and children hold the read end. Children can - * pass the read file descriptor to select() to wake up in case postmaster + * (POSTMASTER_FD_OWN), and children hold the read end. Children can pass + * the read file descriptor to select() to wake up in case postmaster * dies, or check for postmaster death with a (read() == 0). Children must * close the write end as soon as possible after forking, because EOF * won't be signaled in the read end until all processes have closed the @@ -5147,8 +5154,8 @@ InitPostmasterDeathWatchHandle(void) ereport(FATAL, (errcode_for_socket_access(), errmsg_internal("could not set postmaster death monitoring pipe to non-blocking mode: %m"))); - #else + /* * On Windows, we use a process handle for the same purpose. */ diff --git a/src/backend/postmaster/syslogger.c b/src/backend/postmaster/syslogger.c index 352887924342699f9cc07b8a023669b3a2d4f6f0..25d095b158de34ca390f443d0a8e2f3fe1ee2d61 100644 --- a/src/backend/postmaster/syslogger.c +++ b/src/backend/postmaster/syslogger.c @@ -401,7 +401,7 @@ SysLoggerMain(int argc, char *argv[]) /* * Calculate time till next time-based rotation, so that we don't - * sleep longer than that. We assume the value of "now" obtained + * sleep longer than that. We assume the value of "now" obtained * above is still close enough. Note we can't make this calculation * until after calling logfile_rotate(), since it will advance * next_rotation_time. @@ -409,7 +409,7 @@ SysLoggerMain(int argc, char *argv[]) if (Log_RotationAge > 0 && !rotation_disabled) { if (now < next_rotation_time) - cur_timeout = (next_rotation_time - now) * 1000L; /* msec */ + cur_timeout = (next_rotation_time - now) * 1000L; /* msec */ else cur_timeout = 0; cur_flags = WL_TIMEOUT; @@ -632,6 +632,7 @@ SysLogger_Start(void) errmsg("could not redirect stderr: %m"))); close(fd); _setmode(_fileno(stderr), _O_BINARY); + /* * Now we are done with the write end of the pipe. * CloseHandle() must not be called because the preceding diff --git a/src/backend/postmaster/walwriter.c b/src/backend/postmaster/walwriter.c index 77455db1666966b6c076318a231e465e50cfe8e0..b7b85125553cf39bef2bc9e775a2e3c69fad8f98 100644 --- a/src/backend/postmaster/walwriter.c +++ b/src/backend/postmaster/walwriter.c @@ -244,11 +244,11 @@ WalWriterMain(void) */ for (;;) { - long cur_timeout; - int rc; + long cur_timeout; + int rc; /* - * Advertise whether we might hibernate in this cycle. We do this + * Advertise whether we might hibernate in this cycle. We do this * before resetting the latch to ensure that any async commits will * see the flag set if they might possibly need to wake us up, and * that we won't miss any signal they send us. (If we discover work @@ -294,7 +294,7 @@ WalWriterMain(void) * sleep time so as to reduce the server's idle power consumption. */ if (left_till_hibernate > 0) - cur_timeout = WalWriterDelay; /* in ms */ + cur_timeout = WalWriterDelay; /* in ms */ else cur_timeout = WalWriterDelay * HIBERNATE_FACTOR; diff --git a/src/backend/regex/regc_locale.c b/src/backend/regex/regc_locale.c index c0414a24912f2f8e1d73be9cbe6469546873c57b..da597053448195afec86fa8ca635d8881920d1cb 100644 --- a/src/backend/regex/regc_locale.c +++ b/src/backend/regex/regc_locale.c @@ -513,7 +513,7 @@ cclass(struct vars * v, /* context */ { size_t len; struct cvec *cv = NULL; - const char * const *namePtr; + const char *const * namePtr; int i, index; @@ -521,7 +521,7 @@ cclass(struct vars * v, /* context */ * The following arrays define the valid character class names. */ - static const char * const classNames[] = { + static const char *const classNames[] = { "alnum", "alpha", "ascii", "blank", "cntrl", "digit", "graph", "lower", "print", "punct", "space", "upper", "xdigit", NULL }; @@ -562,8 +562,8 @@ cclass(struct vars * v, /* context */ index = (int) CC_ALPHA; /* - * Now compute the character class contents. For classes that are - * based on the behavior of a <wctype.h> or <ctype.h> function, we use + * Now compute the character class contents. For classes that are based + * on the behavior of a <wctype.h> or <ctype.h> function, we use * pg_ctype_get_cache so that we can cache the results. Other classes * have definitions that are hard-wired here, and for those we just * construct a transient cvec on the fly. @@ -605,10 +605,11 @@ cclass(struct vars * v, /* context */ cv = pg_ctype_get_cache(pg_wc_ispunct); break; case CC_XDIGIT: + /* * It's not clear how to define this in non-western locales, and - * even less clear that there's any particular use in trying. - * So just hard-wire the meaning. + * even less clear that there's any particular use in trying. So + * just hard-wire the meaning. */ cv = getcvec(v, 0, 3); if (cv) diff --git a/src/backend/regex/regc_pg_locale.c b/src/backend/regex/regc_pg_locale.c index eac951f200065bada7be27020d10d09ab960ded1..e85c5ddba51845345488d44cc78fe4d0598b3803 100644 --- a/src/backend/regex/regc_pg_locale.c +++ b/src/backend/regex/regc_pg_locale.c @@ -680,9 +680,9 @@ typedef int (*pg_wc_probefunc) (pg_wchar c); typedef struct pg_ctype_cache { - pg_wc_probefunc probefunc; /* pg_wc_isalpha or a sibling */ - Oid collation; /* collation this entry is for */ - struct cvec cv; /* cache entry contents */ + pg_wc_probefunc probefunc; /* pg_wc_isalpha or a sibling */ + Oid collation; /* collation this entry is for */ + struct cvec cv; /* cache entry contents */ struct pg_ctype_cache *next; /* chain link */ } pg_ctype_cache; @@ -730,7 +730,7 @@ store_match(pg_ctype_cache *pcc, pg_wchar chr1, int nchrs) /* * Given a probe function (e.g., pg_wc_isalpha) get a struct cvec for all - * chrs satisfying the probe function. The active collation is the one + * chrs satisfying the probe function. The active collation is the one * previously set by pg_set_regex_collation. Return NULL if out of memory. * * Note that the result must not be freed or modified by caller. @@ -777,7 +777,7 @@ pg_ctype_get_cache(pg_wc_probefunc probefunc) * UTF8 go up to 0x7FF, which is a pretty arbitrary cutoff but we cannot * extend it as far as we'd like (say, 0xFFFF, the end of the Basic * Multilingual Plane) without creating significant performance issues due - * to too many characters being fed through the colormap code. This will + * to too many characters being fed through the colormap code. This will * need redesign to fix reasonably, but at least for the moment we have * all common European languages covered. Otherwise (not C, not UTF8) go * up to 255. These limits are interrelated with restrictions discussed diff --git a/src/backend/regex/regcomp.c b/src/backend/regex/regcomp.c index 7fd0b07e2c5d45405493247f28fd3aac88883572..57055f04abb807466744daf74d8506b9f0e594b8 100644 --- a/src/backend/regex/regcomp.c +++ b/src/backend/regex/regcomp.c @@ -1119,11 +1119,11 @@ parseqatom(struct vars * v, { /* * If there's no backrefs involved, we can turn x{m,n} into - * x{m-1,n-1}x, with capturing parens in only the second x. This - * is valid because we only care about capturing matches from the - * final iteration of the quantifier. It's a win because we can - * implement the backref-free left side as a plain DFA node, since - * we don't really care where its submatches are. + * x{m-1,n-1}x, with capturing parens in only the second x. This is + * valid because we only care about capturing matches from the final + * iteration of the quantifier. It's a win because we can implement + * the backref-free left side as a plain DFA node, since we don't + * really care where its submatches are. */ dupnfa(v->nfa, atom->begin, atom->end, s, atom->begin); assert(m >= 1 && m != INFINITY && n >= 1); diff --git a/src/backend/regex/rege_dfa.c b/src/backend/regex/rege_dfa.c index f4fd41458bd4677196412e6e8d58d982fe97a125..da7a0bf402ff41bace723924fa8a8f0635eb15ec 100644 --- a/src/backend/regex/rege_dfa.c +++ b/src/backend/regex/rege_dfa.c @@ -272,7 +272,7 @@ static struct dfa * newdfa(struct vars * v, struct cnfa * cnfa, struct colormap * cm, - struct smalldfa * sml) /* preallocated space, may be NULL */ + struct smalldfa * sml) /* preallocated space, may be NULL */ { struct dfa *d; size_t nss = cnfa->nstates * 2; diff --git a/src/backend/regex/regerror.c b/src/backend/regex/regerror.c index 9d44eb04cec2e9c138109018901d5ca0bc250945..f6a3f2667f8795cfcf649b2b9d30cea5b9a8ebe4 100644 --- a/src/backend/regex/regerror.c +++ b/src/backend/regex/regerror.c @@ -46,7 +46,7 @@ static struct rerr { /* the actual table is built from regex.h */ -#include "regex/regerrs.h" /* pgrminclude ignore */ +#include "regex/regerrs.h" /* pgrminclude ignore */ { -1, "", "oops" }, /* explanation special-cased in code */ diff --git a/src/backend/regex/regexec.c b/src/backend/regex/regexec.c index 5d7415b3c1ab973d86411206593f2f0a085199c0..3748a9c1714d67f97eee79c3f51baa73069c1327 100644 --- a/src/backend/regex/regexec.c +++ b/src/backend/regex/regexec.c @@ -531,7 +531,7 @@ zaptreesubs(struct vars * v, { if (t->op == '(') { - int n = t->subno; + int n = t->subno; assert(n > 0); if ((size_t) n < v->nmatch) @@ -948,7 +948,7 @@ citerdissect(struct vars * v, } /* - * We need workspace to track the endpoints of each sub-match. Normally + * We need workspace to track the endpoints of each sub-match. Normally * we consider only nonzero-length sub-matches, so there can be at most * end-begin of them. However, if min is larger than that, we will also * consider zero-length sub-matches in order to find enough matches. @@ -977,8 +977,8 @@ citerdissect(struct vars * v, /* * Our strategy is to first find a set of sub-match endpoints that are * valid according to the child node's DFA, and then recursively dissect - * each sub-match to confirm validity. If any validity check fails, - * backtrack the last sub-match and try again. And, when we next try for + * each sub-match to confirm validity. If any validity check fails, + * backtrack the last sub-match and try again. And, when we next try for * a validity check, we need not recheck any successfully verified * sub-matches that we didn't move the endpoints of. nverified remembers * how many sub-matches are currently known okay. @@ -1028,10 +1028,10 @@ citerdissect(struct vars * v, } /* - * We've identified a way to divide the string into k sub-matches - * that works so far as the child DFA can tell. If k is an allowed - * number of matches, start the slow part: recurse to verify each - * sub-match. We always have k <= max_matches, needn't check that. + * We've identified a way to divide the string into k sub-matches that + * works so far as the child DFA can tell. If k is an allowed number + * of matches, start the slow part: recurse to verify each sub-match. + * We always have k <= max_matches, needn't check that. */ if (k < min_matches) goto backtrack; @@ -1065,13 +1065,14 @@ citerdissect(struct vars * v, /* match failed to verify, so backtrack */ backtrack: + /* * Must consider shorter versions of the current sub-match. However, * we'll only ask for a zero-length match if necessary. */ while (k > 0) { - chr *prev_end = endpts[k - 1]; + chr *prev_end = endpts[k - 1]; if (endpts[k] > prev_end) { @@ -1132,7 +1133,7 @@ creviterdissect(struct vars * v, } /* - * We need workspace to track the endpoints of each sub-match. Normally + * We need workspace to track the endpoints of each sub-match. Normally * we consider only nonzero-length sub-matches, so there can be at most * end-begin of them. However, if min is larger than that, we will also * consider zero-length sub-matches in order to find enough matches. @@ -1161,8 +1162,8 @@ creviterdissect(struct vars * v, /* * Our strategy is to first find a set of sub-match endpoints that are * valid according to the child node's DFA, and then recursively dissect - * each sub-match to confirm validity. If any validity check fails, - * backtrack the last sub-match and try again. And, when we next try for + * each sub-match to confirm validity. If any validity check fails, + * backtrack the last sub-match and try again. And, when we next try for * a validity check, we need not recheck any successfully verified * sub-matches that we didn't move the endpoints of. nverified remembers * how many sub-matches are currently known okay. @@ -1214,10 +1215,10 @@ creviterdissect(struct vars * v, } /* - * We've identified a way to divide the string into k sub-matches - * that works so far as the child DFA can tell. If k is an allowed - * number of matches, start the slow part: recurse to verify each - * sub-match. We always have k <= max_matches, needn't check that. + * We've identified a way to divide the string into k sub-matches that + * works so far as the child DFA can tell. If k is an allowed number + * of matches, start the slow part: recurse to verify each sub-match. + * We always have k <= max_matches, needn't check that. */ if (k < min_matches) goto backtrack; @@ -1251,6 +1252,7 @@ creviterdissect(struct vars * v, /* match failed to verify, so backtrack */ backtrack: + /* * Must consider longer versions of the current sub-match. */ diff --git a/src/backend/replication/basebackup.c b/src/backend/replication/basebackup.c index 72e79ce045248e1c5644f13cf6dd0dd0db60b921..0bc88a4040d6938c4619bb6ad9c42a6deea9a03e 100644 --- a/src/backend/replication/basebackup.c +++ b/src/backend/replication/basebackup.c @@ -141,6 +141,7 @@ perform_base_backup(basebackup_options *opt, DIR *tblspcdir) ti->size = opt->progress ? sendDir(linkpath, strlen(linkpath), true) : -1; tablespaces = lappend(tablespaces, ti); #else + /* * If the platform does not have symbolic links, it should not be * possible to have tablespaces - clearly somebody else created @@ -148,7 +149,7 @@ perform_base_backup(basebackup_options *opt, DIR *tblspcdir) */ ereport(WARNING, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("tablespaces are not supported on this platform"))); + errmsg("tablespaces are not supported on this platform"))); #endif } @@ -661,9 +662,9 @@ sendDir(char *path, int basepathlen, bool sizeonly) /* Allow symbolic links in pg_tblspc only */ if (strcmp(path, "./pg_tblspc") == 0 && #ifndef WIN32 - S_ISLNK(statbuf.st_mode) + S_ISLNK(statbuf.st_mode) #else - pgwin32_is_junction(pathbuf) + pgwin32_is_junction(pathbuf) #endif ) { @@ -687,6 +688,7 @@ sendDir(char *path, int basepathlen, bool sizeonly) _tarWriteHeader(pathbuf + basepathlen + 1, linkpath, &statbuf); size += 512; /* Size of the header just added */ #else + /* * If the platform does not have symbolic links, it should not be * possible to have tablespaces - clearly somebody else created @@ -694,9 +696,9 @@ sendDir(char *path, int basepathlen, bool sizeonly) */ ereport(WARNING, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("tablespaces are not supported on this platform"))); + errmsg("tablespaces are not supported on this platform"))); continue; -#endif /* HAVE_READLINK */ +#endif /* HAVE_READLINK */ } else if (S_ISDIR(statbuf.st_mode)) { diff --git a/src/backend/replication/syncrep.c b/src/backend/replication/syncrep.c index 8977327c8c658b70a9047b8754e61d35bd5e3a6c..a2ae8700d12a9cd1603a6775aa94385743a9a302 100644 --- a/src/backend/replication/syncrep.c +++ b/src/backend/replication/syncrep.c @@ -172,10 +172,10 @@ SyncRepWaitForLSN(XLogRecPtr XactCommitLSN) * never update it again, so we can't be seeing a stale value in that * case. * - * Note: on machines with weak memory ordering, the acquisition of - * the lock is essential to avoid race conditions: we cannot be sure - * the sender's state update has reached main memory until we acquire - * the lock. We could get rid of this dance if SetLatch/ResetLatch + * Note: on machines with weak memory ordering, the acquisition of the + * lock is essential to avoid race conditions: we cannot be sure the + * sender's state update has reached main memory until we acquire the + * lock. We could get rid of this dance if SetLatch/ResetLatch * contained memory barriers. */ syncRepState = MyProc->syncRepState; @@ -241,8 +241,8 @@ SyncRepWaitForLSN(XLogRecPtr XactCommitLSN) } /* - * Wait on latch. Any condition that should wake us up will set - * the latch, so no need for timeout. + * Wait on latch. Any condition that should wake us up will set the + * latch, so no need for timeout. */ WaitLatch(&MyProc->procLatch, WL_LATCH_SET | WL_POSTMASTER_DEATH, -1); } @@ -422,8 +422,8 @@ SyncRepReleaseWaiters(void) } /* - * Set the lsn first so that when we wake backends they will release - * up to this location. + * Set the lsn first so that when we wake backends they will release up to + * this location. */ if (XLByteLT(walsndctl->lsn[SYNC_REP_WAIT_WRITE], MyWalSnd->write)) { @@ -477,8 +477,8 @@ SyncRepGetStandbyPriority(void) bool found = false; /* - * Since synchronous cascade replication is not allowed, we always - * set the priority of cascading walsender to zero. + * Since synchronous cascade replication is not allowed, we always set the + * priority of cascading walsender to zero. */ if (am_cascading_walsender) return 0; @@ -517,7 +517,7 @@ SyncRepGetStandbyPriority(void) } /* - * Walk the specified queue from head. Set the state of any backends that + * Walk the specified queue from head. Set the state of any backends that * need to be woken, remove them from the queue, and then wake them. * Pass all = true to wake whole queue; otherwise, just wake up to * the walsender's LSN. @@ -601,7 +601,7 @@ SyncRepUpdateSyncStandbysDefined(void) */ if (!sync_standbys_defined) { - int i; + int i; for (i = 0; i < NUM_SYNC_REP_WAIT_MODE; i++) SyncRepWakeQueue(true, i); diff --git a/src/backend/replication/walreceiver.c b/src/backend/replication/walreceiver.c index d63ff29472cb49e5fe2c1a54499d4100fac354ac..650b74fff7d5ca5b6201abfce8efffd332f04d64 100644 --- a/src/backend/replication/walreceiver.c +++ b/src/backend/replication/walreceiver.c @@ -752,8 +752,8 @@ ProcessWalSndrMessage(XLogRecPtr walEnd, TimestampTz sendTime) if (log_min_messages <= DEBUG2) elog(DEBUG2, "sendtime %s receipttime %s replication apply delay %d ms transfer latency %d ms", - timestamptz_to_str(sendTime), - timestamptz_to_str(lastMsgReceiptTime), - GetReplicationApplyDelay(), - GetReplicationTransferLatency()); + timestamptz_to_str(sendTime), + timestamptz_to_str(lastMsgReceiptTime), + GetReplicationApplyDelay(), + GetReplicationTransferLatency()); } diff --git a/src/backend/replication/walreceiverfuncs.c b/src/backend/replication/walreceiverfuncs.c index f8dd52351858199075b6ce691584d655a0222ff5..876196f9da397fe578a5b1819f6d34ff9d563b3f 100644 --- a/src/backend/replication/walreceiverfuncs.c +++ b/src/backend/replication/walreceiverfuncs.c @@ -252,8 +252,8 @@ GetReplicationApplyDelay(void) XLogRecPtr receivePtr; XLogRecPtr replayPtr; - long secs; - int usecs; + long secs; + int usecs; SpinLockAcquire(&walrcv->mutex); receivePtr = walrcv->receivedUpto; @@ -284,9 +284,9 @@ GetReplicationTransferLatency(void) TimestampTz lastMsgSendTime; TimestampTz lastMsgReceiptTime; - long secs = 0; - int usecs = 0; - int ms; + long secs = 0; + int usecs = 0; + int ms; SpinLockAcquire(&walrcv->mutex); lastMsgSendTime = walrcv->lastMsgSendTime; diff --git a/src/backend/replication/walsender.c b/src/backend/replication/walsender.c index 5f938124e726627f61d0c6649972b8501c8fc263..45a3b2ef2941063d9901d1f060a73c2e3897c751 100644 --- a/src/backend/replication/walsender.c +++ b/src/backend/replication/walsender.c @@ -74,7 +74,8 @@ WalSnd *MyWalSnd = NULL; /* Global state */ bool am_walsender = false; /* Am I a walsender process ? */ -bool am_cascading_walsender = false; /* Am I cascading WAL to another standby ? */ +bool am_cascading_walsender = false; /* Am I cascading WAL to + * another standby ? */ /* User-settable parameters for walsender */ int max_wal_senders = 0; /* the maximum number of concurrent walsenders */ @@ -372,31 +373,31 @@ StartReplication(StartReplicationCmd *cmd) SendPostmasterSignal(PMSIGNAL_ADVANCE_STATE_MACHINE); /* - * When promoting a cascading standby, postmaster sends SIGUSR2 to - * any cascading walsenders to kill them. But there is a corner-case where - * such walsender fails to receive SIGUSR2 and survives a standby promotion - * unexpectedly. This happens when postmaster sends SIGUSR2 before - * the walsender marks itself as a WAL sender, because postmaster sends - * SIGUSR2 to only the processes marked as a WAL sender. + * When promoting a cascading standby, postmaster sends SIGUSR2 to any + * cascading walsenders to kill them. But there is a corner-case where + * such walsender fails to receive SIGUSR2 and survives a standby + * promotion unexpectedly. This happens when postmaster sends SIGUSR2 + * before the walsender marks itself as a WAL sender, because postmaster + * sends SIGUSR2 to only the processes marked as a WAL sender. * * To avoid this corner-case, if recovery is NOT in progress even though * the walsender is cascading one, we do the same thing as SIGUSR2 signal * handler does, i.e., set walsender_ready_to_stop to true. Which causes * the walsender to end later. * - * When terminating cascading walsenders, usually postmaster writes - * the log message announcing the terminations. But there is a race condition - * here. If there is no walsender except this process before reaching here, - * postmaster thinks that there is no walsender and suppresses that + * When terminating cascading walsenders, usually postmaster writes the + * log message announcing the terminations. But there is a race condition + * here. If there is no walsender except this process before reaching + * here, postmaster thinks that there is no walsender and suppresses that * log message. To handle this case, we always emit that log message here. - * This might cause duplicate log messages, but which is less likely to happen, - * so it's not worth writing some code to suppress them. + * This might cause duplicate log messages, but which is less likely to + * happen, so it's not worth writing some code to suppress them. */ if (am_cascading_walsender && !RecoveryInProgress()) { ereport(LOG, - (errmsg("terminating walsender process to force cascaded standby " - "to update timeline and reconnect"))); + (errmsg("terminating walsender process to force cascaded standby " + "to update timeline and reconnect"))); walsender_ready_to_stop = true; } @@ -405,8 +406,8 @@ StartReplication(StartReplicationCmd *cmd) * log-shipping, since this is checked in PostmasterMain(). * * NOTE: wal_level can only change at shutdown, so in most cases it is - * difficult for there to be WAL data that we can still see that was written - * at wal_level='minimal'. + * difficult for there to be WAL data that we can still see that was + * written at wal_level='minimal'. */ /* @@ -693,7 +694,7 @@ ProcessStandbyHSFeedbackMessage(void) * far enough to make reply.xmin wrap around. In that case the xmin we * set here would be "in the future" and have no effect. No point in * worrying about this since it's too late to save the desired data - * anyway. Assuming that the standby sends us an increasing sequence of + * anyway. Assuming that the standby sends us an increasing sequence of * xmins, this could only happen during the first reply cycle, else our * own xmin would prevent nextXid from advancing so far. * @@ -792,8 +793,8 @@ WalSndLoop(void) if (MyWalSnd->state == WALSNDSTATE_CATCHUP) { ereport(DEBUG1, - (errmsg("standby \"%s\" has now caught up with primary", - application_name))); + (errmsg("standby \"%s\" has now caught up with primary", + application_name))); WalSndSetState(WALSNDSTATE_STREAMING); } @@ -810,7 +811,7 @@ WalSndLoop(void) if (caughtup && !pq_is_send_pending()) { walsender_shutdown_requested = true; - continue; /* don't want to wait more */ + continue; /* don't want to wait more */ } } } @@ -825,7 +826,7 @@ WalSndLoop(void) if (caughtup || pq_is_send_pending()) { TimestampTz timeout = 0; - long sleeptime = 10000; /* 10 s */ + long sleeptime = 10000; /* 10 s */ int wakeEvents; wakeEvents = WL_LATCH_SET | WL_POSTMASTER_DEATH | @@ -845,7 +846,7 @@ WalSndLoop(void) if (replication_timeout > 0) { timeout = TimestampTzPlusMilliseconds(last_reply_timestamp, - replication_timeout); + replication_timeout); sleeptime = 1 + (replication_timeout / 10); } @@ -973,9 +974,9 @@ WalSndKill(int code, Datum arg) void XLogRead(char *buf, XLogRecPtr startptr, Size count) { - char *p; + char *p; XLogRecPtr recptr; - Size nbytes; + Size nbytes; uint32 lastRemovedLog; uint32 lastRemovedSeg; uint32 log; @@ -1087,9 +1088,9 @@ retry: } /* - * During recovery, the currently-open WAL file might be replaced with - * the file of the same name retrieved from archive. So we always need - * to check what we read was valid after reading into the buffer. If it's + * During recovery, the currently-open WAL file might be replaced with the + * file of the same name retrieved from archive. So we always need to + * check what we read was valid after reading into the buffer. If it's * invalid, we try to open and read the file again. */ if (am_cascading_walsender) @@ -1294,8 +1295,8 @@ WalSndShutdownHandler(SIGNAL_ARGS) SetLatch(&MyWalSnd->latch); /* - * Set the standard (non-walsender) state as well, so that we can - * abort things like do_pg_stop_backup(). + * Set the standard (non-walsender) state as well, so that we can abort + * things like do_pg_stop_backup(). */ InterruptPending = true; ProcDiePending = true; diff --git a/src/backend/rewrite/rewriteDefine.c b/src/backend/rewrite/rewriteDefine.c index 6e5633dcdb5200c5bce73936c6938d0a11b36908..971d277b7694a1fe689a654f742ee1c769cbd9d0 100644 --- a/src/backend/rewrite/rewriteDefine.c +++ b/src/backend/rewrite/rewriteDefine.c @@ -200,7 +200,7 @@ DefineRule(RuleStmt *stmt, const char *queryString) transformRuleStmt(stmt, queryString, &actions, &whereClause); /* - * Find and lock the relation. Lock level should match + * Find and lock the relation. Lock level should match * DefineQueryRewrite. */ relId = RangeVarGetRelid(stmt->relation, AccessExclusiveLock, false); diff --git a/src/backend/rewrite/rewriteSupport.c b/src/backend/rewrite/rewriteSupport.c index 5990159c62649d8198347a072de9a535f2360945..3f9b5e6063eaebb21f4be57de6c40ffa97aad680 100644 --- a/src/backend/rewrite/rewriteSupport.c +++ b/src/backend/rewrite/rewriteSupport.c @@ -168,8 +168,8 @@ get_rewrite_oid_without_relid(const char *rulename, if (HeapTupleIsValid(htup)) ereport(ERROR, (errcode(ERRCODE_DUPLICATE_OBJECT), - errmsg("there are multiple rules named \"%s\"", rulename), - errhint("Specify a relation name as well as a rule name."))); + errmsg("there are multiple rules named \"%s\"", rulename), + errhint("Specify a relation name as well as a rule name."))); } heap_endscan(scanDesc); heap_close(RewriteRelation, AccessShareLock); diff --git a/src/backend/storage/buffer/bufmgr.c b/src/backend/storage/buffer/bufmgr.c index d46faaf958d4310da39522c4ada193bedfd62df1..78145472e169decd91557181367b4cc68af84578 100644 --- a/src/backend/storage/buffer/bufmgr.c +++ b/src/backend/storage/buffer/bufmgr.c @@ -1325,7 +1325,7 @@ BufferSync(int flags) * This is called periodically by the background writer process. * * Returns true if it's appropriate for the bgwriter process to go into - * low-power hibernation mode. (This happens if the strategy clock sweep + * low-power hibernation mode. (This happens if the strategy clock sweep * has been "lapped" and no buffer allocations have occurred recently, * or if the bgwriter has been effectively disabled by setting * bgwriter_lru_maxpages to 0.) @@ -1510,8 +1510,8 @@ BgBufferSync(void) /* * If recent_alloc remains at zero for many cycles, smoothed_alloc will * eventually underflow to zero, and the underflows produce annoying - * kernel warnings on some platforms. Once upcoming_alloc_est has gone - * to zero, there's no point in tracking smaller and smaller values of + * kernel warnings on some platforms. Once upcoming_alloc_est has gone to + * zero, there's no point in tracking smaller and smaller values of * smoothed_alloc, so just reset it to exactly zero to avoid this * syndrome. It will pop back up as soon as recent_alloc increases. */ @@ -2006,11 +2006,11 @@ BufferIsPermanent(Buffer buffer) Assert(BufferIsPinned(buffer)); /* - * BM_PERMANENT can't be changed while we hold a pin on the buffer, so - * we need not bother with the buffer header spinlock. Even if someone - * else changes the buffer header flags while we're doing this, we assume - * that changing an aligned 2-byte BufFlags value is atomic, so we'll read - * the old value or the new value, but not random garbage. + * BM_PERMANENT can't be changed while we hold a pin on the buffer, so we + * need not bother with the buffer header spinlock. Even if someone else + * changes the buffer header flags while we're doing this, we assume that + * changing an aligned 2-byte BufFlags value is atomic, so we'll read the + * old value or the new value, but not random garbage. */ bufHdr = &BufferDescriptors[buffer - 1]; return (bufHdr->flags & BM_PERMANENT) != 0; @@ -2461,10 +2461,10 @@ SetBufferCommitInfoNeedsSave(Buffer buffer) * tuples. So, be as quick as we can if the buffer is already dirty. We * do this by not acquiring spinlock if it looks like the status bits are * already. Since we make this test unlocked, there's a chance we might - * fail to notice that the flags have just been cleared, and failed to reset - * them, due to memory-ordering issues. But since this function is only - * intended to be used in cases where failing to write out the data would - * be harmless anyway, it doesn't really matter. + * fail to notice that the flags have just been cleared, and failed to + * reset them, due to memory-ordering issues. But since this function is + * only intended to be used in cases where failing to write out the data + * would be harmless anyway, it doesn't really matter. */ if ((bufHdr->flags & (BM_DIRTY | BM_JUST_DIRTIED)) != (BM_DIRTY | BM_JUST_DIRTIED)) diff --git a/src/backend/storage/buffer/freelist.c b/src/backend/storage/buffer/freelist.c index 76a4beca69924876c1b4902feca790ec7b8e82ae..c92774798cfc57d3f35b5a959d384b8ace77d1c7 100644 --- a/src/backend/storage/buffer/freelist.c +++ b/src/backend/storage/buffer/freelist.c @@ -294,7 +294,7 @@ StrategySyncStart(uint32 *complete_passes, uint32 *num_buf_alloc) * StrategyNotifyBgWriter -- set or clear allocation notification latch * * If bgwriterLatch isn't NULL, the next invocation of StrategyGetBuffer will - * set that latch. Pass NULL to clear the pending notification before it + * set that latch. Pass NULL to clear the pending notification before it * happens. This feature is used by the bgwriter process to wake itself up * from hibernation, and is not meant for anybody else to use. */ diff --git a/src/backend/storage/file/fd.c b/src/backend/storage/file/fd.c index fa376ae4bb3d1a56669552e18b174beea9688ef8..f79f4c6a36e7b427d86ddc14d452a4723cf66c01 100644 --- a/src/backend/storage/file/fd.c +++ b/src/backend/storage/file/fd.c @@ -164,7 +164,7 @@ static bool have_pending_fd_cleanup = false; /* * Tracks the total size of all temporary files. Note: when temp_file_limit * is being enforced, this cannot overflow since the limit cannot be more - * than INT_MAX kilobytes. When not enforcing, it could theoretically + * than INT_MAX kilobytes. When not enforcing, it could theoretically * overflow, but we don't care. */ static uint64 temporary_files_size = 0; @@ -685,7 +685,7 @@ LruInsert(File file) /* seek to the right position */ if (vfdP->seekPos != (off_t) 0) { - off_t returnValue PG_USED_FOR_ASSERTS_ONLY; + off_t returnValue PG_USED_FOR_ASSERTS_ONLY; returnValue = lseek(vfdP->fd, vfdP->seekPos, SEEK_SET); Assert(returnValue != (off_t) -1); @@ -1046,7 +1046,7 @@ OpenTemporaryFileInTablespace(Oid tblspcOid, bool rejectError) void FileSetTransient(File file) { - Vfd *vfdP; + Vfd *vfdP; Assert(FileIsValid(file)); @@ -1255,7 +1255,7 @@ FileWrite(File file, char *buffer, int amount) /* * If enforcing temp_file_limit and it's a temp file, check to see if the - * write would overrun temp_file_limit, and throw error if so. Note: it's + * write would overrun temp_file_limit, and throw error if so. Note: it's * really a modularity violation to throw error here; we should set errno * and return -1. However, there's no way to report a suitable error * message if we do that. All current callers would just throw error @@ -1263,18 +1263,18 @@ FileWrite(File file, char *buffer, int amount) */ if (temp_file_limit >= 0 && (VfdCache[file].fdstate & FD_TEMPORARY)) { - off_t newPos = VfdCache[file].seekPos + amount; + off_t newPos = VfdCache[file].seekPos + amount; if (newPos > VfdCache[file].fileSize) { - uint64 newTotal = temporary_files_size; + uint64 newTotal = temporary_files_size; newTotal += newPos - VfdCache[file].fileSize; if (newTotal > (uint64) temp_file_limit * (uint64) 1024) ereport(ERROR, (errcode(ERRCODE_CONFIGURATION_LIMIT_EXCEEDED), - errmsg("temporary file size exceeds temp_file_limit (%dkB)", - temp_file_limit))); + errmsg("temporary file size exceeds temp_file_limit (%dkB)", + temp_file_limit))); } } @@ -1293,7 +1293,7 @@ retry: /* maintain fileSize and temporary_files_size if it's a temp file */ if (VfdCache[file].fdstate & FD_TEMPORARY) { - off_t newPos = VfdCache[file].seekPos; + off_t newPos = VfdCache[file].seekPos; if (newPos > VfdCache[file].fileSize) { @@ -1915,8 +1915,8 @@ CleanupTempFiles(bool isProcExit) /* * If we're in the process of exiting a backend process, * close all temporary files. Otherwise, only close - * temporary files local to the current transaction. - * They should be closed by the ResourceOwner mechanism + * temporary files local to the current transaction. They + * should be closed by the ResourceOwner mechanism * already, so this is just a debugging cross-check. */ if (isProcExit) @@ -1924,7 +1924,7 @@ CleanupTempFiles(bool isProcExit) else if (fdstate & FD_XACT_TEMPORARY) { elog(WARNING, - "temporary file %s not closed at end-of-transaction", + "temporary file %s not closed at end-of-transaction", VfdCache[i].fileName); FileClose(i); } diff --git a/src/backend/storage/ipc/pmsignal.c b/src/backend/storage/ipc/pmsignal.c index 83842d6494c8385f898aaa459e9096760d46d381..8ad7a97eeb139c97bcf96cc2d5f2f12b541e21b3 100644 --- a/src/backend/storage/ipc/pmsignal.c +++ b/src/backend/storage/ipc/pmsignal.c @@ -272,8 +272,8 @@ bool PostmasterIsAlive(void) { #ifndef WIN32 - char c; - ssize_t rc; + char c; + ssize_t rc; rc = read(postmaster_alive_fds[POSTMASTER_FD_WATCH], &c, 1); if (rc < 0) @@ -287,7 +287,6 @@ PostmasterIsAlive(void) elog(FATAL, "unexpected data in postmaster death monitoring pipe"); return false; - #else /* WIN32 */ return (WaitForSingleObject(PostmasterHandle, 0) == WAIT_TIMEOUT); #endif /* WIN32 */ diff --git a/src/backend/storage/ipc/procarray.c b/src/backend/storage/ipc/procarray.c index d986418a10a88dde0bd7a8b90e12f11ea8c6827b..a3b0540aea73655014c9bcaddb83890a498d9acc 100644 --- a/src/backend/storage/ipc/procarray.c +++ b/src/backend/storage/ipc/procarray.c @@ -82,10 +82,10 @@ typedef struct ProcArrayStruct TransactionId lastOverflowedXid; /* - * We declare pgprocnos[] as 1 entry because C wants a fixed-size array, but - * actually it is maxProcs entries long. + * We declare pgprocnos[] as 1 entry because C wants a fixed-size array, + * but actually it is maxProcs entries long. */ - int pgprocnos[1]; /* VARIABLE LENGTH ARRAY */ + int pgprocnos[1]; /* VARIABLE LENGTH ARRAY */ } ProcArrayStruct; static ProcArrayStruct *procArray; @@ -282,22 +282,22 @@ ProcArrayAdd(PGPROC *proc) * locality of references much better. This is useful while traversing the * ProcArray because there is a increased likelihood of finding the next * PGPROC structure in the cache. - * + * * Since the occurrence of adding/removing a proc is much lower than the * access to the ProcArray itself, the overhead should be marginal */ for (index = 0; index < arrayP->numProcs; index++) { /* - * If we are the first PGPROC or if we have found our right position in - * the array, break + * If we are the first PGPROC or if we have found our right position + * in the array, break */ if ((arrayP->pgprocnos[index] == -1) || (arrayP->pgprocnos[index] > proc->pgprocno)) break; } memmove(&arrayP->pgprocnos[index + 1], &arrayP->pgprocnos[index], - (arrayP->numProcs - index) * sizeof (int)); + (arrayP->numProcs - index) * sizeof(int)); arrayP->pgprocnos[index] = proc->pgprocno; arrayP->numProcs++; @@ -349,8 +349,8 @@ ProcArrayRemove(PGPROC *proc, TransactionId latestXid) { /* Keep the PGPROC array sorted. See notes above */ memmove(&arrayP->pgprocnos[index], &arrayP->pgprocnos[index + 1], - (arrayP->numProcs - index - 1) * sizeof (int)); - arrayP->pgprocnos[arrayP->numProcs - 1] = -1; /* for debugging */ + (arrayP->numProcs - index - 1) * sizeof(int)); + arrayP->pgprocnos[arrayP->numProcs - 1] = -1; /* for debugging */ arrayP->numProcs--; LWLockRelease(ProcArrayLock); return; @@ -380,7 +380,7 @@ ProcArrayRemove(PGPROC *proc, TransactionId latestXid) void ProcArrayEndTransaction(PGPROC *proc, TransactionId latestXid) { - PGXACT *pgxact = &allPgXact[proc->pgprocno]; + PGXACT *pgxact = &allPgXact[proc->pgprocno]; if (TransactionIdIsValid(latestXid)) { @@ -399,7 +399,7 @@ ProcArrayEndTransaction(PGPROC *proc, TransactionId latestXid) pgxact->xmin = InvalidTransactionId; /* must be cleared with xid/xmin: */ pgxact->vacuumFlags &= ~PROC_VACUUM_STATE_MASK; - pgxact->inCommit = false; /* be sure this is cleared in abort */ + pgxact->inCommit = false; /* be sure this is cleared in abort */ proc->recoveryConflictPending = false; /* Clear the subtransaction-XID cache too while holding the lock */ @@ -426,7 +426,7 @@ ProcArrayEndTransaction(PGPROC *proc, TransactionId latestXid) pgxact->xmin = InvalidTransactionId; /* must be cleared with xid/xmin: */ pgxact->vacuumFlags &= ~PROC_VACUUM_STATE_MASK; - pgxact->inCommit = false; /* be sure this is cleared in abort */ + pgxact->inCommit = false; /* be sure this is cleared in abort */ proc->recoveryConflictPending = false; Assert(pgxact->nxids == 0); @@ -446,7 +446,7 @@ ProcArrayEndTransaction(PGPROC *proc, TransactionId latestXid) void ProcArrayClearTransaction(PGPROC *proc) { - PGXACT *pgxact = &allPgXact[proc->pgprocno]; + PGXACT *pgxact = &allPgXact[proc->pgprocno]; /* * We can skip locking ProcArrayLock here, because this action does not @@ -511,9 +511,9 @@ ProcArrayApplyRecoveryInfo(RunningTransactions running) /* * If our initial RunningTransactionsData had an overflowed snapshot then * we knew we were missing some subxids from our snapshot. If we continue - * to see overflowed snapshots then we might never be able to start up, - * so we make another test to see if our snapshot is now valid. We know - * that the missing subxids are equal to or earlier than nextXid. After we + * to see overflowed snapshots then we might never be able to start up, so + * we make another test to see if our snapshot is now valid. We know that + * the missing subxids are equal to or earlier than nextXid. After we * initialise we continue to apply changes during recovery, so once the * oldestRunningXid is later than the nextXid from the initial snapshot we * know that we no longer have missing information and can mark the @@ -522,8 +522,8 @@ ProcArrayApplyRecoveryInfo(RunningTransactions running) if (standbyState == STANDBY_SNAPSHOT_PENDING) { /* - * If the snapshot isn't overflowed or if its empty we can - * reset our pending state and use this snapshot instead. + * If the snapshot isn't overflowed or if its empty we can reset our + * pending state and use this snapshot instead. */ if (!running->subxid_overflow || running->xcnt == 0) { @@ -545,8 +545,8 @@ ProcArrayApplyRecoveryInfo(RunningTransactions running) } else elog(trace_recovery(DEBUG1), - "recovery snapshot waiting for non-overflowed snapshot or " - "until oldest active xid on standby is at least %u (now %u)", + "recovery snapshot waiting for non-overflowed snapshot or " + "until oldest active xid on standby is at least %u (now %u)", standbySnapshotPendingXmin, running->oldestRunningXid); return; @@ -673,7 +673,7 @@ ProcArrayApplyRecoveryInfo(RunningTransactions running) * ShmemVariableCache->nextXid must be beyond any observed xid. * * We don't expect anyone else to modify nextXid, hence we don't need to - * hold a lock while examining it. We still acquire the lock to modify + * hold a lock while examining it. We still acquire the lock to modify * it, though. */ nextXid = latestObservedXid; @@ -861,10 +861,10 @@ TransactionIdIsInProgress(TransactionId xid) /* No shortcuts, gotta grovel through the array */ for (i = 0; i < arrayP->numProcs; i++) { - int pgprocno = arrayP->pgprocnos[i]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; - TransactionId pxid; + int pgprocno = arrayP->pgprocnos[i]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; + TransactionId pxid; /* Ignore my own proc --- dealt with it above */ if (proc == MyProc) @@ -1017,10 +1017,10 @@ TransactionIdIsActive(TransactionId xid) for (i = 0; i < arrayP->numProcs; i++) { - int pgprocno = arrayP->pgprocnos[i]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; - TransactionId pxid; + int pgprocno = arrayP->pgprocnos[i]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; + TransactionId pxid; /* Fetch xid just once - see GetNewTransactionId */ pxid = pgxact->xid; @@ -1115,9 +1115,9 @@ GetOldestXmin(bool allDbs, bool ignoreVacuum) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; if (ignoreVacuum && (pgxact->vacuumFlags & PROC_IN_VACUUM)) continue; @@ -1141,7 +1141,7 @@ GetOldestXmin(bool allDbs, bool ignoreVacuum) * have an Xmin but not (yet) an Xid; conversely, if it has an * Xid, that could determine some not-yet-set Xmin. */ - xid = pgxact->xmin; /* Fetch just once */ + xid = pgxact->xmin; /* Fetch just once */ if (TransactionIdIsNormal(xid) && TransactionIdPrecedes(xid, result)) result = xid; @@ -1318,7 +1318,7 @@ GetSnapshotData(Snapshot snapshot) if (!snapshot->takenDuringRecovery) { - int *pgprocnos = arrayP->pgprocnos; + int *pgprocnos = arrayP->pgprocnos; int numProcs; /* @@ -1329,32 +1329,32 @@ GetSnapshotData(Snapshot snapshot) numProcs = arrayP->numProcs; for (index = 0; index < numProcs; index++) { - int pgprocno = pgprocnos[index]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; - TransactionId xid; + int pgprocno = pgprocnos[index]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; + TransactionId xid; /* Ignore procs running LAZY VACUUM */ if (pgxact->vacuumFlags & PROC_IN_VACUUM) continue; /* Update globalxmin to be the smallest valid xmin */ - xid = pgxact->xmin; /* fetch just once */ + xid = pgxact->xmin; /* fetch just once */ if (TransactionIdIsNormal(xid) && NormalTransactionIdPrecedes(xid, globalxmin)) - globalxmin = xid; + globalxmin = xid; /* Fetch xid just once - see GetNewTransactionId */ xid = pgxact->xid; /* - * If the transaction has no XID assigned, we can skip it; it won't - * have sub-XIDs either. If the XID is >= xmax, we can also skip - * it; such transactions will be treated as running anyway (and any - * sub-XIDs will also be >= xmax). + * If the transaction has no XID assigned, we can skip it; it + * won't have sub-XIDs either. If the XID is >= xmax, we can also + * skip it; such transactions will be treated as running anyway + * (and any sub-XIDs will also be >= xmax). */ if (!TransactionIdIsNormal(xid) || !NormalTransactionIdPrecedes(xid, xmax)) - continue; + continue; /* * We don't include our own XIDs (if any) in the snapshot, but we @@ -1394,6 +1394,7 @@ GetSnapshotData(Snapshot snapshot) if (nxids > 0) { volatile PGPROC *proc = &allProcs[pgprocno]; + memcpy(snapshot->subxip + subcount, (void *) proc->subxids.xids, nxids * sizeof(TransactionId)); @@ -1498,23 +1499,23 @@ ProcArrayInstallImportedXmin(TransactionId xmin, TransactionId sourcexid) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; - TransactionId xid; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; + TransactionId xid; /* Ignore procs running LAZY VACUUM */ if (pgxact->vacuumFlags & PROC_IN_VACUUM) continue; - xid = pgxact->xid; /* fetch just once */ + xid = pgxact->xid; /* fetch just once */ if (xid != sourcexid) continue; /* - * We check the transaction's database ID for paranoia's sake: if - * it's in another DB then its xmin does not cover us. Caller should - * have detected this already, so we just treat any funny cases as + * We check the transaction's database ID for paranoia's sake: if it's + * in another DB then its xmin does not cover us. Caller should have + * detected this already, so we just treat any funny cases as * "transaction not found". */ if (proc->databaseId != MyDatabaseId) @@ -1523,7 +1524,7 @@ ProcArrayInstallImportedXmin(TransactionId xmin, TransactionId sourcexid) /* * Likewise, let's just make real sure its xmin does cover us. */ - xid = pgxact->xmin; /* fetch just once */ + xid = pgxact->xmin; /* fetch just once */ if (!TransactionIdIsNormal(xid) || !TransactionIdPrecedesOrEquals(xid, xmin)) continue; @@ -1531,8 +1532,8 @@ ProcArrayInstallImportedXmin(TransactionId xmin, TransactionId sourcexid) /* * We're good. Install the new xmin. As in GetSnapshotData, set * TransactionXmin too. (Note that because snapmgr.c called - * GetSnapshotData first, we'll be overwriting a valid xmin here, - * so we don't check that.) + * GetSnapshotData first, we'll be overwriting a valid xmin here, so + * we don't check that.) */ MyPgXact->xmin = TransactionXmin = xmin; @@ -1626,7 +1627,7 @@ GetRunningTransactionData(void) */ for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGPROC *proc = &allProcs[pgprocno]; volatile PGXACT *pgxact = &allPgXact[pgprocno]; TransactionId xid; @@ -1726,7 +1727,7 @@ GetOldestActiveTransactionId(void) */ for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGXACT *pgxact = &allPgXact[pgprocno]; TransactionId xid; @@ -1783,7 +1784,7 @@ GetTransactionsInCommit(TransactionId **xids_p) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGXACT *pgxact = &allPgXact[pgprocno]; TransactionId pxid; @@ -1820,9 +1821,9 @@ HaveTransactionsInCommit(TransactionId *xids, int nxids) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; - TransactionId pxid; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; + TransactionId pxid; /* Fetch xid just once - see GetNewTransactionId */ pxid = pgxact->xid; @@ -1911,9 +1912,9 @@ BackendXidGetPid(TransactionId xid) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; if (pgxact->xid == xid) { @@ -1981,9 +1982,9 @@ GetCurrentVirtualXIDs(TransactionId limitXmin, bool excludeXmin0, for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; if (proc == MyProc) continue; @@ -2078,9 +2079,9 @@ GetConflictingVirtualXIDs(TransactionId limitXmin, Oid dbOid) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; /* Exclude prepared transactions */ if (proc->pid == 0) @@ -2134,9 +2135,9 @@ CancelVirtualTransaction(VirtualTransactionId vxid, ProcSignalReason sigmode) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - VirtualTransactionId procvxid; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + VirtualTransactionId procvxid; GET_VXID_FROM_PGPROC(procvxid, *proc); @@ -2189,9 +2190,9 @@ MinimumActiveBackends(int min) */ for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; - volatile PGPROC *proc = &allProcs[pgprocno]; - volatile PGXACT *pgxact = &allPgXact[pgprocno]; + int pgprocno = arrayP->pgprocnos[index]; + volatile PGPROC *proc = &allProcs[pgprocno]; + volatile PGXACT *pgxact = &allPgXact[pgprocno]; /* * Since we're not holding a lock, need to check that the pointer is @@ -2237,7 +2238,7 @@ CountDBBackends(Oid databaseid) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGPROC *proc = &allProcs[pgprocno]; if (proc->pid == 0) @@ -2267,7 +2268,7 @@ CancelDBBackends(Oid databaseid, ProcSignalReason sigmode, bool conflictPending) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGPROC *proc = &allProcs[pgprocno]; if (databaseid == InvalidOid || proc->databaseId == databaseid) @@ -2306,7 +2307,7 @@ CountUserBackends(Oid roleid) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGPROC *proc = &allProcs[pgprocno]; if (proc->pid == 0) @@ -2367,7 +2368,7 @@ CountOtherDBBackends(Oid databaseId, int *nbackends, int *nprepared) for (index = 0; index < arrayP->numProcs; index++) { - int pgprocno = arrayP->pgprocnos[index]; + int pgprocno = arrayP->pgprocnos[index]; volatile PGPROC *proc = &allProcs[pgprocno]; volatile PGXACT *pgxact = &allPgXact[pgprocno]; diff --git a/src/backend/storage/ipc/sinval.c b/src/backend/storage/ipc/sinval.c index cd15a2e9cd4ccbd0e9aa6ebf6e02f48ddf680fbe..6d070030b7ad5a1b97192c01402b5bea64879adc 100644 --- a/src/backend/storage/ipc/sinval.c +++ b/src/backend/storage/ipc/sinval.c @@ -22,7 +22,7 @@ #include "utils/inval.h" -uint64 SharedInvalidMessageCounter; +uint64 SharedInvalidMessageCounter; /* diff --git a/src/backend/storage/ipc/sinvaladt.c b/src/backend/storage/ipc/sinvaladt.c index bb727af8b2a6e6fb6a6c6a6df4e845e3712764e5..ec0153e115ae257d8af83b4def7232d27e03b0d8 100644 --- a/src/backend/storage/ipc/sinvaladt.c +++ b/src/backend/storage/ipc/sinvaladt.c @@ -467,15 +467,16 @@ SIInsertDataEntries(const SharedInvalidationMessage *data, int n) } /* - * Now that the maxMsgNum change is globally visible, we give - * everyone a swift kick to make sure they read the newly added - * messages. Releasing SInvalWriteLock will enforce a full memory - * barrier, so these (unlocked) changes will be committed to memory - * before we exit the function. + * Now that the maxMsgNum change is globally visible, we give everyone + * a swift kick to make sure they read the newly added messages. + * Releasing SInvalWriteLock will enforce a full memory barrier, so + * these (unlocked) changes will be committed to memory before we exit + * the function. */ for (i = 0; i < segP->lastBackend; i++) { ProcState *stateP = &segP->procState[i]; + stateP->hasMessages = true; } @@ -524,12 +525,12 @@ SIGetDataEntries(SharedInvalidationMessage *data, int datasize) /* * Before starting to take locks, do a quick, unlocked test to see whether - * there can possibly be anything to read. On a multiprocessor system, - * it's possible that this load could migrate backwards and occur before we - * actually enter this function, so we might miss a sinval message that - * was just added by some other processor. But they can't migrate - * backwards over a preceding lock acquisition, so it should be OK. If - * we haven't acquired a lock preventing against further relevant + * there can possibly be anything to read. On a multiprocessor system, + * it's possible that this load could migrate backwards and occur before + * we actually enter this function, so we might miss a sinval message that + * was just added by some other processor. But they can't migrate + * backwards over a preceding lock acquisition, so it should be OK. If we + * haven't acquired a lock preventing against further relevant * invalidations, any such occurrence is not much different than if the * invalidation had arrived slightly later in the first place. */ diff --git a/src/backend/storage/ipc/standby.c b/src/backend/storage/ipc/standby.c index 3a6831cab0c30129e31d63b2c947241f4d96fb61..8863a5c9a7f57397b4c7bccddf91593e1d2dacfc 100644 --- a/src/backend/storage/ipc/standby.c +++ b/src/backend/storage/ipc/standby.c @@ -467,7 +467,7 @@ SendRecoveryConflictWithBufferPin(ProcSignalReason reason) * determine whether an actual deadlock condition is present: the lock we * need to wait for might be unrelated to any held by the Startup process. * Sooner or later, this mechanism should get ripped out in favor of somehow - * accounting for buffer locks in DeadLockCheck(). However, errors here + * accounting for buffer locks in DeadLockCheck(). However, errors here * seem to be very low-probability in practice, so for now it's not worth * the trouble. */ @@ -658,7 +658,7 @@ StandbyReleaseOldLocks(int nxids, TransactionId *xids) for (cell = list_head(RecoveryLockList); cell; cell = next) { xl_standby_lock *lock = (xl_standby_lock *) lfirst(cell); - bool remove = false; + bool remove = false; next = lnext(cell); @@ -668,8 +668,8 @@ StandbyReleaseOldLocks(int nxids, TransactionId *xids) remove = false; else { - int i; - bool found = false; + int i; + bool found = false; for (i = 0; i < nxids; i++) { @@ -1009,8 +1009,8 @@ LogAccessExclusiveLockPrepare(void) * RecordTransactionAbort() do not optimise away the transaction * completion record which recovery relies upon to release locks. It's a * hack, but for a corner case not worth adding code for into the main - * commit path. Second, we must assign an xid before the lock is - * recorded in shared memory, otherwise a concurrently executing + * commit path. Second, we must assign an xid before the lock is recorded + * in shared memory, otherwise a concurrently executing * GetRunningTransactionLocks() might see a lock associated with an * InvalidTransactionId which we later assert cannot happen. */ diff --git a/src/backend/storage/lmgr/lock.c b/src/backend/storage/lmgr/lock.c index 9717075354fa476bf09b366a5f4e59a43e12bdb5..cfe3954637d723822b95ec4e64e457e193bb0854 100644 --- a/src/backend/storage/lmgr/lock.c +++ b/src/backend/storage/lmgr/lock.c @@ -164,7 +164,7 @@ typedef struct TwoPhaseLockRecord * our locks to the primary lock table, but it can never be lower than the * real value, since only we can acquire locks on our own behalf. */ -static int FastPathLocalUseCount = 0; +static int FastPathLocalUseCount = 0; /* Macros for manipulating proc->fpLockBits */ #define FAST_PATH_BITS_PER_SLOT 3 @@ -186,7 +186,7 @@ static int FastPathLocalUseCount = 0; /* * The fast-path lock mechanism is concerned only with relation locks on - * unshared relations by backends bound to a database. The fast-path + * unshared relations by backends bound to a database. The fast-path * mechanism exists mostly to accelerate acquisition and release of locks * that rarely conflict. Because ShareUpdateExclusiveLock is * self-conflicting, it can't use the fast-path mechanism; but it also does @@ -207,7 +207,7 @@ static int FastPathLocalUseCount = 0; static bool FastPathGrantRelationLock(Oid relid, LOCKMODE lockmode); static bool FastPathUnGrantRelationLock(Oid relid, LOCKMODE lockmode); static bool FastPathTransferRelationLocks(LockMethod lockMethodTable, - const LOCKTAG *locktag, uint32 hashcode); + const LOCKTAG *locktag, uint32 hashcode); static PROCLOCK *FastPathGetRelationLockEntry(LOCALLOCK *locallock); static void VirtualXactLockTableCleanup(void); @@ -234,8 +234,8 @@ static void VirtualXactLockTableCleanup(void); typedef struct { - slock_t mutex; - uint32 count[FAST_PATH_STRONG_LOCK_HASH_PARTITIONS]; + slock_t mutex; + uint32 count[FAST_PATH_STRONG_LOCK_HASH_PARTITIONS]; } FastPathStrongRelationLockData; FastPathStrongRelationLockData *FastPathStrongRelationLocks; @@ -339,7 +339,7 @@ PROCLOCK_PRINT(const char *where, const PROCLOCK *proclockP) static uint32 proclock_hash(const void *key, Size keysize); static void RemoveLocalLock(LOCALLOCK *locallock); static PROCLOCK *SetupLockInTable(LockMethod lockMethodTable, PGPROC *proc, - const LOCKTAG *locktag, uint32 hashcode, LOCKMODE lockmode); + const LOCKTAG *locktag, uint32 hashcode, LOCKMODE lockmode); static void GrantLockLocal(LOCALLOCK *locallock, ResourceOwner owner); static void BeginStrongLockAcquire(LOCALLOCK *locallock, uint32 fasthashcode); static void FinishStrongLockAcquire(void); @@ -425,7 +425,7 @@ InitLocks(void) */ FastPathStrongRelationLocks = ShmemInitStruct("Fast Path Strong Relation Lock Data", - sizeof(FastPathStrongRelationLockData), &found); + sizeof(FastPathStrongRelationLockData), &found); if (!found) SpinLockInit(&FastPathStrongRelationLocks->mutex); @@ -713,12 +713,12 @@ LockAcquireExtended(const LOCKTAG *locktag, if (EligibleForRelationFastPath(locktag, lockmode) && FastPathLocalUseCount < FP_LOCK_SLOTS_PER_BACKEND) { - uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); - bool acquired; + uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); + bool acquired; /* - * LWLockAcquire acts as a memory sequencing point, so it's safe - * to assume that any strong locker whose increment to + * LWLockAcquire acts as a memory sequencing point, so it's safe to + * assume that any strong locker whose increment to * FastPathStrongRelationLocks->counts becomes visible after we test * it has yet to begin to transfer fast-path locks. */ @@ -744,7 +744,7 @@ LockAcquireExtended(const LOCKTAG *locktag, */ if (ConflictsWithRelationFastPath(locktag, lockmode)) { - uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); + uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); BeginStrongLockAcquire(locallock, fasthashcode); if (!FastPathTransferRelationLocks(lockMethodTable, locktag, @@ -762,9 +762,9 @@ LockAcquireExtended(const LOCKTAG *locktag, } /* - * We didn't find the lock in our LOCALLOCK table, and we didn't manage - * to take it via the fast-path, either, so we've got to mess with the - * shared lock table. + * We didn't find the lock in our LOCALLOCK table, and we didn't manage to + * take it via the fast-path, either, so we've got to mess with the shared + * lock table. */ partitionLock = LockHashPartitionLock(hashcode); @@ -1102,7 +1102,8 @@ RemoveLocalLock(LOCALLOCK *locallock) locallock->lockOwners = NULL; if (locallock->holdsStrongLockCount) { - uint32 fasthashcode; + uint32 fasthashcode; + fasthashcode = FastPathStrongLockHashPartition(locallock->hashcode); SpinLockAcquire(&FastPathStrongRelationLocks->mutex); @@ -1367,9 +1368,9 @@ BeginStrongLockAcquire(LOCALLOCK *locallock, uint32 fasthashcode) Assert(locallock->holdsStrongLockCount == FALSE); /* - * Adding to a memory location is not atomic, so we take a - * spinlock to ensure we don't collide with someone else trying - * to bump the count at the same time. + * Adding to a memory location is not atomic, so we take a spinlock to + * ensure we don't collide with someone else trying to bump the count at + * the same time. * * XXX: It might be worth considering using an atomic fetch-and-add * instruction here, on architectures where that is supported. @@ -1399,9 +1400,9 @@ FinishStrongLockAcquire(void) void AbortStrongLockAcquire(void) { - uint32 fasthashcode; + uint32 fasthashcode; LOCALLOCK *locallock = StrongLockInProgress; - + if (locallock == NULL) return; @@ -1699,11 +1700,11 @@ LockRelease(const LOCKTAG *locktag, LOCKMODE lockmode, bool sessionLock) if (EligibleForRelationFastPath(locktag, lockmode) && FastPathLocalUseCount > 0) { - bool released; + bool released; /* - * We might not find the lock here, even if we originally entered - * it here. Another backend may have moved it to the main table. + * We might not find the lock here, even if we originally entered it + * here. Another backend may have moved it to the main table. */ LWLockAcquire(MyProc->backendLock, LW_EXCLUSIVE); released = FastPathUnGrantRelationLock(locktag->locktag_field2, @@ -1816,8 +1817,8 @@ LockReleaseAll(LOCKMETHODID lockmethodid, bool allLocks) #endif /* - * Get rid of our fast-path VXID lock, if appropriate. Note that this - * is the only way that the lock we hold on our own VXID can ever get + * Get rid of our fast-path VXID lock, if appropriate. Note that this is + * the only way that the lock we hold on our own VXID can ever get * released: it is always and only released when a toplevel transaction * ends. */ @@ -1898,8 +1899,8 @@ LockReleaseAll(LOCKMETHODID lockmethodid, bool allLocks) /* * If we don't currently hold the LWLock that protects our - * fast-path data structures, we must acquire it before - * attempting to release the lock via the fast-path. + * fast-path data structures, we must acquire it before attempting + * to release the lock via the fast-path. */ if (!have_fast_path_lwlock) { @@ -1917,7 +1918,7 @@ LockReleaseAll(LOCKMETHODID lockmethodid, bool allLocks) /* * Our lock, originally taken via the fast path, has been - * transferred to the main lock table. That's going to require + * transferred to the main lock table. That's going to require * some extra work, so release our fast-path lock before starting. */ LWLockRelease(MyProc->backendLock); @@ -1926,7 +1927,7 @@ LockReleaseAll(LOCKMETHODID lockmethodid, bool allLocks) /* * Now dump the lock. We haven't got a pointer to the LOCK or * PROCLOCK in this case, so we have to handle this a bit - * differently than a normal lock release. Unfortunately, this + * differently than a normal lock release. Unfortunately, this * requires an extra LWLock acquire-and-release cycle on the * partitionLock, but hopefully it shouldn't happen often. */ @@ -2268,16 +2269,16 @@ FastPathUnGrantRelationLock(Oid relid, LOCKMODE lockmode) */ static bool FastPathTransferRelationLocks(LockMethod lockMethodTable, const LOCKTAG *locktag, - uint32 hashcode) + uint32 hashcode) { - LWLockId partitionLock = LockHashPartitionLock(hashcode); - Oid relid = locktag->locktag_field2; - uint32 i; + LWLockId partitionLock = LockHashPartitionLock(hashcode); + Oid relid = locktag->locktag_field2; + uint32 i; /* - * Every PGPROC that can potentially hold a fast-path lock is present - * in ProcGlobal->allProcs. Prepared transactions are not, but - * any outstanding fast-path locks held by prepared transactions are + * Every PGPROC that can potentially hold a fast-path lock is present in + * ProcGlobal->allProcs. Prepared transactions are not, but any + * outstanding fast-path locks held by prepared transactions are * transferred to the main lock table. */ for (i = 0; i < ProcGlobal->allProcCount; i++) @@ -2288,19 +2289,19 @@ FastPathTransferRelationLocks(LockMethod lockMethodTable, const LOCKTAG *locktag LWLockAcquire(proc->backendLock, LW_EXCLUSIVE); /* - * If the target backend isn't referencing the same database as we are, - * then we needn't examine the individual relation IDs at all; none of - * them can be relevant. + * If the target backend isn't referencing the same database as we + * are, then we needn't examine the individual relation IDs at all; + * none of them can be relevant. * * proc->databaseId is set at backend startup time and never changes * thereafter, so it might be safe to perform this test before * acquiring proc->backendLock. In particular, it's certainly safe to - * assume that if the target backend holds any fast-path locks, it must - * have performed a memory-fencing operation (in particular, an LWLock - * acquisition) since setting proc->databaseId. However, it's less - * clear that our backend is certain to have performed a memory fencing - * operation since the other backend set proc->databaseId. So for now, - * we test it after acquiring the LWLock just to be safe. + * assume that if the target backend holds any fast-path locks, it + * must have performed a memory-fencing operation (in particular, an + * LWLock acquisition) since setting proc->databaseId. However, it's + * less clear that our backend is certain to have performed a memory + * fencing operation since the other backend set proc->databaseId. So + * for now, we test it after acquiring the LWLock just to be safe. */ if (proc->databaseId != MyDatabaseId) { @@ -2319,7 +2320,7 @@ FastPathTransferRelationLocks(LockMethod lockMethodTable, const LOCKTAG *locktag /* Find or create lock object. */ LWLockAcquire(partitionLock, LW_EXCLUSIVE); for (lockmode = FAST_PATH_LOCKNUMBER_OFFSET; - lockmode < FAST_PATH_LOCKNUMBER_OFFSET+FAST_PATH_BITS_PER_SLOT; + lockmode < FAST_PATH_LOCKNUMBER_OFFSET + FAST_PATH_BITS_PER_SLOT; ++lockmode) { PROCLOCK *proclock; @@ -2346,17 +2347,17 @@ FastPathTransferRelationLocks(LockMethod lockMethodTable, const LOCKTAG *locktag /* * FastPathGetLockEntry * Return the PROCLOCK for a lock originally taken via the fast-path, - * transferring it to the primary lock table if necessary. + * transferring it to the primary lock table if necessary. */ static PROCLOCK * FastPathGetRelationLockEntry(LOCALLOCK *locallock) { - LockMethod lockMethodTable = LockMethods[DEFAULT_LOCKMETHOD]; - LOCKTAG *locktag = &locallock->tag.lock; - PROCLOCK *proclock = NULL; - LWLockId partitionLock = LockHashPartitionLock(locallock->hashcode); - Oid relid = locktag->locktag_field2; - uint32 f; + LockMethod lockMethodTable = LockMethods[DEFAULT_LOCKMETHOD]; + LOCKTAG *locktag = &locallock->tag.lock; + PROCLOCK *proclock = NULL; + LWLockId partitionLock = LockHashPartitionLock(locallock->hashcode); + Oid relid = locktag->locktag_field2; + uint32 f; LWLockAcquire(MyProc->backendLock, LW_EXCLUSIVE); @@ -2383,7 +2384,7 @@ FastPathGetRelationLockEntry(LOCALLOCK *locallock) ereport(ERROR, (errcode(ERRCODE_OUT_OF_MEMORY), errmsg("out of shared memory"), - errhint("You might need to increase max_locks_per_transaction."))); + errhint("You might need to increase max_locks_per_transaction."))); } GrantLock(proclock->tag.myLock, proclock, lockmode); FAST_PATH_CLEAR_LOCKMODE(MyProc, f, lockmode); @@ -2397,7 +2398,7 @@ FastPathGetRelationLockEntry(LOCALLOCK *locallock) if (proclock == NULL) { LOCK *lock; - PROCLOCKTAG proclocktag; + PROCLOCKTAG proclocktag; uint32 proclock_hashcode; LWLockAcquire(partitionLock, LW_SHARED); @@ -2495,15 +2496,15 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode) { int i; Oid relid = locktag->locktag_field2; - VirtualTransactionId vxid; + VirtualTransactionId vxid; /* * Iterate over relevant PGPROCs. Anything held by a prepared * transaction will have been transferred to the primary lock table, - * so we need not worry about those. This is all a bit fuzzy, - * because new locks could be taken after we've visited a particular - * partition, but the callers had better be prepared to deal with - * that anyway, since the locks could equally well be taken between the + * so we need not worry about those. This is all a bit fuzzy, because + * new locks could be taken after we've visited a particular + * partition, but the callers had better be prepared to deal with that + * anyway, since the locks could equally well be taken between the * time we return the value and the time the caller does something * with it. */ @@ -2520,8 +2521,8 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode) /* * If the target backend isn't referencing the same database as we - * are, then we needn't examine the individual relation IDs at all; - * none of them can be relevant. + * are, then we needn't examine the individual relation IDs at + * all; none of them can be relevant. * * See FastPathTransferLocks() for discussion of why we do this * test after acquiring the lock. @@ -2545,9 +2546,8 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode) lockmask <<= FAST_PATH_LOCKNUMBER_OFFSET; /* - * There can only be one entry per relation, so if we found - * it and it doesn't conflict, we can skip the rest of the - * slots. + * There can only be one entry per relation, so if we found it + * and it doesn't conflict, we can skip the rest of the slots. */ if ((lockmask & conflictMask) == 0) break; @@ -2621,7 +2621,7 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode) */ if (VirtualTransactionIdIsValid(vxid)) { - int i; + int i; /* Avoid duplicate entries. */ for (i = 0; i < fast_count; ++i) @@ -2650,7 +2650,7 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode) * responsibility to verify that this is a sane thing to do. (For example, it * would be bad to release a lock here if there might still be a LOCALLOCK * object with pointers to it.) - * + * * We currently use this in two situations: first, to release locks held by * prepared transactions on commit (see lock_twophase_postcommit); and second, * to release locks taken via the fast-path, transferred to the main hash @@ -2725,13 +2725,14 @@ LockRefindAndRelease(LockMethod lockMethodTable, PGPROC *proc, LWLockRelease(partitionLock); - /* + /* * Decrement strong lock count. This logic is needed only for 2PC. */ if (decrement_strong_lock_count && ConflictsWithRelationFastPath(&lock->tag, lockmode)) { - uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); + uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); + SpinLockAcquire(&FastPathStrongRelationLocks->mutex); FastPathStrongRelationLocks->count[fasthashcode]--; SpinLockRelease(&FastPathStrongRelationLocks->mutex); @@ -2760,8 +2761,8 @@ AtPrepare_Locks(void) /* * For the most part, we don't need to touch shared memory for this --- * all the necessary state information is in the locallock table. - * Fast-path locks are an exception, however: we move any such locks - * to the main table before allowing PREPARE TRANSACTION to succeed. + * Fast-path locks are an exception, however: we move any such locks to + * the main table before allowing PREPARE TRANSACTION to succeed. */ hash_seq_init(&status, LockMethodLocalHash); @@ -2799,7 +2800,7 @@ AtPrepare_Locks(void) continue; /* - * If we have both session- and transaction-level locks, fail. This + * If we have both session- and transaction-level locks, fail. This * should never happen with regular locks, since we only take those at * session level in some special operations like VACUUM. It's * possible to hit this with advisory locks, though. @@ -2808,7 +2809,7 @@ AtPrepare_Locks(void) * the transactional hold to the prepared xact. However, that would * require two PROCLOCK objects, and we cannot be sure that another * PROCLOCK will be available when it comes time for PostPrepare_Locks - * to do the deed. So for now, we error out while we can still do so + * to do the deed. So for now, we error out while we can still do so * safely. */ if (haveSessionLock) @@ -2819,7 +2820,8 @@ AtPrepare_Locks(void) /* * If the local lock was taken via the fast-path, we need to move it * to the primary lock table, or just get a pointer to the existing - * primary lock table entry if by chance it's already been transferred. + * primary lock table entry if by chance it's already been + * transferred. */ if (locallock->proclock == NULL) { @@ -2829,8 +2831,8 @@ AtPrepare_Locks(void) /* * Arrange to not release any strong lock count held by this lock - * entry. We must retain the count until the prepared transaction - * is committed or rolled back. + * entry. We must retain the count until the prepared transaction is + * committed or rolled back. */ locallock->holdsStrongLockCount = FALSE; @@ -3114,12 +3116,12 @@ GetLockStatusData(void) /* * First, we iterate through the per-backend fast-path arrays, locking - * them one at a time. This might produce an inconsistent picture of the + * them one at a time. This might produce an inconsistent picture of the * system state, but taking all of those LWLocks at the same time seems * impractical (in particular, note MAX_SIMUL_LWLOCKS). It shouldn't * matter too much, because none of these locks can be involved in lock - * conflicts anyway - anything that might must be present in the main - * lock table. + * conflicts anyway - anything that might must be present in the main lock + * table. */ for (i = 0; i < ProcGlobal->allProcCount; ++i) { @@ -3130,7 +3132,7 @@ GetLockStatusData(void) for (f = 0; f < FP_LOCK_SLOTS_PER_BACKEND; ++f) { - LockInstanceData *instance; + LockInstanceData *instance; uint32 lockbits = FAST_PATH_GET_BITS(proc, f); /* Skip unallocated slots. */ @@ -3159,8 +3161,8 @@ GetLockStatusData(void) if (proc->fpVXIDLock) { - VirtualTransactionId vxid; - LockInstanceData *instance; + VirtualTransactionId vxid; + LockInstanceData *instance; if (el >= els) { @@ -3219,7 +3221,7 @@ GetLockStatusData(void) { PGPROC *proc = proclock->tag.myProc; LOCK *lock = proclock->tag.myLock; - LockInstanceData *instance = &data->locks[el]; + LockInstanceData *instance = &data->locks[el]; memcpy(&instance->locktag, &lock->tag, sizeof(LOCKTAG)); instance->holdMask = proclock->holdMask; @@ -3304,10 +3306,10 @@ GetRunningTransactionLocks(int *nlocks) TransactionId xid = pgxact->xid; /* - * Don't record locks for transactions if we know they have already - * issued their WAL record for commit but not yet released lock. - * It is still possible that we see locks held by already complete - * transactions, if they haven't yet zeroed their xids. + * Don't record locks for transactions if we know they have + * already issued their WAL record for commit but not yet released + * lock. It is still possible that we see locks held by already + * complete transactions, if they haven't yet zeroed their xids. */ if (!TransactionIdIsValid(xid)) continue; @@ -3607,13 +3609,14 @@ lock_twophase_recover(TransactionId xid, uint16 info, */ GrantLock(lock, proclock, lockmode); - /* + /* * Bump strong lock count, to make sure any fast-path lock requests won't * be granted without consulting the primary lock table. */ if (ConflictsWithRelationFastPath(&lock->tag, lockmode)) { - uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); + uint32 fasthashcode = FastPathStrongLockHashPartition(hashcode); + SpinLockAcquire(&FastPathStrongRelationLocks->mutex); FastPathStrongRelationLocks->count[fasthashcode]++; SpinLockRelease(&FastPathStrongRelationLocks->mutex); @@ -3701,7 +3704,7 @@ lock_twophase_postabort(TransactionId xid, uint16 info, * as MyProc->lxid, you might wonder if we really need both. The * difference is that MyProc->lxid is set and cleared unlocked, and * examined by procarray.c, while fpLocalTransactionId is protected by - * backendLock and is used only by the locking subsystem. Doing it this + * backendLock and is used only by the locking subsystem. Doing it this * way makes it easier to verify that there are no funny race conditions. * * We don't bother recording this lock in the local lock table, since it's @@ -3734,8 +3737,8 @@ VirtualXactLockTableInsert(VirtualTransactionId vxid) static void VirtualXactLockTableCleanup() { - bool fastpath; - LocalTransactionId lxid; + bool fastpath; + LocalTransactionId lxid; Assert(MyProc->backendId != InvalidBackendId); @@ -3757,8 +3760,8 @@ VirtualXactLockTableCleanup() */ if (!fastpath && LocalTransactionIdIsValid(lxid)) { - VirtualTransactionId vxid; - LOCKTAG locktag; + VirtualTransactionId vxid; + LOCKTAG locktag; vxid.backendId = MyBackendId; vxid.localTransactionId = lxid; @@ -3766,7 +3769,7 @@ VirtualXactLockTableCleanup() LockRefindAndRelease(LockMethods[DEFAULT_LOCKMETHOD], MyProc, &locktag, ExclusiveLock, false); - } + } } /* @@ -3802,8 +3805,8 @@ VirtualXactLock(VirtualTransactionId vxid, bool wait) /* * We must acquire this lock before checking the backendId and lxid - * against the ones we're waiting for. The target backend will only - * set or clear lxid while holding this lock. + * against the ones we're waiting for. The target backend will only set + * or clear lxid while holding this lock. */ LWLockAcquire(proc->backendLock, LW_EXCLUSIVE); @@ -3841,7 +3844,7 @@ VirtualXactLock(VirtualTransactionId vxid, bool wait) ereport(ERROR, (errcode(ERRCODE_OUT_OF_MEMORY), errmsg("out of shared memory"), - errhint("You might need to increase max_locks_per_transaction."))); + errhint("You might need to increase max_locks_per_transaction."))); GrantLock(proclock->tag.myLock, proclock, ExclusiveLock); proc->fpVXIDLock = false; } diff --git a/src/backend/storage/lmgr/lwlock.c b/src/backend/storage/lmgr/lwlock.c index f1523760e5d6761abaaa08970a9b55db1014f0a4..95d4b37bef3ac6ae4a9ed7d18ffecf71cf043fa2 100644 --- a/src/backend/storage/lmgr/lwlock.c +++ b/src/backend/storage/lmgr/lwlock.c @@ -574,7 +574,7 @@ LWLockConditionalAcquire(LWLockId lockid, LWLockMode mode) /* * LWLockAcquireOrWait - Acquire lock, or wait until it's free * - * The semantics of this function are a bit funky. If the lock is currently + * The semantics of this function are a bit funky. If the lock is currently * free, it is acquired in the given mode, and the function returns true. If * the lock isn't immediately free, the function waits until it is released * and returns false, but does not acquire the lock. @@ -769,7 +769,7 @@ LWLockRelease(LWLockId lockid) /* * Remove the to-be-awakened PGPROCs from the queue. */ - bool releaseOK = true; + bool releaseOK = true; proc = head; @@ -797,6 +797,7 @@ LWLockRelease(LWLockId lockid) /* proc is now the last PGPROC to be released */ lock->head = proc->lwWaitLink; proc->lwWaitLink = NULL; + /* * Prevent additional wakeups until retryer gets to run. Backends * that are just waiting for the lock to become free don't retry diff --git a/src/backend/storage/lmgr/predicate.c b/src/backend/storage/lmgr/predicate.c index 6322a608cb184ea116b8d952e8a68bd6b2bac1ec..0c3c7f089bf27c154a4656cd462512140170dadc 100644 --- a/src/backend/storage/lmgr/predicate.c +++ b/src/backend/storage/lmgr/predicate.c @@ -1509,7 +1509,7 @@ GetSafeSnapshot(Snapshot origSnapshot) * one passed to it, but we avoid assuming that here. */ snapshot = GetSerializableTransactionSnapshotInt(origSnapshot, - InvalidTransactionId); + InvalidTransactionId); if (MySerializableXact == InvalidSerializableXact) return snapshot; /* no concurrent r/w xacts; it's safe */ @@ -1600,9 +1600,9 @@ SetSerializableTransactionSnapshot(Snapshot snapshot, /* * We do not allow SERIALIZABLE READ ONLY DEFERRABLE transactions to - * import snapshots, since there's no way to wait for a safe snapshot - * when we're using the snap we're told to. (XXX instead of throwing - * an error, we could just ignore the XactDeferrable flag?) + * import snapshots, since there's no way to wait for a safe snapshot when + * we're using the snap we're told to. (XXX instead of throwing an error, + * we could just ignore the XactDeferrable flag?) */ if (XactReadOnly && XactDeferrable) ereport(ERROR, @@ -1646,11 +1646,11 @@ GetSerializableTransactionSnapshotInt(Snapshot snapshot, * * We must hold SerializableXactHashLock when taking/checking the snapshot * to avoid race conditions, for much the same reasons that - * GetSnapshotData takes the ProcArrayLock. Since we might have to release - * SerializableXactHashLock to call SummarizeOldestCommittedSxact, this - * means we have to create the sxact first, which is a bit annoying (in - * particular, an elog(ERROR) in procarray.c would cause us to leak the - * sxact). Consider refactoring to avoid this. + * GetSnapshotData takes the ProcArrayLock. Since we might have to + * release SerializableXactHashLock to call SummarizeOldestCommittedSxact, + * this means we have to create the sxact first, which is a bit annoying + * (in particular, an elog(ERROR) in procarray.c would cause us to leak + * the sxact). Consider refactoring to avoid this. */ #ifdef TEST_OLDSERXID SummarizeOldestCommittedSxact(); @@ -1678,8 +1678,8 @@ GetSerializableTransactionSnapshotInt(Snapshot snapshot, ereport(ERROR, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), errmsg("could not import the requested snapshot"), - errdetail("The source transaction %u is not running anymore.", - sourcexid))); + errdetail("The source transaction %u is not running anymore.", + sourcexid))); } /* @@ -2704,8 +2704,8 @@ TransferPredicateLocksToNewTarget(PREDICATELOCKTARGETTAG oldtargettag, newpredlock = (PREDICATELOCK *) hash_search_with_hash_value(PredicateLockHash, &newpredlocktag, - PredicateLockHashCodeFromTargetHashCode(&newpredlocktag, - newtargettaghash), + PredicateLockHashCodeFromTargetHashCode(&newpredlocktag, + newtargettaghash), HASH_ENTER_NULL, &found); if (!newpredlock) @@ -2945,8 +2945,8 @@ DropAllPredicateLocksFromTable(Relation relation, bool transfer) newpredlock = (PREDICATELOCK *) hash_search_with_hash_value(PredicateLockHash, &newpredlocktag, - PredicateLockHashCodeFromTargetHashCode(&newpredlocktag, - heaptargettaghash), + PredicateLockHashCodeFromTargetHashCode(&newpredlocktag, + heaptargettaghash), HASH_ENTER, &found); if (!found) @@ -3253,6 +3253,7 @@ ReleasePredicateLocks(bool isCommit) */ MySerializableXact->flags |= SXACT_FLAG_DOOMED; MySerializableXact->flags |= SXACT_FLAG_ROLLED_BACK; + /* * If the transaction was previously prepared, but is now failing due * to a ROLLBACK PREPARED or (hopefully very rare) error after the @@ -3544,9 +3545,9 @@ ClearOldPredicateLocks(void) else { /* - * A read-write transaction can only be partially - * cleared. We need to keep the SERIALIZABLEXACT but - * can release the SIREAD locks and conflicts in. + * A read-write transaction can only be partially cleared. We + * need to keep the SERIALIZABLEXACT but can release the + * SIREAD locks and conflicts in. */ ReleaseOneSerializableXact(finishedSxact, true, false); } @@ -4003,7 +4004,7 @@ CheckForSerializableConflictOut(bool visible, Relation relation, ereport(ERROR, (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE), errmsg("could not serialize access due to read/write dependencies among transactions"), - errdetail_internal("Reason code: Canceled on conflict out to old pivot."), + errdetail_internal("Reason code: Canceled on conflict out to old pivot."), errhint("The transaction might succeed if retried."))); } } @@ -4507,7 +4508,7 @@ OnConflict_CheckForSerializationFailure(const SERIALIZABLEXACT *reader, && (!SxactIsCommitted(writer) || t2->prepareSeqNo <= writer->commitSeqNo) && (!SxactIsReadOnly(reader) - || t2->prepareSeqNo <= reader->SeqNo.lastCommitBeforeSnapshot)) + || t2->prepareSeqNo <= reader->SeqNo.lastCommitBeforeSnapshot)) { failure = true; break; @@ -4552,7 +4553,7 @@ OnConflict_CheckForSerializationFailure(const SERIALIZABLEXACT *reader, && (!SxactIsCommitted(t0) || t0->commitSeqNo >= writer->prepareSeqNo) && (!SxactIsReadOnly(t0) - || t0->SeqNo.lastCommitBeforeSnapshot >= writer->prepareSeqNo)) + || t0->SeqNo.lastCommitBeforeSnapshot >= writer->prepareSeqNo)) { failure = true; break; @@ -4730,10 +4731,10 @@ AtPrepare_PredicateLocks(void) xactRecord->flags = MySerializableXact->flags; /* - * Note that we don't include the list of conflicts in our out in - * the statefile, because new conflicts can be added even after the - * transaction prepares. We'll just make a conservative assumption - * during recovery instead. + * Note that we don't include the list of conflicts in our out in the + * statefile, because new conflicts can be added even after the + * transaction prepares. We'll just make a conservative assumption during + * recovery instead. */ RegisterTwoPhaseRecord(TWOPHASE_RM_PREDICATELOCK_ID, 0, @@ -4891,10 +4892,9 @@ predicatelock_twophase_recover(TransactionId xid, uint16 info, } /* - * We don't know whether the transaction had any conflicts or - * not, so we'll conservatively assume that it had both a - * conflict in and a conflict out, and represent that with the - * summary conflict flags. + * We don't know whether the transaction had any conflicts or not, so + * we'll conservatively assume that it had both a conflict in and a + * conflict out, and represent that with the summary conflict flags. */ SHMQueueInit(&(sxact->outConflicts)); SHMQueueInit(&(sxact->inConflicts)); diff --git a/src/backend/storage/lmgr/proc.c b/src/backend/storage/lmgr/proc.c index 458cd27a3863b3ba90dcd827e7d46c70fc5ddf00..327f43bb352925e78bd49b3830973b7f487870e7 100644 --- a/src/backend/storage/lmgr/proc.c +++ b/src/backend/storage/lmgr/proc.c @@ -70,9 +70,9 @@ PGXACT *MyPgXact = NULL; NON_EXEC_STATIC slock_t *ProcStructLock = NULL; /* Pointers to shared-memory structures */ -PROC_HDR *ProcGlobal = NULL; +PROC_HDR *ProcGlobal = NULL; NON_EXEC_STATIC PGPROC *AuxiliaryProcs = NULL; -PGPROC *PreparedXactProcs = NULL; +PGPROC *PreparedXactProcs = NULL; /* If we are waiting for a lock, this points to the associated LOCALLOCK */ static LOCALLOCK *lockAwaited = NULL; @@ -222,9 +222,9 @@ InitProcGlobal(void) /* Common initialization for all PGPROCs, regardless of type. */ /* - * Set up per-PGPROC semaphore, latch, and backendLock. Prepared - * xact dummy PGPROCs don't need these though - they're never - * associated with a real process + * Set up per-PGPROC semaphore, latch, and backendLock. Prepared xact + * dummy PGPROCs don't need these though - they're never associated + * with a real process */ if (i < MaxBackends + NUM_AUXILIARY_PROCS) { @@ -235,12 +235,12 @@ InitProcGlobal(void) procs[i].pgprocno = i; /* - * Newly created PGPROCs for normal backends or for autovacuum must - * be queued up on the appropriate free list. Because there can only - * ever be a small, fixed number of auxiliary processes, no free - * list is used in that case; InitAuxiliaryProcess() instead uses a - * linear search. PGPROCs for prepared transactions are added to a - * free list by TwoPhaseShmemInit(). + * Newly created PGPROCs for normal backends or for autovacuum must be + * queued up on the appropriate free list. Because there can only + * ever be a small, fixed number of auxiliary processes, no free list + * is used in that case; InitAuxiliaryProcess() instead uses a linear + * search. PGPROCs for prepared transactions are added to a free list + * by TwoPhaseShmemInit(). */ if (i < MaxConnections) { @@ -261,8 +261,8 @@ InitProcGlobal(void) } /* - * Save pointers to the blocks of PGPROC structures reserved for - * auxiliary processes and prepared transactions. + * Save pointers to the blocks of PGPROC structures reserved for auxiliary + * processes and prepared transactions. */ AuxiliaryProcs = &procs[MaxBackends]; PreparedXactProcs = &procs[MaxBackends + NUM_AUXILIARY_PROCS]; @@ -340,8 +340,8 @@ InitProcess(void) MarkPostmasterChildActive(); /* - * Initialize all fields of MyProc, except for those previously initialized - * by InitProcGlobal. + * Initialize all fields of MyProc, except for those previously + * initialized by InitProcGlobal. */ SHMQueueElemInit(&(MyProc->links)); MyProc->waitStatus = STATUS_OK; @@ -366,7 +366,7 @@ InitProcess(void) #ifdef USE_ASSERT_CHECKING if (assert_enabled) { - int i; + int i; /* Last process should have released all locks. */ for (i = 0; i < NUM_LOCK_PARTITIONS; i++) @@ -500,8 +500,8 @@ InitAuxiliaryProcess(void) SpinLockRelease(ProcStructLock); /* - * Initialize all fields of MyProc, except for those previously initialized - * by InitProcGlobal. + * Initialize all fields of MyProc, except for those previously + * initialized by InitProcGlobal. */ SHMQueueElemInit(&(MyProc->links)); MyProc->waitStatus = STATUS_OK; @@ -521,7 +521,7 @@ InitAuxiliaryProcess(void) #ifdef USE_ASSERT_CHECKING if (assert_enabled) { - int i; + int i; /* Last process should have released all locks. */ for (i = 0; i < NUM_LOCK_PARTITIONS; i++) @@ -751,7 +751,7 @@ ProcKill(int code, Datum arg) #ifdef USE_ASSERT_CHECKING if (assert_enabled) { - int i; + int i; /* Last process should have released all locks. */ for (i = 0; i < NUM_LOCK_PARTITIONS; i++) @@ -1031,8 +1031,8 @@ ProcSleep(LOCALLOCK *locallock, LockMethod lockMethodTable) /* * Also, now that we will successfully clean up after an ereport, it's * safe to check to see if there's a buffer pin deadlock against the - * Startup process. Of course, that's only necessary if we're doing - * Hot Standby and are not the Startup process ourselves. + * Startup process. Of course, that's only necessary if we're doing Hot + * Standby and are not the Startup process ourselves. */ if (RecoveryInProgress() && !InRecovery) CheckRecoveryConflictDeadlock(); diff --git a/src/backend/storage/lmgr/s_lock.c b/src/backend/storage/lmgr/s_lock.c index aa99f495f6ef6818cbe1ba2ace95087680dfa4e2..bc8d89f8c1791f628000719648466507f6d8540a 100644 --- a/src/backend/storage/lmgr/s_lock.c +++ b/src/backend/storage/lmgr/s_lock.c @@ -20,7 +20,7 @@ #include "storage/s_lock.h" -slock_t dummy_spinlock; +slock_t dummy_spinlock; static int spins_per_delay = DEFAULT_SPINS_PER_DELAY; diff --git a/src/backend/storage/smgr/md.c b/src/backend/storage/smgr/md.c index eeb20fdf633f4cf62f3e2a7c0ff9ffc8c5101e11..e5dec9d2a329b1a36e82c4ed62f2ba6be48217c6 100644 --- a/src/backend/storage/smgr/md.c +++ b/src/backend/storage/smgr/md.c @@ -325,7 +325,7 @@ mdcreate(SMgrRelation reln, ForkNumber forkNum, bool isRedo) * * All the above applies only to the relation's main fork; other forks can * just be removed immediately, since they are not needed to prevent the - * relfilenode number from being recycled. Also, we do not carefully + * relfilenode number from being recycled. Also, we do not carefully * track whether other forks have been created or not, but just attempt to * unlink them unconditionally; so we should never complain about ENOENT. * @@ -767,9 +767,10 @@ mdnblocks(SMgrRelation reln, ForkNumber forknum) * NOTE: this assumption could only be wrong if another backend has * truncated the relation. We rely on higher code levels to handle that * scenario by closing and re-opening the md fd, which is handled via - * relcache flush. (Since the checkpointer doesn't participate in relcache - * flush, it could have segment chain entries for inactive segments; - * that's OK because the checkpointer never needs to compute relation size.) + * relcache flush. (Since the checkpointer doesn't participate in + * relcache flush, it could have segment chain entries for inactive + * segments; that's OK because the checkpointer never needs to compute + * relation size.) */ while (v->mdfd_chain != NULL) { @@ -1072,12 +1073,13 @@ mdsync(void) * say "but an unreferenced SMgrRelation is still a leak!" Not * really, because the only case in which a checkpoint is done * by a process that isn't about to shut down is in the - * checkpointer, and it will periodically do smgrcloseall(). This - * fact justifies our not closing the reln in the success path - * either, which is a good thing since in non-checkpointer cases - * we couldn't safely do that.) Furthermore, in many cases - * the relation will have been dirtied through this same smgr - * relation, and so we can save a file open/close cycle. + * checkpointer, and it will periodically do smgrcloseall(). + * This fact justifies our not closing the reln in the success + * path either, which is a good thing since in + * non-checkpointer cases we couldn't safely do that.) + * Furthermore, in many cases the relation will have been + * dirtied through this same smgr relation, and so we can save + * a file open/close cycle. */ reln = smgropen(entry->tag.rnode.node, entry->tag.rnode.backend); @@ -1470,8 +1472,8 @@ ForgetRelationFsyncRequests(RelFileNodeBackend rnode, ForkNumber forknum) pg_usleep(10000L); /* 10 msec seems a good number */ /* - * Note we don't wait for the checkpointer to actually absorb the revoke - * message; see mdsync() for the implications. + * Note we don't wait for the checkpointer to actually absorb the + * revoke message; see mdsync() for the implications. */ } } diff --git a/src/backend/storage/smgr/smgr.c b/src/backend/storage/smgr/smgr.c index 60eb81f774f280d378dd9b04acc8a69c1f50be5c..407942ace490e079c3d2766ce69fe5688e291e04 100644 --- a/src/backend/storage/smgr/smgr.c +++ b/src/backend/storage/smgr/smgr.c @@ -405,8 +405,8 @@ smgrdounlinkfork(SMgrRelation reln, ForkNumber forknum, bool isRedo) (*(smgrsw[which].smgr_close)) (reln, forknum); /* - * Get rid of any remaining buffers for the fork. bufmgr will just - * drop them without bothering to write the contents. + * Get rid of any remaining buffers for the fork. bufmgr will just drop + * them without bothering to write the contents. */ DropRelFileNodeBuffers(rnode, forknum, 0); diff --git a/src/backend/tcop/postgres.c b/src/backend/tcop/postgres.c index 89de154bc609829dd3ea9395575b1cd7c0679a68..51b6df54f4c1afe6b94012aea0bcda2b46ff7719 100644 --- a/src/backend/tcop/postgres.c +++ b/src/backend/tcop/postgres.c @@ -350,9 +350,9 @@ SocketBackend(StringInfo inBuf) else { /* - * Can't send DEBUG log messages to client at this point. - * Since we're disconnecting right away, we don't need to - * restore whereToSendOutput. + * Can't send DEBUG log messages to client at this point. Since + * we're disconnecting right away, we don't need to restore + * whereToSendOutput. */ whereToSendOutput = DestNone; ereport(DEBUG1, @@ -393,7 +393,7 @@ SocketBackend(StringInfo inBuf) whereToSendOutput = DestNone; ereport(DEBUG1, (errcode(ERRCODE_CONNECTION_DOES_NOT_EXIST), - errmsg("unexpected EOF on client connection"))); + errmsg("unexpected EOF on client connection"))); } return EOF; } @@ -999,12 +999,12 @@ exec_simple_query(const char *query_string) /* * Start the portal. - * - * If we took a snapshot for parsing/planning, the portal may be - * able to reuse it for the execution phase. Currently, this will only + * + * If we took a snapshot for parsing/planning, the portal may be able + * to reuse it for the execution phase. Currently, this will only * happen in PORTAL_ONE_SELECT mode. But even if PortalStart doesn't - * end up being able to do this, keeping the parse/plan snapshot around - * until after we start the portal doesn't cost much. + * end up being able to do this, keeping the parse/plan snapshot + * around until after we start the portal doesn't cost much. */ PortalStart(portal, NULL, 0, snapshot_set); @@ -1263,8 +1263,8 @@ exec_parse_message(const char *query_string, /* string to execute */ errdetail_abort())); /* - * Create the CachedPlanSource before we do parse analysis, since - * it needs to see the unmodified raw parse tree. + * Create the CachedPlanSource before we do parse analysis, since it + * needs to see the unmodified raw parse tree. */ psrc = CreateCachedPlan(raw_parse_tree, query_string, commandTag); @@ -1325,8 +1325,8 @@ exec_parse_message(const char *query_string, /* string to execute */ /* * CachedPlanSource must be a direct child of MessageContext before we * reparent unnamed_stmt_context under it, else we have a disconnected - * circular subgraph. Klugy, but less so than flipping contexts even - * more above. + * circular subgraph. Klugy, but less so than flipping contexts even more + * above. */ if (unnamed_stmt_context) MemoryContextSetParent(psrc->context, MessageContext); @@ -1549,9 +1549,9 @@ exec_bind_message(StringInfo input_message) /* * Set a snapshot if we have parameters to fetch (since the input * functions might need it) or the query isn't a utility command (and - * hence could require redoing parse analysis and planning). We keep - * the snapshot active till we're done, so that plancache.c doesn't have - * to take new ones. + * hence could require redoing parse analysis and planning). We keep the + * snapshot active till we're done, so that plancache.c doesn't have to + * take new ones. */ if (numParams > 0 || analyze_requires_snapshot(psrc->raw_parse_tree)) { @@ -1687,8 +1687,8 @@ exec_bind_message(StringInfo input_message) params->params[paramno].isnull = isNull; /* - * We mark the params as CONST. This ensures that any custom - * plan makes full use of the parameter values. + * We mark the params as CONST. This ensures that any custom plan + * makes full use of the parameter values. */ params->params[paramno].pflags = PARAM_FLAG_CONST; params->params[paramno].ptype = ptype; @@ -1736,9 +1736,9 @@ exec_bind_message(StringInfo input_message) /* * And we're ready to start portal execution. * - * If we took a snapshot for parsing/planning, we'll try to reuse it - * for query execution (currently, reuse will only occur if - * PORTAL_ONE_SELECT mode is chosen). + * If we took a snapshot for parsing/planning, we'll try to reuse it for + * query execution (currently, reuse will only occur if PORTAL_ONE_SELECT + * mode is chosen). */ PortalStart(portal, params, 0, snapshot_set); @@ -2601,7 +2601,7 @@ die(SIGNAL_ARGS) /* bump holdoff count to make ProcessInterrupts() a no-op */ /* until we are done getting ready for it */ InterruptHoldoffCount++; - LockErrorCleanup(); /* prevent CheckDeadLock from running */ + LockErrorCleanup(); /* prevent CheckDeadLock from running */ DisableNotifyInterrupt(); DisableCatchupInterrupt(); InterruptHoldoffCount--; @@ -2643,7 +2643,7 @@ StatementCancelHandler(SIGNAL_ARGS) /* bump holdoff count to make ProcessInterrupts() a no-op */ /* until we are done getting ready for it */ InterruptHoldoffCount++; - LockErrorCleanup(); /* prevent CheckDeadLock from running */ + LockErrorCleanup(); /* prevent CheckDeadLock from running */ DisableNotifyInterrupt(); DisableCatchupInterrupt(); InterruptHoldoffCount--; @@ -2802,7 +2802,7 @@ RecoveryConflictInterrupt(ProcSignalReason reason) /* bump holdoff count to make ProcessInterrupts() a no-op */ /* until we are done getting ready for it */ InterruptHoldoffCount++; - LockErrorCleanup(); /* prevent CheckDeadLock from running */ + LockErrorCleanup(); /* prevent CheckDeadLock from running */ DisableNotifyInterrupt(); DisableCatchupInterrupt(); InterruptHoldoffCount--; @@ -3269,9 +3269,12 @@ process_postgres_switches(int argc, char *argv[], GucContext ctx) } #ifdef HAVE_INT_OPTERR - /* Turn this off because it's either printed to stderr and not the log - * where we'd want it, or argv[0] is now "--single", which would make for a - * weird error message. We print our own error message below. */ + + /* + * Turn this off because it's either printed to stderr and not the log + * where we'd want it, or argv[0] is now "--single", which would make for + * a weird error message. We print our own error message below. + */ opterr = 0; #endif @@ -3471,7 +3474,7 @@ process_postgres_switches(int argc, char *argv[], GucContext ctx) if (IsUnderPostmaster) ereport(FATAL, (errcode(ERRCODE_SYNTAX_ERROR), - errmsg("invalid command-line argument for server process: %s", argv[optind]), + errmsg("invalid command-line argument for server process: %s", argv[optind]), errhint("Try \"%s --help\" for more information.", progname))); else ereport(FATAL, diff --git a/src/backend/tcop/utility.c b/src/backend/tcop/utility.c index 89f78f2e82b26d4c8c05ff3ec9a3a5445d1c75aa..8b73858300ee157c09f02f87a55a195f81a93913 100644 --- a/src/backend/tcop/utility.c +++ b/src/backend/tcop/utility.c @@ -82,8 +82,8 @@ CheckRelationOwnership(RangeVar *rel, bool noCatalogs) * XXX: This is unsafe in the presence of concurrent DDL, since it is * called before acquiring any lock on the target relation. However, * locking the target relation (especially using something like - * AccessExclusiveLock) before verifying that the user has permissions - * is not appealing either. + * AccessExclusiveLock) before verifying that the user has permissions is + * not appealing either. */ relOid = RangeVarGetRelid(rel, NoLock, false); @@ -634,7 +634,7 @@ standard_ProcessUtility(Node *parsetree, case OBJECT_INDEX: if (((DropStmt *) parsetree)->concurrent) PreventTransactionChain(isTopLevel, - "DROP INDEX CONCURRENTLY"); + "DROP INDEX CONCURRENTLY"); /* fall through */ case OBJECT_TABLE: @@ -712,7 +712,7 @@ standard_ProcessUtility(Node *parsetree, LOCKMODE lockmode; /* - * Figure out lock mode, and acquire lock. This also does + * Figure out lock mode, and acquire lock. This also does * basic permissions checks, so that we won't wait for a lock * on (for example) a relation on which we have no * permissions. @@ -753,8 +753,8 @@ standard_ProcessUtility(Node *parsetree, } else ereport(NOTICE, - (errmsg("relation \"%s\" does not exist, skipping", - atstmt->relation->relname))); + (errmsg("relation \"%s\" does not exist, skipping", + atstmt->relation->relname))); } break; diff --git a/src/backend/tsearch/dict_thesaurus.c b/src/backend/tsearch/dict_thesaurus.c index 37cf575a85cdd1ecffcf14f5e85a189f1ea9d8d7..7e641ef64f55a9f56658127f08b50aa9f11eeee5 100644 --- a/src/backend/tsearch/dict_thesaurus.c +++ b/src/backend/tsearch/dict_thesaurus.c @@ -372,8 +372,8 @@ cmpLexemeQ(const void *a, const void *b) static int cmpTheLexeme(const void *a, const void *b) { - const TheLexeme *la = (const TheLexeme *) a; - const TheLexeme *lb = (const TheLexeme *) b; + const TheLexeme *la = (const TheLexeme *) a; + const TheLexeme *lb = (const TheLexeme *) b; int res; if ((res = cmpLexeme(la, lb)) != 0) diff --git a/src/backend/tsearch/spell.c b/src/backend/tsearch/spell.c index 2fe40ed97358a71fee9fc9944d7fe62bc5f07406..449aa6a0a552c3fc40e4d585e99def7950c426e6 100644 --- a/src/backend/tsearch/spell.c +++ b/src/backend/tsearch/spell.c @@ -148,12 +148,12 @@ static char *VoidString = ""; static int cmpspell(const void *s1, const void *s2) { - return (strcmp((*(SPELL * const *) s1)->word, (*(SPELL * const *) s2)->word)); + return (strcmp((*(SPELL *const *) s1)->word, (*(SPELL *const *) s2)->word)); } static int cmpspellaffix(const void *s1, const void *s2) { - return (strncmp((*(SPELL * const *) s1)->p.flag, (*(SPELL * const *) s2)->p.flag, MAXFLAGLEN)); + return (strncmp((*(SPELL *const *) s1)->p.flag, (*(SPELL *const *) s2)->p.flag, MAXFLAGLEN)); } static char * diff --git a/src/backend/tsearch/to_tsany.c b/src/backend/tsearch/to_tsany.c index da9ae8d5ba8fa2ad717396d17cde3c6e3b7b9aa7..bb886ee5849fb52246f7a1de94a1c79c057f5e11 100644 --- a/src/backend/tsearch/to_tsany.c +++ b/src/backend/tsearch/to_tsany.c @@ -33,8 +33,8 @@ compareWORD(const void *a, const void *b) int res; res = tsCompareString( - ((const ParsedWord *) a)->word, ((const ParsedWord *) a)->len, - ((const ParsedWord *) b)->word, ((const ParsedWord *) b)->len, + ((const ParsedWord *) a)->word, ((const ParsedWord *) a)->len, + ((const ParsedWord *) b)->word, ((const ParsedWord *) b)->len, false); if (res == 0) diff --git a/src/backend/tsearch/ts_utils.c b/src/backend/tsearch/ts_utils.c index 582afde167d0e532b04e5ad20f99c0ad89095604..6a4888e5f469ee15b787e1b818019fab696b908d 100644 --- a/src/backend/tsearch/ts_utils.c +++ b/src/backend/tsearch/ts_utils.c @@ -62,7 +62,7 @@ get_tsearch_config_filename(const char *basename, static int comparestr(const void *a, const void *b) { - return strcmp(*(char * const *) a, *(char * const *) b); + return strcmp(*(char *const *) a, *(char *const *) b); } /* diff --git a/src/backend/utils/adt/acl.c b/src/backend/utils/adt/acl.c index a8bf2bfffb216e495c524c367e91c81d24edd3c4..77322a115f1f7b36e08541ae6ab5e3fd873aea5b 100644 --- a/src/backend/utils/adt/acl.c +++ b/src/backend/utils/adt/acl.c @@ -835,15 +835,15 @@ acldefault(GrantObjectType objtype, Oid ownerId) /* - * SQL-accessible version of acldefault(). Hackish mapping from "char" type to + * SQL-accessible version of acldefault(). Hackish mapping from "char" type to * ACL_OBJECT_* values, but it's only used in the information schema, not * documented for general use. */ Datum acldefault_sql(PG_FUNCTION_ARGS) { - char objtypec = PG_GETARG_CHAR(0); - Oid owner = PG_GETARG_OID(1); + char objtypec = PG_GETARG_CHAR(0); + Oid owner = PG_GETARG_OID(1); GrantObjectType objtype = 0; switch (objtypec) diff --git a/src/backend/utils/adt/array_selfuncs.c b/src/backend/utils/adt/array_selfuncs.c index bc4ebd20749997ff3441b7824118dbde320b5524..1b7d46f85867c18492188f14d18f8107d72118bd 100644 --- a/src/backend/utils/adt/array_selfuncs.c +++ b/src/backend/utils/adt/array_selfuncs.c @@ -172,7 +172,7 @@ scalararraysel_containment(PlannerInfo *root, selec = mcelem_array_contain_overlap_selec(values, nvalues, numbers, nnumbers, &constval, 1, - OID_ARRAY_CONTAINS_OP, + OID_ARRAY_CONTAINS_OP, cmpfunc); else selec = mcelem_array_contained_selec(values, nvalues, @@ -193,7 +193,7 @@ scalararraysel_containment(PlannerInfo *root, selec = mcelem_array_contain_overlap_selec(NULL, 0, NULL, 0, &constval, 1, - OID_ARRAY_CONTAINS_OP, + OID_ARRAY_CONTAINS_OP, cmpfunc); else selec = mcelem_array_contained_selec(NULL, 0, @@ -285,8 +285,8 @@ arraycontsel(PG_FUNCTION_ARGS) } /* - * If var is on the right, commute the operator, so that we can assume - * the var is on the left in what follows. + * If var is on the right, commute the operator, so that we can assume the + * var is on the left in what follows. */ if (!varonleft) { @@ -451,7 +451,7 @@ mcelem_array_selec(ArrayType *array, TypeCacheEntry *typentry, float4 *hist, int nhist, Oid operator, FmgrInfo *cmpfunc) { - Selectivity selec; + Selectivity selec; int num_elems; Datum *elem_values; bool *elem_nulls; @@ -500,7 +500,7 @@ mcelem_array_selec(ArrayType *array, TypeCacheEntry *typentry, if (operator == OID_ARRAY_CONTAINS_OP || operator == OID_ARRAY_OVERLAP_OP) selec = mcelem_array_contain_overlap_selec(mcelem, nmcelem, numbers, nnumbers, - elem_values, nonnull_nitems, + elem_values, nonnull_nitems, operator, cmpfunc); else if (operator == OID_ARRAY_CONTAINED_OP) selec = mcelem_array_contained_selec(mcelem, nmcelem, @@ -626,7 +626,7 @@ mcelem_array_contain_overlap_selec(Datum *mcelem, int nmcelem, else { if (cmp == 0) - match = true; /* mcelem is found */ + match = true; /* mcelem is found */ break; } } @@ -687,7 +687,7 @@ mcelem_array_contain_overlap_selec(Datum *mcelem, int nmcelem, * In the "column @> const" and "column && const" cases, we usually have a * "const" with low number of elements (otherwise we have selectivity close * to 0 or 1 respectively). That's why the effect of dependence related - * to distinct element count distribution is negligible there. In the + * to distinct element count distribution is negligible there. In the * "column <@ const" case, number of elements is usually high (otherwise we * have selectivity close to 0). That's why we should do a correction with * the array distinct element count distribution here. @@ -806,7 +806,7 @@ mcelem_array_contained_selec(Datum *mcelem, int nmcelem, else { if (cmp == 0) - match = true; /* mcelem is found */ + match = true; /* mcelem is found */ break; } } @@ -854,7 +854,7 @@ mcelem_array_contained_selec(Datum *mcelem, int nmcelem, /*---------- * Using the distinct element count histogram requires * O(unique_nitems * (nmcelem + unique_nitems)) - * operations. Beyond a certain computational cost threshold, it's + * operations. Beyond a certain computational cost threshold, it's * reasonable to sacrifice accuracy for decreased planning time. We limit * the number of operations to EFFORT * nmcelem; since nmcelem is limited * by the column's statistics target, the work done is user-controllable. @@ -866,7 +866,7 @@ mcelem_array_contained_selec(Datum *mcelem, int nmcelem, * elements to start with, we'd have to remove any discarded elements' * frequencies from "mult", but since this is only an approximation * anyway, we don't bother with that. Therefore it's sufficient to qsort - * elem_selec[] and take the largest elements. (They will no longer match + * elem_selec[] and take the largest elements. (They will no longer match * up with the elements of array_data[], but we don't care.) *---------- */ @@ -876,11 +876,11 @@ mcelem_array_contained_selec(Datum *mcelem, int nmcelem, unique_nitems > EFFORT * nmcelem / (nmcelem + unique_nitems)) { /* - * Use the quadratic formula to solve for largest allowable N. We + * Use the quadratic formula to solve for largest allowable N. We * have A = 1, B = nmcelem, C = - EFFORT * nmcelem. */ - double b = (double) nmcelem; - int n; + double b = (double) nmcelem; + int n; n = (int) ((sqrt(b * b + 4 * EFFORT * b) - b) / 2); @@ -891,9 +891,9 @@ mcelem_array_contained_selec(Datum *mcelem, int nmcelem, } /* - * Calculate probabilities of each distinct element count for both - * mcelems and constant elements. At this point, assume independent - * element occurrence. + * Calculate probabilities of each distinct element count for both mcelems + * and constant elements. At this point, assume independent element + * occurrence. */ dist = calc_distr(elem_selec, unique_nitems, unique_nitems, 0.0f); mcelem_dist = calc_distr(numbers, nmcelem, unique_nitems, rest); @@ -906,8 +906,8 @@ mcelem_array_contained_selec(Datum *mcelem, int nmcelem, { /* * mult * dist[i] / mcelem_dist[i] gives us probability of qual - * matching from assumption of independent element occurrence with - * the condition that distinct element count = i. + * matching from assumption of independent element occurrence with the + * condition that distinct element count = i. */ if (mcelem_dist[i] > 0) selec += hist_part[i] * mult * dist[i] / mcelem_dist[i]; @@ -951,7 +951,7 @@ calc_hist(const float4 *hist, int nhist, int n) /* * frac is a probability contribution for each interval between histogram - * values. We have nhist - 1 intervals, so contribution of each one will + * values. We have nhist - 1 intervals, so contribution of each one will * be 1 / (nhist - 1). */ frac = 1.0f / ((float) (nhist - 1)); @@ -1018,7 +1018,7 @@ calc_hist(const float4 *hist, int nhist, int n) * "rest" is the sum of the probabilities of all low-probability events not * included in p. * - * Imagine matrix M of size (n + 1) x (m + 1). Element M[i,j] denotes the + * Imagine matrix M of size (n + 1) x (m + 1). Element M[i,j] denotes the * probability that exactly j of first i events occur. Obviously M[0,0] = 1. * For any constant j, each increment of i increases the probability iff the * event occurs. So, by the law of total probability: diff --git a/src/backend/utils/adt/array_typanalyze.c b/src/backend/utils/adt/array_typanalyze.c index fa79d9fa6b453da3548180fbfce597cf291d8819..604b86ca644490212b2bcd5bac1a13d078231b97 100644 --- a/src/backend/utils/adt/array_typanalyze.c +++ b/src/backend/utils/adt/array_typanalyze.c @@ -42,9 +42,9 @@ typedef struct char typalign; /* - * Lookup data for element type's comparison and hash functions (these - * are in the type's typcache entry, which we expect to remain valid - * over the lifespan of the ANALYZE run) + * Lookup data for element type's comparison and hash functions (these are + * in the type's typcache entry, which we expect to remain valid over the + * lifespan of the ANALYZE run) */ FmgrInfo *cmp; FmgrInfo *hash; @@ -149,8 +149,8 @@ array_typanalyze(PG_FUNCTION_ARGS) stats->extra_data = extra_data; /* - * Note we leave stats->minrows set as std_typanalyze set it. Should - * it be increased for array analysis purposes? + * Note we leave stats->minrows set as std_typanalyze set it. Should it + * be increased for array analysis purposes? */ PG_RETURN_BOOL(true); @@ -160,13 +160,13 @@ array_typanalyze(PG_FUNCTION_ARGS) * compute_array_stats() -- compute statistics for a array column * * This function computes statistics useful for determining selectivity of - * the array operators <@, &&, and @>. It is invoked by ANALYZE via the + * the array operators <@, &&, and @>. It is invoked by ANALYZE via the * compute_stats hook after sample rows have been collected. * * We also invoke the standard compute_stats function, which will compute * "scalar" statistics relevant to the btree-style array comparison operators. * However, exact duplicates of an entire array may be rare despite many - * arrays sharing individual elements. This especially afflicts long arrays, + * arrays sharing individual elements. This especially afflicts long arrays, * which are also liable to lack all scalar statistics due to the low * WIDTH_THRESHOLD used in analyze.c. So, in addition to the standard stats, * we find the most common array elements and compute a histogram of distinct @@ -201,7 +201,7 @@ array_typanalyze(PG_FUNCTION_ARGS) * In the absence of a principled basis for other particular values, we * follow ts_typanalyze() and use parameters s = 0.07/K, epsilon = s/10. * But we leave out the correction for stopwords, which do not apply to - * arrays. These parameters give bucket width w = K/0.007 and maximum + * arrays. These parameters give bucket width w = K/0.007 and maximum * expected hashtable size of about 1000 * K. * * Elements may repeat within an array. Since duplicates do not change the @@ -242,8 +242,8 @@ compute_array_stats(VacAttrStats *stats, AnalyzeAttrFetchFunc fetchfunc, /* * Invoke analyze.c's standard analysis function to create scalar-style - * stats for the column. It will expect its own extra_data pointer, - * so temporarily install that. + * stats for the column. It will expect its own extra_data pointer, so + * temporarily install that. */ stats->extra_data = extra_data->std_extra_data; (*extra_data->std_compute_stats) (stats, fetchfunc, samplerows, totalrows); @@ -373,8 +373,8 @@ compute_array_stats(VacAttrStats *stats, AnalyzeAttrFetchFunc fetchfunc, /* The element value is already on the tracking list */ /* - * The operators we assist ignore duplicate array elements, - * so count a given distinct element only once per array. + * The operators we assist ignore duplicate array elements, so + * count a given distinct element only once per array. */ if (item->last_container == array_no) continue; @@ -387,11 +387,11 @@ compute_array_stats(VacAttrStats *stats, AnalyzeAttrFetchFunc fetchfunc, /* Initialize new tracking list element */ /* - * If element type is pass-by-reference, we must copy it - * into palloc'd space, so that we can release the array - * below. (We do this so that the space needed for element - * values is limited by the size of the hashtable; if we - * kept all the array values around, it could be much more.) + * If element type is pass-by-reference, we must copy it into + * palloc'd space, so that we can release the array below. + * (We do this so that the space needed for element values is + * limited by the size of the hashtable; if we kept all the + * array values around, it could be much more.) */ item->key = datumCopy(elem_value, extra_data->typbyval, @@ -623,7 +623,7 @@ compute_array_stats(VacAttrStats *stats, AnalyzeAttrFetchFunc fetchfunc, * (compare the histogram-making loop in compute_scalar_stats()). * But instead of that we have the sorted_count_items[] array, * which holds unique DEC values with their frequencies (that is, - * a run-length-compressed version of the full array). So we + * a run-length-compressed version of the full array). So we * control advancing through sorted_count_items[] with the * variable "frac", which is defined as (x - y) * (num_hist - 1), * where x is the index in the notional DECs array corresponding @@ -655,7 +655,7 @@ compute_array_stats(VacAttrStats *stats, AnalyzeAttrFetchFunc fetchfunc, frac += (int64) sorted_count_items[j]->frequency * (num_hist - 1); } hist[i] = sorted_count_items[j]->count; - frac -= delta; /* update y for upcoming i increment */ + frac -= delta; /* update y for upcoming i increment */ } Assert(j == count_items_count - 1); @@ -775,8 +775,8 @@ trackitem_compare_element(const void *e1, const void *e2) static int countitem_compare_count(const void *e1, const void *e2) { - const DECountItem * const *t1 = (const DECountItem * const *) e1; - const DECountItem * const *t2 = (const DECountItem * const *) e2; + const DECountItem *const * t1 = (const DECountItem *const *) e1; + const DECountItem *const * t2 = (const DECountItem *const *) e2; if ((*t1)->count < (*t2)->count) return -1; diff --git a/src/backend/utils/adt/cash.c b/src/backend/utils/adt/cash.c index 4a2d413ba204c807fd34e316c1b411b0997b67e6..82551c5f30e6b6b4a40bdff891986fd92499ef9d 100644 --- a/src/backend/utils/adt/cash.c +++ b/src/backend/utils/adt/cash.c @@ -133,7 +133,7 @@ cash_in(PG_FUNCTION_ARGS) dsymbol = '.'; if (*lconvert->mon_thousands_sep != '\0') ssymbol = lconvert->mon_thousands_sep; - else /* ssymbol should not equal dsymbol */ + else /* ssymbol should not equal dsymbol */ ssymbol = (dsymbol != ',') ? "," : "."; csymbol = (*lconvert->currency_symbol != '\0') ? lconvert->currency_symbol : "$"; psymbol = (*lconvert->positive_sign != '\0') ? lconvert->positive_sign : "+"; @@ -301,7 +301,7 @@ cash_out(PG_FUNCTION_ARGS) dsymbol = '.'; if (*lconvert->mon_thousands_sep != '\0') ssymbol = lconvert->mon_thousands_sep; - else /* ssymbol should not equal dsymbol */ + else /* ssymbol should not equal dsymbol */ ssymbol = (dsymbol != ',') ? "," : "."; csymbol = (*lconvert->currency_symbol != '\0') ? lconvert->currency_symbol : "$"; diff --git a/src/backend/utils/adt/date.c b/src/backend/utils/adt/date.c index 0fc187e0d6f1b616c3940880f87cd931cd7a38e0..6e29ebb7844deccf40ce3b03a3c08a9c0a5c823d 100644 --- a/src/backend/utils/adt/date.c +++ b/src/backend/utils/adt/date.c @@ -337,7 +337,7 @@ date_fastcmp(Datum x, Datum y, SortSupport ssup) Datum date_sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = date_fastcmp; PG_RETURN_VOID(); diff --git a/src/backend/utils/adt/datetime.c b/src/backend/utils/adt/datetime.c index d5d34da5521fdad898c0eb2113e0ed19a8cfae5c..1c2c39b2e272ca54506e6c220ced07918aacaaf5 100644 --- a/src/backend/utils/adt/datetime.c +++ b/src/backend/utils/adt/datetime.c @@ -4170,7 +4170,7 @@ TemporalTransform(int32 max_precis, Node *node) typmod = (Node *) lsecond(expr->args); - if (IsA(typmod, Const) && !((Const *) typmod)->constisnull) + if (IsA(typmod, Const) &&!((Const *) typmod)->constisnull) { Node *source = (Node *) linitial(expr->args); int32 old_precis = exprTypmod(source); diff --git a/src/backend/utils/adt/dbsize.c b/src/backend/utils/adt/dbsize.c index fd19de72cb678faf3fa43ccb41847115899122c8..2ccdc0cee6e598ac6871a9a094353c32c3817044 100644 --- a/src/backend/utils/adt/dbsize.c +++ b/src/backend/utils/adt/dbsize.c @@ -555,6 +555,7 @@ static char * numeric_to_cstring(Numeric n) { Datum d = NumericGetDatum(n); + return DatumGetCString(DirectFunctionCall1(numeric_out, d)); } @@ -562,6 +563,7 @@ static Numeric int64_to_numeric(int64 v) { Datum d = Int64GetDatum(v); + return DatumGetNumeric(DirectFunctionCall1(int8_numeric, d)); } diff --git a/src/backend/utils/adt/float.c b/src/backend/utils/adt/float.c index ca0042a1762dcf24086ea647dccf6094e957895e..b7ce9357f4a7daf8b0a87c49fb0f16177bd7670a 100644 --- a/src/backend/utils/adt/float.c +++ b/src/backend/utils/adt/float.c @@ -217,7 +217,7 @@ float4in(PG_FUNCTION_ARGS) /* did we not see anything that looks like a double? */ if (endptr == num || errno != 0) { - int save_errno = errno; + int save_errno = errno; /* * C99 requires that strtod() accept NaN and [-]Infinity, but not all @@ -244,9 +244,9 @@ float4in(PG_FUNCTION_ARGS) /* * Some platforms return ERANGE for denormalized numbers (those * that are not zero, but are too close to zero to have full - * precision). We'd prefer not to throw error for that, so try - * to detect whether it's a "real" out-of-range condition by - * checking to see if the result is zero or huge. + * precision). We'd prefer not to throw error for that, so try to + * detect whether it's a "real" out-of-range condition by checking + * to see if the result is zero or huge. */ if (val == 0.0 || val >= HUGE_VAL || val <= -HUGE_VAL) ereport(ERROR, @@ -422,7 +422,7 @@ float8in(PG_FUNCTION_ARGS) /* did we not see anything that looks like a double? */ if (endptr == num || errno != 0) { - int save_errno = errno; + int save_errno = errno; /* * C99 requires that strtod() accept NaN and [-]Infinity, but not all @@ -449,15 +449,15 @@ float8in(PG_FUNCTION_ARGS) /* * Some platforms return ERANGE for denormalized numbers (those * that are not zero, but are too close to zero to have full - * precision). We'd prefer not to throw error for that, so try - * to detect whether it's a "real" out-of-range condition by - * checking to see if the result is zero or huge. + * precision). We'd prefer not to throw error for that, so try to + * detect whether it's a "real" out-of-range condition by checking + * to see if the result is zero or huge. */ if (val == 0.0 || val >= HUGE_VAL || val <= -HUGE_VAL) ereport(ERROR, (errcode(ERRCODE_NUMERIC_VALUE_OUT_OF_RANGE), - errmsg("\"%s\" is out of range for type double precision", - orig_num))); + errmsg("\"%s\" is out of range for type double precision", + orig_num))); } else ereport(ERROR, @@ -973,7 +973,7 @@ btfloat4fastcmp(Datum x, Datum y, SortSupport ssup) Datum btfloat4sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btfloat4fastcmp; PG_RETURN_VOID(); @@ -1087,7 +1087,7 @@ btfloat8fastcmp(Datum x, Datum y, SortSupport ssup) Datum btfloat8sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = btfloat8fastcmp; PG_RETURN_VOID(); @@ -2750,7 +2750,7 @@ width_bucket_float8(PG_FUNCTION_ARGS) if (isnan(operand) || isnan(bound1) || isnan(bound2)) ereport(ERROR, (errcode(ERRCODE_INVALID_ARGUMENT_FOR_WIDTH_BUCKET_FUNCTION), - errmsg("operand, lower bound, and upper bound cannot be NaN"))); + errmsg("operand, lower bound, and upper bound cannot be NaN"))); /* Note that we allow "operand" to be infinite */ if (isinf(bound1) || isinf(bound2)) diff --git a/src/backend/utils/adt/formatting.c b/src/backend/utils/adt/formatting.c index d848739d4aabbd978e6f8cd091f76f2c828c95b3..765c6aa8d5edff3ab08cb7433057176672ba7404 100644 --- a/src/backend/utils/adt/formatting.c +++ b/src/backend/utils/adt/formatting.c @@ -1987,8 +1987,8 @@ static int adjust_partial_year_to_2020(int year) { /* - * Adjust all dates toward 2020; this is effectively what happens - * when we assume '70' is 1970 and '69' is 2069. + * Adjust all dates toward 2020; this is effectively what happens when we + * assume '70' is 1970 and '69' is 2069. */ /* Force 0-69 into the 2000's */ if (year < 70) @@ -4485,7 +4485,7 @@ NUM_processor(FormatNode *node, NUMDesc *Num, char *inout, char *number, */ if (Np->last_relevant && Np->Num->zero_end > Np->num_pre) { - char *last_zero; + char *last_zero; last_zero = Np->number + (Np->Num->zero_end - Np->num_pre); if (Np->last_relevant < last_zero) diff --git a/src/backend/utils/adt/inet_net_pton.c b/src/backend/utils/adt/inet_net_pton.c index 66cdacecb54cad4fcb951c8e24e8e34d5ebdac1c..9064eaf64b0645b58ffc47c11b2386078213bf22 100644 --- a/src/backend/utils/adt/inet_net_pton.c +++ b/src/backend/utils/adt/inet_net_pton.c @@ -30,7 +30,8 @@ static const char rcsid[] = "Id: inet_net_pton.c,v 1.4.2.3 2004/03/17 00:40:11 m #include <assert.h> #include <ctype.h> -#include "utils/builtins.h" /* pgrminclude ignore */ /* needed on some platforms */ +#include "utils/builtins.h" /* pgrminclude ignore */ /* needed on some + * platforms */ #include "utils/inet.h" diff --git a/src/backend/utils/adt/json.c b/src/backend/utils/adt/json.c index 61ae62eb8a9c61cbc4d2d2f6828c59ac1ece5fe6..e494630d60ba43c1bb3a6585906f233526b3fa76 100644 --- a/src/backend/utils/adt/json.c +++ b/src/backend/utils/adt/json.c @@ -42,7 +42,7 @@ typedef struct char *input; char *token_start; char *token_terminator; - JsonValueType token_type; + JsonValueType token_type; int line_number; char *line_start; } JsonLexContext; @@ -60,7 +60,7 @@ typedef enum typedef struct JsonParseStack { - JsonParseState state; + JsonParseState state; } JsonParseStack; typedef enum @@ -80,9 +80,9 @@ static void report_invalid_token(JsonLexContext *lex); static char *extract_mb_char(char *s); static void composite_to_json(Datum composite, StringInfo result, bool use_line_feeds); static void array_dim_to_json(StringInfo result, int dim, int ndims, int *dims, - Datum *vals, bool *nulls, int *valcount, - TYPCATEGORY tcategory, Oid typoutputfunc, - bool use_line_feeds); + Datum *vals, bool *nulls, int *valcount, + TYPCATEGORY tcategory, Oid typoutputfunc, + bool use_line_feeds); static void array_to_json_internal(Datum array, StringInfo result, bool use_line_feeds); /* fake type category for JSON so we can distinguish it in datum_to_json */ @@ -95,7 +95,7 @@ static void array_to_json_internal(Datum array, StringInfo result, bool use_line Datum json_in(PG_FUNCTION_ARGS) { - char *text = PG_GETARG_CSTRING(0); + char *text = PG_GETARG_CSTRING(0); json_validate_cstring(text); @@ -108,7 +108,7 @@ json_in(PG_FUNCTION_ARGS) Datum json_out(PG_FUNCTION_ARGS) { - Datum txt = PG_GETARG_DATUM(0); + Datum txt = PG_GETARG_DATUM(0); PG_RETURN_CSTRING(TextDatumGetCString(txt)); } @@ -120,7 +120,7 @@ Datum json_send(PG_FUNCTION_ARGS) { StringInfoData buf; - text *t = PG_GETARG_TEXT_PP(0); + text *t = PG_GETARG_TEXT_PP(0); pq_begintypsend(&buf); pq_sendtext(&buf, VARDATA_ANY(t), VARSIZE_ANY_EXHDR(t)); @@ -163,10 +163,10 @@ json_recv(PG_FUNCTION_ARGS) static void json_validate_cstring(char *input) { - JsonLexContext lex; + JsonLexContext lex; JsonParseStack *stack, - *stacktop; - int stacksize; + *stacktop; + int stacksize; /* Set up lexing context. */ lex.input = input; @@ -183,7 +183,7 @@ json_validate_cstring(char *input) /* Main parsing loop. */ for (;;) { - JsonStackOp op; + JsonStackOp op; /* Fetch next token. */ json_lex(&lex); @@ -213,7 +213,7 @@ redo: else if (lex.token_start[0] == ']') op = JSON_STACKOP_POP; else if (lex.token_start[0] == '[' - || lex.token_start[0] == '{') + || lex.token_start[0] == '{') { stack->state = JSON_PARSE_ARRAY_NEXT; op = JSON_STACKOP_PUSH_WITH_PUSHBACK; @@ -235,7 +235,7 @@ redo: if (lex.token_type == JSON_VALUE_STRING) stack->state = JSON_PARSE_OBJECT_LABEL; else if (lex.token_type == JSON_VALUE_INVALID - && lex.token_start[0] == '}') + && lex.token_start[0] == '}') op = JSON_STACKOP_POP; else report_parse_error(stack, &lex); @@ -268,7 +268,7 @@ redo: break; default: elog(ERROR, "unexpected json parse state: %d", - (int) stack->state); + (int) stack->state); } /* Push or pop the stack, if needed. */ @@ -279,7 +279,8 @@ redo: ++stack; if (stack >= &stacktop[stacksize]) { - int stackoffset = stack - stacktop; + int stackoffset = stack - stacktop; + stacksize = stacksize + 32; stacktop = repalloc(stacktop, sizeof(JsonParseStack) * stacksize); @@ -362,19 +363,19 @@ json_lex(JsonLexContext *lex) } else { - char *p; + char *p; /* - * We're not dealing with a string, number, legal punctuation mark, - * or end of string. The only legal tokens we might find here are - * true, false, and null, but for error reporting purposes we scan - * until we see a non-alphanumeric character. That way, we can report - * the whole word as an unexpected token, rather than just some + * We're not dealing with a string, number, legal punctuation mark, or + * end of string. The only legal tokens we might find here are true, + * false, and null, but for error reporting purposes we scan until we + * see a non-alphanumeric character. That way, we can report the + * whole word as an unexpected token, rather than just some * unintuitive prefix thereof. */ - for (p = s; (*p >= 'a' && *p <= 'z') || (*p >= 'A' && *p <= 'Z') - || (*p >= '0' && *p <= '9') || *p == '_' || IS_HIGHBIT_SET(*p); - ++p) + for (p = s; (*p >= 'a' && *p <= 'z') || (*p >= 'A' && *p <= 'Z') + || (*p >= '0' && *p <= '9') || *p == '_' || IS_HIGHBIT_SET(*p); + ++p) ; /* @@ -431,7 +432,7 @@ json_lex_string(JsonLexContext *lex) (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("invalid input syntax for type json"), errdetail_internal("line %d: Character with value \"0x%02x\" must be escaped.", - lex->line_number, (unsigned char) *s))); + lex->line_number, (unsigned char) *s))); } else if (*s == '\\') { @@ -444,8 +445,8 @@ json_lex_string(JsonLexContext *lex) } else if (*s == 'u') { - int i; - int ch = 0; + int i; + int ch = 0; for (i = 1; i <= 4; ++i) { @@ -466,7 +467,7 @@ json_lex_string(JsonLexContext *lex) (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("invalid input syntax for type json"), errdetail_internal("line %d: \"\\u\" must be followed by four hexadecimal digits.", - lex->line_number))); + lex->line_number))); } } @@ -479,8 +480,8 @@ json_lex_string(JsonLexContext *lex) ereport(ERROR, (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("invalid input syntax for type json"), - errdetail_internal("line %d: Invalid escape \"\\%s\".", - lex->line_number, extract_mb_char(s)))); + errdetail_internal("line %d: Invalid escape \"\\%s\".", + lex->line_number, extract_mb_char(s)))); } } } @@ -497,17 +498,17 @@ json_lex_string(JsonLexContext *lex) * (1) An optional minus sign ('-'). * * (2) Either a single '0', or a string of one or more digits that does not - * begin with a '0'. + * begin with a '0'. * * (3) An optional decimal part, consisting of a period ('.') followed by - * one or more digits. (Note: While this part can be omitted - * completely, it's not OK to have only the decimal point without - * any digits afterwards.) + * one or more digits. (Note: While this part can be omitted + * completely, it's not OK to have only the decimal point without + * any digits afterwards.) * * (4) An optional exponent part, consisting of 'e' or 'E', optionally - * followed by '+' or '-', followed by one or more digits. (Note: - * As with the decimal part, if 'e' or 'E' is present, it must be - * followed by at least one digit.) + * followed by '+' or '-', followed by one or more digits. (Note: + * As with the decimal part, if 'e' or 'E' is present, it must be + * followed by at least one digit.) * * The 's' argument to this function points to the ostensible beginning * of part 2 - i.e. the character after any optional minus sign, and the @@ -518,8 +519,8 @@ json_lex_string(JsonLexContext *lex) static void json_lex_number(JsonLexContext *lex, char *s) { - bool error = false; - char *p; + bool error = false; + char *p; /* Part (1): leading sign indicator. */ /* Caller already did this for us; so do nothing. */ @@ -571,7 +572,7 @@ json_lex_number(JsonLexContext *lex, char *s) /* Check for trailing garbage. */ for (p = s; (*p >= 'a' && *p <= 'z') || (*p >= 'A' && *p <= 'Z') - || (*p >= '0' && *p <= '9') || *p == '_' || IS_HIGHBIT_SET(*p); ++p) + || (*p >= '0' && *p <= '9') || *p == '_' || IS_HIGHBIT_SET(*p); ++p) ; lex->token_terminator = p; if (p > s || error) @@ -584,17 +585,17 @@ json_lex_number(JsonLexContext *lex, char *s) static void report_parse_error(JsonParseStack *stack, JsonLexContext *lex) { - char *detail = NULL; - char *token = NULL; - int toklen; + char *detail = NULL; + char *token = NULL; + int toklen; /* Handle case where the input ended prematurely. */ if (lex->token_start == NULL) ereport(ERROR, (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("invalid input syntax for type json: \"%s\"", - lex->input), - errdetail_internal("The input string ended unexpectedly."))); + lex->input), + errdetail_internal("The input string ended unexpectedly."))); /* Work out the offending token. */ toklen = lex->token_terminator - lex->token_start; @@ -636,8 +637,8 @@ report_parse_error(JsonParseStack *stack, JsonLexContext *lex) ereport(ERROR, (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("invalid input syntax for type json: \"%s\"", - lex->input), - detail ? errdetail_internal(detail, lex->line_number, token) : 0)); + lex->input), + detail ? errdetail_internal(detail, lex->line_number, token) : 0)); } /* @@ -646,8 +647,8 @@ report_parse_error(JsonParseStack *stack, JsonLexContext *lex) static void report_invalid_token(JsonLexContext *lex) { - char *token; - int toklen; + char *token; + int toklen; toklen = lex->token_terminator - lex->token_start; token = palloc(toklen + 1); @@ -658,7 +659,7 @@ report_invalid_token(JsonLexContext *lex) (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("invalid input syntax for type json"), errdetail_internal("line %d: Token \"%s\" is invalid.", - lex->line_number, token))); + lex->line_number, token))); } /* @@ -667,8 +668,8 @@ report_invalid_token(JsonLexContext *lex) static char * extract_mb_char(char *s) { - char *res; - int len; + char *res; + int len; len = pg_mblen(s); res = palloc(len + 1); @@ -687,11 +688,11 @@ datum_to_json(Datum val, bool is_null, StringInfo result, TYPCATEGORY tcategory, Oid typoutputfunc) { - char *outputstr; + char *outputstr; if (is_null) { - appendStringInfoString(result,"null"); + appendStringInfoString(result, "null"); return; } @@ -705,19 +706,20 @@ datum_to_json(Datum val, bool is_null, StringInfo result, TYPCATEGORY tcategory, break; case TYPCATEGORY_BOOLEAN: if (DatumGetBool(val)) - appendStringInfoString(result,"true"); + appendStringInfoString(result, "true"); else - appendStringInfoString(result,"false"); + appendStringInfoString(result, "false"); break; case TYPCATEGORY_NUMERIC: outputstr = OidOutputFunctionCall(typoutputfunc, val); + /* - * Don't call escape_json here if it's a valid JSON - * number. Numeric output should usually be a valid - * JSON number and JSON numbers shouldn't be quoted. - * Quote cases like "Nan" and "Infinity", however. + * Don't call escape_json here if it's a valid JSON number. + * Numeric output should usually be a valid JSON number and JSON + * numbers shouldn't be quoted. Quote cases like "Nan" and + * "Infinity", however. */ - if (strpbrk(outputstr,NON_NUMERIC_LETTER) == NULL) + if (strpbrk(outputstr, NON_NUMERIC_LETTER) == NULL) appendStringInfoString(result, outputstr); else escape_json(result, outputstr); @@ -742,13 +744,13 @@ datum_to_json(Datum val, bool is_null, StringInfo result, TYPCATEGORY tcategory, * ourselves recursively to process the next dimension. */ static void -array_dim_to_json(StringInfo result, int dim, int ndims,int * dims, Datum *vals, - bool *nulls, int * valcount, TYPCATEGORY tcategory, +array_dim_to_json(StringInfo result, int dim, int ndims, int *dims, Datum *vals, + bool *nulls, int *valcount, TYPCATEGORY tcategory, Oid typoutputfunc, bool use_line_feeds) { - int i; - char *sep; + int i; + char *sep; Assert(dim < ndims); @@ -759,7 +761,7 @@ array_dim_to_json(StringInfo result, int dim, int ndims,int * dims, Datum *vals, for (i = 1; i <= dims[dim]; i++) { if (i > 1) - appendStringInfoString(result,sep); + appendStringInfoString(result, sep); if (dim + 1 == ndims) { @@ -770,10 +772,10 @@ array_dim_to_json(StringInfo result, int dim, int ndims,int * dims, Datum *vals, else { /* - * Do we want line feeds on inner dimensions of arrays? - * For now we'll say no. + * Do we want line feeds on inner dimensions of arrays? For now + * we'll say no. */ - array_dim_to_json(result, dim+1, ndims, dims, vals, nulls, + array_dim_to_json(result, dim + 1, ndims, dims, vals, nulls, valcount, tcategory, typoutputfunc, false); } } @@ -792,9 +794,9 @@ array_to_json_internal(Datum array, StringInfo result, bool use_line_feeds) int *dim; int ndim; int nitems; - int count = 0; + int count = 0; Datum *elements; - bool *nulls; + bool *nulls; int16 typlen; bool typbyval; @@ -810,7 +812,7 @@ array_to_json_internal(Datum array, StringInfo result, bool use_line_feeds) if (nitems <= 0) { - appendStringInfoString(result,"[]"); + appendStringInfoString(result, "[]"); return; } @@ -842,52 +844,54 @@ array_to_json_internal(Datum array, StringInfo result, bool use_line_feeds) static void composite_to_json(Datum composite, StringInfo result, bool use_line_feeds) { - HeapTupleHeader td; - Oid tupType; - int32 tupTypmod; - TupleDesc tupdesc; - HeapTupleData tmptup, *tuple; - int i; - bool needsep = false; - char *sep; + HeapTupleHeader td; + Oid tupType; + int32 tupTypmod; + TupleDesc tupdesc; + HeapTupleData tmptup, + *tuple; + int i; + bool needsep = false; + char *sep; sep = use_line_feeds ? ",\n " : ","; - td = DatumGetHeapTupleHeader(composite); + td = DatumGetHeapTupleHeader(composite); - /* Extract rowtype info and find a tupdesc */ - tupType = HeapTupleHeaderGetTypeId(td); - tupTypmod = HeapTupleHeaderGetTypMod(td); - tupdesc = lookup_rowtype_tupdesc(tupType, tupTypmod); + /* Extract rowtype info and find a tupdesc */ + tupType = HeapTupleHeaderGetTypeId(td); + tupTypmod = HeapTupleHeaderGetTypMod(td); + tupdesc = lookup_rowtype_tupdesc(tupType, tupTypmod); - /* Build a temporary HeapTuple control structure */ - tmptup.t_len = HeapTupleHeaderGetDatumLength(td); - tmptup.t_data = td; + /* Build a temporary HeapTuple control structure */ + tmptup.t_len = HeapTupleHeaderGetDatumLength(td); + tmptup.t_data = td; tuple = &tmptup; - appendStringInfoChar(result,'{'); + appendStringInfoChar(result, '{'); - for (i = 0; i < tupdesc->natts; i++) - { - Datum val, origval; - bool isnull; - char *attname; + for (i = 0; i < tupdesc->natts; i++) + { + Datum val, + origval; + bool isnull; + char *attname; TYPCATEGORY tcategory; Oid typoutput; bool typisvarlena; if (tupdesc->attrs[i]->attisdropped) - continue; + continue; if (needsep) - appendStringInfoString(result,sep); + appendStringInfoString(result, sep); needsep = true; - attname = NameStr(tupdesc->attrs[i]->attname); - escape_json(result,attname); - appendStringInfoChar(result,':'); + attname = NameStr(tupdesc->attrs[i]->attname); + escape_json(result, attname); + appendStringInfoChar(result, ':'); - origval = heap_getattr(tuple, i + 1, tupdesc, &isnull); + origval = heap_getattr(tuple, i + 1, tupdesc, &isnull); if (tupdesc->attrs[i]->atttypid == RECORDARRAYOID) tcategory = TYPCATEGORY_ARRAY; @@ -902,10 +906,10 @@ composite_to_json(Datum composite, StringInfo result, bool use_line_feeds) &typoutput, &typisvarlena); /* - * If we have a toasted datum, forcibly detoast it here to avoid memory - * leakage inside the type's output routine. + * If we have a toasted datum, forcibly detoast it here to avoid + * memory leakage inside the type's output routine. */ - if (typisvarlena && ! isnull) + if (typisvarlena && !isnull) val = PointerGetDatum(PG_DETOAST_DATUM(origval)); else val = origval; @@ -917,8 +921,8 @@ composite_to_json(Datum composite, StringInfo result, bool use_line_feeds) pfree(DatumGetPointer(val)); } - appendStringInfoChar(result,'}'); - ReleaseTupleDesc(tupdesc); + appendStringInfoChar(result, '}'); + ReleaseTupleDesc(tupdesc); } /* @@ -927,7 +931,7 @@ composite_to_json(Datum composite, StringInfo result, bool use_line_feeds) extern Datum array_to_json(PG_FUNCTION_ARGS) { - Datum array = PG_GETARG_DATUM(0); + Datum array = PG_GETARG_DATUM(0); StringInfo result; result = makeStringInfo(); @@ -943,8 +947,8 @@ array_to_json(PG_FUNCTION_ARGS) extern Datum array_to_json_pretty(PG_FUNCTION_ARGS) { - Datum array = PG_GETARG_DATUM(0); - bool use_line_feeds = PG_GETARG_BOOL(1); + Datum array = PG_GETARG_DATUM(0); + bool use_line_feeds = PG_GETARG_BOOL(1); StringInfo result; result = makeStringInfo(); @@ -960,7 +964,7 @@ array_to_json_pretty(PG_FUNCTION_ARGS) extern Datum row_to_json(PG_FUNCTION_ARGS) { - Datum array = PG_GETARG_DATUM(0); + Datum array = PG_GETARG_DATUM(0); StringInfo result; result = makeStringInfo(); @@ -976,8 +980,8 @@ row_to_json(PG_FUNCTION_ARGS) extern Datum row_to_json_pretty(PG_FUNCTION_ARGS) { - Datum array = PG_GETARG_DATUM(0); - bool use_line_feeds = PG_GETARG_BOOL(1); + Datum array = PG_GETARG_DATUM(0); + bool use_line_feeds = PG_GETARG_BOOL(1); StringInfo result; result = makeStringInfo(); @@ -1031,4 +1035,3 @@ escape_json(StringInfo buf, const char *str) } appendStringInfoCharMacro(buf, '\"'); } - diff --git a/src/backend/utils/adt/lockfuncs.c b/src/backend/utils/adt/lockfuncs.c index ca1b1db18a622ddfa8e4ca09e2a4b3c347ee486f..33c5b64f50ac9b8f05a64f4e3a985d4cee2ceaca 100644 --- a/src/backend/utils/adt/lockfuncs.c +++ b/src/backend/utils/adt/lockfuncs.c @@ -160,7 +160,7 @@ pg_lock_status(PG_FUNCTION_ARGS) bool nulls[NUM_LOCK_STATUS_COLUMNS]; HeapTuple tuple; Datum result; - LockInstanceData *instance; + LockInstanceData *instance; instance = &(lockData->locks[mystatus->currIdx]); @@ -375,8 +375,8 @@ pg_lock_status(PG_FUNCTION_ARGS) nulls[11] = true; /* - * Lock mode. Currently all predicate locks are SIReadLocks, which - * are always held (never waiting) and have no fast path + * Lock mode. Currently all predicate locks are SIReadLocks, which are + * always held (never waiting) and have no fast path */ values[12] = CStringGetTextDatum("SIReadLock"); values[13] = BoolGetDatum(true); diff --git a/src/backend/utils/adt/mac.c b/src/backend/utils/adt/mac.c index 958ff54d73ee8b7666db51c7057ad97416066e66..aa9993fa5c6406fa7274ad61de270d5086781a5d 100644 --- a/src/backend/utils/adt/mac.c +++ b/src/backend/utils/adt/mac.c @@ -247,8 +247,8 @@ hashmacaddr(PG_FUNCTION_ARGS) Datum macaddr_not(PG_FUNCTION_ARGS) { - macaddr *addr = PG_GETARG_MACADDR_P(0); - macaddr *result; + macaddr *addr = PG_GETARG_MACADDR_P(0); + macaddr *result; result = (macaddr *) palloc(sizeof(macaddr)); result->a = ~addr->a; @@ -263,9 +263,9 @@ macaddr_not(PG_FUNCTION_ARGS) Datum macaddr_and(PG_FUNCTION_ARGS) { - macaddr *addr1 = PG_GETARG_MACADDR_P(0); - macaddr *addr2 = PG_GETARG_MACADDR_P(1); - macaddr *result; + macaddr *addr1 = PG_GETARG_MACADDR_P(0); + macaddr *addr2 = PG_GETARG_MACADDR_P(1); + macaddr *result; result = (macaddr *) palloc(sizeof(macaddr)); result->a = addr1->a & addr2->a; @@ -280,9 +280,9 @@ macaddr_and(PG_FUNCTION_ARGS) Datum macaddr_or(PG_FUNCTION_ARGS) { - macaddr *addr1 = PG_GETARG_MACADDR_P(0); - macaddr *addr2 = PG_GETARG_MACADDR_P(1); - macaddr *result; + macaddr *addr1 = PG_GETARG_MACADDR_P(0); + macaddr *addr2 = PG_GETARG_MACADDR_P(1); + macaddr *result; result = (macaddr *) palloc(sizeof(macaddr)); result->a = addr1->a | addr2->a; diff --git a/src/backend/utils/adt/misc.c b/src/backend/utils/adt/misc.c index 6bd7d531bbc8cb516e7ce24f80e87929bc744998..96e692766bfd6ffcb922acbe6c9219338425c5f7 100644 --- a/src/backend/utils/adt/misc.c +++ b/src/backend/utils/adt/misc.c @@ -329,14 +329,14 @@ pg_tablespace_databases(PG_FUNCTION_ARGS) Datum pg_tablespace_location(PG_FUNCTION_ARGS) { - Oid tablespaceOid = PG_GETARG_OID(0); - char sourcepath[MAXPGPATH]; - char targetpath[MAXPGPATH]; - int rllen; + Oid tablespaceOid = PG_GETARG_OID(0); + char sourcepath[MAXPGPATH]; + char targetpath[MAXPGPATH]; + int rllen; /* * It's useful to apply this function to pg_class.reltablespace, wherein - * zero means "the database's default tablespace". So, rather than + * zero means "the database's default tablespace". So, rather than * throwing an error for zero, we choose to assume that's what is meant. */ if (tablespaceOid == InvalidOid) @@ -350,9 +350,10 @@ pg_tablespace_location(PG_FUNCTION_ARGS) PG_RETURN_TEXT_P(cstring_to_text("")); #if defined(HAVE_READLINK) || defined(WIN32) + /* - * Find the location of the tablespace by reading the symbolic link that is - * in pg_tblspc/<oid>. + * Find the location of the tablespace by reading the symbolic link that + * is in pg_tblspc/<oid>. */ snprintf(sourcepath, sizeof(sourcepath), "pg_tblspc/%u", tablespaceOid); @@ -510,8 +511,8 @@ pg_typeof(PG_FUNCTION_ARGS) Datum pg_collation_for(PG_FUNCTION_ARGS) { - Oid typeid; - Oid collid; + Oid typeid; + Oid collid; typeid = get_fn_expr_argtype(fcinfo->flinfo, 0); if (!typeid) diff --git a/src/backend/utils/adt/numeric.c b/src/backend/utils/adt/numeric.c index 14bbdad93ba27a9d38c10fa13b1fbf502b06c8f0..68c1f1de3b77ecf5562fed710e4cff8492f27663 100644 --- a/src/backend/utils/adt/numeric.c +++ b/src/backend/utils/adt/numeric.c @@ -718,7 +718,7 @@ numeric_send(PG_FUNCTION_ARGS) * * Flatten calls to numeric's length coercion function that solely represent * increases in allowable precision. Scale changes mutate every datum, so - * they are unoptimizable. Some values, e.g. 1E-1001, can only fit into an + * they are unoptimizable. Some values, e.g. 1E-1001, can only fit into an * unconstrained numeric, so a change from an unconstrained numeric to any * constrained numeric is also unoptimizable. */ @@ -734,7 +734,7 @@ numeric_transform(PG_FUNCTION_ARGS) typmod = (Node *) lsecond(expr->args); - if (IsA(typmod, Const) && !((Const *) typmod)->constisnull) + if (IsA(typmod, Const) &&!((Const *) typmod)->constisnull) { Node *source = (Node *) linitial(expr->args); int32 old_typmod = exprTypmod(source); @@ -748,7 +748,7 @@ numeric_transform(PG_FUNCTION_ARGS) * If new_typmod < VARHDRSZ, the destination is unconstrained; that's * always OK. If old_typmod >= VARHDRSZ, the source is constrained, * and we're OK if the scale is unchanged and the precision is not - * decreasing. See further notes in function header comment. + * decreasing. See further notes in function header comment. */ if (new_typmod < (int32) VARHDRSZ || (old_typmod >= (int32) VARHDRSZ && @@ -1222,7 +1222,7 @@ width_bucket_numeric(PG_FUNCTION_ARGS) NUMERIC_IS_NAN(bound2)) ereport(ERROR, (errcode(ERRCODE_INVALID_ARGUMENT_FOR_WIDTH_BUCKET_FUNCTION), - errmsg("operand, lower bound, and upper bound cannot be NaN"))); + errmsg("operand, lower bound, and upper bound cannot be NaN"))); init_var(&result_var); init_var(&count_var); diff --git a/src/backend/utils/adt/pg_locale.c b/src/backend/utils/adt/pg_locale.c index 0920c13cd9ff7eae4412feb6fd3045628693d83a..de881bf6344908ab861ac36897c9183313b7437d 100644 --- a/src/backend/utils/adt/pg_locale.c +++ b/src/backend/utils/adt/pg_locale.c @@ -224,7 +224,7 @@ pg_perm_setlocale(int category, const char *locale) * Is the locale name valid for the locale category? * * If successful, and canonname isn't NULL, a palloc'd copy of the locale's - * canonical name is stored there. This is especially useful for figuring out + * canonical name is stored there. This is especially useful for figuring out * what locale name "" means (ie, the server environment value). (Actually, * it seems that on most implementations that's the only thing it's good for; * we could wish that setlocale gave back a canonically spelled version of @@ -578,7 +578,7 @@ strftime_win32(char *dst, size_t dstlen, const wchar_t *format, const struct tm len = WideCharToMultiByte(CP_UTF8, 0, wbuf, len, dst, dstlen, NULL, NULL); if (len == 0) elog(ERROR, - "could not convert string to UTF-8: error code %lu", GetLastError()); + "could not convert string to UTF-8: error code %lu", GetLastError()); dst[len] = '\0'; if (encoding != PG_UTF8) @@ -970,7 +970,7 @@ report_newlocale_failure(const char *localename) errdetail("The operating system could not find any locale data for the locale name \"%s\".", localename) : 0))); } -#endif /* HAVE_LOCALE_T */ +#endif /* HAVE_LOCALE_T */ /* diff --git a/src/backend/utils/adt/pgstatfuncs.c b/src/backend/utils/adt/pgstatfuncs.c index 83d0c229917b3e83f74025c6dd9f8a8f1cbcc522..7c0705abcc9eb355e2df71b5ec9e11e81ce9f7ec 100644 --- a/src/backend/utils/adt/pgstatfuncs.c +++ b/src/backend/utils/adt/pgstatfuncs.c @@ -1225,8 +1225,8 @@ pg_stat_get_db_stat_reset_time(PG_FUNCTION_ARGS) Datum pg_stat_get_db_temp_files(PG_FUNCTION_ARGS) { - Oid dbid = PG_GETARG_OID(0); - int64 result; + Oid dbid = PG_GETARG_OID(0); + int64 result; PgStat_StatDBEntry *dbentry; if ((dbentry = pgstat_fetch_stat_dbentry(dbid)) == NULL) @@ -1241,8 +1241,8 @@ pg_stat_get_db_temp_files(PG_FUNCTION_ARGS) Datum pg_stat_get_db_temp_bytes(PG_FUNCTION_ARGS) { - Oid dbid = PG_GETARG_OID(0); - int64 result; + Oid dbid = PG_GETARG_OID(0); + int64 result; PgStat_StatDBEntry *dbentry; if ((dbentry = pgstat_fetch_stat_dbentry(dbid)) == NULL) diff --git a/src/backend/utils/adt/rangetypes.c b/src/backend/utils/adt/rangetypes.c index 0994fa77cf6d475f952df969d0eba7769f0a6fcb..22ceb3c01d4b8e3ddf5577636bbbbc1ba93a036b 100644 --- a/src/backend/utils/adt/rangetypes.c +++ b/src/backend/utils/adt/rangetypes.c @@ -54,19 +54,19 @@ typedef struct RangeIOData static RangeIOData *get_range_io_data(FunctionCallInfo fcinfo, Oid rngtypid, - IOFuncSelector func); + IOFuncSelector func); static char range_parse_flags(const char *flags_str); static void range_parse(const char *input_str, char *flags, char **lbound_str, char **ubound_str); static const char *range_parse_bound(const char *string, const char *ptr, char **bound_str, bool *infinite); static char *range_deparse(char flags, const char *lbound_str, - const char *ubound_str); + const char *ubound_str); static char *range_bound_escape(const char *value); static bool range_contains_internal(TypeCacheEntry *typcache, - RangeType *r1, RangeType *r2); + RangeType *r1, RangeType *r2); static bool range_contains_elem_internal(TypeCacheEntry *typcache, - RangeType *r, Datum val); + RangeType *r, Datum val); static Size datum_compute_size(Size sz, Datum datum, bool typbyval, char typalign, int16 typlen, char typstorage); static Pointer datum_write(Pointer ptr, Datum datum, bool typbyval, @@ -299,10 +299,10 @@ get_range_io_data(FunctionCallInfo fcinfo, Oid rngtypid, IOFuncSelector func) if (cache == NULL || cache->typcache->type_id != rngtypid) { - int16 typlen; - bool typbyval; - char typalign; - char typdelim; + int16 typlen; + bool typbyval; + char typalign; + char typdelim; cache = (RangeIOData *) MemoryContextAlloc(fcinfo->flinfo->fn_mcxt, sizeof(RangeIOData)); @@ -326,13 +326,13 @@ get_range_io_data(FunctionCallInfo fcinfo, Oid rngtypid, IOFuncSelector func) if (func == IOFunc_receive) ereport(ERROR, (errcode(ERRCODE_UNDEFINED_FUNCTION), - errmsg("no binary input function available for type %s", - format_type_be(cache->typcache->rngelemtype->type_id)))); + errmsg("no binary input function available for type %s", + format_type_be(cache->typcache->rngelemtype->type_id)))); else ereport(ERROR, (errcode(ERRCODE_UNDEFINED_FUNCTION), - errmsg("no binary output function available for type %s", - format_type_be(cache->typcache->rngelemtype->type_id)))); + errmsg("no binary output function available for type %s", + format_type_be(cache->typcache->rngelemtype->type_id)))); } fmgr_info_cxt(cache->typiofunc, &cache->proc, fcinfo->flinfo->fn_mcxt); @@ -397,7 +397,7 @@ range_constructor3(PG_FUNCTION_ARGS) if (PG_ARGISNULL(2)) ereport(ERROR, (errcode(ERRCODE_DATA_EXCEPTION), - errmsg("range constructor flags argument must not be NULL"))); + errmsg("range constructor flags argument must not be NULL"))); flags = range_parse_flags(text_to_cstring(PG_GETARG_TEXT_P(2))); @@ -716,9 +716,9 @@ range_adjacent(PG_FUNCTION_ARGS) PG_RETURN_BOOL(false); /* - * Given two ranges A..B and C..D, where B < C, the ranges are adjacent - * if and only if the range B..C is empty, where inclusivity of these two - * bounds is inverted compared to the original bounds. For discrete + * Given two ranges A..B and C..D, where B < C, the ranges are adjacent if + * and only if the range B..C is empty, where inclusivity of these two + * bounds is inverted compared to the original bounds. For discrete * ranges, we have to rely on the canonicalization function to normalize * B..C to empty if it contains no elements of the subtype. (If there is * no canonicalization function, it's impossible for such a range to @@ -920,7 +920,7 @@ range_minus(PG_FUNCTION_ARGS) if (cmp_l1l2 < 0 && cmp_u1u2 > 0) ereport(ERROR, (errcode(ERRCODE_DATA_EXCEPTION), - errmsg("result of range difference would not be contiguous"))); + errmsg("result of range difference would not be contiguous"))); if (cmp_l1u2 > 0 || cmp_u1l2 < 0) PG_RETURN_RANGE(r1); @@ -1180,11 +1180,11 @@ Datum range_typanalyze(PG_FUNCTION_ARGS) { /* - * For the moment, just punt and don't analyze range columns. If we - * get close to release without having a better answer, we could - * consider letting std_typanalyze do what it can ... but those stats - * are probably next door to useless for most activity with range - * columns, so it's not clear it's worth gathering them. + * For the moment, just punt and don't analyze range columns. If we get + * close to release without having a better answer, we could consider + * letting std_typanalyze do what it can ... but those stats are probably + * next door to useless for most activity with range columns, so it's not + * clear it's worth gathering them. */ PG_RETURN_BOOL(false); } @@ -1392,7 +1392,7 @@ tstzrange_subdiff(PG_FUNCTION_ARGS) * * This is for use by range-related functions that follow the convention * of using the fn_extra field as a pointer to the type cache entry for - * the range type. Functions that need to cache more information than + * the range type. Functions that need to cache more information than * that must fend for themselves. */ TypeCacheEntry * @@ -1416,7 +1416,7 @@ range_get_typcache(FunctionCallInfo fcinfo, Oid rngtypid) * range_serialize: construct a range value from bounds and empty-flag * * This does not force canonicalization of the range value. In most cases, - * external callers should only be canonicalization functions. Note that + * external callers should only be canonicalization functions. Note that * we perform some datatype-independent canonicalization checks anyway. */ RangeType * @@ -1753,7 +1753,7 @@ range_cmp_bounds(TypeCacheEntry *typcache, RangeBound *b1, RangeBound *b2) * Compare two range boundary point values, returning <0, 0, or >0 according * to whether b1 is less than, equal to, or greater than b2. * - * This is similar to but simpler than range_cmp_bounds(). We just compare + * This is similar to but simpler than range_cmp_bounds(). We just compare * the values held in b1 and b2, ignoring inclusive/exclusive flags. The * lower/upper flags only matter for infinities, where they tell us if the * infinity is plus or minus. @@ -1971,7 +1971,7 @@ range_parse(const char *string, char *flags, char **lbound_str, } else if (*ptr == ')') ptr++; - else /* must be a comma */ + else /* must be a comma */ ereport(ERROR, (errcode(ERRCODE_INVALID_TEXT_REPRESENTATION), errmsg("malformed range literal: \"%s\"", @@ -2224,7 +2224,7 @@ range_contains_elem_internal(TypeCacheEntry *typcache, RangeType *r, Datum val) /* * datum_compute_size() and datum_write() are used to insert the bound - * values into a range object. They are modeled after heaptuple.c's + * values into a range object. They are modeled after heaptuple.c's * heap_compute_data_size() and heap_fill_tuple(), but we need not handle * null values here. TYPE_IS_PACKABLE must test the same conditions as * heaptuple.c's ATT_IS_PACKABLE macro. diff --git a/src/backend/utils/adt/rangetypes_gist.c b/src/backend/utils/adt/rangetypes_gist.c index 87f71e6812c3587eb854f6f5e1d7cd052b27f279..16103f854bcb92b27cb03ea96a170423164ea5f8 100644 --- a/src/backend/utils/adt/rangetypes_gist.c +++ b/src/backend/utils/adt/rangetypes_gist.c @@ -57,7 +57,7 @@ #define LIMIT_RATIO 0.3 /* Constants for fixed penalty values */ -#define INFINITE_BOUND_PENALTY 2.0 +#define INFINITE_BOUND_PENALTY 2.0 #define CONTAIN_EMPTY_PENALTY 1.0 #define DEFAULT_SUBTYPE_DIFF_PENALTY 1.0 @@ -66,8 +66,8 @@ */ typedef struct { - int index; - RangeBound bound; + int index; + RangeBound bound; } SingleBoundSortItem; /* place on left or right side of split? */ @@ -83,15 +83,15 @@ typedef enum typedef struct { TypeCacheEntry *typcache; /* typcache for range type */ - bool has_subtype_diff; /* does it have subtype_diff? */ + bool has_subtype_diff; /* does it have subtype_diff? */ int entries_count; /* total number of entries being split */ /* Information about currently selected split follows */ bool first; /* true if no split was selected yet */ - RangeBound *left_upper; /* upper bound of left interval */ - RangeBound *right_lower; /* lower bound of right interval */ + RangeBound *left_upper; /* upper bound of left interval */ + RangeBound *right_lower; /* lower bound of right interval */ float4 ratio; /* split ratio */ float4 overlap; /* overlap between left and right predicate */ @@ -146,8 +146,8 @@ typedef struct ((RangeType *) DatumGetPointer(datumCopy(PointerGetDatum(r), \ false, -1))) -static RangeType *range_super_union(TypeCacheEntry *typcache, RangeType * r1, - RangeType * r2); +static RangeType *range_super_union(TypeCacheEntry *typcache, RangeType *r1, + RangeType *r2); static bool range_gist_consistent_int(FmgrInfo *flinfo, StrategyNumber strategy, RangeType *key, Datum query); @@ -155,19 +155,19 @@ static bool range_gist_consistent_leaf(FmgrInfo *flinfo, StrategyNumber strategy, RangeType *key, Datum query); static void range_gist_fallback_split(TypeCacheEntry *typcache, - GistEntryVector *entryvec, - GIST_SPLITVEC *v); + GistEntryVector *entryvec, + GIST_SPLITVEC *v); static void range_gist_class_split(TypeCacheEntry *typcache, - GistEntryVector *entryvec, - GIST_SPLITVEC *v, - SplitLR *classes_groups); + GistEntryVector *entryvec, + GIST_SPLITVEC *v, + SplitLR *classes_groups); static void range_gist_single_sorting_split(TypeCacheEntry *typcache, - GistEntryVector *entryvec, - GIST_SPLITVEC *v, - bool use_upper_bound); + GistEntryVector *entryvec, + GIST_SPLITVEC *v, + bool use_upper_bound); static void range_gist_double_sorting_split(TypeCacheEntry *typcache, - GistEntryVector *entryvec, - GIST_SPLITVEC *v); + GistEntryVector *entryvec, + GIST_SPLITVEC *v); static void range_gist_consider_split(ConsiderSplitContext *context, RangeBound *right_lower, int min_left_count, RangeBound *left_upper, int max_left_count); @@ -177,7 +177,7 @@ static int interval_cmp_lower(const void *a, const void *b, void *arg); static int interval_cmp_upper(const void *a, const void *b, void *arg); static int common_entry_cmp(const void *i1, const void *i2); static float8 call_subtype_diff(TypeCacheEntry *typcache, - Datum val1, Datum val2); + Datum val1, Datum val2); /* GiST query consistency check */ @@ -187,6 +187,7 @@ range_gist_consistent(PG_FUNCTION_ARGS) GISTENTRY *entry = (GISTENTRY *) PG_GETARG_POINTER(0); Datum query = PG_GETARG_DATUM(1); StrategyNumber strategy = (StrategyNumber) PG_GETARG_UINT16(2); + /* Oid subtype = PG_GETARG_OID(3); */ bool *recheck = (bool *) PG_GETARG_POINTER(4); RangeType *key = DatumGetRangeType(entry->key); @@ -280,9 +281,9 @@ range_gist_penalty(PG_FUNCTION_ARGS) range_deserialize(typcache, new, &new_lower, &new_upper, &new_empty); /* - * Distinct branches for handling distinct classes of ranges. Note - * that penalty values only need to be commensurate within the same - * class of new range. + * Distinct branches for handling distinct classes of ranges. Note that + * penalty values only need to be commensurate within the same class of + * new range. */ if (new_empty) { @@ -290,9 +291,9 @@ range_gist_penalty(PG_FUNCTION_ARGS) if (orig_empty) { /* - * The best case is to insert it to empty original - * range. Insertion here means no broadening of original range. - * Also original range is the most narrow. + * The best case is to insert it to empty original range. + * Insertion here means no broadening of original range. Also + * original range is the most narrow. */ *penalty = 0.0; } @@ -309,7 +310,7 @@ range_gist_penalty(PG_FUNCTION_ARGS) else if (orig_lower.infinite && orig_upper.infinite) { /* - * Original range requires broadening. (-inf; +inf) is most far + * Original range requires broadening. (-inf; +inf) is most far * from normal range in this case. */ *penalty = 2 * CONTAIN_EMPTY_PENALTY; @@ -360,8 +361,8 @@ range_gist_penalty(PG_FUNCTION_ARGS) if (RangeIsOrContainsEmpty(orig)) { /* - * Original range is narrower when it doesn't contain empty ranges. - * Add additional penalty otherwise. + * Original range is narrower when it doesn't contain empty + * ranges. Add additional penalty otherwise. */ *penalty += CONTAIN_EMPTY_PENALTY; } @@ -374,11 +375,11 @@ range_gist_penalty(PG_FUNCTION_ARGS) if (orig_upper.infinite) { /* - * (-inf, +inf) range won't be extended by insertion of - * (-inf, x) range. It's a less desirable case than insertion - * to (-inf, y) original range without extension, because in - * that case original range is narrower. But we can't express - * that in single float value. + * (-inf, +inf) range won't be extended by insertion of (-inf, + * x) range. It's a less desirable case than insertion to + * (-inf, y) original range without extension, because in that + * case original range is narrower. But we can't express that + * in single float value. */ *penalty = 0.0; } @@ -387,8 +388,8 @@ range_gist_penalty(PG_FUNCTION_ARGS) if (range_cmp_bounds(typcache, &new_upper, &orig_upper) > 0) { /* - * Get extension of original range using subtype_diff. - * Use constant if subtype_diff unavailable. + * Get extension of original range using subtype_diff. Use + * constant if subtype_diff unavailable. */ if (has_subtype_diff) *penalty = call_subtype_diff(typcache, @@ -407,8 +408,8 @@ range_gist_penalty(PG_FUNCTION_ARGS) else { /* - * If lower bound of original range is not -inf, then extension - * of it is infinity. + * If lower bound of original range is not -inf, then extension of + * it is infinity. */ *penalty = get_float4_infinity(); } @@ -421,11 +422,11 @@ range_gist_penalty(PG_FUNCTION_ARGS) if (orig_lower.infinite) { /* - * (-inf, +inf) range won't be extended by insertion of - * (x, +inf) range. It's a less desirable case than insertion - * to (y, +inf) original range without extension, because in - * that case original range is narrower. But we can't express - * that in single float value. + * (-inf, +inf) range won't be extended by insertion of (x, + * +inf) range. It's a less desirable case than insertion to + * (y, +inf) original range without extension, because in that + * case original range is narrower. But we can't express that + * in single float value. */ *penalty = 0.0; } @@ -434,8 +435,8 @@ range_gist_penalty(PG_FUNCTION_ARGS) if (range_cmp_bounds(typcache, &new_lower, &orig_lower) < 0) { /* - * Get extension of original range using subtype_diff. - * Use constant if subtype_diff unavailable. + * Get extension of original range using subtype_diff. Use + * constant if subtype_diff unavailable. */ if (has_subtype_diff) *penalty = call_subtype_diff(typcache, @@ -454,8 +455,8 @@ range_gist_penalty(PG_FUNCTION_ARGS) else { /* - * If upper bound of original range is not +inf, then extension - * of it is infinity. + * If upper bound of original range is not +inf, then extension of + * it is infinity. */ *penalty = get_float4_infinity(); } @@ -506,7 +507,7 @@ range_gist_penalty(PG_FUNCTION_ARGS) /* * The GiST PickSplit method for ranges * - * Primarily, we try to segregate ranges of different classes. If splitting + * Primarily, we try to segregate ranges of different classes. If splitting * ranges of the same class, use the appropriate split method for that class. */ Datum @@ -541,7 +542,7 @@ range_gist_picksplit(PG_FUNCTION_ARGS) memset(count_in_classes, 0, sizeof(count_in_classes)); for (i = FirstOffsetNumber; i <= maxoff; i = OffsetNumberNext(i)) { - RangeType *range = DatumGetRangeType(entryvec->vector[i].key); + RangeType *range = DatumGetRangeType(entryvec->vector[i].key); count_in_classes[get_gist_range_class(range)]++; } @@ -597,7 +598,7 @@ range_gist_picksplit(PG_FUNCTION_ARGS) * To which side of the split should each class go? Initialize them * all to go to the left side. */ - SplitLR classes_groups[CLS_COUNT]; + SplitLR classes_groups[CLS_COUNT]; memset(classes_groups, 0, sizeof(classes_groups)); @@ -610,16 +611,18 @@ range_gist_picksplit(PG_FUNCTION_ARGS) { /*---------- * Try to split classes in one of two ways: - * 1) containing infinities - not containing infinities - * 2) containing empty - not containing empty + * 1) containing infinities - not containing infinities + * 2) containing empty - not containing empty * * Select the way which balances the ranges between left and right * the best. If split in these ways is not possible, there are at * most 3 classes, so just separate biggest class. *---------- */ - int infCount, nonInfCount; - int emptyCount, nonEmptyCount; + int infCount, + nonInfCount; + int emptyCount, + nonEmptyCount; nonInfCount = count_in_classes[CLS_NORMAL] + @@ -628,7 +631,7 @@ range_gist_picksplit(PG_FUNCTION_ARGS) infCount = total_count - nonInfCount; nonEmptyCount = - count_in_classes[CLS_NORMAL] + + count_in_classes[CLS_NORMAL] + count_in_classes[CLS_LOWER_INF] + count_in_classes[CLS_UPPER_INF] + count_in_classes[CLS_LOWER_INF | CLS_UPPER_INF]; @@ -638,21 +641,22 @@ range_gist_picksplit(PG_FUNCTION_ARGS) (Abs(infCount - nonInfCount) <= Abs(emptyCount - nonEmptyCount))) { - classes_groups[CLS_NORMAL] = SPLIT_RIGHT; + classes_groups[CLS_NORMAL] = SPLIT_RIGHT; classes_groups[CLS_CONTAIN_EMPTY] = SPLIT_RIGHT; - classes_groups[CLS_EMPTY] = SPLIT_RIGHT; + classes_groups[CLS_EMPTY] = SPLIT_RIGHT; } else if (emptyCount > 0 && nonEmptyCount > 0) { - classes_groups[CLS_NORMAL] = SPLIT_RIGHT; - classes_groups[CLS_LOWER_INF] = SPLIT_RIGHT; - classes_groups[CLS_UPPER_INF] = SPLIT_RIGHT; + classes_groups[CLS_NORMAL] = SPLIT_RIGHT; + classes_groups[CLS_LOWER_INF] = SPLIT_RIGHT; + classes_groups[CLS_UPPER_INF] = SPLIT_RIGHT; classes_groups[CLS_LOWER_INF | CLS_UPPER_INF] = SPLIT_RIGHT; } else { /* - * Either total_count == emptyCount or total_count == infCount. + * Either total_count == emptyCount or total_count == + * infCount. */ classes_groups[biggest_class] = SPLIT_RIGHT; } @@ -673,10 +677,10 @@ range_gist_same(PG_FUNCTION_ARGS) bool *result = (bool *) PG_GETARG_POINTER(2); /* - * range_eq will ignore the RANGE_CONTAIN_EMPTY flag, so we have to - * check that for ourselves. More generally, if the entries have been - * properly normalized, then unequal flags bytes must mean unequal ranges - * ... so let's just test all the flag bits at once. + * range_eq will ignore the RANGE_CONTAIN_EMPTY flag, so we have to check + * that for ourselves. More generally, if the entries have been properly + * normalized, then unequal flags bytes must mean unequal ranges ... so + * let's just test all the flag bits at once. */ if (range_get_flags(r1) != range_get_flags(r2)) *result = false; @@ -710,7 +714,7 @@ range_gist_same(PG_FUNCTION_ARGS) * that *all* unions formed within the GiST index must go through here. */ static RangeType * -range_super_union(TypeCacheEntry *typcache, RangeType * r1, RangeType * r2) +range_super_union(TypeCacheEntry *typcache, RangeType *r1, RangeType *r2) { RangeType *result; RangeBound lower1, @@ -862,9 +866,10 @@ range_gist_consistent_int(FmgrInfo *flinfo, StrategyNumber strategy, proc = range_contains; break; case RANGESTRAT_CONTAINED_BY: + /* * Empty ranges are contained by anything, so if key is or - * contains any empty ranges, we must descend into it. Otherwise, + * contains any empty ranges, we must descend into it. Otherwise, * descend only if key overlaps the query. */ if (RangeIsOrContainsEmpty(key)) @@ -875,6 +880,7 @@ range_gist_consistent_int(FmgrInfo *flinfo, StrategyNumber strategy, proc = range_contains_elem; break; case RANGESTRAT_EQ: + /* * If query is empty, descend only if the key is or contains any * empty ranges. Otherwise, descend if key contains query. @@ -959,9 +965,11 @@ range_gist_fallback_split(TypeCacheEntry *typcache, GistEntryVector *entryvec, GIST_SPLITVEC *v) { - RangeType *left_range = NULL; - RangeType *right_range = NULL; - OffsetNumber i, maxoff, split_idx; + RangeType *left_range = NULL; + RangeType *right_range = NULL; + OffsetNumber i, + maxoff, + split_idx; maxoff = entryvec->n - 1; /* Split entries before this to left page, after to right: */ @@ -971,7 +979,7 @@ range_gist_fallback_split(TypeCacheEntry *typcache, v->spl_nright = 0; for (i = FirstOffsetNumber; i <= maxoff; i++) { - RangeType *range = DatumGetRangeType(entryvec->vector[i].key); + RangeType *range = DatumGetRangeType(entryvec->vector[i].key); if (i < split_idx) PLACE_LEFT(range, i); @@ -996,9 +1004,10 @@ range_gist_class_split(TypeCacheEntry *typcache, GIST_SPLITVEC *v, SplitLR *classes_groups) { - RangeType *left_range = NULL; - RangeType *right_range = NULL; - OffsetNumber i, maxoff; + RangeType *left_range = NULL; + RangeType *right_range = NULL; + OffsetNumber i, + maxoff; maxoff = entryvec->n - 1; @@ -1006,8 +1015,8 @@ range_gist_class_split(TypeCacheEntry *typcache, v->spl_nright = 0; for (i = FirstOffsetNumber; i <= maxoff; i = OffsetNumberNext(i)) { - RangeType *range = DatumGetRangeType(entryvec->vector[i].key); - int class; + RangeType *range = DatumGetRangeType(entryvec->vector[i].key); + int class; /* Get class of range */ class = get_gist_range_class(range); @@ -1038,10 +1047,12 @@ range_gist_single_sorting_split(TypeCacheEntry *typcache, GIST_SPLITVEC *v, bool use_upper_bound) { - SingleBoundSortItem *sortItems; - RangeType *left_range = NULL; - RangeType *right_range = NULL; - OffsetNumber i, maxoff, split_idx; + SingleBoundSortItem *sortItems; + RangeType *left_range = NULL; + RangeType *right_range = NULL; + OffsetNumber i, + maxoff, + split_idx; maxoff = entryvec->n - 1; @@ -1053,9 +1064,9 @@ range_gist_single_sorting_split(TypeCacheEntry *typcache, */ for (i = FirstOffsetNumber; i <= maxoff; i = OffsetNumberNext(i)) { - RangeType *range = DatumGetRangeType(entryvec->vector[i].key); - RangeBound bound2; - bool empty; + RangeType *range = DatumGetRangeType(entryvec->vector[i].key); + RangeBound bound2; + bool empty; sortItems[i - 1].index = i; /* Put appropriate bound into array */ @@ -1078,8 +1089,8 @@ range_gist_single_sorting_split(TypeCacheEntry *typcache, for (i = 0; i < maxoff; i++) { - int idx = sortItems[i].index; - RangeType *range = DatumGetRangeType(entryvec->vector[idx].key); + int idx = sortItems[i].index; + RangeType *range = DatumGetRangeType(entryvec->vector[idx].key); if (i < split_idx) PLACE_LEFT(range, idx); @@ -1125,16 +1136,20 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, GIST_SPLITVEC *v) { ConsiderSplitContext context; - OffsetNumber i, maxoff; - RangeType *range, - *left_range = NULL, - *right_range = NULL; - int common_entries_count; + OffsetNumber i, + maxoff; + RangeType *range, + *left_range = NULL, + *right_range = NULL; + int common_entries_count; NonEmptyRange *by_lower, - *by_upper; + *by_upper; CommonEntry *common_entries; - int nentries, i1, i2; - RangeBound *right_lower, *left_upper; + int nentries, + i1, + i2; + RangeBound *right_lower, + *left_upper; memset(&context, 0, sizeof(ConsiderSplitContext)); context.typcache = typcache; @@ -1151,8 +1166,8 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, /* Fill arrays of bounds */ for (i = FirstOffsetNumber; i <= maxoff; i = OffsetNumberNext(i)) { - RangeType *range = DatumGetRangeType(entryvec->vector[i].key); - bool empty; + RangeType *range = DatumGetRangeType(entryvec->vector[i].key); + bool empty; range_deserialize(typcache, range, &by_lower[i - FirstOffsetNumber].lower, @@ -1209,7 +1224,7 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, i1 = 0; i2 = 0; right_lower = &by_lower[i1].lower; - left_upper = &by_upper[i2].lower; + left_upper = &by_upper[i2].lower; while (true) { /* @@ -1229,8 +1244,8 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, right_lower = &by_lower[i1].lower; /* - * Find count of ranges which anyway should be placed to the - * left group. + * Find count of ranges which anyway should be placed to the left + * group. */ while (i2 < nentries && range_cmp_bounds(typcache, &by_upper[i2].upper, @@ -1244,13 +1259,13 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, } /* - * Iterate over upper bound of left group finding greatest possible - * lower bound of right group. + * Iterate over upper bound of left group finding greatest possible lower + * bound of right group. */ i1 = nentries - 1; i2 = nentries - 1; right_lower = &by_lower[i1].upper; - left_upper = &by_upper[i2].upper; + left_upper = &by_upper[i2].upper; while (true) { /* @@ -1270,8 +1285,8 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, left_upper = &by_upper[i2].upper; /* - * Find count of intervals which anyway should be placed to the - * right group. + * Find count of intervals which anyway should be placed to the right + * group. */ while (i1 >= 0 && range_cmp_bounds(typcache, &by_lower[i1].lower, @@ -1295,9 +1310,9 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, } /* - * Ok, we have now selected bounds of the groups. Now we have to distribute - * entries themselves. At first we distribute entries which can be placed - * unambiguously and collect "common entries" to array. + * Ok, we have now selected bounds of the groups. Now we have to + * distribute entries themselves. At first we distribute entries which can + * be placed unambiguously and collect "common entries" to array. */ /* Allocate vectors for results */ @@ -1394,7 +1409,7 @@ range_gist_double_sorting_split(TypeCacheEntry *typcache, */ for (i = 0; i < common_entries_count; i++) { - int idx = common_entries[i].index; + int idx = common_entries[i].index; range = DatumGetRangeType(entryvec->vector[idx].key); @@ -1530,8 +1545,8 @@ get_gist_range_class(RangeType *range) static int single_bound_cmp(const void *a, const void *b, void *arg) { - SingleBoundSortItem *i1 = (SingleBoundSortItem *) a; - SingleBoundSortItem *i2 = (SingleBoundSortItem *) b; + SingleBoundSortItem *i1 = (SingleBoundSortItem *) a; + SingleBoundSortItem *i2 = (SingleBoundSortItem *) b; TypeCacheEntry *typcache = (TypeCacheEntry *) arg; return range_cmp_bounds(typcache, &i1->bound, &i2->bound); diff --git a/src/backend/utils/adt/ruleutils.c b/src/backend/utils/adt/ruleutils.c index 7ad99a0ec32579760f6e10278ec8585a0b2ab855..9ca3b9d0c44c71522309c2da61071e096825cdf0 100644 --- a/src/backend/utils/adt/ruleutils.c +++ b/src/backend/utils/adt/ruleutils.c @@ -73,7 +73,7 @@ #define PRETTYFLAG_PAREN 1 #define PRETTYFLAG_INDENT 2 -#define PRETTY_WRAP_DEFAULT 79 +#define PRETTY_WRAP_DEFAULT 79 /* macro to test if pretty action needed */ #define PRETTY_PAREN(context) ((context)->prettyFlags & PRETTYFLAG_PAREN) @@ -138,7 +138,7 @@ static SPIPlanPtr plan_getrulebyoid = NULL; static const char *query_getrulebyoid = "SELECT * FROM pg_catalog.pg_rewrite WHERE oid = $1"; static SPIPlanPtr plan_getviewrule = NULL; static const char *query_getviewrule = "SELECT * FROM pg_catalog.pg_rewrite WHERE ev_class = $1 AND rulename = $2"; -static int pretty_wrap = PRETTY_WRAP_DEFAULT; +static int pretty_wrap = PRETTY_WRAP_DEFAULT; /* GUC parameters */ bool quote_all_identifiers = false; @@ -388,9 +388,9 @@ pg_get_viewdef_wrap(PG_FUNCTION_ARGS) { /* By OID */ Oid viewoid = PG_GETARG_OID(0); - int wrap = PG_GETARG_INT32(1); + int wrap = PG_GETARG_INT32(1); int prettyFlags; - char *result; + char *result; /* calling this implies we want pretty printing */ prettyFlags = PRETTYFLAG_PAREN | PRETTYFLAG_INDENT; @@ -1335,10 +1335,10 @@ pg_get_constraintdef_worker(Oid constraintId, bool fullCommand, * Now emit the constraint definition, adding NO INHERIT if * necessary. * - * There are cases where - * the constraint expression will be fully parenthesized and - * we don't need the outer parens ... but there are other - * cases where we do need 'em. Be conservative for now. + * There are cases where the constraint expression will be + * fully parenthesized and we don't need the outer parens ... + * but there are other cases where we do need 'em. Be + * conservative for now. * * Note that simply checking for leading '(' and trailing ')' * would NOT be good enough, consider "(x > 0) AND (y > 0)". @@ -1599,7 +1599,7 @@ pg_get_serial_sequence(PG_FUNCTION_ARGS) SysScanDesc scan; HeapTuple tup; - /* Look up table name. Can't lock it - we might not have privileges. */ + /* Look up table name. Can't lock it - we might not have privileges. */ tablerv = makeRangeVarFromNameList(textToQualifiedNameList(tablename)); tableOid = RangeVarGetRelid(tablerv, NoLock, false); @@ -3038,7 +3038,7 @@ get_target_list(List *targetList, deparse_context *context, char *sep; int colno; ListCell *l; - bool last_was_multiline = false; + bool last_was_multiline = false; sep = " "; colno = 0; @@ -3048,9 +3048,9 @@ get_target_list(List *targetList, deparse_context *context, char *colname; char *attname; StringInfoData targetbuf; - int leading_nl_pos = -1; - char *trailing_nl; - int pos; + int leading_nl_pos = -1; + char *trailing_nl; + int pos; if (tle->resjunk) continue; /* ignore junk entries */ @@ -3060,9 +3060,8 @@ get_target_list(List *targetList, deparse_context *context, colno++; /* - * Put the new field spec into targetbuf so we can - * decide after we've got it whether or not it needs - * to go on a new line. + * Put the new field spec into targetbuf so we can decide after we've + * got it whether or not it needs to go on a new line. */ initStringInfo(&targetbuf); @@ -3112,7 +3111,7 @@ get_target_list(List *targetList, deparse_context *context, /* Does the new field start with whitespace plus a new line? */ - for (pos=0; pos < targetbuf.len; pos++) + for (pos = 0; pos < targetbuf.len; pos++) { if (targetbuf.data[pos] == '\n') { @@ -3123,30 +3122,29 @@ get_target_list(List *targetList, deparse_context *context, break; } - /* Locate the start of the current line in the buffer */ + /* Locate the start of the current line in the buffer */ - trailing_nl = (strrchr(buf->data,'\n')); + trailing_nl = (strrchr(buf->data, '\n')); if (trailing_nl == NULL) trailing_nl = buf->data; - else + else trailing_nl++; /* - * If the field we're adding is the first in the list, or it already - * has a leading newline, or wrap mode is disabled (pretty_wrap < 0), - * don't add anything. - * Otherwise, add a newline, plus some indentation, if either the - * new field would cause an overflow or the last field used more than - * one line. + * If the field we're adding is the first in the list, or it already + * has a leading newline, or wrap mode is disabled (pretty_wrap < 0), + * don't add anything. Otherwise, add a newline, plus some + * indentation, if either the new field would cause an overflow or the + * last field used more than one line. */ if (colno > 1 && - leading_nl_pos == -1 && + leading_nl_pos == -1 && pretty_wrap >= 0 && ((strlen(trailing_nl) + strlen(targetbuf.data) > pretty_wrap) || last_was_multiline)) { - appendContextKeyword(context, "", -PRETTYINDENT_STD, + appendContextKeyword(context, "", -PRETTYINDENT_STD, PRETTYINDENT_STD, PRETTYINDENT_VAR); } @@ -3157,12 +3155,12 @@ get_target_list(List *targetList, deparse_context *context, /* Keep track of this field's status for next iteration */ - last_was_multiline = - (strchr(targetbuf.data + leading_nl_pos + 1,'\n') != NULL); + last_was_multiline = + (strchr(targetbuf.data + leading_nl_pos + 1, '\n') != NULL); /* cleanup */ - pfree (targetbuf.data); + pfree(targetbuf.data); } } @@ -4049,7 +4047,7 @@ get_variable(Var *var, int levelsup, bool istoplevel, deparse_context *context) /* - * Get the name of a field of an expression of composite type. The + * Get the name of a field of an expression of composite type. The * expression is usually a Var, but we handle other cases too. * * levelsup is an extra offset to interpret the Var's varlevelsup correctly. @@ -4059,7 +4057,7 @@ get_variable(Var *var, int levelsup, bool istoplevel, deparse_context *context) * could also be RECORD. Since no actual table or view column is allowed to * have type RECORD, a Var of type RECORD must refer to a JOIN or FUNCTION RTE * or to a subquery output. We drill down to find the ultimate defining - * expression and attempt to infer the field name from it. We ereport if we + * expression and attempt to infer the field name from it. We ereport if we * can't determine the name. * * Similarly, a PARAM of type RECORD has to refer to some expression of @@ -4483,7 +4481,7 @@ find_rte_by_refname(const char *refname, deparse_context *context) * reference a parameter supplied by an upper NestLoop or SubPlan plan node. * * If successful, return the expression and set *dpns_p and *ancestor_cell_p - * appropriately for calling push_ancestor_plan(). If no referent can be + * appropriately for calling push_ancestor_plan(). If no referent can be * found, return NULL. */ static Node * @@ -4615,7 +4613,7 @@ get_parameter(Param *param, deparse_context *context) /* * If it's a PARAM_EXEC parameter, try to locate the expression from which - * the parameter was computed. Note that failing to find a referent isn't + * the parameter was computed. Note that failing to find a referent isn't * an error, since the Param might well be a subplan output rather than an * input. */ @@ -6567,10 +6565,10 @@ get_from_clause(Query *query, const char *prefix, deparse_context *context) else { StringInfoData targetbuf; - char *trailing_nl; + char *trailing_nl; appendStringInfoString(buf, ", "); - + initStringInfo(&targetbuf); context->buf = &targetbuf; @@ -6578,33 +6576,33 @@ get_from_clause(Query *query, const char *prefix, deparse_context *context) context->buf = buf; - /* Locate the start of the current line in the buffer */ + /* Locate the start of the current line in the buffer */ - trailing_nl = (strrchr(buf->data,'\n')); + trailing_nl = (strrchr(buf->data, '\n')); if (trailing_nl == NULL) trailing_nl = buf->data; - else + else trailing_nl++; - + /* - * Add a newline, plus some indentation, if pretty_wrap is on and the - * new from-clause item would cause an overflow. + * Add a newline, plus some indentation, if pretty_wrap is on and + * the new from-clause item would cause an overflow. */ - + if (pretty_wrap >= 0 && (strlen(trailing_nl) + strlen(targetbuf.data) > pretty_wrap)) { - appendContextKeyword(context, "", -PRETTYINDENT_STD, + appendContextKeyword(context, "", -PRETTYINDENT_STD, PRETTYINDENT_STD, PRETTYINDENT_VAR); } /* Add the new item */ appendStringInfoString(buf, targetbuf.data); - + /* cleanup */ - pfree (targetbuf.data); + pfree(targetbuf.data); } } diff --git a/src/backend/utils/adt/selfuncs.c b/src/backend/utils/adt/selfuncs.c index 83e43a999722e396ff3466813839d7e069472d07..95e46276f0a8911758f4ec02b993193bf55eee15 100644 --- a/src/backend/utils/adt/selfuncs.c +++ b/src/backend/utils/adt/selfuncs.c @@ -258,7 +258,7 @@ var_eq_const(VariableStatData *vardata, Oid operator, /* * If we matched the var to a unique index or DISTINCT clause, assume - * there is exactly one match regardless of anything else. (This is + * there is exactly one match regardless of anything else. (This is * slightly bogus, since the index or clause's equality operator might be * different from ours, but it's much more likely to be right than * ignoring the information.) @@ -393,7 +393,7 @@ var_eq_non_const(VariableStatData *vardata, Oid operator, /* * If we matched the var to a unique index or DISTINCT clause, assume - * there is exactly one match regardless of anything else. (This is + * there is exactly one match regardless of anything else. (This is * slightly bogus, since the index or clause's equality operator might be * different from ours, but it's much more likely to be right than * ignoring the information.) @@ -1743,8 +1743,8 @@ scalararraysel(PlannerInfo *root, } /* - * If it is equality or inequality, we might be able to estimate this as - * a form of array containment; for instance "const = ANY(column)" can be + * If it is equality or inequality, we might be able to estimate this as a + * form of array containment; for instance "const = ANY(column)" can be * treated as "ARRAY[const] <@ column". scalararraysel_containment tries * that, and returns the selectivity estimate if successful, or -1 if not. */ @@ -1819,7 +1819,7 @@ scalararraysel(PlannerInfo *root, /* * For generic operators, we assume the probability of success is - * independent for each array element. But for "= ANY" or "<> ALL", + * independent for each array element. But for "= ANY" or "<> ALL", * if the array elements are distinct (which'd typically be the case) * then the probabilities are disjoint, and we should just sum them. * @@ -2132,6 +2132,7 @@ eqjoinsel(PG_FUNCTION_ARGS) break; case JOIN_SEMI: case JOIN_ANTI: + /* * Look up the join's inner relation. min_righthand is sufficient * information because neither SEMI nor ANTI joins permit any @@ -2423,7 +2424,7 @@ eqjoinsel_semi(Oid operator, /* * We clamp nd2 to be not more than what we estimate the inner relation's - * size to be. This is intuitively somewhat reasonable since obviously + * size to be. This is intuitively somewhat reasonable since obviously * there can't be more than that many distinct values coming from the * inner rel. The reason for the asymmetry (ie, that we don't clamp nd1 * likewise) is that this is the only pathway by which restriction clauses @@ -3879,7 +3880,7 @@ convert_string_datum(Datum value, Oid typid) { char *xfrmstr; size_t xfrmlen; - size_t xfrmlen2 PG_USED_FOR_ASSERTS_ONLY; + size_t xfrmlen2 PG_USED_FOR_ASSERTS_ONLY; /* * Note: originally we guessed at a suitable output buffer size, and @@ -4475,7 +4476,7 @@ examine_simple_variable(PlannerInfo *root, Var *var, * Punt if subquery uses set operations or GROUP BY, as these will * mash underlying columns' stats beyond recognition. (Set ops are * particularly nasty; if we forged ahead, we would return stats - * relevant to only the leftmost subselect...) DISTINCT is also + * relevant to only the leftmost subselect...) DISTINCT is also * problematic, but we check that later because there is a possibility * of learning something even with it. */ @@ -4496,12 +4497,12 @@ examine_simple_variable(PlannerInfo *root, Var *var, Assert(rel->subroot && IsA(rel->subroot, PlannerInfo)); /* - * Switch our attention to the subquery as mangled by the planner. - * It was okay to look at the pre-planning version for the tests - * above, but now we need a Var that will refer to the subroot's - * live RelOptInfos. For instance, if any subquery pullup happened - * during planning, Vars in the targetlist might have gotten replaced, - * and we need to see the replacement expressions. + * Switch our attention to the subquery as mangled by the planner. It + * was okay to look at the pre-planning version for the tests above, + * but now we need a Var that will refer to the subroot's live + * RelOptInfos. For instance, if any subquery pullup happened during + * planning, Vars in the targetlist might have gotten replaced, and we + * need to see the replacement expressions. */ subquery = rel->subroot->parse; Assert(IsA(subquery, Query)); @@ -4530,13 +4531,13 @@ examine_simple_variable(PlannerInfo *root, Var *var, /* * If the sub-query originated from a view with the security_barrier - * attribute, we must not look at the variable's statistics, though - * it seems all right to notice the existence of a DISTINCT clause. - * So stop here. + * attribute, we must not look at the variable's statistics, though it + * seems all right to notice the existence of a DISTINCT clause. So + * stop here. * * This is probably a harsher restriction than necessary; it's * certainly OK for the selectivity estimator (which is a C function, - * and therefore omnipotent anyway) to look at the statistics. But + * and therefore omnipotent anyway) to look at the statistics. But * many selectivity estimators will happily *invoke the operator * function* to try to work out a good estimate - and that's not OK. * So for now, don't dig down for stats. @@ -4563,7 +4564,7 @@ examine_simple_variable(PlannerInfo *root, Var *var, /* * Otherwise, the Var comes from a FUNCTION, VALUES, or CTE RTE. (We * won't see RTE_JOIN here because join alias Vars have already been - * flattened.) There's not much we can do with function outputs, but + * flattened.) There's not much we can do with function outputs, but * maybe someday try to be smarter about VALUES and/or CTEs. */ } @@ -4679,8 +4680,8 @@ get_variable_numdistinct(VariableStatData *vardata, bool *isdefault) /* * With no data, estimate ndistinct = ntuples if the table is small, else - * use default. We use DEFAULT_NUM_DISTINCT as the cutoff for "small" - * so that the behavior isn't discontinuous. + * use default. We use DEFAULT_NUM_DISTINCT as the cutoff for "small" so + * that the behavior isn't discontinuous. */ if (ntuples < DEFAULT_NUM_DISTINCT) return ntuples; @@ -6094,16 +6095,16 @@ string_to_bytea_const(const char *str, size_t str_len) * ANDing the index predicate with the explicitly given indexquals produces * a more accurate idea of the index's selectivity. However, we need to be * careful not to insert redundant clauses, because clauselist_selectivity() - * is easily fooled into computing a too-low selectivity estimate. Our + * is easily fooled into computing a too-low selectivity estimate. Our * approach is to add only the predicate clause(s) that cannot be proven to - * be implied by the given indexquals. This successfully handles cases such + * be implied by the given indexquals. This successfully handles cases such * as a qual "x = 42" used with a partial index "WHERE x >= 40 AND x < 50". * There are many other cases where we won't detect redundancy, leading to a * too-low selectivity estimate, which will bias the system in favor of using - * partial indexes where possible. That is not necessarily bad though. + * partial indexes where possible. That is not necessarily bad though. * * Note that indexQuals contains RestrictInfo nodes while the indpred - * does not, so the output list will be mixed. This is OK for both + * does not, so the output list will be mixed. This is OK for both * predicate_implied_by() and clauselist_selectivity(), but might be * problematic if the result were passed to other things. */ @@ -6392,7 +6393,7 @@ btcostestimate(PG_FUNCTION_ARGS) * the index scan). Additional quals can suppress visits to the heap, so * it's OK to count them in indexSelectivity, but they should not count * for estimating numIndexTuples. So we must examine the given indexquals - * to find out which ones count as boundary quals. We rely on the + * to find out which ones count as boundary quals. We rely on the * knowledge that they are given in index column order. * * For a RowCompareExpr, we consider only the first column, just as @@ -6531,8 +6532,8 @@ btcostestimate(PG_FUNCTION_ARGS) /* * If the index is partial, AND the index predicate with the - * index-bound quals to produce a more accurate idea of the number - * of rows covered by the bound conditions. + * index-bound quals to produce a more accurate idea of the number of + * rows covered by the bound conditions. */ selectivityQuals = add_predicate_to_quals(index, indexBoundQuals); @@ -6767,17 +6768,17 @@ gincost_pattern(IndexOptInfo *index, int indexcol, int32 i; /* - * Get the operator's strategy number and declared input data types - * within the index opfamily. (We don't need the latter, but we use - * get_op_opfamily_properties because it will throw error if it fails - * to find a matching pg_amop entry.) + * Get the operator's strategy number and declared input data types within + * the index opfamily. (We don't need the latter, but we use + * get_op_opfamily_properties because it will throw error if it fails to + * find a matching pg_amop entry.) */ get_op_opfamily_properties(clause_op, index->opfamily[indexcol], false, &strategy_op, &lefttype, &righttype); /* - * GIN always uses the "default" support functions, which are those - * with lefttype == righttype == the opclass' opcintype (see + * GIN always uses the "default" support functions, which are those with + * lefttype == righttype == the opclass' opcintype (see * IndexSupportInitialize in relcache.c). */ extractProcOid = get_opfamily_proc(index->opfamily[indexcol], @@ -6864,7 +6865,7 @@ gincost_opexpr(IndexOptInfo *index, OpExpr *clause, GinQualCounts *counts) else { elog(ERROR, "could not match index to operand"); - operand = NULL; /* keep compiler quiet */ + operand = NULL; /* keep compiler quiet */ } if (IsA(operand, RelabelType)) @@ -6872,8 +6873,8 @@ gincost_opexpr(IndexOptInfo *index, OpExpr *clause, GinQualCounts *counts) /* * It's impossible to call extractQuery method for unknown operand. So - * unless operand is a Const we can't do much; just assume there will - * be one ordinary search entry from the operand at runtime. + * unless operand is a Const we can't do much; just assume there will be + * one ordinary search entry from the operand at runtime. */ if (!IsA(operand, Const)) { @@ -6901,7 +6902,7 @@ gincost_opexpr(IndexOptInfo *index, OpExpr *clause, GinQualCounts *counts) * each of which involves one value from the RHS array, plus all the * non-array quals (if any). To model this, we average the counts across * the RHS elements, and add the averages to the counts in *counts (which - * correspond to per-indexscan costs). We also multiply counts->arrayScans + * correspond to per-indexscan costs). We also multiply counts->arrayScans * by N, causing gincostestimate to scale up its estimates accordingly. */ static bool @@ -6935,9 +6936,9 @@ gincost_scalararrayopexpr(IndexOptInfo *index, ScalarArrayOpExpr *clause, /* * It's impossible to call extractQuery method for unknown operand. So - * unless operand is a Const we can't do much; just assume there will - * be one ordinary search entry from each array entry at runtime, and - * fall back on a probably-bad estimate of the number of array entries. + * unless operand is a Const we can't do much; just assume there will be + * one ordinary search entry from each array entry at runtime, and fall + * back on a probably-bad estimate of the number of array entries. */ if (!IsA(rightop, Const)) { @@ -7156,7 +7157,7 @@ gincostestimate(PG_FUNCTION_ARGS) else if (IsA(clause, ScalarArrayOpExpr)) { matchPossible = gincost_scalararrayopexpr(index, - (ScalarArrayOpExpr *) clause, + (ScalarArrayOpExpr *) clause, numEntries, &counts); if (!matchPossible) @@ -7194,7 +7195,8 @@ gincostestimate(PG_FUNCTION_ARGS) outer_scans = loop_count; /* - * Compute cost to begin scan, first of all, pay attention to pending list. + * Compute cost to begin scan, first of all, pay attention to pending + * list. */ entryPagesFetched = numPendingPages; @@ -7247,7 +7249,8 @@ gincostestimate(PG_FUNCTION_ARGS) *indexStartupCost = (entryPagesFetched + dataPagesFetched) * spc_random_page_cost; /* - * Now we compute the number of data pages fetched while the scan proceeds. + * Now we compute the number of data pages fetched while the scan + * proceeds. */ /* data pages scanned for each exact (non-partial) matched entry */ diff --git a/src/backend/utils/adt/timestamp.c b/src/backend/utils/adt/timestamp.c index a3e1e94a2b22a253b1372f02230d783834ad9bdb..8593b6b47f4d26710f36c8dc625e9179defb38cf 100644 --- a/src/backend/utils/adt/timestamp.c +++ b/src/backend/utils/adt/timestamp.c @@ -757,7 +757,7 @@ interval_send(PG_FUNCTION_ARGS) /* * The interval typmod stores a "range" in its high 16 bits and a "precision" - * in its low 16 bits. Both contribute to defining the resolution of the + * in its low 16 bits. Both contribute to defining the resolution of the * type. Range addresses resolution granules larger than one second, and * precision specifies resolution below one second. This representation can * express all SQL standard resolutions, but we implement them all in terms of @@ -940,7 +940,7 @@ interval_transform(PG_FUNCTION_ARGS) typmod = (Node *) lsecond(expr->args); - if (IsA(typmod, Const) && !((Const *) typmod)->constisnull) + if (IsA(typmod, Const) &&!((Const *) typmod)->constisnull) { Node *source = (Node *) linitial(expr->args); int32 old_typmod = exprTypmod(source); @@ -965,7 +965,7 @@ interval_transform(PG_FUNCTION_ARGS) /* * Temporally-smaller fields occupy higher positions in the range - * bitmap. Since only the temporally-smallest bit matters for length + * bitmap. Since only the temporally-smallest bit matters for length * coercion purposes, we compare the last-set bits in the ranges. * Precision, which is to say, sub-second precision, only affects * ranges that include SECOND. @@ -974,8 +974,8 @@ interval_transform(PG_FUNCTION_ARGS) old_range_fls = fls(old_range); if (new_typmod < 0 || ((new_range_fls >= SECOND || new_range_fls >= old_range_fls) && - (old_range_fls < SECOND || new_precis >= MAX_INTERVAL_PRECISION || - new_precis >= old_precis))) + (old_range_fls < SECOND || new_precis >= MAX_INTERVAL_PRECISION || + new_precis >= old_precis))) ret = relabel_to_typmod(source, new_typmod); } @@ -1925,7 +1925,7 @@ timestamp_fastcmp(Datum x, Datum y, SortSupport ssup) Datum timestamp_sortsupport(PG_FUNCTION_ARGS) { - SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); + SortSupport ssup = (SortSupport) PG_GETARG_POINTER(0); ssup->comparator = timestamp_fastcmp; PG_RETURN_VOID(); @@ -4067,11 +4067,11 @@ timestamp_part(PG_FUNCTION_ARGS) { case DTK_EPOCH: #ifdef HAVE_INT64_TIMESTAMP - result = (timestamp - SetEpochTimestamp()) / 1000000.0; + result = (timestamp - SetEpochTimestamp()) / 1000000.0; #else - result = timestamp - SetEpochTimestamp(); + result = timestamp - SetEpochTimestamp(); #endif - break; + break; case DTK_DOW: case DTK_ISODOW: diff --git a/src/backend/utils/adt/tsgistidx.c b/src/backend/utils/adt/tsgistidx.c index b427586d185e9d0f5f2903fdef22a7e39cd85a3e..674e48c871dc8f5b4f262374024bfdb9600c6d74 100644 --- a/src/backend/utils/adt/tsgistidx.c +++ b/src/backend/utils/adt/tsgistidx.c @@ -593,8 +593,8 @@ typedef struct static int comparecost(const void *va, const void *vb) { - const SPLITCOST *a = (const SPLITCOST *) va; - const SPLITCOST *b = (const SPLITCOST *) vb; + const SPLITCOST *a = (const SPLITCOST *) va; + const SPLITCOST *b = (const SPLITCOST *) vb; if (a->cost == b->cost) return 0; diff --git a/src/backend/utils/adt/tsquery_util.c b/src/backend/utils/adt/tsquery_util.c index ae00f180b5dfc3751b19931174ccb858773c8b45..0724d33c1d0d4aa55edc06c7c74f7d29a352001a 100644 --- a/src/backend/utils/adt/tsquery_util.c +++ b/src/backend/utils/adt/tsquery_util.c @@ -134,7 +134,7 @@ QTNodeCompare(QTNode *an, QTNode *bn) static int cmpQTN(const void *a, const void *b) { - return QTNodeCompare(*(QTNode * const *) a, *(QTNode * const *) b); + return QTNodeCompare(*(QTNode *const *) a, *(QTNode *const *) b); } void diff --git a/src/backend/utils/adt/tsrank.c b/src/backend/utils/adt/tsrank.c index 960233b63323cdc7703bad4bbc2df1372bac9ac0..a45caf05af57ac655779470a51559df99bc54870 100644 --- a/src/backend/utils/adt/tsrank.c +++ b/src/backend/utils/adt/tsrank.c @@ -134,8 +134,8 @@ static int compareQueryOperand(const void *a, const void *b, void *arg) { char *operand = (char *) arg; - QueryOperand *qa = (*(QueryOperand * const *) a); - QueryOperand *qb = (*(QueryOperand * const *) b); + QueryOperand *qa = (*(QueryOperand *const *) a); + QueryOperand *qb = (*(QueryOperand *const *) b); return tsCompareString(operand + qa->distance, qa->length, operand + qb->distance, qb->length, diff --git a/src/backend/utils/adt/tsvector_op.c b/src/backend/utils/adt/tsvector_op.c index bab6534feaaabf20eae357b8e8849d9edbd2a292..eb5c45b3d81480e7ea56ca7eded8811e7018c477 100644 --- a/src/backend/utils/adt/tsvector_op.c +++ b/src/backend/utils/adt/tsvector_op.c @@ -373,9 +373,9 @@ tsvector_concat(PG_FUNCTION_ARGS) i2 = in2->size; /* - * Conservative estimate of space needed. We might need all the data - * in both inputs, and conceivably add a pad byte before position data - * for each item where there was none before. + * Conservative estimate of space needed. We might need all the data in + * both inputs, and conceivably add a pad byte before position data for + * each item where there was none before. */ output_bytes = VARSIZE(in1) + VARSIZE(in2) + i1 + i2; diff --git a/src/backend/utils/adt/varbit.c b/src/backend/utils/adt/varbit.c index e74e062338d3bb84044cbb29c4d0a6dc4818141c..2bcf5b8aa8c2a184484afa2826af3d6332bbe766 100644 --- a/src/backend/utils/adt/varbit.c +++ b/src/backend/utils/adt/varbit.c @@ -664,7 +664,7 @@ varbit_transform(PG_FUNCTION_ARGS) typmod = (Node *) lsecond(expr->args); - if (IsA(typmod, Const) && !((Const *) typmod)->constisnull) + if (IsA(typmod, Const) &&!((Const *) typmod)->constisnull) { Node *source = (Node *) linitial(expr->args); int32 new_typmod = DatumGetInt32(((Const *) typmod)->constvalue); diff --git a/src/backend/utils/adt/varchar.c b/src/backend/utils/adt/varchar.c index 199330cef295db88fe50131fd154b9623fe37260..4cdb88837ba038f6130d1d152860f8ed6ef660a4 100644 --- a/src/backend/utils/adt/varchar.c +++ b/src/backend/utils/adt/varchar.c @@ -561,7 +561,7 @@ varchar_transform(PG_FUNCTION_ARGS) typmod = (Node *) lsecond(expr->args); - if (IsA(typmod, Const) && !((Const *) typmod)->constisnull) + if (IsA(typmod, Const) &&!((Const *) typmod)->constisnull) { Node *source = (Node *) linitial(expr->args); int32 old_typmod = exprTypmod(source); diff --git a/src/backend/utils/adt/varlena.c b/src/backend/utils/adt/varlena.c index 53989d1ecb36905485108a01c39caa69563240bd..e1b57ba3fc71fa2cb3898503fa380ff4bc21861b 100644 --- a/src/backend/utils/adt/varlena.c +++ b/src/backend/utils/adt/varlena.c @@ -1353,6 +1353,7 @@ varstr_cmp(char *arg1, int len1, char *arg2, int len2, Oid collid) char a2buf[STACKBUFLEN]; char *a1p, *a2p; + #ifdef HAVE_LOCALE_T pg_locale_t mylocale = 0; #endif @@ -1413,8 +1414,8 @@ varstr_cmp(char *arg1, int len1, char *arg2, int len2, Oid collid) (LPWSTR) a1p, a1len / 2); if (!r) ereport(ERROR, - (errmsg("could not convert string to UTF-16: error code %lu", - GetLastError()))); + (errmsg("could not convert string to UTF-16: error code %lu", + GetLastError()))); } ((LPWSTR) a1p)[r] = 0; @@ -1426,8 +1427,8 @@ varstr_cmp(char *arg1, int len1, char *arg2, int len2, Oid collid) (LPWSTR) a2p, a2len / 2); if (!r) ereport(ERROR, - (errmsg("could not convert string to UTF-16: error code %lu", - GetLastError()))); + (errmsg("could not convert string to UTF-16: error code %lu", + GetLastError()))); } ((LPWSTR) a2p)[r] = 0; @@ -4001,7 +4002,7 @@ text_format_string_conversion(StringInfo buf, char conversion, else if (conversion == 'I') ereport(ERROR, (errcode(ERRCODE_NULL_VALUE_NOT_ALLOWED), - errmsg("null values cannot be formatted as an SQL identifier"))); + errmsg("null values cannot be formatted as an SQL identifier"))); return; } diff --git a/src/backend/utils/adt/xml.c b/src/backend/utils/adt/xml.c index c51a9f76cede07354fb67c3bc320deb49c7dba84..44d327d7606df620c01b6e6532e53a9ee7a21f23 100644 --- a/src/backend/utils/adt/xml.c +++ b/src/backend/utils/adt/xml.c @@ -126,8 +126,8 @@ static bool print_xml_decl(StringInfo buf, const xmlChar *version, static xmlDocPtr xml_parse(text *data, XmlOptionType xmloption_arg, bool preserve_whitespace, int encoding); static text *xml_xmlnodetoxmltype(xmlNodePtr cur); -static int xml_xpathobjtoxmlarray(xmlXPathObjectPtr xpathobj, - ArrayBuildState **astate); +static int xml_xpathobjtoxmlarray(xmlXPathObjectPtr xpathobj, + ArrayBuildState **astate); #endif /* USE_LIBXML */ static StringInfo query_to_xml_internal(const char *query, char *tablename, @@ -913,7 +913,7 @@ pg_xml_init_library(void) * pg_xml_init --- set up for use of libxml and register an error handler * * This should be called by each function that is about to use libxml - * facilities and requires error handling. It initializes libxml with + * facilities and requires error handling. It initializes libxml with * pg_xml_init_library() and establishes our libxml error handler. * * strictness determines which errors are reported and which are ignored. @@ -943,9 +943,9 @@ pg_xml_init(PgXmlStrictness strictness) /* * Save original error handler and install ours. libxml originally didn't * distinguish between the contexts for generic and for structured error - * handlers. If we're using an old libxml version, we must thus save - * the generic error context, even though we're using a structured - * error handler. + * handlers. If we're using an old libxml version, we must thus save the + * generic error context, even though we're using a structured error + * handler. */ errcxt->saved_errfunc = xmlStructuredError; @@ -959,7 +959,7 @@ pg_xml_init(PgXmlStrictness strictness) /* * Verify that xmlSetStructuredErrorFunc set the context variable we - * expected it to. If not, the error context pointer we just saved is not + * expected it to. If not, the error context pointer we just saved is not * the correct thing to restore, and since that leaves us without a way to * restore the context in pg_xml_done, we must fail. * @@ -1014,9 +1014,9 @@ pg_xml_done(PgXmlErrorContext *errcxt, bool isError) Assert(!errcxt->err_occurred || isError); /* - * Check that libxml's global state is correct, warn if not. This is - * a real test and not an Assert because it has a higher probability - * of happening. + * Check that libxml's global state is correct, warn if not. This is a + * real test and not an Assert because it has a higher probability of + * happening. */ #ifdef HAVE_XMLSTRUCTUREDERRORCONTEXT cur_errcxt = xmlStructuredErrorContext; @@ -1108,7 +1108,7 @@ parse_xml_decl(const xmlChar *str, size_t *lenp, int utf8len; /* - * Only initialize libxml. We don't need error handling here, but we do + * Only initialize libxml. We don't need error handling here, but we do * need to make sure libxml is initialized before calling any of its * functions. Note that this is safe (and a no-op) if caller has already * done pg_xml_init(). @@ -1516,9 +1516,9 @@ xml_errorHandler(void *data, xmlErrorPtr error) PgXmlErrorContext *xmlerrcxt = (PgXmlErrorContext *) data; xmlParserCtxtPtr ctxt = (xmlParserCtxtPtr) error->ctxt; xmlParserInputPtr input = (ctxt != NULL) ? ctxt->input : NULL; - xmlNodePtr node = error->node; + xmlNodePtr node = error->node; const xmlChar *name = (node != NULL && - node->type == XML_ELEMENT_NODE) ? node->name : NULL; + node->type == XML_ELEMENT_NODE) ? node->name : NULL; int domain = error->domain; int level = error->level; StringInfo errorBuf; @@ -1599,7 +1599,7 @@ xml_errorHandler(void *data, xmlErrorPtr error) if (input != NULL) { xmlGenericErrorFunc errFuncSaved = xmlGenericError; - void *errCtxSaved = xmlGenericErrorContext; + void *errCtxSaved = xmlGenericErrorContext; xmlSetGenericErrorFunc((void *) errorBuf, (xmlGenericErrorFunc) appendStringInfo); @@ -1617,8 +1617,8 @@ xml_errorHandler(void *data, xmlErrorPtr error) chopStringInfoNewlines(errorBuf); /* - * Legacy error handling mode. err_occurred is never set, we just add the - * message to err_buf. This mode exists because the xml2 contrib module + * Legacy error handling mode. err_occurred is never set, we just add the + * message to err_buf. This mode exists because the xml2 contrib module * uses our error-handling infrastructure, but we don't want to change its * behaviour since it's deprecated anyway. This is also why we don't * distinguish between notices, warnings and errors here --- the old-style @@ -3574,7 +3574,7 @@ xml_xmlnodetoxmltype(xmlNodePtr cur) PG_TRY(); { /* Here we rely on XML having the same representation as TEXT */ - char *escaped = escape_xml((char *) str); + char *escaped = escape_xml((char *) str); result = (xmltype *) cstring_to_text(escaped); pfree(escaped); @@ -3623,7 +3623,7 @@ xml_xpathobjtoxmlarray(xmlXPathObjectPtr xpathobj, result = xpathobj->nodesetval->nodeNr; if (astate != NULL) { - int i; + int i; for (i = 0; i < result; i++) { diff --git a/src/backend/utils/cache/catcache.c b/src/backend/utils/cache/catcache.c index ea3daa599ca000ca4cbe491a59ffb377ebdc56cb..0307b9652d4682d140c1ad12ccbce5996271eadc 100644 --- a/src/backend/utils/cache/catcache.c +++ b/src/backend/utils/cache/catcache.c @@ -1637,8 +1637,8 @@ CatalogCacheCreateEntry(CatCache *cache, HeapTuple ntp, /* * If there are any out-of-line toasted fields in the tuple, expand them - * in-line. This saves cycles during later use of the catcache entry, - * and also protects us against the possibility of the toast tuples being + * in-line. This saves cycles during later use of the catcache entry, and + * also protects us against the possibility of the toast tuples being * freed before we attempt to fetch them, in case of something using a * slightly stale catcache entry. */ diff --git a/src/backend/utils/cache/inval.c b/src/backend/utils/cache/inval.c index d5fe85abbff762b431ab20c2316eb17b1f3f875f..9ccfc4f1144ac90365984476a728a164d7807bb1 100644 --- a/src/backend/utils/cache/inval.c +++ b/src/backend/utils/cache/inval.c @@ -820,7 +820,7 @@ ProcessCommittedInvalidationMessages(SharedInvalidationMessage *msgs, * since they'll not have seen our changed tuples anyway. We can forget * about CurrentCmdInvalidMsgs too, since those changes haven't touched * the caches yet. - * + * * In any case, reset the various lists to empty. We need not physically * free memory here, since TopTransactionContext is about to be emptied * anyway. diff --git a/src/backend/utils/cache/lsyscache.c b/src/backend/utils/cache/lsyscache.c index 44dab822648f01ebce59d2c7f34cd88159f0b7ae..64b413bb6aed32073999d1e631d5d5d56b1b9cc8 100644 --- a/src/backend/utils/cache/lsyscache.c +++ b/src/backend/utils/cache/lsyscache.c @@ -283,7 +283,7 @@ get_sort_function_for_ordering_op(Oid opno, Oid *sortfunc, opcintype, opcintype, BTORDER_PROC); - if (!OidIsValid(*sortfunc)) /* should not happen */ + if (!OidIsValid(*sortfunc)) /* should not happen */ elog(ERROR, "missing support function %d(%u,%u) in opfamily %u", BTORDER_PROC, opcintype, opcintype, opfamily); *issupport = false; @@ -1549,7 +1549,7 @@ func_volatile(Oid funcid) /* * get_func_leakproof - * Given procedure id, return the function's leakproof field. + * Given procedure id, return the function's leakproof field. */ bool get_func_leakproof(Oid funcid) @@ -2914,8 +2914,8 @@ get_range_subtype(Oid rangeOid) tp = SearchSysCache1(RANGETYPE, ObjectIdGetDatum(rangeOid)); if (HeapTupleIsValid(tp)) { - Form_pg_range rngtup = (Form_pg_range) GETSTRUCT(tp); - Oid result; + Form_pg_range rngtup = (Form_pg_range) GETSTRUCT(tp); + Oid result; result = rngtup->rngsubtype; ReleaseSysCache(tp); diff --git a/src/backend/utils/cache/plancache.c b/src/backend/utils/cache/plancache.c index 6292f8dc6c99a55e81b4bdd781acdbd1ff96e3cb..c42765c25a782a095cf3e81e45582c9cf353e558 100644 --- a/src/backend/utils/cache/plancache.c +++ b/src/backend/utils/cache/plancache.c @@ -11,7 +11,7 @@ * The logic for choosing generic or custom plans is in choose_custom_plan, * which see for comments. * - * Cache invalidation is driven off sinval events. Any CachedPlanSource + * Cache invalidation is driven off sinval events. Any CachedPlanSource * that matches the event is marked invalid, as is its generic CachedPlan * if it has one. When (and if) the next demand for a cached plan occurs, * parse analysis and rewrite is repeated to build a new valid query tree, @@ -77,9 +77,9 @@ static void ReleaseGenericPlan(CachedPlanSource *plansource); static List *RevalidateCachedQuery(CachedPlanSource *plansource); static bool CheckCachedPlan(CachedPlanSource *plansource); static CachedPlan *BuildCachedPlan(CachedPlanSource *plansource, List *qlist, - ParamListInfo boundParams); + ParamListInfo boundParams); static bool choose_custom_plan(CachedPlanSource *plansource, - ParamListInfo boundParams); + ParamListInfo boundParams); static double cached_plan_cost(CachedPlan *plan); static void AcquireExecutorLocks(List *stmt_list, bool acquire); static void AcquirePlannerLocks(List *stmt_list, bool acquire); @@ -111,7 +111,7 @@ InitPlanCache(void) * CreateCachedPlan: initially create a plan cache entry. * * Creation of a cached plan is divided into two steps, CreateCachedPlan and - * CompleteCachedPlan. CreateCachedPlan should be called after running the + * CompleteCachedPlan. CreateCachedPlan should be called after running the * query through raw_parser, but before doing parse analysis and rewrite; * CompleteCachedPlan is called after that. The reason for this arrangement * is that it can save one round of copying of the raw parse tree, since @@ -198,13 +198,13 @@ CreateCachedPlan(Node *raw_parse_tree, * CompleteCachedPlan: second step of creating a plan cache entry. * * Pass in the analyzed-and-rewritten form of the query, as well as the - * required subsidiary data about parameters and such. All passed values will + * required subsidiary data about parameters and such. All passed values will * be copied into the CachedPlanSource's memory, except as specified below. * After this is called, GetCachedPlan can be called to obtain a plan, and * optionally the CachedPlanSource can be saved using SaveCachedPlan. * * If querytree_context is not NULL, the querytree_list must be stored in that - * context (but the other parameters need not be). The querytree_list is not + * context (but the other parameters need not be). The querytree_list is not * copied, rather the given context is kept as the initial query_context of * the CachedPlanSource. (It should have been created as a child of the * caller's working memory context, but it will now be reparented to belong @@ -277,8 +277,8 @@ CompleteCachedPlan(CachedPlanSource *plansource, /* * Use the planner machinery to extract dependencies. Data is saved in - * query_context. (We assume that not a lot of extra cruft is created - * by this call.) + * query_context. (We assume that not a lot of extra cruft is created by + * this call.) */ extract_query_dependencies((Node *) querytree_list, &plansource->relationOids, @@ -327,7 +327,7 @@ CompleteCachedPlan(CachedPlanSource *plansource, * * This is guaranteed not to throw error; callers typically depend on that * since this is called just before or just after adding a pointer to the - * CachedPlanSource to some permanent data structure of their own. Up until + * CachedPlanSource to some permanent data structure of their own. Up until * this is done, a CachedPlanSource is just transient data that will go away * automatically on transaction abort. */ @@ -341,16 +341,16 @@ SaveCachedPlan(CachedPlanSource *plansource) /* * In typical use, this function would be called before generating any - * plans from the CachedPlanSource. If there is a generic plan, moving - * it into CacheMemoryContext would be pretty risky since it's unclear + * plans from the CachedPlanSource. If there is a generic plan, moving it + * into CacheMemoryContext would be pretty risky since it's unclear * whether the caller has taken suitable care with making references - * long-lived. Best thing to do seems to be to discard the plan. + * long-lived. Best thing to do seems to be to discard the plan. */ ReleaseGenericPlan(plansource); /* - * Reparent the source memory context under CacheMemoryContext so that - * it will live indefinitely. The query_context follows along since it's + * Reparent the source memory context under CacheMemoryContext so that it + * will live indefinitely. The query_context follows along since it's * already a child of the other one. */ MemoryContextSetParent(plansource->context, CacheMemoryContext); @@ -474,8 +474,8 @@ RevalidateCachedQuery(CachedPlanSource *plansource) } /* - * Discard the no-longer-useful query tree. (Note: we don't want to - * do this any earlier, else we'd not have been able to release locks + * Discard the no-longer-useful query tree. (Note: we don't want to do + * this any earlier, else we'd not have been able to release locks * correctly in the race condition case.) */ plansource->is_valid = false; @@ -484,14 +484,14 @@ RevalidateCachedQuery(CachedPlanSource *plansource) plansource->invalItems = NIL; /* - * Free the query_context. We don't really expect MemoryContextDelete to + * Free the query_context. We don't really expect MemoryContextDelete to * fail, but just in case, make sure the CachedPlanSource is left in a - * reasonably sane state. (The generic plan won't get unlinked yet, - * but that's acceptable.) + * reasonably sane state. (The generic plan won't get unlinked yet, but + * that's acceptable.) */ if (plansource->query_context) { - MemoryContext qcxt = plansource->query_context; + MemoryContext qcxt = plansource->query_context; plansource->query_context = NULL; MemoryContextDelete(qcxt); @@ -553,7 +553,7 @@ RevalidateCachedQuery(CachedPlanSource *plansource) PopOverrideSearchPath(); /* - * Check or update the result tupdesc. XXX should we use a weaker + * Check or update the result tupdesc. XXX should we use a weaker * condition than equalTupleDescs() here? * * We assume the parameter types didn't change from the first time, so no @@ -596,8 +596,8 @@ RevalidateCachedQuery(CachedPlanSource *plansource) /* * Use the planner machinery to extract dependencies. Data is saved in - * query_context. (We assume that not a lot of extra cruft is created - * by this call.) + * query_context. (We assume that not a lot of extra cruft is created by + * this call.) */ extract_query_dependencies((Node *) qlist, &plansource->relationOids, @@ -612,12 +612,12 @@ RevalidateCachedQuery(CachedPlanSource *plansource) plansource->query_list = qlist; /* - * Note: we do not reset generic_cost or total_custom_cost, although - * we could choose to do so. If the DDL or statistics change that - * prompted the invalidation meant a significant change in the cost - * estimates, it would be better to reset those variables and start - * fresh; but often it doesn't, and we're better retaining our hard-won - * knowledge about the relative costs. + * Note: we do not reset generic_cost or total_custom_cost, although we + * could choose to do so. If the DDL or statistics change that prompted + * the invalidation meant a significant change in the cost estimates, it + * would be better to reset those variables and start fresh; but often it + * doesn't, and we're better retaining our hard-won knowledge about the + * relative costs. */ plansource->is_valid = true; @@ -728,7 +728,7 @@ BuildCachedPlan(CachedPlanSource *plansource, List *qlist, * we ought to be holding sufficient locks to prevent any invalidation. * However, if we're building a custom plan after having built and * rejected a generic plan, it's possible to reach here with is_valid - * false due to an invalidation while making the generic plan. In theory + * false due to an invalidation while making the generic plan. In theory * the invalidation must be a false positive, perhaps a consequence of an * sinval reset event or the CLOBBER_CACHE_ALWAYS debug code. But for * safety, let's treat it as real and redo the RevalidateCachedQuery call. @@ -770,10 +770,10 @@ BuildCachedPlan(CachedPlanSource *plansource, List *qlist, } /* - * The planner may try to call SPI-using functions, which causes a - * problem if we're already inside one. Rather than expect all - * SPI-using code to do SPI_push whenever a replan could happen, - * it seems best to take care of the case here. + * The planner may try to call SPI-using functions, which causes a problem + * if we're already inside one. Rather than expect all SPI-using code to + * do SPI_push whenever a replan could happen, it seems best to take care + * of the case here. */ spi_pushed = SPI_push_conditional(); @@ -865,8 +865,8 @@ choose_custom_plan(CachedPlanSource *plansource, ParamListInfo boundParams) /* * Prefer generic plan if it's less than 10% more expensive than average * custom plan. This threshold is a bit arbitrary; it'd be better if we - * had some means of comparing planning time to the estimated runtime - * cost differential. + * had some means of comparing planning time to the estimated runtime cost + * differential. * * Note that if generic_cost is -1 (indicating we've not yet determined * the generic plan cost), we'll always prefer generic at this point. @@ -966,7 +966,7 @@ GetCachedPlan(CachedPlanSource *plansource, ParamListInfo boundParams, { /* otherwise, it should be a sibling of the plansource */ MemoryContextSetParent(plan->context, - MemoryContextGetParent(plansource->context)); + MemoryContextGetParent(plansource->context)); } /* Update generic_cost whenever we make a new generic plan */ plansource->generic_cost = cached_plan_cost(plan); @@ -984,7 +984,7 @@ GetCachedPlan(CachedPlanSource *plansource, ParamListInfo boundParams, /* * If we choose to plan again, we need to re-copy the query_list, - * since the planner probably scribbled on it. We can force + * since the planner probably scribbled on it. We can force * BuildCachedPlan to do that by passing NIL. */ qlist = NIL; @@ -1089,7 +1089,7 @@ CachedPlanSetParentContext(CachedPlanSource *plansource, * * This is a convenience routine that does the equivalent of * CreateCachedPlan + CompleteCachedPlan, using the data stored in the - * input CachedPlanSource. The result is therefore "unsaved" (regardless + * input CachedPlanSource. The result is therefore "unsaved" (regardless * of the state of the source), and we don't copy any generic plan either. * The result will be currently valid, or not, the same as the source. */ @@ -1233,7 +1233,7 @@ AcquireExecutorLocks(List *stmt_list, bool acquire) { /* * Ignore utility statements, except those (such as EXPLAIN) that - * contain a parsed-but-not-planned query. Note: it's okay to use + * contain a parsed-but-not-planned query. Note: it's okay to use * ScanQueryForLocks, even though the query hasn't been through * rule rewriting, because rewriting doesn't change the query * representation. @@ -1429,7 +1429,7 @@ plan_list_is_transient(List *stmt_list) /* * PlanCacheComputeResultDesc: given a list of analyzed-and-rewritten Queries, - * determine the result tupledesc it will produce. Returns NULL if the + * determine the result tupledesc it will produce. Returns NULL if the * execution will not return tuples. * * Note: the result is created or copied into current memory context. diff --git a/src/backend/utils/cache/relcache.c b/src/backend/utils/cache/relcache.c index 7f0e20ec1741acd1715e1af33180e5ca5d2d28fe..4cbf812ed594b4ca291d7662fa2bd47f34f131cc 100644 --- a/src/backend/utils/cache/relcache.c +++ b/src/backend/utils/cache/relcache.c @@ -2540,7 +2540,7 @@ RelationBuildLocalRelation(const char *relname, /* * Insert relation physical and logical identifiers (OIDs) into the right - * places. For a mapped relation, we set relfilenode to zero and rely on + * places. For a mapped relation, we set relfilenode to zero and rely on * RelationInitPhysicalAddr to consult the map. */ rel->rd_rel->relisshared = shared_relation; @@ -3365,9 +3365,9 @@ RelationGetIndexList(Relation relation) result = insert_ordered_oid(result, index->indexrelid); /* - * indclass cannot be referenced directly through the C struct, because - * it comes after the variable-width indkey field. Must extract the - * datum the hard way... + * indclass cannot be referenced directly through the C struct, + * because it comes after the variable-width indkey field. Must + * extract the datum the hard way... */ indclassDatum = heap_getattr(htup, Anum_pg_index_indclass, @@ -4514,8 +4514,8 @@ RelationCacheInitFilePreInvalidate(void) /* * The file might not be there if no backend has been started since * the last removal. But complain about failures other than ENOENT. - * Fortunately, it's not too late to abort the transaction if we - * can't get rid of the would-be-obsolete init file. + * Fortunately, it's not too late to abort the transaction if we can't + * get rid of the would-be-obsolete init file. */ if (errno != ENOENT) ereport(ERROR, diff --git a/src/backend/utils/cache/ts_cache.c b/src/backend/utils/cache/ts_cache.c index 4ad5e7fc0b0642398adaaea27f5b8466798d7ef1..b408de0730ebc791c18b6d16f6a42832a5d4f283 100644 --- a/src/backend/utils/cache/ts_cache.c +++ b/src/backend/utils/cache/ts_cache.c @@ -602,10 +602,10 @@ check_TSCurrentConfig(char **newval, void **extra, GucSource source) cfgId = get_ts_config_oid(stringToQualifiedNameList(*newval), true); /* - * When source == PGC_S_TEST, we are checking the argument of an - * ALTER DATABASE SET or ALTER USER SET command. It could be that - * the intended use of the setting is for some other database, so - * we should not error out if the text search configuration is not + * When source == PGC_S_TEST, we are checking the argument of an ALTER + * DATABASE SET or ALTER USER SET command. It could be that the + * intended use of the setting is for some other database, so we + * should not error out if the text search configuration is not * present in the current database. We issue a NOTICE instead. */ if (!OidIsValid(cfgId)) diff --git a/src/backend/utils/error/elog.c b/src/backend/utils/error/elog.c index 65c28a750807aec39cc6adcc9ad506f16fb31f33..a40b343ebcfb478515659bc52d84ecccd8ca97a5 100644 --- a/src/backend/utils/error/elog.c +++ b/src/backend/utils/error/elog.c @@ -114,7 +114,7 @@ int Log_destination = LOG_DESTINATION_STDERR; /* * Max string length to send to syslog(). Note that this doesn't count the * sequence-number prefix we add, and of course it doesn't count the prefix - * added by syslog itself. Solaris and sysklogd truncate the final message + * added by syslog itself. Solaris and sysklogd truncate the final message * at 1024 bytes, so this value leaves 124 bytes for those prefixes. (Most * other syslog implementations seem to have limits of 2KB or so.) */ @@ -1857,8 +1857,8 @@ setup_formatted_log_time(void) stamp_time = (pg_time_t) tv.tv_sec; /* - * Note: we expect that guc.c will ensure that log_timezone is set up - * (at least with a minimal GMT value) before Log_line_prefix can become + * Note: we expect that guc.c will ensure that log_timezone is set up (at + * least with a minimal GMT value) before Log_line_prefix can become * nonempty or CSV mode can be selected. */ pg_strftime(formatted_log_time, FORMATTED_TS_LEN, @@ -1880,8 +1880,8 @@ setup_formatted_start_time(void) pg_time_t stamp_time = (pg_time_t) MyStartTime; /* - * Note: we expect that guc.c will ensure that log_timezone is set up - * (at least with a minimal GMT value) before Log_line_prefix can become + * Note: we expect that guc.c will ensure that log_timezone is set up (at + * least with a minimal GMT value) before Log_line_prefix can become * nonempty or CSV mode can be selected. */ pg_strftime(formatted_start_time, FORMATTED_TS_LEN, @@ -2506,7 +2506,7 @@ send_message_to_server_log(ErrorData *edata) * * Note: when there are multiple backends writing into the syslogger pipe, * it's critical that each write go into the pipe indivisibly, and not - * get interleaved with data from other processes. Fortunately, the POSIX + * get interleaved with data from other processes. Fortunately, the POSIX * spec requires that writes to pipes be atomic so long as they are not * more than PIPE_BUF bytes long. So we divide long messages into chunks * that are no more than that length, and send one chunk per write() call. diff --git a/src/backend/utils/fmgr/fmgr.c b/src/backend/utils/fmgr/fmgr.c index 788f1801a80761379476a958d75f705e89b08950..2ec63fae5682a3cae44d7d2babf27eb7d6d453ec 100644 --- a/src/backend/utils/fmgr/fmgr.c +++ b/src/backend/utils/fmgr/fmgr.c @@ -408,8 +408,8 @@ fmgr_info_other_lang(Oid functionId, FmgrInfo *finfo, HeapTuple procedureTuple) /* * Look up the language's call handler function, ignoring any attributes - * that would normally cause insertion of fmgr_security_definer. We - * need to get back a bare pointer to the actual C-language function. + * that would normally cause insertion of fmgr_security_definer. We need + * to get back a bare pointer to the actual C-language function. */ fmgr_info_cxt_security(languageStruct->lanplcallfoid, &plfinfo, CurrentMemoryContext, true); diff --git a/src/backend/utils/fmgr/funcapi.c b/src/backend/utils/fmgr/funcapi.c index dd914789c07451835c09000c2ba67bcaff55c341..addf95bca9b1b03ffede8318c4f8e49799c4d685 100644 --- a/src/backend/utils/fmgr/funcapi.c +++ b/src/backend/utils/fmgr/funcapi.c @@ -490,9 +490,9 @@ resolve_polymorphic_tupdesc(TupleDesc tupdesc, oidvector *declared_args, ANYARRAYOID); if (OidIsValid(anyrange_type)) { - Oid subtype = resolve_generic_type(ANYELEMENTOID, - anyrange_type, - ANYRANGEOID); + Oid subtype = resolve_generic_type(ANYELEMENTOID, + anyrange_type, + ANYRANGEOID); /* check for inconsistent array and range results */ if (OidIsValid(anyelement_type) && anyelement_type != subtype) @@ -524,8 +524,8 @@ resolve_polymorphic_tupdesc(TupleDesc tupdesc, oidvector *declared_args, /* * Identify the collation to use for polymorphic OUT parameters. (It'll * necessarily be the same for both anyelement and anyarray.) Note that - * range types are not collatable, so any possible internal collation of - * a range type is not considered here. + * range types are not collatable, so any possible internal collation of a + * range type is not considered here. */ if (OidIsValid(anyelement_type)) anycollation = get_typcollation(anyelement_type); @@ -687,9 +687,9 @@ resolve_polymorphic_argtypes(int numargs, Oid *argtypes, char *argmodes, ANYARRAYOID); if (OidIsValid(anyrange_type)) { - Oid subtype = resolve_generic_type(ANYELEMENTOID, - anyrange_type, - ANYRANGEOID); + Oid subtype = resolve_generic_type(ANYELEMENTOID, + anyrange_type, + ANYRANGEOID); /* check for inconsistent array and range results */ if (OidIsValid(anyelement_type) && anyelement_type != subtype) diff --git a/src/backend/utils/init/miscinit.c b/src/backend/utils/init/miscinit.c index 0f734260c16d71136055344b3195c070d458d0e1..fb376a0d271730f2813bb1d1f2e15562e82ff865 100644 --- a/src/backend/utils/init/miscinit.c +++ b/src/backend/utils/init/miscinit.c @@ -631,7 +631,7 @@ GetUserNameFromId(Oid roleid) * ($DATADIR/postmaster.pid) and a Unix-socket-file lockfile ($SOCKFILE.lock). * Both kinds of files contain the same info initially, although we can add * more information to a data-directory lockfile after it's created, using - * AddToDataDirLockFile(). See miscadmin.h for documentation of the contents + * AddToDataDirLockFile(). See miscadmin.h for documentation of the contents * of these lockfiles. * * On successful lockfile creation, a proc_exit callback to remove the diff --git a/src/backend/utils/mb/wchar.c b/src/backend/utils/mb/wchar.c index 7de460e0dd3c5fbb912d4b43c8ec8cd4d771af55..03d68119d1ce2fc482d4c00d7cb4b5448a2b8ed4 100644 --- a/src/backend/utils/mb/wchar.c +++ b/src/backend/utils/mb/wchar.c @@ -1341,7 +1341,7 @@ pg_utf8_islegal(const unsigned char *source, int length) * * Not knowing anything about the properties of the encoding in use, we just * keep incrementing the last byte until we get a validly-encoded result, - * or we run out of values to try. We don't bother to try incrementing + * or we run out of values to try. We don't bother to try incrementing * higher-order bytes, so there's no growth in runtime for wider characters. * (If we did try to do that, we'd need to consider the likelihood that 255 * is not a valid final byte in the encoding.) @@ -1371,7 +1371,7 @@ pg_generic_charinc(unsigned char *charptr, int len) * For a one-byte character less than 0x7F, we just increment the byte. * * For a multibyte character, every byte but the first must fall between 0x80 - * and 0xBF; and the first byte must be between 0xC0 and 0xF4. We increment + * and 0xBF; and the first byte must be between 0xC0 and 0xF4. We increment * the last byte that's not already at its maximum value. If we can't find a * byte that's less than the maximum allowable value, we simply fail. We also * need some special-case logic to skip regions used for surrogate pair @@ -1530,7 +1530,7 @@ pg_eucjp_increment(unsigned char *charptr, int length) return false; } else - { /* ASCII, single byte */ + { /* ASCII, single byte */ if (c1 > 0x7e) return false; (*charptr)++; @@ -1540,8 +1540,7 @@ pg_eucjp_increment(unsigned char *charptr, int length) return true; } - -#endif /* !FRONTEND */ +#endif /* !FRONTEND */ /* @@ -1675,8 +1674,8 @@ mbcharacter_incrementer pg_database_encoding_character_incrementer(void) { /* - * Eventually it might be best to add a field to pg_wchar_table[], - * but for now we just use a switch. + * Eventually it might be best to add a field to pg_wchar_table[], but for + * now we just use a switch. */ switch (GetDatabaseEncoding()) { @@ -1878,10 +1877,10 @@ report_untranslatable_char(int src_encoding, int dest_encoding, ereport(ERROR, (errcode(ERRCODE_UNTRANSLATABLE_CHARACTER), - errmsg("character with byte sequence %s in encoding \"%s\" has no equivalent in encoding \"%s\"", - buf, - pg_enc2name_tbl[src_encoding].name, - pg_enc2name_tbl[dest_encoding].name))); + errmsg("character with byte sequence %s in encoding \"%s\" has no equivalent in encoding \"%s\"", + buf, + pg_enc2name_tbl[src_encoding].name, + pg_enc2name_tbl[dest_encoding].name))); } -#endif /* !FRONTEND */ +#endif /* !FRONTEND */ diff --git a/src/backend/utils/misc/guc.c b/src/backend/utils/misc/guc.c index d75ab430296897a48d6b8da18ffe947e5cb20528..b756e58a3690eb05ab12b0ca2e1c12c0b2d743d8 100644 --- a/src/backend/utils/misc/guc.c +++ b/src/backend/utils/misc/guc.c @@ -3333,7 +3333,7 @@ static void InitializeOneGUCOption(struct config_generic * gconf); static void push_old_value(struct config_generic * gconf, GucAction action); static void ReportGUCOption(struct config_generic * record); static void reapply_stacked_values(struct config_generic * variable, - struct config_string *pHolder, + struct config_string * pHolder, GucStack *stack, const char *curvalue, GucContext curscontext, GucSource cursource); @@ -4143,8 +4143,8 @@ SelectConfigFiles(const char *userDoption, const char *progname) /* * If timezone_abbreviations wasn't set in the configuration file, install - * the default value. We do it this way because we can't safely install - * a "real" value until my_exec_path is set, which may not have happened + * the default value. We do it this way because we can't safely install a + * "real" value until my_exec_path is set, which may not have happened * when InitializeGUCOptions runs, so the bootstrap default value cannot * be the real desired default. */ @@ -4415,7 +4415,7 @@ NewGUCNestLevel(void) /* * Do GUC processing at transaction or subtransaction commit or abort, or * when exiting a function that has proconfig settings, or when undoing a - * transient assignment to some GUC variables. (The name is thus a bit of + * transient assignment to some GUC variables. (The name is thus a bit of * a misnomer; perhaps it should be ExitGUCNestLevel or some such.) * During abort, we discard all GUC settings that were applied at nesting * levels >= nestLevel. nestLevel == 1 corresponds to the main transaction. @@ -5106,7 +5106,7 @@ config_enum_get_options(struct config_enum * record, const char *prefix, * * Return value: * +1: the value is valid and was successfully applied. - * 0: the name or value is invalid (but see below). + * 0: the name or value is invalid (but see below). * -1: the value was not applied because of context, priority, or changeVal. * * If there is an error (non-existing option, invalid value) then an @@ -6441,7 +6441,7 @@ define_custom_variable(struct config_generic * variable) * variable. Essentially, we need to duplicate all the active and stacked * values, but with appropriate validation and datatype adjustment. * - * If an assignment fails, we report a WARNING and keep going. We don't + * If an assignment fails, we report a WARNING and keep going. We don't * want to throw ERROR for bad values, because it'd bollix the add-on * module that's presumably halfway through getting loaded. In such cases * the default or previous state will become active instead. @@ -6469,7 +6469,7 @@ define_custom_variable(struct config_generic * variable) /* * Free up as much as we conveniently can of the placeholder structure. * (This neglects any stack items, so it's possible for some memory to be - * leaked. Since this can only happen once per session per variable, it + * leaked. Since this can only happen once per session per variable, it * doesn't seem worth spending much code on.) */ set_string_field(pHolder, pHolder->variable, NULL); @@ -6487,7 +6487,7 @@ define_custom_variable(struct config_generic * variable) */ static void reapply_stacked_values(struct config_generic * variable, - struct config_string *pHolder, + struct config_string * pHolder, GucStack *stack, const char *curvalue, GucContext curscontext, GucSource cursource) @@ -6526,7 +6526,7 @@ reapply_stacked_values(struct config_generic * variable, case GUC_SET_LOCAL: /* first, apply the masked value as SET */ (void) set_config_option(name, stack->masked.val.stringval, - stack->masked_scontext, PGC_S_SESSION, + stack->masked_scontext, PGC_S_SESSION, GUC_ACTION_SET, true, WARNING); /* then apply the current value as LOCAL */ (void) set_config_option(name, curvalue, @@ -6542,7 +6542,7 @@ reapply_stacked_values(struct config_generic * variable, else { /* - * We are at the end of the stack. If the active/previous value is + * We are at the end of the stack. If the active/previous value is * different from the reset value, it must represent a previously * committed session value. Apply it, and then drop the stack entry * that set_config_option will have created under the impression that @@ -8028,8 +8028,8 @@ validate_option_array_item(const char *name, const char *value, * * name is not known, but exists or can be created as a placeholder (i.e., * it has a prefixed name). We allow this case if you're a superuser, - * otherwise not. Superusers are assumed to know what they're doing. - * We can't allow it for other users, because when the placeholder is + * otherwise not. Superusers are assumed to know what they're doing. We + * can't allow it for other users, because when the placeholder is * resolved it might turn out to be a SUSET variable; * define_custom_variable assumes we checked that. * diff --git a/src/backend/utils/mmgr/portalmem.c b/src/backend/utils/mmgr/portalmem.c index cfb73c1b09057145752ea9b67126269898e047b5..5713bbe12ce585f756ceef6f03f671c144664fca 100644 --- a/src/backend/utils/mmgr/portalmem.c +++ b/src/backend/utils/mmgr/portalmem.c @@ -487,7 +487,7 @@ PortalDrop(Portal portal, bool isTopCommit) * during transaction abort. * * Note: in most paths of control, this will have been done already in - * MarkPortalDone or MarkPortalFailed. We're just making sure. + * MarkPortalDone or MarkPortalFailed. We're just making sure. */ if (PointerIsValid(portal->cleanup)) { diff --git a/src/backend/utils/sort/sortsupport.c b/src/backend/utils/sort/sortsupport.c index 7f388fd9bfce6660843d282cfefa4fc61a8b4e3f..b6d916d3e43d5068cd0c36892338ec72a77a15fe 100644 --- a/src/backend/utils/sort/sortsupport.c +++ b/src/backend/utils/sort/sortsupport.c @@ -24,7 +24,7 @@ typedef struct { FunctionCallInfoData fcinfo; /* reusable callinfo structure */ - FmgrInfo flinfo; /* lookup data for comparison function */ + FmgrInfo flinfo; /* lookup data for comparison function */ } SortShimExtra; @@ -70,7 +70,6 @@ ApplySortComparator(Datum datum1, bool isNull1, return compare; } - #endif /* ! USE_INLINE */ /* @@ -108,7 +107,7 @@ comparison_shim(Datum x, Datum y, SortSupport ssup) void PrepareSortSupportComparisonShim(Oid cmpFunc, SortSupport ssup) { - SortShimExtra *extra; + SortShimExtra *extra; extra = (SortShimExtra *) MemoryContextAlloc(ssup->ssup_cxt, sizeof(SortShimExtra)); diff --git a/src/backend/utils/sort/tuplesort.c b/src/backend/utils/sort/tuplesort.c index 89698181dbf514456234d059bd448ca4aed04d32..d5a2003e5b896d27bdc46bcab75dc3c38eb48382 100644 --- a/src/backend/utils/sort/tuplesort.c +++ b/src/backend/utils/sort/tuplesort.c @@ -195,8 +195,8 @@ typedef enum #define TAPE_BUFFER_OVERHEAD (BLCKSZ * 3) #define MERGE_BUFFER_SIZE (BLCKSZ * 32) -typedef int (*SortTupleComparator) (const SortTuple *a, const SortTuple *b, - Tuplesortstate *state); +typedef int (*SortTupleComparator) (const SortTuple *a, const SortTuple *b, + Tuplesortstate *state); /* * Private state of a Tuplesort operation. @@ -226,7 +226,7 @@ struct Tuplesortstate * <0, 0, >0 according as a<b, a=b, a>b. The API must match * qsort_arg_comparator. */ - SortTupleComparator comparetup; + SortTupleComparator comparetup; /* * Function to copy a supplied input tuple into palloc'd space and set up @@ -342,13 +342,13 @@ struct Tuplesortstate * tuplesort_begin_heap and used only by the MinimalTuple routines. */ TupleDesc tupDesc; - SortSupport sortKeys; /* array of length nKeys */ + SortSupport sortKeys; /* array of length nKeys */ /* * This variable is shared by the single-key MinimalTuple case and the * Datum case (which both use qsort_ssup()). Otherwise it's NULL. */ - SortSupport onlyKey; + SortSupport onlyKey; /* * These variables are specific to the CLUSTER case; they are set by @@ -634,7 +634,7 @@ tuplesort_begin_heap(TupleDesc tupDesc, for (i = 0; i < nkeys; i++) { - SortSupport sortKey = state->sortKeys + i; + SortSupport sortKey = state->sortKeys + i; AssertArg(attNums[i] != 0); AssertArg(sortOperators[i] != 0); @@ -2685,7 +2685,7 @@ inlineApplySortFunction(FmgrInfo *sortFunction, int sk_flags, Oid collation, static int comparetup_heap(const SortTuple *a, const SortTuple *b, Tuplesortstate *state) { - SortSupport sortKey = state->sortKeys; + SortSupport sortKey = state->sortKeys; HeapTupleData ltup; HeapTupleData rtup; TupleDesc tupDesc; @@ -2806,7 +2806,7 @@ readtup_heap(Tuplesortstate *state, SortTuple *stup, static void reversedirection_heap(Tuplesortstate *state) { - SortSupport sortKey = state->sortKeys; + SortSupport sortKey = state->sortKeys; int nkey; for (nkey = 0; nkey < state->nKeys; nkey++, sortKey++) @@ -3076,9 +3076,10 @@ comparetup_index_btree(const SortTuple *a, const SortTuple *b, bool isnull[INDEX_MAX_KEYS]; /* - * Some rather brain-dead implementations of qsort (such as the one in QNX 4) - * will sometimes call the comparison routine to compare a value to itself, - * but we always use our own implementation, which does not. + * Some rather brain-dead implementations of qsort (such as the one in + * QNX 4) will sometimes call the comparison routine to compare a + * value to itself, but we always use our own implementation, which + * does not. */ Assert(tuple1 != tuple2); @@ -3094,8 +3095,8 @@ comparetup_index_btree(const SortTuple *a, const SortTuple *b, /* * If key values are equal, we sort on ItemPointer. This does not affect - * validity of the finished index, but it may be useful to have index scans - * in physical order. + * validity of the finished index, but it may be useful to have index + * scans in physical order. */ { BlockNumber blk1 = ItemPointerGetBlockNumber(&tuple1->t_tid); @@ -3140,8 +3141,8 @@ comparetup_index_hash(const SortTuple *a, const SortTuple *b, /* * If hash values are equal, we sort on ItemPointer. This does not affect - * validity of the finished index, but it may be useful to have index scans - * in physical order. + * validity of the finished index, but it may be useful to have index + * scans in physical order. */ tuple1 = (IndexTuple) a->tuple; tuple2 = (IndexTuple) b->tuple; diff --git a/src/backend/utils/sort/tuplestore.c b/src/backend/utils/sort/tuplestore.c index 2d30f09ac14d350de4423fcd36951bb4a37acd0a..8a7931b856605fc46c8f9ea886cabe88fbb3e9ca 100644 --- a/src/backend/utils/sort/tuplestore.c +++ b/src/backend/utils/sort/tuplestore.c @@ -569,7 +569,7 @@ tuplestore_puttuple(Tuplestorestate *state, HeapTuple tuple) MemoryContext oldcxt = MemoryContextSwitchTo(state->context); /* - * Copy the tuple. (Must do this even in WRITEFILE case. Note that + * Copy the tuple. (Must do this even in WRITEFILE case. Note that * COPYTUP includes USEMEM, so we needn't do that here.) */ tuple = COPYTUP(state, tuple); diff --git a/src/backend/utils/time/snapmgr.c b/src/backend/utils/time/snapmgr.c index 574099dc9a38b9792cd65ebc09cba0e53117c907..7187ca7c98d7da16ec7fdadb8d839ca995056b58 100644 --- a/src/backend/utils/time/snapmgr.c +++ b/src/backend/utils/time/snapmgr.c @@ -11,7 +11,7 @@ * regd_count and count it in RegisteredSnapshots, but this reference is not * tracked by a resource owner. We used to use the TopTransactionResourceOwner * to track this snapshot reference, but that introduces logical circularity - * and thus makes it impossible to clean up in a sane fashion. It's better to + * and thus makes it impossible to clean up in a sane fashion. It's better to * handle this reference as an internally-tracked registration, so that this * module is entirely lower-level than ResourceOwners. * @@ -113,7 +113,7 @@ static int RegisteredSnapshots = 0; bool FirstSnapshotSet = false; /* - * Remember the serializable transaction snapshot, if any. We cannot trust + * Remember the serializable transaction snapshot, if any. We cannot trust * FirstSnapshotSet in combination with IsolationUsesXactSnapshot(), because * GUC may be reset before us, changing the value of IsolationUsesXactSnapshot. */ @@ -269,23 +269,23 @@ SetTransactionSnapshot(Snapshot sourcesnap, TransactionId sourcexid) * Now we have to fix what GetSnapshotData did with MyPgXact->xmin and * TransactionXmin. There is a race condition: to make sure we are not * causing the global xmin to go backwards, we have to test that the - * source transaction is still running, and that has to be done atomically. - * So let procarray.c do it. + * source transaction is still running, and that has to be done + * atomically. So let procarray.c do it. * - * Note: in serializable mode, predicate.c will do this a second time. - * It doesn't seem worth contorting the logic here to avoid two calls, + * Note: in serializable mode, predicate.c will do this a second time. It + * doesn't seem worth contorting the logic here to avoid two calls, * especially since it's not clear that predicate.c *must* do this. */ if (!ProcArrayInstallImportedXmin(CurrentSnapshot->xmin, sourcexid)) ereport(ERROR, (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE), errmsg("could not import the requested snapshot"), - errdetail("The source transaction %u is not running anymore.", - sourcexid))); + errdetail("The source transaction %u is not running anymore.", + sourcexid))); /* * In transaction-snapshot mode, the first snapshot must live until end of - * xact, so we must make a copy of it. Furthermore, if we're running in + * xact, so we must make a copy of it. Furthermore, if we're running in * serializable mode, predicate.c needs to do its own processing. */ if (IsolationUsesXactSnapshot()) @@ -647,8 +647,8 @@ AtEOXact_Snapshot(bool isCommit) * RegisteredSnapshots to keep the check below happy. But we don't bother * to do FreeSnapshot, for two reasons: the memory will go away with * TopTransactionContext anyway, and if someone has left the snapshot - * stacked as active, we don't want the code below to be chasing through - * a dangling pointer. + * stacked as active, we don't want the code below to be chasing through a + * dangling pointer. */ if (FirstXactSnapshot != NULL) { @@ -668,9 +668,9 @@ AtEOXact_Snapshot(bool isCommit) char buf[MAXPGPATH]; /* - * Get rid of the files. Unlink failure is only a WARNING because - * (1) it's too late to abort the transaction, and (2) leaving a - * leaked file around has little real consequence anyway. + * Get rid of the files. Unlink failure is only a WARNING because (1) + * it's too late to abort the transaction, and (2) leaving a leaked + * file around has little real consequence anyway. */ for (i = 1; i <= list_length(exportedSnapshots); i++) { @@ -745,17 +745,17 @@ ExportSnapshot(Snapshot snapshot) char pathtmp[MAXPGPATH]; /* - * It's tempting to call RequireTransactionChain here, since it's not - * very useful to export a snapshot that will disappear immediately - * afterwards. However, we haven't got enough information to do that, - * since we don't know if we're at top level or not. For example, we - * could be inside a plpgsql function that is going to fire off other - * transactions via dblink. Rather than disallow perfectly legitimate - * usages, don't make a check. + * It's tempting to call RequireTransactionChain here, since it's not very + * useful to export a snapshot that will disappear immediately afterwards. + * However, we haven't got enough information to do that, since we don't + * know if we're at top level or not. For example, we could be inside a + * plpgsql function that is going to fire off other transactions via + * dblink. Rather than disallow perfectly legitimate usages, don't make a + * check. * * Also note that we don't make any restriction on the transaction's - * isolation level; however, importers must check the level if they - * are serializable. + * isolation level; however, importers must check the level if they are + * serializable. */ /* @@ -798,8 +798,8 @@ ExportSnapshot(Snapshot snapshot) /* * Fill buf with a text serialization of the snapshot, plus identification - * data about this transaction. The format expected by ImportSnapshot - * is pretty rigid: each line must be fieldname:value. + * data about this transaction. The format expected by ImportSnapshot is + * pretty rigid: each line must be fieldname:value. */ initStringInfo(&buf); @@ -830,8 +830,8 @@ ExportSnapshot(Snapshot snapshot) appendStringInfo(&buf, "xip:%u\n", topXid); /* - * Similarly, we add our subcommitted child XIDs to the subxid data. - * Here, we have to cope with possible overflow. + * Similarly, we add our subcommitted child XIDs to the subxid data. Here, + * we have to cope with possible overflow. */ if (snapshot->suboverflowed || snapshot->subxcnt + nchildren > GetMaxSnapshotSubxidCount()) @@ -963,16 +963,16 @@ parseXidFromText(const char *prefix, char **s, const char *filename) /* * ImportSnapshot - * Import a previously exported snapshot. The argument should be a - * filename in SNAPSHOT_EXPORT_DIR. Load the snapshot from that file. - * This is called by "SET TRANSACTION SNAPSHOT 'foo'". + * Import a previously exported snapshot. The argument should be a + * filename in SNAPSHOT_EXPORT_DIR. Load the snapshot from that file. + * This is called by "SET TRANSACTION SNAPSHOT 'foo'". */ void ImportSnapshot(const char *idstr) { char path[MAXPGPATH]; FILE *f; - struct stat stat_buf; + struct stat stat_buf; char *filebuf; int xcnt; int i; @@ -985,19 +985,19 @@ ImportSnapshot(const char *idstr) /* * Must be at top level of a fresh transaction. Note in particular that * we check we haven't acquired an XID --- if we have, it's conceivable - * that the snapshot would show it as not running, making for very - * screwy behavior. + * that the snapshot would show it as not running, making for very screwy + * behavior. */ if (FirstSnapshotSet || GetTopTransactionIdIfAny() != InvalidTransactionId || IsSubTransaction()) ereport(ERROR, (errcode(ERRCODE_ACTIVE_SQL_TRANSACTION), - errmsg("SET TRANSACTION SNAPSHOT must be called before any query"))); + errmsg("SET TRANSACTION SNAPSHOT must be called before any query"))); /* - * If we are in read committed mode then the next query would execute - * with a new snapshot thus making this function call quite useless. + * If we are in read committed mode then the next query would execute with + * a new snapshot thus making this function call quite useless. */ if (!IsolationUsesXactSnapshot()) ereport(ERROR, @@ -1100,8 +1100,8 @@ ImportSnapshot(const char *idstr) /* * If we're serializable, the source transaction must be too, otherwise - * predicate.c has problems (SxactGlobalXmin could go backwards). Also, - * a non-read-only transaction can't adopt a snapshot from a read-only + * predicate.c has problems (SxactGlobalXmin could go backwards). Also, a + * non-read-only transaction can't adopt a snapshot from a read-only * transaction, as predicate.c handles the cases very differently. */ if (IsolationIsSerializable()) @@ -1120,15 +1120,15 @@ ImportSnapshot(const char *idstr) * We cannot import a snapshot that was taken in a different database, * because vacuum calculates OldestXmin on a per-database basis; so the * source transaction's xmin doesn't protect us from data loss. This - * restriction could be removed if the source transaction were to mark - * its xmin as being globally applicable. But that would require some + * restriction could be removed if the source transaction were to mark its + * xmin as being globally applicable. But that would require some * additional syntax, since that has to be known when the snapshot is * initially taken. (See pgsql-hackers discussion of 2011-10-21.) */ if (src_dbid != MyDatabaseId) ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("cannot import a snapshot from a different database"))); + errmsg("cannot import a snapshot from a different database"))); /* OK, install the snapshot */ SetTransactionSnapshot(&snapshot, src_xid); diff --git a/src/backend/utils/time/tqual.c b/src/backend/utils/time/tqual.c index 01f73980aff140b68e46d14811e2926df18b616e..4caef9ca25e5322ba12926f8527bcf94a26745ee 100644 --- a/src/backend/utils/time/tqual.c +++ b/src/backend/utils/time/tqual.c @@ -1226,7 +1226,7 @@ HeapTupleSatisfiesVacuum(HeapTupleHeader tuple, TransactionId OldestXmin, * in lieu of HeapTupleSatisifesVacuum when the tuple has just been * tested by HeapTupleSatisfiesMVCC and, therefore, any hint bits that * can be set should already be set. We assume that if no hint bits - * either for xmin or xmax, the transaction is still running. This is + * either for xmin or xmax, the transaction is still running. This is * therefore faster than HeapTupleSatisfiesVacuum, because we don't * consult CLOG (and also because we don't need to give an exact answer, * just whether or not the tuple is surely dead). @@ -1235,10 +1235,10 @@ bool HeapTupleIsSurelyDead(HeapTupleHeader tuple, TransactionId OldestXmin) { /* - * If the inserting transaction is marked invalid, then it aborted, - * and the tuple is definitely dead. If it's marked neither committed - * nor invalid, then we assume it's still alive (since the presumption - * is that all relevant hint bits were just set moments ago). + * If the inserting transaction is marked invalid, then it aborted, and + * the tuple is definitely dead. If it's marked neither committed nor + * invalid, then we assume it's still alive (since the presumption is that + * all relevant hint bits were just set moments ago). */ if (!(tuple->t_infomask & HEAP_XMIN_COMMITTED)) return (tuple->t_infomask & HEAP_XMIN_INVALID) != 0 ? true : false; diff --git a/src/bin/initdb/findtimezone.c b/src/bin/initdb/findtimezone.c index 1ebad182b2d7f14dd7d4cb574d0a26019d83dc33..6d6f96add06dfd5f6aaa3ca5b570d2d41978ac8d 100644 --- a/src/bin/initdb/findtimezone.c +++ b/src/bin/initdb/findtimezone.c @@ -52,7 +52,7 @@ pg_TZDIR(void) * This is simpler than the backend function of the same name because * we assume that the input string has the correct case already, so there * is no need for case-folding. (This is obviously true if we got the file - * name from the filesystem to start with. The only other place it can come + * name from the filesystem to start with. The only other place it can come * from is the environment variable TZ, and there seems no need to allow * case variation in that; other programs aren't likely to.) * @@ -571,7 +571,6 @@ scan_available_timezones(char *tzdir, char *tzdirsub, struct tztry * tt, pgfnames_cleanup(names); } - #else /* WIN32 */ static const struct diff --git a/src/bin/initdb/initdb.c b/src/bin/initdb/initdb.c index 2707334dd52c0dee4609129b59fe430e1a2e488a..3789948cdfb82af769295e6cec88dca7a427cbc2 100644 --- a/src/bin/initdb/initdb.c +++ b/src/bin/initdb/initdb.c @@ -66,32 +66,32 @@ extern const char *select_default_timezone(const char *share_path); static const char *auth_methods_host[] = {"trust", "reject", "md5", "password", "ident", "radius", #ifdef ENABLE_GSS - "gss", + "gss", #endif #ifdef ENABLE_SSPI - "sspi", + "sspi", #endif #ifdef KRB5 - "krb5", + "krb5", #endif #ifdef USE_PAM - "pam", "pam ", + "pam", "pam ", #endif #ifdef USE_LDAP - "ldap", + "ldap", #endif #ifdef USE_SSL - "cert", + "cert", #endif - NULL}; +NULL}; static const char *auth_methods_local[] = {"trust", "reject", "md5", "password", "peer", "radius", #ifdef USE_PAM - "pam", "pam ", + "pam", "pam ", #endif #ifdef USE_LDAP - "ldap", + "ldap", #endif - NULL}; +NULL}; /* * these values are passed in by makefile defines @@ -214,7 +214,7 @@ static void check_ok(void); static char *escape_quotes(const char *src); static int locale_date_order(const char *locale); static bool check_locale_name(int category, const char *locale, - char **canonname); + char **canonname); static bool check_locale_encoding(const char *locale, int encoding); static void setlocales(void); static void usage(const char *progname); @@ -2244,7 +2244,7 @@ locale_date_order(const char *locale) * Is the locale name valid for the locale category? * * If successful, and canonname isn't NULL, a malloc'd copy of the locale's - * canonical name is stored there. This is especially useful for figuring out + * canonical name is stored there. This is especially useful for figuring out * what locale name "" means (ie, the environment value). (Actually, * it seems that on most implementations that's the only thing it's good for; * we could wish that setlocale gave back a canonically spelled version of @@ -2548,7 +2548,7 @@ check_authmethod_unspecified(const char **authmethod) { authwarning = _("\nWARNING: enabling \"trust\" authentication for local connections\n" "You can change this by editing pg_hba.conf or using the option -A, or\n" - "--auth-local and --auth-host, the next time you run initdb.\n"); + "--auth-local and --auth-host, the next time you run initdb.\n"); *authmethod = "trust"; } } @@ -2676,6 +2676,7 @@ main(int argc, char *argv[]) { case 'A': authmethodlocal = authmethodhost = xstrdup(optarg); + /* * When ident is specified, use peer for local connections. * Mirrored, when peer is specified, use ident for TCP/IP @@ -2760,9 +2761,9 @@ main(int argc, char *argv[]) } - /* - * Non-option argument specifies data directory - * as long as it wasn't already specified with -D / --pgdata + /* + * Non-option argument specifies data directory as long as it wasn't + * already specified with -D / --pgdata */ if (optind < argc && strlen(pg_data) == 0) { @@ -3017,18 +3018,18 @@ main(int argc, char *argv[]) */ #ifdef WIN32 printf(_("Encoding \"%s\" implied by locale is not allowed as a server-side encoding.\n" - "The default database encoding will be set to \"%s\" instead.\n"), + "The default database encoding will be set to \"%s\" instead.\n"), pg_encoding_to_char(ctype_enc), pg_encoding_to_char(PG_UTF8)); ctype_enc = PG_UTF8; encodingid = encodingid_to_string(ctype_enc); #else fprintf(stderr, - _("%s: locale \"%s\" requires unsupported encoding \"%s\"\n"), + _("%s: locale \"%s\" requires unsupported encoding \"%s\"\n"), progname, lc_ctype, pg_encoding_to_char(ctype_enc)); fprintf(stderr, - _("Encoding \"%s\" is not allowed as a server-side encoding.\n" - "Rerun %s with a different locale selection.\n"), + _("Encoding \"%s\" is not allowed as a server-side encoding.\n" + "Rerun %s with a different locale selection.\n"), pg_encoding_to_char(ctype_enc), progname); exit(1); #endif diff --git a/src/bin/pg_basebackup/pg_basebackup.c b/src/bin/pg_basebackup/pg_basebackup.c index c3a0d89897ad553d7d28bb12ea20d431c3c46d34..66a281477a738bf789928be0d9ddff382a5718ec 100644 --- a/src/bin/pg_basebackup/pg_basebackup.c +++ b/src/bin/pg_basebackup/pg_basebackup.c @@ -63,6 +63,7 @@ static pid_t bgchild = -1; /* End position for xlog streaming, empty string if unknown yet */ static XLogRecPtr xlogendptr; + #ifndef WIN32 static int has_xlogendptr = 0; #else @@ -223,14 +224,14 @@ typedef struct char xlogdir[MAXPGPATH]; char *sysidentifier; int timeline; -} logstreamer_param; +} logstreamer_param; static int -LogStreamerMain(logstreamer_param * param) +LogStreamerMain(logstreamer_param *param) { if (!ReceiveXlogStream(param->bgconn, param->startptr, param->timeline, param->sysidentifier, param->xlogdir, - reached_end_position, standby_message_timeout, true)) + reached_end_position, standby_message_timeout, true)) /* * Any errors will already have been reported in the function process, @@ -1092,7 +1093,7 @@ BaseBackup(void) int status; int r; #else - DWORD status; + DWORD status; #endif if (verbose) diff --git a/src/bin/pg_basebackup/pg_receivexlog.c b/src/bin/pg_basebackup/pg_receivexlog.c index 67a70bcf713e4b7b1c618fde9035705851df6df7..1acb7ccb56f845ce37bd034af384ee68a62acfee 100644 --- a/src/bin/pg_basebackup/pg_receivexlog.c +++ b/src/bin/pg_basebackup/pg_receivexlog.c @@ -92,7 +92,7 @@ stop_streaming(XLogRecPtr segendpos, uint32 timeline, bool segment_finished) /* * Determine starting location for streaming, based on: * 1. If there are existing xlog segments, start at the end of the last one - * that is complete (size matches XLogSegSize) + * that is complete (size matches XLogSegSize) * 2. If no valid xlog exists, start from the beginning of the current * WAL segment. */ @@ -190,9 +190,10 @@ FindStreamingStart(XLogRecPtr currentpos, uint32 currenttimeline) if (high_log > 0 || high_seg > 0) { XLogRecPtr high_ptr; + /* - * Move the starting pointer to the start of the next segment, - * since the highest one we've seen was completed. + * Move the starting pointer to the start of the next segment, since + * the highest one we've seen was completed. */ NextLogSeg(high_log, high_seg); @@ -284,7 +285,6 @@ sigint_handler(int signum) { time_to_abort = true; } - #endif int @@ -413,9 +413,10 @@ main(int argc, char **argv) { StreamLog(); if (time_to_abort) + /* - * We've been Ctrl-C'ed. That's not an error, so exit without - * an errorcode. + * We've been Ctrl-C'ed. That's not an error, so exit without an + * errorcode. */ exit(0); else if (noloop) diff --git a/src/bin/pg_basebackup/receivelog.c b/src/bin/pg_basebackup/receivelog.c index a51a40edfd17fb892e5faefef5d94b63684dc600..9dd94e1140683ee7ec2c44cb8d1866b980ff7b46 100644 --- a/src/bin/pg_basebackup/receivelog.c +++ b/src/bin/pg_basebackup/receivelog.c @@ -52,7 +52,7 @@ open_walfile(XLogRecPtr startpoint, uint32 timeline, char *basedir, char *namebu { int f; char fn[MAXPGPATH]; - struct stat statbuf; + struct stat statbuf; char *zerobuf; int bytes; @@ -80,7 +80,7 @@ open_walfile(XLogRecPtr startpoint, uint32 timeline, char *basedir, char *namebu return -1; } if (statbuf.st_size == XLogSegSize) - return f; /* File is open and ready to use */ + return f; /* File is open and ready to use */ if (statbuf.st_size != 0) { fprintf(stderr, _("%s: WAL segment %s is %d bytes, should be 0 or %d\n"), @@ -147,8 +147,8 @@ close_walfile(int walfile, char *basedir, char *walname, bool segment_complete) } /* - * Rename the .partial file only if we've completed writing the - * whole segment or segment_complete is true. + * Rename the .partial file only if we've completed writing the whole + * segment or segment_complete is true. */ if (currpos == XLOG_SEG_SIZE || segment_complete) { @@ -202,7 +202,7 @@ localGetCurrentTimestamp(void) */ static void localTimestampDifference(TimestampTz start_time, TimestampTz stop_time, - long *secs, int *microsecs) + long *secs, int *microsecs) { TimestampTz diff = stop_time - start_time; @@ -229,8 +229,8 @@ localTimestampDifference(TimestampTz start_time, TimestampTz stop_time, */ static bool localTimestampDifferenceExceeds(TimestampTz start_time, - TimestampTz stop_time, - int msec) + TimestampTz stop_time, + int msec) { TimestampTz diff = stop_time - start_time; @@ -392,14 +392,14 @@ ReceiveXlogStream(PGconn *conn, XLogRecPtr startpos, uint32 timeline, char *sysi FD_SET(PQsocket(conn), &input_mask); if (standby_message_timeout) { - TimestampTz targettime; + TimestampTz targettime; targettime = TimestampTzPlusMilliseconds(last_status, - standby_message_timeout - 1); + standby_message_timeout - 1); localTimestampDifference(now, targettime, &timeout.tv_sec, - (int *)&timeout.tv_usec); + (int *) &timeout.tv_usec); if (timeout.tv_sec <= 0) timeout.tv_sec = 1; /* Always sleep at least 1 sec */ timeoutptr = &timeout; @@ -444,9 +444,8 @@ ReceiveXlogStream(PGconn *conn, XLogRecPtr startpos, uint32 timeline, char *sysi if (copybuf[0] == 'k') { /* - * keepalive message, sent in 9.2 and newer. We just ignore - * this message completely, but need to skip past it in the - * stream. + * keepalive message, sent in 9.2 and newer. We just ignore this + * message completely, but need to skip past it in the stream. */ if (r != STREAMING_KEEPALIVE_SIZE) { diff --git a/src/bin/pg_basebackup/receivelog.h b/src/bin/pg_basebackup/receivelog.h index 0a803ee4ac1381acddfe248e4f64ddbea18db5fa..7176a68beaab6d9f1d39d19a7e8cd35037e80b0c 100644 --- a/src/bin/pg_basebackup/receivelog.h +++ b/src/bin/pg_basebackup/receivelog.h @@ -4,13 +4,13 @@ * Called before trying to read more data or when a segment is * finished. Return true to stop streaming. */ -typedef bool (*stream_stop_callback)(XLogRecPtr segendpos, uint32 timeline, bool segment_finished); +typedef bool (*stream_stop_callback) (XLogRecPtr segendpos, uint32 timeline, bool segment_finished); extern bool ReceiveXlogStream(PGconn *conn, - XLogRecPtr startpos, - uint32 timeline, - char *sysidentifier, - char *basedir, - stream_stop_callback stream_stop, - int standby_message_timeout, - bool rename_partial); + XLogRecPtr startpos, + uint32 timeline, + char *sysidentifier, + char *basedir, + stream_stop_callback stream_stop, + int standby_message_timeout, + bool rename_partial); diff --git a/src/bin/pg_basebackup/streamutil.c b/src/bin/pg_basebackup/streamutil.c index 1416faa2e3c11437e27b59de1f4df3e6ee2b036b..1b4a9d240b3d57e4c84c123cac966f9e76269870 100644 --- a/src/bin/pg_basebackup/streamutil.c +++ b/src/bin/pg_basebackup/streamutil.c @@ -164,8 +164,8 @@ GetConnection(void) free(keywords); /* - * Ensure we have the same value of integer timestamps as the - * server we are connecting to. + * Ensure we have the same value of integer timestamps as the server + * we are connecting to. */ tmpparam = PQparameterStatus(tmpconn, "integer_datetimes"); if (!tmpparam) diff --git a/src/bin/pg_ctl/pg_ctl.c b/src/bin/pg_ctl/pg_ctl.c index 766ba95be1fb65fd66cc8fb042c7fe92993fd362..38828ec9142bca8ec0b121c0b3ec89550cbe92f3 100644 --- a/src/bin/pg_ctl/pg_ctl.c +++ b/src/bin/pg_ctl/pg_ctl.c @@ -1170,7 +1170,7 @@ do_status(void) } } else - /* must be a postmaster */ + /* must be a postmaster */ { if (postmaster_is_alive((pid_t) pid)) { @@ -1188,9 +1188,12 @@ do_status(void) } } printf(_("%s: no server running\n"), progname); + /* - * The Linux Standard Base Core Specification 3.1 says this should return '3' - * http://refspecs.freestandards.org/LSB_3.1.1/LSB-Core-generic/LSB-Core-generic/iniscrptact.html + * The Linux Standard Base Core Specification 3.1 says this should return + * '3' + * http://refspecs.freestandards.org/LSB_3.1.1/LSB-Core-generic/LSB-Core-ge + * neric/iniscrptact.html */ exit(3); } @@ -1851,7 +1854,7 @@ set_sig(char *signame) sig = SIGABRT; #if 0 /* probably should NOT provide SIGKILL */ - else if (strcmp(signame,"KILL") == 0) + else if (strcmp(signame, "KILL") == 0) sig = SIGKILL; #endif else if (strcmp(signame, "TERM") == 0) @@ -1894,7 +1897,9 @@ set_starttype(char *starttypeopt) static void adjust_data_dir(void) { - char cmd[MAXPGPATH], filename[MAXPGPATH], *my_exec_path; + char cmd[MAXPGPATH], + filename[MAXPGPATH], + *my_exec_path; FILE *fd; /* If there is no postgresql.conf, it can't be a config-only dir */ @@ -2184,7 +2189,7 @@ main(int argc, char **argv) } adjust_data_dir(); - + if (pg_config == NULL && ctl_command != KILL_COMMAND && ctl_command != UNREGISTER_COMMAND) { diff --git a/src/bin/pg_dump/common.c b/src/bin/pg_dump/common.c index 42d3645a6d4f1270856036f497965f41adf31eac..b02217e81d5bebdf6e8258c1a223d1d3944b047a 100644 --- a/src/bin/pg_dump/common.c +++ b/src/bin/pg_dump/common.c @@ -597,8 +597,8 @@ buildIndexArray(void *objArray, int numObjs, Size objSize) static int DOCatalogIdCompare(const void *p1, const void *p2) { - const DumpableObject *obj1 = *(DumpableObject * const *) p1; - const DumpableObject *obj2 = *(DumpableObject * const *) p2; + const DumpableObject *obj1 = *(DumpableObject *const *) p1; + const DumpableObject *obj2 = *(DumpableObject *const *) p2; int cmpval; /* diff --git a/src/bin/pg_dump/dumputils.c b/src/bin/pg_dump/dumputils.c index f5d7b5d85947feb3b63a11048f32e3b47d7a428c..8574380b9cf10000513f8baadd574eaef6f2490b 100644 --- a/src/bin/pg_dump/dumputils.c +++ b/src/bin/pg_dump/dumputils.c @@ -33,11 +33,11 @@ const char *progname = NULL; static struct { - on_exit_nicely_callback function; + on_exit_nicely_callback function; void *arg; -} on_exit_nicely_list[MAX_ON_EXIT_NICELY]; +} on_exit_nicely_list[MAX_ON_EXIT_NICELY]; -static int on_exit_nicely_index; +static int on_exit_nicely_index; #define supports_grant_options(version) ((version) >= 70400) @@ -1221,9 +1221,9 @@ emitShSecLabels(PGconn *conn, PGresult *res, PQExpBuffer buffer, int i; for (i = 0; i < PQntuples(res); i++) - { - char *provider = PQgetvalue(res, i, 0); - char *label = PQgetvalue(res, i, 1); + { + char *provider = PQgetvalue(res, i, 0); + char *label = PQgetvalue(res, i, 1); /* must use fmtId result before calling it again */ appendPQExpBuffer(buffer, @@ -1233,7 +1233,7 @@ emitShSecLabels(PGconn *conn, PGresult *res, PQExpBuffer buffer, " %s IS ", fmtId(objname)); appendStringLiteralConn(buffer, label, conn); - appendPQExpBuffer(buffer, ";\n"); + appendPQExpBuffer(buffer, ";\n"); } } @@ -1252,11 +1252,11 @@ set_dump_section(const char *arg, int *dumpSections) if (*dumpSections == DUMP_UNSECTIONED) *dumpSections = 0; - if (strcmp(arg,"pre-data") == 0) + if (strcmp(arg, "pre-data") == 0) *dumpSections |= DUMP_PRE_DATA; - else if (strcmp(arg,"data") == 0) + else if (strcmp(arg, "data") == 0) *dumpSections |= DUMP_DATA; - else if (strcmp(arg,"post-data") == 0) + else if (strcmp(arg, "post-data") == 0) *dumpSections |= DUMP_POST_DATA; else { @@ -1304,7 +1304,7 @@ vwrite_msg(const char *modulename, const char *fmt, va_list ap) /* - * Fail and die, with a message to stderr. Parameters as for write_msg. + * Fail and die, with a message to stderr. Parameters as for write_msg. */ void exit_horribly(const char *modulename, const char *fmt,...) @@ -1336,11 +1336,11 @@ on_exit_nicely(on_exit_nicely_callback function, void *arg) void exit_nicely(int code) { - int i; + int i; for (i = on_exit_nicely_index - 1; i >= 0; i--) - (*on_exit_nicely_list[i].function)(code, - on_exit_nicely_list[i].arg); + (*on_exit_nicely_list[i].function) (code, + on_exit_nicely_list[i].arg); #ifdef WIN32 if (parallel_init_done && GetCurrentThreadId() != mainThreadId) diff --git a/src/bin/pg_dump/dumputils.h b/src/bin/pg_dump/dumputils.h index 3d1ed9570abbbbee3e4856268052c4fb4f474465..4ef8cb3a4903bc2e9bdca967d76dd051bcb6e692 100644 --- a/src/bin/pg_dump/dumputils.h +++ b/src/bin/pg_dump/dumputils.h @@ -63,12 +63,15 @@ extern void buildShSecLabelQuery(PGconn *conn, const char *catalog_name, extern void emitShSecLabels(PGconn *conn, PGresult *res, PQExpBuffer buffer, const char *target, const char *objname); extern void set_dump_section(const char *arg, int *dumpSections); -extern void write_msg(const char *modulename, const char *fmt,...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); -extern void vwrite_msg(const char *modulename, const char *fmt, va_list ap) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 0))); -extern void exit_horribly(const char *modulename, const char *fmt,...) - __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3), noreturn)); +extern void +write_msg(const char *modulename, const char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); +extern void +vwrite_msg(const char *modulename, const char *fmt, va_list ap) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 0))); +extern void +exit_horribly(const char *modulename, const char *fmt,...) +__attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3), noreturn)); extern void on_exit_nicely(on_exit_nicely_callback function, void *arg); extern void exit_nicely(int code) __attribute__((noreturn)); diff --git a/src/bin/pg_dump/pg_backup.h b/src/bin/pg_dump/pg_backup.h index bf7cc1c1ac8bed9e76f19ef5cd5146f3af99d518..3b49395ecbc2b2adc2f9f459dbd00263f229824d 100644 --- a/src/bin/pg_dump/pg_backup.h +++ b/src/bin/pg_dump/pg_backup.h @@ -110,7 +110,7 @@ typedef struct _restoreOptions const char *filename; int dataOnly; int schemaOnly; - int dumpSections; + int dumpSections; int verbose; int aclsSkip; int tocSummary; diff --git a/src/bin/pg_dump/pg_backup_archiver.c b/src/bin/pg_dump/pg_backup_archiver.c index c049becf126cccbd034474a09f818b0daa0c9e09..5826bace685d9e489be5611f701936e11547db23 100644 --- a/src/bin/pg_dump/pg_backup_archiver.c +++ b/src/bin/pg_dump/pg_backup_archiver.c @@ -95,7 +95,7 @@ typedef struct _parallel_slot typedef struct ShutdownInformation { ParallelState *pstate; - Archive *AHX; + Archive *AHX; } ShutdownInformation; static ShutdownInformation shutdown_info; @@ -529,8 +529,8 @@ restore_toc_entry(ArchiveHandle *AH, TocEntry *te, /* * Ignore DATABASE entry unless we should create it. We must check this - * here, not in _tocEntryRequired, because the createDB option should - * not affect emitting a DATABASE entry to an archive file. + * here, not in _tocEntryRequired, because the createDB option should not + * affect emitting a DATABASE entry to an archive file. */ if (!ropt->createDB && strcmp(te->desc, "DATABASE") == 0) reqs = 0; @@ -1296,7 +1296,7 @@ RestoreOutput(ArchiveHandle *AH, OutputContext savedContext) if (res != 0) exit_horribly(modulename, "could not close output file: %s\n", - strerror(errno)); + strerror(errno)); AH->gzOut = savedContext.gzOut; AH->OF = savedContext.OF; @@ -1317,8 +1317,8 @@ ahprintf(ArchiveHandle *AH, const char *fmt,...) /* * This is paranoid: deal with the possibility that vsnprintf is willing - * to ignore trailing null or returns > 0 even if string does not fit. - * It may be the case that it returns cnt = bufsize. + * to ignore trailing null or returns > 0 even if string does not fit. It + * may be the case that it returns cnt = bufsize. */ while (cnt < 0 || cnt >= (bSize - 1)) { @@ -1456,7 +1456,7 @@ ahwrite(const void *ptr, size_t size, size_t nmemb, ArchiveHandle *AH) res = fwrite(ptr, size, nmemb, AH->OF); if (res != nmemb) exit_horribly(modulename, "could not write to output file: %s\n", - strerror(errno)); + strerror(errno)); return res; } } @@ -1465,7 +1465,7 @@ ahwrite(const void *ptr, size_t size, size_t nmemb, ArchiveHandle *AH) /* on some error, we may decide to go on... */ void warn_or_exit_horribly(ArchiveHandle *AH, - const char *modulename, const char *fmt,...) + const char *modulename, const char *fmt,...) { va_list ap; @@ -1549,7 +1549,7 @@ _moveBefore(ArchiveHandle *AH, TocEntry *pos, TocEntry *te) * items. * * The arrays are indexed by dump ID (so entry zero is unused). Note that the - * array entries run only up to maxDumpId. We might see dependency dump IDs + * array entries run only up to maxDumpId. We might see dependency dump IDs * beyond that (if the dump was partial); so always check the array bound * before trying to touch an array entry. */ @@ -1573,7 +1573,7 @@ buildTocEntryArrays(ArchiveHandle *AH) /* * tableDataId provides the TABLE DATA item's dump ID for each TABLE - * TOC entry that has a DATA item. We compute this by reversing the + * TOC entry that has a DATA item. We compute this by reversing the * TABLE DATA item's dependency, knowing that a TABLE DATA item has * just one dependency and it is the TABLE item. */ @@ -1925,8 +1925,8 @@ _discoverArchiveFormat(ArchiveHandle *AH) else { /* - * *Maybe* we have a tar archive format file or a text dump ... - * So, read first 512 byte header... + * *Maybe* we have a tar archive format file or a text dump ... So, + * read first 512 byte header... */ cnt = fread(&AH->lookahead[AH->lookaheadLen], 1, 512 - AH->lookaheadLen, fh); AH->lookaheadLen += cnt; @@ -1935,7 +1935,10 @@ _discoverArchiveFormat(ArchiveHandle *AH) (strncmp(AH->lookahead, TEXT_DUMP_HEADER, strlen(TEXT_DUMP_HEADER)) == 0 || strncmp(AH->lookahead, TEXT_DUMPALL_HEADER, strlen(TEXT_DUMPALL_HEADER)) == 0)) { - /* looks like it's probably a text format dump. so suggest they try psql */ + /* + * looks like it's probably a text format dump. so suggest they + * try psql + */ exit_horribly(modulename, "input file appears to be a text format dump. Please use psql.\n"); } @@ -2217,7 +2220,7 @@ ReadToc(ArchiveHandle *AH) /* Sanity check */ if (te->dumpId <= 0) exit_horribly(modulename, - "entry ID %d out of range -- perhaps a corrupt TOC\n", + "entry ID %d out of range -- perhaps a corrupt TOC\n", te->dumpId); te->hadDumper = ReadInt(AH); @@ -2835,8 +2838,8 @@ _selectTablespace(ArchiveHandle *AH, const char *tablespace) if (!res || PQresultStatus(res) != PGRES_COMMAND_OK) warn_or_exit_horribly(AH, modulename, - "could not set default_tablespace to %s: %s", - fmtId(want), PQerrorMessage(AH->connection)); + "could not set default_tablespace to %s: %s", + fmtId(want), PQerrorMessage(AH->connection)); PQclear(res); } @@ -3043,7 +3046,7 @@ _printTocEntry(ArchiveHandle *AH, TocEntry *te, RestoreOptions *ropt, bool isDat if (te->tablespace && !ropt->noTablespace) { - char *sanitized_tablespace; + char *sanitized_tablespace; sanitized_tablespace = replace_line_endings(te->tablespace); ahprintf(AH, "; Tablespace: %s", sanitized_tablespace); @@ -3150,8 +3153,8 @@ _printTocEntry(ArchiveHandle *AH, TocEntry *te, RestoreOptions *ropt, bool isDat static char * replace_line_endings(const char *str) { - char *result; - char *s; + char *result; + char *s; result = pg_strdup(str); @@ -3381,7 +3384,7 @@ unsetProcessIdentifier(ParallelStateEntry *pse) static ParallelStateEntry * GetMyPSEntry(ParallelState *pstate) { - int i; + int i; for (i = 0; i < pstate->numWorkers; i++) #ifdef WIN32 @@ -3509,8 +3512,8 @@ restore_toc_entries_parallel(ArchiveHandle *AH) DisconnectDatabase(&AH->public); /* - * Set the pstate in the shutdown_info. The exit handler uses pstate if set - * and falls back to AHX otherwise. + * Set the pstate in the shutdown_info. The exit handler uses pstate if + * set and falls back to AHX otherwise. */ shutdown_info.pstate = pstate; diff --git a/src/bin/pg_dump/pg_backup_archiver.h b/src/bin/pg_dump/pg_backup_archiver.h index 4361805baab8eea273a7bcc74d0396704709d9e9..8859bd9776e33a7943c6b5f00951033a4f035371 100644 --- a/src/bin/pg_dump/pg_backup_archiver.h +++ b/src/bin/pg_dump/pg_backup_archiver.h @@ -256,7 +256,7 @@ typedef struct _archiveHandle DumpId maxDumpId; /* largest DumpId among all TOC entries */ /* arrays created after the TOC list is complete: */ - struct _tocEntry **tocsByDumpId; /* TOCs indexed by dumpId */ + struct _tocEntry **tocsByDumpId; /* TOCs indexed by dumpId */ DumpId *tableDataId; /* TABLE DATA ids, indexed by table dumpId */ struct _tocEntry *currToc; /* Used when dumping data */ diff --git a/src/bin/pg_dump/pg_backup_custom.c b/src/bin/pg_dump/pg_backup_custom.c index 2156d0f619c2b540f2532c100ba6e4b2212f600e..204309d6ad63db65dd9e6f4b1525958de0c0686a 100644 --- a/src/bin/pg_dump/pg_backup_custom.c +++ b/src/bin/pg_dump/pg_backup_custom.c @@ -466,7 +466,7 @@ _PrintTocData(ArchiveHandle *AH, TocEntry *te, RestoreOptions *ropt) else if (!ctx->hasSeek) exit_horribly(modulename, "could not find block ID %d in archive -- " "possibly due to out-of-order restore request, " - "which cannot be handled due to non-seekable input file\n", + "which cannot be handled due to non-seekable input file\n", te->dumpId); else /* huh, the dataPos led us to EOF? */ exit_horribly(modulename, "could not find block ID %d in archive -- " @@ -572,10 +572,10 @@ _skipData(ArchiveHandle *AH) { if (feof(AH->FH)) exit_horribly(modulename, - "could not read from input file: end of file\n"); + "could not read from input file: end of file\n"); else exit_horribly(modulename, - "could not read from input file: %s\n", strerror(errno)); + "could not read from input file: %s\n", strerror(errno)); } ctx->filePos += blkLen; @@ -646,7 +646,7 @@ _WriteBuf(ArchiveHandle *AH, const void *buf, size_t len) if (res != len) exit_horribly(modulename, - "could not write to output file: %s\n", strerror(errno)); + "could not write to output file: %s\n", strerror(errno)); ctx->filePos += res; return res; diff --git a/src/bin/pg_dump/pg_backup_db.c b/src/bin/pg_dump/pg_backup_db.c index b315e681926dc343481e49822d9cb2b71dcb788f..d912aaf77f309ec8308c6796085c40894a17d394 100644 --- a/src/bin/pg_dump/pg_backup_db.c +++ b/src/bin/pg_dump/pg_backup_db.c @@ -182,7 +182,7 @@ _connectDB(ArchiveHandle *AH, const char *reqdb, const char *requser) { if (!PQconnectionNeedsPassword(newConn)) exit_horribly(modulename, "could not reconnect to database: %s", - PQerrorMessage(newConn)); + PQerrorMessage(newConn)); PQfinish(newConn); if (password) @@ -300,7 +300,7 @@ ConnectDatabase(Archive *AHX, /* check to see that the backend connection was successfully made */ if (PQstatus(AH->connection) == CONNECTION_BAD) exit_horribly(modulename, "connection to database \"%s\" failed: %s", - PQdb(AH->connection), PQerrorMessage(AH->connection)); + PQdb(AH->connection), PQerrorMessage(AH->connection)); /* check for version mismatch */ _check_database_version(AH); @@ -313,7 +313,7 @@ DisconnectDatabase(Archive *AHX) { ArchiveHandle *AH = (ArchiveHandle *) AHX; - PQfinish(AH->connection); /* noop if AH->connection is NULL */ + PQfinish(AH->connection); /* noop if AH->connection is NULL */ AH->connection = NULL; } @@ -343,7 +343,7 @@ die_on_query_failure(ArchiveHandle *AH, const char *modulename, const char *quer void ExecuteSqlStatement(Archive *AHX, const char *query) { - ArchiveHandle *AH = (ArchiveHandle *) AHX; + ArchiveHandle *AH = (ArchiveHandle *) AHX; PGresult *res; res = PQexec(AH->connection, query); @@ -355,7 +355,7 @@ ExecuteSqlStatement(Archive *AHX, const char *query) PGresult * ExecuteSqlQuery(Archive *AHX, const char *query, ExecStatusType status) { - ArchiveHandle *AH = (ArchiveHandle *) AHX; + ArchiveHandle *AH = (ArchiveHandle *) AHX; PGresult *res; res = PQexec(AH->connection, query); @@ -436,7 +436,7 @@ ExecuteInsertCommands(ArchiveHandle *AH, const char *buf, size_t bufLen) for (; qry < eos; qry++) { - char ch = *qry; + char ch = *qry; /* For neatness, we skip any newlines between commands */ if (!(ch == '\n' && AH->sqlparse.curCmd->len == 0)) @@ -526,7 +526,7 @@ ExecuteSqlCommandBuf(ArchiveHandle *AH, const char *buf, size_t bufLen) ExecuteSqlCommand(AH, buf, "could not execute query"); else { - char *str = (char *) pg_malloc(bufLen + 1); + char *str = (char *) pg_malloc(bufLen + 1); memcpy(str, buf, bufLen); str[bufLen] = '\0'; diff --git a/src/bin/pg_dump/pg_backup_directory.c b/src/bin/pg_dump/pg_backup_directory.c index 629e309abe57007d95ffcfbd68b0351f20cea74e..32dcb12fdf8c9240a6b26f793fc5437ba8833b2b 100644 --- a/src/bin/pg_dump/pg_backup_directory.c +++ b/src/bin/pg_dump/pg_backup_directory.c @@ -178,7 +178,7 @@ InitArchiveFmt_Directory(ArchiveHandle *AH) /* Nothing else in the file, so close it again... */ if (cfclose(tocFH) != 0) exit_horribly(modulename, "could not close TOC file: %s\n", - strerror(errno)); + strerror(errno)); ctx->dataFH = NULL; } } @@ -347,7 +347,7 @@ _PrintFileData(ArchiveHandle *AH, char *filename, RestoreOptions *ropt) if (!cfp) exit_horribly(modulename, "could not open input file \"%s\": %s\n", - filename, strerror(errno)); + filename, strerror(errno)); buf = pg_malloc(ZLIB_OUT_SIZE); buflen = ZLIB_OUT_SIZE; @@ -356,9 +356,9 @@ _PrintFileData(ArchiveHandle *AH, char *filename, RestoreOptions *ropt) ahwrite(buf, 1, cnt, AH); free(buf); - if (cfclose(cfp) != 0) + if (cfclose(cfp) !=0) exit_horribly(modulename, "could not close data file: %s\n", - strerror(errno)); + strerror(errno)); } /* @@ -417,7 +417,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt) } if (!cfeof(ctx->blobsTocFH)) exit_horribly(modulename, "error reading large object TOC file \"%s\"\n", - fname); + fname); if (cfclose(ctx->blobsTocFH) != 0) exit_horribly(modulename, "could not close large object TOC file \"%s\": %s\n", @@ -478,7 +478,7 @@ _WriteBuf(ArchiveHandle *AH, const void *buf, size_t len) res = cfwrite(buf, len, ctx->dataFH); if (res != len) exit_horribly(modulename, "could not write to output file: %s\n", - strerror(errno)); + strerror(errno)); return res; } @@ -589,7 +589,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid) if (ctx->dataFH == NULL) exit_horribly(modulename, "could not open output file \"%s\": %s\n", - fname, strerror(errno)); + fname, strerror(errno)); } /* diff --git a/src/bin/pg_dump/pg_backup_tar.c b/src/bin/pg_dump/pg_backup_tar.c index 9fe2b14df5560e127562b7e01fcb68d671c39a56..c5e19968b7c61efe85d5059b94d62287c9d795ec 100644 --- a/src/bin/pg_dump/pg_backup_tar.c +++ b/src/bin/pg_dump/pg_backup_tar.c @@ -111,7 +111,7 @@ static void tarClose(ArchiveHandle *AH, TAR_MEMBER *TH); #ifdef __NOT_USED__ static char *tarGets(char *buf, size_t len, TAR_MEMBER *th); #endif -static int tarPrintf(ArchiveHandle *AH, TAR_MEMBER *th, const char *fmt, ...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 3, 4))); +static int tarPrintf(ArchiveHandle *AH, TAR_MEMBER *th, const char *fmt,...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 3, 4))); static void _tarAddFile(ArchiveHandle *AH, TAR_MEMBER *th); static int _tarChecksum(char *th); @@ -177,7 +177,7 @@ InitArchiveFmt_Tar(ArchiveHandle *AH) ctx->tarFH = fopen(AH->fSpec, PG_BINARY_W); if (ctx->tarFH == NULL) exit_horribly(modulename, - "could not open TOC file \"%s\" for output: %s\n", + "could not open TOC file \"%s\" for output: %s\n", AH->fSpec, strerror(errno)); } else @@ -213,7 +213,7 @@ InitArchiveFmt_Tar(ArchiveHandle *AH) */ if (AH->compression != 0) exit_horribly(modulename, - "compression is not supported by tar archive format\n"); + "compression is not supported by tar archive format\n"); } else { /* Read Mode */ @@ -585,7 +585,7 @@ tarWrite(const void *buf, size_t len, TAR_MEMBER *th) if (res != len) exit_horribly(modulename, - "could not write to output file: %s\n", strerror(errno)); + "could not write to output file: %s\n", strerror(errno)); th->pos += res; return res; @@ -1230,7 +1230,7 @@ _tarGetHeader(ArchiveHandle *AH, TAR_MEMBER *th) snprintf(buf2, sizeof(buf2), INT64_FORMAT, (int64) ftello(ctx->tarFHpos)); exit_horribly(modulename, "mismatch in actual vs. predicted file position (%s vs. %s)\n", - buf1, buf2); + buf1, buf2); } #endif @@ -1245,7 +1245,7 @@ _tarGetHeader(ArchiveHandle *AH, TAR_MEMBER *th) if (len != 512) exit_horribly(modulename, ngettext("incomplete tar header found (%lu byte)\n", - "incomplete tar header found (%lu bytes)\n", + "incomplete tar header found (%lu bytes)\n", len), (unsigned long) len); diff --git a/src/bin/pg_dump/pg_dump.c b/src/bin/pg_dump/pg_dump.c index 3461f3e34c39ab7be68c205e271331043cabc72e..d9aeee3c8d252922711ba3da1993af243d2953e8 100644 --- a/src/bin/pg_dump/pg_dump.c +++ b/src/bin/pg_dump/pg_dump.c @@ -221,9 +221,9 @@ static char *format_function_arguments_old(Archive *fout, char **argmodes, char **argnames); static char *format_function_signature(Archive *fout, - FuncInfo *finfo, bool honor_quotes); + FuncInfo *finfo, bool honor_quotes); static const char *convertRegProcReference(Archive *fout, - const char *proc); + const char *proc); static const char *convertOperatorReference(Archive *fout, const char *opr); static const char *convertTSFunction(Archive *fout, Oid funcOid); static Oid findLastBuiltinOid_V71(Archive *fout, const char *); @@ -232,7 +232,7 @@ static void selectSourceSchema(Archive *fout, const char *schemaName); static char *getFormattedTypeName(Archive *fout, Oid oid, OidOptions opts); static char *myFormatType(const char *typname, int32 typmod); static const char *fmtQualifiedId(Archive *fout, - const char *schema, const char *id); + const char *schema, const char *id); static void getBlobs(Archive *fout); static void dumpBlob(Archive *fout, BlobInfo *binfo); static int dumpBlobs(Archive *fout, void *arg); @@ -285,7 +285,7 @@ main(int argc, char **argv) RestoreOptions *ropt; ArchiveFormat archiveFormat = archUnknown; ArchiveMode archiveMode; - Archive *fout; /* the script file */ + Archive *fout; /* the script file */ static int disable_triggers = 0; static int outputNoTablespaces = 0; @@ -495,7 +495,7 @@ main(int argc, char **argv) use_role = optarg; break; - case 4: /* exclude table(s) data */ + case 4: /* exclude table(s) data */ simple_string_list_append(&tabledata_exclude_patterns, optarg); break; @@ -605,7 +605,7 @@ main(int argc, char **argv) "SERIALIZABLE, READ ONLY, DEFERRABLE"); else ExecuteSqlStatement(fout, - "SET TRANSACTION ISOLATION LEVEL " + "SET TRANSACTION ISOLATION LEVEL " "REPEATABLE READ"); } else @@ -625,7 +625,7 @@ main(int argc, char **argv) { if (fout->remoteVersion >= 70100) g_last_builtin_oid = findLastBuiltinOid_V71(fout, - PQdb(GetConnection(fout))); + PQdb(GetConnection(fout))); else g_last_builtin_oid = findLastBuiltinOid_V70(fout); if (g_verbose) @@ -748,7 +748,7 @@ main(int argc, char **argv) else ropt->compression = compressLevel; - ropt->suppressDumpWarnings = true; /* We've already shown them */ + ropt->suppressDumpWarnings = true; /* We've already shown them */ SetArchiveRestoreOptions(fout, ropt); @@ -1123,6 +1123,7 @@ selectDumpableType(TypeInfo *tyinfo) if (tyinfo->isArray) { tyinfo->dobj.objType = DO_DUMMY_TYPE; + /* * Fall through to set the dump flag; we assume that the subsequent * rules will do the same thing as they would for the array's base @@ -2666,7 +2667,7 @@ findNamespace(Archive *fout, Oid nsoid, Oid objoid) else { /* This code depends on the dummy objects set up by getNamespaces. */ - Oid i; + Oid i; if (objoid > g_last_builtin_oid) i = 0; /* user object */ @@ -2938,7 +2939,7 @@ getTypes(Archive *fout, int *numTypes) /* * If it's a base type, make a DumpableObject representing a shell * definition of the type. We will need to dump that ahead of the I/O - * functions for the type. Similarly, range types need a shell + * functions for the type. Similarly, range types need a shell * definition in case they have a canonicalize function. * * Note: the shell type doesn't have a catId. You might think it @@ -3972,7 +3973,7 @@ getTables(Archive *fout, int *numTables) "SELECT c.tableoid, c.oid, c.relname, " "c.relacl, c.relkind, c.relnamespace, " "(%s c.relowner) AS rolname, " - "c.relchecks, (c.reltriggers <> 0) AS relhastriggers, " + "c.relchecks, (c.reltriggers <> 0) AS relhastriggers, " "c.relhasindex, c.relhasrules, c.relhasoids, " "c.relfrozenxid, tc.oid AS toid, " "tc.relfrozenxid AS tfrozenxid, " @@ -4278,9 +4279,9 @@ getTables(Archive *fout, int *numTables) resetPQExpBuffer(query); appendPQExpBuffer(query, "LOCK TABLE %s IN ACCESS SHARE MODE", - fmtQualifiedId(fout, + fmtQualifiedId(fout, tblinfo[i].dobj.namespace->dobj.name, - tblinfo[i].dobj.name)); + tblinfo[i].dobj.name)); ExecuteSqlStatement(fout, query->data); } @@ -4879,7 +4880,7 @@ getDomainConstraints(Archive *fout, TypeInfo *tyinfo) for (i = 0; i < ntups; i++) { - bool validated = PQgetvalue(res, i, 4)[0] == 't'; + bool validated = PQgetvalue(res, i, 4)[0] == 't'; constrinfo[i].dobj.objType = DO_CONSTRAINT; constrinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); @@ -4901,7 +4902,7 @@ getDomainConstraints(Archive *fout, TypeInfo *tyinfo) /* * Make the domain depend on the constraint, ensuring it won't be - * output till any constraint dependencies are OK. If the constraint + * output till any constraint dependencies are OK. If the constraint * has not been validated, it's going to be dumped after the domain * anyway, so this doesn't matter. */ @@ -5625,11 +5626,11 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) "pg_catalog.format_type(t.oid,a.atttypmod) AS atttypname, " "array_to_string(a.attoptions, ', ') AS attoptions, " "CASE WHEN a.attcollation <> t.typcollation " - "THEN a.attcollation ELSE 0 END AS attcollation, " + "THEN a.attcollation ELSE 0 END AS attcollation, " "pg_catalog.array_to_string(ARRAY(" "SELECT pg_catalog.quote_ident(option_name) || " "' ' || pg_catalog.quote_literal(option_value) " - "FROM pg_catalog.pg_options_to_table(attfdwoptions) " + "FROM pg_catalog.pg_options_to_table(attfdwoptions) " "ORDER BY option_name" "), E',\n ') AS attfdwoptions " "FROM pg_catalog.pg_attribute a LEFT JOIN pg_catalog.pg_type t " @@ -5654,7 +5655,7 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) "pg_catalog.format_type(t.oid,a.atttypmod) AS atttypname, " "array_to_string(a.attoptions, ', ') AS attoptions, " "CASE WHEN a.attcollation <> t.typcollation " - "THEN a.attcollation ELSE 0 END AS attcollation, " + "THEN a.attcollation ELSE 0 END AS attcollation, " "NULL AS attfdwoptions " "FROM pg_catalog.pg_attribute a LEFT JOIN pg_catalog.pg_type t " "ON a.atttypid = t.oid " @@ -5898,8 +5899,8 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) /* * Defaults on a VIEW must always be dumped as separate ALTER * TABLE commands. Defaults on regular tables are dumped as - * part of the CREATE TABLE if possible, which it won't be - * if the column is not going to be emitted explicitly. + * part of the CREATE TABLE if possible, which it won't be if + * the column is not going to be emitted explicitly. */ if (tbinfo->relkind == RELKIND_VIEW) { @@ -5919,6 +5920,7 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) else { attrdefs[j].separate = false; + /* * Mark the default as needing to appear before the table, * so that any dependencies it has must be emitted before @@ -6051,7 +6053,7 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) for (j = 0; j < numConstrs; j++) { - bool validated = PQgetvalue(res, j, 5)[0] == 't'; + bool validated = PQgetvalue(res, j, 5)[0] == 't'; constrs[j].dobj.objType = DO_CONSTRAINT; constrs[j].dobj.catId.tableoid = atooid(PQgetvalue(res, j, 0)); @@ -6068,6 +6070,7 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) constrs[j].condeferrable = false; constrs[j].condeferred = false; constrs[j].conislocal = (PQgetvalue(res, j, 4)[0] == 't'); + /* * An unvalidated constraint needs to be dumped separately, so * that potentially-violating existing data is loaded before @@ -6081,10 +6084,10 @@ getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) * Mark the constraint as needing to appear before the table * --- this is so that any other dependencies of the * constraint will be emitted before we try to create the - * table. If the constraint is to be dumped separately, it will be - * dumped after data is loaded anyway, so don't do it. (There's - * an automatic dependency in the opposite direction anyway, so - * don't need to add one manually here.) + * table. If the constraint is to be dumped separately, it + * will be dumped after data is loaded anyway, so don't do it. + * (There's an automatic dependency in the opposite direction + * anyway, so don't need to add one manually here.) */ if (!constrs[j].separate) addObjectDependency(&tbinfo->dobj, @@ -6597,7 +6600,7 @@ getForeignServers(Archive *fout, int *numForeignServers) } /* Make sure we are in proper schema */ - selectSourceSchema(fout,"pg_catalog"); + selectSourceSchema(fout, "pg_catalog"); appendPQExpBuffer(query, "SELECT tableoid, oid, srvname, " "(%s srvowner) AS rolname, " @@ -7531,7 +7534,7 @@ dumpRangeType(Archive *fout, TypeInfo *tyinfo) selectSourceSchema(fout, tyinfo->dobj.namespace->dobj.name); appendPQExpBuffer(query, - "SELECT pg_catalog.format_type(rngsubtype, NULL) AS rngsubtype, " + "SELECT pg_catalog.format_type(rngsubtype, NULL) AS rngsubtype, " "opc.opcname AS opcname, " "(SELECT nspname FROM pg_catalog.pg_namespace nsp " " WHERE nsp.oid = opc.opcnamespace) AS opcnsp, " @@ -7570,8 +7573,8 @@ dumpRangeType(Archive *fout, TypeInfo *tyinfo) /* print subtype_opclass only if not default for subtype */ if (PQgetvalue(res, 0, PQfnumber(res, "opcdefault"))[0] != 't') { - char *opcname = PQgetvalue(res, 0, PQfnumber(res, "opcname")); - char *nspname = PQgetvalue(res, 0, PQfnumber(res, "opcnsp")); + char *opcname = PQgetvalue(res, 0, PQfnumber(res, "opcname")); + char *nspname = PQgetvalue(res, 0, PQfnumber(res, "opcnsp")); /* always schema-qualify, don't try to be smart */ appendPQExpBuffer(q, ",\n subtype_opclass = %s.", @@ -9409,12 +9412,12 @@ dumpCast(Archive *fout, CastInfo *cast) labelq = createPQExpBuffer(); appendPQExpBuffer(delqry, "DROP CAST (%s AS %s);\n", - getFormattedTypeName(fout, cast->castsource, zeroAsNone), - getFormattedTypeName(fout, cast->casttarget, zeroAsNone)); + getFormattedTypeName(fout, cast->castsource, zeroAsNone), + getFormattedTypeName(fout, cast->casttarget, zeroAsNone)); appendPQExpBuffer(defqry, "CREATE CAST (%s AS %s) ", - getFormattedTypeName(fout, cast->castsource, zeroAsNone), - getFormattedTypeName(fout, cast->casttarget, zeroAsNone)); + getFormattedTypeName(fout, cast->castsource, zeroAsNone), + getFormattedTypeName(fout, cast->casttarget, zeroAsNone)); switch (cast->castmethod) { @@ -9427,14 +9430,15 @@ dumpCast(Archive *fout, CastInfo *cast) case COERCION_METHOD_FUNCTION: if (funcInfo) { - char *fsig = format_function_signature(fout, funcInfo, true); + char *fsig = format_function_signature(fout, funcInfo, true); /* * Always qualify the function name, in case it is not in - * pg_catalog schema (format_function_signature won't qualify it). + * pg_catalog schema (format_function_signature won't qualify + * it). */ appendPQExpBuffer(defqry, "WITH FUNCTION %s.%s", - fmtId(funcInfo->dobj.namespace->dobj.name), fsig); + fmtId(funcInfo->dobj.namespace->dobj.name), fsig); free(fsig); } else @@ -9451,8 +9455,8 @@ dumpCast(Archive *fout, CastInfo *cast) appendPQExpBuffer(defqry, ";\n"); appendPQExpBuffer(labelq, "CAST (%s AS %s)", - getFormattedTypeName(fout, cast->castsource, zeroAsNone), - getFormattedTypeName(fout, cast->casttarget, zeroAsNone)); + getFormattedTypeName(fout, cast->castsource, zeroAsNone), + getFormattedTypeName(fout, cast->casttarget, zeroAsNone)); if (binary_upgrade) binary_upgrade_extension_member(defqry, &cast->dobj, labelq->data); @@ -11715,7 +11719,7 @@ dumpACL(Archive *fout, CatalogId objCatId, DumpId objDumpId, if (!buildACLCommands(name, subname, type, acls, owner, "", fout->remoteVersion, sql)) exit_horribly(NULL, - "could not parse ACL list (%s) for object \"%s\" (%s)\n", + "could not parse ACL list (%s) for object \"%s\" (%s)\n", acls, name, type); if (sql->len > 0) @@ -12157,10 +12161,10 @@ dumpTableSchema(Archive *fout, TableInfo *tbinfo) { if (PQntuples(res) < 1) exit_horribly(NULL, "query to obtain definition of view \"%s\" returned no data\n", - tbinfo->dobj.name); + tbinfo->dobj.name); else exit_horribly(NULL, "query to obtain definition of view \"%s\" returned more than one definition\n", - tbinfo->dobj.name); + tbinfo->dobj.name); } viewdef = PQgetvalue(res, 0, 0); @@ -12207,7 +12211,7 @@ dumpTableSchema(Archive *fout, TableInfo *tbinfo) "pg_catalog.array_to_string(ARRAY(" "SELECT pg_catalog.quote_ident(option_name) || " "' ' || pg_catalog.quote_literal(option_value) " - "FROM pg_catalog.pg_options_to_table(ftoptions) " + "FROM pg_catalog.pg_options_to_table(ftoptions) " "ORDER BY option_name" "), E',\n ') AS ftoptions " "FROM pg_catalog.pg_foreign_table ft " @@ -13152,7 +13156,7 @@ findLastBuiltinOid_V70(Archive *fout) int last_oid; res = ExecuteSqlQueryForSingleRow(fout, - "SELECT oid FROM pg_class WHERE relname = 'pg_indexes'"); + "SELECT oid FROM pg_class WHERE relname = 'pg_indexes'"); last_oid = atooid(PQgetvalue(res, 0, PQfnumber(res, "oid"))); PQclear(res); return last_oid; @@ -13882,8 +13886,8 @@ getExtensionMembership(Archive *fout, ExtensionInfo extinfo[], continue; /* - * Note: config tables are dumped without OIDs regardless - * of the --oids setting. This is because row filtering + * Note: config tables are dumped without OIDs regardless of + * the --oids setting. This is because row filtering * conditions aren't compatible with dumping OIDs. */ makeTableDataInfo(configtbl, false); @@ -14284,7 +14288,7 @@ ExecuteSqlQueryForSingleRow(Archive *fout, char *query) exit_horribly(NULL, ngettext("query returned %d row instead of one: %s\n", "query returned %d rows instead of one: %s\n", - ntups), + ntups), ntups, query); return res; diff --git a/src/bin/pg_dump/pg_dump_sort.c b/src/bin/pg_dump/pg_dump_sort.c index 2a1b4299ca89dcd0967ce4971dd9277c83d828fe..9a82e4b6c58a6d22f6785d261ddf9fe1a8c34e96 100644 --- a/src/bin/pg_dump/pg_dump_sort.c +++ b/src/bin/pg_dump/pg_dump_sort.c @@ -111,7 +111,7 @@ static bool TopoSort(DumpableObject **objs, static void addHeapElement(int val, int *heap, int heapLength); static int removeHeapElement(int *heap, int heapLength); static void findDependencyLoops(DumpableObject **objs, int nObjs, int totObjs); -static int findLoop(DumpableObject *obj, +static int findLoop(DumpableObject *obj, DumpId startPoint, bool *processed, DumpableObject **workspace, @@ -139,8 +139,8 @@ sortDumpableObjectsByTypeName(DumpableObject **objs, int numObjs) static int DOTypeNameCompare(const void *p1, const void *p2) { - DumpableObject *obj1 = *(DumpableObject * const *) p1; - DumpableObject *obj2 = *(DumpableObject * const *) p2; + DumpableObject *obj1 = *(DumpableObject *const *) p1; + DumpableObject *obj2 = *(DumpableObject *const *) p2; int cmpval; /* Sort by type */ @@ -171,8 +171,8 @@ DOTypeNameCompare(const void *p1, const void *p2) /* To have a stable sort order, break ties for some object types */ if (obj1->objType == DO_FUNC || obj1->objType == DO_AGG) { - FuncInfo *fobj1 = *(FuncInfo * const *) p1; - FuncInfo *fobj2 = *(FuncInfo * const *) p2; + FuncInfo *fobj1 = *(FuncInfo *const *) p1; + FuncInfo *fobj2 = *(FuncInfo *const *) p2; cmpval = fobj1->nargs - fobj2->nargs; if (cmpval != 0) @@ -180,8 +180,8 @@ DOTypeNameCompare(const void *p1, const void *p2) } else if (obj1->objType == DO_OPERATOR) { - OprInfo *oobj1 = *(OprInfo * const *) p1; - OprInfo *oobj2 = *(OprInfo * const *) p2; + OprInfo *oobj1 = *(OprInfo *const *) p1; + OprInfo *oobj2 = *(OprInfo *const *) p2; /* oprkind is 'l', 'r', or 'b'; this sorts prefix, postfix, infix */ cmpval = (oobj2->oprkind - oobj1->oprkind); @@ -190,8 +190,8 @@ DOTypeNameCompare(const void *p1, const void *p2) } else if (obj1->objType == DO_ATTRDEF) { - AttrDefInfo *adobj1 = *(AttrDefInfo * const *) p1; - AttrDefInfo *adobj2 = *(AttrDefInfo * const *) p2; + AttrDefInfo *adobj1 = *(AttrDefInfo *const *) p1; + AttrDefInfo *adobj2 = *(AttrDefInfo *const *) p2; cmpval = (adobj1->adnum - adobj2->adnum); if (cmpval != 0) @@ -220,8 +220,8 @@ sortDumpableObjectsByTypeOid(DumpableObject **objs, int numObjs) static int DOTypeOidCompare(const void *p1, const void *p2) { - DumpableObject *obj1 = *(DumpableObject * const *) p1; - DumpableObject *obj2 = *(DumpableObject * const *) p2; + DumpableObject *obj1 = *(DumpableObject *const *) p1; + DumpableObject *obj2 = *(DumpableObject *const *) p2; int cmpval; cmpval = oldObjectTypePriority[obj1->objType] - @@ -545,7 +545,7 @@ findDependencyLoops(DumpableObject **objs, int nObjs, int totObjs) { /* * There's no loop starting at this object, but mark it processed - * anyway. This is not necessary for correctness, but saves later + * anyway. This is not necessary for correctness, but saves later * invocations of findLoop() from uselessly chasing references to * such an object. */ @@ -587,7 +587,7 @@ findLoop(DumpableObject *obj, int i; /* - * Reject if obj is already processed. This test prevents us from finding + * Reject if obj is already processed. This test prevents us from finding * loops that overlap previously-processed loops. */ if (processed[obj->dumpId]) @@ -645,7 +645,7 @@ findLoop(DumpableObject *obj, * A user-defined datatype will have a dependency loop with each of its * I/O functions (since those have the datatype as input or output). * Similarly, a range type will have a loop with its canonicalize function, - * if any. Break the loop by making the function depend on the associated + * if any. Break the loop by making the function depend on the associated * shell type, instead. */ static void diff --git a/src/bin/pg_dump/pg_dumpall.c b/src/bin/pg_dump/pg_dumpall.c index 34d6920364432043522e94d6343a0976e5b55537..053e5fd36ae4c0860861368a69a36799bd639437 100644 --- a/src/bin/pg_dump/pg_dumpall.c +++ b/src/bin/pg_dump/pg_dumpall.c @@ -52,8 +52,8 @@ static void doShellQuoting(PQExpBuffer buf, const char *str); static int runPgDump(const char *dbname); static void buildShSecLabels(PGconn *conn, const char *catalog_name, - uint32 objectId, PQExpBuffer buffer, - const char *target, const char *objname); + uint32 objectId, PQExpBuffer buffer, + const char *target, const char *objname); static PGconn *connectDatabase(const char *dbname, const char *pghost, const char *pgport, const char *pguser, enum trivalue prompt_password, bool fail_on_error); static PGresult *executeQuery(PGconn *conn, const char *query); @@ -1663,7 +1663,7 @@ static void buildShSecLabels(PGconn *conn, const char *catalog_name, uint32 objectId, PQExpBuffer buffer, const char *target, const char *objname) { - PQExpBuffer sql = createPQExpBuffer(); + PQExpBuffer sql = createPQExpBuffer(); PGresult *res; buildShSecLabelQuery(conn, catalog_name, objectId, sql); diff --git a/src/bin/pgevent/pgevent.c b/src/bin/pgevent/pgevent.c index 669be05c221366ac5b8b95fc2ae86fbeb14fb205..91d35b4daf14d506066e4097e1c557641ef70ffd 100644 --- a/src/bin/pgevent/pgevent.c +++ b/src/bin/pgevent/pgevent.c @@ -113,7 +113,7 @@ DllRegisterServer(void) "TypesSupported", 0, REG_DWORD, - (LPBYTE) & data, + (LPBYTE) &data, sizeof(DWORD))) { MessageBox(NULL, "Could not set the supported types.", "PostgreSQL error", MB_OK | MB_ICONSTOP); diff --git a/src/bin/psql/command.c b/src/bin/psql/command.c index 8544d15109a3cf2116eb954590f1cce3c71934de..5614120255d12e403eb8b10c8e7cc7fde037fac9 100644 --- a/src/bin/psql/command.c +++ b/src/bin/psql/command.c @@ -777,7 +777,7 @@ exec_command(const char *cmd, /* \i and \ir include files */ else if (strcmp(cmd, "i") == 0 || strcmp(cmd, "include") == 0 - || strcmp(cmd, "ir") == 0 || strcmp(cmd, "include_relative") == 0) + || strcmp(cmd, "ir") == 0 || strcmp(cmd, "include_relative") == 0) { char *fname = psql_scan_slash_option(scan_state, OT_NORMAL, NULL, true); @@ -789,7 +789,7 @@ exec_command(const char *cmd, } else { - bool include_relative; + bool include_relative; include_relative = (strcmp(cmd, "ir") == 0 || strcmp(cmd, "include_relative") == 0); @@ -1103,16 +1103,16 @@ exec_command(const char *cmd, else if (strcmp(cmd, "setenv") == 0) { char *envvar = psql_scan_slash_option(scan_state, - OT_NORMAL, NULL, false); + OT_NORMAL, NULL, false); char *envval = psql_scan_slash_option(scan_state, - OT_NORMAL, NULL, false); + OT_NORMAL, NULL, false); if (!envvar) { psql_error("\\%s: missing required argument\n", cmd); success = false; } - else if (strchr(envvar,'=') != NULL) + else if (strchr(envvar, '=') != NULL) { psql_error("\\%s: environment variable name must not contain \"=\"\n", cmd); @@ -1127,16 +1127,17 @@ exec_command(const char *cmd, else { /* Set variable to the value of the next argument */ - int len = strlen(envvar) + strlen(envval) + 1; + int len = strlen(envvar) + strlen(envval) + 1; char *newval = pg_malloc(len + 1); - snprintf(newval, len+1, "%s=%s", envvar, envval); + snprintf(newval, len + 1, "%s=%s", envvar, envval); putenv(newval); success = true; + /* - * Do not free newval here, it will screw up the environment - * if you do. See putenv man page for details. That means we - * leak a bit of memory here, but not enough to worry about. + * Do not free newval here, it will screw up the environment if + * you do. See putenv man page for details. That means we leak a + * bit of memory here, but not enough to worry about. */ } free(envvar); @@ -2046,9 +2047,9 @@ process_file(char *filename, bool single_txn, bool use_relative_path) /* * If we were asked to resolve the pathname relative to the location - * of the currently executing script, and there is one, and this is - * a relative pathname, then prepend all but the last pathname - * component of the current script to this pathname. + * of the currently executing script, and there is one, and this is a + * relative pathname, then prepend all but the last pathname component + * of the current script to this pathname. */ if (use_relative_path && pset.inputfile && !is_absolute_path(filename) && !has_drive_prefix(filename)) diff --git a/src/bin/psql/common.c b/src/bin/psql/common.c index 33dc97e95f249576165adb28d7c59b556a2ea08b..3691b507a4d168539812daf7097a4b57fd5f96c9 100644 --- a/src/bin/psql/common.c +++ b/src/bin/psql/common.c @@ -707,7 +707,7 @@ ProcessResult(PGresult **results) /* * Call PQgetResult() once more. In the typical case of a - * single-command string, it will return NULL. Otherwise, we'll + * single-command string, it will return NULL. Otherwise, we'll * have other results to process that may include other COPYs. */ PQclear(*results); @@ -982,11 +982,12 @@ SendQuery(const char *query) break; case PQTRANS_INTRANS: + /* * Do nothing if they are messing with savepoints themselves: - * If the user did RELEASE or ROLLBACK, our savepoint is - * gone. If they issued a SAVEPOINT, releasing ours would - * remove theirs. + * If the user did RELEASE or ROLLBACK, our savepoint is gone. + * If they issued a SAVEPOINT, releasing ours would remove + * theirs. */ if (results && (strcmp(PQcmdStatus(results), "SAVEPOINT") == 0 || diff --git a/src/bin/psql/copy.c b/src/bin/psql/copy.c index a1dea9502c27304d15f0a9cf3227da41299cb35f..22fcc5975e5c42ecf3b20f0de586549397f8b4dd 100644 --- a/src/bin/psql/copy.c +++ b/src/bin/psql/copy.c @@ -394,7 +394,7 @@ handleCopyOut(PGconn *conn, FILE *copystream) /* * Check command status and return to normal libpq state. After a * client-side error, the server will remain ready to deliver data. The - * cleanest thing is to fully drain and discard that data. If the + * cleanest thing is to fully drain and discard that data. If the * client-side error happened early in a large file, this takes a long * time. Instead, take advantage of the fact that PQexec() will silently * end any ongoing PGRES_COPY_OUT state. This does cause us to lose the @@ -405,7 +405,7 @@ handleCopyOut(PGconn *conn, FILE *copystream) * We must not ever return with the status still PGRES_COPY_OUT. Our * caller is unable to distinguish that situation from reaching the next * COPY in a command string that happened to contain two consecutive COPY - * TO STDOUT commands. We trust that no condition can make PQexec() fail + * TO STDOUT commands. We trust that no condition can make PQexec() fail * indefinitely while retaining status PGRES_COPY_OUT. */ while (res = PQgetResult(conn), PQresultStatus(res) == PGRES_COPY_OUT) @@ -584,6 +584,7 @@ handleCopyIn(PGconn *conn, FILE *copystream, bool isbinary) OK = false; copyin_cleanup: + /* * Check command status and return to normal libpq state * diff --git a/src/bin/psql/describe.c b/src/bin/psql/describe.c index 2cfacd34e38233cedcf28c55f3fee2a45d5a9dbd..9170dc6982a51642507337de4284b9fa7bb4c39a 100644 --- a/src/bin/psql/describe.c +++ b/src/bin/psql/describe.c @@ -142,15 +142,15 @@ describeTablespaces(const char *pattern, bool verbose) if (pset.sversion >= 90200) printfPQExpBuffer(&buf, "SELECT spcname AS \"%s\",\n" - " pg_catalog.pg_get_userbyid(spcowner) AS \"%s\",\n" - " pg_catalog.pg_tablespace_location(oid) AS \"%s\"", + " pg_catalog.pg_get_userbyid(spcowner) AS \"%s\",\n" + " pg_catalog.pg_tablespace_location(oid) AS \"%s\"", gettext_noop("Name"), gettext_noop("Owner"), gettext_noop("Location")); else printfPQExpBuffer(&buf, "SELECT spcname AS \"%s\",\n" - " pg_catalog.pg_get_userbyid(spcowner) AS \"%s\",\n" + " pg_catalog.pg_get_userbyid(spcowner) AS \"%s\",\n" " spclocation AS \"%s\"", gettext_noop("Name"), gettext_noop("Owner"), @@ -910,7 +910,7 @@ objectDescription(const char *pattern, bool showSystem) if (!showSystem && !pattern) appendPQExpBuffer(&buf, " AND n.nspname <> 'pg_catalog'\n" - " AND n.nspname <> 'information_schema'\n"); + " AND n.nspname <> 'information_schema'\n"); processSQLNamePattern(pset.db, &buf, pattern, true, false, "n.nspname", "o.opcname", NULL, @@ -926,7 +926,7 @@ objectDescription(const char *pattern, bool showSystem) /* Operator family descriptions */ appendPQExpBuffer(&buf, "UNION ALL\n" - " SELECT opf.oid as oid, opf.tableoid as tableoid,\n" + " SELECT opf.oid as oid, opf.tableoid as tableoid,\n" " n.nspname as nspname,\n" " CAST(opf.opfname AS pg_catalog.text) AS name,\n" " CAST('%s' AS pg_catalog.text) as object\n" @@ -939,7 +939,7 @@ objectDescription(const char *pattern, bool showSystem) if (!showSystem && !pattern) appendPQExpBuffer(&buf, " AND n.nspname <> 'pg_catalog'\n" - " AND n.nspname <> 'information_schema'\n"); + " AND n.nspname <> 'information_schema'\n"); processSQLNamePattern(pset.db, &buf, pattern, true, false, "n.nspname", "opf.opfname", NULL, @@ -1294,14 +1294,15 @@ describeOneTableDetails(const char *schemaname, appendPQExpBuffer(&buf, ",\n NULL AS indexdef"); if (tableinfo.relkind == 'f' && pset.sversion >= 90200) appendPQExpBuffer(&buf, ",\n CASE WHEN attfdwoptions IS NULL THEN '' ELSE " - " '(' || array_to_string(ARRAY(SELECT quote_ident(option_name) || ' ' || quote_literal(option_value) FROM " - " pg_options_to_table(attfdwoptions)), ', ') || ')' END AS attfdwoptions"); + " '(' || array_to_string(ARRAY(SELECT quote_ident(option_name) || ' ' || quote_literal(option_value) FROM " + " pg_options_to_table(attfdwoptions)), ', ') || ')' END AS attfdwoptions"); else appendPQExpBuffer(&buf, ",\n NULL AS attfdwoptions"); if (verbose) { appendPQExpBuffer(&buf, ",\n a.attstorage"); appendPQExpBuffer(&buf, ",\n CASE WHEN a.attstattarget=-1 THEN NULL ELSE a.attstattarget END AS attstattarget"); + /* * In 9.0+, we have column comments for: relations, views, composite * types, and foreign tables (c.f. CommentObject() in comment.c). @@ -1416,7 +1417,7 @@ describeOneTableDetails(const char *schemaname, PGresult *result; printfPQExpBuffer(&buf, - "SELECT pg_catalog.pg_get_viewdef('%s'::pg_catalog.oid, true);", + "SELECT pg_catalog.pg_get_viewdef('%s'::pg_catalog.oid, true);", oid); result = PSQLexec(buf.data, false); if (!result) @@ -1651,13 +1652,13 @@ describeOneTableDetails(const char *schemaname, "\n pg_catalog.quote_ident(relname) || '.' ||" "\n pg_catalog.quote_ident(attname)" "\nFROM pg_catalog.pg_class c" - "\nINNER JOIN pg_catalog.pg_depend d ON c.oid=d.refobjid" - "\nINNER JOIN pg_catalog.pg_namespace n ON n.oid=c.relnamespace" + "\nINNER JOIN pg_catalog.pg_depend d ON c.oid=d.refobjid" + "\nINNER JOIN pg_catalog.pg_namespace n ON n.oid=c.relnamespace" "\nINNER JOIN pg_catalog.pg_attribute a ON (" "\n a.attrelid=c.oid AND" "\n a.attnum=d.refobjsubid)" - "\nWHERE d.classid='pg_catalog.pg_class'::pg_catalog.regclass" - "\n AND d.refclassid='pg_catalog.pg_class'::pg_catalog.regclass" + "\nWHERE d.classid='pg_catalog.pg_class'::pg_catalog.regclass" + "\n AND d.refclassid='pg_catalog.pg_class'::pg_catalog.regclass" "\n AND d.objid=%s" "\n AND d.deptype='a'", oid); @@ -1671,10 +1672,11 @@ describeOneTableDetails(const char *schemaname, PQgetvalue(result, 0, 0)); printTableAddFooter(&cont, buf.data); } + /* - * If we get no rows back, don't show anything (obviously). - * We should never get more than one row back, but if we do, - * just ignore it and don't print anything. + * If we get no rows back, don't show anything (obviously). We should + * never get more than one row back, but if we do, just ignore it and + * don't print anything. */ PQclear(result); } @@ -1711,7 +1713,7 @@ describeOneTableDetails(const char *schemaname, " LEFT JOIN pg_catalog.pg_constraint con ON (conrelid = i.indrelid AND conindid = i.indexrelid AND contype IN ('p','u','x'))\n"); appendPQExpBuffer(&buf, "WHERE c.oid = '%s' AND c.oid = i.indrelid AND i.indexrelid = c2.oid\n" - "ORDER BY i.indisprimary DESC, i.indisunique DESC, c2.relname;", + "ORDER BY i.indisprimary DESC, i.indisunique DESC, c2.relname;", oid); result = PSQLexec(buf.data, false); if (!result) @@ -1823,7 +1825,7 @@ describeOneTableDetails(const char *schemaname, "SELECT conname,\n" " pg_catalog.pg_get_constraintdef(r.oid, true) as condef\n" "FROM pg_catalog.pg_constraint r\n" - "WHERE r.conrelid = '%s' AND r.contype = 'f' ORDER BY 1;", + "WHERE r.conrelid = '%s' AND r.contype = 'f' ORDER BY 1;", oid); result = PSQLexec(buf.data, false); if (!result) @@ -1854,7 +1856,7 @@ describeOneTableDetails(const char *schemaname, "SELECT conname, conrelid::pg_catalog.regclass,\n" " pg_catalog.pg_get_constraintdef(c.oid, true) as condef\n" "FROM pg_catalog.pg_constraint c\n" - "WHERE c.confrelid = '%s' AND c.contype = 'f' ORDER BY 1;", + "WHERE c.confrelid = '%s' AND c.contype = 'f' ORDER BY 1;", oid); result = PSQLexec(buf.data, false); if (!result) @@ -2105,7 +2107,7 @@ describeOneTableDetails(const char *schemaname, /* print foreign server name */ if (tableinfo.relkind == 'f') { - char *ftoptions; + char *ftoptions; /* Footer information about foreign table */ printfPQExpBuffer(&buf, @@ -2113,7 +2115,7 @@ describeOneTableDetails(const char *schemaname, " array_to_string(ARRAY(SELECT " " quote_ident(option_name) || ' ' || " " quote_literal(option_value) FROM " - " pg_options_to_table(ftoptions)), ', ') " + " pg_options_to_table(ftoptions)), ', ') " "FROM pg_catalog.pg_foreign_table f,\n" " pg_catalog.pg_foreign_server s\n" "WHERE f.ftrelid = %s AND s.oid = f.ftserver;", @@ -2841,7 +2843,7 @@ listDomains(const char *pattern, bool verbose, bool showSystem) appendPQExpBuffer(&buf, "\nFROM pg_catalog.pg_type t\n" - " LEFT JOIN pg_catalog.pg_namespace n ON n.oid = t.typnamespace\n"); + " LEFT JOIN pg_catalog.pg_namespace n ON n.oid = t.typnamespace\n"); if (verbose) appendPQExpBuffer(&buf, @@ -3769,7 +3771,7 @@ listForeignDataWrappers(const char *pattern, bool verbose) initPQExpBuffer(&buf); printfPQExpBuffer(&buf, "SELECT fdw.fdwname AS \"%s\",\n" - " pg_catalog.pg_get_userbyid(fdw.fdwowner) AS \"%s\",\n", + " pg_catalog.pg_get_userbyid(fdw.fdwowner) AS \"%s\",\n", gettext_noop("Name"), gettext_noop("Owner")); if (pset.sversion >= 90100) diff --git a/src/bin/psql/help.c b/src/bin/psql/help.c index eff0ea53b69052ddec01e25b8f6fcc1e18af7db1..4a37c3414caa294bd737bffc1920297213baa401 100644 --- a/src/bin/psql/help.c +++ b/src/bin/psql/help.c @@ -124,9 +124,9 @@ usage(void) printf(_(" -T, --table-attr=TEXT set HTML table tag attributes (e.g., width, border)\n")); printf(_(" -x, --expanded turn on expanded table output\n")); printf(_(" -z, --field-separator-zero\n" - " set field separator to zero byte\n")); + " set field separator to zero byte\n")); printf(_(" -0, --record-separator-zero\n" - " set record separator to zero byte\n")); + " set record separator to zero byte\n")); printf(_("\nConnection options:\n")); /* Display default host */ @@ -247,7 +247,7 @@ slashUsage(unsigned short int pager) ON(pset.popt.topt.tuples_only)); fprintf(output, _(" \\T [STRING] set HTML <table> tag attributes, or unset if none\n")); fprintf(output, _(" \\x [on|off|auto] toggle expanded output (currently %s)\n"), - pset.popt.topt.expanded == 2 ? "auto" : ON(pset.popt.topt.expanded)); + pset.popt.topt.expanded == 2 ? "auto" : ON(pset.popt.topt.expanded)); fprintf(output, "\n"); fprintf(output, _("Connection\n")); diff --git a/src/bin/psql/input.c b/src/bin/psql/input.c index 880e7e6511d7d47e3bfbd3992265412656aa1003..1a446e2afef45ba037c1f809cd6b14620f8736bf 100644 --- a/src/bin/psql/input.c +++ b/src/bin/psql/input.c @@ -288,7 +288,8 @@ initializeInput(int flags) if (histfile == NULL) { - char * envhist; + char *envhist; + envhist = getenv("PSQL_HISTORY"); if (envhist != NULL && strlen(envhist) > 0) histfile = envhist; diff --git a/src/bin/psql/print.c b/src/bin/psql/print.c index c431f6a437a3a052af4784f7701bc66c8b684feb..8fa5e371284e67c3236d54a8d3329ec83019e06e 100644 --- a/src/bin/psql/print.c +++ b/src/bin/psql/print.c @@ -44,8 +44,8 @@ static char *decimal_point; static char *grouping; static char *thousands_sep; -static char default_footer[100]; -static printTableFooter default_footer_cell = { default_footer, NULL }; +static char default_footer[100]; +static printTableFooter default_footer_cell = {default_footer, NULL}; /* Line style control structures */ const printTextFormat pg_asciiformat = @@ -283,7 +283,7 @@ print_separator(struct separator sep, FILE *fout) /* * Return the list of explicitly-requested footers or, when applicable, the - * default "(xx rows)" footer. Always omit the default footer when given + * default "(xx rows)" footer. Always omit the default footer when given * non-default footers, "\pset footer off", or a specific instruction to that * effect from a calling backslash command. Vertical formats number each row, * making the default footer redundant; they do not call this function. @@ -388,6 +388,7 @@ print_unaligned_text(const printTableContent *cont, FILE *fout) need_recordsep = true; } } + /* * The last record is terminated by a newline, independent of the set * record separator. But when the record separator is a zero byte, we diff --git a/src/bin/psql/print.h b/src/bin/psql/print.h index 25adfc58138852374daf23f82b56177554bdee02..2b2ad0ba4eda535b28debb025695397154ffa764 100644 --- a/src/bin/psql/print.h +++ b/src/bin/psql/print.h @@ -89,7 +89,7 @@ typedef struct printTableOpt unsigned long prior_records; /* start offset for record counters */ const printTextFormat *line_style; /* line style (NULL for default) */ struct separator fieldSep; /* field separator for unaligned text mode */ - struct separator recordSep; /* record separator for unaligned text mode */ + struct separator recordSep; /* record separator for unaligned text mode */ bool numericLocale; /* locale-aware numeric units separator and * decimal marker */ char *tableAttr; /* attributes for HTML <table ...> */ @@ -162,9 +162,9 @@ extern void printTableInit(printTableContent *const content, const printTableOpt *opt, const char *title, const int ncolumns, const int nrows); extern void printTableAddHeader(printTableContent *const content, - char *header, const bool translate, const char align); + char *header, const bool translate, const char align); extern void printTableAddCell(printTableContent *const content, - char *cell, const bool translate, const bool mustfree); + char *cell, const bool translate, const bool mustfree); extern void printTableAddFooter(printTableContent *const content, const char *footer); extern void printTableSetFooter(printTableContent *const content, diff --git a/src/bin/psql/startup.c b/src/bin/psql/startup.c index 1c2a5b3577e245ef41af687962496a93ab7835f8..9a6306b8cf2479039c1b3982a9c018eadbff04b0 100644 --- a/src/bin/psql/startup.c +++ b/src/bin/psql/startup.c @@ -591,7 +591,7 @@ process_psqlrc(char *argv0) char rc_file[MAXPGPATH]; char my_exec_path[MAXPGPATH]; char etc_path[MAXPGPATH]; - char *envrc; + char *envrc; find_my_exec(argv0, my_exec_path); get_etc_path(my_exec_path, etc_path); @@ -600,7 +600,7 @@ process_psqlrc(char *argv0) process_psqlrc_file(rc_file); envrc = getenv("PSQLRC"); - + if (envrc != NULL && strlen(envrc) > 0) { expand_tilde(&envrc); @@ -618,7 +618,8 @@ process_psqlrc(char *argv0) static void process_psqlrc_file(char *filename) { - char *psqlrc_minor, *psqlrc_major; + char *psqlrc_minor, + *psqlrc_major; #if defined(WIN32) && (!defined(__MINGW32__)) #define R_OK 4 diff --git a/src/bin/psql/stringutils.c b/src/bin/psql/stringutils.c index 77387dcf3deaab0bac0b58751e8359279afe4134..b557c5a6bacb96504f053fed5a88eedb140418b4 100644 --- a/src/bin/psql/stringutils.c +++ b/src/bin/psql/stringutils.c @@ -277,7 +277,7 @@ strip_quotes(char *source, char quote, char escape, int encoding) /* * quote_if_needed * - * Opposite of strip_quotes(). If "source" denotes itself literally without + * Opposite of strip_quotes(). If "source" denotes itself literally without * quoting or escaping, returns NULL. Otherwise, returns a malloc'd copy with * quoting and escaping applied: * @@ -303,7 +303,7 @@ quote_if_needed(const char *source, const char *entails_quote, psql_assert(quote); src = source; - dst = ret = pg_malloc(2 * strlen(src) + 3); /* excess */ + dst = ret = pg_malloc(2 * strlen(src) + 3); /* excess */ *dst++ = quote; diff --git a/src/bin/psql/tab-complete.c b/src/bin/psql/tab-complete.c index a50e7356f1d25a8872584c6c437aae60ff9f87e2..061acd13b2c50be852928ce6f2fd181f1a05e16f 100644 --- a/src/bin/psql/tab-complete.c +++ b/src/bin/psql/tab-complete.c @@ -132,7 +132,7 @@ static const char *const * completion_charpp; /* to pass a list of strings */ static const char *completion_info_charp; /* to pass a second string */ static const char *completion_info_charp2; /* to pass a third string */ static const SchemaQuery *completion_squery; /* to pass a SchemaQuery */ -static bool completion_case_sensitive; /* completion is case sensitive */ +static bool completion_case_sensitive; /* completion is case sensitive */ /* * A few macros to ease typing. You can use these to complete the given @@ -790,9 +790,9 @@ psql_completion(char *text, int start, int end) completion_info_charp2 = NULL; /* - * Scan the input line before our current position for the last few - * words. According to those we'll make some smart decisions on what the - * user is probably intending to type. + * Scan the input line before our current position for the last few words. + * According to those we'll make some smart decisions on what the user is + * probably intending to type. */ get_previous_words(start, previous_words, lengthof(previous_words)); @@ -1041,7 +1041,7 @@ psql_completion(char *text, int start, int end) "ENCRYPTED", "INHERIT", "LOGIN", "NOCREATEDB", "NOCREATEROLE", "NOCREATEUSER", "NOINHERIT", "NOLOGIN", "NOREPLICATION", "NOSUPERUSER", "RENAME TO", "REPLICATION", "RESET", "SET", - "SUPERUSER", "UNENCRYPTED", "VALID UNTIL", NULL}; + "SUPERUSER", "UNENCRYPTED", "VALID UNTIL", NULL}; COMPLETE_WITH_LIST(list_ALTERUSER_WITH); } @@ -2017,7 +2017,7 @@ psql_completion(char *text, int start, int end) "ENCRYPTED", "IN", "INHERIT", "LOGIN", "NOCREATEDB", "NOCREATEROLE", "NOCREATEUSER", "NOINHERIT", "NOLOGIN", "NOREPLICATION", "NOSUPERUSER", "REPLICATION", "ROLE", - "SUPERUSER", "SYSID", "UNENCRYPTED", "VALID UNTIL", NULL}; + "SUPERUSER", "SYSID", "UNENCRYPTED", "VALID UNTIL", NULL}; COMPLETE_WITH_LIST(list_CREATEROLE_WITH); } @@ -2317,7 +2317,11 @@ psql_completion(char *text, int start, int end) " UNION SELECT 'USAGE'" " UNION SELECT 'ALL'"); } - /* Complete GRANT/REVOKE <privilege> with "ON", GRANT/REVOKE <role> with TO/FROM */ + + /* + * Complete GRANT/REVOKE <privilege> with "ON", GRANT/REVOKE <role> with + * TO/FROM + */ else if (pg_strcasecmp(prev2_wd, "GRANT") == 0 || pg_strcasecmp(prev2_wd, "REVOKE") == 0) { @@ -2901,8 +2905,11 @@ psql_completion(char *text, int start, int end) COMPLETE_WITH_SCHEMA_QUERY(Query_for_list_of_tables, NULL); /* WITH [RECURSIVE] */ - /* Only match when WITH is the first word, as WITH may appear in many other - contexts. */ + + /* + * Only match when WITH is the first word, as WITH may appear in many + * other contexts. + */ else if (pg_strcasecmp(prev_wd, "WITH") == 0 && prev2_wd[0] == '\0') COMPLETE_WITH_CONST("RECURSIVE"); @@ -3029,7 +3036,7 @@ psql_completion(char *text, int start, int end) strcmp(prev_wd, "\\e") == 0 || strcmp(prev_wd, "\\edit") == 0 || strcmp(prev_wd, "\\g") == 0 || strcmp(prev_wd, "\\i") == 0 || strcmp(prev_wd, "\\include") == 0 || - strcmp(prev_wd, "\\ir") == 0 || strcmp(prev_wd, "\\include_relative") == 0 || + strcmp(prev_wd, "\\ir") == 0 || strcmp(prev_wd, "\\include_relative") == 0 || strcmp(prev_wd, "\\o") == 0 || strcmp(prev_wd, "\\out") == 0 || strcmp(prev_wd, "\\s") == 0 || strcmp(prev_wd, "\\w") == 0 || strcmp(prev_wd, "\\write") == 0 @@ -3412,8 +3419,11 @@ complete_from_list(const char *text, int state) if (completion_case_sensitive) return pg_strdup(item); else - /* If case insensitive matching was requested initially, adjust - * the case according to setting. */ + + /* + * If case insensitive matching was requested initially, + * adjust the case according to setting. + */ return pg_strdup_keyword_case(item, text); } } @@ -3451,8 +3461,11 @@ complete_from_const(const char *text, int state) if (completion_case_sensitive) return pg_strdup(completion_charp); else - /* If case insensitive matching was requested initially, adjust the - * case according to setting. */ + + /* + * If case insensitive matching was requested initially, adjust + * the case according to setting. + */ return pg_strdup_keyword_case(completion_charp, text); } else @@ -3500,7 +3513,7 @@ complete_from_variables(char *text, const char *prefix, const char *suffix) } varnames[nvars] = NULL; - COMPLETE_WITH_LIST_CS((const char * const *) varnames); + COMPLETE_WITH_LIST_CS((const char *const *) varnames); for (i = 0; i < nvars; i++) free(varnames[i]); @@ -3567,9 +3580,10 @@ complete_from_files(const char *text, int state) static char * pg_strdup_keyword_case(const char *s, const char *ref) { - char *ret, *p; + char *ret, + *p; unsigned char first = ref[0]; - int tocase; + int tocase; const char *varval; varval = GetVariable(pset.vars, "COMP_KEYWORD_CASE"); @@ -3635,7 +3649,7 @@ exec_query(const char *query) /* - * Return the nwords word(s) before point. Words are returned right to left, + * Return the nwords word(s) before point. Words are returned right to left, * that is, previous_words[0] gets the last word before point. * If we run out of words, remaining array elements are set to empty strings. * Each array element is filled with a malloc'd string. diff --git a/src/bin/psql/variables.c b/src/bin/psql/variables.c index 33d08176d0231fec596772971fc23f644c960ab2..5e41efc5bdeab5fefc7ed5c72ecca869608eb15a 100644 --- a/src/bin/psql/variables.c +++ b/src/bin/psql/variables.c @@ -15,7 +15,7 @@ * Check whether a variable's name is allowed. * * We allow any non-ASCII character, as well as ASCII letters, digits, and - * underscore. Keep this in sync with the definition of variable_char in + * underscore. Keep this in sync with the definition of variable_char in * psqlscan.l. */ static bool diff --git a/src/bin/scripts/clusterdb.c b/src/bin/scripts/clusterdb.c index 0f711e870b313a565fc4b6a1c387e705ec878db8..76de70b6efc99fdc419315c4928b6e4e2bee5f1b 100644 --- a/src/bin/scripts/clusterdb.c +++ b/src/bin/scripts/clusterdb.c @@ -112,9 +112,9 @@ main(int argc, char *argv[]) } } - /* - * Non-option argument specifies database name - * as long as it wasn't already specified with -d / --dbname + /* + * Non-option argument specifies database name as long as it wasn't + * already specified with -d / --dbname */ if (optind < argc && dbname == NULL) { diff --git a/src/bin/scripts/common.c b/src/bin/scripts/common.c index 5406a98c8387c3af5f81d2c14d923cc42b894b66..0ae708b21ea6f182970a2aebf67976e58529f4af 100644 --- a/src/bin/scripts/common.c +++ b/src/bin/scripts/common.c @@ -185,7 +185,7 @@ connectMaintenanceDatabase(const char *maintenance_db, const char *pghost, enum trivalue prompt_password, const char *progname) { - PGconn *conn; + PGconn *conn; /* If a maintenance database name was specified, just connect to it. */ if (maintenance_db) diff --git a/src/bin/scripts/common.h b/src/bin/scripts/common.h index 229b8dc596776dff69a540b728ecc16904b837be..854bc2f03a2b1016d678dac5dfa3063be59611d5 100644 --- a/src/bin/scripts/common.h +++ b/src/bin/scripts/common.h @@ -10,8 +10,8 @@ #define COMMON_H #include "libpq-fe.h" -#include "getopt_long.h" /* pgrminclude ignore */ -#include "pqexpbuffer.h" /* pgrminclude ignore */ +#include "getopt_long.h" /* pgrminclude ignore */ +#include "pqexpbuffer.h" /* pgrminclude ignore */ enum trivalue { @@ -34,8 +34,8 @@ extern PGconn *connectDatabase(const char *dbname, const char *pghost, bool fail_ok); extern PGconn *connectMaintenanceDatabase(const char *maintenance_db, - const char *pghost, const char *pgport, const char *pguser, - enum trivalue prompt_password, const char *progname); + const char *pghost, const char *pgport, const char *pguser, + enum trivalue prompt_password, const char *progname); extern PGresult *executeQuery(PGconn *conn, const char *query, const char *progname, bool echo); diff --git a/src/bin/scripts/createlang.c b/src/bin/scripts/createlang.c index cc671a49767dd7f0eb7610f5c37209e8141da9da..8268eda6ed314e6177abb97015405a5f0702423b 100644 --- a/src/bin/scripts/createlang.c +++ b/src/bin/scripts/createlang.c @@ -92,10 +92,9 @@ main(int argc, char *argv[]) } /* - * We set dbname from positional arguments if it is not - * already set by option arguments -d. If not doing - * listlangs, positional dbname must follow positional - * langname. + * We set dbname from positional arguments if it is not already set by + * option arguments -d. If not doing listlangs, positional dbname must + * follow positional langname. */ if (argc - optind > 0) diff --git a/src/bin/scripts/dropdb.c b/src/bin/scripts/dropdb.c index 444dce500e3ed5794adc5a7cbed968ad1ec1ef4e..10ed67df8b27d0f250f0a81d50e2c2bd59a4d8eb 100644 --- a/src/bin/scripts/dropdb.c +++ b/src/bin/scripts/dropdb.c @@ -21,7 +21,7 @@ static void help(const char *progname); int main(int argc, char *argv[]) { - static int if_exists = 0; + static int if_exists = 0; static struct option long_options[] = { {"host", required_argument, NULL, 'h'}, @@ -129,7 +129,7 @@ main(int argc, char *argv[]) maintenance_db = "template1"; conn = connectMaintenanceDatabase(maintenance_db, - host, port, username, prompt_password, progname); + host, port, username, prompt_password, progname); if (echo) printf("%s", sql.data); diff --git a/src/bin/scripts/droplang.c b/src/bin/scripts/droplang.c index 47ec37f5e95464bc4519dd900266f08046b04338..74553c19fddc325caef5ce4e136e31c9351c0c8d 100644 --- a/src/bin/scripts/droplang.c +++ b/src/bin/scripts/droplang.c @@ -91,10 +91,9 @@ main(int argc, char *argv[]) } /* - * We set dbname from positional arguments if it is not - * already set by option arguments -d. If not doing - * listlangs, positional dbname must follow positional - * langname. + * We set dbname from positional arguments if it is not already set by + * option arguments -d. If not doing listlangs, positional dbname must + * follow positional langname. */ if (argc - optind > 0) diff --git a/src/bin/scripts/reindexdb.c b/src/bin/scripts/reindexdb.c index 614a6885a0b944b4949d11b8bdd3dd0178dc9410..35254f20cb241bcabc16780f6e58c4818df01df3 100644 --- a/src/bin/scripts/reindexdb.c +++ b/src/bin/scripts/reindexdb.c @@ -122,9 +122,9 @@ main(int argc, char *argv[]) } } - /* - * Non-option argument specifies database name - * as long as it wasn't already specified with -d / --dbname + /* + * Non-option argument specifies database name as long as it wasn't + * already specified with -d / --dbname */ if (optind < argc && dbname == NULL) { diff --git a/src/bin/scripts/vacuumdb.c b/src/bin/scripts/vacuumdb.c index fe303ad8853aa9b3535681ee43eca23d1e96d8c9..6221bdc30d35b290ebc5d298cac77673f275b80d 100644 --- a/src/bin/scripts/vacuumdb.c +++ b/src/bin/scripts/vacuumdb.c @@ -135,10 +135,10 @@ main(int argc, char *argv[]) } } - - /* - * Non-option argument specifies database name - * as long as it wasn't already specified with -d / --dbname + + /* + * Non-option argument specifies database name as long as it wasn't + * already specified with -d / --dbname */ if (optind < argc && dbname == NULL) { @@ -312,7 +312,7 @@ vacuum_all_databases(bool full, bool verbose, bool and_analyze, bool analyze_onl int i; conn = connectMaintenanceDatabase(maintenance_db, host, port, - username, prompt_password, progname); + username, prompt_password, progname); result = executeQuery(conn, "SELECT datname FROM pg_database WHERE datallowconn ORDER BY 1;", progname, echo); PQfinish(conn); diff --git a/src/include/access/gist_private.h b/src/include/access/gist_private.h index 5ad9858c2252f39b7a4314627ac5020a84f38da1..9af9a0cf8c14ca160ffadbeb98c9f582b4530383 100644 --- a/src/include/access/gist_private.h +++ b/src/include/access/gist_private.h @@ -408,7 +408,7 @@ typedef struct GiSTOptions int32 vl_len_; /* varlena header (do not touch directly!) */ int fillfactor; /* page fill factor in percent (0..100) */ int bufferingModeOffset; /* use buffering build? */ -} GiSTOptions; +} GiSTOptions; /* gist.c */ extern Datum gistbuildempty(PG_FUNCTION_ARGS); diff --git a/src/include/access/heapam.h b/src/include/access/heapam.h index d554392e5ae9722e8b77cf46e71cc13fb9f68736..026a19fa741e1d4e1c7ab4e7198903d4eec51d92 100644 --- a/src/include/access/heapam.h +++ b/src/include/access/heapam.h @@ -113,7 +113,7 @@ extern HTSU_Result heap_lock_tuple(Relation relation, HeapTuple tuple, extern void heap_inplace_update(Relation relation, HeapTuple tuple); extern bool heap_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid); extern bool heap_tuple_needs_freeze(HeapTupleHeader tuple, TransactionId cutoff_xid, - Buffer buf); + Buffer buf); extern Oid simple_heap_insert(Relation relation, HeapTuple tup); extern void simple_heap_delete(Relation relation, ItemPointer tid); diff --git a/src/include/access/htup.h b/src/include/access/htup.h index 39213ff849b439ad41bae851b0950e065c149d69..b289e149269e180b7ee2e95893ee65071d0b606b 100644 --- a/src/include/access/htup.h +++ b/src/include/access/htup.h @@ -608,7 +608,7 @@ typedef HeapTupleData *HeapTuple; /* 0x20 is free, was XLOG_HEAP2_CLEAN_MOVE */ #define XLOG_HEAP2_CLEANUP_INFO 0x30 #define XLOG_HEAP2_VISIBLE 0x40 -#define XLOG_HEAP2_MULTI_INSERT 0x50 +#define XLOG_HEAP2_MULTI_INSERT 0x50 /* * All what we need to find changed tuple @@ -671,7 +671,7 @@ typedef struct xl_heap_insert typedef struct xl_heap_multi_insert { RelFileNode node; - BlockNumber blkno; + BlockNumber blkno; bool all_visible_cleared; uint16 ntuples; OffsetNumber offsets[1]; @@ -683,7 +683,7 @@ typedef struct xl_heap_multi_insert typedef struct xl_multi_insert_tuple { - uint16 datalen; /* size of tuple data that follows */ + uint16 datalen; /* size of tuple data that follows */ uint16 t_infomask2; uint16 t_infomask; uint8 t_hoff; diff --git a/src/include/access/nbtree.h b/src/include/access/nbtree.h index cae51a384d4760595a7ad4677ca0369d13b2948a..f23ac3559ad6cc509fd6708cecdb5f64c348e444 100644 --- a/src/include/access/nbtree.h +++ b/src/include/access/nbtree.h @@ -418,12 +418,12 @@ typedef struct xl_btree_newroot /* * When a new operator class is declared, we require that the user * supply us with an amproc procedure (BTORDER_PROC) for determining - * whether, for two keys a and b, a < b, a = b, or a > b. This routine + * whether, for two keys a and b, a < b, a = b, or a > b. This routine * must return < 0, 0, > 0, respectively, in these three cases. (It must * not return INT_MIN, since we may negate the result before using it.) * * To facilitate accelerated sorting, an operator class may choose to - * offer a second procedure (BTSORTSUPPORT_PROC). For full details, see + * offer a second procedure (BTSORTSUPPORT_PROC). For full details, see * src/include/utils/sortsupport.h. */ @@ -551,7 +551,7 @@ typedef struct BTScanOpaqueData int numArrayKeys; /* number of equality-type array keys (-1 if * there are any unsatisfiable array keys) */ BTArrayKeyInfo *arrayKeys; /* info about each equality-type array key */ - MemoryContext arrayContext; /* scan-lifespan context for array data */ + MemoryContext arrayContext; /* scan-lifespan context for array data */ /* info about killed items if any (killedItems is NULL if never used) */ int *killedItems; /* currPos.items indexes of killed items */ @@ -559,8 +559,8 @@ typedef struct BTScanOpaqueData /* * If we are doing an index-only scan, these are the tuple storage - * workspaces for the currPos and markPos respectively. Each is of - * size BLCKSZ, so it can hold as much as a full page's worth of tuples. + * workspaces for the currPos and markPos respectively. Each is of size + * BLCKSZ, so it can hold as much as a full page's worth of tuples. */ char *currTuples; /* tuple storage for currPos */ char *markTuples; /* tuple storage for markPos */ diff --git a/src/include/access/slru.h b/src/include/access/slru.h index 41cd484d808d903bcd9cf7c1c7990e12bc65f6a9..711601ae6260fbf50581533ee70255c2d13aa155 100644 --- a/src/include/access/slru.h +++ b/src/include/access/slru.h @@ -147,7 +147,7 @@ extern void SimpleLruFlush(SlruCtl ctl, bool checkpoint); extern void SimpleLruTruncate(SlruCtl ctl, int cutoffPage); typedef bool (*SlruScanCallback) (SlruCtl ctl, char *filename, int segpage, - void *data); + void *data); extern bool SlruScanDirectory(SlruCtl ctl, SlruScanCallback callback, void *data); /* SlruScanDirectory public callbacks */ diff --git a/src/include/access/spgist.h b/src/include/access/spgist.h index 8d0205e691f2d0335c08738b723fd4504ca3eef7..50cac280a5463fbffb29b653ccb05648ce656358 100644 --- a/src/include/access/spgist.h +++ b/src/include/access/spgist.h @@ -78,25 +78,25 @@ typedef struct spgChooseOut { struct /* results for spgMatchNode */ { - int nodeN; /* descend to this node (index from 0) */ - int levelAdd; /* increment level by this much */ - Datum restDatum; /* new leaf datum */ + int nodeN; /* descend to this node (index from 0) */ + int levelAdd; /* increment level by this much */ + Datum restDatum; /* new leaf datum */ } matchNode; struct /* results for spgAddNode */ { - Datum nodeLabel; /* new node's label */ - int nodeN; /* where to insert it (index from 0) */ + Datum nodeLabel; /* new node's label */ + int nodeN; /* where to insert it (index from 0) */ } addNode; struct /* results for spgSplitTuple */ { /* Info to form new inner tuple with one node */ - bool prefixHasPrefix; /* tuple should have a prefix? */ - Datum prefixPrefixDatum; /* if so, its value */ - Datum nodeLabel; /* node's label */ + bool prefixHasPrefix; /* tuple should have a prefix? */ + Datum prefixPrefixDatum; /* if so, its value */ + Datum nodeLabel; /* node's label */ /* Info to form new lower-level inner tuple with all old nodes */ - bool postfixHasPrefix; /* tuple should have a prefix? */ - Datum postfixPrefixDatum; /* if so, its value */ + bool postfixHasPrefix; /* tuple should have a prefix? */ + Datum postfixPrefixDatum; /* if so, its value */ } splitTuple; } result; } spgChooseOut; @@ -119,7 +119,7 @@ typedef struct spgPickSplitOut int nNodes; /* number of nodes for new inner tuple */ Datum *nodeLabels; /* their labels (or NULL for no labels) */ - int *mapTuplesToNodes; /* node index for each leaf tuple */ + int *mapTuplesToNodes; /* node index for each leaf tuple */ Datum *leafTupleDatums; /* datum to store in each new leaf tuple */ } spgPickSplitOut; diff --git a/src/include/access/spgist_private.h b/src/include/access/spgist_private.h index aa5a6024189413c399d09627e064ec00cf3e811d..74267a439002af2bc58806b0898a02098c2b5465 100644 --- a/src/include/access/spgist_private.h +++ b/src/include/access/spgist_private.h @@ -24,7 +24,7 @@ #define SPGIST_METAPAGE_BLKNO (0) /* metapage */ #define SPGIST_ROOT_BLKNO (1) /* root for normal entries */ #define SPGIST_NULL_BLKNO (2) /* root for null-value entries */ -#define SPGIST_LAST_FIXED_BLKNO SPGIST_NULL_BLKNO +#define SPGIST_LAST_FIXED_BLKNO SPGIST_NULL_BLKNO #define SpGistBlockIsRoot(blkno) \ ((blkno) == SPGIST_ROOT_BLKNO || (blkno) == SPGIST_NULL_BLKNO) @@ -40,7 +40,7 @@ typedef struct SpGistPageOpaqueData uint16 nRedirection; /* number of redirection tuples on page */ uint16 nPlaceholder; /* number of placeholder tuples on page */ /* note there's no count of either LIVE or DEAD tuples ... */ - uint16 spgist_page_id; /* for identification of SP-GiST indexes */ + uint16 spgist_page_id; /* for identification of SP-GiST indexes */ } SpGistPageOpaqueData; typedef SpGistPageOpaqueData *SpGistPageOpaque; @@ -91,7 +91,7 @@ typedef struct SpGistLUPCache typedef struct SpGistMetaPageData { uint32 magicNumber; /* for identity cross-check */ - SpGistLUPCache lastUsedPages; /* shared storage of last-used info */ + SpGistLUPCache lastUsedPages; /* shared storage of last-used info */ } SpGistMetaPageData; #define SPGIST_MAGIC_NUMBER (0xBA0BABEE) @@ -116,11 +116,11 @@ typedef struct SpGistState { spgConfigOut config; /* filled in by opclass config method */ - SpGistTypeDesc attType; /* type of input data and leaf values */ - SpGistTypeDesc attPrefixType; /* type of inner-tuple prefix values */ + SpGistTypeDesc attType; /* type of input data and leaf values */ + SpGistTypeDesc attPrefixType; /* type of inner-tuple prefix values */ SpGistTypeDesc attLabelType; /* type of node label values */ - char *deadTupleStorage; /* workspace for spgFormDeadTuple */ + char *deadTupleStorage; /* workspace for spgFormDeadTuple */ TransactionId myXid; /* XID to use when creating a redirect tuple */ bool isBuild; /* true if doing index build */ @@ -136,7 +136,7 @@ typedef struct SpGistScanOpaqueData /* Control flags showing whether to search nulls and/or non-nulls */ bool searchNulls; /* scan matches (all) null entries */ - bool searchNonNulls; /* scan matches (some) non-null entries */ + bool searchNonNulls; /* scan matches (some) non-null entries */ /* Index quals to be passed to opclass (null-related quals removed) */ int numberOfKeys; /* number of index qualifier conditions */ @@ -154,14 +154,14 @@ typedef struct SpGistScanOpaqueData TupleDesc indexTupDesc; /* if so, tuple descriptor for them */ int nPtrs; /* number of TIDs found on current page */ int iPtr; /* index for scanning through same */ - ItemPointerData heapPtrs[MaxIndexTuplesPerPage]; /* TIDs from cur page */ - bool recheck[MaxIndexTuplesPerPage]; /* their recheck flags */ - IndexTuple indexTups[MaxIndexTuplesPerPage]; /* reconstructed tuples */ + ItemPointerData heapPtrs[MaxIndexTuplesPerPage]; /* TIDs from cur page */ + bool recheck[MaxIndexTuplesPerPage]; /* their recheck flags */ + IndexTuple indexTups[MaxIndexTuplesPerPage]; /* reconstructed tuples */ /* * Note: using MaxIndexTuplesPerPage above is a bit hokey since - * SpGistLeafTuples aren't exactly IndexTuples; however, they are - * larger, so this is safe. + * SpGistLeafTuples aren't exactly IndexTuples; however, they are larger, + * so this is safe. */ } SpGistScanOpaqueData; @@ -175,17 +175,17 @@ typedef struct SpGistCache { spgConfigOut config; /* filled in by opclass config method */ - SpGistTypeDesc attType; /* type of input data and leaf values */ - SpGistTypeDesc attPrefixType; /* type of inner-tuple prefix values */ + SpGistTypeDesc attType; /* type of input data and leaf values */ + SpGistTypeDesc attPrefixType; /* type of inner-tuple prefix values */ SpGistTypeDesc attLabelType; /* type of node label values */ - SpGistLUPCache lastUsedPages; /* local storage of last-used info */ + SpGistLUPCache lastUsedPages; /* local storage of last-used info */ } SpGistCache; /* - * SPGiST tuple types. Note: inner, leaf, and dead tuple structs - * must have the same tupstate field in the same position! Real inner and + * SPGiST tuple types. Note: inner, leaf, and dead tuple structs + * must have the same tupstate field in the same position! Real inner and * leaf tuples always have tupstate = LIVE; if the state is something else, * use the SpGistDeadTuple struct to inspect the tuple. */ @@ -353,7 +353,7 @@ typedef SpGistDeadTupleData *SpGistDeadTuple; * ACCEPT_RDATA_* can only use fixed-length rdata arrays, because of lengthof */ -#define ACCEPT_RDATA_DATA(p, s, i) \ +#define ACCEPT_RDATA_DATA(p, s, i) \ do { \ Assert((i) < lengthof(rdata)); \ rdata[i].data = (char *) (p); \ @@ -387,7 +387,7 @@ typedef SpGistDeadTupleData *SpGistDeadTuple; #define XLOG_SPGIST_PICKSPLIT 0x50 #define XLOG_SPGIST_VACUUM_LEAF 0x60 #define XLOG_SPGIST_VACUUM_ROOT 0x70 -#define XLOG_SPGIST_VACUUM_REDIRECT 0x80 +#define XLOG_SPGIST_VACUUM_REDIRECT 0x80 /* * Some redo functions need an SpGistState, although only a few of its fields @@ -415,7 +415,7 @@ typedef struct spgxlogAddLeaf bool newPage; /* init dest page? */ bool storesNulls; /* page is in the nulls tree? */ OffsetNumber offnumLeaf; /* offset where leaf tuple gets placed */ - OffsetNumber offnumHeadLeaf; /* offset of head tuple in chain, if any */ + OffsetNumber offnumHeadLeaf; /* offset of head tuple in chain, if any */ BlockNumber blknoParent; /* where the parent downlink is, if any */ OffsetNumber offnumParent; @@ -589,7 +589,7 @@ typedef struct spgxlogVacuumRedirect RelFileNode node; BlockNumber blkno; /* block number to clean */ - uint16 nToPlaceholder; /* number of redirects to make placeholders */ + uint16 nToPlaceholder; /* number of redirects to make placeholders */ OffsetNumber firstPlaceholder; /* first placeholder tuple to remove */ /* offsets of redirect tuples to make placeholders follow */ @@ -620,24 +620,24 @@ extern void initSpGistState(SpGistState *state, Relation index); extern Buffer SpGistNewBuffer(Relation index); extern void SpGistUpdateMetaPage(Relation index); extern Buffer SpGistGetBuffer(Relation index, int flags, - int needSpace, bool *isNew); + int needSpace, bool *isNew); extern void SpGistSetLastUsedPage(Relation index, Buffer buffer); extern void SpGistInitPage(Page page, uint16 f); extern void SpGistInitBuffer(Buffer b, uint16 f); extern void SpGistInitMetapage(Page page); extern unsigned int SpGistGetTypeSize(SpGistTypeDesc *att, Datum datum); extern SpGistLeafTuple spgFormLeafTuple(SpGistState *state, - ItemPointer heapPtr, - Datum datum, bool isnull); + ItemPointer heapPtr, + Datum datum, bool isnull); extern SpGistNodeTuple spgFormNodeTuple(SpGistState *state, - Datum label, bool isnull); + Datum label, bool isnull); extern SpGistInnerTuple spgFormInnerTuple(SpGistState *state, - bool hasPrefix, Datum prefix, - int nNodes, SpGistNodeTuple *nodes); + bool hasPrefix, Datum prefix, + int nNodes, SpGistNodeTuple *nodes); extern SpGistDeadTuple spgFormDeadTuple(SpGistState *state, int tupstate, BlockNumber blkno, OffsetNumber offnum); extern Datum *spgExtractNodeLabels(SpGistState *state, - SpGistInnerTuple innerTuple); + SpGistInnerTuple innerTuple); extern OffsetNumber SpGistPageAddNewItem(SpGistState *state, Page page, Item item, Size size, OffsetNumber *startOffset, @@ -645,12 +645,12 @@ extern OffsetNumber SpGistPageAddNewItem(SpGistState *state, Page page, /* spgdoinsert.c */ extern void spgUpdateNodeLink(SpGistInnerTuple tup, int nodeN, - BlockNumber blkno, OffsetNumber offset); + BlockNumber blkno, OffsetNumber offset); extern void spgPageIndexMultiDelete(SpGistState *state, Page page, OffsetNumber *itemnos, int nitems, int firststate, int reststate, BlockNumber blkno, OffsetNumber offnum); extern void spgdoinsert(Relation index, SpGistState *state, - ItemPointer heapPtr, Datum datum, bool isnull); + ItemPointer heapPtr, Datum datum, bool isnull); #endif /* SPGIST_PRIVATE_H */ diff --git a/src/include/access/xact.h b/src/include/access/xact.h index 50f181307ff583f645d50e51574499cc8517ac77..b12d2a0068531d311137ed5b18c8cd17e015cd0f 100644 --- a/src/include/access/xact.h +++ b/src/include/access/xact.h @@ -55,7 +55,8 @@ typedef enum { SYNCHRONOUS_COMMIT_OFF, /* asynchronous commit */ SYNCHRONOUS_COMMIT_LOCAL_FLUSH, /* wait for local flush only */ - SYNCHRONOUS_COMMIT_REMOTE_WRITE, /* wait for local flush and remote write */ + SYNCHRONOUS_COMMIT_REMOTE_WRITE, /* wait for local flush and remote + * write */ SYNCHRONOUS_COMMIT_REMOTE_FLUSH /* wait for local and remote flush */ } SyncCommitLevel; diff --git a/src/include/access/xlog_internal.h b/src/include/access/xlog_internal.h index 2020a3b41fe5d28ea72ba8fbce7cce0a73aefdf7..3328a50faba0064a58a58ad8ff9dfd445f13e5fb 100644 --- a/src/include/access/xlog_internal.h +++ b/src/include/access/xlog_internal.h @@ -157,8 +157,8 @@ typedef XLogLongPageHeaderData *XLogLongPageHeader; #define NextLogPage(recptr) \ do { \ if ((recptr).xrecoff % XLOG_BLCKSZ != 0) \ - (recptr).xrecoff += \ - (XLOG_BLCKSZ - (recptr).xrecoff % XLOG_BLCKSZ); \ + (recptr).xrecoff += \ + (XLOG_BLCKSZ - (recptr).xrecoff % XLOG_BLCKSZ); \ if ((recptr).xrecoff >= XLogFileSize) \ { \ ((recptr).xlogid)++; \ diff --git a/src/include/catalog/catalog.h b/src/include/catalog/catalog.h index 97c79841a61b3db6416ac6f9a8b183e5758bfc5e..678a9452717de715aa4acb6b4be1f8bc676b77de 100644 --- a/src/include/catalog/catalog.h +++ b/src/include/catalog/catalog.h @@ -18,7 +18,7 @@ * 'pgrminclude ignore' needed here because CppAsString2() does not throw * an error if the symbol is not defined. */ -#include "catalog/catversion.h" /* pgrminclude ignore */ +#include "catalog/catversion.h" /* pgrminclude ignore */ #include "catalog/pg_class.h" #include "storage/relfilenode.h" #include "utils/relcache.h" diff --git a/src/include/catalog/genbki.h b/src/include/catalog/genbki.h index bcf31e6c6bdb5a3c448d79de5ffda5b07bfb16c3..f973580e5f9e00034e7649ef39cb517353ae6c14 100644 --- a/src/include/catalog/genbki.h +++ b/src/include/catalog/genbki.h @@ -27,7 +27,7 @@ * * Variable-length catalog fields (except possibly the first not nullable one) * should not be visible in C structures, so they are made invisible by #ifdefs - * of an undefined symbol. See also MARKNOTNULL in bootstrap.c for how this is + * of an undefined symbol. See also MARKNOTNULL in bootstrap.c for how this is * handled. */ #undef CATALOG_VARLEN diff --git a/src/include/catalog/index.h b/src/include/catalog/index.h index 3f73a6c58c28a120dcbf312a7c9e9c38c023d724..7c8198f31ee1519cc087192c5fa57d28196298d5 100644 --- a/src/include/catalog/index.h +++ b/src/include/catalog/index.h @@ -99,6 +99,6 @@ extern bool reindex_relation(Oid relid, int flags); extern bool ReindexIsProcessingHeap(Oid heapOid); extern bool ReindexIsProcessingIndex(Oid indexOid); -extern Oid IndexGetRelation(Oid indexId, bool missing_ok); +extern Oid IndexGetRelation(Oid indexId, bool missing_ok); #endif /* INDEX_H */ diff --git a/src/include/catalog/namespace.h b/src/include/catalog/namespace.h index fa3ba5bd102e6320346a9c9cf54ad8366b4cb737..76215dc8a1e7534d879d9f407378f1c1e75428f0 100644 --- a/src/include/catalog/namespace.h +++ b/src/include/catalog/namespace.h @@ -47,18 +47,18 @@ typedef struct OverrideSearchPath bool addTemp; /* implicitly prepend temp schema? */ } OverrideSearchPath; -typedef void (*RangeVarGetRelidCallback)(const RangeVar *relation, Oid relId, - Oid oldRelId, void *callback_arg); +typedef void (*RangeVarGetRelidCallback) (const RangeVar *relation, Oid relId, + Oid oldRelId, void *callback_arg); #define RangeVarGetRelid(relation, lockmode, missing_ok) \ RangeVarGetRelidExtended(relation, lockmode, missing_ok, false, NULL, NULL) -extern Oid RangeVarGetRelidExtended(const RangeVar *relation, +extern Oid RangeVarGetRelidExtended(const RangeVar *relation, LOCKMODE lockmode, bool missing_ok, bool nowait, RangeVarGetRelidCallback callback, void *callback_arg); extern Oid RangeVarGetCreationNamespace(const RangeVar *newRelation); -extern Oid RangeVarGetAndCheckCreationNamespace(RangeVar *newRelation, +extern Oid RangeVarGetAndCheckCreationNamespace(RangeVar *newRelation, LOCKMODE lockmode, Oid *existing_relation_id); extern void RangeVarAdjustRelationPersistence(RangeVar *newRelation, Oid nspid); diff --git a/src/include/catalog/objectaccess.h b/src/include/catalog/objectaccess.h index a5158e2bc9c1c2be95db3bc26211797c0ca6ee5e..3b40dbc492329b90e449df3386d154253ace570c 100644 --- a/src/include/catalog/objectaccess.h +++ b/src/include/catalog/objectaccess.h @@ -36,10 +36,10 @@ typedef enum ObjectAccessType typedef struct { /* - * Flags to inform extensions the context of this deletion. - * Also see PERFORM_DELETION_* in dependency.h + * Flags to inform extensions the context of this deletion. Also see + * PERFORM_DELETION_* in dependency.h */ - int dropflags; + int dropflags; } ObjectAccessDrop; /* diff --git a/src/include/catalog/objectaddress.h b/src/include/catalog/objectaddress.h index 9c039968856f998080f441903c77fe061d72e55f..0af09c616deefbf3371bd2ca2be3a56ebb0663fb 100644 --- a/src/include/catalog/objectaddress.h +++ b/src/include/catalog/objectaddress.h @@ -28,8 +28,8 @@ typedef struct ObjectAddress } ObjectAddress; extern ObjectAddress get_object_address(ObjectType objtype, List *objname, - List *objargs, Relation *relp, - LOCKMODE lockmode, bool missing_ok); + List *objargs, Relation *relp, + LOCKMODE lockmode, bool missing_ok); extern void check_object_ownership(Oid roleid, ObjectType objtype, ObjectAddress address, diff --git a/src/include/catalog/pg_aggregate.h b/src/include/catalog/pg_aggregate.h index 461772c27b0683bdd126263f38331a8def4f700a..4f44c46f4f220b16bb503074fd48532b8643e0ef 100644 --- a/src/include/catalog/pg_aggregate.h +++ b/src/include/catalog/pg_aggregate.h @@ -44,6 +44,7 @@ CATALOG(pg_aggregate,2600) BKI_WITHOUT_OIDS regproc aggfinalfn; Oid aggsortop; Oid aggtranstype; + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text agginitval; #endif diff --git a/src/include/catalog/pg_attrdef.h b/src/include/catalog/pg_attrdef.h index ad770e410fa17a8fdc61b333502dc3d43929ad96..b92fd1593f9adb7815b5777fb4fd1e84fcf964b9 100644 --- a/src/include/catalog/pg_attrdef.h +++ b/src/include/catalog/pg_attrdef.h @@ -32,6 +32,7 @@ CATALOG(pg_attrdef,2604) { Oid adrelid; /* OID of table containing attribute */ int2 adnum; /* attnum of attribute */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ pg_node_tree adbin; /* nodeToString representation of default */ text adsrc; /* human-readable representation of default */ diff --git a/src/include/catalog/pg_attribute.h b/src/include/catalog/pg_attribute.h index 45e38e4dfc0c26407efe879e4a17c6c920230a0a..4ee1d90ce324ed06b5082811e47d4fccb663be9e 100644 --- a/src/include/catalog/pg_attribute.h +++ b/src/include/catalog/pg_attribute.h @@ -201,7 +201,7 @@ typedef FormData_pg_attribute *Form_pg_attribute; #define Anum_pg_attribute_attcollation 18 #define Anum_pg_attribute_attacl 19 #define Anum_pg_attribute_attoptions 20 -#define Anum_pg_attribute_attfdwoptions 21 +#define Anum_pg_attribute_attfdwoptions 21 /* ---------------- diff --git a/src/include/catalog/pg_constraint.h b/src/include/catalog/pg_constraint.h index 3a77124b00fc780f513d89f6886a5de0dcebfb65..b9e4bf41f93f659a1edcb0d857b587ae117489ad 100644 --- a/src/include/catalog/pg_constraint.h +++ b/src/include/catalog/pg_constraint.h @@ -92,6 +92,7 @@ CATALOG(pg_constraint,2606) bool connoinherit; #ifdef CATALOG_VARLEN /* variable-length fields start here */ + /* * Columns of conrelid that the constraint applies to, if known (this is * NULL for trigger constraints) diff --git a/src/include/catalog/pg_control.h b/src/include/catalog/pg_control.h index 1031e5651274268b8433f0632022f8aa49ef9829..5cff39608bae71caa16e6f99039baa60c7491737 100644 --- a/src/include/catalog/pg_control.h +++ b/src/include/catalog/pg_control.h @@ -33,7 +33,7 @@ typedef struct CheckPoint XLogRecPtr redo; /* next RecPtr available when we began to * create CheckPoint (i.e. REDO start point) */ TimeLineID ThisTimeLineID; /* current TLI */ - bool fullPageWrites; /* current full_page_writes */ + bool fullPageWrites; /* current full_page_writes */ uint32 nextXidEpoch; /* higher-order bits of nextXid */ TransactionId nextXid; /* next free XID */ Oid nextOid; /* next free OID */ @@ -140,11 +140,11 @@ typedef struct ControlFileData * record, to make sure the end-of-backup record corresponds the base * backup we're recovering from. * - * backupEndPoint is the backup end location, if we are recovering from - * an online backup which was taken from the standby and haven't reached - * the end of backup yet. It is initialized to the minimum recovery point - * in pg_control which was backed up last. It is reset to zero when - * the end of backup is reached, and we mustn't start up before that. + * backupEndPoint is the backup end location, if we are recovering from an + * online backup which was taken from the standby and haven't reached the + * end of backup yet. It is initialized to the minimum recovery point in + * pg_control which was backed up last. It is reset to zero when the end + * of backup is reached, and we mustn't start up before that. * * If backupEndRequired is true, we know for sure that we're restoring * from a backup, and must see a backup-end record before we can safely diff --git a/src/include/catalog/pg_database.h b/src/include/catalog/pg_database.h index e8509f59bbbfd13cd4e1c045838d4086e7a14279..af803bba89b238ec504f086eaf435228f3cd2d85 100644 --- a/src/include/catalog/pg_database.h +++ b/src/include/catalog/pg_database.h @@ -42,6 +42,7 @@ CATALOG(pg_database,1262) BKI_SHARED_RELATION BKI_ROWTYPE_OID(1248) BKI_SCHEMA_M Oid datlastsysoid; /* highest OID to consider a system OID */ TransactionId datfrozenxid; /* all Xids < this are frozen in this DB */ Oid dattablespace; /* default table space for this DB */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem datacl[1]; /* access permissions */ #endif diff --git a/src/include/catalog/pg_db_role_setting.h b/src/include/catalog/pg_db_role_setting.h index c6e2f3b4dca5cd60dd50fb3489b90179dbf72f79..c6a69c5a6dace6c162285692747ff26a0e0815b6 100644 --- a/src/include/catalog/pg_db_role_setting.h +++ b/src/include/catalog/pg_db_role_setting.h @@ -35,6 +35,7 @@ CATALOG(pg_db_role_setting,2964) BKI_SHARED_RELATION BKI_WITHOUT_OIDS { Oid setdatabase; /* database */ Oid setrole; /* role */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text setconfig[1]; /* GUC settings to apply at login */ #endif diff --git a/src/include/catalog/pg_default_acl.h b/src/include/catalog/pg_default_acl.h index f36ce52519b74074939db20e3dd61ed1b1b59120..d7421007af6eb9af808d439c6d32834a5a2b901e 100644 --- a/src/include/catalog/pg_default_acl.h +++ b/src/include/catalog/pg_default_acl.h @@ -32,6 +32,7 @@ CATALOG(pg_default_acl,826) Oid defaclrole; /* OID of role owning this ACL */ Oid defaclnamespace; /* OID of namespace, or 0 for all */ char defaclobjtype; /* see DEFACLOBJ_xxx constants below */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem defaclacl[1]; /* permissions to add at CREATE time */ #endif diff --git a/src/include/catalog/pg_description.h b/src/include/catalog/pg_description.h index 352c51736982125d382c60c5b10052aad72abfe0..a454194893055f176a6197f784bb8f594f31c94d 100644 --- a/src/include/catalog/pg_description.h +++ b/src/include/catalog/pg_description.h @@ -50,6 +50,7 @@ CATALOG(pg_description,2609) BKI_WITHOUT_OIDS Oid objoid; /* OID of object itself */ Oid classoid; /* OID of table containing object */ int4 objsubid; /* column number, or 0 if not used */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text description; /* description of object */ #endif diff --git a/src/include/catalog/pg_extension.h b/src/include/catalog/pg_extension.h index b07b43cf7e65ae484a541e1a4797ea08f786297b..4807c6a7338668a44412bf7dd27f67e7832b9397 100644 --- a/src/include/catalog/pg_extension.h +++ b/src/include/catalog/pg_extension.h @@ -34,7 +34,8 @@ CATALOG(pg_extension,3079) Oid extowner; /* extension owner */ Oid extnamespace; /* namespace of contained objects */ bool extrelocatable; /* if true, allow ALTER EXTENSION SET SCHEMA */ -#ifdef CATALOG_VARLEN /* variable-length fields start here */ + +#ifdef CATALOG_VARLEN /* variable-length fields start here */ /* extversion should never be null, but the others can be. */ text extversion; /* extension version name */ Oid extconfig[1]; /* dumpable configuration tables */ diff --git a/src/include/catalog/pg_foreign_data_wrapper.h b/src/include/catalog/pg_foreign_data_wrapper.h index b6dd8eb7cc81fa20519abfb24843e2c0799eff12..18c538f49c036b1943902a1d5976d57294148b6f 100644 --- a/src/include/catalog/pg_foreign_data_wrapper.h +++ b/src/include/catalog/pg_foreign_data_wrapper.h @@ -34,6 +34,7 @@ CATALOG(pg_foreign_data_wrapper,2328) Oid fdwowner; /* FDW owner */ Oid fdwhandler; /* handler function, or 0 if none */ Oid fdwvalidator; /* option validation function, or 0 if none */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem fdwacl[1]; /* access permissions */ text fdwoptions[1]; /* FDW options */ diff --git a/src/include/catalog/pg_foreign_server.h b/src/include/catalog/pg_foreign_server.h index dd1e65e02aa65b62d1388f4850e32c98d7004403..38830af29eb1ca694b360f8d21629b4d371e48fb 100644 --- a/src/include/catalog/pg_foreign_server.h +++ b/src/include/catalog/pg_foreign_server.h @@ -31,6 +31,7 @@ CATALOG(pg_foreign_server,1417) NameData srvname; /* foreign server name */ Oid srvowner; /* server owner */ Oid srvfdw; /* server FDW */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text srvtype; text srvversion; diff --git a/src/include/catalog/pg_foreign_table.h b/src/include/catalog/pg_foreign_table.h index 9af983eb93bb4404e77a6545d3cfc032849f15f1..186c49d4254bbb25fcd3e0b3bbd63823825ac801 100644 --- a/src/include/catalog/pg_foreign_table.h +++ b/src/include/catalog/pg_foreign_table.h @@ -30,6 +30,7 @@ CATALOG(pg_foreign_table,3118) BKI_WITHOUT_OIDS { Oid ftrelid; /* OID of foreign table */ Oid ftserver; /* OID of foreign server */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text ftoptions[1]; /* FDW-specific options */ #endif diff --git a/src/include/catalog/pg_index.h b/src/include/catalog/pg_index.h index 92ca22c686ad98f80577295306342263a6115fc2..9a86121cf9e6dc6693e8cd20f86e5b120b989986 100644 --- a/src/include/catalog/pg_index.h +++ b/src/include/catalog/pg_index.h @@ -44,6 +44,7 @@ CATALOG(pg_index,2610) BKI_WITHOUT_OIDS BKI_SCHEMA_MACRO /* variable-length fields start here, but we allow direct access to indkey */ int2vector indkey; /* column numbers of indexed cols, or 0 */ + #ifdef CATALOG_VARLEN oidvector indcollation; /* collation identifiers */ oidvector indclass; /* opclass identifiers */ diff --git a/src/include/catalog/pg_language.h b/src/include/catalog/pg_language.h index eb4ae5ab2d32a766823f0566423fc7b361cea39a..79cfa09d02c488c80ffe6186888afbd3abd5e17f 100644 --- a/src/include/catalog/pg_language.h +++ b/src/include/catalog/pg_language.h @@ -37,6 +37,7 @@ CATALOG(pg_language,2612) Oid lanplcallfoid; /* Call handler for PL */ Oid laninline; /* Optional anonymous-block handler function */ Oid lanvalidator; /* Optional validation function */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem lanacl[1]; /* Access privileges */ #endif diff --git a/src/include/catalog/pg_largeobject.h b/src/include/catalog/pg_largeobject.h index b89d4ec1a90e0dd38fb63a334105fea394887d45..d442ec4e4a2d364e7119d53b2cbc90e69eccf210 100644 --- a/src/include/catalog/pg_largeobject.h +++ b/src/include/catalog/pg_largeobject.h @@ -32,6 +32,7 @@ CATALOG(pg_largeobject,2613) BKI_WITHOUT_OIDS { Oid loid; /* Identifier of large object */ int4 pageno; /* Page number (starting from 0) */ + /* data has variable length, but we allow direct access; see inv_api.c */ bytea data; /* Data for page (may be zero-length) */ } FormData_pg_largeobject; diff --git a/src/include/catalog/pg_largeobject_metadata.h b/src/include/catalog/pg_largeobject_metadata.h index c280176491963981994c5e5a963598bf764d415d..768497eb5256e34889940ea717c2f0e338e579f2 100644 --- a/src/include/catalog/pg_largeobject_metadata.h +++ b/src/include/catalog/pg_largeobject_metadata.h @@ -31,6 +31,7 @@ CATALOG(pg_largeobject_metadata,2995) { Oid lomowner; /* OID of the largeobject owner */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem lomacl[1]; /* access permissions */ #endif diff --git a/src/include/catalog/pg_namespace.h b/src/include/catalog/pg_namespace.h index 1daba477b409a453bfbe19885fd57455c56245cc..e25392127842afc5de27201a22063c5bb0853644 100644 --- a/src/include/catalog/pg_namespace.h +++ b/src/include/catalog/pg_namespace.h @@ -37,6 +37,7 @@ CATALOG(pg_namespace,2615) { NameData nspname; Oid nspowner; + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem nspacl[1]; #endif diff --git a/src/include/catalog/pg_opclass.h b/src/include/catalog/pg_opclass.h index 96eaa600b3f219b15c7ef13fb249c5d9d3f9d5be..638f8088c7be4e1f6fa30b514da2b416279c8f68 100644 --- a/src/include/catalog/pg_opclass.h +++ b/src/include/catalog/pg_opclass.h @@ -134,7 +134,7 @@ DATA(insert ( 405 macaddr_ops PGNSP PGUID 1985 829 t 0 )); */ DATA(insert ( 403 name_ops PGNSP PGUID 1986 19 t 2275 )); DATA(insert ( 405 name_ops PGNSP PGUID 1987 19 t 0 )); -DATA(insert OID = 3125 ( 403 numeric_ops PGNSP PGUID 1988 1700 t 0 )); +DATA(insert OID = 3125 ( 403 numeric_ops PGNSP PGUID 1988 1700 t 0 )); #define NUMERIC_BTREE_OPS_OID 3125 DATA(insert ( 405 numeric_ops PGNSP PGUID 1998 1700 t 0 )); DATA(insert OID = 1981 ( 403 oid_ops PGNSP PGUID 1989 26 t 0 )); @@ -148,7 +148,7 @@ DATA(insert OID = 3126 ( 403 text_ops PGNSP PGUID 1994 25 t 0 )); DATA(insert ( 405 text_ops PGNSP PGUID 1995 25 t 0 )); DATA(insert ( 403 time_ops PGNSP PGUID 1996 1083 t 0 )); DATA(insert ( 405 time_ops PGNSP PGUID 1997 1083 t 0 )); -DATA(insert OID = 3127 ( 403 timestamptz_ops PGNSP PGUID 434 1184 t 0 )); +DATA(insert OID = 3127 ( 403 timestamptz_ops PGNSP PGUID 434 1184 t 0 )); #define TIMESTAMPTZ_BTREE_OPS_OID 3127 DATA(insert ( 405 timestamptz_ops PGNSP PGUID 1999 1184 t 0 )); DATA(insert ( 403 timetz_ops PGNSP PGUID 2000 1266 t 0 )); diff --git a/src/include/catalog/pg_operator.h b/src/include/catalog/pg_operator.h index 48ddd16a94d8cc0afe2288d62d68b937b4ee1317..94702541f55642862367cb505ef1b919fd5851e0 100644 --- a/src/include/catalog/pg_operator.h +++ b/src/include/catalog/pg_operator.h @@ -1702,7 +1702,7 @@ DATA(insert OID = 3895 ( "&<" PGNSP PGUID b f f 3831 3831 16 0 0 range_overl DESCR("overlaps or is left of"); DATA(insert OID = 3896 ( "&>" PGNSP PGUID b f f 3831 3831 16 0 0 range_overright scalargtsel scalargtjoinsel )); DESCR("overlaps or is right of"); -DATA(insert OID = 3897 ( "-|-" PGNSP PGUID b f f 3831 3831 16 3897 0 range_adjacent contsel contjoinsel )); +DATA(insert OID = 3897 ( "-|-" PGNSP PGUID b f f 3831 3831 16 3897 0 range_adjacent contsel contjoinsel )); DESCR("is adjacent to"); DATA(insert OID = 3898 ( "+" PGNSP PGUID b f f 3831 3831 3831 3898 0 range_union - - )); DESCR("range union"); diff --git a/src/include/catalog/pg_pltemplate.h b/src/include/catalog/pg_pltemplate.h index 00abd533708666b5aba7cec099f3c49d07dcd795..d8927adcbe24d9e716ad8c39971f306cc61e05d5 100644 --- a/src/include/catalog/pg_pltemplate.h +++ b/src/include/catalog/pg_pltemplate.h @@ -33,6 +33,7 @@ CATALOG(pg_pltemplate,1136) BKI_SHARED_RELATION BKI_WITHOUT_OIDS NameData tmplname; /* name of PL */ bool tmpltrusted; /* PL is trusted? */ bool tmpldbacreate; /* PL is installable by db owner? */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text tmplhandler; /* name of call handler function */ text tmplinline; /* name of anonymous-block handler, or NULL */ diff --git a/src/include/catalog/pg_proc.h b/src/include/catalog/pg_proc.h index 34b77f01919e9259576cc7efa03a93f269b89973..1e097ddbe6c420abcc89409925347fb643d3c152 100644 --- a/src/include/catalog/pg_proc.h +++ b/src/include/catalog/pg_proc.h @@ -54,8 +54,12 @@ CATALOG(pg_proc,1255) BKI_BOOTSTRAP BKI_ROWTYPE_OID(81) BKI_SCHEMA_MACRO int2 pronargdefaults; /* number of arguments with defaults */ Oid prorettype; /* OID of result type */ - /* variable-length fields start here, but we allow direct access to proargtypes */ + /* + * variable-length fields start here, but we allow direct access to + * proargtypes + */ oidvector proargtypes; /* parameter types (excludes OUT params) */ + #ifdef CATALOG_VARLEN Oid proallargtypes[1]; /* all param types (NULL if IN only) */ char proargmodes[1]; /* parameter modes (NULL if IN only) */ @@ -2664,7 +2668,7 @@ DATA(insert OID = 3151 ( pg_stat_get_db_temp_bytes PGNSP PGUID 12 1 0 0 0 f f f DESCR("statistics: number of bytes in temporary files written"); DATA(insert OID = 2844 ( pg_stat_get_db_blk_read_time PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 701 "26" _null_ _null_ _null_ _null_ pg_stat_get_db_blk_read_time _null_ _null_ _null_ )); DESCR("statistics: block read time, in msec"); -DATA(insert OID = 2845 ( pg_stat_get_db_blk_write_time PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 701 "26" _null_ _null_ _null_ _null_ pg_stat_get_db_blk_write_time _null_ _null_ _null_ )); +DATA(insert OID = 2845 ( pg_stat_get_db_blk_write_time PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 701 "26" _null_ _null_ _null_ _null_ pg_stat_get_db_blk_write_time _null_ _null_ _null_ )); DESCR("statistics: block write time, in msec"); DATA(insert OID = 2769 ( pg_stat_get_bgwriter_timed_checkpoints PGNSP PGUID 12 1 0 0 0 f f f f t f s 0 0 20 "" _null_ _null_ _null_ _null_ pg_stat_get_bgwriter_timed_checkpoints _null_ _null_ _null_ )); DESCR("statistics: number of timed checkpoints started by the bgwriter"); @@ -2904,7 +2908,7 @@ DATA(insert OID = 2082 ( pg_operator_is_visible PGNSP PGUID 12 1 0 0 0 f f f f DESCR("is operator visible in search path?"); DATA(insert OID = 2083 ( pg_opclass_is_visible PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 16 "26" _null_ _null_ _null_ _null_ pg_opclass_is_visible _null_ _null_ _null_ )); DESCR("is opclass visible in search path?"); -DATA(insert OID = 3829 ( pg_opfamily_is_visible PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 16 "26" _null_ _null_ _null_ _null_ pg_opfamily_is_visible _null_ _null_ _null_ )); +DATA(insert OID = 3829 ( pg_opfamily_is_visible PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 16 "26" _null_ _null_ _null_ _null_ pg_opfamily_is_visible _null_ _null_ _null_ )); DESCR("is opfamily visible in search path?"); DATA(insert OID = 2093 ( pg_conversion_is_visible PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 16 "26" _null_ _null_ _null_ _null_ pg_conversion_is_visible _null_ _null_ _null_ )); DESCR("is conversion visible in search path?"); @@ -4067,9 +4071,9 @@ DATA(insert OID = 323 ( json_recv PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 DESCR("I/O"); DATA(insert OID = 324 ( json_send PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 17 "114" _null_ _null_ _null_ _null_ json_send _null_ _null_ _null_ )); DESCR("I/O"); -DATA(insert OID = 3153 ( array_to_json PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 114 "2277" _null_ _null_ _null_ _null_ array_to_json _null_ _null_ _null_ )); +DATA(insert OID = 3153 ( array_to_json PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 114 "2277" _null_ _null_ _null_ _null_ array_to_json _null_ _null_ _null_ )); DESCR("map array to json"); -DATA(insert OID = 3154 ( array_to_json PGNSP PGUID 12 1 0 0 0 f f f f t f s 2 0 114 "2277 16" _null_ _null_ _null_ _null_ array_to_json_pretty _null_ _null_ _null_ )); +DATA(insert OID = 3154 ( array_to_json PGNSP PGUID 12 1 0 0 0 f f f f t f s 2 0 114 "2277 16" _null_ _null_ _null_ _null_ array_to_json_pretty _null_ _null_ _null_ )); DESCR("map array to json with optional pretty printing"); DATA(insert OID = 3155 ( row_to_json PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 114 "2249" _null_ _null_ _null_ _null_ row_to_json _null_ _null_ _null_ )); DESCR("map row to json"); @@ -4466,13 +4470,13 @@ DATA(insert OID = 3849 ( upper PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 2283 " DESCR("upper bound of range"); DATA(insert OID = 3850 ( isempty PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_empty _null_ _null_ _null_ )); DESCR("is the range empty?"); -DATA(insert OID = 3851 ( lower_inc PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_lower_inc _null_ _null_ _null_ )); +DATA(insert OID = 3851 ( lower_inc PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_lower_inc _null_ _null_ _null_ )); DESCR("is the range's lower bound inclusive?"); -DATA(insert OID = 3852 ( upper_inc PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_upper_inc _null_ _null_ _null_ )); +DATA(insert OID = 3852 ( upper_inc PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_upper_inc _null_ _null_ _null_ )); DESCR("is the range's upper bound inclusive?"); -DATA(insert OID = 3853 ( lower_inf PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_lower_inf _null_ _null_ _null_ )); +DATA(insert OID = 3853 ( lower_inf PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_lower_inf _null_ _null_ _null_ )); DESCR("is the range's lower bound infinite?"); -DATA(insert OID = 3854 ( upper_inf PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_upper_inf _null_ _null_ _null_ )); +DATA(insert OID = 3854 ( upper_inf PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 16 "3831" _null_ _null_ _null_ _null_ range_upper_inf _null_ _null_ _null_ )); DESCR("is the range's upper bound infinite?"); DATA(insert OID = 3855 ( range_eq PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "3831 3831" _null_ _null_ _null_ _null_ range_eq _null_ _null_ _null_ )); DESCR("implementation of = operator"); @@ -4504,19 +4508,19 @@ DATA(insert OID = 3868 ( range_intersect PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 DESCR("implementation of * operator"); DATA(insert OID = 3869 ( range_minus PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 3831 "3831 3831" _null_ _null_ _null_ _null_ range_minus _null_ _null_ _null_ )); DESCR("implementation of - operator"); -DATA(insert OID = 3870 ( range_cmp PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 23 "3831 3831" _null_ _null_ _null_ _null_ range_cmp _null_ _null_ _null_ )); +DATA(insert OID = 3870 ( range_cmp PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 23 "3831 3831" _null_ _null_ _null_ _null_ range_cmp _null_ _null_ _null_ )); DESCR("less-equal-greater"); DATA(insert OID = 3871 ( range_lt PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "3831 3831" _null_ _null_ _null_ _null_ range_lt _null_ _null_ _null_ )); DATA(insert OID = 3872 ( range_le PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "3831 3831" _null_ _null_ _null_ _null_ range_le _null_ _null_ _null_ )); DATA(insert OID = 3873 ( range_ge PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "3831 3831" _null_ _null_ _null_ _null_ range_ge _null_ _null_ _null_ )); DATA(insert OID = 3874 ( range_gt PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "3831 3831" _null_ _null_ _null_ _null_ range_gt _null_ _null_ _null_ )); -DATA(insert OID = 3875 ( range_gist_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 5 0 16 "2281 3831 23 26 2281" _null_ _null_ _null_ _null_ range_gist_consistent _null_ _null_ _null_ )); +DATA(insert OID = 3875 ( range_gist_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 5 0 16 "2281 3831 23 26 2281" _null_ _null_ _null_ _null_ range_gist_consistent _null_ _null_ _null_ )); DESCR("GiST support"); DATA(insert OID = 3876 ( range_gist_union PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2281 "2281 2281" _null_ _null_ _null_ _null_ range_gist_union _null_ _null_ _null_ )); DESCR("GiST support"); DATA(insert OID = 3877 ( range_gist_compress PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 2281 "2281" _null_ _null_ _null_ _null_ range_gist_compress _null_ _null_ _null_ )); DESCR("GiST support"); -DATA(insert OID = 3878 ( range_gist_decompress PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 2281 "2281" _null_ _null_ _null_ _null_ range_gist_decompress _null_ _null_ _null_ )); +DATA(insert OID = 3878 ( range_gist_decompress PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 2281 "2281" _null_ _null_ _null_ _null_ range_gist_decompress _null_ _null_ _null_ )); DESCR("GiST support"); DATA(insert OID = 3879 ( range_gist_penalty PGNSP PGUID 12 1 0 0 0 f f f f t f i 3 0 2281 "2281 2281 2281" _null_ _null_ _null_ _null_ range_gist_penalty _null_ _null_ _null_ )); DESCR("GiST support"); @@ -4524,7 +4528,7 @@ DATA(insert OID = 3880 ( range_gist_picksplit PGNSP PGUID 12 1 0 0 0 f f f f t DESCR("GiST support"); DATA(insert OID = 3881 ( range_gist_same PGNSP PGUID 12 1 0 0 0 f f f f t f i 3 0 2281 "3831 3831 2281" _null_ _null_ _null_ _null_ range_gist_same _null_ _null_ _null_ )); DESCR("GiST support"); -DATA(insert OID = 3902 ( hash_range PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 23 "3831" _null_ _null_ _null_ _null_ hash_range _null_ _null_ _null_ )); +DATA(insert OID = 3902 ( hash_range PGNSP PGUID 12 1 0 0 0 f f f f t f i 1 0 23 "3831" _null_ _null_ _null_ _null_ hash_range _null_ _null_ _null_ )); DESCR("hash a range"); DATA(insert OID = 3916 ( range_typanalyze PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 16 "2281" _null_ _null_ _null_ _null_ range_typanalyze _null_ _null_ _null_ )); DESCR("range typanalyze"); @@ -4548,9 +4552,9 @@ DESCR("float8 difference of two timestamp values"); DATA(insert OID = 3930 ( tstzrange_subdiff PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 701 "1184 1184" _null_ _null_ _null_ _null_ tstzrange_subdiff _null_ _null_ _null_ )); DESCR("float8 difference of two timestamp with time zone values"); -DATA(insert OID = 3840 ( int4range PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3904 "23 23" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); +DATA(insert OID = 3840 ( int4range PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3904 "23 23" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); DESCR("int4range constructor"); -DATA(insert OID = 3841 ( int4range PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3904 "23 23 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); +DATA(insert OID = 3841 ( int4range PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3904 "23 23 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); DESCR("int4range constructor"); DATA(insert OID = 3844 ( numrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3906 "1700 1700" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); DESCR("numrange constructor"); @@ -4560,17 +4564,17 @@ DATA(insert OID = 3933 ( tsrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3908 DESCR("tsrange constructor"); DATA(insert OID = 3934 ( tsrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3908 "1114 1114 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); DESCR("tsrange constructor"); -DATA(insert OID = 3937 ( tstzrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3910 "1184 1184" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); +DATA(insert OID = 3937 ( tstzrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3910 "1184 1184" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); DESCR("tstzrange constructor"); -DATA(insert OID = 3938 ( tstzrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3910 "1184 1184 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); +DATA(insert OID = 3938 ( tstzrange PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3910 "1184 1184 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); DESCR("tstzrange constructor"); -DATA(insert OID = 3941 ( daterange PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3912 "1082 1082" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); +DATA(insert OID = 3941 ( daterange PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3912 "1082 1082" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); DESCR("daterange constructor"); -DATA(insert OID = 3942 ( daterange PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3912 "1082 1082 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); +DATA(insert OID = 3942 ( daterange PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3912 "1082 1082 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); DESCR("daterange constructor"); -DATA(insert OID = 3945 ( int8range PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3926 "20 20" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); +DATA(insert OID = 3945 ( int8range PGNSP PGUID 12 1 0 0 0 f f f f f f i 2 0 3926 "20 20" _null_ _null_ _null_ _null_ range_constructor2 _null_ _null_ _null_ )); DESCR("int8range constructor"); -DATA(insert OID = 3946 ( int8range PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3926 "20 20 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); +DATA(insert OID = 3946 ( int8range PGNSP PGUID 12 1 0 0 0 f f f f f f i 3 0 3926 "20 20 25" _null_ _null_ _null_ _null_ range_constructor3 _null_ _null_ _null_ )); DESCR("int8range constructor"); /* spgist support functions */ @@ -4596,7 +4600,7 @@ DATA(insert OID = 4010 ( spgbuildempty PGNSP PGUID 12 1 0 0 0 f f f f t f v DESCR("spgist(internal)"); DATA(insert OID = 4011 ( spgbulkdelete PGNSP PGUID 12 1 0 0 0 f f f f t f v 4 0 2281 "2281 2281 2281 2281" _null_ _null_ _null_ _null_ spgbulkdelete _null_ _null_ _null_ )); DESCR("spgist(internal)"); -DATA(insert OID = 4012 ( spgvacuumcleanup PGNSP PGUID 12 1 0 0 0 f f f f t f v 2 0 2281 "2281 2281" _null_ _null_ _null_ _null_ spgvacuumcleanup _null_ _null_ _null_ )); +DATA(insert OID = 4012 ( spgvacuumcleanup PGNSP PGUID 12 1 0 0 0 f f f f t f v 2 0 2281 "2281 2281" _null_ _null_ _null_ _null_ spgvacuumcleanup _null_ _null_ _null_ )); DESCR("spgist(internal)"); DATA(insert OID = 4032 ( spgcanreturn PGNSP PGUID 12 1 0 0 0 f f f f t f s 1 0 16 "2281" _null_ _null_ _null_ _null_ spgcanreturn _null_ _null_ _null_ )); DESCR("spgist(internal)"); @@ -4612,14 +4616,14 @@ DATA(insert OID = 4019 ( spg_quad_choose PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 DESCR("SP-GiST support for quad tree over point"); DATA(insert OID = 4020 ( spg_quad_picksplit PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_quad_picksplit _null_ _null_ _null_ )); DESCR("SP-GiST support for quad tree over point"); -DATA(insert OID = 4021 ( spg_quad_inner_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_quad_inner_consistent _null_ _null_ _null_ )); +DATA(insert OID = 4021 ( spg_quad_inner_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_quad_inner_consistent _null_ _null_ _null_ )); DESCR("SP-GiST support for quad tree over point"); DATA(insert OID = 4022 ( spg_quad_leaf_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "2281 2281" _null_ _null_ _null_ _null_ spg_quad_leaf_consistent _null_ _null_ _null_ )); DESCR("SP-GiST support for quad tree and k-d tree over point"); -DATA(insert OID = 4023 ( spg_kd_config PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_kd_config _null_ _null_ _null_ )); +DATA(insert OID = 4023 ( spg_kd_config PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_kd_config _null_ _null_ _null_ )); DESCR("SP-GiST support for k-d tree over point"); -DATA(insert OID = 4024 ( spg_kd_choose PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_kd_choose _null_ _null_ _null_ )); +DATA(insert OID = 4024 ( spg_kd_choose PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_kd_choose _null_ _null_ _null_ )); DESCR("SP-GiST support for k-d tree over point"); DATA(insert OID = 4025 ( spg_kd_picksplit PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_kd_picksplit _null_ _null_ _null_ )); DESCR("SP-GiST support for k-d tree over point"); @@ -4632,7 +4636,7 @@ DATA(insert OID = 4028 ( spg_text_choose PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 DESCR("SP-GiST support for suffix tree over text"); DATA(insert OID = 4029 ( spg_text_picksplit PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_text_picksplit _null_ _null_ _null_ )); DESCR("SP-GiST support for suffix tree over text"); -DATA(insert OID = 4030 ( spg_text_inner_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_text_inner_consistent _null_ _null_ _null_ )); +DATA(insert OID = 4030 ( spg_text_inner_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 2278 "2281 2281" _null_ _null_ _null_ _null_ spg_text_inner_consistent _null_ _null_ _null_ )); DESCR("SP-GiST support for suffix tree over text"); DATA(insert OID = 4031 ( spg_text_leaf_consistent PGNSP PGUID 12 1 0 0 0 f f f f t f i 2 0 16 "2281 2281" _null_ _null_ _null_ _null_ spg_text_leaf_consistent _null_ _null_ _null_ )); DESCR("SP-GiST support for suffix tree over text"); @@ -4662,4 +4666,3 @@ DESCR("SP-GiST support for suffix tree over text"); #define PROARGMODE_TABLE 't' #endif /* PG_PROC_H */ - diff --git a/src/include/catalog/pg_range.h b/src/include/catalog/pg_range.h index d66e83a017a4fa31ccbb0e9e6a25b66089a973f3..9b2be92fc93bec2181e59e19e19da6bdb1cd655e 100644 --- a/src/include/catalog/pg_range.h +++ b/src/include/catalog/pg_range.h @@ -29,7 +29,7 @@ * typedef struct FormData_pg_range * ---------------- */ -#define RangeRelationId 3541 +#define RangeRelationId 3541 CATALOG(pg_range,3541) BKI_WITHOUT_OIDS { @@ -65,12 +65,12 @@ typedef FormData_pg_range *Form_pg_range; * initial contents of pg_range * ---------------- */ -DATA(insert ( 3904 23 0 1978 int4range_canonical int4range_subdiff)); +DATA(insert ( 3904 23 0 1978 int4range_canonical int4range_subdiff)); DATA(insert ( 3906 1700 0 3125 - numrange_subdiff)); DATA(insert ( 3908 1114 0 3128 - tsrange_subdiff)); DATA(insert ( 3910 1184 0 3127 - tstzrange_subdiff)); DATA(insert ( 3912 1082 0 3122 daterange_canonical daterange_subdiff)); -DATA(insert ( 3926 20 0 3124 int8range_canonical int8range_subdiff)); +DATA(insert ( 3926 20 0 3124 int8range_canonical int8range_subdiff)); /* diff --git a/src/include/catalog/pg_rewrite.h b/src/include/catalog/pg_rewrite.h index e04ba81b3c2cccd931d74ca505db05a878043cd5..5171522da080511a9209d30eaf60ac8fb0e35f03 100644 --- a/src/include/catalog/pg_rewrite.h +++ b/src/include/catalog/pg_rewrite.h @@ -39,6 +39,7 @@ CATALOG(pg_rewrite,2618) char ev_type; char ev_enabled; bool is_instead; + #ifdef CATALOG_VARLEN /* variable-length fields start here */ pg_node_tree ev_qual; pg_node_tree ev_action; diff --git a/src/include/catalog/pg_seclabel.h b/src/include/catalog/pg_seclabel.h index 101ec3c111b45f8899ef36d23a38007d2d6b8a93..917efcf8ca9126f952600a6940edaca71212c7a5 100644 --- a/src/include/catalog/pg_seclabel.h +++ b/src/include/catalog/pg_seclabel.h @@ -25,6 +25,7 @@ CATALOG(pg_seclabel,3596) BKI_WITHOUT_OIDS Oid objoid; /* OID of the object itself */ Oid classoid; /* OID of table containing the object */ int4 objsubid; /* column number, or 0 if not used */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text provider; /* name of label provider */ text label; /* security label of the object */ diff --git a/src/include/catalog/pg_shdescription.h b/src/include/catalog/pg_shdescription.h index 377f4337d9a3810ec7ee1e67ed9a81cf92221ce3..acd529b875c671fbb81ece439ecd729d9eda5384 100644 --- a/src/include/catalog/pg_shdescription.h +++ b/src/include/catalog/pg_shdescription.h @@ -42,6 +42,7 @@ CATALOG(pg_shdescription,2396) BKI_SHARED_RELATION BKI_WITHOUT_OIDS { Oid objoid; /* OID of object itself */ Oid classoid; /* OID of table containing object */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text description; /* description of object */ #endif diff --git a/src/include/catalog/pg_shseclabel.h b/src/include/catalog/pg_shseclabel.h index d7c49e78bb5c1a38e5e03f93a5e1240e8ada20ac..3d7a013e8c72f6bd4df45679b2d68ed124d0e1d5 100644 --- a/src/include/catalog/pg_shseclabel.h +++ b/src/include/catalog/pg_shseclabel.h @@ -1,7 +1,7 @@ /* ------------------------------------------------------------------------- * * pg_shseclabel.h - * definition of the system "security label" relation (pg_shseclabel) + * definition of the system "security label" relation (pg_shseclabel) * * Portions Copyright (c) 1996-2012, PostgreSQL Global Development Group * Portions Copyright (c) 1994, Regents of the University of California @@ -22,11 +22,12 @@ CATALOG(pg_shseclabel,3592) BKI_SHARED_RELATION BKI_WITHOUT_OIDS { - Oid objoid; /* OID of the shared object itself */ - Oid classoid; /* OID of table containing the shared object */ -#ifdef CATALOG_VARLEN /* variable-length fields start here */ - text provider; /* name of label provider */ - text label; /* security label of the object */ + Oid objoid; /* OID of the shared object itself */ + Oid classoid; /* OID of table containing the shared object */ + +#ifdef CATALOG_VARLEN /* variable-length fields start here */ + text provider; /* name of label provider */ + text label; /* security label of the object */ #endif } FormData_pg_shseclabel; @@ -40,4 +41,4 @@ CATALOG(pg_shseclabel,3592) BKI_SHARED_RELATION BKI_WITHOUT_OIDS #define Anum_pg_shseclabel_provider 3 #define Anum_pg_shseclabel_label 4 -#endif /* PG_SHSECLABEL_H */ +#endif /* PG_SHSECLABEL_H */ diff --git a/src/include/catalog/pg_statistic.h b/src/include/catalog/pg_statistic.h index 383cc014159e366ef704fd710e568f3bc907e5ea..3ad0c281102f8071fc97620ee04a6b811282277a 100644 --- a/src/include/catalog/pg_statistic.h +++ b/src/include/catalog/pg_statistic.h @@ -105,7 +105,7 @@ CATALOG(pg_statistic,2619) BKI_WITHOUT_OIDS /* * Values in these arrays are values of the column's data type, or of some - * related type such as an array element type. We presently have to cheat + * related type such as an array element type. We presently have to cheat * quite a bit to allow polymorphic arrays of this kind, but perhaps * someday it'll be a less bogus facility. */ @@ -258,7 +258,7 @@ typedef FormData_pg_statistic *Form_pg_statistic; /* * A "distinct elements count histogram" slot describes the distribution of * the number of distinct element values present in each row of an array-type - * column. Only non-null rows are considered, and only non-null elements. + * column. Only non-null rows are considered, and only non-null elements. * staop contains the equality operator appropriate to the element type. * stavalues is not used and should be NULL. The last member of stanumbers is * the average count of distinct element values over all non-null rows. The @@ -266,6 +266,6 @@ typedef FormData_pg_statistic *Form_pg_statistic; * distinct-elements counts into M-1 bins of approximately equal population. * The first of these is the minimum observed count, and the last the maximum. */ -#define STATISTIC_KIND_DECHIST 5 +#define STATISTIC_KIND_DECHIST 5 #endif /* PG_STATISTIC_H */ diff --git a/src/include/catalog/pg_tablespace.h b/src/include/catalog/pg_tablespace.h index 0650a5fbae49a1c83241905270b8a06b0cdbb3cb..777a8a1778764fae7bef65f30d7e925e759f081c 100644 --- a/src/include/catalog/pg_tablespace.h +++ b/src/include/catalog/pg_tablespace.h @@ -32,6 +32,7 @@ CATALOG(pg_tablespace,1213) BKI_SHARED_RELATION { NameData spcname; /* tablespace name */ Oid spcowner; /* owner of tablespace */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ aclitem spcacl[1]; /* access permissions */ text spcoptions[1]; /* per-tablespace options */ diff --git a/src/include/catalog/pg_trigger.h b/src/include/catalog/pg_trigger.h index 0ee5b8af8d1b83aab5b3e6e95b09a44b663f32dd..71afab58e56a4396b0d1de9b113562a03103602d 100644 --- a/src/include/catalog/pg_trigger.h +++ b/src/include/catalog/pg_trigger.h @@ -50,9 +50,12 @@ CATALOG(pg_trigger,2620) bool tginitdeferred; /* constraint trigger is deferred initially */ int2 tgnargs; /* # of extra arguments in tgargs */ - /* Variable-length fields start here, but we allow direct access to tgattr. - * Note: tgattr and tgargs must not be null. */ + /* + * Variable-length fields start here, but we allow direct access to + * tgattr. Note: tgattr and tgargs must not be null. + */ int2vector tgattr; /* column numbers, if trigger is on columns */ + #ifdef CATALOG_VARLEN bytea tgargs; /* first\000second\000tgnargs\000 */ pg_node_tree tgqual; /* WHEN expression, or NULL if none */ diff --git a/src/include/catalog/pg_ts_dict.h b/src/include/catalog/pg_ts_dict.h index 31fcdd8896488ed6b63ff552ceb8b394fc48f3bc..677a870b77a28974ab6359a4582340aa079406b2 100644 --- a/src/include/catalog/pg_ts_dict.h +++ b/src/include/catalog/pg_ts_dict.h @@ -36,6 +36,7 @@ CATALOG(pg_ts_dict,3600) Oid dictnamespace; /* name space */ Oid dictowner; /* owner */ Oid dicttemplate; /* dictionary's template */ + #ifdef CATALOG_VARLEN /* variable-length fields start here */ text dictinitoption; /* options passed to dict_init() */ #endif diff --git a/src/include/catalog/pg_type.h b/src/include/catalog/pg_type.h index 507819da80466df95c9d680a45961daaf7f28080..25c664b7c90ca3abb6730c91cdc7d984ee8bee67 100644 --- a/src/include/catalog/pg_type.h +++ b/src/include/catalog/pg_type.h @@ -61,9 +61,8 @@ CATALOG(pg_type,1247) BKI_BOOTSTRAP BKI_ROWTYPE_OID(71) BKI_SCHEMA_MACRO /* * typtype is 'b' for a base type, 'c' for a composite type (e.g., a - * table's rowtype), 'd' for a domain, 'e' for an enum type, - * 'p' for a pseudo-type, or 'r' for a range type. - * (Use the TYPTYPE macros below.) + * table's rowtype), 'd' for a domain, 'e' for an enum type, 'p' for a + * pseudo-type, or 'r' for a range type. (Use the TYPTYPE macros below.) * * If typtype is 'c', typrelid is the OID of the class' entry in pg_class. */ @@ -201,6 +200,7 @@ CATALOG(pg_type,1247) BKI_BOOTSTRAP BKI_ROWTYPE_OID(71) BKI_SCHEMA_MACRO Oid typcollation; #ifdef CATALOG_VARLEN /* variable-length fields start here */ + /* * If typdefaultbin is not NULL, it is the nodeToString representation of * a default expression for the type. Currently this is only used for diff --git a/src/include/commands/createas.h b/src/include/commands/createas.h index ed65ccd8ee3fde3f2198bec39df25387853d3f11..946c7e2cf1e2b7e67eab14e8be8d4c64f60a5750 100644 --- a/src/include/commands/createas.h +++ b/src/include/commands/createas.h @@ -20,7 +20,7 @@ extern void ExecCreateTableAs(CreateTableAsStmt *stmt, const char *queryString, - ParamListInfo params, char *completionTag); + ParamListInfo params, char *completionTag); extern int GetIntoRelEFlags(IntoClause *intoClause); diff --git a/src/include/commands/defrem.h b/src/include/commands/defrem.h index 163b2ea002b7f4312c446413fb4e8e097aa9b2e0..8f3d2c358dceb73587a934284fb6340b91d91cb4 100644 --- a/src/include/commands/defrem.h +++ b/src/include/commands/defrem.h @@ -149,7 +149,7 @@ extern List *deserialize_deflist(Datum txt); extern void RenameForeignServer(const char *oldname, const char *newname); extern void RenameForeignDataWrapper(const char *oldname, const char *newname); extern void AlterForeignServerOwner(const char *name, Oid newOwnerId); -extern void AlterForeignServerOwner_oid(Oid , Oid newOwnerId); +extern void AlterForeignServerOwner_oid(Oid, Oid newOwnerId); extern void AlterForeignDataWrapperOwner(const char *name, Oid newOwnerId); extern void AlterForeignDataWrapperOwner_oid(Oid fwdId, Oid newOwnerId); extern void CreateForeignDataWrapper(CreateFdwStmt *stmt); diff --git a/src/include/commands/explain.h b/src/include/commands/explain.h index e4e98bfb0435ad05b9fc2f515f5d0ad93bcd1479..cd1d56d6e8d0e57394779179ce863925def3300d 100644 --- a/src/include/commands/explain.h +++ b/src/include/commands/explain.h @@ -42,7 +42,7 @@ typedef struct ExplainState /* Hook for plugins to get control in ExplainOneQuery() */ typedef void (*ExplainOneQuery_hook_type) (Query *query, - IntoClause *into, + IntoClause *into, ExplainState *es, const char *queryString, ParamListInfo params); diff --git a/src/include/commands/tablecmds.h b/src/include/commands/tablecmds.h index 47b0cddc9bb8634c1a0f655b4db695676dd7f984..9ceb086f681420be255ae8790988fb074b6495ff 100644 --- a/src/include/commands/tablecmds.h +++ b/src/include/commands/tablecmds.h @@ -75,6 +75,6 @@ extern void AtEOSubXact_on_commit_actions(bool isCommit, SubTransactionId parentSubid); extern void RangeVarCallbackOwnsTable(const RangeVar *relation, - Oid relId, Oid oldRelId, void *arg); + Oid relId, Oid oldRelId, void *arg); #endif /* TABLECMDS_H */ diff --git a/src/include/commands/typecmds.h b/src/include/commands/typecmds.h index bb4a7c32bc2b4d5693c6f8db1ca2ed8ecfc87689..b72cfc4fd94457f2a6419472330fcdcee0a068c2 100644 --- a/src/include/commands/typecmds.h +++ b/src/include/commands/typecmds.h @@ -34,7 +34,7 @@ extern void AlterDomainNotNull(List *names, bool notNull); extern void AlterDomainAddConstraint(List *names, Node *constr); extern void AlterDomainValidateConstraint(List *names, char *constrName); extern void AlterDomainDropConstraint(List *names, const char *constrName, - DropBehavior behavior, bool missing_ok); + DropBehavior behavior, bool missing_ok); extern void checkDomainOwner(HeapTuple tup); diff --git a/src/include/commands/vacuum.h b/src/include/commands/vacuum.h index 3c95dec3bc61ad0da5486fe1310212d045d021a2..7a50d2fcb306bb9a5070b8f85f6e49d58e2f5ee7 100644 --- a/src/include/commands/vacuum.h +++ b/src/include/commands/vacuum.h @@ -62,9 +62,9 @@ typedef Datum (*AnalyzeAttrFetchFunc) (VacAttrStatsP stats, int rownum, bool *isNull); typedef void (*AnalyzeAttrComputeStatsFunc) (VacAttrStatsP stats, - AnalyzeAttrFetchFunc fetchfunc, - int samplerows, - double totalrows); + AnalyzeAttrFetchFunc fetchfunc, + int samplerows, + double totalrows); typedef struct VacAttrStats { diff --git a/src/include/datatype/timestamp.h b/src/include/datatype/timestamp.h index 0583d45fad17e582f62df66bf88b13daa0be5945..706b4480b3aaef5486a0974744f463148fa1bf7b 100644 --- a/src/include/datatype/timestamp.h +++ b/src/include/datatype/timestamp.h @@ -109,10 +109,10 @@ typedef struct * We allow numeric timezone offsets up to 15:59:59 either way from Greenwich. * Currently, the record holders for wackiest offsets in actual use are zones * Asia/Manila, at -15:56:00 until 1844, and America/Metlakatla, at +15:13:42 - * until 1867. If we were to reject such values we would fail to dump and + * until 1867. If we were to reject such values we would fail to dump and * restore old timestamptz values with these zone settings. */ -#define MAX_TZDISP_HOUR 15 /* maximum allowed hour part */ +#define MAX_TZDISP_HOUR 15 /* maximum allowed hour part */ #define TZDISP_LIMIT ((MAX_TZDISP_HOUR + 1) * SECS_PER_HOUR) /* @@ -121,7 +121,7 @@ typedef struct #ifdef HAVE_INT64_TIMESTAMP #define DT_NOBEGIN (-INT64CONST(0x7fffffffffffffff) - 1) #define DT_NOEND (INT64CONST(0x7fffffffffffffff)) -#else /* !HAVE_INT64_TIMESTAMP */ +#else /* !HAVE_INT64_TIMESTAMP */ #ifdef HUGE_VAL #define DT_NOBEGIN (-HUGE_VAL) #define DT_NOEND (HUGE_VAL) @@ -164,7 +164,7 @@ typedef struct || ((m) == JULIAN_MINMONTH && (d) >= JULIAN_MINDAY)))) \ && (y) < JULIAN_MAXYEAR) -#define JULIAN_MAX (2147483494) /* == date2j(JULIAN_MAXYEAR, 1, 1) */ +#define JULIAN_MAX (2147483494) /* == date2j(JULIAN_MAXYEAR, 1, 1) */ /* Julian-date equivalents of Day 0 in Unix and Postgres reckoning */ #define UNIX_EPOCH_JDATE 2440588 /* == date2j(1970, 1, 1) */ diff --git a/src/include/executor/executor.h b/src/include/executor/executor.h index f5503a566341550b5d49b3c31eb0a3bb7aee9513..075bbe8b57d8819a37bb2f95e9444b0da75bddfe 100644 --- a/src/include/executor/executor.h +++ b/src/include/executor/executor.h @@ -51,7 +51,7 @@ * is responsible for there being a trigger context for them to be queued in. * * WITH/WITHOUT_OIDS tell the executor to emit tuples with or without space - * for OIDs, respectively. These are currently used only for CREATE TABLE AS. + * for OIDs, respectively. These are currently used only for CREATE TABLE AS. * If neither is set, the plan may or may not produce tuples including OIDs. */ #define EXEC_FLAG_EXPLAIN_ONLY 0x0001 /* EXPLAIN, no ANALYZE */ diff --git a/src/include/executor/instrument.h b/src/include/executor/instrument.h index fe64369527e12bc8817bf375ef7678c0eb046b55..e6dd03c2d77166bc617c8212858f95be5c5784b4 100644 --- a/src/include/executor/instrument.h +++ b/src/include/executor/instrument.h @@ -18,18 +18,18 @@ typedef struct BufferUsage { - long shared_blks_hit; /* # of shared buffer hits */ + long shared_blks_hit; /* # of shared buffer hits */ long shared_blks_read; /* # of shared disk blocks read */ long shared_blks_dirtied; /* # of shared blocks dirtied */ long shared_blks_written; /* # of shared disk blocks written */ - long local_blks_hit; /* # of local buffer hits */ - long local_blks_read; /* # of local disk blocks read */ + long local_blks_hit; /* # of local buffer hits */ + long local_blks_read; /* # of local disk blocks read */ long local_blks_dirtied; /* # of shared blocks dirtied */ long local_blks_written; /* # of local disk blocks written */ - long temp_blks_read; /* # of temp blocks read */ + long temp_blks_read; /* # of temp blocks read */ long temp_blks_written; /* # of temp blocks written */ - instr_time blk_read_time; /* time spent reading */ - instr_time blk_write_time; /* time spent writing */ + instr_time blk_read_time; /* time spent reading */ + instr_time blk_write_time; /* time spent writing */ } BufferUsage; /* Flag bits included in InstrAlloc's instrument_options bitmask */ @@ -44,7 +44,7 @@ typedef enum InstrumentOption typedef struct Instrumentation { /* Parameters set at node creation: */ - bool need_timer; /* TRUE if we need timer data */ + bool need_timer; /* TRUE if we need timer data */ bool need_bufusage; /* TRUE if we need buffer usage data */ /* Info about current plan cycle: */ bool running; /* TRUE if we've completed first tuple */ diff --git a/src/include/executor/spi_priv.h b/src/include/executor/spi_priv.h index 900cebb00078df5c614a00e25cd4e02d1cea2ba1..4fbb548af4b93d360b6cd48f8f03993bd383594f 100644 --- a/src/include/executor/spi_priv.h +++ b/src/include/executor/spi_priv.h @@ -48,7 +48,7 @@ typedef struct * adequate locks to prevent other backends from messing with the tables. * * For a saved plan, the plancxt is made a child of CacheMemoryContext - * since it should persist until explicitly destroyed. Likewise, the + * since it should persist until explicitly destroyed. Likewise, the * plancache entries will be under CacheMemoryContext since we tell * plancache.c to save them. We rely on plancache.c to keep the cache * entries up-to-date as needed in the face of invalidation events. diff --git a/src/include/foreign/fdwapi.h b/src/include/foreign/fdwapi.h index 0a09c94932692171f78b14bfd0a80316bf4e53be..721cd254368075cbc5317a993e9fd3238b102e23 100644 --- a/src/include/foreign/fdwapi.h +++ b/src/include/foreign/fdwapi.h @@ -24,19 +24,19 @@ struct ExplainState; */ typedef void (*GetForeignRelSize_function) (PlannerInfo *root, - RelOptInfo *baserel, - Oid foreigntableid); + RelOptInfo *baserel, + Oid foreigntableid); typedef void (*GetForeignPaths_function) (PlannerInfo *root, - RelOptInfo *baserel, - Oid foreigntableid); + RelOptInfo *baserel, + Oid foreigntableid); typedef ForeignScan *(*GetForeignPlan_function) (PlannerInfo *root, - RelOptInfo *baserel, - Oid foreigntableid, - ForeignPath *best_path, - List *tlist, - List *scan_clauses); + RelOptInfo *baserel, + Oid foreigntableid, + ForeignPath *best_path, + List *tlist, + List *scan_clauses); typedef void (*ExplainForeignScan_function) (ForeignScanState *node, struct ExplainState *es); @@ -51,13 +51,13 @@ typedef void (*ReScanForeignScan_function) (ForeignScanState *node); typedef void (*EndForeignScan_function) (ForeignScanState *node); typedef int (*AcquireSampleRowsFunc) (Relation relation, int elevel, - HeapTuple *rows, int targrows, - double *totalrows, - double *totaldeadrows); + HeapTuple *rows, int targrows, + double *totalrows, + double *totaldeadrows); typedef bool (*AnalyzeForeignTable_function) (Relation relation, - AcquireSampleRowsFunc *func, - BlockNumber *totalpages); + AcquireSampleRowsFunc *func, + BlockNumber *totalpages); /* * FdwRoutine is the struct returned by a foreign-data wrapper's handler @@ -86,8 +86,8 @@ typedef struct FdwRoutine EndForeignScan_function EndForeignScan; /* - * These functions are optional. Set the pointer to NULL for any - * that are not provided. + * These functions are optional. Set the pointer to NULL for any that are + * not provided. */ AnalyzeForeignTable_function AnalyzeForeignTable; } FdwRoutine; diff --git a/src/include/lib/stringinfo.h b/src/include/lib/stringinfo.h index 8e3a7afb47b53f05020e2f935df6453ffb62a505..dbf9277abe763f4d71ad6d0ea40301ed5e0b95aa 100644 --- a/src/include/lib/stringinfo.h +++ b/src/include/lib/stringinfo.h @@ -105,7 +105,8 @@ __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); * without modifying str. Typically the caller would enlarge str and retry * on false return --- see appendStringInfo for standard usage pattern. */ -extern bool appendStringInfoVA(StringInfo str, const char *fmt, va_list args) +extern bool +appendStringInfoVA(StringInfo str, const char *fmt, va_list args) __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 0))); /*------------------------ diff --git a/src/include/libpq/hba.h b/src/include/libpq/hba.h index c5a77c2a5af66371f8578e19df992e3ea4dc5201..f3b8be6a0ccd27ef3bf9427a3c3ef3d9e5ae200c 100644 --- a/src/include/libpq/hba.h +++ b/src/include/libpq/hba.h @@ -11,7 +11,7 @@ #ifndef HBA_H #define HBA_H -#include "libpq/pqcomm.h" /* pgrminclude ignore */ /* needed for NetBSD */ +#include "libpq/pqcomm.h" /* pgrminclude ignore */ /* needed for NetBSD */ #include "nodes/pg_list.h" diff --git a/src/include/libpq/ip.h b/src/include/libpq/ip.h index a81234df45a4ef16708ee397244a1782383dcaf9..0ea57461fc17ae110f6fd0bcdc9983d3a52ea8df 100644 --- a/src/include/libpq/ip.h +++ b/src/include/libpq/ip.h @@ -15,8 +15,8 @@ #ifndef IP_H #define IP_H -#include "getaddrinfo.h" /* pgrminclude ignore */ -#include "libpq/pqcomm.h" /* pgrminclude ignore */ +#include "getaddrinfo.h" /* pgrminclude ignore */ +#include "libpq/pqcomm.h" /* pgrminclude ignore */ #ifdef HAVE_UNIX_SOCKETS diff --git a/src/include/nodes/execnodes.h b/src/include/nodes/execnodes.h index b48a03b4b259043fbca8777421280823012962d1..6fe8c2303a21850e9cbf2aa7247a2e82cc7d7911 100644 --- a/src/include/nodes/execnodes.h +++ b/src/include/nodes/execnodes.h @@ -1100,7 +1100,7 @@ typedef struct MergeAppendState PlanState **mergeplans; /* array of PlanStates for my inputs */ int ms_nplans; int ms_nkeys; - SortSupport ms_sortkeys; /* array of length ms_nkeys */ + SortSupport ms_sortkeys; /* array of length ms_nkeys */ TupleTableSlot **ms_slots; /* array of length ms_nplans */ int *ms_heap; /* array of length ms_nplans */ int ms_heap_size; /* current active length of ms_heap[] */ diff --git a/src/include/nodes/parsenodes.h b/src/include/nodes/parsenodes.h index 13b95e11aaf0121120ec8189bc868cbf4ecd4ee6..deff1a374c05fb799ea054039861f253728513fe 100644 --- a/src/include/nodes/parsenodes.h +++ b/src/include/nodes/parsenodes.h @@ -706,7 +706,7 @@ typedef struct RangeTblEntry * Fields valid for a subquery RTE (else NULL): */ Query *subquery; /* the sub-query */ - bool security_barrier; /* subquery from security_barrier view */ + bool security_barrier; /* subquery from security_barrier view */ /* * Fields valid for a join RTE (else NULL/zero): @@ -1171,7 +1171,7 @@ typedef struct AlterTableStmt RangeVar *relation; /* table to work on */ List *cmds; /* list of subcommands */ ObjectType relkind; /* type of object */ - bool missing_ok; /* skip error if table missing */ + bool missing_ok; /* skip error if table missing */ } AlterTableStmt; typedef enum AlterTableType @@ -1193,14 +1193,14 @@ typedef enum AlterTableType AT_AddConstraint, /* add constraint */ AT_AddConstraintRecurse, /* internal to commands/tablecmds.c */ AT_ValidateConstraint, /* validate constraint */ - AT_ValidateConstraintRecurse, /* internal to commands/tablecmds.c */ + AT_ValidateConstraintRecurse, /* internal to commands/tablecmds.c */ AT_ProcessedConstraint, /* pre-processed add constraint (local in * parser/parse_utilcmd.c) */ AT_AddIndexConstraint, /* add constraint using existing index */ AT_DropConstraint, /* drop constraint */ AT_DropConstraintRecurse, /* internal to commands/tablecmds.c */ AT_AlterColumnType, /* alter column type */ - AT_AlterColumnGenericOptions, /* alter column OPTIONS (...) */ + AT_AlterColumnGenericOptions, /* alter column OPTIONS (...) */ AT_ChangeOwner, /* change owner */ AT_ClusterOn, /* CLUSTER ON */ AT_DropCluster, /* SET WITHOUT CLUSTER */ @@ -1477,7 +1477,7 @@ typedef struct CreateStmt * * If skip_validation is true then we skip checking that the existing rows * in the table satisfy the constraint, and just install the catalog entries - * for the constraint. A new FK constraint is marked as valid iff + * for the constraint. A new FK constraint is marked as valid iff * initially_valid is true. (Usually skip_validation and initially_valid * are inverses, but we can set both true if the table is known empty.) * @@ -1967,7 +1967,7 @@ typedef struct SecLabelStmt #define CURSOR_OPT_HOLD 0x0010 /* WITH HOLD */ /* these planner-control flags do not correspond to any SQL grammar: */ #define CURSOR_OPT_FAST_PLAN 0x0020 /* prefer fast-start plan */ -#define CURSOR_OPT_GENERIC_PLAN 0x0040 /* force use of generic plan */ +#define CURSOR_OPT_GENERIC_PLAN 0x0040 /* force use of generic plan */ #define CURSOR_OPT_CUSTOM_PLAN 0x0080 /* force use of custom plan */ typedef struct DeclareCursorStmt @@ -2122,7 +2122,7 @@ typedef struct RenameStmt * trigger, etc) */ char *newname; /* the new name */ DropBehavior behavior; /* RESTRICT or CASCADE behavior */ - bool missing_ok; /* skip error if missing? */ + bool missing_ok; /* skip error if missing? */ } RenameStmt; /* ---------------------- @@ -2138,7 +2138,7 @@ typedef struct AlterObjectSchemaStmt List *objarg; /* argument types, if applicable */ char *addname; /* additional name if needed */ char *newschema; /* the new schema */ - bool missing_ok; /* skip error if missing? */ + bool missing_ok; /* skip error if missing? */ } AlterObjectSchemaStmt; /* ---------------------- @@ -2413,7 +2413,7 @@ typedef struct CreateTableAsStmt NodeTag type; Node *query; /* the query (see comments above) */ IntoClause *into; /* destination table */ - bool is_select_into; /* it was written as SELECT INTO */ + bool is_select_into; /* it was written as SELECT INTO */ } CreateTableAsStmt; /* ---------------------- diff --git a/src/include/nodes/primnodes.h b/src/include/nodes/primnodes.h index 50831eebf8ca3d2300c6198a8e238ccd2a122802..cd4561dcf494e1972eed0683c05a0ca8fa2c392a 100644 --- a/src/include/nodes/primnodes.h +++ b/src/include/nodes/primnodes.h @@ -125,9 +125,9 @@ typedef struct Expr * The code doesn't really need varnoold/varoattno, but they are very useful * for debugging and interpreting completed plans, so we keep them around. */ -#define INNER_VAR 65000 /* reference to inner subplan */ -#define OUTER_VAR 65001 /* reference to outer subplan */ -#define INDEX_VAR 65002 /* reference to index column */ +#define INNER_VAR 65000 /* reference to inner subplan */ +#define OUTER_VAR 65001 /* reference to outer subplan */ +#define INDEX_VAR 65002 /* reference to index column */ #define IS_SPECIAL_VARNO(varno) ((varno) >= INNER_VAR) @@ -847,13 +847,13 @@ typedef struct ArrayExpr * the same as the number of columns logically present in the rowtype. * * colnames provides field names in cases where the names can't easily be - * obtained otherwise. Names *must* be provided if row_typeid is RECORDOID. + * obtained otherwise. Names *must* be provided if row_typeid is RECORDOID. * If row_typeid identifies a known composite type, colnames can be NIL to * indicate the type's cataloged field names apply. Note that colnames can * be non-NIL even for a composite type, and typically is when the RowExpr * was created by expanding a whole-row Var. This is so that we can retain * the column alias names of the RTE that the Var referenced (which would - * otherwise be very difficult to extract from the parsetree). Like the + * otherwise be very difficult to extract from the parsetree). Like the * args list, colnames is one-for-one with physical fields of the rowtype. */ typedef struct RowExpr diff --git a/src/include/nodes/relation.h b/src/include/nodes/relation.h index e1d5fc03192ec1ad9a2e6475419c24b8eb3dc2c8..cf0bbd9f159e6fc168bce1a448527c3f901fff71 100644 --- a/src/include/nodes/relation.h +++ b/src/include/nodes/relation.h @@ -147,8 +147,8 @@ typedef struct PlannerInfo /* * all_baserels is a Relids set of all base relids (but not "other" - * relids) in the query; that is, the Relids identifier of the final - * join we need to form. + * relids) in the query; that is, the Relids identifier of the final join + * we need to form. */ Relids all_baserels; @@ -423,7 +423,7 @@ typedef struct RelOptInfo struct Plan *subplan; /* if subquery */ PlannerInfo *subroot; /* if subquery */ /* use "struct FdwRoutine" to avoid including fdwapi.h here */ - struct FdwRoutine *fdwroutine; /* if foreign table */ + struct FdwRoutine *fdwroutine; /* if foreign table */ void *fdw_private; /* if foreign table */ /* used by various scans and joins: */ @@ -575,7 +575,7 @@ typedef struct EquivalenceClass * * em_is_child signifies that this element was built by transposing a member * for an appendrel parent relation to represent the corresponding expression - * for an appendrel child. These members are used for determining the + * for an appendrel child. These members are used for determining the * pathkeys of scans on the child relation and for explicitly sorting the * child when necessary to build a MergeAppend path for the whole appendrel * tree. An em_is_child member has no impact on the properties of the EC as a @@ -668,7 +668,7 @@ typedef struct ParamPathInfo * "param_info", if not NULL, links to a ParamPathInfo that identifies outer * relation(s) that provide parameter values to each scan of this path. * That means this path can only be joined to those rels by means of nestloop - * joins with this path on the inside. Also note that a parameterized path + * joins with this path on the inside. Also note that a parameterized path * is responsible for testing all "movable" joinclauses involving this rel * and the specified outer rel(s). * @@ -832,7 +832,7 @@ typedef struct TidPath /* * ForeignPath represents a potential scan of a foreign table * - * fdw_private stores FDW private data about the scan. While fdw_private is + * fdw_private stores FDW private data about the scan. While fdw_private is * not actually touched by the core code during normal operations, it's * generally a good idea to use a representation that can be dumped by * nodeToString(), so that you can examine the structure during debugging @@ -1483,7 +1483,7 @@ typedef struct MinMaxAggInfo * value in the Var will always be zero. * * A PlaceHolderVar: this works much like the Var case, except that the - * entry is a PlaceHolderVar node with a contained expression. The PHV + * entry is a PlaceHolderVar node with a contained expression. The PHV * will have phlevelsup = 0, and the contained expression is adjusted * to match in level. * diff --git a/src/include/optimizer/cost.h b/src/include/optimizer/cost.h index c197e7c0c1f9b2e34e251ffbb3b1697faa3be1a4..b2cdb3d62e96cb363118cebf0091f0c00a917730 100644 --- a/src/include/optimizer/cost.h +++ b/src/include/optimizer/cost.h @@ -67,9 +67,9 @@ extern double clamp_row_est(double nrows); extern double index_pages_fetched(double tuples_fetched, BlockNumber pages, double index_pages, PlannerInfo *root); extern void cost_seqscan(Path *path, PlannerInfo *root, RelOptInfo *baserel, - ParamPathInfo *param_info); + ParamPathInfo *param_info); extern void cost_index(IndexPath *path, PlannerInfo *root, - double loop_count); + double loop_count); extern void cost_bitmap_heap_scan(Path *path, PlannerInfo *root, RelOptInfo *baserel, ParamPathInfo *param_info, Path *bitmapqual, double loop_count); @@ -79,7 +79,7 @@ extern void cost_bitmap_tree_node(Path *path, Cost *cost, Selectivity *selec); extern void cost_tidscan(Path *path, PlannerInfo *root, RelOptInfo *baserel, List *tidquals); extern void cost_subqueryscan(Path *path, PlannerInfo *root, - RelOptInfo *baserel, ParamPathInfo *param_info); + RelOptInfo *baserel, ParamPathInfo *param_info); extern void cost_functionscan(Path *path, PlannerInfo *root, RelOptInfo *baserel); extern void cost_valuesscan(Path *path, PlannerInfo *root, @@ -153,14 +153,14 @@ extern void compute_semi_anti_join_factors(PlannerInfo *root, SemiAntiJoinFactors *semifactors); extern void set_baserel_size_estimates(PlannerInfo *root, RelOptInfo *rel); extern double get_parameterized_baserel_size(PlannerInfo *root, - RelOptInfo *rel, - List *param_clauses); + RelOptInfo *rel, + List *param_clauses); extern double get_parameterized_joinrel_size(PlannerInfo *root, - RelOptInfo *rel, - double outer_rows, - double inner_rows, - SpecialJoinInfo *sjinfo, - List *restrict_clauses); + RelOptInfo *rel, + double outer_rows, + double inner_rows, + SpecialJoinInfo *sjinfo, + List *restrict_clauses); extern void set_joinrel_size_estimates(PlannerInfo *root, RelOptInfo *rel, RelOptInfo *outer_rel, RelOptInfo *inner_rel, diff --git a/src/include/optimizer/pathnode.h b/src/include/optimizer/pathnode.h index 4b2483be6030a7921247e23bc50b5a61a9ce8b7b..385bae6eb80e737741001d978139dcdeb15fdb1e 100644 --- a/src/include/optimizer/pathnode.h +++ b/src/include/optimizer/pathnode.h @@ -31,7 +31,7 @@ extern bool add_path_precheck(RelOptInfo *parent_rel, List *pathkeys, Relids required_outer); extern Path *create_seqscan_path(PlannerInfo *root, RelOptInfo *rel, - Relids required_outer); + Relids required_outer); extern IndexPath *create_index_path(PlannerInfo *root, IndexOptInfo *index, List *indexclauses, @@ -57,7 +57,7 @@ extern BitmapOrPath *create_bitmap_or_path(PlannerInfo *root, extern TidPath *create_tidscan_path(PlannerInfo *root, RelOptInfo *rel, List *tidquals); extern AppendPath *create_append_path(RelOptInfo *rel, List *subpaths, - Relids required_outer); + Relids required_outer); extern MergeAppendPath *create_merge_append_path(PlannerInfo *root, RelOptInfo *rel, List *subpaths, @@ -68,7 +68,7 @@ extern MaterialPath *create_material_path(RelOptInfo *rel, Path *subpath); extern UniquePath *create_unique_path(PlannerInfo *root, RelOptInfo *rel, Path *subpath, SpecialJoinInfo *sjinfo); extern Path *create_subqueryscan_path(PlannerInfo *root, RelOptInfo *rel, - List *pathkeys, Relids required_outer); + List *pathkeys, Relids required_outer); extern Path *create_functionscan_path(PlannerInfo *root, RelOptInfo *rel); extern Path *create_valuesscan_path(PlannerInfo *root, RelOptInfo *rel); extern Path *create_ctescan_path(PlannerInfo *root, RelOptInfo *rel); @@ -139,18 +139,18 @@ extern RelOptInfo *build_join_rel(PlannerInfo *root, SpecialJoinInfo *sjinfo, List **restrictlist_ptr); extern AppendRelInfo *find_childrel_appendrelinfo(PlannerInfo *root, - RelOptInfo *rel); + RelOptInfo *rel); extern ParamPathInfo *get_baserel_parampathinfo(PlannerInfo *root, - RelOptInfo *baserel, - Relids required_outer); + RelOptInfo *baserel, + Relids required_outer); extern ParamPathInfo *get_joinrel_parampathinfo(PlannerInfo *root, - RelOptInfo *joinrel, - Path *outer_path, - Path *inner_path, - SpecialJoinInfo *sjinfo, - Relids required_outer, - List **restrict_clauses); + RelOptInfo *joinrel, + Path *outer_path, + Path *inner_path, + SpecialJoinInfo *sjinfo, + Relids required_outer, + List **restrict_clauses); extern ParamPathInfo *get_appendrel_parampathinfo(RelOptInfo *appendrel, - Relids required_outer); + Relids required_outer); #endif /* PATHNODE_H */ diff --git a/src/include/optimizer/paths.h b/src/include/optimizer/paths.h index b3a2dc1d2de028382f9a5b6c24608cc980090775..b6fb8ee5ce93cda5de8a824528378921f43f0584 100644 --- a/src/include/optimizer/paths.h +++ b/src/include/optimizer/paths.h @@ -50,8 +50,8 @@ extern bool relation_has_unique_index_for(PlannerInfo *root, RelOptInfo *rel, List *restrictlist, List *exprlist, List *oprlist); extern bool eclass_member_matches_indexcol(EquivalenceClass *ec, - EquivalenceMember *em, - IndexOptInfo *index, int indexcol); + EquivalenceMember *em, + IndexOptInfo *index, int indexcol); extern bool match_index_to_operand(Node *operand, int indexcol, IndexOptInfo *index); extern void expand_indexqual_conditions(IndexOptInfo *index, diff --git a/src/include/optimizer/prep.h b/src/include/optimizer/prep.h index fb03acc2b4482f383a4e44087359b95316986cbb..47a27b66e8a4e6379d07f71a2899358fb2477df8 100644 --- a/src/include/optimizer/prep.h +++ b/src/include/optimizer/prep.h @@ -53,6 +53,6 @@ extern Plan *plan_set_operations(PlannerInfo *root, double tuple_fraction, extern void expand_inherited_tables(PlannerInfo *root); extern Node *adjust_appendrel_attrs(PlannerInfo *root, Node *node, - AppendRelInfo *appinfo); + AppendRelInfo *appinfo); #endif /* PREP_H */ diff --git a/src/include/optimizer/subselect.h b/src/include/optimizer/subselect.h index 06b50624a103f163bd31efed1ded001482941617..90fe8fc9c0ed9329cca969ee21f1bdb04908d006 100644 --- a/src/include/optimizer/subselect.h +++ b/src/include/optimizer/subselect.h @@ -31,7 +31,7 @@ extern Param *SS_make_initplan_from_plan(PlannerInfo *root, Plan *plan, Oid resulttype, int32 resulttypmod, Oid resultcollation); extern Param *assign_nestloop_param_var(PlannerInfo *root, Var *var); extern Param *assign_nestloop_param_placeholdervar(PlannerInfo *root, - PlaceHolderVar *phv); + PlaceHolderVar *phv); extern int SS_assign_special_param(PlannerInfo *root); #endif /* SUBSELECT_H */ diff --git a/src/include/parser/analyze.h b/src/include/parser/analyze.h index fe7f80a5aaa7a7d46370b07b3e66656d3b44bf20..5fbf5209922b8d59d0a775764172bac4324dba0d 100644 --- a/src/include/parser/analyze.h +++ b/src/include/parser/analyze.h @@ -18,7 +18,7 @@ /* Hook for plugins to get control at end of parse analysis */ typedef void (*post_parse_analyze_hook_type) (ParseState *pstate, - Query *query); + Query *query); extern PGDLLIMPORT post_parse_analyze_hook_type post_parse_analyze_hook; diff --git a/src/include/pg_config_manual.h b/src/include/pg_config_manual.h index ac45ee6426eb32e3dd893f7e3e82d2b98dc9fc03..f29f9e64e3432141bb461691f846076ec26d42d0 100644 --- a/src/include/pg_config_manual.h +++ b/src/include/pg_config_manual.h @@ -182,7 +182,7 @@ * which should be safe in nearly all cases. You might want to override * this if you are building 32-bit code for a known-recent PPC machine. */ -#ifdef HAVE_PPC_LWARX_MUTEX_HINT /* must have assembler support in any case */ +#ifdef HAVE_PPC_LWARX_MUTEX_HINT /* must have assembler support in any case */ #if defined(__ppc64__) || defined(__powerpc64__) #define USE_PPC_LWARX_MUTEX_HINT #endif @@ -190,7 +190,7 @@ /* * On PPC machines, decide whether to use LWSYNC instructions in place of - * ISYNC and SYNC. This provides slightly better performance, but will + * ISYNC and SYNC. This provides slightly better performance, but will * result in illegal-instruction failures on some pre-POWER4 machines. * By default we use LWSYNC when building for 64-bit PPC, which should be * safe in nearly all cases. diff --git a/src/include/pg_trace.h b/src/include/pg_trace.h index 049f1b29d828f1d3adbbf719a901619d235b0a71..45d495fc0db0ecf76b8a6517260bef76b7ddb00c 100644 --- a/src/include/pg_trace.h +++ b/src/include/pg_trace.h @@ -12,6 +12,6 @@ #ifndef PG_TRACE_H #define PG_TRACE_H -#include "utils/probes.h" /* pgrminclude ignore */ +#include "utils/probes.h" /* pgrminclude ignore */ #endif /* PG_TRACE_H */ diff --git a/src/include/pgstat.h b/src/include/pgstat.h index 3583bbe3014cf259164c1b2a5808e6a235cdd40a..dd978d79c3dad830a5925341177a08cc57eeef61 100644 --- a/src/include/pgstat.h +++ b/src/include/pgstat.h @@ -233,8 +233,8 @@ typedef struct PgStat_MsgTabstat int m_nentries; int m_xact_commit; int m_xact_rollback; - PgStat_Counter m_block_read_time; /* times in microseconds */ - PgStat_Counter m_block_write_time; + PgStat_Counter m_block_read_time; /* times in microseconds */ + PgStat_Counter m_block_write_time; PgStat_TableEntry m_entry[PGSTAT_NUM_TABENTRIES]; } PgStat_MsgTabstat; @@ -429,7 +429,7 @@ typedef struct PgStat_FunctionEntry { Oid f_id; PgStat_Counter f_numcalls; - PgStat_Counter f_total_time; /* times in microseconds */ + PgStat_Counter f_total_time; /* times in microseconds */ PgStat_Counter f_self_time; } PgStat_FunctionEntry; @@ -540,7 +540,7 @@ typedef struct PgStat_StatDBEntry PgStat_Counter n_temp_files; PgStat_Counter n_temp_bytes; PgStat_Counter n_deadlocks; - PgStat_Counter n_block_read_time; /* times in microseconds */ + PgStat_Counter n_block_read_time; /* times in microseconds */ PgStat_Counter n_block_write_time; TimestampTz stat_reset_timestamp; @@ -600,7 +600,7 @@ typedef struct PgStat_StatFuncEntry PgStat_Counter f_numcalls; - PgStat_Counter f_total_time; /* times in microseconds */ + PgStat_Counter f_total_time; /* times in microseconds */ PgStat_Counter f_self_time; } PgStat_StatFuncEntry; @@ -629,7 +629,8 @@ typedef struct PgStat_GlobalStats * Backend states * ---------- */ -typedef enum BackendState { +typedef enum BackendState +{ STATE_UNDEFINED, STATE_IDLE, STATE_RUNNING, @@ -674,7 +675,7 @@ typedef struct PgBackendStatus TimestampTz st_proc_start_timestamp; TimestampTz st_xact_start_timestamp; TimestampTz st_activity_start_timestamp; - TimestampTz st_state_start_timestamp; + TimestampTz st_state_start_timestamp; /* Database OID, owning user's OID, connection client address */ Oid st_databaseid; @@ -685,8 +686,8 @@ typedef struct PgBackendStatus /* Is backend currently waiting on an lmgr lock? */ bool st_waiting; - /* current state */ - BackendState st_state; + /* current state */ + BackendState st_state; /* application name; MUST be null-terminated */ char *st_appname; diff --git a/src/include/port.h b/src/include/port.h index 9f06f1a34d86026d480655105bd41d94a314910c..25c4e9883d01365598c6033c5b91d364ac283b35 100644 --- a/src/include/port.h +++ b/src/include/port.h @@ -244,7 +244,6 @@ __attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); extern char *pgwin32_setlocale(int category, const char *locale); #define setlocale(a,b) pgwin32_setlocale(a,b) - #endif /* WIN32 */ /* Portable prompt handling */ @@ -378,7 +377,7 @@ extern long pg_lrand48(void); extern void pg_srand48(long seed); #ifndef HAVE_FLS -extern int fls(int mask); +extern int fls(int mask); #endif #ifndef HAVE_FSEEKO diff --git a/src/include/port/win32.h b/src/include/port/win32.h index 287da98ecf5e0ec87d1399f4ec6110a5bee7cdb4..a00ec897d296582df2396aacc2a6626153aaee1f 100644 --- a/src/include/port/win32.h +++ b/src/include/port/win32.h @@ -319,7 +319,7 @@ typedef int pid_t; #define ECONNREFUSED WSAECONNREFUSED #define EOPNOTSUPP WSAEOPNOTSUPP #pragma warning(default:4005) -#endif +#endif /* * Extended locale functions with gratuitous underscore prefixes. diff --git a/src/include/postgres.h b/src/include/postgres.h index 94c0218cd1b29d50582f5d497cabed69c11ea768..63203bec9a59cb9cff25824dc4b8c60985e44f7b 100644 --- a/src/include/postgres.h +++ b/src/include/postgres.h @@ -685,6 +685,6 @@ extern PGDLLIMPORT bool assert_enabled; extern void ExceptionalCondition(const char *conditionName, const char *errorType, - const char *fileName, int lineNumber) __attribute__((noreturn)); + const char *fileName, int lineNumber) __attribute__((noreturn)); #endif /* POSTGRES_H */ diff --git a/src/include/postmaster/postmaster.h b/src/include/postmaster/postmaster.h index dded0e623e1bdbc32850c8e812f706e0676462c7..683ce3c4074775d173de8de20049b59109e55d4d 100644 --- a/src/include/postmaster/postmaster.h +++ b/src/include/postmaster/postmaster.h @@ -33,13 +33,15 @@ extern bool restart_after_crash; #ifdef WIN32 extern HANDLE PostmasterHandle; #else -extern int postmaster_alive_fds[2]; +extern int postmaster_alive_fds[2]; + /* * Constants that represent which of postmaster_alive_fds is held by * postmaster, and which is used in children to check for postmaster death. */ -#define POSTMASTER_FD_WATCH 0 /* used in children to check for postmaster death */ -#define POSTMASTER_FD_OWN 1 /* kept open by postmaster only */ +#define POSTMASTER_FD_WATCH 0 /* used in children to check for + * postmaster death */ +#define POSTMASTER_FD_OWN 1 /* kept open by postmaster only */ #endif extern const char *progname; diff --git a/src/include/regex/regguts.h b/src/include/regex/regguts.h index 65b8d178da86d244080560e9afd5499a4b500b24..e8415799ec6da88945c5b93832e39f98b176fb4f 100644 --- a/src/include/regex/regguts.h +++ b/src/include/regex/regguts.h @@ -186,12 +186,12 @@ union tree * * If "sub" is not NOSUB then it is the number of the color's current * subcolor, i.e. we are in process of dividing this color (character - * equivalence class) into two colors. See src/backend/regex/README for + * equivalence class) into two colors. See src/backend/regex/README for * discussion of subcolors. * * Currently-unused colors have the FREECOL bit set and are linked into a * freelist using their "sub" fields, but only if their color numbers are - * less than colormap.max. Any array entries beyond "max" are just garbage. + * less than colormap.max. Any array entries beyond "max" are just garbage. */ struct colordesc { diff --git a/src/include/replication/walprotocol.h b/src/include/replication/walprotocol.h index 01cb208d6f4714574980f5d0b7da5b5b6a95581e..0305fb7e59e9031811f36f7b8eb5a78efd956b92 100644 --- a/src/include/replication/walprotocol.h +++ b/src/include/replication/walprotocol.h @@ -59,7 +59,7 @@ typedef struct * * Note that the data length is not specified here. */ -typedef WalSndrMessage PrimaryKeepaliveMessage; +typedef WalSndrMessage PrimaryKeepaliveMessage; /* * Reply message from standby (message type 'r'). This is wrapped within diff --git a/src/include/replication/walreceiver.h b/src/include/replication/walreceiver.h index 68c864751e22984b61fe11732a2d507265dd4e4b..d21ec94a45511b3432515f3b7175da883a112503 100644 --- a/src/include/replication/walreceiver.h +++ b/src/include/replication/walreceiver.h @@ -118,7 +118,7 @@ extern void ShutdownWalRcv(void); extern bool WalRcvInProgress(void); extern void RequestXLogStreaming(XLogRecPtr recptr, const char *conninfo); extern XLogRecPtr GetWalRcvWriteRecPtr(XLogRecPtr *latestChunkStart); -extern int GetReplicationApplyDelay(void); -extern int GetReplicationTransferLatency(void); +extern int GetReplicationApplyDelay(void); +extern int GetReplicationTransferLatency(void); #endif /* _WALRECEIVER_H */ diff --git a/src/include/replication/walsender_private.h b/src/include/replication/walsender_private.h index 183bf19f6df171623fddd8437d44b8851b72c643..66234cd8b5925d81b535522983fdc75da1d40c75 100644 --- a/src/include/replication/walsender_private.h +++ b/src/include/replication/walsender_private.h @@ -35,7 +35,8 @@ typedef struct WalSnd pid_t pid; /* this walsender's process id, or 0 */ WalSndState state; /* this walsender's state */ XLogRecPtr sentPtr; /* WAL has been sent up to this point */ - bool needreload; /* does currently-open file need to be reloaded? */ + bool needreload; /* does currently-open file need to be + * reloaded? */ /* * The xlog locations that have been written, flushed, and applied by diff --git a/src/include/rewrite/rewriteSupport.h b/src/include/rewrite/rewriteSupport.h index 653bc444cab544fdcae2cf1ce55e36ae2234f112..e61fc0582d576fc294ac0de6829a191ddd9ccd63 100644 --- a/src/include/rewrite/rewriteSupport.h +++ b/src/include/rewrite/rewriteSupport.h @@ -23,7 +23,7 @@ extern void SetRelationRuleStatus(Oid relationId, bool relHasRules, bool relIsBecomingView); extern Oid get_rewrite_oid(Oid relid, const char *rulename, bool missing_ok); -extern Oid get_rewrite_oid_without_relid(const char *rulename, - Oid *relid, bool missing_ok); +extern Oid get_rewrite_oid_without_relid(const char *rulename, + Oid *relid, bool missing_ok); #endif /* REWRITESUPPORT_H */ diff --git a/src/include/snowball/header.h b/src/include/snowball/header.h index 0c280a076c397568bffa32dfb5990892b7c297e6..159aa4d8e87a4de63e6101ba291cc4a2eb89781f 100644 --- a/src/include/snowball/header.h +++ b/src/include/snowball/header.h @@ -33,7 +33,7 @@ #endif /* Now we can include the original Snowball header.h */ -#include "snowball/libstemmer/header.h" /* pgrminclude ignore */ +#include "snowball/libstemmer/header.h" /* pgrminclude ignore */ /* * Redefine standard memory allocation interface to pgsql's one. diff --git a/src/include/storage/barrier.h b/src/include/storage/barrier.h index 57f03ecf2aaed6cf21d785d1ef8f170b07009141..50378709917049fde75e7ff7a2ffecf2bfd5c879 100644 --- a/src/include/storage/barrier.h +++ b/src/include/storage/barrier.h @@ -15,7 +15,7 @@ #include "storage/s_lock.h" -extern slock_t dummy_spinlock; +extern slock_t dummy_spinlock; /* * A compiler barrier need not (and preferably should not) emit any actual @@ -30,10 +30,10 @@ extern slock_t dummy_spinlock; * loads and stores are totally ordered (which is not the case on most * architectures) this requires issuing some sort of memory fencing * instruction. - * + * * A read barrier must act as a compiler barrier, and in addition must * guarantee that any loads issued prior to the barrier are completed before - * any loads issued after the barrier. Similarly, a write barrier acts + * any loads issued after the barrier. Similarly, a write barrier acts * as a compiler barrier, and also orders stores. Read and write barriers * are thus weaker than a full memory barrier, but stronger than a compiler * barrier. In practice, on machines with strong memory ordering, read and @@ -48,7 +48,6 @@ extern slock_t dummy_spinlock; /* * Fall through to the spinlock-based implementation. */ - #elif defined(__INTEL_COMPILER) /* @@ -56,7 +55,6 @@ extern slock_t dummy_spinlock; */ #define pg_memory_barrier() _mm_mfence() #define pg_compiler_barrier() __memory_barrier() - #elif defined(__GNUC__) /* This works on any architecture, since it's only talking to GCC itself. */ @@ -75,7 +73,6 @@ extern slock_t dummy_spinlock; __asm__ __volatile__ ("lock; addl $0,0(%%esp)" : : : "memory") #define pg_read_barrier() pg_compiler_barrier() #define pg_write_barrier() pg_compiler_barrier() - #elif defined(__x86_64__) /* 64 bit x86 */ /* @@ -90,7 +87,6 @@ extern slock_t dummy_spinlock; __asm__ __volatile__ ("lock; addl $0,0(%%rsp)" : : : "memory") #define pg_read_barrier() pg_compiler_barrier() #define pg_write_barrier() pg_compiler_barrier() - #elif defined(__ia64__) || defined(__ia64) /* @@ -98,7 +94,6 @@ extern slock_t dummy_spinlock; * fence. */ #define pg_memory_barrier() __asm__ __volatile__ ("mf" : : : "memory") - #elif defined(__ppc__) || defined(__powerpc__) || defined(__ppc64__) || defined(__powerpc64__) /* @@ -109,8 +104,7 @@ extern slock_t dummy_spinlock; #define pg_memory_barrier() __asm__ __volatile__ ("sync" : : : "memory") #define pg_read_barrier() __asm__ __volatile__ ("lwsync" : : : "memory") #define pg_write_barrier() __asm__ __volatile__ ("lwsync" : : : "memory") - -#elif defined(__alpha) || defined(__alpha__) /* Alpha */ +#elif defined(__alpha) || defined(__alpha__) /* Alpha */ /* * Unlike all other known architectures, Alpha allows dependent reads to be @@ -120,7 +114,6 @@ extern slock_t dummy_spinlock; #define pg_memory_barrier() __asm__ __volatile__ ("mb" : : : "memory") #define pg_read_barrier() __asm__ __volatile__ ("rmb" : : : "memory") #define pg_write_barrier() __asm__ __volatile__ ("wmb" : : : "memory") - #elif __GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 1) /* @@ -129,14 +122,11 @@ extern slock_t dummy_spinlock; * own definitions where possible, and use this only as a fallback. */ #define pg_memory_barrier() __sync_synchronize() - #endif - #elif defined(__ia64__) || defined(__ia64) #define pg_compiler_barrier() _Asm_sched_fence() #define pg_memory_barrier() _Asm_mf() - #elif defined(WIN32_ONLY_COMPILER) /* Should work on both MSVC and Borland. */ @@ -144,7 +134,6 @@ extern slock_t dummy_spinlock; #pragma intrinsic(_ReadWriteBarrier) #define pg_compiler_barrier() _ReadWriteBarrier() #define pg_memory_barrier() MemoryBarrier() - #endif /* diff --git a/src/include/storage/latch.h b/src/include/storage/latch.h index 6a7df38d1a2591599afe3602f3fdcc3276114be5..71fb4868a000ccdd1afc6dd38bbbc8102c07f87f 100644 --- a/src/include/storage/latch.h +++ b/src/include/storage/latch.h @@ -68,7 +68,7 @@ * than an ad-hoc shared latch for signaling auxiliary processes. This is * because generic signal handlers will call SetLatch on the process latch * only, so using any latch other than the process latch effectively precludes - * ever registering a generic handler. Since signals have the potential to + * ever registering a generic handler. Since signals have the potential to * invalidate the latch timeout on some platforms, resulting in a * denial-of-service, it is important to verify that all signal handlers * within all WaitLatch-calling processes call SetLatch. @@ -102,10 +102,10 @@ typedef struct } Latch; /* Bitmasks for events that may wake-up WaitLatch() clients */ -#define WL_LATCH_SET (1 << 0) -#define WL_SOCKET_READABLE (1 << 1) +#define WL_LATCH_SET (1 << 0) +#define WL_SOCKET_READABLE (1 << 1) #define WL_SOCKET_WRITEABLE (1 << 2) -#define WL_TIMEOUT (1 << 3) +#define WL_TIMEOUT (1 << 3) #define WL_POSTMASTER_DEATH (1 << 4) /* @@ -115,7 +115,7 @@ extern void InitLatch(volatile Latch *latch); extern void InitSharedLatch(volatile Latch *latch); extern void OwnLatch(volatile Latch *latch); extern void DisownLatch(volatile Latch *latch); -extern int WaitLatch(volatile Latch *latch, int wakeEvents, long timeout); +extern int WaitLatch(volatile Latch *latch, int wakeEvents, long timeout); extern int WaitLatchOrSocket(volatile Latch *latch, int wakeEvents, pgsocket sock, long timeout); extern void SetLatch(volatile Latch *latch); diff --git a/src/include/storage/lock.h b/src/include/storage/lock.h index 17b894285bac87df8078f17118410b4cb896902f..d629ac2ad2e7cfb11988284d7df78644de46ab3a 100644 --- a/src/include/storage/lock.h +++ b/src/include/storage/lock.h @@ -428,7 +428,7 @@ typedef struct LockInstanceData LOCKMASK holdMask; /* locks held by this PGPROC */ LOCKMODE waitLockMode; /* lock awaited by this PGPROC, if any */ BackendId backend; /* backend ID of this PGPROC */ - LocalTransactionId lxid; /* local transaction ID of this PGPROC */ + LocalTransactionId lxid; /* local transaction ID of this PGPROC */ int pid; /* pid of this PGPROC */ bool fastpath; /* taken via fastpath? */ } LockInstanceData; @@ -436,7 +436,7 @@ typedef struct LockInstanceData typedef struct LockData { int nelements; /* The length of the array */ - LockInstanceData *locks; + LockInstanceData *locks; } LockData; diff --git a/src/include/storage/lwlock.h b/src/include/storage/lwlock.h index 6b59efcbb1501ead369be6dff8eeb521bbb5ffb4..82d8ec4edc262f476f17f8e271dbd001a4c19742 100644 --- a/src/include/storage/lwlock.h +++ b/src/include/storage/lwlock.h @@ -95,9 +95,9 @@ typedef enum LWLockMode { LW_EXCLUSIVE, LW_SHARED, - LW_WAIT_UNTIL_FREE /* A special mode used in PGPROC->lwlockMode, when - * waiting for lock to become free. Not to be used - * as LWLockAcquire argument */ + LW_WAIT_UNTIL_FREE /* A special mode used in PGPROC->lwlockMode, + * when waiting for lock to become free. Not + * to be used as LWLockAcquire argument */ } LWLockMode; diff --git a/src/include/storage/predicate.h b/src/include/storage/predicate.h index 6ea70ea845928631a9bc0d523b77dc0ec8d2dd42..7ec79e077d2c8aa6533527b2826f0690fac31413 100644 --- a/src/include/storage/predicate.h +++ b/src/include/storage/predicate.h @@ -44,7 +44,7 @@ extern bool PageIsPredicateLocked(Relation relation, BlockNumber blkno); /* predicate lock maintenance */ extern Snapshot GetSerializableTransactionSnapshot(Snapshot snapshot); extern void SetSerializableTransactionSnapshot(Snapshot snapshot, - TransactionId sourcexid); + TransactionId sourcexid); extern void RegisterPredicateLockingXid(TransactionId xid); extern void PredicateLockRelation(Relation relation, Snapshot snapshot); extern void PredicateLockPage(Relation relation, BlockNumber blkno, Snapshot snapshot); diff --git a/src/include/storage/proc.h b/src/include/storage/proc.h index 618a02f42bea354bddb05a294d85498e358417f9..71413aaf5910b21c7e8bd467e2cdeb08387552be 100644 --- a/src/include/storage/proc.h +++ b/src/include/storage/proc.h @@ -131,14 +131,15 @@ struct PGPROC struct XidCache subxids; /* cache for subtransaction XIDs */ - /* Per-backend LWLock. Protects fields below. */ + /* Per-backend LWLock. Protects fields below. */ LWLockId backendLock; /* protects the fields below */ /* Lock manager data, recording fast-path locks taken by this backend. */ uint64 fpLockBits; /* lock modes held for each fast-path slot */ - Oid fpRelId[FP_LOCK_SLOTS_PER_BACKEND]; /* slots for rel oids */ + Oid fpRelId[FP_LOCK_SLOTS_PER_BACKEND]; /* slots for rel oids */ bool fpVXIDLock; /* are we holding a fast-path VXID lock? */ - LocalTransactionId fpLocalTransactionId; /* lxid for fast-path VXID lock */ + LocalTransactionId fpLocalTransactionId; /* lxid for fast-path VXID + * lock */ }; /* NOTE: "typedef struct PGPROC PGPROC" appears in storage/lock.h. */ @@ -149,7 +150,7 @@ extern PGDLLIMPORT struct PGXACT *MyPgXact; /* * Prior to PostgreSQL 9.2, the fields below were stored as part of the - * PGPROC. However, benchmarking revealed that packing these particular + * PGPROC. However, benchmarking revealed that packing these particular * members into a separate array as tightly as possible sped up GetSnapshotData * considerably on systems with many CPU cores, by reducing the number of * cache lines needing to be fetched. Thus, think very carefully before adding diff --git a/src/include/storage/procarray.h b/src/include/storage/procarray.h index 0b0aa35ee1c4a103d6c372df43e26a457eda0e99..5b4cab926f58dda816b6b9dfa9b93fedb6e33fce 100644 --- a/src/include/storage/procarray.h +++ b/src/include/storage/procarray.h @@ -43,7 +43,7 @@ extern int GetMaxSnapshotSubxidCount(void); extern Snapshot GetSnapshotData(Snapshot snapshot); extern bool ProcArrayInstallImportedXmin(TransactionId xmin, - TransactionId sourcexid); + TransactionId sourcexid); extern RunningTransactions GetRunningTransactionData(void); diff --git a/src/include/storage/sinval.h b/src/include/storage/sinval.h index 7fdfdbe7c477273488fa00deedeb5e6fc9ce1772..bcf2c8111dd9faa477442156fc4881860538007e 100644 --- a/src/include/storage/sinval.h +++ b/src/include/storage/sinval.h @@ -33,8 +33,8 @@ * updates and deletions in system catalogs (see CacheInvalidateHeapTuple). * An update can generate two inval events, one for the old tuple and one for * the new, but this is reduced to one event if the tuple's hash key doesn't - * change. Note that the inval events themselves don't actually say whether - * the tuple is being inserted or deleted. Also, since we transmit only a + * change. Note that the inval events themselves don't actually say whether + * the tuple is being inserted or deleted. Also, since we transmit only a * hash key, there is a small risk of unnecessary invalidations due to chance * matches of hash keys. * diff --git a/src/include/storage/smgr.h b/src/include/storage/smgr.h index f1e1b8cdefd2d27dde164a945c2b2e9bb04717b1..f8fc2b2d6e82857ed2483dc0a3444e470ffcd43a 100644 --- a/src/include/storage/smgr.h +++ b/src/include/storage/smgr.h @@ -60,7 +60,7 @@ typedef struct SMgrRelationData * submodules. Do not touch them from elsewhere. */ int smgr_which; /* storage manager selector */ - bool smgr_transient; /* T if files are to be closed at EOXact */ + bool smgr_transient; /* T if files are to be closed at EOXact */ /* for md.c; NULL for forks that are not open */ struct _MdfdVec *md_fd[MAX_FORKNUM + 1]; diff --git a/src/include/tsearch/ts_public.h b/src/include/tsearch/ts_public.h index 380df6ca5f5ea295d43a055115b3de8adf574300..d5c18f2cf04074a15d3f92039158c65c8f5f4152 100644 --- a/src/include/tsearch/ts_public.h +++ b/src/include/tsearch/ts_public.h @@ -91,11 +91,11 @@ typedef struct * and ( fot, ball, klubb ). So, dictionary should return: * * nvariant lexeme - * 1 fotball - * 1 klubb - * 2 fot - * 2 ball - * 2 klubb + * 1 fotball + * 1 klubb + * 2 fot + * 2 ball + * 2 klubb * * In general, a TSLexeme will be considered to belong to the same split * variant as the previous one if they have the same nvariant value. diff --git a/src/include/utils/acl.h b/src/include/utils/acl.h index ff3c6aa1f62aefee9563c2e8ef204406c392edc7..6de39b21cf4f01972a7c5b965af65fd29a3ee5cd 100644 --- a/src/include/utils/acl.h +++ b/src/include/utils/acl.h @@ -278,7 +278,7 @@ extern AclMode pg_foreign_data_wrapper_aclmask(Oid fdw_oid, Oid roleid, extern AclMode pg_foreign_server_aclmask(Oid srv_oid, Oid roleid, AclMode mask, AclMaskHow how); extern AclMode pg_type_aclmask(Oid type_oid, Oid roleid, - AclMode mask, AclMaskHow how); + AclMode mask, AclMaskHow how); extern AclResult pg_attribute_aclcheck(Oid table_oid, AttrNumber attnum, Oid roleid, AclMode mode); diff --git a/src/include/utils/builtins.h b/src/include/utils/builtins.h index f246f117ba3db02beb4e624010ef8ed6320757a4..d1e83707606610f65afa8ca4630af73c3f91757d 100644 --- a/src/include/utils/builtins.h +++ b/src/include/utils/builtins.h @@ -311,7 +311,7 @@ extern Datum btnamecmp(PG_FUNCTION_ARGS); extern Datum bttextcmp(PG_FUNCTION_ARGS); /* - * Per-opclass sort support functions for new btrees. Like the + * Per-opclass sort support functions for new btrees. Like the * functions above, these are stored in pg_amproc; most are defined in * access/nbtree/nbtcompare.c */ diff --git a/src/include/utils/guc.h b/src/include/utils/guc.h index 38ed54891a6e95ef6b782b518dd2227aed53ab3d..6810387755499e6b86d6f5569571255911e6b19e 100644 --- a/src/include/utils/guc.h +++ b/src/include/utils/guc.h @@ -313,7 +313,7 @@ extern void ParseLongOption(const char *string, char **name, char **value); extern bool parse_int(const char *value, int *result, int flags, const char **hintmsg); extern bool parse_real(const char *value, double *result); -extern int set_config_option(const char *name, const char *value, +extern int set_config_option(const char *name, const char *value, GucContext context, GucSource source, GucAction action, bool changeVal, int elevel); extern char *GetConfigOptionByName(const char *name, const char **varname); diff --git a/src/include/utils/guc_tables.h b/src/include/utils/guc_tables.h index eaea4862c24a01f52d740881a5f65fb5d55038eb..5d1ca06b2a04dd212910d3ad7d019b2c0f72b1c7 100644 --- a/src/include/utils/guc_tables.h +++ b/src/include/utils/guc_tables.h @@ -120,7 +120,7 @@ typedef struct guc_stack GucSource source; /* source of the prior value */ /* masked value's source must be PGC_S_SESSION, so no need to store it */ GucContext scontext; /* context that set the prior value */ - GucContext masked_scontext; /* context that set the masked value */ + GucContext masked_scontext; /* context that set the masked value */ config_var_value prior; /* previous value of variable */ config_var_value masked; /* SET value in a GUC_SET_LOCAL entry */ } GucStack; @@ -152,7 +152,7 @@ struct config_generic GucSource source; /* source of the current actual value */ GucSource reset_source; /* source of the reset_value */ GucContext scontext; /* context that set the current value */ - GucContext reset_scontext; /* context that set the reset value */ + GucContext reset_scontext; /* context that set the reset value */ GucStack *stack; /* stacked prior values */ void *extra; /* "extra" pointer for current actual value */ char *sourcefile; /* file current setting is from (NULL if not diff --git a/src/include/utils/json.h b/src/include/utils/json.h index c355e0f73eef5876f4c3fc1ce136bc467ca664ed..0f38147acb8fb940e8d3ca5b1c780d02b89cd0d3 100644 --- a/src/include/utils/json.h +++ b/src/include/utils/json.h @@ -25,6 +25,6 @@ extern Datum array_to_json(PG_FUNCTION_ARGS); extern Datum array_to_json_pretty(PG_FUNCTION_ARGS); extern Datum row_to_json(PG_FUNCTION_ARGS); extern Datum row_to_json_pretty(PG_FUNCTION_ARGS); -extern void escape_json(StringInfo buf, const char *str); +extern void escape_json(StringInfo buf, const char *str); #endif /* JSON_H */ diff --git a/src/include/utils/lsyscache.h b/src/include/utils/lsyscache.h index 696ca77307f1d4a1decf51023046916fad7ee7a5..4bf1029e109f5356e1a1d5261c8f1953076ab1fb 100644 --- a/src/include/utils/lsyscache.h +++ b/src/include/utils/lsyscache.h @@ -18,10 +18,10 @@ /* Result list element for get_op_btree_interpretation */ typedef struct OpBtreeInterpretation { - Oid opfamily_id; /* btree opfamily containing operator */ - int strategy; /* its strategy number */ - Oid oplefttype; /* declared left input datatype */ - Oid oprighttype; /* declared right input datatype */ + Oid opfamily_id; /* btree opfamily containing operator */ + int strategy; /* its strategy number */ + Oid oplefttype; /* declared left input datatype */ + Oid oprighttype; /* declared right input datatype */ } OpBtreeInterpretation; /* I/O function selector for get_type_io_data */ @@ -149,7 +149,7 @@ extern void free_attstatsslot(Oid atttype, Datum *values, int nvalues, float4 *numbers, int nnumbers); extern char *get_namespace_name(Oid nspid); -extern Oid get_range_subtype(Oid rangeOid); +extern Oid get_range_subtype(Oid rangeOid); #define type_is_array(typid) (get_element_type(typid) != InvalidOid) /* type_is_array_domain accepts both plain arrays and domains over arrays */ diff --git a/src/include/utils/memutils.h b/src/include/utils/memutils.h index a2bb3d9c72005649f58b0758932bac0975d55f59..06c8afdd0aa34cc88918aca1a38f795d4fa44309 100644 --- a/src/include/utils/memutils.h +++ b/src/include/utils/memutils.h @@ -91,7 +91,7 @@ extern void MemoryContextResetChildren(MemoryContext context); extern void MemoryContextDeleteChildren(MemoryContext context); extern void MemoryContextResetAndDeleteChildren(MemoryContext context); extern void MemoryContextSetParent(MemoryContext context, - MemoryContext new_parent); + MemoryContext new_parent); extern Size GetMemoryChunkSpace(void *pointer); extern MemoryContext GetMemoryChunkContext(void *pointer); extern MemoryContext MemoryContextGetParent(MemoryContext context); diff --git a/src/include/utils/pg_crc_tables.h b/src/include/utils/pg_crc_tables.h index 524410fffdfb64fe1ca89211ce8fe2f037acee00..43052aa8987a95b85ee148df4eccd430da22d5b4 100644 --- a/src/include/utils/pg_crc_tables.h +++ b/src/include/utils/pg_crc_tables.h @@ -511,7 +511,6 @@ const uint64 pg_crc64_table[256] = { UINT64CONST(0xD80C07CD676F8394), UINT64CONST(0x9AFCE626CE85B507) }; #endif /* SIZEOF_VOID_P < 8 */ - #endif /* PROVIDE_64BIT_CRC */ #endif /* PG_CRC_TABLES_H */ diff --git a/src/include/utils/plancache.h b/src/include/utils/plancache.h index 9b1f7e46ec6acd0852f35be262d5715d6d090c17..413e8462a6c6e50241574d6ebcb2f38837b0af25 100644 --- a/src/include/utils/plancache.h +++ b/src/include/utils/plancache.h @@ -30,7 +30,7 @@ * the analyzed-and-rewritten query tree, and rebuild it when next needed. * * An actual execution plan, represented by CachedPlan, is derived from the - * CachedPlanSource when we need to execute the query. The plan could be + * CachedPlanSource when we need to execute the query. The plan could be * either generic (usable with any set of plan parameters) or custom (for a * specific set of parameters). plancache.c contains the logic that decides * which way to do it for any particular execution. If we are using a generic @@ -84,7 +84,7 @@ typedef struct CachedPlanSource List *query_list; /* list of Query nodes, or NIL if not valid */ List *relationOids; /* OIDs of relations the queries depend on */ List *invalItems; /* other dependencies, as PlanInvalItems */ - MemoryContext query_context; /* context holding the above, or NULL */ + MemoryContext query_context; /* context holding the above, or NULL */ /* If we have a generic plan, this is a reference-counted link to it: */ struct CachedPlan *gplan; /* generic plan, or NULL if not valid */ /* Some state flags: */ @@ -93,26 +93,26 @@ typedef struct CachedPlanSource bool is_valid; /* is the query_list currently valid? */ int generation; /* increments each time we create a plan */ /* If CachedPlanSource has been saved, it is a member of a global list */ - struct CachedPlanSource *next_saved; /* list link, if so */ + struct CachedPlanSource *next_saved; /* list link, if so */ /* State kept to help decide whether to use custom or generic plans: */ double generic_cost; /* cost of generic plan, or -1 if not known */ - double total_custom_cost; /* total cost of custom plans so far */ - int num_custom_plans; /* number of plans included in total */ + double total_custom_cost; /* total cost of custom plans so far */ + int num_custom_plans; /* number of plans included in total */ } CachedPlanSource; /* * CachedPlan represents an execution plan derived from a CachedPlanSource. * The reference count includes both the link from the parent CachedPlanSource * (if any), and any active plan executions, so the plan can be discarded - * exactly when refcount goes to zero. Both the struct itself and the + * exactly when refcount goes to zero. Both the struct itself and the * subsidiary data live in the context denoted by the context field. * This makes it easy to free a no-longer-needed cached plan. */ typedef struct CachedPlan { int magic; /* should equal CACHEDPLAN_MAGIC */ - List *stmt_list; /* list of statement nodes (PlannedStmts - * and bare utility statements) */ + List *stmt_list; /* list of statement nodes (PlannedStmts and + * bare utility statements) */ bool is_saved; /* is CachedPlan in a long-lived context? */ bool is_valid; /* is the stmt_list currently valid? */ TransactionId saved_xmin; /* if valid, replan when TransactionXmin @@ -130,20 +130,20 @@ extern CachedPlanSource *CreateCachedPlan(Node *raw_parse_tree, const char *query_string, const char *commandTag); extern void CompleteCachedPlan(CachedPlanSource *plansource, - List *querytree_list, - MemoryContext querytree_context, - Oid *param_types, - int num_params, - ParserSetupHook parserSetup, - void *parserSetupArg, - int cursor_options, - bool fixed_result); + List *querytree_list, + MemoryContext querytree_context, + Oid *param_types, + int num_params, + ParserSetupHook parserSetup, + void *parserSetupArg, + int cursor_options, + bool fixed_result); extern void SaveCachedPlan(CachedPlanSource *plansource); extern void DropCachedPlan(CachedPlanSource *plansource); extern void CachedPlanSetParentContext(CachedPlanSource *plansource, - MemoryContext newcontext); + MemoryContext newcontext); extern CachedPlanSource *CopyCachedPlan(CachedPlanSource *plansource); diff --git a/src/include/utils/rangetypes.h b/src/include/utils/rangetypes.h index 7b6ff198d472d0bdfc31518069fbef25730cadd4..ad72df57dd61f1acb18b8ceb9e48bf28ffe51600 100644 --- a/src/include/utils/rangetypes.h +++ b/src/include/utils/rangetypes.h @@ -30,7 +30,7 @@ typedef struct } RangeType; /* Use this macro in preference to fetching rangetypid field directly */ -#define RangeTypeGetOid(r) ((r)->rangetypid) +#define RangeTypeGetOid(r) ((r)->rangetypid) /* A range's flags byte contains these bits: */ #define RANGE_EMPTY 0x01 /* range is empty */ @@ -40,8 +40,8 @@ typedef struct #define RANGE_UB_INF 0x10 /* upper bound is +infinity */ #define RANGE_LB_NULL 0x20 /* lower bound is null (NOT USED) */ #define RANGE_UB_NULL 0x40 /* upper bound is null (NOT USED) */ -#define RANGE_CONTAIN_EMPTY 0x80 /* marks a GiST internal-page entry whose - * subtree contains some empty ranges */ +#define RANGE_CONTAIN_EMPTY 0x80/* marks a GiST internal-page entry whose + * subtree contains some empty ranges */ #define RANGE_HAS_LBOUND(flags) (!((flags) & (RANGE_EMPTY | \ RANGE_LB_NULL | \ @@ -149,18 +149,18 @@ extern Datum tstzrange_subdiff(PG_FUNCTION_ARGS); /* assorted support functions */ extern TypeCacheEntry *range_get_typcache(FunctionCallInfo fcinfo, - Oid rngtypid); + Oid rngtypid); extern RangeType *range_serialize(TypeCacheEntry *typcache, RangeBound *lower, - RangeBound *upper, bool empty); + RangeBound *upper, bool empty); extern void range_deserialize(TypeCacheEntry *typcache, RangeType *range, - RangeBound *lower, RangeBound *upper, - bool *empty); + RangeBound *lower, RangeBound *upper, + bool *empty); extern char range_get_flags(RangeType *range); extern void range_set_contain_empty(RangeType *range); extern RangeType *make_range(TypeCacheEntry *typcache, RangeBound *lower, - RangeBound *upper, bool empty); + RangeBound *upper, bool empty); extern int range_cmp_bounds(TypeCacheEntry *typcache, RangeBound *b1, - RangeBound *b2); + RangeBound *b2); extern int range_cmp_bound_values(TypeCacheEntry *typcache, RangeBound *b1, RangeBound *b2); extern RangeType *make_empty_range(TypeCacheEntry *typcache); diff --git a/src/include/utils/rel.h b/src/include/utils/rel.h index d404c2adb516d95d5737b850bfad030de2acf43b..4669d8a67ef4298c591accbc81710c76e1efa535 100644 --- a/src/include/utils/rel.h +++ b/src/include/utils/rel.h @@ -196,7 +196,7 @@ typedef struct StdRdOptions int32 vl_len_; /* varlena header (do not touch directly!) */ int fillfactor; /* page fill factor in percent (0..100) */ AutoVacOpts autovacuum; /* autovacuum-related options */ - bool security_barrier; /* for views */ + bool security_barrier; /* for views */ } StdRdOptions; #define HEAP_MIN_FILLFACTOR 10 diff --git a/src/include/utils/selfuncs.h b/src/include/utils/selfuncs.h index 4529f276839a076d4735131eb37873495de88137..87c6554b327d45cc2d82ab50c3a97341333939a2 100644 --- a/src/include/utils/selfuncs.h +++ b/src/include/utils/selfuncs.h @@ -121,7 +121,7 @@ extern void get_join_variables(PlannerInfo *root, List *args, VariableStatData *vardata2, bool *join_is_reversed); extern double get_variable_numdistinct(VariableStatData *vardata, - bool *isdefault); + bool *isdefault); extern double mcv_selectivity(VariableStatData *vardata, FmgrInfo *opproc, Datum constval, bool varonleft, double *sumcommonp); diff --git a/src/include/utils/sortsupport.h b/src/include/utils/sortsupport.h index ef8d8534934b3e5f3e9e84a98e7c0f166e337a0e..720a54c0d769cb11ac2a588f29f28041e8d49fb2 100644 --- a/src/include/utils/sortsupport.h +++ b/src/include/utils/sortsupport.h @@ -33,7 +33,7 @@ * * Note: since pg_amproc functions are indexed by (lefttype, righttype) * it is possible to associate a BTSORTSUPPORT function with a cross-type - * comparison. This could sensibly be used to provide a fast comparator + * comparison. This could sensibly be used to provide a fast comparator * function for such cases, but probably not any other acceleration method. * * @@ -57,28 +57,28 @@ typedef struct SortSupportData * These fields are initialized before calling the BTSORTSUPPORT function * and should not be changed later. */ - MemoryContext ssup_cxt; /* Context containing sort info */ - Oid ssup_collation; /* Collation to use, or InvalidOid */ + MemoryContext ssup_cxt; /* Context containing sort info */ + Oid ssup_collation; /* Collation to use, or InvalidOid */ /* - * Additional sorting parameters; but unlike ssup_collation, these can - * be changed after BTSORTSUPPORT is called, so don't use them in - * selecting sort support functions. + * Additional sorting parameters; but unlike ssup_collation, these can be + * changed after BTSORTSUPPORT is called, so don't use them in selecting + * sort support functions. */ - bool ssup_reverse; /* descending-order sort? */ + bool ssup_reverse; /* descending-order sort? */ bool ssup_nulls_first; /* sort nulls first? */ /* * These fields are workspace for callers, and should not be touched by * opclass-specific functions. */ - AttrNumber ssup_attno; /* column number to sort */ + AttrNumber ssup_attno; /* column number to sort */ /* - * ssup_extra is zeroed before calling the BTSORTSUPPORT function, and - * is not touched subsequently by callers. + * ssup_extra is zeroed before calling the BTSORTSUPPORT function, and is + * not touched subsequently by callers. */ - void *ssup_extra; /* Workspace for opclass functions */ + void *ssup_extra; /* Workspace for opclass functions */ /* * Function pointers are zeroed before calling the BTSORTSUPPORT function, @@ -140,13 +140,11 @@ ApplySortComparator(Datum datum1, bool isNull1, return compare; } - #else -extern int ApplySortComparator(Datum datum1, bool isNull1, +extern int ApplySortComparator(Datum datum1, bool isNull1, Datum datum2, bool isNull2, SortSupport ssup); - #endif /* USE_INLINE */ /* Other functions in utils/sort/sortsupport.c */ diff --git a/src/include/utils/timestamp.h b/src/include/utils/timestamp.h index 56699247668e62fa1fd92946428044aa15b3b32f..665e969498cffe3df94e17985987311049a6fbbe 100644 --- a/src/include/utils/timestamp.h +++ b/src/include/utils/timestamp.h @@ -42,7 +42,6 @@ #define PG_RETURN_TIMESTAMP(x) return TimestampGetDatum(x) #define PG_RETURN_TIMESTAMPTZ(x) return TimestampTzGetDatum(x) #define PG_RETURN_INTERVAL_P(x) return IntervalPGetDatum(x) - #else /* !HAVE_INT64_TIMESTAMP */ #define DatumGetTimestamp(X) ((Timestamp) DatumGetFloat8(X)) @@ -60,7 +59,6 @@ #define PG_RETURN_TIMESTAMP(x) return TimestampGetDatum(x) #define PG_RETURN_TIMESTAMPTZ(x) return TimestampTzGetDatum(x) #define PG_RETURN_INTERVAL_P(x) return IntervalPGetDatum(x) - #endif /* HAVE_INT64_TIMESTAMP */ diff --git a/src/include/utils/tqual.h b/src/include/utils/tqual.h index 3d8a480d8106bf9e02722247cf2cb74ffcc6e5e5..ff74f868a6fbe1d76d21f32176076a3d20d95293 100644 --- a/src/include/utils/tqual.h +++ b/src/include/utils/tqual.h @@ -84,7 +84,7 @@ extern HTSU_Result HeapTupleSatisfiesUpdate(HeapTupleHeader tuple, extern HTSV_Result HeapTupleSatisfiesVacuum(HeapTupleHeader tuple, TransactionId OldestXmin, Buffer buffer); extern bool HeapTupleIsSurelyDead(HeapTupleHeader tuple, - TransactionId OldestXmin); + TransactionId OldestXmin); extern void HeapTupleSetHintBits(HeapTupleHeader tuple, Buffer buffer, uint16 infomask, TransactionId xid); diff --git a/src/include/utils/typcache.h b/src/include/utils/typcache.h index 935fb864a38c166c11118360edd2d126cda46f16..12fb4f8310169a0b4ff442f9778a1250a988842c 100644 --- a/src/include/utils/typcache.h +++ b/src/include/utils/typcache.h @@ -73,13 +73,13 @@ typedef struct TypeCacheEntry TupleDesc tupDesc; /* - * Fields computed when TYPECACHE_RANGE_INFO is requested. Zeroes if - * not a range type or information hasn't yet been requested. Note that + * Fields computed when TYPECACHE_RANGE_INFO is requested. Zeroes if not + * a range type or information hasn't yet been requested. Note that * rng_cmp_proc_finfo could be different from the element type's default * btree comparison function. */ - struct TypeCacheEntry *rngelemtype; /* range's element type */ - Oid rng_collation; /* collation for comparisons, if any */ + struct TypeCacheEntry *rngelemtype; /* range's element type */ + Oid rng_collation; /* collation for comparisons, if any */ FmgrInfo rng_cmp_proc_finfo; /* comparison function */ FmgrInfo rng_canonical_finfo; /* canonicalization function, if any */ FmgrInfo rng_subdiff_finfo; /* difference function, if any */ diff --git a/src/include/utils/xml.h b/src/include/utils/xml.h index a36e26f2ed0e7e4336a2e1186b7517f1b0a651e7..a645af918c35f09b42aaa732a693f4dde8b2a5be 100644 --- a/src/include/utils/xml.h +++ b/src/include/utils/xml.h @@ -37,11 +37,11 @@ typedef enum typedef enum { - PG_XML_STRICTNESS_LEGACY, /* ignore errors unless function result - * indicates error condition */ - PG_XML_STRICTNESS_WELLFORMED, /* ignore non-parser messages */ - PG_XML_STRICTNESS_ALL /* report all notices/warnings/errors */ -} PgXmlStrictness; + PG_XML_STRICTNESS_LEGACY, /* ignore errors unless function result + * indicates error condition */ + PG_XML_STRICTNESS_WELLFORMED, /* ignore non-parser messages */ + PG_XML_STRICTNESS_ALL /* report all notices/warnings/errors */ +} PgXmlStrictness; /* struct PgXmlErrorContext is private to xml.c */ typedef struct PgXmlErrorContext PgXmlErrorContext; @@ -90,7 +90,7 @@ extern PgXmlErrorContext *pg_xml_init(PgXmlStrictness strictness); extern void pg_xml_done(PgXmlErrorContext *errcxt, bool isError); extern bool pg_xml_error_occurred(PgXmlErrorContext *errcxt); extern void xml_ereport(PgXmlErrorContext *errcxt, int level, int sqlcode, - const char *msg); + const char *msg); extern xmltype *xmlconcat(List *args); extern xmltype *xmlelement(XmlExprState *xmlExpr, ExprContext *econtext); diff --git a/src/interfaces/ecpg/ecpglib/connect.c b/src/interfaces/ecpg/ecpglib/connect.c index b54b1f5503626abae63b984f74c3bd1164bc19a0..49f2d546bbe1bb5f1dbbca66b77b0c399f51d6cd 100644 --- a/src/interfaces/ecpg/ecpglib/connect.c +++ b/src/interfaces/ecpg/ecpglib/connect.c @@ -267,7 +267,8 @@ ECPGconnect(int lineno, int c, const char *name, const char *user, const char *p struct sqlca_t *sqlca = ECPGget_sqlca(); enum COMPAT_MODE compat = c; struct connection *this; - int i, connect_params = 0; + int i, + connect_params = 0; char *dbname = name ? ecpg_strdup(name, lineno) : NULL, *host = NULL, *tmp, @@ -505,10 +506,10 @@ ECPGconnect(int lineno, int c, const char *name, const char *user, const char *p connect_params++; /* allocate enough space for all connection parameters */ - conn_keywords = (const char **) ecpg_alloc((connect_params + 1) * sizeof (char *), lineno); - conn_values = (const char **) ecpg_alloc(connect_params * sizeof (char *), lineno); + conn_keywords = (const char **) ecpg_alloc((connect_params + 1) * sizeof(char *), lineno); + conn_values = (const char **) ecpg_alloc(connect_params * sizeof(char *), lineno); if (conn_keywords == NULL || conn_values == NULL) - { + { if (host) ecpg_free(host); if (port) @@ -560,23 +561,25 @@ ECPGconnect(int lineno, int c, const char *name, const char *user, const char *p } if (options) { - char *str; + char *str; /* options look like this "option1 = value1 option2 = value2 ... */ /* we have to break up the string into single options */ for (str = options; *str;) { - int e, a; - char *token1, *token2; + int e, + a; + char *token1, + *token2; - for (token1 = str; *token1 && *token1 == ' '; token1++); + for (token1 = str; *token1 && *token1 == ' '; token1++); for (e = 0; token1[e] && token1[e] != '='; e++); - if (token1[e]) /* found "=" */ + if (token1[e]) /* found "=" */ { token1[e] = '\0'; for (token2 = token1 + e + 1; *token2 && *token2 == ' '; token2++); for (a = 0; token2[a] && token2[a] != '&'; a++); - if (token2[a]) /* found "&" => another option follows */ + if (token2[a]) /* found "&" => another option follows */ { token2[a] = '\0'; str = token2 + a + 1; @@ -587,10 +590,10 @@ ECPGconnect(int lineno, int c, const char *name, const char *user, const char *p conn_keywords[i] = token1; conn_values[i] = token2; i++; - } + } else /* the parser should not be able to create this invalid option */ - str = token1 + e; + str = token1 + e; } } diff --git a/src/interfaces/ecpg/ecpglib/execute.c b/src/interfaces/ecpg/ecpglib/execute.c index 50a2d95347edc4a656e7db2a1191e292052feff1..1a7876ecf2d7db90abda37e289f2b38d9e15f1b5 100644 --- a/src/interfaces/ecpg/ecpglib/execute.c +++ b/src/interfaces/ecpg/ecpglib/execute.c @@ -1441,7 +1441,7 @@ ecpg_execute(struct statement * stmt) ecpg_log("ecpg_execute on line %d: query: %s; with %d parameter(s) on connection %s\n", stmt->lineno, stmt->command, nParams, stmt->connection->name); if (stmt->statement_type == ECPGst_execute) { - results = PQexecPrepared(stmt->connection->connection, stmt->name, nParams, (const char *const*) paramValues, NULL, NULL, 0); + results = PQexecPrepared(stmt->connection->connection, stmt->name, nParams, (const char *const *) paramValues, NULL, NULL, 0); ecpg_log("ecpg_execute on line %d: using PQexecPrepared for \"%s\"\n", stmt->lineno, stmt->command); } else @@ -1453,7 +1453,7 @@ ecpg_execute(struct statement * stmt) } else { - results = PQexecParams(stmt->connection->connection, stmt->command, nParams, NULL, (const char *const*) paramValues, NULL, NULL, 0); + results = PQexecParams(stmt->connection->connection, stmt->command, nParams, NULL, (const char *const *) paramValues, NULL, NULL, 0); ecpg_log("ecpg_execute on line %d: using PQexecParams\n", stmt->lineno); } } diff --git a/src/interfaces/ecpg/ecpglib/extern.h b/src/interfaces/ecpg/ecpglib/extern.h index bd1ffb096c0a09b287fd4f6c3aa4cd087e4d0d4e..835e70c38f4d0767baa837d3265ae0633c90a2d3 100644 --- a/src/interfaces/ecpg/ecpglib/extern.h +++ b/src/interfaces/ecpg/ecpglib/extern.h @@ -123,8 +123,8 @@ struct variable struct var_list { - int number; - void *pointer; + int number; + void *pointer; struct var_list *next; }; @@ -170,7 +170,7 @@ void ecpg_raise(int line, int code, const char *sqlstate, const char *str); void ecpg_raise_backend(int line, PGresult *result, PGconn *conn, int compat); char *ecpg_prepared(const char *, struct connection *); bool ecpg_deallocate_all_conn(int lineno, enum COMPAT_MODE c, struct connection * conn); -void ecpg_log(const char *format, ...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); +void ecpg_log(const char *format,...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 1, 2))); bool ecpg_auto_prepare(int, const char *, const int, char **, const char *); void ecpg_init_sqlca(struct sqlca_t * sqlca); diff --git a/src/interfaces/ecpg/pgtypeslib/dt.h b/src/interfaces/ecpg/pgtypeslib/dt.h index 269af7822b15d57287109aa5f97ee52aedfcf0c5..dfe6f9e687276fdacc6d70f68b2fc7d5c0eb4224 100644 --- a/src/interfaces/ecpg/pgtypeslib/dt.h +++ b/src/interfaces/ecpg/pgtypeslib/dt.h @@ -334,12 +334,12 @@ do { \ int DecodeInterval(char **, int *, int, int *, struct tm *, fsec_t *); int DecodeTime(char *, int *, struct tm *, fsec_t *); -int EncodeDateTime(struct tm *tm, fsec_t fsec, bool print_tz, int tz, const char *tzn, int style, char *str, bool EuroDates); -int EncodeInterval(struct tm *tm, fsec_t fsec, int style, char *str); +int EncodeDateTime(struct tm * tm, fsec_t fsec, bool print_tz, int tz, const char *tzn, int style, char *str, bool EuroDates); +int EncodeInterval(struct tm * tm, fsec_t fsec, int style, char *str); int tm2timestamp(struct tm *, fsec_t, int *, timestamp *); int DecodeUnits(int field, char *lowtoken, int *val); bool CheckDateTokenTables(void); -int EncodeDateOnly(struct tm *tm, int style, char *str, bool EuroDates); +int EncodeDateOnly(struct tm * tm, int style, char *str, bool EuroDates); int GetEpochTime(struct tm *); int ParseDateTime(char *, char *, char **, int *, int *, char **); int DecodeDateTime(char **, int *, int, int *, struct tm *, fsec_t *, bool); diff --git a/src/interfaces/ecpg/preproc/type.c b/src/interfaces/ecpg/preproc/type.c index ff1224754ea8529971ec022295c71b0689b39448..c743616a6caa539533dbfcc4a97bf3a23ff34267 100644 --- a/src/interfaces/ecpg/preproc/type.c +++ b/src/interfaces/ecpg/preproc/type.c @@ -396,7 +396,10 @@ ECPGdump_a_simple(FILE *o, const char *name, enum ECPGttype type, else sprintf(variable, "&(%s%s)", prefix ? prefix : "", name); - /* If we created a varchar structure atomatically, counter is greater than 0. */ + /* + * If we created a varchar structure atomatically, counter is + * greater than 0. + */ if (counter) sprintf(offset, "sizeof(struct varchar_%d)", counter); else diff --git a/src/interfaces/libpq/fe-connect.c b/src/interfaces/libpq/fe-connect.c index 33daaec7be47ab6113ad4de295a4292f9c84290f..7c9fa34560a3502557c01794eb5bf7de872a8568 100644 --- a/src/interfaces/libpq/fe-connect.c +++ b/src/interfaces/libpq/fe-connect.c @@ -298,7 +298,7 @@ static void closePGconn(PGconn *conn); static PQconninfoOption *conninfo_init(PQExpBuffer errorMessage); static PQconninfoOption *parse_connection_string(const char *conninfo, PQExpBuffer errorMessage, bool use_defaults); -static int uri_prefix_length(const char *connstr); +static int uri_prefix_length(const char *connstr); static bool recognized_connection_string(const char *connstr); static PQconninfoOption *conninfo_parse(const char *conninfo, PQExpBuffer errorMessage, bool use_defaults); @@ -308,7 +308,7 @@ static PQconninfoOption *conninfo_array_parse(const char *const * keywords, static bool conninfo_add_defaults(PQconninfoOption *options, PQExpBuffer errorMessage); static PQconninfoOption *conninfo_uri_parse(const char *uri, - PQExpBuffer errorMessage, bool use_defaults); + PQExpBuffer errorMessage, bool use_defaults); static bool conninfo_uri_parse_options(PQconninfoOption *options, const char *uri, PQExpBuffer errorMessage); static bool conninfo_uri_parse_params(char *params, @@ -319,8 +319,8 @@ static bool get_hexdigit(char digit, int *value); static const char *conninfo_getval(PQconninfoOption *connOptions, const char *keyword); static PQconninfoOption *conninfo_storeval(PQconninfoOption *connOptions, - const char *keyword, const char *value, - PQExpBuffer errorMessage, bool ignoreMissing, bool uri_decode); + const char *keyword, const char *value, + PQExpBuffer errorMessage, bool ignoreMissing, bool uri_decode); static PQconninfoOption *conninfo_find(PQconninfoOption *connOptions, const char *keyword); static void defaultNoticeReceiver(void *arg, const PGresult *res); @@ -605,7 +605,7 @@ PQconnectStart(const char *conninfo) static void fillPGconn(PGconn *conn, PQconninfoOption *connOptions) { - const char *tmp; + const char *tmp; /* * Move option values into conn structure @@ -903,8 +903,8 @@ PQsetdbLogin(const char *pghost, const char *pgport, const char *pgoptions, return NULL; /* - * If the dbName parameter contains what looks like a connection - * string, parse it into conn struct using connectOptions1. + * If the dbName parameter contains what looks like a connection string, + * parse it into conn struct using connectOptions1. */ if (dbName && recognized_connection_string(dbName)) { @@ -2094,7 +2094,7 @@ keep_going: /* We will come back to here until there is { /* * Server failure of some sort, such as failure to - * fork a backend process. We need to process and + * fork a backend process. We need to process and * report the error message, which might be formatted * according to either protocol 2 or protocol 3. * Rather than duplicate the code for that, we flip @@ -4305,9 +4305,9 @@ conninfo_array_parse(const char *const * keywords, const char *const * values, if (strcmp(pname, "dbname") == 0 && pvalue) { /* - * If value is a connection string, parse it, but do not use defaults - * here -- those get picked up later. We only want to override for - * those parameters actually passed. + * If value is a connection string, parse it, but do not use + * defaults here -- those get picked up later. We only want to + * override for those parameters actually passed. */ if (recognized_connection_string(pvalue)) { @@ -4558,14 +4558,15 @@ static bool conninfo_uri_parse_options(PQconninfoOption *options, const char *uri, PQExpBuffer errorMessage) { - int prefix_len; - char *p; - char *buf = strdup(uri); /* need a modifiable copy of the input URI */ - char *start = buf; - char prevchar = '\0'; - char *user = NULL; - char *host = NULL; - bool retval = false; + int prefix_len; + char *p; + char *buf = strdup(uri); /* need a modifiable copy of the input + * URI */ + char *start = buf; + char prevchar = '\0'; + char *user = NULL; + char *host = NULL; + bool retval = false; if (buf == NULL) { @@ -4657,7 +4658,7 @@ conninfo_uri_parse_options(PQconninfoOption *options, const char *uri, if (p == host) { printfPQExpBuffer(errorMessage, - libpq_gettext("IPv6 host address may not be empty in URI: %s\n"), + libpq_gettext("IPv6 host address may not be empty in URI: %s\n"), uri); goto cleanup; } @@ -4683,8 +4684,8 @@ conninfo_uri_parse_options(PQconninfoOption *options, const char *uri, host = p; /* - * Look for port specifier (colon) or end of host specifier - * (slash), or query (question mark). + * Look for port specifier (colon) or end of host specifier (slash), + * or query (question mark). */ while (*p && *p != ':' && *p != '/' && *p != '?') ++p; @@ -4718,7 +4719,7 @@ conninfo_uri_parse_options(PQconninfoOption *options, const char *uri, if (prevchar && prevchar != '?') { - const char *dbname = ++p; /* advance past host terminator */ + const char *dbname = ++p; /* advance past host terminator */ /* Look for query parameters */ while (*p && *p != '?') @@ -4740,7 +4741,7 @@ conninfo_uri_parse_options(PQconninfoOption *options, const char *uri, if (prevchar) { - ++p; /* advance past terminator */ + ++p; /* advance past terminator */ if (!conninfo_uri_parse_params(p, options, errorMessage)) goto cleanup; @@ -4758,7 +4759,7 @@ cleanup: * Connection URI parameters parser routine * * If successful, returns true while connOptions is filled with parsed - * parameters. Otherwise, returns false and fills errorMessage appropriately. + * parameters. Otherwise, returns false and fills errorMessage appropriately. * * Destructively modifies 'params' buffer. */ @@ -4769,10 +4770,10 @@ conninfo_uri_parse_params(char *params, { while (*params) { - char *keyword = params; - char *value = NULL; - char *p = params; - bool malloced = false; + char *keyword = params; + char *value = NULL; + char *p = params; + bool malloced = false; /* * Scan the params string for '=' and '&', marking the end of keyword @@ -4796,7 +4797,7 @@ conninfo_uri_parse_params(char *params, } else if (*p == '&' || *p == '\0') { - char prevchar; + char prevchar; /* Cut off value, remember old value */ prevchar = *p; @@ -4810,6 +4811,7 @@ conninfo_uri_parse_params(char *params, params); return false; } + /* * If not at the end, advance; now pointing to start of the * next parameter, if any. @@ -4876,7 +4878,7 @@ conninfo_uri_parse_params(char *params, printfPQExpBuffer(errorMessage, libpq_gettext( - "invalid URI query parameter \"%s\"\n"), + "invalid URI query parameter \"%s\"\n"), keyword); return false; } @@ -4908,8 +4910,8 @@ conninfo_uri_parse_params(char *params, static char * conninfo_uri_decode(const char *str, PQExpBuffer errorMessage) { - char *buf = malloc(strlen(str) + 1); - char *p = buf; + char *buf = malloc(strlen(str) + 1); + char *p = buf; const char *q = str; if (buf == NULL) @@ -4928,20 +4930,20 @@ conninfo_uri_decode(const char *str, PQExpBuffer errorMessage) } else { - int hi; - int lo; - int c; + int hi; + int lo; + int c; - ++q; /* skip the percent sign itself */ + ++q; /* skip the percent sign itself */ /* - * Possible EOL will be caught by the first call to get_hexdigit(), - * so we never dereference an invalid q pointer. + * Possible EOL will be caught by the first call to + * get_hexdigit(), so we never dereference an invalid q pointer. */ if (!(get_hexdigit(*q++, &hi) && get_hexdigit(*q++, &lo))) { printfPQExpBuffer(errorMessage, - libpq_gettext("invalid percent-encoded token: %s\n"), + libpq_gettext("invalid percent-encoded token: %s\n"), str); free(buf); return NULL; @@ -5025,14 +5027,14 @@ conninfo_storeval(PQconninfoOption *connOptions, bool uri_decode) { PQconninfoOption *option; - char *value_copy; + char *value_copy; option = conninfo_find(connOptions, keyword); if (option == NULL) { if (!ignoreMissing) printfPQExpBuffer(errorMessage, - libpq_gettext("invalid connection option \"%s\"\n"), + libpq_gettext("invalid connection option \"%s\"\n"), keyword); return NULL; } diff --git a/src/interfaces/libpq/fe-exec.c b/src/interfaces/libpq/fe-exec.c index c643b8e69caf28808afaa8aa233e94fd686aac96..badc0b32a8e8f3e33e42729a8f8899475a0de31b 100644 --- a/src/interfaces/libpq/fe-exec.c +++ b/src/interfaces/libpq/fe-exec.c @@ -1052,7 +1052,7 @@ pqStdRowProcessor(PGresult *res, const PGdataValue *columns, * copy the data over. * * Note: on malloc failure, we return -1 leaving *errmsgp still NULL, - * which caller will take to mean "out of memory". This is preferable to + * which caller will take to mean "out of memory". This is preferable to * trying to set up such a message here, because evidently there's not * enough memory for gettext() to do anything. */ @@ -1063,7 +1063,7 @@ pqStdRowProcessor(PGresult *res, const PGdataValue *columns, for (i = 0; i < nfields; i++) { - int clen = columns[i].len; + int clen = columns[i].len; if (clen < 0) { @@ -1743,7 +1743,7 @@ PQgetResult(PGconn *conn) * * This is mainly useful for cleaning up after a longjmp out of a row * processor, when resuming processing of the current query result isn't - * wanted. Note that this is of little value in an async-style application, + * wanted. Note that this is of little value in an async-style application, * since any preceding calls to PQisBusy would have already called the regular * row processor. */ diff --git a/src/interfaces/libpq/fe-protocol2.c b/src/interfaces/libpq/fe-protocol2.c index 43f9954dd1c01c4a8716a3d69bff5f644a8e49ea..8dbd6b6982395850fc167eb20cb8d5590d33122f 100644 --- a/src/interfaces/libpq/fe-protocol2.c +++ b/src/interfaces/libpq/fe-protocol2.c @@ -773,10 +773,11 @@ getRowDescriptions(PGconn *conn) goto set_error_result; advance_and_error: + /* - * Discard the failed message. Unfortunately we don't know for sure - * where the end is, so just throw away everything in the input buffer. - * This is not very desirable but it's the best we can do in protocol v2. + * Discard the failed message. Unfortunately we don't know for sure where + * the end is, so just throw away everything in the input buffer. This is + * not very desirable but it's the best we can do in protocol v2. */ conn->inStart = conn->inEnd; @@ -829,6 +830,7 @@ getAnotherTuple(PGconn *conn, bool binary) int nfields = result->numAttributes; const char *errmsg; PGdataValue *rowbuf; + /* the backend sends us a bitmap of which attributes are null */ char std_bitmap[64]; /* used unless it doesn't fit */ char *bitmap = std_bitmap; @@ -962,10 +964,11 @@ getAnotherTuple(PGconn *conn, bool binary) goto set_error_result; advance_and_error: + /* - * Discard the failed message. Unfortunately we don't know for sure - * where the end is, so just throw away everything in the input buffer. - * This is not very desirable but it's the best we can do in protocol v2. + * Discard the failed message. Unfortunately we don't know for sure where + * the end is, so just throw away everything in the input buffer. This is + * not very desirable but it's the best we can do in protocol v2. */ conn->inStart = conn->inEnd; diff --git a/src/interfaces/libpq/fe-protocol3.c b/src/interfaces/libpq/fe-protocol3.c index a773d7a5246f201c2bc11b5c98433222e0b335ad..173af2e0a79ef3b443de515cda851e277deaed2d 100644 --- a/src/interfaces/libpq/fe-protocol3.c +++ b/src/interfaces/libpq/fe-protocol3.c @@ -627,7 +627,7 @@ set_error_result: pqSaveErrorResult(conn); /* - * Return zero to allow input parsing to continue. Subsequent "D" + * Return zero to allow input parsing to continue. Subsequent "D" * messages will be ignored until we get to end of data, since an error * result is already set up. */ @@ -829,7 +829,7 @@ set_error_result: pqSaveErrorResult(conn); /* - * Return zero to allow input parsing to continue. Subsequent "D" + * Return zero to allow input parsing to continue. Subsequent "D" * messages will be ignored until we get to end of data, since an error * result is already set up. */ diff --git a/src/interfaces/libpq/fe-secure.c b/src/interfaces/libpq/fe-secure.c index 5c4d73c3acf4a214185bdf6c49881e050fa74e66..b1ad776a234280e2921bf2378627070cc6005166 100644 --- a/src/interfaces/libpq/fe-secure.c +++ b/src/interfaces/libpq/fe-secure.c @@ -361,19 +361,19 @@ rloop: result_errno == ECONNRESET) printfPQExpBuffer(&conn->errorMessage, libpq_gettext( - "server closed the connection unexpectedly\n" - "\tThis probably means the server terminated abnormally\n" - "\tbefore or while processing the request.\n")); + "server closed the connection unexpectedly\n" + "\tThis probably means the server terminated abnormally\n" + "\tbefore or while processing the request.\n")); else printfPQExpBuffer(&conn->errorMessage, - libpq_gettext("SSL SYSCALL error: %s\n"), + libpq_gettext("SSL SYSCALL error: %s\n"), SOCK_STRERROR(result_errno, - sebuf, sizeof(sebuf))); + sebuf, sizeof(sebuf))); } else { printfPQExpBuffer(&conn->errorMessage, - libpq_gettext("SSL SYSCALL error: EOF detected\n")); + libpq_gettext("SSL SYSCALL error: EOF detected\n")); /* assume the connection is broken */ result_errno = ECONNRESET; n = -1; @@ -392,6 +392,7 @@ rloop: break; } case SSL_ERROR_ZERO_RETURN: + /* * Per OpenSSL documentation, this error code is only returned * for a clean connection closure, so we should not report it @@ -415,7 +416,7 @@ rloop: RESTORE_SIGPIPE(conn, spinfo); } else -#endif /* USE_SSL */ +#endif /* USE_SSL */ { n = recv(conn->sock, ptr, len, 0); @@ -440,15 +441,15 @@ rloop: case ECONNRESET: printfPQExpBuffer(&conn->errorMessage, libpq_gettext( - "server closed the connection unexpectedly\n" - "\tThis probably means the server terminated abnormally\n" - "\tbefore or while processing the request.\n")); + "server closed the connection unexpectedly\n" + "\tThis probably means the server terminated abnormally\n" + "\tbefore or while processing the request.\n")); break; #endif default: printfPQExpBuffer(&conn->errorMessage, - libpq_gettext("could not receive data from server: %s\n"), + libpq_gettext("could not receive data from server: %s\n"), SOCK_STRERROR(result_errno, sebuf, sizeof(sebuf))); break; @@ -521,19 +522,19 @@ pqsecure_write(PGconn *conn, const void *ptr, size_t len) result_errno == ECONNRESET) printfPQExpBuffer(&conn->errorMessage, libpq_gettext( - "server closed the connection unexpectedly\n" - "\tThis probably means the server terminated abnormally\n" - "\tbefore or while processing the request.\n")); + "server closed the connection unexpectedly\n" + "\tThis probably means the server terminated abnormally\n" + "\tbefore or while processing the request.\n")); else printfPQExpBuffer(&conn->errorMessage, - libpq_gettext("SSL SYSCALL error: %s\n"), + libpq_gettext("SSL SYSCALL error: %s\n"), SOCK_STRERROR(result_errno, - sebuf, sizeof(sebuf))); + sebuf, sizeof(sebuf))); } else { printfPQExpBuffer(&conn->errorMessage, - libpq_gettext("SSL SYSCALL error: EOF detected\n")); + libpq_gettext("SSL SYSCALL error: EOF detected\n")); /* assume the connection is broken */ result_errno = ECONNRESET; n = -1; @@ -552,6 +553,7 @@ pqsecure_write(PGconn *conn, const void *ptr, size_t len) break; } case SSL_ERROR_ZERO_RETURN: + /* * Per OpenSSL documentation, this error code is only returned * for a clean connection closure, so we should not report it @@ -573,7 +575,7 @@ pqsecure_write(PGconn *conn, const void *ptr, size_t len) } } else -#endif /* USE_SSL */ +#endif /* USE_SSL */ { int flags = 0; @@ -629,14 +631,14 @@ retry_masked: #endif printfPQExpBuffer(&conn->errorMessage, libpq_gettext( - "server closed the connection unexpectedly\n" - "\tThis probably means the server terminated abnormally\n" - "\tbefore or while processing the request.\n")); + "server closed the connection unexpectedly\n" + "\tThis probably means the server terminated abnormally\n" + "\tbefore or while processing the request.\n")); break; default: printfPQExpBuffer(&conn->errorMessage, - libpq_gettext("could not send data to server: %s\n"), + libpq_gettext("could not send data to server: %s\n"), SOCK_STRERROR(result_errno, sebuf, sizeof(sebuf))); break; @@ -1346,11 +1348,12 @@ initialize_SSL(PGconn *conn) } /* - * If the OpenSSL version used supports it (from 1.0.0 on) - * and the user requested it, disable SSL compression. + * If the OpenSSL version used supports it (from 1.0.0 on) and the user + * requested it, disable SSL compression. */ #ifdef SSL_OP_NO_COMPRESSION - if (conn->sslcompression && conn->sslcompression[0] == '0') { + if (conn->sslcompression && conn->sslcompression[0] == '0') + { SSL_set_options(conn->ssl, SSL_OP_NO_COMPRESSION); } #endif diff --git a/src/interfaces/libpq/libpq-fe.h b/src/interfaces/libpq/libpq-fe.h index 32b466e245cc1587fbb8e7164314d473aeac1225..67db6119bbaa35ae31fb58bb902a455b093ea23f 100644 --- a/src/interfaces/libpq/libpq-fe.h +++ b/src/interfaces/libpq/libpq-fe.h @@ -163,7 +163,7 @@ typedef struct pgNotify /* Function type for row-processor callback */ typedef int (*PQrowProcessor) (PGresult *res, const PGdataValue *columns, - const char **errmsgp, void *param); + const char **errmsgp, void *param); /* Function types for notice-handling callbacks */ typedef void (*PQnoticeReceiver) (void *arg, const PGresult *res); diff --git a/src/interfaces/libpq/libpq-int.h b/src/interfaces/libpq/libpq-int.h index 0b6e6769c018ccc48dfd97d2412a54fc3121290e..4bc89269fababe5e4d8ecbf6e80ca1a8625d4bd5 100644 --- a/src/interfaces/libpq/libpq-int.h +++ b/src/interfaces/libpq/libpq-int.h @@ -310,7 +310,7 @@ struct pg_conn char *keepalives_count; /* maximum number of TCP keepalive * retransmits */ char *sslmode; /* SSL mode (require,prefer,allow,disable) */ - char *sslcompression; /* SSL compression (0 or 1) */ + char *sslcompression; /* SSL compression (0 or 1) */ char *sslkey; /* client key filename */ char *sslcert; /* client certificate filename */ char *sslrootcert; /* root certificate filename */ @@ -326,7 +326,7 @@ struct pg_conn /* Callback procedure for per-row processing */ PQrowProcessor rowProcessor; /* function pointer */ - void *rowProcessorParam; /* passthrough argument */ + void *rowProcessorParam; /* passthrough argument */ /* Callback procedures for notice message processing */ PGNoticeHooks noticeHooks; diff --git a/src/interfaces/libpq/test/uri-regress.c b/src/interfaces/libpq/test/uri-regress.c index 17fcce9fb27a66332d1f4352327c80b18dee5529..a0ba9e4583bffc6ac80fea2757d7edea431b9914 100644 --- a/src/interfaces/libpq/test/uri-regress.c +++ b/src/interfaces/libpq/test/uri-regress.c @@ -1,8 +1,8 @@ /* * uri-regress.c - * A test program for libpq URI format + * A test program for libpq URI format * - * This is a helper for libpq conninfo regression testing. It takes a single + * This is a helper for libpq conninfo regression testing. It takes a single * conninfo string as a parameter, parses it using PQconninfoParse, and then * prints out the values from the parsed PQconninfoOption struct that differ * from the defaults (obtained from PQconndefaults). @@ -10,7 +10,7 @@ * Portions Copyright (c) 2012, PostgreSQL Global Development Group * * IDENTIFICATION - * src/interfaces/libpq/test/uri-regress.c + * src/interfaces/libpq/test/uri-regress.c */ #include "postgres_fe.h" diff --git a/src/pl/plperl/plperl.c b/src/pl/plperl/plperl.c index 7a92f3d8e28374c694ff51b27db0b0203a750e8d..db584c4e7e4743a8368cc08a5777c8377a66be69 100644 --- a/src/pl/plperl/plperl.c +++ b/src/pl/plperl/plperl.c @@ -1259,19 +1259,19 @@ plperl_sv_to_datum(SV *sv, Oid typid, int32 typmod, if (!type_is_rowtype(typid)) ereport(ERROR, (errcode(ERRCODE_DATATYPE_MISMATCH), - errmsg("cannot convert Perl hash to non-composite type %s", - format_type_be(typid)))); + errmsg("cannot convert Perl hash to non-composite type %s", + format_type_be(typid)))); td = lookup_rowtype_tupdesc_noerror(typid, typmod, true); if (td == NULL) { /* Try to look it up based on our result type */ if (fcinfo == NULL || - get_call_result_type(fcinfo, NULL, &td) != TYPEFUNC_COMPOSITE) + get_call_result_type(fcinfo, NULL, &td) != TYPEFUNC_COMPOSITE) ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("function returning record called in context " - "that cannot accept type record"))); + errmsg("function returning record called in context " + "that cannot accept type record"))); } ret = plperl_hash_to_datum(sv, td); diff --git a/src/pl/plperl/plperl_helpers.h b/src/pl/plperl/plperl_helpers.h index 6b714e52a14014dbd65ba3a399026a4000ce9b27..1b6648be1d169daf3869ca4135598af7e37d23ca 100644 --- a/src/pl/plperl/plperl_helpers.h +++ b/src/pl/plperl/plperl_helpers.h @@ -7,15 +7,15 @@ static inline char * utf_u2e(const char *utf8_str, size_t len) { - int enc = GetDatabaseEncoding(); + int enc = GetDatabaseEncoding(); char *ret = (char *) pg_do_encoding_conversion((unsigned char *) utf8_str, len, PG_UTF8, enc); /* - * when we are a PG_UTF8 or SQL_ASCII database - * pg_do_encoding_conversion() will not do any conversion or - * verification. we need to do it manually instead. - */ + * when we are a PG_UTF8 or SQL_ASCII database pg_do_encoding_conversion() + * will not do any conversion or verification. we need to do it manually + * instead. + */ if (enc == PG_UTF8 || enc == PG_SQL_ASCII) pg_verify_mbstr_len(PG_UTF8, utf8_str, len, false); @@ -45,7 +45,8 @@ utf_e2u(const char *str) static inline char * sv2cstr(SV *sv) { - char *val, *res; + char *val, + *res; STRLEN len; /* @@ -54,23 +55,26 @@ sv2cstr(SV *sv) * SvPVutf8() croaks nastily on certain things, like typeglobs and * readonly objects such as $^V. That's a perl bug - it's not supposed to * happen. To avoid crashing the backend, we make a copy of the sv before - * passing it to SvPVutf8(). The copy is garbage collected - * when we're done with it. + * passing it to SvPVutf8(). The copy is garbage collected when we're done + * with it. */ if (SvREADONLY(sv) || isGV_with_GP(sv) || (SvTYPE(sv) > SVt_PVLV && SvTYPE(sv) != SVt_PVFM)) sv = newSVsv(sv); else - /* increase the reference count so we can just SvREFCNT_dec() it when - * we are done */ + + /* + * increase the reference count so we can just SvREFCNT_dec() it when + * we are done + */ SvREFCNT_inc_simple_void(sv); val = SvPVutf8(sv, len); /* - * we use perl's length in the event we had an embedded null byte to ensure - * we error out properly + * we use perl's length in the event we had an embedded null byte to + * ensure we error out properly */ res = utf_u2e(val, len); diff --git a/src/pl/plpgsql/src/pl_comp.c b/src/pl/plpgsql/src/pl_comp.c index d43b8e0b1a149b8b602d608169c6bfc684364b6c..5d2f818dacb83b3e64673ddd7a89b6be05757ff9 100644 --- a/src/pl/plpgsql/src/pl_comp.c +++ b/src/pl/plpgsql/src/pl_comp.c @@ -77,7 +77,7 @@ typedef struct } ExceptionLabelMap; static const ExceptionLabelMap exception_label_map[] = { -#include "plerrcodes.h" /* pgrminclude ignore */ +#include "plerrcodes.h" /* pgrminclude ignore */ {NULL, 0} }; diff --git a/src/pl/plpgsql/src/pl_exec.c b/src/pl/plpgsql/src/pl_exec.c index 95e74b38dc4d9c64c656b877d0d74ac9fa165c37..8ca791ce3f42486a8dc47634397600d4b2b7c915 100644 --- a/src/pl/plpgsql/src/pl_exec.c +++ b/src/pl/plpgsql/src/pl_exec.c @@ -881,8 +881,8 @@ copy_plpgsql_datum(PLpgSQL_datum *datum) /* * These datum records are read-only at runtime, so no need to - * copy them (well, ARRAYELEM contains some cached type data, - * but we'd just as soon centralize the caching anyway) + * copy them (well, ARRAYELEM contains some cached type data, but + * we'd just as soon centralize the caching anyway) */ result = datum; break; @@ -1441,8 +1441,8 @@ exec_stmt_getdiag(PLpgSQL_execstate *estate, PLpgSQL_stmt_getdiag *stmt) */ if (stmt->is_stacked && estate->cur_error == NULL) ereport(ERROR, - (errcode(ERRCODE_STACKED_DIAGNOSTICS_ACCESSED_WITHOUT_ACTIVE_HANDLER), - errmsg("GET STACKED DIAGNOSTICS cannot be used outside an exception handler"))); + (errcode(ERRCODE_STACKED_DIAGNOSTICS_ACCESSED_WITHOUT_ACTIVE_HANDLER), + errmsg("GET STACKED DIAGNOSTICS cannot be used outside an exception handler"))); foreach(lc, stmt->diag_items) { @@ -1481,7 +1481,7 @@ exec_stmt_getdiag(PLpgSQL_execstate *estate, PLpgSQL_stmt_getdiag *stmt) case PLPGSQL_GETDIAG_RETURNED_SQLSTATE: exec_assign_c_string(estate, var, - unpack_sql_state(estate->cur_error->sqlerrcode)); + unpack_sql_state(estate->cur_error->sqlerrcode)); break; case PLPGSQL_GETDIAG_MESSAGE_TEXT: @@ -2676,8 +2676,8 @@ exec_stmt_raise(PLpgSQL_execstate *estate, PLpgSQL_stmt_raise *stmt) ReThrowError(estate->cur_error); /* oops, we're not inside a handler */ ereport(ERROR, - (errcode(ERRCODE_STACKED_DIAGNOSTICS_ACCESSED_WITHOUT_ACTIVE_HANDLER), - errmsg("RAISE without parameters cannot be used outside an exception handler"))); + (errcode(ERRCODE_STACKED_DIAGNOSTICS_ACCESSED_WITHOUT_ACTIVE_HANDLER), + errmsg("RAISE without parameters cannot be used outside an exception handler"))); } if (stmt->condname) @@ -3036,7 +3036,7 @@ exec_stmt_execsql(PLpgSQL_execstate *estate, foreach(l2, plansource->query_list) { - Query *q = (Query *) lfirst(l2); + Query *q = (Query *) lfirst(l2); Assert(IsA(q, Query)); if (q->canSetTag) @@ -3288,9 +3288,9 @@ exec_stmt_dynexecute(PLpgSQL_execstate *estate, * a functional limitation because CREATE TABLE AS is allowed. */ ereport(ERROR, - (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), - errmsg("EXECUTE of SELECT ... INTO is not implemented"), - errhint("You might want to use EXECUTE ... INTO or EXECUTE CREATE TABLE ... AS instead."))); + (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), + errmsg("EXECUTE of SELECT ... INTO is not implemented"), + errhint("You might want to use EXECUTE ... INTO or EXECUTE CREATE TABLE ... AS instead."))); break; /* Some SPI errors deserve specific error messages */ @@ -3771,8 +3771,8 @@ exec_assign_value(PLpgSQL_execstate *estate, /* * If type is by-reference, copy the new value (which is - * probably in the eval_econtext) into the procedure's - * memory context. + * probably in the eval_econtext) into the procedure's memory + * context. */ if (!var->datatype->typbyval && !*isNull) newvalue = datumCopy(newvalue, @@ -4051,7 +4051,7 @@ exec_assign_value(PLpgSQL_execstate *estate, if (!OidIsValid(elemtypoid)) ereport(ERROR, (errcode(ERRCODE_DATATYPE_MISMATCH), - errmsg("subscripted object is not an array"))); + errmsg("subscripted object is not an array"))); /* Collect needed data about the types */ arraytyplen = get_typlen(arraytypoid); @@ -4124,7 +4124,7 @@ exec_assign_value(PLpgSQL_execstate *estate, * array, either, so that's a no-op too. This is all ugly but * corresponds to the current behavior of ExecEvalArrayRef(). */ - if (arrayelem->arraytyplen > 0 && /* fixed-length array? */ + if (arrayelem->arraytyplen > 0 && /* fixed-length array? */ (oldarrayisnull || *isNull)) return; @@ -5358,7 +5358,7 @@ convert_value_to_string(PLpgSQL_execstate *estate, Datum value, Oid valtype) * * Note: the estate's eval_econtext is used for temporary storage, and may * also contain the result Datum if we have to do a conversion to a pass- - * by-reference data type. Be sure to do an exec_eval_cleanup() call when + * by-reference data type. Be sure to do an exec_eval_cleanup() call when * done with the result. * ---------- */ @@ -5708,8 +5708,8 @@ exec_simple_check_plan(PLpgSQL_expr *expr) /* * Initialize to "not simple", and remember the plan generation number we - * last checked. (If we don't get as far as obtaining a plan to check, - * we just leave expr_simple_generation set to 0.) + * last checked. (If we don't get as far as obtaining a plan to check, we + * just leave expr_simple_generation set to 0.) */ expr->expr_simple_expr = NULL; expr->expr_simple_generation = 0; @@ -5722,12 +5722,12 @@ exec_simple_check_plan(PLpgSQL_expr *expr) plansource = (CachedPlanSource *) linitial(expr->plan->plancache_list); /* - * Do some checking on the analyzed-and-rewritten form of the query. - * These checks are basically redundant with the tests in + * Do some checking on the analyzed-and-rewritten form of the query. These + * checks are basically redundant with the tests in * exec_simple_recheck_plan, but the point is to avoid building a plan if - * possible. Since this function is only - * called immediately after creating the CachedPlanSource, we need not - * worry about the query being stale. + * possible. Since this function is only called immediately after + * creating the CachedPlanSource, we need not worry about the query being + * stale. */ /* diff --git a/src/pl/plpython/plpy_cursorobject.c b/src/pl/plpython/plpy_cursorobject.c index e8240b63c901e8cefa232d0011dd5eeb17655559..910e63b19954cae697043d1641365de9b480cdc0 100644 --- a/src/pl/plpython/plpy_cursorobject.c +++ b/src/pl/plpython/plpy_cursorobject.c @@ -40,7 +40,7 @@ static PyMethodDef PLy_cursor_methods[] = { static PyTypeObject PLy_CursorType = { PyVarObject_HEAD_INIT(NULL, 0) - "PLyCursor", /* tp_name */ + "PLyCursor", /* tp_name */ sizeof(PLyCursorObject), /* tp_size */ 0, /* tp_itemsize */ @@ -103,7 +103,7 @@ PLy_cursor(PyObject *self, PyObject *args) static PyObject * PLy_cursor_query(const char *query) { - PLyCursorObject *cursor; + PLyCursorObject *cursor; volatile MemoryContext oldcontext; volatile ResourceOwner oldowner; @@ -120,7 +120,7 @@ PLy_cursor_query(const char *query) PG_TRY(); { - PLyExecutionContext *exec_ctx = PLy_current_execution_context(); + PLyExecutionContext *exec_ctx = PLy_current_execution_context(); SPIPlanPtr plan; Portal portal; @@ -157,7 +157,7 @@ PLy_cursor_query(const char *query) static PyObject * PLy_cursor_plan(PyObject *ob, PyObject *args) { - PLyCursorObject *cursor; + PLyCursorObject *cursor; volatile int nargs; int i; PLyPlanObject *plan; @@ -187,8 +187,8 @@ PLy_cursor_plan(PyObject *ob, PyObject *args) PLy_elog(ERROR, "could not execute plan"); sv = PyString_AsString(so); PLy_exception_set_plural(PyExc_TypeError, - "Expected sequence of %d argument, got %d: %s", - "Expected sequence of %d arguments, got %d: %s", + "Expected sequence of %d argument, got %d: %s", + "Expected sequence of %d arguments, got %d: %s", plan->nargs, plan->nargs, nargs, sv); Py_DECREF(so); @@ -305,7 +305,7 @@ static void PLy_cursor_dealloc(PyObject *arg) { PLyCursorObject *cursor; - Portal portal; + Portal portal; cursor = (PLyCursorObject *) arg; @@ -328,10 +328,10 @@ static PyObject * PLy_cursor_iternext(PyObject *self) { PLyCursorObject *cursor; - PyObject *ret; + PyObject *ret; volatile MemoryContext oldcontext; volatile ResourceOwner oldowner; - Portal portal; + Portal portal; cursor = (PLyCursorObject *) self; @@ -391,11 +391,11 @@ static PyObject * PLy_cursor_fetch(PyObject *self, PyObject *args) { PLyCursorObject *cursor; - int count; - PLyResultObject *ret; + int count; + PLyResultObject *ret; volatile MemoryContext oldcontext; volatile ResourceOwner oldowner; - Portal portal; + Portal portal; if (!PyArg_ParseTuple(args, "i", &count)) return NULL; @@ -440,7 +440,7 @@ PLy_cursor_fetch(PyObject *self, PyObject *args) if (SPI_processed != 0) { - int i; + int i; Py_DECREF(ret->rows); ret->rows = PyList_New(SPI_processed); @@ -450,6 +450,7 @@ PLy_cursor_fetch(PyObject *self, PyObject *args) PyObject *row = PLyDict_FromTuple(&cursor->result, SPI_tuptable->vals[i], SPI_tuptable->tupdesc); + PyList_SetItem(ret->rows, i, row); } } @@ -477,12 +478,12 @@ PLy_cursor_close(PyObject *self, PyObject *unused) if (!cursor->closed) { - Portal portal = GetPortalByName(cursor->portalname); + Portal portal = GetPortalByName(cursor->portalname); if (!PortalIsValid(portal)) { PLy_exception_set(PyExc_ValueError, - "closing a cursor in an aborted subtransaction"); + "closing a cursor in an aborted subtransaction"); return NULL; } diff --git a/src/pl/plpython/plpy_cursorobject.h b/src/pl/plpython/plpy_cursorobject.h index 1dd9d48fd58c5049ea0a3e5fdf64cb6f1149a693..3c28f4f8e716eb2f97704447ae62ff20ea789f69 100644 --- a/src/pl/plpython/plpy_cursorobject.h +++ b/src/pl/plpython/plpy_cursorobject.h @@ -11,7 +11,7 @@ typedef struct PLyCursorObject { PyObject_HEAD - char *portalname; + char *portalname; PLyTypeInfo result; bool closed; } PLyCursorObject; @@ -19,4 +19,4 @@ typedef struct PLyCursorObject extern void PLy_cursor_init_type(void); extern PyObject *PLy_cursor(PyObject *self, PyObject *args); -#endif /* PLPY_CURSOROBJECT_H */ +#endif /* PLPY_CURSOROBJECT_H */ diff --git a/src/pl/plpython/plpy_elog.c b/src/pl/plpython/plpy_elog.c index f7d321289d4acec776dca8da684f08e6532fb157..c375ac07fa8f12a2f7189a702e777fba956933e3 100644 --- a/src/pl/plpython/plpy_elog.c +++ b/src/pl/plpython/plpy_elog.c @@ -16,15 +16,15 @@ #include "plpy_procedure.h" -PyObject *PLy_exc_error = NULL; -PyObject *PLy_exc_fatal = NULL; -PyObject *PLy_exc_spi_error = NULL; +PyObject *PLy_exc_error = NULL; +PyObject *PLy_exc_fatal = NULL; +PyObject *PLy_exc_spi_error = NULL; static void PLy_traceback(char **xmsg, char **tbmsg, int *tb_depth); static void PLy_get_spi_error_data(PyObject *exc, int *sqlerrcode, char **detail, - char **hint, char **query, int *position); -static char * get_source_line(const char *src, int lineno); + char **hint, char **query, int *position); +static char *get_source_line(const char *src, int lineno); /* @@ -46,7 +46,7 @@ PLy_elog(int elevel, const char *fmt,...) *val, *tb; const char *primary = NULL; - int sqlerrcode = 0; + int sqlerrcode = 0; char *detail = NULL; char *hint = NULL; char *query = NULL; @@ -98,7 +98,7 @@ PLy_elog(int elevel, const char *fmt,...) { ereport(elevel, (errcode(sqlerrcode ? sqlerrcode : ERRCODE_INTERNAL_ERROR), - errmsg_internal("%s", primary ? primary : "no exception data"), + errmsg_internal("%s", primary ? primary : "no exception data"), (detail) ? errdetail_internal("%s", detail) : 0, (tb_depth > 0 && tbmsg) ? errcontext("%s", tbmsg) : 0, (hint) ? errhint("%s", hint) : 0, @@ -256,7 +256,7 @@ PLy_traceback(char **xmsg, char **tbmsg, int *tb_depth) /* The first frame always points at <module>, skip it. */ if (*tb_depth > 0) { - PLyExecutionContext *exec_ctx = PLy_current_execution_context(); + PLyExecutionContext *exec_ctx = PLy_current_execution_context(); char *proname; char *fname; char *line; diff --git a/src/pl/plpython/plpy_elog.h b/src/pl/plpython/plpy_elog.h index f7223b005683561ee651691a93314967ff98e7c6..6b8d485625be54d9bdc6ff120bdf48bea3da0d05 100644 --- a/src/pl/plpython/plpy_elog.h +++ b/src/pl/plpython/plpy_elog.h @@ -10,15 +10,18 @@ extern PyObject *PLy_exc_error; extern PyObject *PLy_exc_fatal; extern PyObject *PLy_exc_spi_error; -extern void PLy_elog(int elevel, const char *fmt,...) +extern void +PLy_elog(int elevel, const char *fmt,...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); -extern void PLy_exception_set(PyObject *exc, const char *fmt,...) +extern void +PLy_exception_set(PyObject *exc, const char *fmt,...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 3))); -extern void PLy_exception_set_plural(PyObject *exc, const char *fmt_singular, const char *fmt_plural, - unsigned long n,...) +extern void +PLy_exception_set_plural(PyObject *exc, const char *fmt_singular, const char *fmt_plural, + unsigned long n,...) __attribute__((format(PG_PRINTF_ATTRIBUTE, 2, 5))) __attribute__((format(PG_PRINTF_ATTRIBUTE, 3, 5))); -#endif /* PLPY_ELOG_H */ +#endif /* PLPY_ELOG_H */ diff --git a/src/pl/plpython/plpy_exec.c b/src/pl/plpython/plpy_exec.c index ad30fc0065bf27672f1e056dab0ac8a9d5af3edf..96ee26c35c7327d95a9f434dbf4eb55071db2070 100644 --- a/src/pl/plpython/plpy_exec.c +++ b/src/pl/plpython/plpy_exec.c @@ -30,9 +30,9 @@ static void PLy_function_delete_args(PLyProcedure *proc); static void plpython_return_error_callback(void *arg); static PyObject *PLy_trigger_build_args(FunctionCallInfo fcinfo, PLyProcedure *proc, - HeapTuple *rv); + HeapTuple *rv); static HeapTuple PLy_modify_tuple(PLyProcedure *proc, PyObject *pltd, - TriggerData *tdata, HeapTuple otup); + TriggerData *tdata, HeapTuple otup); static void plpython_trigger_error_callback(void *arg); static PyObject *PLy_procedure_call(PLyProcedure *proc, char *kargs, PyObject *vargs); @@ -180,7 +180,7 @@ PLy_exec_function(FunctionCallInfo fcinfo, PLyProcedure *proc) } else if (proc->result.is_rowtype >= 1) { - TupleDesc desc; + TupleDesc desc; /* make sure it's not an unnamed record */ Assert((proc->result.out.d.typoid == RECORDOID && diff --git a/src/pl/plpython/plpy_exec.h b/src/pl/plpython/plpy_exec.h index f3dec074c137da0eb3d17f2c7a6bb55e75528a22..439a1d801f6942994ae74dfa6db4b6eaeef83a3d 100644 --- a/src/pl/plpython/plpy_exec.h +++ b/src/pl/plpython/plpy_exec.h @@ -10,4 +10,4 @@ extern Datum PLy_exec_function(FunctionCallInfo fcinfo, PLyProcedure *proc); extern HeapTuple PLy_exec_trigger(FunctionCallInfo fcinfo, PLyProcedure *proc); -#endif /* PLPY_EXEC_H */ +#endif /* PLPY_EXEC_H */ diff --git a/src/pl/plpython/plpy_main.c b/src/pl/plpython/plpy_main.c index c126db995ae3470a405d4f1d96a7a3c638be6937..494ec37ea7cfada29752f2f75b9225880b9aae15 100644 --- a/src/pl/plpython/plpy_main.c +++ b/src/pl/plpython/plpy_main.c @@ -73,7 +73,7 @@ static void PLy_pop_execution_context(void); static const int plpython_python_version = PY_MAJOR_VERSION; /* initialize global variables */ -PyObject *PLy_interp_globals = NULL; +PyObject *PLy_interp_globals = NULL; /* this doesn't need to be global; use PLy_current_execution_context() */ static PLyExecutionContext *PLy_execution_contexts = NULL; @@ -284,8 +284,8 @@ plpython_inline_handler(PG_FUNCTION_ARGS) * Push execution context onto stack. It is important that this get * popped again, so avoid putting anything that could throw error between * here and the PG_TRY. (plpython_inline_error_callback doesn't currently - * need the stack entry, but for consistency with plpython_call_handler - * we do it in this order.) + * need the stack entry, but for consistency with plpython_call_handler we + * do it in this order.) */ exec_ctx = PLy_push_execution_context(); @@ -330,7 +330,8 @@ plpython2_inline_handler(PG_FUNCTION_ARGS) } #endif /* PY_MAJOR_VERSION < 3 */ -static bool PLy_procedure_is_trigger(Form_pg_proc procStruct) +static bool +PLy_procedure_is_trigger(Form_pg_proc procStruct) { return (procStruct->prorettype == TRIGGEROID || (procStruct->prorettype == OPAQUEOID && @@ -365,7 +366,7 @@ PLy_current_execution_context(void) static PLyExecutionContext * PLy_push_execution_context(void) { - PLyExecutionContext *context = PLy_malloc(sizeof(PLyExecutionContext)); + PLyExecutionContext *context = PLy_malloc(sizeof(PLyExecutionContext)); context->curr_proc = NULL; context->scratch_ctx = AllocSetContextCreate(TopTransactionContext, @@ -381,7 +382,7 @@ PLy_push_execution_context(void) static void PLy_pop_execution_context(void) { - PLyExecutionContext *context = PLy_execution_contexts; + PLyExecutionContext *context = PLy_execution_contexts; if (context == NULL) elog(ERROR, "no Python function is currently executing"); diff --git a/src/pl/plpython/plpy_main.h b/src/pl/plpython/plpy_main.h index cb214bf83c77c7e9173a3da98b1ca437eada1b77..b13e2c21a111b045f61fdc4da23ba583003e0ed4 100644 --- a/src/pl/plpython/plpy_main.h +++ b/src/pl/plpython/plpy_main.h @@ -17,12 +17,12 @@ extern PyObject *PLy_interp_globals; */ typedef struct PLyExecutionContext { - PLyProcedure *curr_proc; /* the currently executing procedure */ - MemoryContext scratch_ctx; /* a context for things like type I/O */ + PLyProcedure *curr_proc; /* the currently executing procedure */ + MemoryContext scratch_ctx; /* a context for things like type I/O */ struct PLyExecutionContext *next; /* previous stack level */ } PLyExecutionContext; /* Get the current execution context */ extern PLyExecutionContext *PLy_current_execution_context(void); -#endif /* PLPY_MAIN_H */ +#endif /* PLPY_MAIN_H */ diff --git a/src/pl/plpython/plpy_planobject.h b/src/pl/plpython/plpy_planobject.h index febc5c25ef61caa3c4ad5a22a6485abc18f99caa..7a89ffc2c18090355b500434545b7555c92ccec2 100644 --- a/src/pl/plpython/plpy_planobject.h +++ b/src/pl/plpython/plpy_planobject.h @@ -23,4 +23,4 @@ extern void PLy_plan_init_type(void); extern PyObject *PLy_plan_new(void); extern bool is_PLyPlanObject(PyObject *ob); -#endif /* PLPY_PLANOBJECT_H */ +#endif /* PLPY_PLANOBJECT_H */ diff --git a/src/pl/plpython/plpy_plpymodule.c b/src/pl/plpython/plpy_plpymodule.c index bc0b9e6f8423bec53bf15f8d0c2687ed8c6ed295..37ea2a490d963eecdbebd07889ad9a4e0aa0f7c1 100644 --- a/src/pl/plpython/plpy_plpymodule.c +++ b/src/pl/plpython/plpy_plpymodule.c @@ -21,7 +21,7 @@ #include "plpy_subxactobject.h" -HTAB *PLy_spi_exceptions = NULL; +HTAB *PLy_spi_exceptions = NULL; static void PLy_add_exceptions(PyObject *plpy); @@ -137,7 +137,7 @@ PyInit_plpy(void) return m; } -#endif /* PY_MAJOR_VERSION >= 3 */ +#endif /* PY_MAJOR_VERSION >= 3 */ void PLy_init_plpy(void) @@ -145,6 +145,7 @@ PLy_init_plpy(void) PyObject *main_mod, *main_dict, *plpy_mod; + #if PY_MAJOR_VERSION < 3 PyObject *plpy; #endif diff --git a/src/pl/plpython/plpy_plpymodule.h b/src/pl/plpython/plpy_plpymodule.h index 930ecfd1b1728848dba527ab52ff2626501fd27a..ee089b78a167078b3a85c92407d4221cf89812c0 100644 --- a/src/pl/plpython/plpy_plpymodule.h +++ b/src/pl/plpython/plpy_plpymodule.h @@ -16,4 +16,4 @@ PyMODINIT_FUNC PyInit_plpy(void); #endif extern void PLy_init_plpy(void); -#endif /* PLPY_PLPYMODULE_H */ +#endif /* PLPY_PLPYMODULE_H */ diff --git a/src/pl/plpython/plpy_procedure.h b/src/pl/plpython/plpy_procedure.h index c7405e064ec1b8e6646c0d515f610c4d5295551e..40a0314cdfb8e5f271df378d4de606c3ccad9815 100644 --- a/src/pl/plpython/plpy_procedure.h +++ b/src/pl/plpython/plpy_procedure.h @@ -45,4 +45,4 @@ extern PLyProcedure *PLy_procedure_get(Oid fn_oid, bool is_trigger); extern void PLy_procedure_compile(PLyProcedure *proc, const char *src); extern void PLy_procedure_delete(PLyProcedure *proc); -#endif /* PLPY_PROCEDURE_H */ +#endif /* PLPY_PROCEDURE_H */ diff --git a/src/pl/plpython/plpy_resultobject.c b/src/pl/plpython/plpy_resultobject.c index deaddb7980fb5f4884bf879e6cbef0917e8f8e8b..6b01e717c47024ee057a9c65dfe03e0a5efe316e 100644 --- a/src/pl/plpython/plpy_resultobject.c +++ b/src/pl/plpython/plpy_resultobject.c @@ -24,7 +24,7 @@ static PyObject *PLy_result_slice(PyObject *arg, Py_ssize_t lidx, Py_ssize_t hid static int PLy_result_ass_item(PyObject *arg, Py_ssize_t idx, PyObject *item); static int PLy_result_ass_slice(PyObject *rg, Py_ssize_t lidx, Py_ssize_t hidx, PyObject *slice); static PyObject *PLy_result_subscript(PyObject *arg, PyObject *item); -static int PLy_result_ass_subscript(PyObject* self, PyObject* item, PyObject* value); +static int PLy_result_ass_subscript(PyObject *self, PyObject *item, PyObject *value); static char PLy_result_doc[] = { "Results of a PostgreSQL query" @@ -263,7 +263,7 @@ PLy_result_ass_slice(PyObject *arg, Py_ssize_t lidx, Py_ssize_t hidx, PyObject * static PyObject * PLy_result_subscript(PyObject *arg, PyObject *item) { - PLyResultObject *ob = (PLyResultObject *) arg; + PLyResultObject *ob = (PLyResultObject *) arg; return PyObject_GetItem(ob->rows, item); } @@ -271,7 +271,7 @@ PLy_result_subscript(PyObject *arg, PyObject *item) static int PLy_result_ass_subscript(PyObject *arg, PyObject *item, PyObject *value) { - PLyResultObject *ob = (PLyResultObject *) arg; + PLyResultObject *ob = (PLyResultObject *) arg; return PyObject_SetItem(ob->rows, item, value); } diff --git a/src/pl/plpython/plpy_resultobject.h b/src/pl/plpython/plpy_resultobject.h index c5ba99988747a750922bcbf7b7a561c8a4a5bbdd..314510c40fffacd1aa3ec2151aaa90806297ff1e 100644 --- a/src/pl/plpython/plpy_resultobject.h +++ b/src/pl/plpython/plpy_resultobject.h @@ -13,7 +13,8 @@ typedef struct PLyResultObject PyObject_HEAD /* HeapTuple *tuples; */ PyObject *nrows; /* number of rows returned by query */ - PyObject *rows; /* data rows, or empty list if no data returned */ + PyObject *rows; /* data rows, or empty list if no data + * returned */ PyObject *status; /* query status, SPI_OK_*, or SPI_ERR_* */ TupleDesc tupdesc; } PLyResultObject; @@ -21,4 +22,4 @@ typedef struct PLyResultObject extern void PLy_result_init_type(void); extern PyObject *PLy_result_new(void); -#endif /* PLPY_RESULTOBJECT_H */ +#endif /* PLPY_RESULTOBJECT_H */ diff --git a/src/pl/plpython/plpy_spi.c b/src/pl/plpython/plpy_spi.c index 4bc3d96d5891408738c139a812d8faf59e2da8c1..00156e6658ec4509b0660eef03be3566d3d9a625 100644 --- a/src/pl/plpython/plpy_spi.c +++ b/src/pl/plpython/plpy_spi.c @@ -350,7 +350,7 @@ PLy_spi_execute_query(char *query, long limit) PG_TRY(); { - PLyExecutionContext *exec_ctx = PLy_current_execution_context(); + PLyExecutionContext *exec_ctx = PLy_current_execution_context(); pg_verifymbstr(query, strlen(query), false); rv = SPI_execute(query, exec_ctx->curr_proc->fn_readonly, limit); @@ -456,22 +456,22 @@ PLy_spi_execute_fetch_result(SPITupleTable *tuptable, int rows, int status) * * Usage: * - * MemoryContext oldcontext = CurrentMemoryContext; - * ResourceOwner oldowner = CurrentResourceOwner; + * MemoryContext oldcontext = CurrentMemoryContext; + * ResourceOwner oldowner = CurrentResourceOwner; * - * PLy_spi_subtransaction_begin(oldcontext, oldowner); - * PG_TRY(); - * { - * <call SPI functions> - * PLy_spi_subtransaction_commit(oldcontext, oldowner); - * } - * PG_CATCH(); - * { - * <do cleanup> - * PLy_spi_subtransaction_abort(oldcontext, oldowner); - * return NULL; - * } - * PG_END_TRY(); + * PLy_spi_subtransaction_begin(oldcontext, oldowner); + * PG_TRY(); + * { + * <call SPI functions> + * PLy_spi_subtransaction_commit(oldcontext, oldowner); + * } + * PG_CATCH(); + * { + * <do cleanup> + * PLy_spi_subtransaction_abort(oldcontext, oldowner); + * return NULL; + * } + * PG_END_TRY(); * * These utilities take care of restoring connection to the SPI manager and * setting a Python exception in case of an abort. @@ -493,8 +493,8 @@ PLy_spi_subtransaction_commit(MemoryContext oldcontext, ResourceOwner oldowner) CurrentResourceOwner = oldowner; /* - * AtEOSubXact_SPI() should not have popped any SPI context, but just - * in case it did, make sure we remain connected. + * AtEOSubXact_SPI() should not have popped any SPI context, but just in + * case it did, make sure we remain connected. */ SPI_restore_connection(); } @@ -517,8 +517,8 @@ PLy_spi_subtransaction_abort(MemoryContext oldcontext, ResourceOwner oldowner) CurrentResourceOwner = oldowner; /* - * If AtEOSubXact_SPI() popped any SPI context of the subxact, it will have - * left us in a disconnected state. We need this hack to return to + * If AtEOSubXact_SPI() popped any SPI context of the subxact, it will + * have left us in a disconnected state. We need this hack to return to * connected state. */ SPI_restore_connection(); diff --git a/src/pl/plpython/plpy_spi.h b/src/pl/plpython/plpy_spi.h index f8d31638ec8209ab84c147ff867b90e097d36d98..b0427947ef42541efeb5de33f3cc51f514e2c048 100644 --- a/src/pl/plpython/plpy_spi.h +++ b/src/pl/plpython/plpy_spi.h @@ -22,4 +22,4 @@ extern void PLy_spi_subtransaction_begin(MemoryContext oldcontext, ResourceOwner extern void PLy_spi_subtransaction_commit(MemoryContext oldcontext, ResourceOwner oldowner); extern void PLy_spi_subtransaction_abort(MemoryContext oldcontext, ResourceOwner oldowner); -#endif /* PLPY_SPI_H */ +#endif /* PLPY_SPI_H */ diff --git a/src/pl/plpython/plpy_subxactobject.c b/src/pl/plpython/plpy_subxactobject.c index 9feeddb7231fc3338440c1c6043cea11e1523138..2e7ec4fdab49945c3e2edc5e2cb8c934c4dab1df 100644 --- a/src/pl/plpython/plpy_subxactobject.c +++ b/src/pl/plpython/plpy_subxactobject.c @@ -16,7 +16,7 @@ #include "plpy_elog.h" -List *explicit_subtransactions = NIL; +List *explicit_subtransactions = NIL; static void PLy_subtransaction_dealloc(PyObject *subxact); diff --git a/src/pl/plpython/plpy_subxactobject.h b/src/pl/plpython/plpy_subxactobject.h index 7e3002fc2fd039414c0cdadad720c39b689d7f0f..b8591c7bf07336c4a93de2b6009ccf8d030fd454 100644 --- a/src/pl/plpython/plpy_subxactobject.h +++ b/src/pl/plpython/plpy_subxactobject.h @@ -26,4 +26,4 @@ typedef struct PLySubtransactionData extern void PLy_subtransaction_init_type(void); extern PyObject *PLy_subtransaction_new(PyObject *self, PyObject *unused); -#endif /* PLPY_SUBXACTOBJECT */ +#endif /* PLPY_SUBXACTOBJECT */ diff --git a/src/pl/plpython/plpy_typeio.c b/src/pl/plpython/plpy_typeio.c index c5f6c4e5a3e55cc6fe319d5f01f6cb521230497a..2cc7bbbd4deb343dec49f8e563ca50f9afed5493 100644 --- a/src/pl/plpython/plpy_typeio.c +++ b/src/pl/plpython/plpy_typeio.c @@ -293,8 +293,8 @@ PLyDict_FromTuple(PLyTypeInfo *info, HeapTuple tuple, TupleDesc desc) PG_TRY(); { /* - * Do the work in the scratch context to avoid leaking memory from - * the datatype output function calls. + * Do the work in the scratch context to avoid leaking memory from the + * datatype output function calls. */ MemoryContextSwitchTo(exec_ctx->scratch_ctx); for (i = 0; i < info->in.r.natts; i++) @@ -341,7 +341,7 @@ PLyDict_FromTuple(PLyTypeInfo *info, HeapTuple tuple, TupleDesc desc) Datum PLyObject_ToCompositeDatum(PLyTypeInfo *info, TupleDesc desc, PyObject *plrv) { - Datum datum; + Datum datum; if (PyString_Check(plrv) || PyUnicode_Check(plrv)) datum = PLyString_ToComposite(info, desc, plrv); diff --git a/src/pl/plpython/plpy_typeio.h b/src/pl/plpython/plpy_typeio.h index 11532b8c2012a29fceb655160119ac7d071f7d80..d2dfa66e0b2fea60274b3ca8323b976c826c867d 100644 --- a/src/pl/plpython/plpy_typeio.h +++ b/src/pl/plpython/plpy_typeio.h @@ -104,4 +104,4 @@ extern Datum PLyObject_ToCompositeDatum(PLyTypeInfo *info, TupleDesc desc, PyObj /* conversion from heap tuples to Python dictionaries */ extern PyObject *PLyDict_FromTuple(PLyTypeInfo *info, HeapTuple tuple, TupleDesc desc); -#endif /* PLPY_TYPEIO_H */ +#endif /* PLPY_TYPEIO_H */ diff --git a/src/pl/plpython/plpy_util.c b/src/pl/plpython/plpy_util.c index 414b9d5445ae5360125206537980f3636df9146f..9a4901ecb2f5e2de38b5e9bc781a7bbe6f62d9ea 100644 --- a/src/pl/plpython/plpy_util.c +++ b/src/pl/plpython/plpy_util.c @@ -122,4 +122,5 @@ PLyUnicode_FromString(const char *s) return o; } + #endif /* PY_MAJOR_VERSION >= 3 */ diff --git a/src/pl/plpython/plpy_util.h b/src/pl/plpython/plpy_util.h index 9b9eca0050c003ddec809ae6a271325f8bfb78b7..f93e8379fb2e58ac56ec40ab6d3543f19e7240b4 100644 --- a/src/pl/plpython/plpy_util.h +++ b/src/pl/plpython/plpy_util.h @@ -18,4 +18,4 @@ extern char *PLyUnicode_AsString(PyObject *unicode); extern PyObject *PLyUnicode_FromString(const char *s); #endif -#endif /* PLPY_UTIL_H */ +#endif /* PLPY_UTIL_H */ diff --git a/src/pl/plpython/plpython.h b/src/pl/plpython/plpython.h index 15ec85e8057db70f2322b8a57125faa3b9347759..e788cd9a897f11bd0185250ff2ca51ea583b8c1a 100644 --- a/src/pl/plpython/plpython.h +++ b/src/pl/plpython/plpython.h @@ -132,11 +132,11 @@ typedef int Py_ssize_t; #undef vsnprintf #endif #ifdef __GNUC__ -#define vsnprintf(...) pg_vsnprintf(__VA_ARGS__) -#define snprintf(...) pg_snprintf(__VA_ARGS__) +#define vsnprintf(...) pg_vsnprintf(__VA_ARGS__) +#define snprintf(...) pg_snprintf(__VA_ARGS__) #else -#define vsnprintf pg_vsnprintf -#define snprintf pg_snprintf +#define vsnprintf pg_vsnprintf +#define snprintf pg_snprintf #endif /* __GNUC__ */ #endif /* USE_REPL_SNPRINTF */ diff --git a/src/port/erand48.c b/src/port/erand48.c index 9d471197c354056c8903a5a9e1c6b0023419f1d1..524911edd12fe409cd878e29255a43c7af401d45 100644 --- a/src/port/erand48.c +++ b/src/port/erand48.c @@ -5,7 +5,7 @@ * This file supplies pg_erand48(), pg_lrand48(), and pg_srand48(), which * are just like erand48(), lrand48(), and srand48() except that we use * our own implementation rather than the one provided by the operating - * system. We used to test for an operating system version rather than + * system. We used to test for an operating system version rather than * unconditionally using our own, but (1) some versions of Cygwin have a * buggy erand48() that always returns zero and (2) as of 2011, glibc's * erand48() is strangely coded to be almost-but-not-quite thread-safe, diff --git a/src/port/fls.c b/src/port/fls.c index 4a2d6737cf930244985c4ffc8eef428efae1d1ef..8be2c51ef3d28c79ec3abe6891adf4dfc3499183 100644 --- a/src/port/fls.c +++ b/src/port/fls.c @@ -10,7 +10,7 @@ * src/port/fls.c * * This file was taken from FreeBSD to provide an implementation of fls() - * for platforms that lack it. Note that the operating system's version may + * for platforms that lack it. Note that the operating system's version may * be substantially more efficient than ours, since some platforms have an * assembly instruction that does exactly this. * @@ -25,18 +25,18 @@ * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. + * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in the - * documentation and/or other materials provided with the distribution. + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. * 4. Neither the name of the University nor the names of its contributors - * may be used to endorse or promote products derived from this software - * without specific prior written permission. + * may be used to endorse or promote products derived from this software + * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE - * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE + * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) @@ -54,11 +54,11 @@ int fls(int mask) { - int bit; + int bit; if (mask == 0) return (0); for (bit = 1; mask != 1; bit++) - mask = (unsigned int)mask >> 1; + mask = (unsigned int) mask >> 1; return (bit); } diff --git a/src/port/getaddrinfo.c b/src/port/getaddrinfo.c index 579d855648035b0ed3e3d8f1721b01c042c78dbd..c117012ec7eadf8ea87c254a55c08a4372e3d9a2 100644 --- a/src/port/getaddrinfo.c +++ b/src/port/getaddrinfo.c @@ -328,7 +328,7 @@ gai_strerror(int errcode) case EAI_MEMORY: return "Not enough memory"; #endif -#if defined(EAI_NODATA) && EAI_NODATA != EAI_NONAME /* MSVC/WIN64 duplicate */ +#if defined(EAI_NODATA) && EAI_NODATA != EAI_NONAME /* MSVC/WIN64 duplicate */ case EAI_NODATA: return "No host data of that type was found"; #endif diff --git a/src/port/path.c b/src/port/path.c index be55e4af60d7c3866d4c853805c7b380a545b9ac..738b5cc547c183eff96b56de82063616bcab0397 100644 --- a/src/port/path.c +++ b/src/port/path.c @@ -212,7 +212,7 @@ join_path_components(char *ret_path, } if (*tail) snprintf(ret_path + strlen(ret_path), MAXPGPATH - strlen(ret_path), - /* only add slash if there is something already in head */ + /* only add slash if there is something already in head */ "%s%s", head[0] ? "/" : "", tail); } diff --git a/src/port/win32setlocale.c b/src/port/win32setlocale.c index f8b1762371764f2db8e4a609c29d2038990dfe87..844891df537089e2d222c5b5e2f28e008fcae30e 100644 --- a/src/port/win32setlocale.c +++ b/src/port/win32setlocale.c @@ -27,12 +27,11 @@ struct locale_map { - const char *locale_name_part; /* string in locale name to replace */ - const char *replacement; /* string to replace it with */ + const char *locale_name_part; /* string in locale name to replace */ + const char *replacement; /* string to replace it with */ }; static const struct locale_map locale_map_list[] = { - /* * "HKG" is listed here: * http://msdn.microsoft.com/en-us/library/cdax410z%28v=vs.71%29.aspx @@ -41,26 +40,26 @@ static const struct locale_map locale_map_list[] = { * "ARE" is the ISO-3166 three-letter code for U.A.E. It is not on the * above list, but seems to work anyway. */ - { "Hong Kong S.A.R.", "HKG" }, - { "U.A.E.", "ARE" }, + {"Hong Kong S.A.R.", "HKG"}, + {"U.A.E.", "ARE"}, /* * The ISO-3166 country code for Macau S.A.R. is MAC, but Windows doesn't - * seem to recognize that. And Macau isn't listed in the table of - * accepted abbreviations linked above. Fortunately, "ZHM" seems to be - * accepted as an alias for "Chinese (Traditional)_Macau S.A.R..950". I'm - * not sure where "ZHM" comes from, must be some legacy naming scheme. But - * hey, it works. + * seem to recognize that. And Macau isn't listed in the table of accepted + * abbreviations linked above. Fortunately, "ZHM" seems to be accepted as + * an alias for "Chinese (Traditional)_Macau S.A.R..950". I'm not sure + * where "ZHM" comes from, must be some legacy naming scheme. But hey, it + * works. * * Note that unlike HKG and ARE, ZHM is an alias for the *whole* locale * name, not just the country part. * * Some versions of Windows spell it "Macau", others "Macao". */ - { "Chinese (Traditional)_Macau S.A.R..950", "ZHM" }, - { "Chinese_Macau S.A.R..950", "ZHM" }, - { "Chinese (Traditional)_Macao S.A.R..950", "ZHM" }, - { "Chinese_Macao S.A.R..950", "ZHM" } + {"Chinese (Traditional)_Macau S.A.R..950", "ZHM"}, + {"Chinese_Macau S.A.R..950", "ZHM"}, + {"Chinese (Traditional)_Macao S.A.R..950", "ZHM"}, + {"Chinese_Macao S.A.R..950", "ZHM"} }; char * @@ -85,10 +84,10 @@ pgwin32_setlocale(int category, const char *locale) if (match != NULL) { /* Found a match. Replace the matched string. */ - int matchpos = match - locale; - int replacementlen = strlen(replacement); - char *rest = match + strlen(needle); - int restlen = strlen(rest); + int matchpos = match - locale; + int replacementlen = strlen(replacement); + char *rest = match + strlen(needle); + int restlen = strlen(rest); alias = malloc(matchpos + replacementlen + restlen + 1); if (!alias) @@ -96,7 +95,8 @@ pgwin32_setlocale(int category, const char *locale) memcpy(&alias[0], &locale[0], matchpos); memcpy(&alias[matchpos], replacement, replacementlen); - memcpy(&alias[matchpos + replacementlen], rest, restlen + 1); /* includes null terminator */ + memcpy(&alias[matchpos + replacementlen], rest, restlen + 1); /* includes null + * terminator */ break; } diff --git a/src/test/isolation/isolationtester.c b/src/test/isolation/isolationtester.c index 0e681639ba19753606a8783ae1a1d2ba3c5143b3..98f89da6bff842bfbc21be3f92691561c9b23f3a 100644 --- a/src/test/isolation/isolationtester.c +++ b/src/test/isolation/isolationtester.c @@ -18,7 +18,6 @@ #ifdef HAVE_GETOPT_H #include <getopt.h> #endif - #else int getopt(int argc, char *const argv[], const char *optstring); #endif /* ! WIN32 */ @@ -47,16 +46,16 @@ static int nconns = 0; /* In dry run only output permutations to be run by the tester. */ static int dry_run = false; -static void run_testspec(TestSpec *testspec); +static void run_testspec(TestSpec * testspec); static void run_all_permutations(TestSpec * testspec); static void run_all_permutations_recurse(TestSpec * testspec, int nsteps, Step ** steps); static void run_named_permutations(TestSpec * testspec); static void run_permutation(TestSpec * testspec, int nsteps, Step ** steps); -#define STEP_NONBLOCK 0x1 /* return 0 as soon as cmd waits for a lock */ -#define STEP_RETRY 0x2 /* this is a retry of a previously-waiting cmd */ -static bool try_complete_step(Step *step, int flags); +#define STEP_NONBLOCK 0x1 /* return 0 as soon as cmd waits for a lock */ +#define STEP_RETRY 0x2 /* this is a retry of a previously-waiting cmd */ +static bool try_complete_step(Step * step, int flags); static int step_qsort_cmp(const void *a, const void *b); static int step_bsearch_cmp(const void *a, const void *b); @@ -82,7 +81,7 @@ main(int argc, char **argv) int i; PGresult *res; PQExpBufferData wait_query; - int opt; + int opt; while ((opt = getopt(argc, argv, "n")) != -1) { @@ -99,8 +98,8 @@ main(int argc, char **argv) /* * If the user supplies a non-option parameter on the command line, use it - * as the conninfo string; otherwise default to setting dbname=postgres and - * using environment variables or defaults for all other connection + * as the conninfo string; otherwise default to setting dbname=postgres + * and using environment variables or defaults for all other connection * parameters. */ if (argc > optind) @@ -125,8 +124,8 @@ main(int argc, char **argv) printf("Parsed test spec with %d sessions\n", testspec->nsessions); /* - * Establish connections to the database, one for each session and an extra - * for lock wait detection and global work. + * Establish connections to the database, one for each session and an + * extra for lock wait detection and global work. */ nconns = 1 + testspec->nsessions; conns = calloc(nconns, sizeof(PGconn *)); @@ -187,7 +186,7 @@ main(int argc, char **argv) /* * Build the query we'll use to detect lock contention among sessions in - * the test specification. Most of the time, we could get away with + * the test specification. Most of the time, we could get away with * simply checking whether a session is waiting for *any* lock: we don't * exactly expect concurrent use of test tables. However, autovacuum will * occasionally take AccessExclusiveLock to truncate a table, and we must @@ -254,16 +253,16 @@ main(int argc, char **argv) "'ExclusiveLock'," "'AccessExclusiveLock'] END) " - "AND holder.locktype IS NOT DISTINCT FROM waiter.locktype " - "AND holder.database IS NOT DISTINCT FROM waiter.database " - "AND holder.relation IS NOT DISTINCT FROM waiter.relation " + "AND holder.locktype IS NOT DISTINCT FROM waiter.locktype " + "AND holder.database IS NOT DISTINCT FROM waiter.database " + "AND holder.relation IS NOT DISTINCT FROM waiter.relation " "AND holder.page IS NOT DISTINCT FROM waiter.page " "AND holder.tuple IS NOT DISTINCT FROM waiter.tuple " - "AND holder.virtualxid IS NOT DISTINCT FROM waiter.virtualxid " - "AND holder.transactionid IS NOT DISTINCT FROM waiter.transactionid " - "AND holder.classid IS NOT DISTINCT FROM waiter.classid " + "AND holder.virtualxid IS NOT DISTINCT FROM waiter.virtualxid " + "AND holder.transactionid IS NOT DISTINCT FROM waiter.transactionid " + "AND holder.classid IS NOT DISTINCT FROM waiter.classid " "AND holder.objid IS NOT DISTINCT FROM waiter.objid " - "AND holder.objsubid IS NOT DISTINCT FROM waiter.objsubid "); + "AND holder.objsubid IS NOT DISTINCT FROM waiter.objsubid "); res = PQprepare(conns[0], PREP_WAITING, wait_query.data, 0, NULL); if (PQresultStatus(res) != PGRES_COMMAND_OK) @@ -294,7 +293,7 @@ static int *piles; * explicitly specified. */ static void -run_testspec(TestSpec *testspec) +run_testspec(TestSpec * testspec) { if (testspec->permutations) run_named_permutations(testspec); @@ -400,9 +399,10 @@ run_named_permutations(TestSpec * testspec) /* Find all the named steps using the lookup table */ for (j = 0; j < p->nsteps; j++) { - Step **this = (Step **) bsearch(p->stepnames[j], allsteps, - nallsteps, sizeof(Step *), - &step_bsearch_cmp); + Step **this = (Step **) bsearch(p->stepnames[j], allsteps, + nallsteps, sizeof(Step *), + &step_bsearch_cmp); + if (this == NULL) { fprintf(stderr, "undefined step \"%s\" specified in permutation\n", @@ -441,7 +441,7 @@ step_bsearch_cmp(const void *a, const void *b) * If a step caused an error to be reported, print it out and clear it. */ static void -report_error_message(Step *step) +report_error_message(Step * step) { if (step->errormsg) { @@ -458,9 +458,9 @@ report_error_message(Step *step) * one fails due to a timeout such as deadlock timeout. */ static void -report_two_error_messages(Step *step1, Step *step2) +report_two_error_messages(Step * step1, Step * step2) { - char *prefix; + char *prefix; prefix = malloc(strlen(step1->name) + strlen(step2->name) + 2); sprintf(prefix, "%s %s", step1->name, step2->name); @@ -494,8 +494,8 @@ run_permutation(TestSpec * testspec, int nsteps, Step ** steps) Step *waiting = NULL; /* - * In dry run mode, just display the permutation in the same format used by - * spec files, and return. + * In dry run mode, just display the permutation in the same format used + * by spec files, and return. */ if (dry_run) { @@ -547,21 +547,22 @@ run_permutation(TestSpec * testspec, int nsteps, Step ** steps) /* Perform steps */ for (i = 0; i < nsteps; i++) { - Step *step = steps[i]; - PGconn *conn = conns[1 + step->session]; + Step *step = steps[i]; + PGconn *conn = conns[1 + step->session]; if (waiting != NULL && step->session == waiting->session) { - PGcancel *cancel; - PGresult *res; - int j; + PGcancel *cancel; + PGresult *res; + int j; /* * This permutation is invalid: it can never happen in real life. * - * A session is blocked on an earlier step (waiting) and no further - * steps from this session can run until it is unblocked, but it - * can only be unblocked by running steps from other sessions. + * A session is blocked on an earlier step (waiting) and no + * further steps from this session can run until it is unblocked, + * but it can only be unblocked by running steps from other + * sessions. */ fprintf(stderr, "invalid permutation detected\n"); @@ -569,7 +570,7 @@ run_permutation(TestSpec * testspec, int nsteps, Step ** steps) cancel = PQgetCancel(conn); if (cancel != NULL) { - char buf[256]; + char buf[256]; PQcancel(cancel, buf, sizeof(buf)); @@ -673,11 +674,11 @@ teardown: /* * Our caller already sent the query associated with this step. Wait for it * to either complete or (if given the STEP_NONBLOCK flag) to block while - * waiting for a lock. We assume that any lock wait will persist until we + * waiting for a lock. We assume that any lock wait will persist until we * have executed additional steps in the permutation. * * When calling this function on behalf of a given step for a second or later - * time, pass the STEP_RETRY flag. This only affects the messages printed. + * time, pass the STEP_RETRY flag. This only affects the messages printed. * * If the connection returns an error, the message is saved in step->errormsg. * Caller should call report_error_message shortly after this, to have it @@ -687,7 +688,7 @@ teardown: * a lock, returns true. Otherwise, returns false. */ static bool -try_complete_step(Step *step, int flags) +try_complete_step(Step * step, int flags) { PGconn *conn = conns[1 + step->session]; fd_set read_set; @@ -705,12 +706,12 @@ try_complete_step(Step *step, int flags) timeout.tv_usec = 10000; /* Check for lock waits every 10ms. */ ret = select(sock + 1, &read_set, NULL, NULL, &timeout); - if (ret < 0) /* error in select() */ + if (ret < 0) /* error in select() */ { fprintf(stderr, "select failed: %s\n", strerror(errno)); exit_nicely(); } - else if (ret == 0) /* select() timeout: check for lock wait */ + else if (ret == 0) /* select() timeout: check for lock wait */ { int ntuples; @@ -765,9 +766,9 @@ try_complete_step(Step *step, int flags) } /* Detail may contain xid values, so just show primary. */ step->errormsg = malloc(5 + - strlen(PQresultErrorField(res, PG_DIAG_SEVERITY)) + + strlen(PQresultErrorField(res, PG_DIAG_SEVERITY)) + strlen(PQresultErrorField(res, - PG_DIAG_MESSAGE_PRIMARY))); + PG_DIAG_MESSAGE_PRIMARY))); sprintf(step->errormsg, "%s: %s", PQresultErrorField(res, PG_DIAG_SEVERITY), PQresultErrorField(res, PG_DIAG_MESSAGE_PRIMARY)); diff --git a/src/test/regress/pg_regress.c b/src/test/regress/pg_regress.c index 2931a73fdbe94618f1231dba91d6424a1198ee83..7d89318b270f7061953c0acbed95cc4e8b11ba7d 100644 --- a/src/test/regress/pg_regress.c +++ b/src/test/regress/pg_regress.c @@ -475,7 +475,7 @@ convert_sourcefiles_in(char *source_subdir, char *dest_dir, char *dest_subdir, c /* build the full actual paths to open */ snprintf(prefix, strlen(*name) - 6, "%s", *name); snprintf(srcfile, MAXPGPATH, "%s/%s", indir, *name); - snprintf(destfile, MAXPGPATH, "%s/%s/%s.%s", dest_dir, dest_subdir, + snprintf(destfile, MAXPGPATH, "%s/%s/%s.%s", dest_dir, dest_subdir, prefix, suffix); infile = fopen(srcfile, "r"); diff --git a/src/test/thread/thread_test.c b/src/test/thread/thread_test.c index bb5b92f142124efa93f679cce03e32d41ec55948..9041928f08f941d4bd3e517eae1961446bac2c6d 100644 --- a/src/test/thread/thread_test.c +++ b/src/test/thread/thread_test.c @@ -114,8 +114,10 @@ static bool platform_is_threadsafe = true; int main(int argc, char *argv[]) { - pthread_t thread1, thread2; + pthread_t thread1, + thread2; int rc; + #ifdef WIN32 WSADATA wsaData; int err; @@ -199,7 +201,7 @@ main(int argc, char *argv[]) #endif /* close down threads */ - + pthread_mutex_unlock(&init_mutex); /* let children exit */ pthread_join(thread1, NULL); /* clean up children */ @@ -277,7 +279,7 @@ func_call_1(void) #ifdef WIN32 HANDLE h1; #else - int fd; + int fd; #endif unlink(TEMP_FILENAME_1); @@ -285,10 +287,11 @@ func_call_1(void) /* Set errno = EEXIST */ /* create, then try to fail on exclusive create open */ + /* * It would be great to check errno here but if errno is not thread-safe - * we might get a value from the other thread and mis-report the cause - * of the failure. + * we might get a value from the other thread and mis-report the cause of + * the failure. */ #ifdef WIN32 if ((h1 = CreateFile(TEMP_FILENAME_1, GENERIC_WRITE, 0, NULL, OPEN_ALWAYS, 0, NULL)) == @@ -301,7 +304,7 @@ func_call_1(void) TEMP_FILENAME_1); exit(1); } - + #ifdef WIN32 if (CreateFile(TEMP_FILENAME_1, GENERIC_WRITE, 0, NULL, CREATE_NEW, 0, NULL) != INVALID_HANDLE_VALUE) @@ -346,6 +349,7 @@ func_call_1(void) unlink(TEMP_FILENAME_1); #ifndef HAVE_STRERROR_R + /* * If strerror() uses sys_errlist, the pointer might change for different * errno values, so we don't check to see if it varies within the thread. @@ -428,6 +432,7 @@ func_call_2(void) } #ifndef HAVE_STRERROR_R + /* * If strerror() uses sys_errlist, the pointer might change for different * errno values, so we don't check to see if it varies within the thread. diff --git a/src/timezone/pgtz.c b/src/timezone/pgtz.c index 36e20b834ee09e8650956b9560504f4416dcb5d7..3dae9e5e9fd711a40ca4508b04a403da58caa397 100644 --- a/src/timezone/pgtz.c +++ b/src/timezone/pgtz.c @@ -295,7 +295,7 @@ pg_tzset(const char *name) * This is called before GUC variable initialization begins. Its purpose * is to ensure that log_timezone has a valid value before any logging GUC * variables could become set to values that require elog.c to provide - * timestamps (e.g., log_line_prefix). We may as well initialize + * timestamps (e.g., log_line_prefix). We may as well initialize * session_timestamp to something valid, too. */ void @@ -303,9 +303,9 @@ pg_timezone_initialize(void) { /* * We may not yet know where PGSHAREDIR is (in particular this is true in - * an EXEC_BACKEND subprocess). So use "GMT", which pg_tzset forces to - * be interpreted without reference to the filesystem. This corresponds - * to the bootstrap default for these variables in guc.c, although in + * an EXEC_BACKEND subprocess). So use "GMT", which pg_tzset forces to be + * interpreted without reference to the filesystem. This corresponds to + * the bootstrap default for these variables in guc.c, although in * principle it could be different. */ session_timezone = pg_tzset("GMT"); diff --git a/src/tools/msvc/Install.pm b/src/tools/msvc/Install.pm index 7c743d17b4f81004a9cba2270911e95adf720095..058fab3e5ab91207b46b62b0602f1ec0b68d5adf 100644 --- a/src/tools/msvc/Install.pm +++ b/src/tools/msvc/Install.pm @@ -19,593 +19,615 @@ our (@ISA,@EXPORT_OK); sub lcopy { - my $src = shift; - my $target = shift; + my $src = shift; + my $target = shift; - if (-f $target) - { - unlink $target || confess "Could not delete $target\n"; - } + if (-f $target) + { + unlink $target || confess "Could not delete $target\n"; + } - copy($src,$target) - || confess "Could not copy $src to $target\n"; + copy($src,$target) + || confess "Could not copy $src to $target\n"; } sub Install { - $| = 1; - - my $target = shift; - our $config; - require "config_default.pl"; - require "config.pl" if (-f "config.pl"); - - chdir("../../..") if (-f "../../../configure"); - chdir("../../../..") if (-f "../../../../configure"); - my $conf = ""; - if (-d "debug") - { - $conf = "debug"; - } - if (-d "release") - { - $conf = "release"; - } - die "Could not find debug or release binaries" if ($conf eq ""); - my $majorver = DetermineMajorVersion(); - print "Installing version $majorver for $conf in $target\n"; - - EnsureDirectories($target, 'bin', 'lib', 'share','share/timezonesets','share/extension', - 'share/contrib','doc','doc/extension', 'doc/contrib','symbols', 'share/tsearch_data'); - - CopySolutionOutput($conf, $target); - lcopy($target . '/lib/libpq.dll', $target . '/bin/libpq.dll'); - my $sample_files = []; - File::Find::find( - { - wanted =>sub { - /^.*\.sample\z/s - &&push(@$sample_files, $File::Find::name); - } - }, - "src" - ); - CopySetOfFiles('config files', $sample_files, $target . '/share/'); - CopyFiles( - 'Import libraries', - $target .'/lib/', - "$conf\\", "postgres\\postgres.lib","libpq\\libpq.lib", "libecpg\\libecpg.lib", - "libpgport\\libpgport.lib" - ); - CopySetOfFiles( - 'timezone names', - [ glob('src\timezone\tznames\*.txt') ], - $target . '/share/timezonesets/' - ); - CopyFiles( - 'timezone sets', - $target . '/share/timezonesets/', - 'src/timezone/tznames/', 'Default','Australia','India' - ); - CopySetOfFiles('BKI files', [ glob("src\\backend\\catalog\\postgres.*") ],$target .'/share/'); - CopySetOfFiles('SQL files', [ glob("src\\backend\\catalog\\*.sql") ],$target . '/share/'); - CopyFiles( - 'Information schema data', - $target . '/share/', - 'src/backend/catalog/', 'sql_features.txt' - ); - GenerateConversionScript($target); - GenerateTimezoneFiles($target,$conf); - GenerateTsearchFiles($target); - CopySetOfFiles( - 'Stopword files', - [ glob("src\\backend\\snowball\\stopwords\\*.stop") ], - $target . '/share/tsearch_data/' - ); - CopySetOfFiles( - 'Dictionaries sample files', - [ glob("src\\backend\\tsearch\\*_sample.*") ], - $target . '/share/tsearch_data/' - ); - CopyContribFiles($config,$target); - CopyIncludeFiles($target); - - my $pl_extension_files = []; - my @pldirs = ('src/pl/plpgsql/src'); - push @pldirs,"src/pl/plperl" if $config->{perl}; - push @pldirs,"src/pl/plpython" if $config->{python}; - push @pldirs,"src/pl/tcl" if $config->{tcl}; - File::Find::find( - { - wanted =>sub { - /^(.*--.*\.sql|.*\.control)\z/s - &&push(@$pl_extension_files, $File::Find::name); - } - }, - @pldirs - ); - CopySetOfFiles('PL Extension files', $pl_extension_files,$target . '/share/extension/'); - - GenerateNLSFiles($target,$config->{nls},$majorver) if ($config->{nls}); - - print "Installation complete.\n"; + $| = 1; + + my $target = shift; + our $config; + require "config_default.pl"; + require "config.pl" if (-f "config.pl"); + + chdir("../../..") if (-f "../../../configure"); + chdir("../../../..") if (-f "../../../../configure"); + my $conf = ""; + if (-d "debug") + { + $conf = "debug"; + } + if (-d "release") + { + $conf = "release"; + } + die "Could not find debug or release binaries" if ($conf eq ""); + my $majorver = DetermineMajorVersion(); + print "Installing version $majorver for $conf in $target\n"; + + EnsureDirectories($target, 'bin', 'lib', 'share','share/timezonesets','share/extension', + 'share/contrib','doc','doc/extension', 'doc/contrib','symbols', + 'share/tsearch_data'); + + CopySolutionOutput($conf, $target); + lcopy($target . '/lib/libpq.dll', $target . '/bin/libpq.dll'); + my $sample_files = []; + File::Find::find( + { + wanted =>sub { + /^.*\.sample\z/s + &&push(@$sample_files, $File::Find::name); + } + }, + "src" + ); + CopySetOfFiles('config files', $sample_files, $target . '/share/'); + CopyFiles( + 'Import libraries', + $target .'/lib/', + "$conf\\", "postgres\\postgres.lib","libpq\\libpq.lib", "libecpg\\libecpg.lib", + "libpgport\\libpgport.lib" + ); + CopySetOfFiles( + 'timezone names', + [ glob('src\timezone\tznames\*.txt') ], + $target . '/share/timezonesets/' + ); + CopyFiles( + 'timezone sets', + $target . '/share/timezonesets/', + 'src/timezone/tznames/', 'Default','Australia','India' + ); + CopySetOfFiles( + 'BKI files', + [ glob("src\\backend\\catalog\\postgres.*") ], + $target .'/share/' + ); + CopySetOfFiles('SQL files', [ glob("src\\backend\\catalog\\*.sql") ],$target . '/share/'); + CopyFiles( + 'Information schema data',$target . '/share/', + 'src/backend/catalog/', 'sql_features.txt' + ); + GenerateConversionScript($target); + GenerateTimezoneFiles($target,$conf); + GenerateTsearchFiles($target); + CopySetOfFiles( + 'Stopword files', + [ glob("src\\backend\\snowball\\stopwords\\*.stop") ], + $target . '/share/tsearch_data/' + ); + CopySetOfFiles( + 'Dictionaries sample files', + [ glob("src\\backend\\tsearch\\*_sample.*") ], + $target . '/share/tsearch_data/' + ); + CopyContribFiles($config,$target); + CopyIncludeFiles($target); + + my $pl_extension_files = []; + my @pldirs = ('src/pl/plpgsql/src'); + push @pldirs,"src/pl/plperl" if $config->{perl}; + push @pldirs,"src/pl/plpython" if $config->{python}; + push @pldirs,"src/pl/tcl" if $config->{tcl}; + File::Find::find( + { + wanted =>sub { + /^(.*--.*\.sql|.*\.control)\z/s + &&push(@$pl_extension_files, + $File::Find::name); + } + }, + @pldirs + ); + CopySetOfFiles('PL Extension files', $pl_extension_files,$target . '/share/extension/'); + + GenerateNLSFiles($target,$config->{nls},$majorver) if ($config->{nls}); + + print "Installation complete.\n"; } sub EnsureDirectories { - my $target = shift; - mkdir $target unless -d ($target); - while (my $d = shift) - { - mkdir $target . '/' . $d unless -d ($target . '/' . $d); - } + my $target = shift; + mkdir $target unless -d ($target); + while (my $d = shift) + { + mkdir $target . '/' . $d unless -d ($target . '/' . $d); + } } sub CopyFiles { - my $what = shift; - my $target = shift; - my $basedir = shift; - - print "Copying $what"; - while (my $f = shift) - { - print "."; - $f = $basedir . $f; - die "No file $f\n" if (!-f $f); - lcopy($f, $target . basename($f)); - } - print "\n"; + my $what = shift; + my $target = shift; + my $basedir = shift; + + print "Copying $what"; + while (my $f = shift) + { + print "."; + $f = $basedir . $f; + die "No file $f\n" if (!-f $f); + lcopy($f, $target . basename($f)); + } + print "\n"; } sub CopySetOfFiles { - my $what = shift; - my $flist = shift; - my $target = shift; - print "Copying $what" if $what; - foreach (@$flist) - { - next if /regress/; # Skip temporary install in regression subdir - next if /ecpg.test/; # Skip temporary install in regression subdir - my $tgt = $target . basename($_); - print "."; - lcopy($_, $tgt) || croak "Could not copy $_: $!\n"; - } - print "\n"; + my $what = shift; + my $flist = shift; + my $target = shift; + print "Copying $what" if $what; + foreach (@$flist) + { + next if /regress/; # Skip temporary install in regression subdir + next if /ecpg.test/; # Skip temporary install in regression subdir + my $tgt = $target . basename($_); + print "."; + lcopy($_, $tgt) || croak "Could not copy $_: $!\n"; + } + print "\n"; } sub CopySolutionOutput { - my $conf = shift; - my $target = shift; - my $rem = qr{Project\("{8BC9CEB8-8B4A-11D0-8D11-00A0C91BC942}"\) = "([^"]+)"}; - - my $sln = read_file("pgsql.sln") || croak "Could not open pgsql.sln\n"; - - my $vcproj = 'vcproj'; - if ($sln =~ /Microsoft Visual Studio Solution File, Format Version (\d+)\.\d+/ && $1 >= 11) - { - $vcproj = 'vcxproj'; - } - - print "Copying build output files..."; - while ($sln =~ $rem) - { - my $pf = $1; - my $dir; - my $ext; - - $sln =~ s/$rem//; - - my $proj = read_file("$pf.$vcproj") || croak "Could not open $pf.$vcproj\n"; - if ($vcproj eq 'vcproj' && $proj =~ qr{ConfigurationType="([^"]+)"}) - { - if ($1 == 1) - { - $dir = "bin"; - $ext = "exe"; - } - elsif ($1 == 2) - { - $dir = "lib"; - $ext = "dll"; - } - else - { - - # Static lib, such as libpgport, only used internally during build, don't install - next; - } - } - elsif ($vcproj eq 'vcxproj' && $proj =~ qr{<ConfigurationType>(\w+)</ConfigurationType>}) - { - if ($1 eq 'Application') - { - $dir = "bin"; - $ext = "exe"; - } - elsif ($1 eq 'DynamicLibrary') - { - $dir = "lib"; - $ext = "dll"; - } - else # 'StaticLibrary' - { - - # Static lib, such as libpgport, only used internally during build, don't install - next; - } - } - else - { - croak "Could not parse $pf.$vcproj\n"; - } - lcopy("$conf\\$pf\\$pf.$ext","$target\\$dir\\$pf.$ext") - || croak "Could not copy $pf.$ext\n"; - lcopy("$conf\\$pf\\$pf.pdb","$target\\symbols\\$pf.pdb") - || croak "Could not copy $pf.pdb\n"; - print "."; - } - print "\n"; + my $conf = shift; + my $target = shift; + my $rem = qr{Project\("{8BC9CEB8-8B4A-11D0-8D11-00A0C91BC942}"\) = "([^"]+)"}; + + my $sln = read_file("pgsql.sln") || croak "Could not open pgsql.sln\n"; + + my $vcproj = 'vcproj'; + if ($sln =~ /Microsoft Visual Studio Solution File, Format Version (\d+)\.\d+/ && $1 >= 11) + { + $vcproj = 'vcxproj'; + } + + print "Copying build output files..."; + while ($sln =~ $rem) + { + my $pf = $1; + my $dir; + my $ext; + + $sln =~ s/$rem//; + + my $proj = read_file("$pf.$vcproj") || croak "Could not open $pf.$vcproj\n"; + if ($vcproj eq 'vcproj' && $proj =~ qr{ConfigurationType="([^"]+)"}) + { + if ($1 == 1) + { + $dir = "bin"; + $ext = "exe"; + } + elsif ($1 == 2) + { + $dir = "lib"; + $ext = "dll"; + } + else + { + + # Static lib, such as libpgport, only used internally during build, don't install + next; + } + } + elsif ( $vcproj eq 'vcxproj' + && $proj =~ qr{<ConfigurationType>(\w+)</ConfigurationType>}) + { + if ($1 eq 'Application') + { + $dir = "bin"; + $ext = "exe"; + } + elsif ($1 eq 'DynamicLibrary') + { + $dir = "lib"; + $ext = "dll"; + } + else # 'StaticLibrary' + { + + # Static lib, such as libpgport, only used internally during build, don't install + next; + } + } + else + { + croak "Could not parse $pf.$vcproj\n"; + } + lcopy("$conf\\$pf\\$pf.$ext","$target\\$dir\\$pf.$ext") + || croak "Could not copy $pf.$ext\n"; + lcopy("$conf\\$pf\\$pf.pdb","$target\\symbols\\$pf.pdb") + || croak "Could not copy $pf.pdb\n"; + print "."; + } + print "\n"; } sub GenerateConversionScript { - my $target = shift; - my $sql = ""; - my $F; - - print "Generating conversion proc script..."; - my $mf = read_file('src/backend/utils/mb/conversion_procs/Makefile'); - $mf =~ s{\\\s*[\r\n]+}{}mg; - $mf =~ /^CONVERSIONS\s*=\s*(.*)$/m - || die "Could not find CONVERSIONS line in conversions Makefile\n"; - my @pieces = split /\s+/,$1; - while ($#pieces > 0) - { - my $name = shift @pieces; - my $se = shift @pieces; - my $de = shift @pieces; - my $func = shift @pieces; - my $obj = shift @pieces; - $sql .= "-- $se --> $de\n"; - $sql .= + my $target = shift; + my $sql = ""; + my $F; + + print "Generating conversion proc script..."; + my $mf = read_file('src/backend/utils/mb/conversion_procs/Makefile'); + $mf =~ s{\\\s*[\r\n]+}{}mg; + $mf =~ /^CONVERSIONS\s*=\s*(.*)$/m + || die "Could not find CONVERSIONS line in conversions Makefile\n"; + my @pieces = split /\s+/,$1; + while ($#pieces > 0) + { + my $name = shift @pieces; + my $se = shift @pieces; + my $de = shift @pieces; + my $func = shift @pieces; + my $obj = shift @pieces; + $sql .= "-- $se --> $de\n"; + $sql .= "CREATE OR REPLACE FUNCTION $func (INTEGER, INTEGER, CSTRING, INTERNAL, INTEGER) RETURNS VOID AS '\$libdir/$obj', '$func' LANGUAGE C STRICT;\n"; - $sql .= + $sql .= "COMMENT ON FUNCTION $func(INTEGER, INTEGER, CSTRING, INTERNAL, INTEGER) IS 'internal conversion function for $se to $de';\n"; - $sql .= "DROP CONVERSION pg_catalog.$name;\n"; - $sql .= "CREATE DEFAULT CONVERSION pg_catalog.$name FOR '$se' TO '$de' FROM $func;\n"; - $sql .= "COMMENT ON CONVERSION pg_catalog.$name IS 'conversion for $se to $de';\n"; - } - open($F,">$target/share/conversion_create.sql") - || die "Could not write to conversion_create.sql\n"; - print $F $sql; - close($F); - print "\n"; + $sql .= "DROP CONVERSION pg_catalog.$name;\n"; + $sql .= + "CREATE DEFAULT CONVERSION pg_catalog.$name FOR '$se' TO '$de' FROM $func;\n"; + $sql .= "COMMENT ON CONVERSION pg_catalog.$name IS 'conversion for $se to $de';\n"; + } + open($F,">$target/share/conversion_create.sql") + || die "Could not write to conversion_create.sql\n"; + print $F $sql; + close($F); + print "\n"; } sub GenerateTimezoneFiles { - my $target = shift; - my $conf = shift; - my $mf = read_file("src/timezone/Makefile"); - $mf =~ s{\\\s*[\r\n]+}{}mg; - $mf =~ /^TZDATA\s*:?=\s*(.*)$/m || die "Could not find TZDATA row in timezone makefile\n"; - my @tzfiles = split /\s+/,$1; - unshift @tzfiles,''; - print "Generating timezone files..."; - system( - "$conf\\zic\\zic -d \"$target/share/timezone\" " . join(" src/timezone/data/", @tzfiles)); - print "\n"; + my $target = shift; + my $conf = shift; + my $mf = read_file("src/timezone/Makefile"); + $mf =~ s{\\\s*[\r\n]+}{}mg; + $mf =~ /^TZDATA\s*:?=\s*(.*)$/m || die "Could not find TZDATA row in timezone makefile\n"; + my @tzfiles = split /\s+/,$1; + unshift @tzfiles,''; + print "Generating timezone files..."; + system("$conf\\zic\\zic -d \"$target/share/timezone\" " + . join(" src/timezone/data/", @tzfiles)); + print "\n"; } sub GenerateTsearchFiles { - my $target = shift; - - print "Generating tsearch script..."; - my $F; - my $tmpl = read_file('src/backend/snowball/snowball.sql.in'); - my $mf = read_file('src/backend/snowball/Makefile'); - $mf =~ s{\\\s*[\r\n]+}{}mg; - $mf =~ /^LANGUAGES\s*=\s*(.*)$/m - || die "Could not find LANGUAGES line in snowball Makefile\n"; - my @pieces = split /\s+/,$1; - open($F,">$target/share/snowball_create.sql") - || die "Could not write snowball_create.sql"; - print $F read_file('src/backend/snowball/snowball_func.sql.in'); - - while ($#pieces > 0) - { - my $lang = shift @pieces || last; - my $asclang = shift @pieces || last; - my $txt = $tmpl; - my $stop = ''; - - if (-s "src/backend/snowball/stopwords/$lang.stop") - { - $stop = ", StopWords=$lang"; - } - - $txt =~ s#_LANGNAME_#${lang}#gs; - $txt =~ s#_DICTNAME_#${lang}_stem#gs; - $txt =~ s#_CFGNAME_#${lang}#gs; - $txt =~ s#_ASCDICTNAME_#${asclang}_stem#gs; - $txt =~ s#_NONASCDICTNAME_#${lang}_stem#gs; - $txt =~ s#_STOPWORDS_#$stop#gs; - print $F $txt; - print "."; - } - close($F); - print "\n"; + my $target = shift; + + print "Generating tsearch script..."; + my $F; + my $tmpl = read_file('src/backend/snowball/snowball.sql.in'); + my $mf = read_file('src/backend/snowball/Makefile'); + $mf =~ s{\\\s*[\r\n]+}{}mg; + $mf =~ /^LANGUAGES\s*=\s*(.*)$/m + || die "Could not find LANGUAGES line in snowball Makefile\n"; + my @pieces = split /\s+/,$1; + open($F,">$target/share/snowball_create.sql") + || die "Could not write snowball_create.sql"; + print $F read_file('src/backend/snowball/snowball_func.sql.in'); + + while ($#pieces > 0) + { + my $lang = shift @pieces || last; + my $asclang = shift @pieces || last; + my $txt = $tmpl; + my $stop = ''; + + if (-s "src/backend/snowball/stopwords/$lang.stop") + { + $stop = ", StopWords=$lang"; + } + + $txt =~ s#_LANGNAME_#${lang}#gs; + $txt =~ s#_DICTNAME_#${lang}_stem#gs; + $txt =~ s#_CFGNAME_#${lang}#gs; + $txt =~ s#_ASCDICTNAME_#${asclang}_stem#gs; + $txt =~ s#_NONASCDICTNAME_#${lang}_stem#gs; + $txt =~ s#_STOPWORDS_#$stop#gs; + print $F $txt; + print "."; + } + close($F); + print "\n"; } sub CopyContribFiles { - my $config = shift; - my $target = shift; - - print "Copying contrib data files..."; - my $D; - opendir($D, 'contrib') || croak "Could not opendir on contrib!\n"; - while (my $d = readdir($D)) - { - next if ($d =~ /^\./); - next unless (-f "contrib/$d/Makefile"); - next if ($d eq "uuid-ossp"&& !defined($config->{uuid})); - next if ($d eq "sslinfo" && !defined($config->{openssl})); - next if ($d eq "xml2" && !defined($config->{xml})); - next if ($d eq "sepgsql"); - - my $mf = read_file("contrib/$d/Makefile"); - $mf =~ s{\\s*[\r\n]+}{}mg; - - # Note: we currently don't support setting MODULEDIR in the makefile - my $moduledir = 'contrib'; - - my $flist = ''; - if ($mf =~ /^EXTENSION\s*=\s*(.*)$/m) {$flist .= $1} - if ($flist ne '') - { - $moduledir = 'extension'; - $flist = ParseAndCleanRule($flist, $mf); - - foreach my $f (split /\s+/,$flist) - { - lcopy( - 'contrib/' . $d . '/' . $f . '.control', - $target . '/share/extension/' . $f . '.control' - )|| croak("Could not copy file $f.control in contrib $d"); - print '.'; - } - } - - $flist = ''; - if ($mf =~ /^DATA_built\s*=\s*(.*)$/m) {$flist .= $1} - if ($mf =~ /^DATA\s*=\s*(.*)$/m) {$flist .= " $1"} - $flist =~ s/^\s*//; # Remove leading spaces if we had only DATA_built - - if ($flist ne '') - { - $flist = ParseAndCleanRule($flist, $mf); - - foreach my $f (split /\s+/,$flist) - { - lcopy('contrib/' . $d . '/' . $f, - $target . '/share/' . $moduledir . '/' . basename($f)) - || croak("Could not copy file $f in contrib $d"); - print '.'; - } - } - - $flist = ''; - if ($mf =~ /^DATA_TSEARCH\s*=\s*(.*)$/m) {$flist .= $1} - if ($flist ne '') - { - $flist = ParseAndCleanRule($flist, $mf); - - foreach my $f (split /\s+/,$flist) - { - lcopy('contrib/' . $d . '/' . $f,$target . '/share/tsearch_data/' . basename($f)) - || croak("Could not copy file $f in contrib $d"); - print '.'; - } - } - - $flist = ''; - if ($mf =~ /^DOCS\s*=\s*(.*)$/mg) {$flist .= $1} - if ($flist ne '') - { - $flist = ParseAndCleanRule($flist, $mf); - - # Special case for contrib/spi - $flist = + my $config = shift; + my $target = shift; + + print "Copying contrib data files..."; + my $D; + opendir($D, 'contrib') || croak "Could not opendir on contrib!\n"; + while (my $d = readdir($D)) + { + next if ($d =~ /^\./); + next unless (-f "contrib/$d/Makefile"); + next if ($d eq "uuid-ossp"&& !defined($config->{uuid})); + next if ($d eq "sslinfo" && !defined($config->{openssl})); + next if ($d eq "xml2" && !defined($config->{xml})); + next if ($d eq "sepgsql"); + + my $mf = read_file("contrib/$d/Makefile"); + $mf =~ s{\\s*[\r\n]+}{}mg; + + # Note: we currently don't support setting MODULEDIR in the makefile + my $moduledir = 'contrib'; + + my $flist = ''; + if ($mf =~ /^EXTENSION\s*=\s*(.*)$/m) {$flist .= $1} + if ($flist ne '') + { + $moduledir = 'extension'; + $flist = ParseAndCleanRule($flist, $mf); + + foreach my $f (split /\s+/,$flist) + { + lcopy( + 'contrib/' . $d . '/' . $f . '.control', + $target . '/share/extension/' . $f . '.control' + )|| croak("Could not copy file $f.control in contrib $d"); + print '.'; + } + } + + $flist = ''; + if ($mf =~ /^DATA_built\s*=\s*(.*)$/m) {$flist .= $1} + if ($mf =~ /^DATA\s*=\s*(.*)$/m) {$flist .= " $1"} + $flist =~ s/^\s*//; # Remove leading spaces if we had only DATA_built + + if ($flist ne '') + { + $flist = ParseAndCleanRule($flist, $mf); + + foreach my $f (split /\s+/,$flist) + { + lcopy('contrib/' . $d . '/' . $f, + $target . '/share/' . $moduledir . '/' . basename($f)) + || croak("Could not copy file $f in contrib $d"); + print '.'; + } + } + + $flist = ''; + if ($mf =~ /^DATA_TSEARCH\s*=\s*(.*)$/m) {$flist .= $1} + if ($flist ne '') + { + $flist = ParseAndCleanRule($flist, $mf); + + foreach my $f (split /\s+/,$flist) + { + lcopy('contrib/' . $d . '/' . $f, + $target . '/share/tsearch_data/' . basename($f)) + || croak("Could not copy file $f in contrib $d"); + print '.'; + } + } + + $flist = ''; + if ($mf =~ /^DOCS\s*=\s*(.*)$/mg) {$flist .= $1} + if ($flist ne '') + { + $flist = ParseAndCleanRule($flist, $mf); + + # Special case for contrib/spi + $flist = "autoinc.example insert_username.example moddatetime.example refint.example timetravel.example" - if ($d eq 'spi'); - foreach my $f (split /\s+/,$flist) - { - lcopy('contrib/' . $d . '/' . $f,$target . '/doc/' . $moduledir . '/' . $f) - || croak("Could not copy file $f in contrib $d"); - print '.'; - } - } - } - closedir($D); - print "\n"; + if ($d eq 'spi'); + foreach my $f (split /\s+/,$flist) + { + lcopy('contrib/' . $d . '/' . $f, + $target . '/doc/' . $moduledir . '/' . $f) + || croak("Could not copy file $f in contrib $d"); + print '.'; + } + } + } + closedir($D); + print "\n"; } sub ParseAndCleanRule { - my $flist = shift; - my $mf = shift; - - # Strip out $(addsuffix) rules - if (index($flist, '$(addsuffix ') >= 0) - { - my $pcount = 0; - my $i; - for ($i = index($flist, '$(addsuffix ') + 12; $i < length($flist); $i++) - { - $pcount++ if (substr($flist, $i, 1) eq '('); - $pcount-- if (substr($flist, $i, 1) eq ')'); - last if ($pcount < 0); - } - $flist = substr($flist, 0, index($flist, '$(addsuffix ')) . substr($flist, $i+1); - } - return $flist; + my $flist = shift; + my $mf = shift; + + # Strip out $(addsuffix) rules + if (index($flist, '$(addsuffix ') >= 0) + { + my $pcount = 0; + my $i; + for ($i = index($flist, '$(addsuffix ') + 12; $i < length($flist); $i++) + { + $pcount++ if (substr($flist, $i, 1) eq '('); + $pcount-- if (substr($flist, $i, 1) eq ')'); + last if ($pcount < 0); + } + $flist = substr($flist, 0, index($flist, '$(addsuffix ')) . substr($flist, $i+1); + } + return $flist; } sub CopyIncludeFiles { - my $target = shift; - - EnsureDirectories($target, 'include', 'include/libpq','include/internal', - 'include/internal/libpq','include/server', 'include/server/parser'); - - CopyFiles( - 'Public headers', - $target . '/include/', - 'src/include/', 'postgres_ext.h', 'pg_config.h', 'pg_config_os.h', 'pg_config_manual.h' - ); - lcopy('src/include/libpq/libpq-fs.h', $target . '/include/libpq/') - || croak 'Could not copy libpq-fs.h'; - - CopyFiles( - 'Libpq headers', - $target . '/include/', - 'src/interfaces/libpq/','libpq-fe.h', 'libpq-events.h' - ); - CopyFiles( - 'Libpq internal headers', - $target .'/include/internal/', - 'src/interfaces/libpq/', 'libpq-int.h', 'pqexpbuffer.h' - ); - - CopyFiles( - 'Internal headers', - $target . '/include/internal/', - 'src/include/', 'c.h', 'port.h', 'postgres_fe.h' - ); - lcopy('src/include/libpq/pqcomm.h', $target . '/include/internal/libpq/') - || croak 'Could not copy pqcomm.h'; - - CopyFiles( - 'Server headers', - $target . '/include/server/', - 'src/include/', 'pg_config.h', 'pg_config_os.h' - ); - CopyFiles('Grammar header', $target . '/include/server/parser/','src/backend/parser/','gram.h'); - CopySetOfFiles('',[ glob("src\\include\\*.h") ],$target . '/include/server/'); - my $D; - opendir($D, 'src/include') || croak "Could not opendir on src/include!\n"; - - CopyFiles('PL/pgSQL header', $target . '/include/server/','src/pl/plpgsql/src/', 'plpgsql.h'); - - # some xcopy progs don't like mixed slash style paths - (my $ctarget = $target) =~ s!/!\\!g; - while (my $d = readdir($D)) - { - next if ($d =~ /^\./); - next if ($d eq '.git'); - next if ($d eq 'CVS'); - next unless (-d "src/include/$d"); - - EnsureDirectories("$target/include/server/$d"); - system(qq{xcopy /s /i /q /r /y src\\include\\$d\\*.h "$ctarget\\include\\server\\$d\\"}) - && croak("Failed to copy include directory $d\n"); - } - closedir($D); - - my $mf = read_file('src/interfaces/ecpg/include/Makefile'); - $mf =~ s{\\s*[\r\n]+}{}mg; - $mf =~ /^ecpg_headers\s*=\s*(.*)$/m || croak "Could not find ecpg_headers line\n"; - CopyFiles( - 'ECPG headers', - $target . '/include/', - 'src/interfaces/ecpg/include/', - 'ecpg_config.h', split /\s+/,$1 - ); - $mf =~ /^informix_headers\s*=\s*(.*)$/m || croak "Could not find informix_headers line\n"; - EnsureDirectories($target . '/include', 'informix', 'informix/esql'); - CopyFiles( - 'ECPG informix headers', - $target .'/include/informix/esql/', - 'src/interfaces/ecpg/include/', - split /\s+/,$1 - ); + my $target = shift; + + EnsureDirectories($target, 'include', 'include/libpq','include/internal', + 'include/internal/libpq','include/server', 'include/server/parser'); + + CopyFiles( + 'Public headers', + $target . '/include/', + 'src/include/', 'postgres_ext.h', 'pg_config.h', 'pg_config_os.h', + 'pg_config_manual.h' + ); + lcopy('src/include/libpq/libpq-fs.h', $target . '/include/libpq/') + || croak 'Could not copy libpq-fs.h'; + + CopyFiles( + 'Libpq headers', + $target . '/include/', + 'src/interfaces/libpq/','libpq-fe.h', 'libpq-events.h' + ); + CopyFiles( + 'Libpq internal headers', + $target .'/include/internal/', + 'src/interfaces/libpq/', 'libpq-int.h', 'pqexpbuffer.h' + ); + + CopyFiles( + 'Internal headers', + $target . '/include/internal/', + 'src/include/', 'c.h', 'port.h', 'postgres_fe.h' + ); + lcopy('src/include/libpq/pqcomm.h', $target . '/include/internal/libpq/') + || croak 'Could not copy pqcomm.h'; + + CopyFiles( + 'Server headers', + $target . '/include/server/', + 'src/include/', 'pg_config.h', 'pg_config_os.h' + ); + CopyFiles( + 'Grammar header', + $target . '/include/server/parser/', + 'src/backend/parser/','gram.h' + ); + CopySetOfFiles('',[ glob("src\\include\\*.h") ],$target . '/include/server/'); + my $D; + opendir($D, 'src/include') || croak "Could not opendir on src/include!\n"; + + CopyFiles( + 'PL/pgSQL header', + $target . '/include/server/', + 'src/pl/plpgsql/src/', 'plpgsql.h' + ); + + # some xcopy progs don't like mixed slash style paths + (my $ctarget = $target) =~ s!/!\\!g; + while (my $d = readdir($D)) + { + next if ($d =~ /^\./); + next if ($d eq '.git'); + next if ($d eq 'CVS'); + next unless (-d "src/include/$d"); + + EnsureDirectories("$target/include/server/$d"); + system( +qq{xcopy /s /i /q /r /y src\\include\\$d\\*.h "$ctarget\\include\\server\\$d\\"} + )&& croak("Failed to copy include directory $d\n"); + } + closedir($D); + + my $mf = read_file('src/interfaces/ecpg/include/Makefile'); + $mf =~ s{\\s*[\r\n]+}{}mg; + $mf =~ /^ecpg_headers\s*=\s*(.*)$/m || croak "Could not find ecpg_headers line\n"; + CopyFiles( + 'ECPG headers', + $target . '/include/', + 'src/interfaces/ecpg/include/', + 'ecpg_config.h', split /\s+/,$1 + ); + $mf =~ /^informix_headers\s*=\s*(.*)$/m || croak "Could not find informix_headers line\n"; + EnsureDirectories($target . '/include', 'informix', 'informix/esql'); + CopyFiles( + 'ECPG informix headers', + $target .'/include/informix/esql/', + 'src/interfaces/ecpg/include/', + split /\s+/,$1 + ); } sub GenerateNLSFiles { - my $target = shift; - my $nlspath = shift; - my $majorver = shift; - - print "Installing NLS files..."; - EnsureDirectories($target, "share/locale"); - my @flist; - File::Find::find( - { - wanted =>sub { - /^nls\.mk\z/s - &&!push(@flist, $File::Find::name); - } - }, - "src" - ); - foreach (@flist) - { - my $prgm = DetermineCatalogName($_); - s/nls.mk/po/; - my $dir = $_; - next unless ($dir =~ /([^\/]+)\/po$/); - foreach (glob("$dir/*.po")) - { - my $lang; - next unless /([^\/]+)\.po/; - $lang = $1; - - EnsureDirectories($target, "share/locale/$lang", "share/locale/$lang/LC_MESSAGES"); - system( + my $target = shift; + my $nlspath = shift; + my $majorver = shift; + + print "Installing NLS files..."; + EnsureDirectories($target, "share/locale"); + my @flist; + File::Find::find( + { + wanted =>sub { + /^nls\.mk\z/s + &&!push(@flist, $File::Find::name); + } + }, + "src" + ); + foreach (@flist) + { + my $prgm = DetermineCatalogName($_); + s/nls.mk/po/; + my $dir = $_; + next unless ($dir =~ /([^\/]+)\/po$/); + foreach (glob("$dir/*.po")) + { + my $lang; + next unless /([^\/]+)\.po/; + $lang = $1; + + EnsureDirectories($target, "share/locale/$lang", + "share/locale/$lang/LC_MESSAGES"); + system( "\"$nlspath\\bin\\msgfmt\" -o \"$target\\share\\locale\\$lang\\LC_MESSAGES\\$prgm-$majorver.mo\" $_" - )&& croak("Could not run msgfmt on $dir\\$_"); - print "."; - } - } - print "\n"; + )&& croak("Could not run msgfmt on $dir\\$_"); + print "."; + } + } + print "\n"; } sub DetermineMajorVersion { - my $f = read_file('src/include/pg_config.h') || croak 'Could not open pg_config.h'; - $f =~ /^#define\s+PG_MAJORVERSION\s+"([^"]+)"/m || croak 'Could not determine major version'; - return $1; + my $f = read_file('src/include/pg_config.h') || croak 'Could not open pg_config.h'; + $f =~ /^#define\s+PG_MAJORVERSION\s+"([^"]+)"/m + || croak 'Could not determine major version'; + return $1; } sub DetermineCatalogName { - my $filename = shift; + my $filename = shift; - my $f = read_file($filename) || croak "Could not open $filename"; - $f =~ /CATALOG_NAME\s*\:?=\s*(\S+)/m || croak "Could not determine catalog name in $filename"; - return $1; + my $f = read_file($filename) || croak "Could not open $filename"; + $f =~ /CATALOG_NAME\s*\:?=\s*(\S+)/m + || croak "Could not determine catalog name in $filename"; + return $1; } sub read_file { - my $filename = shift; - my $F; - my $t = $/; + my $filename = shift; + my $F; + my $t = $/; - undef $/; - open($F, $filename) || die "Could not open file $filename\n"; - my $txt = <$F>; - close($F); - $/ = $t; + undef $/; + open($F, $filename) || die "Could not open file $filename\n"; + my $txt = <$F>; + close($F); + $/ = $t; - return $txt; + return $txt; } 1; diff --git a/src/tools/msvc/MSBuildProject.pm b/src/tools/msvc/MSBuildProject.pm index fcce9ebb68dc7ac3a6907616e895cb078f1c6604..4e6ea1f74091cbc778d342bfe3850e5e50745510 100644 --- a/src/tools/msvc/MSBuildProject.pm +++ b/src/tools/msvc/MSBuildProject.pm @@ -13,158 +13,159 @@ use base qw(Project); sub _new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{filenameExtension} = '.vcxproj'; + $self->{filenameExtension} = '.vcxproj'; - return $self; + return $self; } sub WriteHeader { - my ($self, $f) = @_; + my ($self, $f) = @_; - print $f <<EOF; + print $f <<EOF; <?xml version="1.0" encoding="Windows-1252"?> <Project DefaultTargets="Build" ToolsVersion="4.0" xmlns="http://schemas.microsoft.com/developer/msbuild/2003"> <ItemGroup Label="ProjectConfigurations"> EOF - $self->WriteConfigurationHeader($f, 'Debug'); - $self->WriteConfigurationHeader($f, 'Release'); - print $f <<EOF; + $self->WriteConfigurationHeader($f, 'Debug'); + $self->WriteConfigurationHeader($f, 'Release'); + print $f <<EOF; </ItemGroup> <PropertyGroup Label="Globals"> <ProjectGuid>$self->{guid}</ProjectGuid> </PropertyGroup> <Import Project="\$(VCTargetsPath)\\Microsoft.Cpp.Default.props" /> EOF - $self->WriteConfigurationPropertyGroup($f, 'Release',{ wholeopt=>'false' }); - $self->WriteConfigurationPropertyGroup($f, 'Debug',{ wholeopt=>'false' }); - print $f <<EOF; + $self->WriteConfigurationPropertyGroup($f, 'Release',{wholeopt=>'false'}); + $self->WriteConfigurationPropertyGroup($f, 'Debug',{wholeopt=>'false'}); + print $f <<EOF; <Import Project="\$(VCTargetsPath)\\Microsoft.Cpp.props" /> <ImportGroup Label="ExtensionSettings"> </ImportGroup> EOF - $self->WritePropertySheetsPropertyGroup($f, 'Release'); - $self->WritePropertySheetsPropertyGroup($f, 'Debug'); - print $f <<EOF; + $self->WritePropertySheetsPropertyGroup($f, 'Release'); + $self->WritePropertySheetsPropertyGroup($f, 'Debug'); + print $f <<EOF; <PropertyGroup Label="UserMacros" /> <PropertyGroup> <_ProjectFileVersion>10.0.30319.1</_ProjectFileVersion> EOF - $self->WriteAdditionalProperties($f, 'Debug'); - $self->WriteAdditionalProperties($f, 'Release'); - print $f <<EOF; + $self->WriteAdditionalProperties($f, 'Debug'); + $self->WriteAdditionalProperties($f, 'Release'); + print $f <<EOF; </PropertyGroup> EOF - $self->WriteItemDefinitionGroup( - $f, 'Debug', - { - defs=>'_DEBUG;DEBUG=1;', - opt=>'Disabled', - strpool=>'false', - runtime=>'MultiThreadedDebugDLL' - } - ); - $self->WriteItemDefinitionGroup($f, 'Release', - { defs=>'', opt=>'Full', strpool=>'true', runtime=>'MultiThreadedDLL' }); + $self->WriteItemDefinitionGroup( + $f, 'Debug', + { + defs=>'_DEBUG;DEBUG=1;', + opt=>'Disabled', + strpool=>'false', + runtime=>'MultiThreadedDebugDLL' + } + ); + $self->WriteItemDefinitionGroup($f, 'Release', + {defs=>'', opt=>'Full', strpool=>'true', runtime=>'MultiThreadedDLL'}); } sub AddDefine { - my ($self, $def) = @_; + my ($self, $def) = @_; - $self->{defines} .= $def . ';'; + $self->{defines} .= $def . ';'; } sub WriteReferences { - my ($self, $f) = @_; + my ($self, $f) = @_; - my @references = @{$self->{references}}; + my @references = @{$self->{references}}; - if (scalar(@references)) - { - print $f <<EOF; + if (scalar(@references)) + { + print $f <<EOF; <ItemGroup> EOF - foreach my $ref (@references) - { - print $f <<EOF; + foreach my $ref (@references) + { + print $f <<EOF; <ProjectReference Include="$ref->{name}$ref->{filenameExtension}"> <Project>$ref->{guid}</Project> </ProjectReference> EOF - } - print $f <<EOF; + } + print $f <<EOF; </ItemGroup> EOF - } + } } sub WriteFiles { - my ($self, $f) = @_; - print $f <<EOF; + my ($self, $f) = @_; + print $f <<EOF; <ItemGroup> EOF - my @grammarFiles = (); - my @resourceFiles = (); - my %uniquefiles; - foreach my $fileNameWithPath (sort keys %{ $self->{files} }) - { - confess "Bad format filename '$fileNameWithPath'\n" - unless ($fileNameWithPath =~ /^(.*)\\([^\\]+)\.[r]?[cyl]$/); - my $dir = $1; - my $fileName = $2; - if ($fileNameWithPath =~ /\.y$/ or $fileNameWithPath =~ /\.l$/) - { - push @grammarFiles, $fileNameWithPath; - } - elsif ($fileNameWithPath =~ /\.rc$/) - { - push @resourceFiles, $fileNameWithPath; - } - elsif (defined($uniquefiles{$fileName})) - { - - # File already exists, so fake a new name - my $obj = $dir; - $obj =~ s/\\/_/g; - - print $f <<EOF; + my @grammarFiles = (); + my @resourceFiles = (); + my %uniquefiles; + foreach my $fileNameWithPath (sort keys %{$self->{files}}) + { + confess "Bad format filename '$fileNameWithPath'\n" + unless ($fileNameWithPath =~ /^(.*)\\([^\\]+)\.[r]?[cyl]$/); + my $dir = $1; + my $fileName = $2; + if ($fileNameWithPath =~ /\.y$/ or $fileNameWithPath =~ /\.l$/) + { + push @grammarFiles, $fileNameWithPath; + } + elsif ($fileNameWithPath =~ /\.rc$/) + { + push @resourceFiles, $fileNameWithPath; + } + elsif (defined($uniquefiles{$fileName})) + { + + # File already exists, so fake a new name + my $obj = $dir; + $obj =~ s/\\/_/g; + + print $f <<EOF; <ClCompile Include="$fileNameWithPath"> <ObjectFileName Condition="'\$(Configuration)|\$(Platform)'=='Debug|$self->{platform}'">.\\debug\\$self->{name}\\${obj}_$fileName.obj</ObjectFileName> <ObjectFileName Condition="'\$(Configuration)|\$(Platform)'=='Release|$self->{platform}'">.\\release\\$self->{name}\\${obj}_$fileName.obj</ObjectFileName> </ClCompile> EOF - } - else - { - $uniquefiles{$fileName} = 1; - print $f <<EOF; + } + else + { + $uniquefiles{$fileName} = 1; + print $f <<EOF; <ClCompile Include="$fileNameWithPath" /> EOF - } + } - } - print $f <<EOF; + } + print $f <<EOF; </ItemGroup> EOF - if (scalar(@grammarFiles)) - { - print $f <<EOF; + if (scalar(@grammarFiles)) + { + print $f <<EOF; <ItemGroup> EOF - foreach my $grammarFile (@grammarFiles) - { - (my $outputFile = $grammarFile) =~ s/\.(y|l)$/.c/; - if ($grammarFile =~ /\.y$/) - { - $outputFile =~ s{^src\\pl\\plpgsql\\src\\gram.c$}{src\\pl\\plpgsql\\src\\pl_gram.c}; - print $f <<EOF; + foreach my $grammarFile (@grammarFiles) + { + (my $outputFile = $grammarFile) =~ s/\.(y|l)$/.c/; + if ($grammarFile =~ /\.y$/) + { + $outputFile =~ +s{^src\\pl\\plpgsql\\src\\gram.c$}{src\\pl\\plpgsql\\src\\pl_gram.c}; + print $f <<EOF; <CustomBuild Include="$grammarFile"> <Message Condition="'\$(Configuration)|\$(Platform)'=='Debug|$self->{platform}'">Running bison on $grammarFile</Message> <Command Condition="'\$(Configuration)|\$(Platform)'=='Debug|$self->{platform}'">perl "src\\tools\\msvc\\pgbison.pl" "$grammarFile"</Command> @@ -176,10 +177,10 @@ EOF <Outputs Condition="'\$(Configuration)|\$(Platform)'=='Release|$self->{platform}'">$outputFile;%(Outputs)</Outputs> </CustomBuild> EOF - } - else #if ($grammarFile =~ /\.l$/) - { - print $f <<EOF; + } + else #if ($grammarFile =~ /\.l$/) + { + print $f <<EOF; <CustomBuild Include="$grammarFile"> <Message Condition="'\$(Configuration)|\$(Platform)'=='Debug|$self->{platform}'">Running flex on $grammarFile</Message> <Command Condition="'\$(Configuration)|\$(Platform)'=='Debug|$self->{platform}'">perl "src\\tools\\msvc\\pgflex.pl" "$grammarFile"</Command> @@ -191,33 +192,33 @@ EOF <Outputs Condition="'\$(Configuration)|\$(Platform)'=='Release|$self->{platform}'">$outputFile;%(Outputs)</Outputs> </CustomBuild> EOF - } - } - print $f <<EOF; + } + } + print $f <<EOF; </ItemGroup> EOF - } - if (scalar(@resourceFiles)) - { - print $f <<EOF; + } + if (scalar(@resourceFiles)) + { + print $f <<EOF; <ItemGroup> EOF - foreach my $rcFile (@resourceFiles) - { - print $f <<EOF; + foreach my $rcFile (@resourceFiles) + { + print $f <<EOF; <ResourceCompile Include="$rcFile" /> EOF - } - print $f <<EOF; + } + print $f <<EOF; </ItemGroup> EOF - } + } } sub WriteConfigurationHeader { - my ($self, $f, $cfgname) = @_; - print $f <<EOF; + my ($self, $f, $cfgname) = @_; + print $f <<EOF; <ProjectConfiguration Include="$cfgname|$self->{platform}"> <Configuration>$cfgname</Configuration> <Platform>$self->{platform}</Platform> @@ -227,13 +228,13 @@ EOF sub WriteConfigurationPropertyGroup { - my ($self, $f, $cfgname, $p) = @_; - my $cfgtype = - ($self->{type} eq "exe") - ?'Application' - :($self->{type} eq "dll"?'DynamicLibrary':'StaticLibrary'); + my ($self, $f, $cfgname, $p) = @_; + my $cfgtype = + ($self->{type} eq "exe") + ?'Application' + :($self->{type} eq "dll"?'DynamicLibrary':'StaticLibrary'); - print $f <<EOF; + print $f <<EOF; <PropertyGroup Condition="'\$(Configuration)|\$(Platform)'=='$cfgname|$self->{platform}'" Label="Configuration"> <ConfigurationType>$cfgtype</ConfigurationType> <UseOfMfc>false</UseOfMfc> @@ -245,8 +246,8 @@ EOF sub WritePropertySheetsPropertyGroup { - my ($self, $f, $cfgname) = @_; - print $f <<EOF; + my ($self, $f, $cfgname) = @_; + print $f <<EOF; <ImportGroup Condition="'\$(Configuration)|\$(Platform)'=='$cfgname|$self->{platform}'" Label="PropertySheets"> <Import Project="\$(UserRootDir)\\Microsoft.Cpp.\$(Platform).user.props" Condition="exists('\$(UserRootDir)\\Microsoft.Cpp.\$(Platform).user.props')" Label="LocalAppDataPlatform" /> </ImportGroup> @@ -255,8 +256,8 @@ EOF sub WriteAdditionalProperties { - my ($self, $f, $cfgname) = @_; - print $f <<EOF; + my ($self, $f, $cfgname) = @_; + print $f <<EOF; <OutDir Condition="'\$(Configuration)|\$(Platform)'=='$cfgname|$self->{platform}'">.\\$cfgname\\$self->{name}\\</OutDir> <IntDir Condition="'\$(Configuration)|\$(Platform)'=='$cfgname|$self->{platform}'">.\\$cfgname\\$self->{name}\\</IntDir> <LinkIncremental Condition="'\$(Configuration)|\$(Platform)'=='$cfgname|$self->{platform}'">false</LinkIncremental> @@ -265,21 +266,21 @@ EOF sub WriteItemDefinitionGroup { - my ($self, $f, $cfgname, $p) = @_; - my $cfgtype = - ($self->{type} eq "exe") - ?'Application' - :($self->{type} eq "dll"?'DynamicLibrary':'StaticLibrary'); - my $libs = $self->GetAdditionalLinkerDependencies($cfgname, ';'); - - my $targetmachine = $self->{platform} eq 'Win32' ? 'MachineX86' : 'MachineX64'; - - my $includes = $self->{includes}; - unless ($includes eq '' or $includes =~ /;$/) - { - $includes .= ';'; - } - print $f <<EOF; + my ($self, $f, $cfgname, $p) = @_; + my $cfgtype = + ($self->{type} eq "exe") + ?'Application' + :($self->{type} eq "dll"?'DynamicLibrary':'StaticLibrary'); + my $libs = $self->GetAdditionalLinkerDependencies($cfgname, ';'); + + my $targetmachine = $self->{platform} eq 'Win32' ? 'MachineX86' : 'MachineX64'; + + my $includes = $self->{includes}; + unless ($includes eq '' or $includes =~ /;$/) + { + $includes .= ';'; + } + print $f <<EOF; <ItemDefinitionGroup Condition="'\$(Configuration)|\$(Platform)'=='$cfgname|$self->{platform}'"> <ClCompile> <Optimization>$p->{opt}</Optimization> @@ -314,49 +315,49 @@ sub WriteItemDefinitionGroup <SubSystem>Console</SubSystem> <TargetMachine>$targetmachine</TargetMachine> EOF - if ($self->{disablelinkerwarnings}) - { - print $f + if ($self->{disablelinkerwarnings}) + { + print $f " <AdditionalOptions>/ignore:$self->{disablelinkerwarnings} \%(AdditionalOptions)</AdditionalOptions>\n"; - } - if ($self->{implib}) - { - my $l = $self->{implib}; - $l =~ s/__CFGNAME__/$cfgname/g; - print $f " <ImportLibrary>$l</ImportLibrary>\n"; - } - if ($self->{def}) - { - my $d = $self->{def}; - $d =~ s/__CFGNAME__/$cfgname/g; - print $f " <ModuleDefinitionFile>$d</ModuleDefinitionFile>\n"; - } - print $f <<EOF; + } + if ($self->{implib}) + { + my $l = $self->{implib}; + $l =~ s/__CFGNAME__/$cfgname/g; + print $f " <ImportLibrary>$l</ImportLibrary>\n"; + } + if ($self->{def}) + { + my $d = $self->{def}; + $d =~ s/__CFGNAME__/$cfgname/g; + print $f " <ModuleDefinitionFile>$d</ModuleDefinitionFile>\n"; + } + print $f <<EOF; </Link> <ResourceCompile> <AdditionalIncludeDirectories>src\\include;\%(AdditionalIncludeDirectories)</AdditionalIncludeDirectories> </ResourceCompile> EOF - if ($self->{builddef}) - { - print $f <<EOF; + if ($self->{builddef}) + { + print $f <<EOF; <PreLinkEvent> <Message>Generate DEF file</Message> <Command>perl src\\tools\\msvc\\gendef.pl $cfgname\\$self->{name} $self->{platform}</Command> </PreLinkEvent> EOF - } - print $f <<EOF; + } + print $f <<EOF; </ItemDefinitionGroup> EOF } sub Footer { - my ($self, $f) = @_; - $self->WriteReferences($f); + my ($self, $f) = @_; + $self->WriteReferences($f); - print $f <<EOF; + print $f <<EOF; <Import Project="\$(VCTargetsPath)\\Microsoft.Cpp.targets" /> <ImportGroup Label="ExtensionTargets"> </ImportGroup> @@ -376,13 +377,13 @@ use base qw(MSBuildProject); sub new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{vcver} = '10.00'; + $self->{vcver} = '10.00'; - return $self; + return $self; } 1; diff --git a/src/tools/msvc/Mkvcbuild.pm b/src/tools/msvc/Mkvcbuild.pm index f0fad43feadf755899ebe4e2d3dfa0d83c3506bf..23023e54b7e3d0b02b2d98e3b0a4a0152bcb89d5 100644 --- a/src/tools/msvc/Mkvcbuild.pm +++ b/src/tools/msvc/Mkvcbuild.pm @@ -30,488 +30,506 @@ my $libpq; my $contrib_defines = {'refint' => 'REFINT_VERBOSE'}; my @contrib_uselibpq = ('dblink', 'oid2name', 'pgbench', 'pg_upgrade','vacuumlo'); my @contrib_uselibpgport =( - 'oid2name', 'pgbench', 'pg_standby','pg_archivecleanup', - 'pg_test_fsync', 'pg_test_timing', 'pg_upgrade', 'vacuumlo' + 'oid2name', 'pgbench', 'pg_standby','pg_archivecleanup', + 'pg_test_fsync', 'pg_test_timing', 'pg_upgrade', 'vacuumlo' ); my $contrib_extralibs = {'pgbench' => ['wsock32.lib']}; my $contrib_extraincludes = {'tsearch2' => ['contrib/tsearch2'], 'dblink' => ['src/backend']}; my $contrib_extrasource = { - 'cube' => ['cubescan.l','cubeparse.y'], - 'seg' => ['segscan.l','segparse.y'] + 'cube' => ['cubescan.l','cubeparse.y'], + 'seg' => ['segscan.l','segparse.y'] }; my @contrib_excludes = ('pgcrypto','intagg','sepgsql'); sub mkvcbuild { - our $config = shift; - - chdir('..\..\..') if (-d '..\msvc' && -d '..\..\..\src'); - die 'Must run from root or msvc directory' unless (-d 'src\tools\msvc' && -d 'src'); - - my $vsVersion = DetermineVisualStudioVersion(); - - $solution = CreateSolution($vsVersion, $config); - - our @pgportfiles = qw( - chklocale.c crypt.c fls.c fseeko.c getrusage.c inet_aton.c random.c - srandom.c getaddrinfo.c gettimeofday.c inet_net_ntop.c kill.c open.c - erand48.c snprintf.c strlcat.c strlcpy.c dirmod.c exec.c noblock.c path.c - pgcheckdir.c pg_crc.c pgmkdirp.c pgsleep.c pgstrcasecmp.c qsort.c qsort_arg.c - sprompt.c thread.c getopt.c getopt_long.c dirent.c rint.c win32env.c - win32error.c win32setlocale.c); - - $libpgport = $solution->AddProject('libpgport','lib','misc'); - $libpgport->AddDefine('FRONTEND'); - $libpgport->AddFiles('src\port',@pgportfiles); - - $postgres = $solution->AddProject('postgres','exe','','src\backend'); - $postgres->AddIncludeDir('src\backend'); - $postgres->AddDir('src\backend\port\win32'); - $postgres->AddFile('src\backend\utils\fmgrtab.c'); - $postgres->ReplaceFile('src\backend\port\dynloader.c','src\backend\port\dynloader\win32.c'); - $postgres->ReplaceFile('src\backend\port\pg_sema.c','src\backend\port\win32_sema.c'); - $postgres->ReplaceFile('src\backend\port\pg_shmem.c','src\backend\port\win32_shmem.c'); - $postgres->ReplaceFile('src\backend\port\pg_latch.c','src\backend\port\win32_latch.c'); - $postgres->AddFiles('src\port',@pgportfiles); - $postgres->AddDir('src\timezone'); - $postgres->AddFiles('src\backend\parser','scan.l','gram.y'); - $postgres->AddFiles('src\backend\bootstrap','bootscanner.l','bootparse.y'); - $postgres->AddFiles('src\backend\utils\misc','guc-file.l'); - $postgres->AddFiles('src\backend\replication', 'repl_scanner.l', 'repl_gram.y'); - $postgres->AddDefine('BUILDING_DLL'); - $postgres->AddLibrary('wsock32.lib'); - $postgres->AddLibrary('ws2_32.lib'); - $postgres->AddLibrary('secur32.lib'); - $postgres->AddLibrary('wldap32.lib') if ($solution->{options}->{ldap}); - $postgres->FullExportDLL('postgres.lib'); - - my $snowball = $solution->AddProject('dict_snowball','dll','','src\backend\snowball'); - $snowball->RelocateFiles( - 'src\backend\snowball\libstemmer', - sub { - return shift !~ /dict_snowball.c$/; - } - ); - $snowball->AddIncludeDir('src\include\snowball'); - $snowball->AddReference($postgres); - - my $plpgsql = $solution->AddProject('plpgsql','dll','PLs','src\pl\plpgsql\src'); - $plpgsql->AddFiles('src\pl\plpgsql\src', 'gram.y'); - $plpgsql->AddReference($postgres); - - if ($solution->{options}->{perl}) - { - my $plperlsrc = "src\\pl\\plperl\\"; - my $plperl = $solution->AddProject('plperl','dll','PLs','src\pl\plperl'); - $plperl->AddIncludeDir($solution->{options}->{perl} . '/lib/CORE'); - $plperl->AddDefine('PLPERL_HAVE_UID_GID'); - foreach my $xs ('SPI.xs', 'Util.xs') - { - (my $xsc = $xs) =~ s/\.xs/.c/; - if (Solution::IsNewer("$plperlsrc$xsc","$plperlsrc$xs")) - { - my $xsubppdir = first { -e "$_\\ExtUtils\\xsubpp" } @INC; - print "Building $plperlsrc$xsc...\n"; - system( $solution->{options}->{perl} - . '/bin/perl ' - . "$xsubppdir/ExtUtils/xsubpp -typemap " - . $solution->{options}->{perl} - . '/lib/ExtUtils/typemap ' - . "$plperlsrc$xs " - . ">$plperlsrc$xsc"); - if ((!(-f "$plperlsrc$xsc")) || -z "$plperlsrc$xsc") - { - unlink("$plperlsrc$xsc"); # if zero size - die "Failed to create $xsc.\n"; - } - } - } - if ( Solution::IsNewer('src\pl\plperl\perlchunks.h','src\pl\plperl\plc_perlboot.pl') - ||Solution::IsNewer('src\pl\plperl\perlchunks.h','src\pl\plperl\plc_trusted.pl')) - { - print 'Building src\pl\plperl\perlchunks.h ...' . "\n"; - my $basedir = getcwd; - chdir 'src\pl\plperl'; - system( $solution->{options}->{perl} - . '/bin/perl ' - . 'text2macro.pl ' - . '--strip="^(\#.*|\s*)$$" ' - . 'plc_perlboot.pl plc_trusted.pl ' - . '>perlchunks.h'); - chdir $basedir; - if ((!(-f 'src\pl\plperl\perlchunks.h')) || -z 'src\pl\plperl\perlchunks.h') - { - unlink('src\pl\plperl\perlchunks.h'); # if zero size - die 'Failed to create perlchunks.h' . "\n"; - } - } - if ( Solution::IsNewer('src\pl\plperl\plperl_opmask.h','src\pl\plperl\plperl_opmask.pl')) - { - print 'Building src\pl\plperl\plperl_opmask.h ...' . "\n"; - my $basedir = getcwd; - chdir 'src\pl\plperl'; - system( $solution->{options}->{perl} - . '/bin/perl ' - . 'plperl_opmask.pl ' - . 'plperl_opmask.h'); - chdir $basedir; - if ((!(-f 'src\pl\plperl\plperl_opmask.h')) || -z 'src\pl\plperl\plperl_opmask.h') - { - unlink('src\pl\plperl\plperl_opmask.h'); # if zero size - die 'Failed to create plperl_opmask.h' . "\n"; - } - } - $plperl->AddReference($postgres); - my @perl_libs = - grep {/perl\d+.lib$/ }glob($solution->{options}->{perl} . '\lib\CORE\perl*.lib'); - if (@perl_libs == 1) - { - $plperl->AddLibrary($perl_libs[0]); - } - else - { - die "could not identify perl library version"; - } - } - - if ($solution->{options}->{python}) - { - - # Attempt to get python version and location. - # Assume python.exe in specified dir. - open(P, - $solution->{options}->{python} - . "\\python -c \"import sys;print(sys.prefix);print(str(sys.version_info[0])+str(sys.version_info[1]))\" |" - ) || die "Could not query for python version!\n"; - my $pyprefix = <P>; - chomp($pyprefix); - my $pyver = <P>; - chomp($pyver); - close(P); - - # Sometimes (always?) if python is not present, the execution - # appears to work, but gives no data... - die "Failed to query python for version information\n" - if (!(defined($pyprefix) && defined($pyver))); - - my $pymajorver = substr($pyver, 0, 1); - my $plpython = - $solution->AddProject('plpython' . $pymajorver, 'dll','PLs', 'src\pl\plpython'); - $plpython->AddIncludeDir($pyprefix . '\include'); - $plpython->AddLibrary($pyprefix . "\\Libs\\python$pyver.lib"); - $plpython->AddReference($postgres); - } - - if ($solution->{options}->{tcl}) - { - my $pltcl = $solution->AddProject('pltcl','dll','PLs','src\pl\tcl'); - $pltcl->AddIncludeDir($solution->{options}->{tcl} . '\include'); - $pltcl->AddReference($postgres); - if (-e $solution->{options}->{tcl} . '\lib\tcl85.lib') - { - $pltcl->AddLibrary($solution->{options}->{tcl} . '\lib\tcl85.lib'); - } - else - { - $pltcl->AddLibrary($solution->{options}->{tcl} . '\lib\tcl84.lib'); - } - } - - $libpq = $solution->AddProject('libpq','dll','interfaces','src\interfaces\libpq'); - $libpq->AddDefine('FRONTEND'); - $libpq->AddDefine('UNSAFE_STAT_OK'); - $libpq->AddIncludeDir('src\port'); - $libpq->AddLibrary('wsock32.lib'); - $libpq->AddLibrary('secur32.lib'); - $libpq->AddLibrary('ws2_32.lib'); - $libpq->AddLibrary('wldap32.lib') if ($solution->{options}->{ldap}); - $libpq->UseDef('src\interfaces\libpq\libpqdll.def'); - $libpq->ReplaceFile('src\interfaces\libpq\libpqrc.c','src\interfaces\libpq\libpq.rc'); - $libpq->AddReference($libpgport); - - my $libpqwalreceiver = $solution->AddProject('libpqwalreceiver', 'dll', '', - 'src\backend\replication\libpqwalreceiver'); - $libpqwalreceiver->AddIncludeDir('src\interfaces\libpq'); - $libpqwalreceiver->AddReference($postgres,$libpq); - - my $pgtypes = - $solution->AddProject('libpgtypes','dll','interfaces','src\interfaces\ecpg\pgtypeslib'); - $pgtypes->AddDefine('FRONTEND'); - $pgtypes->AddReference($libpgport); - $pgtypes->UseDef('src\interfaces\ecpg\pgtypeslib\pgtypeslib.def'); - $pgtypes->AddIncludeDir('src\interfaces\ecpg\include'); - - my $libecpg =$solution->AddProject('libecpg','dll','interfaces','src\interfaces\ecpg\ecpglib'); - $libecpg->AddDefine('FRONTEND'); - $libecpg->AddIncludeDir('src\interfaces\ecpg\include'); - $libecpg->AddIncludeDir('src\interfaces\libpq'); - $libecpg->AddIncludeDir('src\port'); - $libecpg->UseDef('src\interfaces\ecpg\ecpglib\ecpglib.def'); - $libecpg->AddLibrary('wsock32.lib'); - $libecpg->AddReference($libpq,$pgtypes,$libpgport); - - my $libecpgcompat = - $solution->AddProject('libecpg_compat','dll','interfaces','src\interfaces\ecpg\compatlib'); - $libecpgcompat->AddIncludeDir('src\interfaces\ecpg\include'); - $libecpgcompat->AddIncludeDir('src\interfaces\libpq'); - $libecpgcompat->UseDef('src\interfaces\ecpg\compatlib\compatlib.def'); - $libecpgcompat->AddReference($pgtypes,$libecpg,$libpgport); - - my $ecpg = $solution->AddProject('ecpg','exe','interfaces','src\interfaces\ecpg\preproc'); - $ecpg->AddIncludeDir('src\interfaces\ecpg\include'); - $ecpg->AddIncludeDir('src\interfaces\libpq'); - $ecpg->AddPrefixInclude('src\interfaces\ecpg\preproc'); - $ecpg->AddFiles('src\interfaces\ecpg\preproc','pgc.l','preproc.y'); - $ecpg->AddDefine('MAJOR_VERSION=4'); - $ecpg->AddDefine('MINOR_VERSION=2'); - $ecpg->AddDefine('PATCHLEVEL=1'); - $ecpg->AddDefine('ECPG_COMPILE'); - $ecpg->AddReference($libpgport); - - my $pgregress_ecpg = $solution->AddProject('pg_regress_ecpg','exe','misc'); - $pgregress_ecpg->AddFile('src\interfaces\ecpg\test\pg_regress_ecpg.c'); - $pgregress_ecpg->AddFile('src\test\regress\pg_regress.c'); - $pgregress_ecpg->AddIncludeDir('src\port'); - $pgregress_ecpg->AddIncludeDir('src\test\regress'); - $pgregress_ecpg->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); - $pgregress_ecpg->AddDefine('FRONTEND'); - $pgregress_ecpg->AddReference($libpgport); - - my $isolation_tester = $solution->AddProject('isolationtester','exe','misc'); - $isolation_tester->AddFile('src\test\isolation\isolationtester.c'); - $isolation_tester->AddFile('src\test\isolation\specparse.y'); - $isolation_tester->AddFile('src\test\isolation\specscanner.l'); - $isolation_tester->AddFile('src\test\isolation\specparse.c'); - $isolation_tester->AddIncludeDir('src\test\isolation'); - $isolation_tester->AddIncludeDir('src\port'); - $isolation_tester->AddIncludeDir('src\test\regress'); - $isolation_tester->AddIncludeDir('src\interfaces\libpq'); - $isolation_tester->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); - $isolation_tester->AddDefine('FRONTEND'); - $isolation_tester->AddLibrary('wsock32.lib'); - $isolation_tester->AddReference($libpq, $libpgport); - - my $pgregress_isolation = $solution->AddProject('pg_isolation_regress','exe','misc'); - $pgregress_isolation->AddFile('src\test\isolation\isolation_main.c'); - $pgregress_isolation->AddFile('src\test\regress\pg_regress.c'); - $pgregress_isolation->AddIncludeDir('src\port'); - $pgregress_isolation->AddIncludeDir('src\test\regress'); - $pgregress_isolation->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); - $pgregress_isolation->AddDefine('FRONTEND'); - $pgregress_isolation->AddReference($libpgport); - - # src/bin - my $initdb = AddSimpleFrontend('initdb'); - $initdb->AddIncludeDir('src\interfaces\libpq'); - $initdb->AddIncludeDir('src\timezone'); - $initdb->AddDefine('FRONTEND'); - $initdb->AddLibrary('wsock32.lib'); - $initdb->AddLibrary('ws2_32.lib'); - - my $pgbasebackup = AddSimpleFrontend('pg_basebackup', 1); - $pgbasebackup->AddFile('src\bin\pg_basebackup\pg_basebackup.c'); - $pgbasebackup->AddLibrary('ws2_32.lib'); - - my $pgreceivexlog = AddSimpleFrontend('pg_basebackup', 1); - $pgreceivexlog->{name} = 'pg_receivexlog'; - $pgreceivexlog->AddFile('src\bin\pg_basebackup\pg_receivexlog.c'); - $pgreceivexlog->AddLibrary('ws2_32.lib'); - - my $pgconfig = AddSimpleFrontend('pg_config'); - - my $pgcontrol = AddSimpleFrontend('pg_controldata'); - - my $pgctl = AddSimpleFrontend('pg_ctl', 1); - - my $pgreset = AddSimpleFrontend('pg_resetxlog'); - - my $pgevent = $solution->AddProject('pgevent','dll','bin'); - $pgevent->AddFiles('src\bin\pgevent','pgevent.c','pgmsgevent.rc'); - $pgevent->AddResourceFile('src\bin\pgevent','Eventlog message formatter'); - $pgevent->RemoveFile('src\bin\pgevent\win32ver.rc'); - $pgevent->UseDef('src\bin\pgevent\pgevent.def'); - $pgevent->DisableLinkerWarnings('4104'); - - my $psql = AddSimpleFrontend('psql', 1); - $psql->AddIncludeDir('src\bin\pg_dump'); - $psql->AddIncludeDir('src\backend'); - $psql->AddFile('src\bin\psql\psqlscan.l'); - - my $pgdump = AddSimpleFrontend('pg_dump', 1); - $pgdump->AddIncludeDir('src\backend'); - $pgdump->AddFile('src\bin\pg_dump\pg_dump.c'); - $pgdump->AddFile('src\bin\pg_dump\common.c'); - $pgdump->AddFile('src\bin\pg_dump\pg_dump_sort.c'); - $pgdump->AddFile('src\bin\pg_dump\keywords.c'); - $pgdump->AddFile('src\backend\parser\kwlookup.c'); - - my $pgdumpall = AddSimpleFrontend('pg_dump', 1); - - # pg_dumpall doesn't use the files in the Makefile's $(OBJS), unlike - # pg_dump and pg_restore. - # So remove their sources from the object, keeping the other setup that - # AddSimpleFrontend() has done. - my @nodumpall = grep { m/src\\bin\\pg_dump\\.*\.c$/ } - keys %{$pgdumpall->{files}}; - delete @{$pgdumpall->{files}}{@nodumpall}; - $pgdumpall->{name} = 'pg_dumpall'; - $pgdumpall->AddIncludeDir('src\backend'); - $pgdumpall->AddFile('src\bin\pg_dump\pg_dumpall.c'); - $pgdumpall->AddFile('src\bin\pg_dump\dumputils.c'); - $pgdumpall->AddFile('src\bin\pg_dump\dumpmem.c'); - $pgdumpall->AddFile('src\bin\pg_dump\keywords.c'); - $pgdumpall->AddFile('src\backend\parser\kwlookup.c'); - - my $pgrestore = AddSimpleFrontend('pg_dump', 1); - $pgrestore->{name} = 'pg_restore'; - $pgrestore->AddIncludeDir('src\backend'); - $pgrestore->AddFile('src\bin\pg_dump\pg_restore.c'); - $pgrestore->AddFile('src\bin\pg_dump\keywords.c'); - $pgrestore->AddFile('src\backend\parser\kwlookup.c'); - - my $zic = $solution->AddProject('zic','exe','utils'); - $zic->AddFiles('src\timezone','zic.c','ialloc.c','scheck.c','localtime.c'); - $zic->AddReference($libpgport); - - if ($solution->{options}->{xml}) - { - $contrib_extraincludes->{'pgxml'} = [ - $solution->{options}->{xml} . '\include', - $solution->{options}->{xslt} . '\include', - $solution->{options}->{iconv} . '\include' - ]; - - $contrib_extralibs->{'pgxml'} = [ - $solution->{options}->{xml} . '\lib\libxml2.lib', - $solution->{options}->{xslt} . '\lib\libxslt.lib' - ]; - } - else - { - push @contrib_excludes,'xml2'; - } - - if (!$solution->{options}->{openssl}) - { - push @contrib_excludes,'sslinfo'; - } - - if ($solution->{options}->{uuid}) - { - $contrib_extraincludes->{'uuid-ossp'} = [ $solution->{options}->{uuid} . '\include' ]; - $contrib_extralibs->{'uuid-ossp'} = [ $solution->{options}->{uuid} . '\lib\uuid.lib' ]; - } - else - { - push @contrib_excludes,'uuid-ossp'; - } - - # Pgcrypto makefile too complex to parse.... - my $pgcrypto = $solution->AddProject('pgcrypto','dll','crypto'); - $pgcrypto->AddFiles( - 'contrib\pgcrypto','pgcrypto.c','px.c','px-hmac.c', - 'px-crypt.c','crypt-gensalt.c','crypt-blowfish.c','crypt-des.c', - 'crypt-md5.c','mbuf.c','pgp.c','pgp-armor.c', - 'pgp-cfb.c','pgp-compress.c','pgp-decrypt.c','pgp-encrypt.c', - 'pgp-info.c','pgp-mpi.c','pgp-pubdec.c','pgp-pubenc.c', - 'pgp-pubkey.c','pgp-s2k.c','pgp-pgsql.c' - ); - if ($solution->{options}->{openssl}) - { - $pgcrypto->AddFiles('contrib\pgcrypto', 'openssl.c','pgp-mpi-openssl.c'); - } - else - { - $pgcrypto->AddFiles( - 'contrib\pgcrypto', 'md5.c','sha1.c','sha2.c', - 'internal.c','internal-sha2.c','blf.c','rijndael.c', - 'fortuna.c','random.c','pgp-mpi-internal.c','imath.c' - ); - } - $pgcrypto->AddReference($postgres); - $pgcrypto->AddLibrary('wsock32.lib'); - my $mf = Project::read_file('contrib/pgcrypto/Makefile'); - GenerateContribSqlFiles('pgcrypto', $mf); - - my $D; - opendir($D, 'contrib') || croak "Could not opendir on contrib!\n"; - while (my $d = readdir($D)) - { - next if ($d =~ /^\./); - next unless (-f "contrib/$d/Makefile"); - next if (grep {/^$d$/} @contrib_excludes); - AddContrib($d); - } - closedir($D); - - $mf = Project::read_file('src\backend\utils\mb\conversion_procs\Makefile'); - $mf =~ s{\\s*[\r\n]+}{}mg; - $mf =~ m{SUBDIRS\s*=\s*(.*)$}m || die 'Could not match in conversion makefile' . "\n"; - foreach my $sub (split /\s+/,$1) - { - my $mf = Project::read_file('src\backend\utils\mb\conversion_procs\\' . $sub . '\Makefile'); - my $p = $solution->AddProject($sub, 'dll', 'conversion procs'); - $p->AddFile('src\backend\utils\mb\conversion_procs\\' . $sub . '\\' . $sub . '.c'); - if ($mf =~ m{^SRCS\s*\+=\s*(.*)$}m) - { - $p->AddFile('src\backend\utils\mb\conversion_procs\\' . $sub . '\\' . $1); - } - $p->AddReference($postgres); - } - - $mf = Project::read_file('src\bin\scripts\Makefile'); - $mf =~ s{\\s*[\r\n]+}{}mg; - $mf =~ m{PROGRAMS\s*=\s*(.*)$}m || die 'Could not match in bin\scripts\Makefile' . "\n"; - foreach my $prg (split /\s+/,$1) - { - my $proj = $solution->AddProject($prg,'exe','bin'); - $mf =~ m{$prg\s*:\s*(.*)$}m || die 'Could not find script define for $prg' . "\n"; - my @files = split /\s+/,$1; - foreach my $f (@files) - { - $f =~ s/\.o$/\.c/; - if ($f eq 'keywords.c') - { - $proj->AddFile('src\bin\pg_dump\keywords.c'); - } - elsif ($f eq 'kwlookup.c') - { - $proj->AddFile('src\backend\parser\kwlookup.c'); - } - elsif ($f eq 'dumputils.c') - { - $proj->AddFile('src\bin\pg_dump\dumputils.c'); - } - elsif ($f =~ /print\.c$/) - { # Also catches mbprint.c - $proj->AddFile('src\bin\psql\\' . $f); - } - elsif ($f =~ /\.c$/) - { - $proj->AddFile('src\bin\scripts\\' . $f); - } - } - $proj->AddIncludeDir('src\interfaces\libpq'); - $proj->AddIncludeDir('src\bin\pg_dump'); - $proj->AddIncludeDir('src\bin\psql'); - $proj->AddReference($libpq,$libpgport); - $proj->AddResourceFile('src\bin\scripts','PostgreSQL Utility'); - } - - # Regression DLL and EXE - my $regress = $solution->AddProject('regress','dll','misc'); - $regress->AddFile('src\test\regress\regress.c'); - $regress->AddReference($postgres); - - my $pgregress = $solution->AddProject('pg_regress','exe','misc'); - $pgregress->AddFile('src\test\regress\pg_regress.c'); - $pgregress->AddFile('src\test\regress\pg_regress_main.c'); - $pgregress->AddIncludeDir('src\port'); - $pgregress->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); - $pgregress->AddReference($libpgport); - - $solution->Save(); - return $solution->{vcver}; + our $config = shift; + + chdir('..\..\..') if (-d '..\msvc' && -d '..\..\..\src'); + die 'Must run from root or msvc directory' unless (-d 'src\tools\msvc' && -d 'src'); + + my $vsVersion = DetermineVisualStudioVersion(); + + $solution = CreateSolution($vsVersion, $config); + + our @pgportfiles = qw( + chklocale.c crypt.c fls.c fseeko.c getrusage.c inet_aton.c random.c + srandom.c getaddrinfo.c gettimeofday.c inet_net_ntop.c kill.c open.c + erand48.c snprintf.c strlcat.c strlcpy.c dirmod.c exec.c noblock.c path.c + pgcheckdir.c pg_crc.c pgmkdirp.c pgsleep.c pgstrcasecmp.c qsort.c qsort_arg.c + sprompt.c thread.c getopt.c getopt_long.c dirent.c rint.c win32env.c + win32error.c win32setlocale.c); + + $libpgport = $solution->AddProject('libpgport','lib','misc'); + $libpgport->AddDefine('FRONTEND'); + $libpgport->AddFiles('src\port',@pgportfiles); + + $postgres = $solution->AddProject('postgres','exe','','src\backend'); + $postgres->AddIncludeDir('src\backend'); + $postgres->AddDir('src\backend\port\win32'); + $postgres->AddFile('src\backend\utils\fmgrtab.c'); + $postgres->ReplaceFile('src\backend\port\dynloader.c','src\backend\port\dynloader\win32.c'); + $postgres->ReplaceFile('src\backend\port\pg_sema.c','src\backend\port\win32_sema.c'); + $postgres->ReplaceFile('src\backend\port\pg_shmem.c','src\backend\port\win32_shmem.c'); + $postgres->ReplaceFile('src\backend\port\pg_latch.c','src\backend\port\win32_latch.c'); + $postgres->AddFiles('src\port',@pgportfiles); + $postgres->AddDir('src\timezone'); + $postgres->AddFiles('src\backend\parser','scan.l','gram.y'); + $postgres->AddFiles('src\backend\bootstrap','bootscanner.l','bootparse.y'); + $postgres->AddFiles('src\backend\utils\misc','guc-file.l'); + $postgres->AddFiles('src\backend\replication', 'repl_scanner.l', 'repl_gram.y'); + $postgres->AddDefine('BUILDING_DLL'); + $postgres->AddLibrary('wsock32.lib'); + $postgres->AddLibrary('ws2_32.lib'); + $postgres->AddLibrary('secur32.lib'); + $postgres->AddLibrary('wldap32.lib') if ($solution->{options}->{ldap}); + $postgres->FullExportDLL('postgres.lib'); + + my $snowball = $solution->AddProject('dict_snowball','dll','','src\backend\snowball'); + $snowball->RelocateFiles( + 'src\backend\snowball\libstemmer', + sub { + return shift !~ /dict_snowball.c$/; + } + ); + $snowball->AddIncludeDir('src\include\snowball'); + $snowball->AddReference($postgres); + + my $plpgsql = $solution->AddProject('plpgsql','dll','PLs','src\pl\plpgsql\src'); + $plpgsql->AddFiles('src\pl\plpgsql\src', 'gram.y'); + $plpgsql->AddReference($postgres); + + if ($solution->{options}->{perl}) + { + my $plperlsrc = "src\\pl\\plperl\\"; + my $plperl = $solution->AddProject('plperl','dll','PLs','src\pl\plperl'); + $plperl->AddIncludeDir($solution->{options}->{perl} . '/lib/CORE'); + $plperl->AddDefine('PLPERL_HAVE_UID_GID'); + foreach my $xs ('SPI.xs', 'Util.xs') + { + (my $xsc = $xs) =~ s/\.xs/.c/; + if (Solution::IsNewer("$plperlsrc$xsc","$plperlsrc$xs")) + { + my $xsubppdir = first { -e "$_\\ExtUtils\\xsubpp" } @INC; + print "Building $plperlsrc$xsc...\n"; + system( $solution->{options}->{perl} + . '/bin/perl ' + . "$xsubppdir/ExtUtils/xsubpp -typemap " + . $solution->{options}->{perl} + . '/lib/ExtUtils/typemap ' + . "$plperlsrc$xs " + . ">$plperlsrc$xsc"); + if ((!(-f "$plperlsrc$xsc")) || -z "$plperlsrc$xsc") + { + unlink("$plperlsrc$xsc"); # if zero size + die "Failed to create $xsc.\n"; + } + } + } + if ( + Solution::IsNewer('src\pl\plperl\perlchunks.h', + 'src\pl\plperl\plc_perlboot.pl') + ||Solution::IsNewer( + 'src\pl\plperl\perlchunks.h','src\pl\plperl\plc_trusted.pl' + ) + ) + { + print 'Building src\pl\plperl\perlchunks.h ...' . "\n"; + my $basedir = getcwd; + chdir 'src\pl\plperl'; + system( $solution->{options}->{perl} + . '/bin/perl ' + . 'text2macro.pl ' + . '--strip="^(\#.*|\s*)$$" ' + . 'plc_perlboot.pl plc_trusted.pl ' + . '>perlchunks.h'); + chdir $basedir; + if ((!(-f 'src\pl\plperl\perlchunks.h')) || -z 'src\pl\plperl\perlchunks.h') + { + unlink('src\pl\plperl\perlchunks.h'); # if zero size + die 'Failed to create perlchunks.h' . "\n"; + } + } + if ( + Solution::IsNewer( + 'src\pl\plperl\plperl_opmask.h', + 'src\pl\plperl\plperl_opmask.pl' + ) + ) + { + print 'Building src\pl\plperl\plperl_opmask.h ...' . "\n"; + my $basedir = getcwd; + chdir 'src\pl\plperl'; + system( $solution->{options}->{perl} + . '/bin/perl ' + . 'plperl_opmask.pl ' + . 'plperl_opmask.h'); + chdir $basedir; + if ((!(-f 'src\pl\plperl\plperl_opmask.h')) + || -z 'src\pl\plperl\plperl_opmask.h') + { + unlink('src\pl\plperl\plperl_opmask.h'); # if zero size + die 'Failed to create plperl_opmask.h' . "\n"; + } + } + $plperl->AddReference($postgres); + my @perl_libs = + grep {/perl\d+.lib$/ }glob($solution->{options}->{perl} . '\lib\CORE\perl*.lib'); + if (@perl_libs == 1) + { + $plperl->AddLibrary($perl_libs[0]); + } + else + { + die "could not identify perl library version"; + } + } + + if ($solution->{options}->{python}) + { + + # Attempt to get python version and location. + # Assume python.exe in specified dir. + open(P, + $solution->{options}->{python} + . "\\python -c \"import sys;print(sys.prefix);print(str(sys.version_info[0])+str(sys.version_info[1]))\" |" + ) || die "Could not query for python version!\n"; + my $pyprefix = <P>; + chomp($pyprefix); + my $pyver = <P>; + chomp($pyver); + close(P); + + # Sometimes (always?) if python is not present, the execution + # appears to work, but gives no data... + die "Failed to query python for version information\n" + if (!(defined($pyprefix) && defined($pyver))); + + my $pymajorver = substr($pyver, 0, 1); + my $plpython = + $solution->AddProject('plpython' . $pymajorver, 'dll','PLs', 'src\pl\plpython'); + $plpython->AddIncludeDir($pyprefix . '\include'); + $plpython->AddLibrary($pyprefix . "\\Libs\\python$pyver.lib"); + $plpython->AddReference($postgres); + } + + if ($solution->{options}->{tcl}) + { + my $pltcl = $solution->AddProject('pltcl','dll','PLs','src\pl\tcl'); + $pltcl->AddIncludeDir($solution->{options}->{tcl} . '\include'); + $pltcl->AddReference($postgres); + if (-e $solution->{options}->{tcl} . '\lib\tcl85.lib') + { + $pltcl->AddLibrary($solution->{options}->{tcl} . '\lib\tcl85.lib'); + } + else + { + $pltcl->AddLibrary($solution->{options}->{tcl} . '\lib\tcl84.lib'); + } + } + + $libpq = $solution->AddProject('libpq','dll','interfaces','src\interfaces\libpq'); + $libpq->AddDefine('FRONTEND'); + $libpq->AddDefine('UNSAFE_STAT_OK'); + $libpq->AddIncludeDir('src\port'); + $libpq->AddLibrary('wsock32.lib'); + $libpq->AddLibrary('secur32.lib'); + $libpq->AddLibrary('ws2_32.lib'); + $libpq->AddLibrary('wldap32.lib') if ($solution->{options}->{ldap}); + $libpq->UseDef('src\interfaces\libpq\libpqdll.def'); + $libpq->ReplaceFile('src\interfaces\libpq\libpqrc.c','src\interfaces\libpq\libpq.rc'); + $libpq->AddReference($libpgport); + + my $libpqwalreceiver = $solution->AddProject('libpqwalreceiver', 'dll', '', + 'src\backend\replication\libpqwalreceiver'); + $libpqwalreceiver->AddIncludeDir('src\interfaces\libpq'); + $libpqwalreceiver->AddReference($postgres,$libpq); + + my $pgtypes = + $solution->AddProject('libpgtypes','dll','interfaces','src\interfaces\ecpg\pgtypeslib'); + $pgtypes->AddDefine('FRONTEND'); + $pgtypes->AddReference($libpgport); + $pgtypes->UseDef('src\interfaces\ecpg\pgtypeslib\pgtypeslib.def'); + $pgtypes->AddIncludeDir('src\interfaces\ecpg\include'); + + my $libecpg = + $solution->AddProject('libecpg','dll','interfaces','src\interfaces\ecpg\ecpglib'); + $libecpg->AddDefine('FRONTEND'); + $libecpg->AddIncludeDir('src\interfaces\ecpg\include'); + $libecpg->AddIncludeDir('src\interfaces\libpq'); + $libecpg->AddIncludeDir('src\port'); + $libecpg->UseDef('src\interfaces\ecpg\ecpglib\ecpglib.def'); + $libecpg->AddLibrary('wsock32.lib'); + $libecpg->AddReference($libpq,$pgtypes,$libpgport); + + my $libecpgcompat =$solution->AddProject('libecpg_compat','dll','interfaces', + 'src\interfaces\ecpg\compatlib'); + $libecpgcompat->AddIncludeDir('src\interfaces\ecpg\include'); + $libecpgcompat->AddIncludeDir('src\interfaces\libpq'); + $libecpgcompat->UseDef('src\interfaces\ecpg\compatlib\compatlib.def'); + $libecpgcompat->AddReference($pgtypes,$libecpg,$libpgport); + + my $ecpg = $solution->AddProject('ecpg','exe','interfaces','src\interfaces\ecpg\preproc'); + $ecpg->AddIncludeDir('src\interfaces\ecpg\include'); + $ecpg->AddIncludeDir('src\interfaces\libpq'); + $ecpg->AddPrefixInclude('src\interfaces\ecpg\preproc'); + $ecpg->AddFiles('src\interfaces\ecpg\preproc','pgc.l','preproc.y'); + $ecpg->AddDefine('MAJOR_VERSION=4'); + $ecpg->AddDefine('MINOR_VERSION=2'); + $ecpg->AddDefine('PATCHLEVEL=1'); + $ecpg->AddDefine('ECPG_COMPILE'); + $ecpg->AddReference($libpgport); + + my $pgregress_ecpg = $solution->AddProject('pg_regress_ecpg','exe','misc'); + $pgregress_ecpg->AddFile('src\interfaces\ecpg\test\pg_regress_ecpg.c'); + $pgregress_ecpg->AddFile('src\test\regress\pg_regress.c'); + $pgregress_ecpg->AddIncludeDir('src\port'); + $pgregress_ecpg->AddIncludeDir('src\test\regress'); + $pgregress_ecpg->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); + $pgregress_ecpg->AddDefine('FRONTEND'); + $pgregress_ecpg->AddReference($libpgport); + + my $isolation_tester = $solution->AddProject('isolationtester','exe','misc'); + $isolation_tester->AddFile('src\test\isolation\isolationtester.c'); + $isolation_tester->AddFile('src\test\isolation\specparse.y'); + $isolation_tester->AddFile('src\test\isolation\specscanner.l'); + $isolation_tester->AddFile('src\test\isolation\specparse.c'); + $isolation_tester->AddIncludeDir('src\test\isolation'); + $isolation_tester->AddIncludeDir('src\port'); + $isolation_tester->AddIncludeDir('src\test\regress'); + $isolation_tester->AddIncludeDir('src\interfaces\libpq'); + $isolation_tester->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); + $isolation_tester->AddDefine('FRONTEND'); + $isolation_tester->AddLibrary('wsock32.lib'); + $isolation_tester->AddReference($libpq, $libpgport); + + my $pgregress_isolation = $solution->AddProject('pg_isolation_regress','exe','misc'); + $pgregress_isolation->AddFile('src\test\isolation\isolation_main.c'); + $pgregress_isolation->AddFile('src\test\regress\pg_regress.c'); + $pgregress_isolation->AddIncludeDir('src\port'); + $pgregress_isolation->AddIncludeDir('src\test\regress'); + $pgregress_isolation->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); + $pgregress_isolation->AddDefine('FRONTEND'); + $pgregress_isolation->AddReference($libpgport); + + # src/bin + my $initdb = AddSimpleFrontend('initdb'); + $initdb->AddIncludeDir('src\interfaces\libpq'); + $initdb->AddIncludeDir('src\timezone'); + $initdb->AddDefine('FRONTEND'); + $initdb->AddLibrary('wsock32.lib'); + $initdb->AddLibrary('ws2_32.lib'); + + my $pgbasebackup = AddSimpleFrontend('pg_basebackup', 1); + $pgbasebackup->AddFile('src\bin\pg_basebackup\pg_basebackup.c'); + $pgbasebackup->AddLibrary('ws2_32.lib'); + + my $pgreceivexlog = AddSimpleFrontend('pg_basebackup', 1); + $pgreceivexlog->{name} = 'pg_receivexlog'; + $pgreceivexlog->AddFile('src\bin\pg_basebackup\pg_receivexlog.c'); + $pgreceivexlog->AddLibrary('ws2_32.lib'); + + my $pgconfig = AddSimpleFrontend('pg_config'); + + my $pgcontrol = AddSimpleFrontend('pg_controldata'); + + my $pgctl = AddSimpleFrontend('pg_ctl', 1); + + my $pgreset = AddSimpleFrontend('pg_resetxlog'); + + my $pgevent = $solution->AddProject('pgevent','dll','bin'); + $pgevent->AddFiles('src\bin\pgevent','pgevent.c','pgmsgevent.rc'); + $pgevent->AddResourceFile('src\bin\pgevent','Eventlog message formatter'); + $pgevent->RemoveFile('src\bin\pgevent\win32ver.rc'); + $pgevent->UseDef('src\bin\pgevent\pgevent.def'); + $pgevent->DisableLinkerWarnings('4104'); + + my $psql = AddSimpleFrontend('psql', 1); + $psql->AddIncludeDir('src\bin\pg_dump'); + $psql->AddIncludeDir('src\backend'); + $psql->AddFile('src\bin\psql\psqlscan.l'); + + my $pgdump = AddSimpleFrontend('pg_dump', 1); + $pgdump->AddIncludeDir('src\backend'); + $pgdump->AddFile('src\bin\pg_dump\pg_dump.c'); + $pgdump->AddFile('src\bin\pg_dump\common.c'); + $pgdump->AddFile('src\bin\pg_dump\pg_dump_sort.c'); + $pgdump->AddFile('src\bin\pg_dump\keywords.c'); + $pgdump->AddFile('src\backend\parser\kwlookup.c'); + + my $pgdumpall = AddSimpleFrontend('pg_dump', 1); + + # pg_dumpall doesn't use the files in the Makefile's $(OBJS), unlike + # pg_dump and pg_restore. + # So remove their sources from the object, keeping the other setup that + # AddSimpleFrontend() has done. + my @nodumpall = grep { m/src\\bin\\pg_dump\\.*\.c$/ } + keys %{$pgdumpall->{files}}; + delete @{$pgdumpall->{files}}{@nodumpall}; + $pgdumpall->{name} = 'pg_dumpall'; + $pgdumpall->AddIncludeDir('src\backend'); + $pgdumpall->AddFile('src\bin\pg_dump\pg_dumpall.c'); + $pgdumpall->AddFile('src\bin\pg_dump\dumputils.c'); + $pgdumpall->AddFile('src\bin\pg_dump\dumpmem.c'); + $pgdumpall->AddFile('src\bin\pg_dump\keywords.c'); + $pgdumpall->AddFile('src\backend\parser\kwlookup.c'); + + my $pgrestore = AddSimpleFrontend('pg_dump', 1); + $pgrestore->{name} = 'pg_restore'; + $pgrestore->AddIncludeDir('src\backend'); + $pgrestore->AddFile('src\bin\pg_dump\pg_restore.c'); + $pgrestore->AddFile('src\bin\pg_dump\keywords.c'); + $pgrestore->AddFile('src\backend\parser\kwlookup.c'); + + my $zic = $solution->AddProject('zic','exe','utils'); + $zic->AddFiles('src\timezone','zic.c','ialloc.c','scheck.c','localtime.c'); + $zic->AddReference($libpgport); + + if ($solution->{options}->{xml}) + { + $contrib_extraincludes->{'pgxml'} = [ + $solution->{options}->{xml} . '\include', + $solution->{options}->{xslt} . '\include', + $solution->{options}->{iconv} . '\include' + ]; + + $contrib_extralibs->{'pgxml'} = [ + $solution->{options}->{xml} . '\lib\libxml2.lib', + $solution->{options}->{xslt} . '\lib\libxslt.lib' + ]; + } + else + { + push @contrib_excludes,'xml2'; + } + + if (!$solution->{options}->{openssl}) + { + push @contrib_excludes,'sslinfo'; + } + + if ($solution->{options}->{uuid}) + { + $contrib_extraincludes->{'uuid-ossp'} = + [ $solution->{options}->{uuid} . '\include' ]; + $contrib_extralibs->{'uuid-ossp'} = + [ $solution->{options}->{uuid} . '\lib\uuid.lib' ]; + } + else + { + push @contrib_excludes,'uuid-ossp'; + } + + # Pgcrypto makefile too complex to parse.... + my $pgcrypto = $solution->AddProject('pgcrypto','dll','crypto'); + $pgcrypto->AddFiles( + 'contrib\pgcrypto','pgcrypto.c','px.c','px-hmac.c', + 'px-crypt.c','crypt-gensalt.c','crypt-blowfish.c','crypt-des.c', + 'crypt-md5.c','mbuf.c','pgp.c','pgp-armor.c', + 'pgp-cfb.c','pgp-compress.c','pgp-decrypt.c','pgp-encrypt.c', + 'pgp-info.c','pgp-mpi.c','pgp-pubdec.c','pgp-pubenc.c', + 'pgp-pubkey.c','pgp-s2k.c','pgp-pgsql.c' + ); + if ($solution->{options}->{openssl}) + { + $pgcrypto->AddFiles('contrib\pgcrypto', 'openssl.c','pgp-mpi-openssl.c'); + } + else + { + $pgcrypto->AddFiles( + 'contrib\pgcrypto', 'md5.c', + 'sha1.c','sha2.c', + 'internal.c','internal-sha2.c', + 'blf.c','rijndael.c', + 'fortuna.c','random.c', + 'pgp-mpi-internal.c','imath.c' + ); + } + $pgcrypto->AddReference($postgres); + $pgcrypto->AddLibrary('wsock32.lib'); + my $mf = Project::read_file('contrib/pgcrypto/Makefile'); + GenerateContribSqlFiles('pgcrypto', $mf); + + my $D; + opendir($D, 'contrib') || croak "Could not opendir on contrib!\n"; + while (my $d = readdir($D)) + { + next if ($d =~ /^\./); + next unless (-f "contrib/$d/Makefile"); + next if (grep {/^$d$/} @contrib_excludes); + AddContrib($d); + } + closedir($D); + + $mf = Project::read_file('src\backend\utils\mb\conversion_procs\Makefile'); + $mf =~ s{\\s*[\r\n]+}{}mg; + $mf =~ m{SUBDIRS\s*=\s*(.*)$}m || die 'Could not match in conversion makefile' . "\n"; + foreach my $sub (split /\s+/,$1) + { + my $mf = Project::read_file( + 'src\backend\utils\mb\conversion_procs\\' . $sub . '\Makefile'); + my $p = $solution->AddProject($sub, 'dll', 'conversion procs'); + $p->AddFile('src\backend\utils\mb\conversion_procs\\' . $sub . '\\' . $sub . '.c'); + if ($mf =~ m{^SRCS\s*\+=\s*(.*)$}m) + { + $p->AddFile('src\backend\utils\mb\conversion_procs\\' . $sub . '\\' . $1); + } + $p->AddReference($postgres); + } + + $mf = Project::read_file('src\bin\scripts\Makefile'); + $mf =~ s{\\s*[\r\n]+}{}mg; + $mf =~ m{PROGRAMS\s*=\s*(.*)$}m || die 'Could not match in bin\scripts\Makefile' . "\n"; + foreach my $prg (split /\s+/,$1) + { + my $proj = $solution->AddProject($prg,'exe','bin'); + $mf =~ m{$prg\s*:\s*(.*)$}m || die 'Could not find script define for $prg' . "\n"; + my @files = split /\s+/,$1; + foreach my $f (@files) + { + $f =~ s/\.o$/\.c/; + if ($f eq 'keywords.c') + { + $proj->AddFile('src\bin\pg_dump\keywords.c'); + } + elsif ($f eq 'kwlookup.c') + { + $proj->AddFile('src\backend\parser\kwlookup.c'); + } + elsif ($f eq 'dumputils.c') + { + $proj->AddFile('src\bin\pg_dump\dumputils.c'); + } + elsif ($f =~ /print\.c$/) + { # Also catches mbprint.c + $proj->AddFile('src\bin\psql\\' . $f); + } + elsif ($f =~ /\.c$/) + { + $proj->AddFile('src\bin\scripts\\' . $f); + } + } + $proj->AddIncludeDir('src\interfaces\libpq'); + $proj->AddIncludeDir('src\bin\pg_dump'); + $proj->AddIncludeDir('src\bin\psql'); + $proj->AddReference($libpq,$libpgport); + $proj->AddResourceFile('src\bin\scripts','PostgreSQL Utility'); + } + + # Regression DLL and EXE + my $regress = $solution->AddProject('regress','dll','misc'); + $regress->AddFile('src\test\regress\regress.c'); + $regress->AddReference($postgres); + + my $pgregress = $solution->AddProject('pg_regress','exe','misc'); + $pgregress->AddFile('src\test\regress\pg_regress.c'); + $pgregress->AddFile('src\test\regress\pg_regress_main.c'); + $pgregress->AddIncludeDir('src\port'); + $pgregress->AddDefine('HOST_TUPLE="i686-pc-win32vc"'); + $pgregress->AddReference($libpgport); + + $solution->Save(); + return $solution->{vcver}; } ##################### @@ -521,174 +539,178 @@ sub mkvcbuild # Add a simple frontend project (exe) sub AddSimpleFrontend { - my $n = shift; - my $uselibpq= shift; - - my $p = $solution->AddProject($n,'exe','bin'); - $p->AddDir('src\bin\\' . $n); - $p->AddReference($libpgport); - if ($uselibpq) - { - $p->AddIncludeDir('src\interfaces\libpq'); - $p->AddReference($libpq); - } - return $p; + my $n = shift; + my $uselibpq= shift; + + my $p = $solution->AddProject($n,'exe','bin'); + $p->AddDir('src\bin\\' . $n); + $p->AddReference($libpgport); + if ($uselibpq) + { + $p->AddIncludeDir('src\interfaces\libpq'); + $p->AddReference($libpq); + } + return $p; } # Add a simple contrib project sub AddContrib { - my $n = shift; - my $mf = Project::read_file('contrib\\' . $n . '\Makefile'); - - if ($mf =~ /^MODULE_big\s*=\s*(.*)$/mg) - { - my $dn = $1; - $mf =~ s{\\\s*[\r\n]+}{}mg; - my $proj = $solution->AddProject($dn, 'dll', 'contrib'); - $mf =~ /^OBJS\s*=\s*(.*)$/gm || croak "Could not find objects in MODULE_big for $n\n"; - my $objs = $1; - while ($objs =~ /\b([\w-]+\.o)\b/g) - { - my $o = $1; - $o =~ s/\.o$/.c/; - $proj->AddFile('contrib\\' . $n . '\\' . $o); - } - $proj->AddReference($postgres); - if ($mf =~ /^SUBDIRS\s*:?=\s*(.*)$/mg) - { - foreach my $d (split /\s+/, $1) - { - my $mf2 = Project::read_file('contrib\\' . $n . '\\' . $d . '\Makefile'); - $mf2 =~ s{\\\s*[\r\n]+}{}mg; - $mf2 =~ /^SUBOBJS\s*=\s*(.*)$/gm - || croak "Could not find objects in MODULE_big for $n, subdir $d\n"; - $objs = $1; - while ($objs =~ /\b([\w-]+\.o)\b/g) - { - my $o = $1; - $o =~ s/\.o$/.c/; - $proj->AddFile('contrib\\' . $n . '\\' . $d . '\\' . $o); - } - } - } - AdjustContribProj($proj); - } - elsif ($mf =~ /^MODULES\s*=\s*(.*)$/mg) - { - foreach my $mod (split /\s+/, $1) - { - my $proj = $solution->AddProject($mod, 'dll', 'contrib'); - $proj->AddFile('contrib\\' . $n . '\\' . $mod . '.c'); - $proj->AddReference($postgres); - AdjustContribProj($proj); - } - } - elsif ($mf =~ /^PROGRAM\s*=\s*(.*)$/mg) - { - my $proj = $solution->AddProject($1, 'exe', 'contrib'); - $mf =~ s{\\\s*[\r\n]+}{}mg; - $mf =~ /^OBJS\s*=\s*(.*)$/gm || croak "Could not find objects in PROGRAM for $n\n"; - my $objs = $1; - while ($objs =~ /\b([\w-]+\.o)\b/g) - { - my $o = $1; - $o =~ s/\.o$/.c/; - $proj->AddFile('contrib\\' . $n . '\\' . $o); - } - AdjustContribProj($proj); - } - else - { - croak "Could not determine contrib module type for $n\n"; - } - - # Are there any output data files to build? - GenerateContribSqlFiles($n, $mf); + my $n = shift; + my $mf = Project::read_file('contrib\\' . $n . '\Makefile'); + + if ($mf =~ /^MODULE_big\s*=\s*(.*)$/mg) + { + my $dn = $1; + $mf =~ s{\\\s*[\r\n]+}{}mg; + my $proj = $solution->AddProject($dn, 'dll', 'contrib'); + $mf =~ /^OBJS\s*=\s*(.*)$/gm + || croak "Could not find objects in MODULE_big for $n\n"; + my $objs = $1; + while ($objs =~ /\b([\w-]+\.o)\b/g) + { + my $o = $1; + $o =~ s/\.o$/.c/; + $proj->AddFile('contrib\\' . $n . '\\' . $o); + } + $proj->AddReference($postgres); + if ($mf =~ /^SUBDIRS\s*:?=\s*(.*)$/mg) + { + foreach my $d (split /\s+/, $1) + { + my $mf2 = + Project::read_file('contrib\\' . $n . '\\' . $d . '\Makefile'); + $mf2 =~ s{\\\s*[\r\n]+}{}mg; + $mf2 =~ /^SUBOBJS\s*=\s*(.*)$/gm + || croak + "Could not find objects in MODULE_big for $n, subdir $d\n"; + $objs = $1; + while ($objs =~ /\b([\w-]+\.o)\b/g) + { + my $o = $1; + $o =~ s/\.o$/.c/; + $proj->AddFile('contrib\\' . $n . '\\' . $d . '\\' . $o); + } + } + } + AdjustContribProj($proj); + } + elsif ($mf =~ /^MODULES\s*=\s*(.*)$/mg) + { + foreach my $mod (split /\s+/, $1) + { + my $proj = $solution->AddProject($mod, 'dll', 'contrib'); + $proj->AddFile('contrib\\' . $n . '\\' . $mod . '.c'); + $proj->AddReference($postgres); + AdjustContribProj($proj); + } + } + elsif ($mf =~ /^PROGRAM\s*=\s*(.*)$/mg) + { + my $proj = $solution->AddProject($1, 'exe', 'contrib'); + $mf =~ s{\\\s*[\r\n]+}{}mg; + $mf =~ /^OBJS\s*=\s*(.*)$/gm || croak "Could not find objects in PROGRAM for $n\n"; + my $objs = $1; + while ($objs =~ /\b([\w-]+\.o)\b/g) + { + my $o = $1; + $o =~ s/\.o$/.c/; + $proj->AddFile('contrib\\' . $n . '\\' . $o); + } + AdjustContribProj($proj); + } + else + { + croak "Could not determine contrib module type for $n\n"; + } + + # Are there any output data files to build? + GenerateContribSqlFiles($n, $mf); } sub GenerateContribSqlFiles { - my $n = shift; - my $mf = shift; - if ($mf =~ /^DATA_built\s*=\s*(.*)$/mg) - { - my $l = $1; - - # Strip out $(addsuffix) rules - if (index($l, '$(addsuffix ') >= 0) - { - my $pcount = 0; - my $i; - for ($i = index($l, '$(addsuffix ') + 12; $i < length($l); $i++) - { - $pcount++ if (substr($l, $i, 1) eq '('); - $pcount-- if (substr($l, $i, 1) eq ')'); - last if ($pcount < 0); - } - $l = substr($l, 0, index($l, '$(addsuffix ')) . substr($l, $i+1); - } - - foreach my $d (split /\s+/, $l) - { - my $in = "$d.in"; - my $out = "$d"; - - if (Solution::IsNewer("contrib/$n/$out", "contrib/$n/$in")) - { - print "Building $out from $in (contrib/$n)...\n"; - my $cont = Project::read_file("contrib/$n/$in"); - my $dn = $out; - $dn =~ s/\.sql$//; - $cont =~ s/MODULE_PATHNAME/\$libdir\/$dn/g; - my $o; - open($o,">contrib/$n/$out") || croak "Could not write to contrib/$n/$d"; - print $o $cont; - close($o); - } - } - } + my $n = shift; + my $mf = shift; + if ($mf =~ /^DATA_built\s*=\s*(.*)$/mg) + { + my $l = $1; + + # Strip out $(addsuffix) rules + if (index($l, '$(addsuffix ') >= 0) + { + my $pcount = 0; + my $i; + for ($i = index($l, '$(addsuffix ') + 12; $i < length($l); $i++) + { + $pcount++ if (substr($l, $i, 1) eq '('); + $pcount-- if (substr($l, $i, 1) eq ')'); + last if ($pcount < 0); + } + $l = substr($l, 0, index($l, '$(addsuffix ')) . substr($l, $i+1); + } + + foreach my $d (split /\s+/, $l) + { + my $in = "$d.in"; + my $out = "$d"; + + if (Solution::IsNewer("contrib/$n/$out", "contrib/$n/$in")) + { + print "Building $out from $in (contrib/$n)...\n"; + my $cont = Project::read_file("contrib/$n/$in"); + my $dn = $out; + $dn =~ s/\.sql$//; + $cont =~ s/MODULE_PATHNAME/\$libdir\/$dn/g; + my $o; + open($o,">contrib/$n/$out") + || croak "Could not write to contrib/$n/$d"; + print $o $cont; + close($o); + } + } + } } sub AdjustContribProj { - my $proj = shift; - my $n = $proj->{name}; - - if ($contrib_defines->{$n}) - { - foreach my $d ($contrib_defines->{$n}) - { - $proj->AddDefine($d); - } - } - if (grep {/^$n$/} @contrib_uselibpq) - { - $proj->AddIncludeDir('src\interfaces\libpq'); - $proj->AddReference($libpq); - } - if (grep {/^$n$/} @contrib_uselibpgport) - { - $proj->AddReference($libpgport); - } - if ($contrib_extralibs->{$n}) - { - foreach my $l (@{$contrib_extralibs->{$n}}) - { - $proj->AddLibrary($l); - } - } - if ($contrib_extraincludes->{$n}) - { - foreach my $i (@{$contrib_extraincludes->{$n}}) - { - $proj->AddIncludeDir($i); - } - } - if ($contrib_extrasource->{$n}) - { - $proj->AddFiles('contrib\\' . $n, @{$contrib_extrasource->{$n}}); - } + my $proj = shift; + my $n = $proj->{name}; + + if ($contrib_defines->{$n}) + { + foreach my $d ($contrib_defines->{$n}) + { + $proj->AddDefine($d); + } + } + if (grep {/^$n$/} @contrib_uselibpq) + { + $proj->AddIncludeDir('src\interfaces\libpq'); + $proj->AddReference($libpq); + } + if (grep {/^$n$/} @contrib_uselibpgport) + { + $proj->AddReference($libpgport); + } + if ($contrib_extralibs->{$n}) + { + foreach my $l (@{$contrib_extralibs->{$n}}) + { + $proj->AddLibrary($l); + } + } + if ($contrib_extraincludes->{$n}) + { + foreach my $i (@{$contrib_extraincludes->{$n}}) + { + $proj->AddIncludeDir($i); + } + } + if ($contrib_extrasource->{$n}) + { + $proj->AddFiles('contrib\\' . $n, @{$contrib_extrasource->{$n}}); + } } 1; diff --git a/src/tools/msvc/Project.pm b/src/tools/msvc/Project.pm index 98db076e58c78333d36ca72ad6a471243ee13a06..53cfdb1753845963c018fd6ce8261dbec23686ec 100644 --- a/src/tools/msvc/Project.pm +++ b/src/tools/msvc/Project.pm @@ -12,393 +12,394 @@ use File::Basename; sub _new { - my ($classname, $name, $type, $solution) = @_; - my $good_types = { - lib => 1, - exe => 1, - dll => 1, - }; - confess("Bad project type: $type\n") unless exists $good_types->{$type}; - my $self = { - name => $name, - type => $type, - guid => Win32::GuidGen(), - files => {}, - references => [], - libraries => [], - suffixlib => [], - includes => '', - prefixincludes => '', - defines => ';', - solution => $solution, - disablewarnings => '4018;4244;4273;4102;4090;4267', - disablelinkerwarnings => '', - platform => $solution->{platform}, - }; - - bless($self, $classname); - return $self; + my ($classname, $name, $type, $solution) = @_; + my $good_types = { + lib => 1, + exe => 1, + dll => 1, + }; + confess("Bad project type: $type\n") unless exists $good_types->{$type}; + my $self = { + name => $name, + type => $type, + guid => Win32::GuidGen(), + files => {}, + references => [], + libraries => [], + suffixlib => [], + includes => '', + prefixincludes => '', + defines => ';', + solution => $solution, + disablewarnings => '4018;4244;4273;4102;4090;4267', + disablelinkerwarnings => '', + platform => $solution->{platform}, + }; + + bless($self, $classname); + return $self; } sub AddFile { - my ($self, $filename) = @_; + my ($self, $filename) = @_; - $self->{files}->{$filename} = 1; + $self->{files}->{$filename} = 1; } sub AddFiles { - my $self = shift; - my $dir = shift; + my $self = shift; + my $dir = shift; - while (my $f = shift) - { - $self->{files}->{$dir . "\\" . $f} = 1; - } + while (my $f = shift) + { + $self->{files}->{$dir . "\\" . $f} = 1; + } } sub ReplaceFile { - my ($self, $filename, $newname) = @_; - my $re = "\\\\$filename\$"; - - foreach my $file ( keys %{ $self->{files} } ) - { - - # Match complete filename - if ($filename =~ /\\/) - { - if ($file eq $filename) - { - delete $self->{files}{$file}; - $self->{files}{$newname} = 1; - return; - } - } - elsif ($file =~ m/($re)/) - { - delete $self->{files}{$file}; - $self->{files}{"$newname\\$filename"} = 1; - return; - } - } - confess("Could not find file $filename to replace\n"); + my ($self, $filename, $newname) = @_; + my $re = "\\\\$filename\$"; + + foreach my $file (keys %{$self->{files}}) + { + + # Match complete filename + if ($filename =~ /\\/) + { + if ($file eq $filename) + { + delete $self->{files}{$file}; + $self->{files}{$newname} = 1; + return; + } + } + elsif ($file =~ m/($re)/) + { + delete $self->{files}{$file}; + $self->{files}{"$newname\\$filename"} = 1; + return; + } + } + confess("Could not find file $filename to replace\n"); } sub RemoveFile { - my ($self, $filename) = @_; - my $orig = scalar keys %{ $self->{files} }; - delete $self->{files}->{$filename}; - if ($orig > scalar keys %{$self->{files}} ) - { - return; - } - confess("Could not find file $filename to remove\n"); + my ($self, $filename) = @_; + my $orig = scalar keys %{$self->{files}}; + delete $self->{files}->{$filename}; + if ($orig > scalar keys %{$self->{files}}) + { + return; + } + confess("Could not find file $filename to remove\n"); } sub RelocateFiles { - my ($self, $targetdir, $proc) = @_; - foreach my $f (keys %{$self->{files}}) - { - my $r = &$proc($f); - if ($r) - { - $self->RemoveFile($f); - $self->AddFile($targetdir . '\\' . basename($f)); - } - } + my ($self, $targetdir, $proc) = @_; + foreach my $f (keys %{$self->{files}}) + { + my $r = &$proc($f); + if ($r) + { + $self->RemoveFile($f); + $self->AddFile($targetdir . '\\' . basename($f)); + } + } } sub AddReference { - my $self = shift; + my $self = shift; - while (my $ref = shift) - { - push @{$self->{references}},$ref; - $self->AddLibrary("__CFGNAME__\\" . $ref->{name} . "\\" . $ref->{name} . ".lib"); - } + while (my $ref = shift) + { + push @{$self->{references}},$ref; + $self->AddLibrary("__CFGNAME__\\" . $ref->{name} . "\\" . $ref->{name} . ".lib"); + } } sub AddLibrary { - my ($self, $lib, $dbgsuffix) = @_; - - if ($lib =~ m/\s/) - { - $lib = '"' . $lib . """; - } - - push @{$self->{libraries}}, $lib; - if ($dbgsuffix) - { - push @{$self->{suffixlib}}, $lib; - } + my ($self, $lib, $dbgsuffix) = @_; + + if ($lib =~ m/\s/) + { + $lib = '"' . $lib . """; + } + + push @{$self->{libraries}}, $lib; + if ($dbgsuffix) + { + push @{$self->{suffixlib}}, $lib; + } } sub AddIncludeDir { - my ($self, $inc) = @_; + my ($self, $inc) = @_; - if ($self->{includes} ne '') - { - $self->{includes} .= ';'; - } - $self->{includes} .= $inc; + if ($self->{includes} ne '') + { + $self->{includes} .= ';'; + } + $self->{includes} .= $inc; } sub AddPrefixInclude { - my ($self, $inc) = @_; + my ($self, $inc) = @_; - $self->{prefixincludes} = $inc . ';' . $self->{prefixincludes}; + $self->{prefixincludes} = $inc . ';' . $self->{prefixincludes}; } sub AddDefine { - my ($self, $def) = @_; + my ($self, $def) = @_; - $def =~ s/"/""/g; - $self->{defines} .= $def . ';'; + $def =~ s/"/""/g; + $self->{defines} .= $def . ';'; } sub FullExportDLL { - my ($self, $libname) = @_; + my ($self, $libname) = @_; - $self->{builddef} = 1; - $self->{def} = ".\\__CFGNAME__\\$self->{name}\\$self->{name}.def"; - $self->{implib} = "__CFGNAME__\\$self->{name}\\$libname"; + $self->{builddef} = 1; + $self->{def} = ".\\__CFGNAME__\\$self->{name}\\$self->{name}.def"; + $self->{implib} = "__CFGNAME__\\$self->{name}\\$libname"; } sub UseDef { - my ($self, $def) = @_; + my ($self, $def) = @_; - $self->{def} = $def; + $self->{def} = $def; } sub AddDir { - my ($self, $reldir) = @_; - my $MF; - - my $t = $/; - undef $/; - open($MF,"$reldir\\Makefile") - || open($MF,"$reldir\\GNUMakefile") - || croak "Could not open $reldir\\Makefile\n"; - my $mf = <$MF>; - close($MF); - - $mf =~ s{\\\s*[\r\n]+}{}mg; - if ($mf =~ m{^(?:SUB)?DIRS[^=]*=\s*(.*)$}mg) - { - foreach my $subdir (split /\s+/,$1) - { - next - if $subdir eq "\$(top_builddir)/src/timezone"; #special case for non-standard include - next - if $reldir . "\\" . $subdir eq "src\\backend\\port\\darwin"; - - $self->AddDir($reldir . "\\" . $subdir); - } - } - while ($mf =~ m{^(?:EXTRA_)?OBJS[^=]*=\s*(.*)$}m) - { - my $s = $1; - my $filter_re = qr{\$\(filter ([^,]+),\s+\$\(([^\)]+)\)\)}; - while ($s =~ /$filter_re/) - { - - # Process $(filter a b c, $(VAR)) expressions - my $list = $1; - my $filter = $2; - $list =~ s/\.o/\.c/g; - my @pieces = split /\s+/, $list; - my $matches = ""; - foreach my $p (@pieces) - { - - if ($filter eq "LIBOBJS") - { - if (grep(/$p/, @main::pgportfiles) == 1) - { - $p =~ s/\.c/\.o/; - $matches .= $p . " "; - } - } - else - { - confess "Unknown filter $filter\n"; - } - } - $s =~ s/$filter_re/$matches/; - } - foreach my $f (split /\s+/,$s) - { - next if $f =~ /^\s*$/; - next if $f eq "\\"; - next if $f =~ /\/SUBSYS.o$/; - $f =~ s/,$//; # Remove trailing comma that can show up from filter stuff - next unless $f =~ /.*\.o$/; - $f =~ s/\.o$/\.c/; - if ($f =~ /^\$\(top_builddir\)\/(.*)/) - { - $f = $1; - $f =~ s/\//\\/g; - $self->{files}->{$f} = 1; - } - else - { - $f =~ s/\//\\/g; - $self->{files}->{"$reldir\\$f"} = 1; - } - } - $mf =~ s{OBJS[^=]*=\s*(.*)$}{}m; - } - - # Match rules that pull in source files from different directories, eg - # pgstrcasecmp.c rint.c snprintf.c: % : $(top_srcdir)/src/port/% - my $replace_re = qr{^([^:\n\$]+\.c)\s*:\s*(?:%\s*: )?\$(\([^\)]+\))\/(.*)\/[^\/]+$}m; - while ($mf =~ m{$replace_re}m) - { - my $match = $1; - my $top = $2; - my $target = $3; - $target =~ s{/}{\\}g; - my @pieces = split /\s+/,$match; - foreach my $fn (@pieces) - { - if ($top eq "(top_srcdir)") - { - eval { $self->ReplaceFile($fn, $target) }; - } - elsif ($top eq "(backend_src)") - { - eval { $self->ReplaceFile($fn, "src\\backend\\$target") }; - } - else - { - confess "Bad replacement top: $top, on line $_\n"; - } - } - $mf =~ s{$replace_re}{}m; - } - - # See if this Makefile contains a description, and should have a RC file - if ($mf =~ /^PGFILEDESC\s*=\s*\"([^\"]+)\"/m) - { - my $desc = $1; - my $ico; - if ($mf =~ /^PGAPPICON\s*=\s*(.*)$/m) { $ico = $1; } - $self->AddResourceFile($reldir,$desc,$ico); - } - $/ = $t; + my ($self, $reldir) = @_; + my $MF; + + my $t = $/; + undef $/; + open($MF,"$reldir\\Makefile") + || open($MF,"$reldir\\GNUMakefile") + || croak "Could not open $reldir\\Makefile\n"; + my $mf = <$MF>; + close($MF); + + $mf =~ s{\\\s*[\r\n]+}{}mg; + if ($mf =~ m{^(?:SUB)?DIRS[^=]*=\s*(.*)$}mg) + { + foreach my $subdir (split /\s+/,$1) + { + next + if $subdir eq "\$(top_builddir)/src/timezone" + ; #special case for non-standard include + next + if $reldir . "\\" . $subdir eq "src\\backend\\port\\darwin"; + + $self->AddDir($reldir . "\\" . $subdir); + } + } + while ($mf =~ m{^(?:EXTRA_)?OBJS[^=]*=\s*(.*)$}m) + { + my $s = $1; + my $filter_re = qr{\$\(filter ([^,]+),\s+\$\(([^\)]+)\)\)}; + while ($s =~ /$filter_re/) + { + + # Process $(filter a b c, $(VAR)) expressions + my $list = $1; + my $filter = $2; + $list =~ s/\.o/\.c/g; + my @pieces = split /\s+/, $list; + my $matches = ""; + foreach my $p (@pieces) + { + + if ($filter eq "LIBOBJS") + { + if (grep(/$p/, @main::pgportfiles) == 1) + { + $p =~ s/\.c/\.o/; + $matches .= $p . " "; + } + } + else + { + confess "Unknown filter $filter\n"; + } + } + $s =~ s/$filter_re/$matches/; + } + foreach my $f (split /\s+/,$s) + { + next if $f =~ /^\s*$/; + next if $f eq "\\"; + next if $f =~ /\/SUBSYS.o$/; + $f =~ s/,$//; # Remove trailing comma that can show up from filter stuff + next unless $f =~ /.*\.o$/; + $f =~ s/\.o$/\.c/; + if ($f =~ /^\$\(top_builddir\)\/(.*)/) + { + $f = $1; + $f =~ s/\//\\/g; + $self->{files}->{$f} = 1; + } + else + { + $f =~ s/\//\\/g; + $self->{files}->{"$reldir\\$f"} = 1; + } + } + $mf =~ s{OBJS[^=]*=\s*(.*)$}{}m; + } + + # Match rules that pull in source files from different directories, eg + # pgstrcasecmp.c rint.c snprintf.c: % : $(top_srcdir)/src/port/% + my $replace_re = qr{^([^:\n\$]+\.c)\s*:\s*(?:%\s*: )?\$(\([^\)]+\))\/(.*)\/[^\/]+$}m; + while ($mf =~ m{$replace_re}m) + { + my $match = $1; + my $top = $2; + my $target = $3; + $target =~ s{/}{\\}g; + my @pieces = split /\s+/,$match; + foreach my $fn (@pieces) + { + if ($top eq "(top_srcdir)") + { + eval { $self->ReplaceFile($fn, $target) }; + } + elsif ($top eq "(backend_src)") + { + eval { $self->ReplaceFile($fn, "src\\backend\\$target") }; + } + else + { + confess "Bad replacement top: $top, on line $_\n"; + } + } + $mf =~ s{$replace_re}{}m; + } + + # See if this Makefile contains a description, and should have a RC file + if ($mf =~ /^PGFILEDESC\s*=\s*\"([^\"]+)\"/m) + { + my $desc = $1; + my $ico; + if ($mf =~ /^PGAPPICON\s*=\s*(.*)$/m) { $ico = $1; } + $self->AddResourceFile($reldir,$desc,$ico); + } + $/ = $t; } sub AddResourceFile { - my ($self, $dir, $desc, $ico) = @_; - - my ($sec,$min,$hour,$mday,$mon,$year,$wday,$yday,$isdst) = localtime(time); - my $d = ($year - 100) . "$yday"; - - if (Solution::IsNewer("$dir\\win32ver.rc",'src\port\win32ver.rc')) - { - print "Generating win32ver.rc for $dir\n"; - open(I,'src\port\win32ver.rc') || confess "Could not open win32ver.rc"; - open(O,">$dir\\win32ver.rc") || confess "Could not write win32ver.rc"; - my $icostr = $ico?"IDI_ICON ICON \"src/port/$ico.ico\"":""; - while (<I>) - { - s/FILEDESC/"$desc"/gm; - s/_ICO_/$icostr/gm; - s/(VERSION.*),0/$1,$d/; - if ($self->{type} eq "dll") - { - s/VFT_APP/VFT_DLL/gm; - } - print O; - } - } - close(O); - close(I); - $self->AddFile("$dir\\win32ver.rc"); + my ($self, $dir, $desc, $ico) = @_; + + my ($sec,$min,$hour,$mday,$mon,$year,$wday,$yday,$isdst) = localtime(time); + my $d = ($year - 100) . "$yday"; + + if (Solution::IsNewer("$dir\\win32ver.rc",'src\port\win32ver.rc')) + { + print "Generating win32ver.rc for $dir\n"; + open(I,'src\port\win32ver.rc') || confess "Could not open win32ver.rc"; + open(O,">$dir\\win32ver.rc") || confess "Could not write win32ver.rc"; + my $icostr = $ico?"IDI_ICON ICON \"src/port/$ico.ico\"":""; + while (<I>) + { + s/FILEDESC/"$desc"/gm; + s/_ICO_/$icostr/gm; + s/(VERSION.*),0/$1,$d/; + if ($self->{type} eq "dll") + { + s/VFT_APP/VFT_DLL/gm; + } + print O; + } + } + close(O); + close(I); + $self->AddFile("$dir\\win32ver.rc"); } sub DisableLinkerWarnings { - my ($self, $warnings) = @_; + my ($self, $warnings) = @_; - $self->{disablelinkerwarnings} .= ',' unless ($self->{disablelinkerwarnings} eq ''); - $self->{disablelinkerwarnings} .= $warnings; + $self->{disablelinkerwarnings} .= ',' unless ($self->{disablelinkerwarnings} eq ''); + $self->{disablelinkerwarnings} .= $warnings; } sub Save { - my ($self) = @_; - - # If doing DLL and haven't specified a DEF file, do a full export of all symbols - # in the project. - if ($self->{type} eq "dll" && !$self->{def}) - { - $self->FullExportDLL($self->{name} . ".lib"); - } - - # Warning 4197 is about double exporting, disable this per - # http://connect.microsoft.com/VisualStudio/feedback/ViewFeedback.aspx?FeedbackID=99193 - $self->DisableLinkerWarnings('4197') if ($self->{platform} eq 'x64'); - - # Dump the project - open(F, ">$self->{name}$self->{filenameExtension}") - || croak("Could not write to $self->{name}$self->{filenameExtension}\n"); - $self->WriteHeader(*F); - $self->WriteFiles(*F); - $self->Footer(*F); - close(F); + my ($self) = @_; + + # If doing DLL and haven't specified a DEF file, do a full export of all symbols + # in the project. + if ($self->{type} eq "dll" && !$self->{def}) + { + $self->FullExportDLL($self->{name} . ".lib"); + } + + # Warning 4197 is about double exporting, disable this per + # http://connect.microsoft.com/VisualStudio/feedback/ViewFeedback.aspx?FeedbackID=99193 + $self->DisableLinkerWarnings('4197') if ($self->{platform} eq 'x64'); + + # Dump the project + open(F, ">$self->{name}$self->{filenameExtension}") + || croak("Could not write to $self->{name}$self->{filenameExtension}\n"); + $self->WriteHeader(*F); + $self->WriteFiles(*F); + $self->Footer(*F); + close(F); } sub GetAdditionalLinkerDependencies { - my ($self, $cfgname, $seperator) = @_; - my $libcfg = (uc $cfgname eq "RELEASE")?"MD":"MDd"; - my $libs = ''; - foreach my $lib (@{$self->{libraries}}) - { - my $xlib = $lib; - foreach my $slib (@{$self->{suffixlib}}) - { - if ($slib eq $lib) - { - $xlib =~ s/\.lib$/$libcfg.lib/; - last; - } - } - $libs .= $xlib . $seperator; - } - $libs =~ s/.$//; - $libs =~ s/__CFGNAME__/$cfgname/g; - return $libs; + my ($self, $cfgname, $seperator) = @_; + my $libcfg = (uc $cfgname eq "RELEASE")?"MD":"MDd"; + my $libs = ''; + foreach my $lib (@{$self->{libraries}}) + { + my $xlib = $lib; + foreach my $slib (@{$self->{suffixlib}}) + { + if ($slib eq $lib) + { + $xlib =~ s/\.lib$/$libcfg.lib/; + last; + } + } + $libs .= $xlib . $seperator; + } + $libs =~ s/.$//; + $libs =~ s/__CFGNAME__/$cfgname/g; + return $libs; } # Utility function that loads a complete file sub read_file { - my $filename = shift; - my $F; - my $t = $/; + my $filename = shift; + my $F; + my $t = $/; - undef $/; - open($F, $filename) || croak "Could not open file $filename\n"; - my $txt = <$F>; - close($F); - $/ = $t; + undef $/; + open($F, $filename) || croak "Could not open file $filename\n"; + my $txt = <$F>; + close($F); + $/ = $t; - return $txt; + return $txt; } 1; diff --git a/src/tools/msvc/Solution.pm b/src/tools/msvc/Solution.pm index e1d85c85ad030b8481d9729bf10a0c60e7bd9ac0..0c50c05734786af3ca8c5991dd0e5c889dbeef1c 100644 --- a/src/tools/msvc/Solution.pm +++ b/src/tools/msvc/Solution.pm @@ -12,346 +12,359 @@ use VSObjectFactory; sub _new { - my $classname = shift; - my $options = shift; - my $self = { - projects => {}, - options => $options, - numver => '', - strver => '', - vcver => undef, - platform => undef, - }; - bless($self, $classname); - - # integer_datetimes is now the default - $options->{integer_datetimes} = 1 - unless exists $options->{integer_datetimes}; - $options->{float4byval} = 1 - unless exists $options->{float4byval}; - if ($options->{xml}) - { - if (!($options->{xslt} && $options->{iconv})) - { - die "XML requires both XSLT and ICONV\n"; - } - } - $options->{blocksize} = 8 - unless $options->{blocksize}; # undef or 0 means default - die "Bad blocksize $options->{blocksize}" - unless grep {$_ == $options->{blocksize}} (1,2,4,8,16,32); - $options->{segsize} = 1 - unless $options->{segsize}; # undef or 0 means default - # only allow segsize 1 for now, as we can't do large files yet in windows - die "Bad segsize $options->{segsize}" - unless $options->{segsize} == 1; - $options->{wal_blocksize} = 8 - unless $options->{wal_blocksize}; # undef or 0 means default - die "Bad wal_blocksize $options->{wal_blocksize}" - unless grep {$_ == $options->{wal_blocksize}} (1,2,4,8,16,32,64); - $options->{wal_segsize} = 16 - unless $options->{wal_segsize}; # undef or 0 means default - die "Bad wal_segsize $options->{wal_segsize}" - unless grep {$_ == $options->{wal_segsize}} (1,2,4,8,16,32,64); - - $self->DeterminePlatform(); - - return $self; + my $classname = shift; + my $options = shift; + my $self = { + projects => {}, + options => $options, + numver => '', + strver => '', + vcver => undef, + platform => undef, + }; + bless($self, $classname); + + # integer_datetimes is now the default + $options->{integer_datetimes} = 1 + unless exists $options->{integer_datetimes}; + $options->{float4byval} = 1 + unless exists $options->{float4byval}; + if ($options->{xml}) + { + if (!($options->{xslt} && $options->{iconv})) + { + die "XML requires both XSLT and ICONV\n"; + } + } + $options->{blocksize} = 8 + unless $options->{blocksize}; # undef or 0 means default + die "Bad blocksize $options->{blocksize}" + unless grep {$_ == $options->{blocksize}} (1,2,4,8,16,32); + $options->{segsize} = 1 + unless $options->{segsize}; # undef or 0 means default + # only allow segsize 1 for now, as we can't do large files yet in windows + die "Bad segsize $options->{segsize}" + unless $options->{segsize} == 1; + $options->{wal_blocksize} = 8 + unless $options->{wal_blocksize}; # undef or 0 means default + die "Bad wal_blocksize $options->{wal_blocksize}" + unless grep {$_ == $options->{wal_blocksize}} (1,2,4,8,16,32,64); + $options->{wal_segsize} = 16 + unless $options->{wal_segsize}; # undef or 0 means default + die "Bad wal_segsize $options->{wal_segsize}" + unless grep {$_ == $options->{wal_segsize}} (1,2,4,8,16,32,64); + + $self->DeterminePlatform(); + + return $self; } sub DeterminePlatform { - my $self = shift; - - # Determine if we are in 32 or 64-bit mode. Do this by seeing if CL has - # 64-bit only parameters. - $self->{platform} = 'Win32'; - open(P,"cl /? 2>NUL|") || die "cl command not found"; - while (<P>) - { - if (/^\/favor:</) - { - $self->{platform} = 'x64'; - last; - } - } - close(P); - print "Detected hardware platform: $self->{platform}\n"; + my $self = shift; + + # Determine if we are in 32 or 64-bit mode. Do this by seeing if CL has + # 64-bit only parameters. + $self->{platform} = 'Win32'; + open(P,"cl /? 2>NUL|") || die "cl command not found"; + while (<P>) + { + if (/^\/favor:</) + { + $self->{platform} = 'x64'; + last; + } + } + close(P); + print "Detected hardware platform: $self->{platform}\n"; } # Return 1 if $oldfile is newer than $newfile, or if $newfile doesn't exist. # Special case - if config.pl has changed, always return 1 sub IsNewer { - my ($newfile, $oldfile) = @_; - if ($oldfile ne 'src\tools\msvc\config.pl' && $oldfile ne 'src\tools\msvc\config_default.pl') - { - return 1 - if (-f 'src\tools\msvc\config.pl') && IsNewer($newfile, 'src\tools\msvc\config.pl'); - return 1 - if (-f 'src\tools\msvc\config_default.pl') - && IsNewer($newfile, 'src\tools\msvc\config_default.pl'); - } - return 1 if (!(-e $newfile)); - my @nstat = stat($newfile); - my @ostat = stat($oldfile); - return 1 if ($nstat[9] < $ostat[9]); - return 0; + my ($newfile, $oldfile) = @_; + if ( $oldfile ne 'src\tools\msvc\config.pl' + && $oldfile ne 'src\tools\msvc\config_default.pl') + { + return 1 + if (-f 'src\tools\msvc\config.pl') + && IsNewer($newfile, 'src\tools\msvc\config.pl'); + return 1 + if (-f 'src\tools\msvc\config_default.pl') + && IsNewer($newfile, 'src\tools\msvc\config_default.pl'); + } + return 1 if (!(-e $newfile)); + my @nstat = stat($newfile); + my @ostat = stat($oldfile); + return 1 if ($nstat[9] < $ostat[9]); + return 0; } # Copy a file, *not* preserving date. Only works for text files. sub copyFile { - my ($src, $dest) = @_; - open(I,$src) || croak "Could not open $src"; - open(O,">$dest") || croak "Could not open $dest"; - while (<I>) - { - print O; - } - close(I); - close(O); + my ($src, $dest) = @_; + open(I,$src) || croak "Could not open $src"; + open(O,">$dest") || croak "Could not open $dest"; + while (<I>) + { + print O; + } + close(I); + close(O); } sub GenerateFiles { - my $self = shift; - my $bits = $self->{platform} eq 'Win32' ? 32 : 64; - - # Parse configure.in to get version numbers - open(C,"configure.in") || confess("Could not open configure.in for reading\n"); - while (<C>) - { - if (/^AC_INIT\(\[PostgreSQL\], \[([^\]]+)\]/) - { - $self->{strver} = $1; - if ($self->{strver} !~ /^(\d+)\.(\d+)(?:\.(\d+))?/) - { - confess "Bad format of version: $self->{strver}\n"; - } - $self->{numver} = sprintf("%d%02d%02d", $1, $2, $3?$3:0); - $self->{majorver} = sprintf("%d.%d", $1, $2); - } - } - close(C); - confess "Unable to parse configure.in for all variables!" - if ($self->{strver} eq '' || $self->{numver} eq ''); - - if (IsNewer("src\\include\\pg_config_os.h","src\\include\\port\\win32.h")) - { - print "Copying pg_config_os.h...\n"; - copyFile("src\\include\\port\\win32.h","src\\include\\pg_config_os.h"); - } - - if (IsNewer("src\\include\\pg_config.h","src\\include\\pg_config.h.win32")) - { - print "Generating pg_config.h...\n"; - open(I,"src\\include\\pg_config.h.win32") || confess "Could not open pg_config.h.win32\n"; - open(O,">src\\include\\pg_config.h") || confess "Could not write to pg_config.h\n"; - while (<I>) - { - s{PG_VERSION "[^"]+"}{PG_VERSION "$self->{strver}"}; - s{PG_VERSION_NUM \d+}{PG_VERSION_NUM $self->{numver}}; + my $self = shift; + my $bits = $self->{platform} eq 'Win32' ? 32 : 64; + + # Parse configure.in to get version numbers + open(C,"configure.in") || confess("Could not open configure.in for reading\n"); + while (<C>) + { + if (/^AC_INIT\(\[PostgreSQL\], \[([^\]]+)\]/) + { + $self->{strver} = $1; + if ($self->{strver} !~ /^(\d+)\.(\d+)(?:\.(\d+))?/) + { + confess "Bad format of version: $self->{strver}\n"; + } + $self->{numver} = sprintf("%d%02d%02d", $1, $2, $3?$3:0); + $self->{majorver} = sprintf("%d.%d", $1, $2); + } + } + close(C); + confess "Unable to parse configure.in for all variables!" + if ($self->{strver} eq '' || $self->{numver} eq ''); + + if (IsNewer("src\\include\\pg_config_os.h","src\\include\\port\\win32.h")) + { + print "Copying pg_config_os.h...\n"; + copyFile("src\\include\\port\\win32.h","src\\include\\pg_config_os.h"); + } + + if (IsNewer("src\\include\\pg_config.h","src\\include\\pg_config.h.win32")) + { + print "Generating pg_config.h...\n"; + open(I,"src\\include\\pg_config.h.win32") + || confess "Could not open pg_config.h.win32\n"; + open(O,">src\\include\\pg_config.h") || confess "Could not write to pg_config.h\n"; + while (<I>) + { + s{PG_VERSION "[^"]+"}{PG_VERSION "$self->{strver}"}; + s{PG_VERSION_NUM \d+}{PG_VERSION_NUM $self->{numver}}; s{PG_VERSION_STR "[^"]+"}{__STRINGIFY(x) #x\n#define __STRINGIFY2(z) __STRINGIFY(z)\n#define PG_VERSION_STR "PostgreSQL $self->{strver}, compiled by Visual C++ build " __STRINGIFY2(_MSC_VER) ", $bits-bit"}; - print O; - } - print O "#define PG_MAJORVERSION \"$self->{majorver}\"\n"; - print O "#define LOCALEDIR \"/share/locale\"\n" if ($self->{options}->{nls}); - print O "/* defines added by config steps */\n"; - print O "#ifndef IGNORE_CONFIGURED_SETTINGS\n"; - print O "#define USE_ASSERT_CHECKING 1\n" if ($self->{options}->{asserts}); - print O "#define USE_INTEGER_DATETIMES 1\n" if ($self->{options}->{integer_datetimes}); - print O "#define USE_LDAP 1\n" if ($self->{options}->{ldap}); - print O "#define HAVE_LIBZ 1\n" if ($self->{options}->{zlib}); - print O "#define USE_SSL 1\n" if ($self->{options}->{openssl}); - print O "#define ENABLE_NLS 1\n" if ($self->{options}->{nls}); - - print O "#define BLCKSZ ",1024 * $self->{options}->{blocksize},"\n"; - print O "#define RELSEG_SIZE ", - (1024 / $self->{options}->{blocksize}) *$self->{options}->{segsize} * 1024, "\n"; - print O "#define XLOG_BLCKSZ ",1024 * $self->{options}->{wal_blocksize},"\n"; - print O "#define XLOG_SEG_SIZE (",$self->{options}->{wal_segsize}," * 1024 * 1024)\n"; - - if ($self->{options}->{float4byval}) - { - print O "#define USE_FLOAT4_BYVAL 1\n"; - print O "#define FLOAT4PASSBYVAL true\n"; - } - else - { - print O "#define FLOAT4PASSBYVAL false\n"; - } - if ($self->{options}->{float8byval}) - { - print O "#define USE_FLOAT8_BYVAL 1\n"; - print O "#define FLOAT8PASSBYVAL true\n"; - } - else - { - print O "#define FLOAT8PASSBYVAL false\n"; - } - - if ($self->{options}->{uuid}) - { - print O "#define HAVE_UUID_H\n"; - } - if ($self->{options}->{xml}) - { - print O "#define HAVE_LIBXML2\n"; - print O "#define USE_LIBXML\n"; - } - if ($self->{options}->{xslt}) - { - print O "#define HAVE_LIBXSLT\n"; - print O "#define USE_LIBXSLT\n"; - } - if ($self->{options}->{krb5}) - { - print O "#define KRB5 1\n"; - print O "#define HAVE_KRB5_ERROR_TEXT_DATA 1\n"; - print O "#define HAVE_KRB5_TICKET_ENC_PART2 1\n"; - print O "#define HAVE_KRB5_FREE_UNPARSED_NAME 1\n"; - print O "#define ENABLE_GSS 1\n"; - } - if (my $port = $self->{options}->{"--with-pgport"}) - { - print O "#undef DEF_PGPORT\n"; - print O "#undef DEF_PGPORT_STR\n"; - print O "#define DEF_PGPORT $port\n"; - print O "#define DEF_PGPORT_STR \"$port\"\n"; - } - print O "#define VAL_CONFIGURE \"" . $self->GetFakeConfigure() . "\"\n"; - print O "#endif /* IGNORE_CONFIGURED_SETTINGS */\n"; - close(O); - close(I); - } - - $self->GenerateDefFile("src\\interfaces\\libpq\\libpqdll.def", - "src\\interfaces\\libpq\\exports.txt","LIBPQ"); - $self->GenerateDefFile( - "src\\interfaces\\ecpg\\ecpglib\\ecpglib.def", - "src\\interfaces\\ecpg\\ecpglib\\exports.txt", - "LIBECPG" - ); - $self->GenerateDefFile( - "src\\interfaces\\ecpg\\compatlib\\compatlib.def", - "src\\interfaces\\ecpg\\compatlib\\exports.txt", - "LIBECPG_COMPAT" - ); - $self->GenerateDefFile( - "src\\interfaces\\ecpg\\pgtypeslib\\pgtypeslib.def", - "src\\interfaces\\ecpg\\pgtypeslib\\exports.txt", - "LIBPGTYPES" - ); - - if (IsNewer('src\backend\utils\fmgrtab.c','src\include\catalog\pg_proc.h')) - { - print "Generating fmgrtab.c and fmgroids.h...\n"; - chdir('src\backend\utils'); - system("perl -I ../catalog Gen_fmgrtab.pl ../../../src/include/catalog/pg_proc.h"); - chdir('..\..\..'); - copyFile('src\backend\utils\fmgroids.h','src\include\utils\fmgroids.h'); - } - - if (IsNewer('src\include\utils\probes.h','src\backend\utils\probes.d')) - { - print "Generating probes.h...\n"; - system( + print O; + } + print O "#define PG_MAJORVERSION \"$self->{majorver}\"\n"; + print O "#define LOCALEDIR \"/share/locale\"\n" if ($self->{options}->{nls}); + print O "/* defines added by config steps */\n"; + print O "#ifndef IGNORE_CONFIGURED_SETTINGS\n"; + print O "#define USE_ASSERT_CHECKING 1\n" if ($self->{options}->{asserts}); + print O "#define USE_INTEGER_DATETIMES 1\n" + if ($self->{options}->{integer_datetimes}); + print O "#define USE_LDAP 1\n" if ($self->{options}->{ldap}); + print O "#define HAVE_LIBZ 1\n" if ($self->{options}->{zlib}); + print O "#define USE_SSL 1\n" if ($self->{options}->{openssl}); + print O "#define ENABLE_NLS 1\n" if ($self->{options}->{nls}); + + print O "#define BLCKSZ ",1024 * $self->{options}->{blocksize},"\n"; + print O "#define RELSEG_SIZE ", + (1024 / $self->{options}->{blocksize}) *$self->{options}->{segsize} * 1024, "\n"; + print O "#define XLOG_BLCKSZ ",1024 * $self->{options}->{wal_blocksize},"\n"; + print O "#define XLOG_SEG_SIZE (",$self->{options}->{wal_segsize}, + " * 1024 * 1024)\n"; + + if ($self->{options}->{float4byval}) + { + print O "#define USE_FLOAT4_BYVAL 1\n"; + print O "#define FLOAT4PASSBYVAL true\n"; + } + else + { + print O "#define FLOAT4PASSBYVAL false\n"; + } + if ($self->{options}->{float8byval}) + { + print O "#define USE_FLOAT8_BYVAL 1\n"; + print O "#define FLOAT8PASSBYVAL true\n"; + } + else + { + print O "#define FLOAT8PASSBYVAL false\n"; + } + + if ($self->{options}->{uuid}) + { + print O "#define HAVE_UUID_H\n"; + } + if ($self->{options}->{xml}) + { + print O "#define HAVE_LIBXML2\n"; + print O "#define USE_LIBXML\n"; + } + if ($self->{options}->{xslt}) + { + print O "#define HAVE_LIBXSLT\n"; + print O "#define USE_LIBXSLT\n"; + } + if ($self->{options}->{krb5}) + { + print O "#define KRB5 1\n"; + print O "#define HAVE_KRB5_ERROR_TEXT_DATA 1\n"; + print O "#define HAVE_KRB5_TICKET_ENC_PART2 1\n"; + print O "#define HAVE_KRB5_FREE_UNPARSED_NAME 1\n"; + print O "#define ENABLE_GSS 1\n"; + } + if (my $port = $self->{options}->{"--with-pgport"}) + { + print O "#undef DEF_PGPORT\n"; + print O "#undef DEF_PGPORT_STR\n"; + print O "#define DEF_PGPORT $port\n"; + print O "#define DEF_PGPORT_STR \"$port\"\n"; + } + print O "#define VAL_CONFIGURE \"" . $self->GetFakeConfigure() . "\"\n"; + print O "#endif /* IGNORE_CONFIGURED_SETTINGS */\n"; + close(O); + close(I); + } + + $self->GenerateDefFile("src\\interfaces\\libpq\\libpqdll.def", + "src\\interfaces\\libpq\\exports.txt","LIBPQ"); + $self->GenerateDefFile( + "src\\interfaces\\ecpg\\ecpglib\\ecpglib.def", + "src\\interfaces\\ecpg\\ecpglib\\exports.txt", + "LIBECPG" + ); + $self->GenerateDefFile( + "src\\interfaces\\ecpg\\compatlib\\compatlib.def", + "src\\interfaces\\ecpg\\compatlib\\exports.txt", + "LIBECPG_COMPAT" + ); + $self->GenerateDefFile( + "src\\interfaces\\ecpg\\pgtypeslib\\pgtypeslib.def", + "src\\interfaces\\ecpg\\pgtypeslib\\exports.txt", + "LIBPGTYPES" + ); + + if (IsNewer('src\backend\utils\fmgrtab.c','src\include\catalog\pg_proc.h')) + { + print "Generating fmgrtab.c and fmgroids.h...\n"; + chdir('src\backend\utils'); + system("perl -I ../catalog Gen_fmgrtab.pl ../../../src/include/catalog/pg_proc.h"); + chdir('..\..\..'); + copyFile('src\backend\utils\fmgroids.h','src\include\utils\fmgroids.h'); + } + + if (IsNewer('src\include\utils\probes.h','src\backend\utils\probes.d')) + { + print "Generating probes.h...\n"; + system( 'psed -f src\backend\utils\Gen_dummy_probes.sed src\backend\utils\probes.d > src\include\utils\probes.h' - ); - } - - if ($self->{options}->{python} - && IsNewer('src\pl\plpython\spiexceptions.h','src\include\backend\errcodes.txt')) - { - print "Generating spiexceptions.h...\n"; - system( + ); + } + + if ($self->{options}->{python} + && IsNewer('src\pl\plpython\spiexceptions.h','src\include\backend\errcodes.txt')) + { + print "Generating spiexceptions.h...\n"; + system( 'perl src\pl\plpython\generate-spiexceptions.pl src\backend\utils\errcodes.txt > src\pl\plpython\spiexceptions.h' - ); - } + ); + } - if (IsNewer('src\include\utils\errcodes.h','src\backend\utils\errcodes.txt')) - { - print "Generating errcodes.h...\n"; - system( + if (IsNewer('src\include\utils\errcodes.h','src\backend\utils\errcodes.txt')) + { + print "Generating errcodes.h...\n"; + system( 'perl src\backend\utils\generate-errcodes.pl src\backend\utils\errcodes.txt > src\backend\utils\errcodes.h' - ); - copyFile('src\backend\utils\errcodes.h','src\include\utils\errcodes.h'); - } - - if (IsNewer('src\pl\plpgsql\src\plerrcodes.h','src\backend\utils\errcodes.txt')) - { - print "Generating plerrcodes.h...\n"; - system( + ); + copyFile('src\backend\utils\errcodes.h','src\include\utils\errcodes.h'); + } + + if (IsNewer('src\pl\plpgsql\src\plerrcodes.h','src\backend\utils\errcodes.txt')) + { + print "Generating plerrcodes.h...\n"; + system( 'perl src\pl\plpgsql\src\generate-plerrcodes.pl src\backend\utils\errcodes.txt > src\pl\plpgsql\src\plerrcodes.h' - ); - } - - if (IsNewer('src\backend\utils\sort\qsort_tuple.c','src\backend\utils\sort\gen_qsort_tuple.pl')) - { - print "Generating qsort_tuple.c...\n"; - system( + ); + } + + if ( + IsNewer( + 'src\backend\utils\sort\qsort_tuple.c', + 'src\backend\utils\sort\gen_qsort_tuple.pl' + ) + ) + { + print "Generating qsort_tuple.c...\n"; + system( 'perl src\backend\utils\sort\gen_qsort_tuple.pl > src\backend\utils\sort\qsort_tuple.c' - ); - } - - if (IsNewer('src\interfaces\libpq\libpq.rc','src\interfaces\libpq\libpq.rc.in')) - { - print "Generating libpq.rc...\n"; - my ($sec,$min,$hour,$mday,$mon,$year,$wday,$yday,$isdst) = localtime(time); - my $d = ($year - 100) . "$yday"; - open(I,'<', 'src\interfaces\libpq\libpq.rc.in') || confess "Could not open libpq.rc.in"; - open(O,'>', 'src\interfaces\libpq\libpq.rc') || confess "Could not open libpq.rc"; - while (<I>) - { - s/(VERSION.*),0/$1,$d/; - print O; - } - close(I); - close(O); - } - - if (IsNewer('src\bin\psql\sql_help.h','src\bin\psql\create_help.pl')) - { - print "Generating sql_help.h...\n"; - chdir('src\bin\psql'); - system("perl create_help.pl ../../../doc/src/sgml/ref sql_help"); - chdir('..\..\..'); - } - - if (IsNewer('src\interfaces\ecpg\preproc\preproc.y','src\backend\parser\gram.y')) - { - print "Generating preproc.y...\n"; - chdir('src\interfaces\ecpg\preproc'); - system('perl parse.pl < ..\..\..\backend\parser\gram.y > preproc.y'); - chdir('..\..\..\..'); - } - - if ( - IsNewer( - 'src\interfaces\ecpg\include\ecpg_config.h', - 'src\interfaces\ecpg\include\ecpg_config.h.in' - ) - ) - { - print "Generating ecpg_config.h...\n"; - open(O,'>','src\interfaces\ecpg\include\ecpg_config.h') - || confess "Could not open ecpg_config.h"; - print O <<EOF; + ); + } + + if (IsNewer('src\interfaces\libpq\libpq.rc','src\interfaces\libpq\libpq.rc.in')) + { + print "Generating libpq.rc...\n"; + my ($sec,$min,$hour,$mday,$mon,$year,$wday,$yday,$isdst) = localtime(time); + my $d = ($year - 100) . "$yday"; + open(I,'<', 'src\interfaces\libpq\libpq.rc.in') + || confess "Could not open libpq.rc.in"; + open(O,'>', 'src\interfaces\libpq\libpq.rc') || confess "Could not open libpq.rc"; + while (<I>) + { + s/(VERSION.*),0/$1,$d/; + print O; + } + close(I); + close(O); + } + + if (IsNewer('src\bin\psql\sql_help.h','src\bin\psql\create_help.pl')) + { + print "Generating sql_help.h...\n"; + chdir('src\bin\psql'); + system("perl create_help.pl ../../../doc/src/sgml/ref sql_help"); + chdir('..\..\..'); + } + + if (IsNewer('src\interfaces\ecpg\preproc\preproc.y','src\backend\parser\gram.y')) + { + print "Generating preproc.y...\n"; + chdir('src\interfaces\ecpg\preproc'); + system('perl parse.pl < ..\..\..\backend\parser\gram.y > preproc.y'); + chdir('..\..\..\..'); + } + + if ( + IsNewer( + 'src\interfaces\ecpg\include\ecpg_config.h', + 'src\interfaces\ecpg\include\ecpg_config.h.in' + ) + ) + { + print "Generating ecpg_config.h...\n"; + open(O,'>','src\interfaces\ecpg\include\ecpg_config.h') + || confess "Could not open ecpg_config.h"; + print O <<EOF; #if (_MSC_VER > 1200) #define HAVE_LONG_LONG_INT_64 #define ENABLE_THREAD_SAFETY 1 EOF - print O "#define USE_INTEGER_DATETIMES 1\n" if ($self->{options}->{integer_datetimes}); - print O "#endif\n"; - close(O); - } - - unless (-f "src\\port\\pg_config_paths.h") - { - print "Generating pg_config_paths.h...\n"; - open(O,'>', 'src\port\pg_config_paths.h') || confess "Could not open pg_config_paths.h"; - print O <<EOF; + print O "#define USE_INTEGER_DATETIMES 1\n" + if ($self->{options}->{integer_datetimes}); + print O "#endif\n"; + close(O); + } + + unless (-f "src\\port\\pg_config_paths.h") + { + print "Generating pg_config_paths.h...\n"; + open(O,'>', 'src\port\pg_config_paths.h') + || confess "Could not open pg_config_paths.h"; + print O <<EOF; #define PGBINDIR "/bin" #define PGSHAREDIR "/share" #define SYSCONFDIR "/etc" @@ -365,149 +378,150 @@ EOF #define HTMLDIR "/doc" #define MANDIR "/man" EOF - close(O); - } - - my $mf = Project::read_file('src\backend\catalog\Makefile'); - $mf =~ s{\\s*[\r\n]+}{}mg; - $mf =~ /^POSTGRES_BKI_SRCS\s*:?=[^,]+,(.*)\)$/gm - || croak "Could not find POSTGRES_BKI_SRCS in Makefile\n"; - my @allbki = split /\s+/, $1; - foreach my $bki (@allbki) - { - next if $bki eq ""; - if (IsNewer('src/backend/catalog/postgres.bki', "src/include/catalog/$bki")) - { - print "Generating postgres.bki and schemapg.h...\n"; - chdir('src\backend\catalog'); - my $bki_srcs = join(' ../../../src/include/catalog/', @allbki); - system( + close(O); + } + + my $mf = Project::read_file('src\backend\catalog\Makefile'); + $mf =~ s{\\s*[\r\n]+}{}mg; + $mf =~ /^POSTGRES_BKI_SRCS\s*:?=[^,]+,(.*)\)$/gm + || croak "Could not find POSTGRES_BKI_SRCS in Makefile\n"; + my @allbki = split /\s+/, $1; + foreach my $bki (@allbki) + { + next if $bki eq ""; + if (IsNewer('src/backend/catalog/postgres.bki', "src/include/catalog/$bki")) + { + print "Generating postgres.bki and schemapg.h...\n"; + chdir('src\backend\catalog'); + my $bki_srcs = join(' ../../../src/include/catalog/', @allbki); + system( "perl genbki.pl -I../../../src/include/catalog --set-version=$self->{majorver} $bki_srcs" - ); - chdir('..\..\..'); - copyFile('src\backend\catalog\schemapg.h', 'src\include\catalog\schemapg.h'); - last; - } - } - - open(O, ">doc/src/sgml/version.sgml") || croak "Could not write to version.sgml\n"; - print O <<EOF; + ); + chdir('..\..\..'); + copyFile('src\backend\catalog\schemapg.h', + 'src\include\catalog\schemapg.h'); + last; + } + } + + open(O, ">doc/src/sgml/version.sgml") || croak "Could not write to version.sgml\n"; + print O <<EOF; <!ENTITY version "$self->{strver}"> <!ENTITY majorversion "$self->{majorver}"> EOF - close(O); + close(O); } sub GenerateDefFile { - my ($self, $deffile, $txtfile, $libname) = @_; - - if (IsNewer($deffile,$txtfile)) - { - print "Generating $deffile...\n"; - open(I,$txtfile) || confess("Could not open $txtfile\n"); - open(O,">$deffile") || confess("Could not open $deffile\n"); - print O "LIBRARY $libname\nEXPORTS\n"; - while (<I>) - { - next if (/^#/); - next if (/^\s*$/); - my ($f, $o) = split; - print O " $f @ $o\n"; - } - close(O); - close(I); - } + my ($self, $deffile, $txtfile, $libname) = @_; + + if (IsNewer($deffile,$txtfile)) + { + print "Generating $deffile...\n"; + open(I,$txtfile) || confess("Could not open $txtfile\n"); + open(O,">$deffile") || confess("Could not open $deffile\n"); + print O "LIBRARY $libname\nEXPORTS\n"; + while (<I>) + { + next if (/^#/); + next if (/^\s*$/); + my ($f, $o) = split; + print O " $f @ $o\n"; + } + close(O); + close(I); + } } sub AddProject { - my ($self, $name, $type, $folder, $initialdir) = @_; - - my $proj = VSObjectFactory::CreateProject($self->{vcver}, $name, $type, $self); - push @{$self->{projects}->{$folder}}, $proj; - $proj->AddDir($initialdir) if ($initialdir); - if ($self->{options}->{zlib}) - { - $proj->AddIncludeDir($self->{options}->{zlib} . '\include'); - $proj->AddLibrary($self->{options}->{zlib} . '\lib\zdll.lib'); - } - if ($self->{options}->{openssl}) - { - $proj->AddIncludeDir($self->{options}->{openssl} . '\include'); - $proj->AddLibrary($self->{options}->{openssl} . '\lib\VC\ssleay32.lib', 1); - $proj->AddLibrary($self->{options}->{openssl} . '\lib\VC\libeay32.lib', 1); - } - if ($self->{options}->{nls}) - { - $proj->AddIncludeDir($self->{options}->{nls} . '\include'); - $proj->AddLibrary($self->{options}->{nls} . '\lib\libintl.lib'); - } - if ($self->{options}->{krb5}) - { - $proj->AddIncludeDir($self->{options}->{krb5} . '\inc\krb5'); - $proj->AddLibrary($self->{options}->{krb5} . '\lib\i386\krb5_32.lib'); - $proj->AddLibrary($self->{options}->{krb5} . '\lib\i386\comerr32.lib'); - $proj->AddLibrary($self->{options}->{krb5} . '\lib\i386\gssapi32.lib'); - } - if ($self->{options}->{iconv}) - { - $proj->AddIncludeDir($self->{options}->{iconv} . '\include'); - $proj->AddLibrary($self->{options}->{iconv} . '\lib\iconv.lib'); - } - if ($self->{options}->{xml}) - { - $proj->AddIncludeDir($self->{options}->{xml} . '\include'); - $proj->AddLibrary($self->{options}->{xml} . '\lib\libxml2.lib'); - } - if ($self->{options}->{xslt}) - { - $proj->AddIncludeDir($self->{options}->{xslt} . '\include'); - $proj->AddLibrary($self->{options}->{xslt} . '\lib\libxslt.lib'); - } - return $proj; + my ($self, $name, $type, $folder, $initialdir) = @_; + + my $proj = VSObjectFactory::CreateProject($self->{vcver}, $name, $type, $self); + push @{$self->{projects}->{$folder}}, $proj; + $proj->AddDir($initialdir) if ($initialdir); + if ($self->{options}->{zlib}) + { + $proj->AddIncludeDir($self->{options}->{zlib} . '\include'); + $proj->AddLibrary($self->{options}->{zlib} . '\lib\zdll.lib'); + } + if ($self->{options}->{openssl}) + { + $proj->AddIncludeDir($self->{options}->{openssl} . '\include'); + $proj->AddLibrary($self->{options}->{openssl} . '\lib\VC\ssleay32.lib', 1); + $proj->AddLibrary($self->{options}->{openssl} . '\lib\VC\libeay32.lib', 1); + } + if ($self->{options}->{nls}) + { + $proj->AddIncludeDir($self->{options}->{nls} . '\include'); + $proj->AddLibrary($self->{options}->{nls} . '\lib\libintl.lib'); + } + if ($self->{options}->{krb5}) + { + $proj->AddIncludeDir($self->{options}->{krb5} . '\inc\krb5'); + $proj->AddLibrary($self->{options}->{krb5} . '\lib\i386\krb5_32.lib'); + $proj->AddLibrary($self->{options}->{krb5} . '\lib\i386\comerr32.lib'); + $proj->AddLibrary($self->{options}->{krb5} . '\lib\i386\gssapi32.lib'); + } + if ($self->{options}->{iconv}) + { + $proj->AddIncludeDir($self->{options}->{iconv} . '\include'); + $proj->AddLibrary($self->{options}->{iconv} . '\lib\iconv.lib'); + } + if ($self->{options}->{xml}) + { + $proj->AddIncludeDir($self->{options}->{xml} . '\include'); + $proj->AddLibrary($self->{options}->{xml} . '\lib\libxml2.lib'); + } + if ($self->{options}->{xslt}) + { + $proj->AddIncludeDir($self->{options}->{xslt} . '\include'); + $proj->AddLibrary($self->{options}->{xslt} . '\lib\libxslt.lib'); + } + return $proj; } sub Save { - my ($self) = @_; - my %flduid; - - $self->GenerateFiles(); - foreach my $fld (keys %{$self->{projects}}) - { - foreach my $proj (@{$self->{projects}->{$fld}}) - { - $proj->Save(); - } - } - - open(SLN,">pgsql.sln") || croak "Could not write to pgsql.sln\n"; - print SLN <<EOF; + my ($self) = @_; + my %flduid; + + $self->GenerateFiles(); + foreach my $fld (keys %{$self->{projects}}) + { + foreach my $proj (@{$self->{projects}->{$fld}}) + { + $proj->Save(); + } + } + + open(SLN,">pgsql.sln") || croak "Could not write to pgsql.sln\n"; + print SLN <<EOF; Microsoft Visual Studio Solution File, Format Version $self->{solutionFileVersion} # $self->{visualStudioName} EOF - foreach my $fld (keys %{$self->{projects}}) - { - foreach my $proj (@{$self->{projects}->{$fld}}) - { - print SLN <<EOF; + foreach my $fld (keys %{$self->{projects}}) + { + foreach my $proj (@{$self->{projects}->{$fld}}) + { + print SLN <<EOF; Project("{8BC9CEB8-8B4A-11D0-8D11-00A0C91BC942}") = "$proj->{name}", "$proj->{name}$proj->{filenameExtension}", "$proj->{guid}" EndProject EOF - } - if ($fld ne "") - { - $flduid{$fld} = Win32::GuidGen(); - print SLN <<EOF; + } + if ($fld ne "") + { + $flduid{$fld} = Win32::GuidGen(); + print SLN <<EOF; Project("{2150E333-8FDC-42A3-9474-1A3956D46DE8}") = "$fld", "$fld", "$flduid{$fld}" EndProject EOF - } - } + } + } - print SLN <<EOF; + print SLN <<EOF; Global GlobalSection(SolutionConfigurationPlatforms) = preSolution Debug|$self->{platform}= Debug|$self->{platform} @@ -516,20 +530,20 @@ Global GlobalSection(ProjectConfigurationPlatforms) = postSolution EOF - foreach my $fld (keys %{$self->{projects}}) - { - foreach my $proj (@{$self->{projects}->{$fld}}) - { - print SLN <<EOF; + foreach my $fld (keys %{$self->{projects}}) + { + foreach my $proj (@{$self->{projects}->{$fld}}) + { + print SLN <<EOF; $proj->{guid}.Debug|$self->{platform}.ActiveCfg = Debug|$self->{platform} $proj->{guid}.Debug|$self->{platform}.Build.0 = Debug|$self->{platform} $proj->{guid}.Release|$self->{platform}.ActiveCfg = Release|$self->{platform} $proj->{guid}.Release|$self->{platform}.Build.0 = Release|$self->{platform} EOF - } - } + } + } - print SLN <<EOF; + print SLN <<EOF; EndGlobalSection GlobalSection(SolutionProperties) = preSolution HideSolutionNode = FALSE @@ -537,42 +551,42 @@ EOF GlobalSection(NestedProjects) = preSolution EOF - foreach my $fld (keys %{$self->{projects}}) - { - next if ($fld eq ""); - foreach my $proj (@{$self->{projects}->{$fld}}) - { - print SLN "\t\t$proj->{guid} = $flduid{$fld}\n"; - } - } + foreach my $fld (keys %{$self->{projects}}) + { + next if ($fld eq ""); + foreach my $proj (@{$self->{projects}->{$fld}}) + { + print SLN "\t\t$proj->{guid} = $flduid{$fld}\n"; + } + } - print SLN <<EOF; + print SLN <<EOF; EndGlobalSection EndGlobal EOF - close(SLN); + close(SLN); } sub GetFakeConfigure { - my $self = shift; - - my $cfg = '--enable-thread-safety'; - $cfg .= ' --enable-cassert' if ($self->{options}->{asserts}); - $cfg .= ' --enable-integer-datetimes' if ($self->{options}->{integer_datetimes}); - $cfg .= ' --enable-nls' if ($self->{options}->{nls}); - $cfg .= ' --with-ldap' if ($self->{options}->{ldap}); - $cfg .= ' --without-zlib' unless ($self->{options}->{zlib}); - $cfg .= ' --with-openssl' if ($self->{options}->{ssl}); - $cfg .= ' --with-ossp-uuid' if ($self->{options}->{uuid}); - $cfg .= ' --with-libxml' if ($self->{options}->{xml}); - $cfg .= ' --with-libxslt' if ($self->{options}->{xslt}); - $cfg .= ' --with-krb5' if ($self->{options}->{krb5}); - $cfg .= ' --with-tcl' if ($self->{options}->{tcl}); - $cfg .= ' --with-perl' if ($self->{options}->{perl}); - $cfg .= ' --with-python' if ($self->{options}->{python}); - - return $cfg; + my $self = shift; + + my $cfg = '--enable-thread-safety'; + $cfg .= ' --enable-cassert' if ($self->{options}->{asserts}); + $cfg .= ' --enable-integer-datetimes' if ($self->{options}->{integer_datetimes}); + $cfg .= ' --enable-nls' if ($self->{options}->{nls}); + $cfg .= ' --with-ldap' if ($self->{options}->{ldap}); + $cfg .= ' --without-zlib' unless ($self->{options}->{zlib}); + $cfg .= ' --with-openssl' if ($self->{options}->{ssl}); + $cfg .= ' --with-ossp-uuid' if ($self->{options}->{uuid}); + $cfg .= ' --with-libxml' if ($self->{options}->{xml}); + $cfg .= ' --with-libxslt' if ($self->{options}->{xslt}); + $cfg .= ' --with-krb5' if ($self->{options}->{krb5}); + $cfg .= ' --with-tcl' if ($self->{options}->{tcl}); + $cfg .= ' --with-perl' if ($self->{options}->{perl}); + $cfg .= ' --with-python' if ($self->{options}->{python}); + + return $cfg; } package VS2005Solution; @@ -587,15 +601,15 @@ use base qw(Solution); sub new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{solutionFileVersion} = '9.00'; - $self->{vcver} = '8.00'; - $self->{visualStudioName} = 'Visual Studio 2005'; + $self->{solutionFileVersion} = '9.00'; + $self->{vcver} = '8.00'; + $self->{visualStudioName} = 'Visual Studio 2005'; - return $self; + return $self; } package VS2008Solution; @@ -610,15 +624,15 @@ use base qw(Solution); sub new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{solutionFileVersion} = '10.00'; - $self->{vcver} = '9.00'; - $self->{visualStudioName} = 'Visual Studio 2008'; + $self->{solutionFileVersion} = '10.00'; + $self->{vcver} = '9.00'; + $self->{visualStudioName} = 'Visual Studio 2008'; - return $self; + return $self; } package VS2010Solution; @@ -634,15 +648,15 @@ use base qw(Solution); sub new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{solutionFileVersion} = '11.00'; - $self->{vcver} = '10.00'; - $self->{visualStudioName} = 'Visual Studio 2010'; + $self->{solutionFileVersion} = '11.00'; + $self->{vcver} = '10.00'; + $self->{visualStudioName} = 'Visual Studio 2010'; - return $self; + return $self; } 1; diff --git a/src/tools/msvc/VCBuildProject.pm b/src/tools/msvc/VCBuildProject.pm index 97439d9d5c49374bd3a048814a11aaefecfff980..a7fd0c0d9d53d199e5ee24f29b7deba1a0cc73e1 100644 --- a/src/tools/msvc/VCBuildProject.pm +++ b/src/tools/msvc/VCBuildProject.pm @@ -13,124 +13,124 @@ use base qw(Project); sub _new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{filenameExtension} = '.vcproj'; + $self->{filenameExtension} = '.vcproj'; - return $self; + return $self; } sub WriteHeader { - my ($self, $f) = @_; + my ($self, $f) = @_; - print $f <<EOF; + print $f <<EOF; <?xml version="1.0" encoding="Windows-1252"?> <VisualStudioProject ProjectType="Visual C++" Version="$self->{vcver}" Name="$self->{name}" ProjectGUID="$self->{guid}"> <Platforms><Platform Name="$self->{platform}"/></Platforms> <Configurations> EOF - $self->WriteConfiguration($f, 'Debug', - { defs=>'_DEBUG;DEBUG=1;', wholeopt=>0, opt=>0, strpool=>'false', runtime=>3 }); - $self->WriteConfiguration($f, 'Release', - { defs=>'', wholeopt=>0, opt=>3, strpool=>'true', runtime=>2 }); - print $f <<EOF; + $self->WriteConfiguration($f, 'Debug', + {defs=>'_DEBUG;DEBUG=1;', wholeopt=>0, opt=>0, strpool=>'false', runtime=>3}); + $self->WriteConfiguration($f, 'Release', + {defs=>'', wholeopt=>0, opt=>3, strpool=>'true', runtime=>2}); + print $f <<EOF; </Configurations> EOF - $self->WriteReferences($f); + $self->WriteReferences($f); } sub WriteFiles { - my ($self, $f) = @_; - print $f <<EOF; + my ($self, $f) = @_; + print $f <<EOF; <Files> EOF - my @dirstack = (); - my %uniquefiles; - foreach my $fileNameWithPath (sort keys %{ $self->{files} }) - { - confess "Bad format filename '$fileNameWithPath'\n" - unless ($fileNameWithPath =~ /^(.*)\\([^\\]+)\.[r]?[cyl]$/); - my $dir = $1; - my $file = $2; - - # Walk backwards down the directory stack and close any dirs we're done with - while ($#dirstack >= 0) - { - if (join('\\',@dirstack) eq substr($dir, 0, length(join('\\',@dirstack)))) - { - last if (length($dir) == length(join('\\',@dirstack))); - last if (substr($dir, length(join('\\',@dirstack)),1) eq '\\'); - } - print $f ' ' x $#dirstack . " </Filter>\n"; - pop @dirstack; - } - - # Now walk forwards and create whatever directories are needed - while (join('\\',@dirstack) ne $dir) - { - my $left = substr($dir, length(join('\\',@dirstack))); - $left =~ s/^\\//; - my @pieces = split /\\/, $left; - push @dirstack, $pieces[0]; - print $f ' ' x $#dirstack . " <Filter Name=\"$pieces[0]\" Filter=\"\">\n"; - } - - print $f ' ' x $#dirstack . " <File RelativePath=\"$fileNameWithPath\""; - if ($fileNameWithPath =~ /\.y$/) - { - my $of = $fileNameWithPath; - $of =~ s/\.y$/.c/; - $of =~ s{^src\\pl\\plpgsql\\src\\gram.c$}{src\\pl\\plpgsql\\src\\pl_gram.c}; - print $f '>' - . $self->GenerateCustomTool('Running bison on ' . $fileNameWithPath, - "perl src\\tools\\msvc\\pgbison.pl $fileNameWithPath", $of) - . '</File>' . "\n"; - } - elsif ($fileNameWithPath =~ /\.l$/) - { - my $of = $fileNameWithPath; - $of =~ s/\.l$/.c/; - print $f '>' - . $self->GenerateCustomTool('Running flex on ' . $fileNameWithPath, - "perl src\\tools\\msvc\\pgflex.pl $fileNameWithPath", $of) - . '</File>' . "\n"; - } - elsif (defined($uniquefiles{$file})) - { - - # File already exists, so fake a new name - my $obj = $dir; - $obj =~ s/\\/_/g; - print $f + my @dirstack = (); + my %uniquefiles; + foreach my $fileNameWithPath (sort keys %{$self->{files}}) + { + confess "Bad format filename '$fileNameWithPath'\n" + unless ($fileNameWithPath =~ /^(.*)\\([^\\]+)\.[r]?[cyl]$/); + my $dir = $1; + my $file = $2; + + # Walk backwards down the directory stack and close any dirs we're done with + while ($#dirstack >= 0) + { + if (join('\\',@dirstack) eq substr($dir, 0, length(join('\\',@dirstack)))) + { + last if (length($dir) == length(join('\\',@dirstack))); + last if (substr($dir, length(join('\\',@dirstack)),1) eq '\\'); + } + print $f ' ' x $#dirstack . " </Filter>\n"; + pop @dirstack; + } + + # Now walk forwards and create whatever directories are needed + while (join('\\',@dirstack) ne $dir) + { + my $left = substr($dir, length(join('\\',@dirstack))); + $left =~ s/^\\//; + my @pieces = split /\\/, $left; + push @dirstack, $pieces[0]; + print $f ' ' x $#dirstack . " <Filter Name=\"$pieces[0]\" Filter=\"\">\n"; + } + + print $f ' ' x $#dirstack . " <File RelativePath=\"$fileNameWithPath\""; + if ($fileNameWithPath =~ /\.y$/) + { + my $of = $fileNameWithPath; + $of =~ s/\.y$/.c/; + $of =~ s{^src\\pl\\plpgsql\\src\\gram.c$}{src\\pl\\plpgsql\\src\\pl_gram.c}; + print $f '>' + . $self->GenerateCustomTool('Running bison on ' . $fileNameWithPath, + "perl src\\tools\\msvc\\pgbison.pl $fileNameWithPath", $of) + . '</File>' . "\n"; + } + elsif ($fileNameWithPath =~ /\.l$/) + { + my $of = $fileNameWithPath; + $of =~ s/\.l$/.c/; + print $f '>' + . $self->GenerateCustomTool('Running flex on ' . $fileNameWithPath, + "perl src\\tools\\msvc\\pgflex.pl $fileNameWithPath", $of) + . '</File>' . "\n"; + } + elsif (defined($uniquefiles{$file})) + { + + # File already exists, so fake a new name + my $obj = $dir; + $obj =~ s/\\/_/g; + print $f "><FileConfiguration Name=\"Debug|$self->{platform}\"><Tool Name=\"VCCLCompilerTool\" ObjectFile=\".\\debug\\$self->{name}\\$obj" - . "_$file.obj\" /></FileConfiguration><FileConfiguration Name=\"Release|$self->{platform}\"><Tool Name=\"VCCLCompilerTool\" ObjectFile=\".\\release\\$self->{name}\\$obj" - . "_$file.obj\" /></FileConfiguration></File>\n"; - } - else - { - $uniquefiles{$file} = 1; - print $f " />\n"; - } - } - while ($#dirstack >= 0) - { - print $f ' ' x $#dirstack . " </Filter>\n"; - pop @dirstack; - } - print $f <<EOF; + . "_$file.obj\" /></FileConfiguration><FileConfiguration Name=\"Release|$self->{platform}\"><Tool Name=\"VCCLCompilerTool\" ObjectFile=\".\\release\\$self->{name}\\$obj" + . "_$file.obj\" /></FileConfiguration></File>\n"; + } + else + { + $uniquefiles{$file} = 1; + print $f " />\n"; + } + } + while ($#dirstack >= 0) + { + print $f ' ' x $#dirstack . " </Filter>\n"; + pop @dirstack; + } + print $f <<EOF; </Files> EOF } sub Footer { - my ($self, $f) = @_; + my ($self, $f) = @_; - print $f <<EOF; + print $f <<EOF; <Globals/> </VisualStudioProject> EOF @@ -138,13 +138,13 @@ EOF sub WriteConfiguration { - my ($self, $f, $cfgname, $p) = @_; - my $cfgtype = ($self->{type} eq "exe")?1:($self->{type} eq "dll"?2:4); - my $libs = $self->GetAdditionalLinkerDependencies($cfgname, ' '); + my ($self, $f, $cfgname, $p) = @_; + my $cfgtype = ($self->{type} eq "exe")?1:($self->{type} eq "dll"?2:4); + my $libs = $self->GetAdditionalLinkerDependencies($cfgname, ' '); - my $targetmachine = $self->{platform} eq 'Win32' ? 1 : 17; + my $targetmachine = $self->{platform} eq 'Win32' ? 1 : 17; - print $f <<EOF; + print $f <<EOF; <Configuration Name="$cfgname|$self->{platform}" OutputDirectory=".\\$cfgname\\$self->{name}" IntermediateDirectory=".\\$cfgname\\$self->{name}" ConfigurationType="$cfgtype" UseOfMFC="0" ATLMinimizesCRunTimeLibraryUsage="FALSE" CharacterSet="2" WholeProgramOptimization="$p->{wholeopt}"> <Tool Name="VCCLCompilerTool" Optimization="$p->{opt}" @@ -154,7 +154,7 @@ sub WriteConfiguration RuntimeLibrary="$p->{runtime}" DisableSpecificWarnings="$self->{disablewarnings}" AdditionalOptions="/MP" EOF - print $f <<EOF; + print $f <<EOF; AssemblerOutput="0" AssemblerListingLocation=".\\$cfgname\\$self->{name}\\" ObjectFile=".\\$cfgname\\$self->{name}\\" ProgramDataBaseFileName=".\\$cfgname\\$self->{name}\\" BrowseInformation="0" WarningLevel="3" SuppressStartupBanner="TRUE" DebugInformationFormat="3" CompileAs="0"/> @@ -166,59 +166,59 @@ EOF GenerateMapFile="FALSE" MapFileName=".\\$cfgname\\$self->{name}\\$self->{name}.map" SubSystem="1" TargetMachine="$targetmachine" EOF - if ($self->{disablelinkerwarnings}) - { - print $f "\t\tAdditionalOptions=\"/ignore:$self->{disablelinkerwarnings}\"\n"; - } - if ($self->{implib}) - { - my $l = $self->{implib}; - $l =~ s/__CFGNAME__/$cfgname/g; - print $f "\t\tImportLibrary=\"$l\"\n"; - } - if ($self->{def}) - { - my $d = $self->{def}; - $d =~ s/__CFGNAME__/$cfgname/g; - print $f "\t\tModuleDefinitionFile=\"$d\"\n"; - } - - print $f "\t/>\n"; - print $f + if ($self->{disablelinkerwarnings}) + { + print $f "\t\tAdditionalOptions=\"/ignore:$self->{disablelinkerwarnings}\"\n"; + } + if ($self->{implib}) + { + my $l = $self->{implib}; + $l =~ s/__CFGNAME__/$cfgname/g; + print $f "\t\tImportLibrary=\"$l\"\n"; + } + if ($self->{def}) + { + my $d = $self->{def}; + $d =~ s/__CFGNAME__/$cfgname/g; + print $f "\t\tModuleDefinitionFile=\"$d\"\n"; + } + + print $f "\t/>\n"; + print $f "\t<Tool Name=\"VCLibrarianTool\" OutputFile=\".\\$cfgname\\$self->{name}\\$self->{name}.lib\" IgnoreDefaultLibraryNames=\"libc\" />\n"; - print $f - "\t<Tool Name=\"VCResourceCompilerTool\" AdditionalIncludeDirectories=\"src\\include\" />\n"; - if ($self->{builddef}) - { - print $f + print $f +"\t<Tool Name=\"VCResourceCompilerTool\" AdditionalIncludeDirectories=\"src\\include\" />\n"; + if ($self->{builddef}) + { + print $f "\t<Tool Name=\"VCPreLinkEventTool\" Description=\"Generate DEF file\" CommandLine=\"perl src\\tools\\msvc\\gendef.pl $cfgname\\$self->{name} $self->{platform}\" />\n"; - } - print $f <<EOF; + } + print $f <<EOF; </Configuration> EOF } sub WriteReferences { - my ($self, $f) = @_; - print $f " <References>\n"; - foreach my $ref (@{$self->{references}}) - { - print $f + my ($self, $f) = @_; + print $f " <References>\n"; + foreach my $ref (@{$self->{references}}) + { + print $f " <ProjectReference ReferencedProjectIdentifier=\"$ref->{guid}\" Name=\"$ref->{name}\" />\n"; - } - print $f " </References>\n"; + } + print $f " </References>\n"; } sub GenerateCustomTool { - my ($self, $desc, $tool, $output, $cfg) = @_; - if (!defined($cfg)) - { - return $self->GenerateCustomTool($desc, $tool, $output, 'Debug') - .$self->GenerateCustomTool($desc, $tool, $output, 'Release'); - } - return + my ($self, $desc, $tool, $output, $cfg) = @_; + if (!defined($cfg)) + { + return $self->GenerateCustomTool($desc, $tool, $output, 'Debug') + .$self->GenerateCustomTool($desc, $tool, $output, 'Release'); + } + return "<FileConfiguration Name=\"$cfg|$self->{platform}\"><Tool Name=\"VCCustomBuildTool\" Description=\"$desc\" CommandLine=\"$tool\" AdditionalDependencies=\"\" Outputs=\"$output\" /></FileConfiguration>"; } @@ -234,13 +234,13 @@ use base qw(VCBuildProject); sub new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{vcver} = '8.00'; + $self->{vcver} = '8.00'; - return $self; + return $self; } package VC2008Project; @@ -255,13 +255,13 @@ use base qw(VCBuildProject); sub new { - my $classname = shift; - my $self = $classname->SUPER::_new(@_); - bless($self, $classname); + my $classname = shift; + my $self = $classname->SUPER::_new(@_); + bless($self, $classname); - $self->{vcver} = '9.00'; + $self->{vcver} = '9.00'; - return $self; + return $self; } 1; diff --git a/src/tools/msvc/VSObjectFactory.pm b/src/tools/msvc/VSObjectFactory.pm index 3bed9229918157a89c7d57e8429fcf91b23ec081..e222b04c68111e7e2b9e9f867f17bc42f3f32ad9 100644 --- a/src/tools/msvc/VSObjectFactory.pm +++ b/src/tools/msvc/VSObjectFactory.pm @@ -22,101 +22,103 @@ our (@ISA, @EXPORT); sub CreateSolution { - my $visualStudioVersion = shift; + my $visualStudioVersion = shift; - if (!defined($visualStudioVersion)) - { - $visualStudioVersion = DetermineVisualStudioVersion(); - } + if (!defined($visualStudioVersion)) + { + $visualStudioVersion = DetermineVisualStudioVersion(); + } - if ($visualStudioVersion eq '8.00') - { - return new VS2005Solution(@_); - } - elsif ($visualStudioVersion eq '9.00') - { - return new VS2008Solution(@_); - } - elsif ($visualStudioVersion eq '10.00') - { - return new VS2010Solution(@_); - } - else - { - croak "The requested Visual Studio version is not supported."; - } + if ($visualStudioVersion eq '8.00') + { + return new VS2005Solution(@_); + } + elsif ($visualStudioVersion eq '9.00') + { + return new VS2008Solution(@_); + } + elsif ($visualStudioVersion eq '10.00') + { + return new VS2010Solution(@_); + } + else + { + croak "The requested Visual Studio version is not supported."; + } } sub CreateProject { - my $visualStudioVersion = shift; + my $visualStudioVersion = shift; - if (!defined($visualStudioVersion)) - { - $visualStudioVersion = DetermineVisualStudioVersion(); - } + if (!defined($visualStudioVersion)) + { + $visualStudioVersion = DetermineVisualStudioVersion(); + } - if ($visualStudioVersion eq '8.00') - { - return new VC2005Project(@_); - } - elsif ($visualStudioVersion eq '9.00') - { - return new VC2008Project(@_); - } - elsif ($visualStudioVersion eq '10.00') - { - return new VC2010Project(@_); - } - else - { - croak "The requested Visual Studio version is not supported."; - } + if ($visualStudioVersion eq '8.00') + { + return new VC2005Project(@_); + } + elsif ($visualStudioVersion eq '9.00') + { + return new VC2008Project(@_); + } + elsif ($visualStudioVersion eq '10.00') + { + return new VC2010Project(@_); + } + else + { + croak "The requested Visual Studio version is not supported."; + } } sub DetermineVisualStudioVersion { - my $nmakeVersion = shift; + my $nmakeVersion = shift; - if (!defined($nmakeVersion)) - { + if (!defined($nmakeVersion)) + { - # Determine version of nmake command, to set proper version of visual studio - # we use nmake as it has existed for a long time and still exists in visual studio 2010 - open(P,"nmake /? 2>&1 |") - || croak "Unable to determine Visual Studio version: The nmake command wasn't found."; - while(<P>) - { - chomp; - if (/(\d+)\.(\d+)\.\d+(\.\d+)?$/) - { - return _GetVisualStudioVersion($1, $2); - } - } - close(P); - } - elsif($nmakeVersion =~ /(\d+)\.(\d+)\.\d+(\.\d+)?$/) - { - return _GetVisualStudioVersion($1, $2); - } - croak "Unable to determine Visual Studio version: The nmake version could not be determined."; + # Determine version of nmake command, to set proper version of visual studio + # we use nmake as it has existed for a long time and still exists in visual studio 2010 + open(P,"nmake /? 2>&1 |") + || croak + "Unable to determine Visual Studio version: The nmake command wasn't found."; + while(<P>) + { + chomp; + if (/(\d+)\.(\d+)\.\d+(\.\d+)?$/) + { + return _GetVisualStudioVersion($1, $2); + } + } + close(P); + } + elsif($nmakeVersion =~ /(\d+)\.(\d+)\.\d+(\.\d+)?$/) + { + return _GetVisualStudioVersion($1, $2); + } + croak + "Unable to determine Visual Studio version: The nmake version could not be determined."; } sub _GetVisualStudioVersion { - my($major, $minor) = @_; - if ($major > 10) - { - carp + my($major, $minor) = @_; + if ($major > 10) + { + carp "The determined version of Visual Studio is newer than the latest supported version. Returning the latest supported version instead."; - return '10.00'; - } - elsif ($major < 6) - { - croak + return '10.00'; + } + elsif ($major < 6) + { + croak "Unable to determine Visual Studio version: Visual Studio versions before 6.0 aren't supported."; - } - return "$major.$minor"; + } + return "$major.$minor"; } 1; diff --git a/src/tools/msvc/build.pl b/src/tools/msvc/build.pl index 151849bba512bcef94cce11c3d47dbf2efad7df1..4fa309738b41eba0d81ee979da35d58ff52b5f62 100644 --- a/src/tools/msvc/build.pl +++ b/src/tools/msvc/build.pl @@ -5,7 +5,7 @@ BEGIN { - chdir("../../..") if (-d "../msvc" && -d "../../../src"); + chdir("../../..") if (-d "../msvc" && -d "../../../src"); } @@ -19,13 +19,13 @@ use Mkvcbuild; # it should contain lines like: # $ENV{PATH} = "c:/path/to/bison/bin;$ENV{PATH}"; -if ( -e "src/tools/msvc/buildenv.pl") +if (-e "src/tools/msvc/buildenv.pl") { - require "src/tools/msvc/buildenv.pl"; + require "src/tools/msvc/buildenv.pl"; } -elsif (-e "./buildenv.pl" ) +elsif (-e "./buildenv.pl") { - require "./buildenv.pl"; + require "./buildenv.pl"; } # set up the project @@ -41,26 +41,26 @@ my $bconf = $ENV{CONFIG} || "Release"; my $buildwhat = $ARGV[1] || ""; if ($ARGV[0] eq 'DEBUG') { - $bconf = "Debug"; + $bconf = "Debug"; } elsif ($ARGV[0] ne "RELEASE") { - $buildwhat = $ARGV[0] || ""; + $buildwhat = $ARGV[0] || ""; } # ... and do it if ($buildwhat and $vcver eq '10.00') { - system("msbuild $buildwhat.vcxproj /verbosity:detailed /p:Configuration=$bconf"); + system("msbuild $buildwhat.vcxproj /verbosity:detailed /p:Configuration=$bconf"); } elsif ($buildwhat) { - system("vcbuild $buildwhat.vcproj $bconf"); + system("vcbuild $buildwhat.vcproj $bconf"); } else { - system("msbuild pgsql.sln /verbosity:detailed /p:Configuration=$bconf"); + system("msbuild pgsql.sln /verbosity:detailed /p:Configuration=$bconf"); } # report status diff --git a/src/tools/msvc/builddoc.pl b/src/tools/msvc/builddoc.pl index 8d8d8a35ccf902ff300ec55b831468d0c57d455f..b567f542a744bbb866c546b7886160797f16deca 100644 --- a/src/tools/msvc/builddoc.pl +++ b/src/tools/msvc/builddoc.pl @@ -26,7 +26,7 @@ die "bad DOCROOT '$docroot'" unless ($docroot && -d $docroot); my @notfound; foreach my $dir ('docbook', $openjade, $dsssl) { - push(@notfound,$dir) unless -d "$docroot/$dir"; + push(@notfound,$dir) unless -d "$docroot/$dir"; } missing() if @notfound; @@ -94,28 +94,28 @@ exit; sub renamefiles { - # Rename ISO entity files - my $savedir = getcwd(); - chdir "$docroot/docbook"; - foreach my $f (glob('ISO*')) - { - next if $f =~ /\.gml$/i; - my $nf = $f; - $nf =~ s/ISO(.*)/ISO-$1.gml/; - move $f, $nf; - } - chdir $savedir; + # Rename ISO entity files + my $savedir = getcwd(); + chdir "$docroot/docbook"; + foreach my $f (glob('ISO*')) + { + next if $f =~ /\.gml$/i; + my $nf = $f; + $nf =~ s/ISO(.*)/ISO-$1.gml/; + move $f, $nf; + } + chdir $savedir; } sub missing { - print STDERR "could not find $docroot/$_\n" foreach (@notfound); - exit 1; + print STDERR "could not find $docroot/$_\n" foreach (@notfound); + exit 1; } sub noversion { - print STDERR "Could not find version.sgml. ","Please run mkvcbuild.pl first!\n"; - exit 1; + print STDERR "Could not find version.sgml. ","Please run mkvcbuild.pl first!\n"; + exit 1; } diff --git a/src/tools/msvc/config_default.pl b/src/tools/msvc/config_default.pl index 971d740a9e5ea5a156f2c100d923af77a8e3c698..95e9cd93dab4a9b04db1982b4a3a47bb606275ea 100644 --- a/src/tools/msvc/config_default.pl +++ b/src/tools/msvc/config_default.pl @@ -3,25 +3,25 @@ use strict; use warnings; our $config = { - asserts=>0, # --enable-cassert - # integer_datetimes=>1, # --enable-integer-datetimes - on is now default - # float4byval=>1, # --disable-float4-byval, on by default - # float8byval=>0, # --disable-float8-byval, off by default - # blocksize => 8, # --with-blocksize, 8kB by default - # wal_blocksize => 8, # --with-wal-blocksize, 8kB by default - # wal_segsize => 16, # --with-wal-segsize, 16MB by default - ldap=>1, # --with-ldap - nls=>undef, # --enable-nls=<path> - tcl=>undef, # --with-tls=<path> - perl=>undef, # --with-perl - python=>undef, # --with-python=<path> - krb5=>undef, # --with-krb5=<path> - openssl=>undef, # --with-ssl=<path> - uuid=>undef, # --with-ossp-uuid - xml=>undef, # --with-libxml=<path> - xslt=>undef, # --with-libxslt=<path> - iconv=>undef, # (not in configure, path to iconv) - zlib=>undef # --with-zlib=<path> + asserts=>0, # --enable-cassert + # integer_datetimes=>1, # --enable-integer-datetimes - on is now default + # float4byval=>1, # --disable-float4-byval, on by default + # float8byval=>0, # --disable-float8-byval, off by default + # blocksize => 8, # --with-blocksize, 8kB by default + # wal_blocksize => 8, # --with-wal-blocksize, 8kB by default + # wal_segsize => 16, # --with-wal-segsize, 16MB by default + ldap=>1, # --with-ldap + nls=>undef, # --enable-nls=<path> + tcl=>undef, # --with-tls=<path> + perl=>undef, # --with-perl + python=>undef, # --with-python=<path> + krb5=>undef, # --with-krb5=<path> + openssl=>undef, # --with-ssl=<path> + uuid=>undef, # --with-ossp-uuid + xml=>undef, # --with-libxml=<path> + xslt=>undef, # --with-libxslt=<path> + iconv=>undef, # (not in configure, path to iconv) + zlib=>undef # --with-zlib=<path> }; 1; diff --git a/src/tools/msvc/gendef.pl b/src/tools/msvc/gendef.pl index b8538dd79b8baf21ede87b2ec1aba0276fd3b3d9..2fc8c4a290343f7f3eb234eddbee38a992726ed7 100644 --- a/src/tools/msvc/gendef.pl +++ b/src/tools/msvc/gendef.pl @@ -13,40 +13,40 @@ my $platform = $ARGV[1]; if (-f "$ARGV[0]/$defname.def") { - print "Not re-generating $defname.DEF, file already exists.\n"; - exit(0); + print "Not re-generating $defname.DEF, file already exists.\n"; + exit(0); } print "Generating $defname.DEF from directory $ARGV[0], platform $platform\n"; while (<$ARGV[0]/*.obj>) { - my $symfile = $_; - $symfile=~ s/\.obj$/.sym/i; - print "."; - system("dumpbin /symbols /out:symbols.out $_ >NUL") && die "Could not call dumpbin"; - open(F, "<symbols.out") || die "Could not open symbols.out for $_\n"; - while (<F>) - { - s/\(\)//g; - my @pieces = split; - next unless $pieces[0] =~ /^[A-F0-9]{3,}$/; - next unless $pieces[6]; - next if ($pieces[2] eq "UNDEF"); - next unless ($pieces[4] eq "External"); - next if $pieces[6] =~ /^@/; - next if $pieces[6] =~ /^\(/; - next if $pieces[6] =~ /^__real/; - next if $pieces[6] =~ /^__imp/; - next if $pieces[6] =~ /NULL_THUNK_DATA$/; - next if $pieces[6] =~ /^__IMPORT_DESCRIPTOR/; - next if $pieces[6] =~ /^__NULL_IMPORT/; - next if $pieces[6] =~ /^\?\?_C/; + my $symfile = $_; + $symfile=~ s/\.obj$/.sym/i; + print "."; + system("dumpbin /symbols /out:symbols.out $_ >NUL") && die "Could not call dumpbin"; + open(F, "<symbols.out") || die "Could not open symbols.out for $_\n"; + while (<F>) + { + s/\(\)//g; + my @pieces = split; + next unless $pieces[0] =~ /^[A-F0-9]{3,}$/; + next unless $pieces[6]; + next if ($pieces[2] eq "UNDEF"); + next unless ($pieces[4] eq "External"); + next if $pieces[6] =~ /^@/; + next if $pieces[6] =~ /^\(/; + next if $pieces[6] =~ /^__real/; + next if $pieces[6] =~ /^__imp/; + next if $pieces[6] =~ /NULL_THUNK_DATA$/; + next if $pieces[6] =~ /^__IMPORT_DESCRIPTOR/; + next if $pieces[6] =~ /^__NULL_IMPORT/; + next if $pieces[6] =~ /^\?\?_C/; - push @def, $pieces[6]; - } - close(F); - rename("symbols.out",$symfile); + push @def, $pieces[6]; + } + close(F); + rename("symbols.out",$symfile); } print "\n"; @@ -56,13 +56,13 @@ my $i = 0; my $last = ""; foreach my $f (sort @def) { - next if ($f eq $last); - $last = $f; - $f =~ s/^_// unless ($platform eq "x64"); # win64 has new format of exports - $i++; + next if ($f eq $last); + $last = $f; + $f =~ s/^_// unless ($platform eq "x64"); # win64 has new format of exports + $i++; - # print DEF " $f \@ $i\n"; # ordinaled exports? - print DEF " $f\n"; + # print DEF " $f \@ $i\n"; # ordinaled exports? + print DEF " $f\n"; } close(DEF); print "Generated $i symbols\n"; diff --git a/src/tools/msvc/install.pl b/src/tools/msvc/install.pl index 28563a930d5a7d15ff79deb8fd06e5e6672eace9..f27a7b3f1629f0479afea6b2daf5e3edf73ed14a 100755 --- a/src/tools/msvc/install.pl +++ b/src/tools/msvc/install.pl @@ -13,6 +13,6 @@ Install($target); sub Usage { - print "Usage: install.pl <targetdir>\n"; - exit(1); + print "Usage: install.pl <targetdir>\n"; + exit(1); } diff --git a/src/tools/msvc/pgbison.pl b/src/tools/msvc/pgbison.pl index c48863aff6072fc2893c32115a3b20c183a4abf7..f0c9e26007a857eb267c7744cbaf7ba1c3591ec5 100644 --- a/src/tools/msvc/pgbison.pl +++ b/src/tools/msvc/pgbison.pl @@ -14,21 +14,21 @@ $bisonver=(split(/\s+/,$bisonver))[3]; # grab version number unless ($bisonver eq '1.875' || $bisonver ge '2.2') { - print "WARNING! Bison install not found, or unsupported Bison version.\n"; - print "echo Attempting to build without.\n"; - exit 0; + print "WARNING! Bison install not found, or unsupported Bison version.\n"; + print "echo Attempting to build without.\n"; + exit 0; } my $input = shift; if ($input !~ /\.y$/) { - print "Input must be a .y file\n"; - exit 1; + print "Input must be a .y file\n"; + exit 1; } elsif (!-e $input) { - print "Input file $input not found\n"; - exit 1; + print "Input file $input not found\n"; + exit 1; } (my $output = $input) =~ s/\.y$/.c/; diff --git a/src/tools/msvc/pgflex.pl b/src/tools/msvc/pgflex.pl index 7b9021d2ba279c8320ce8b08daf138e6d58888f2..551b8f67ae86e8a9d6758825cf39bce02037828d 100644 --- a/src/tools/msvc/pgflex.pl +++ b/src/tools/msvc/pgflex.pl @@ -13,26 +13,26 @@ use File::Basename; require 'src/tools/msvc/buildenv.pl' if -e 'src/tools/msvc/buildenv.pl'; my ($flexver) = `flex -V`; # grab first line -$flexver=(split(/\s+/,$flexver))[1]; -$flexver =~ s/[^0-9.]//g; +$flexver=(split(/\s+/,$flexver))[1]; +$flexver =~ s/[^0-9.]//g; my @verparts = split(/\./,$flexver); unless ($verparts[0] == 2 && $verparts[1] == 5 && $verparts[2] >= 31) { - print "WARNING! Flex install not found, or unsupported Flex version.\n"; - print "echo Attempting to build without.\n"; - exit 0; + print "WARNING! Flex install not found, or unsupported Flex version.\n"; + print "echo Attempting to build without.\n"; + exit 0; } my $input = shift; if ($input !~ /\.l$/) { - print "Input must be a .l file\n"; - exit 1; + print "Input must be a .l file\n"; + exit 1; } elsif (!-e $input) { - print "Input file $input not found\n"; - exit 1; + print "Input file $input not found\n"; + exit 1; } (my $output = $input) =~ s/\.l$/.c/; @@ -50,25 +50,25 @@ system("flex $flexflags -o$output $input"); if ($? == 0) { - # For non-reentrant scanners we need to fix up the yywrap macro definition - # to keep the MS compiler happy. - # For reentrant scanners (like the core scanner) we do not - # need to (and must not) change the yywrap definition. - my $lfile; - open($lfile,$input) || die "opening $input for reading: $!"; - my $lcode = <$lfile>; - close($lfile); - if ($lcode !~ /\%option\sreentrant/) - { - my $cfile; - open($cfile,$output) || die "opening $output for reading: $!"; - my $ccode = <$cfile>; - close($cfile); - $ccode =~ s/yywrap\(n\)/yywrap()/; - open($cfile,">$output") || die "opening $output for reading: $!"; - print $cfile $ccode; - close($cfile); - } + # For non-reentrant scanners we need to fix up the yywrap macro definition + # to keep the MS compiler happy. + # For reentrant scanners (like the core scanner) we do not + # need to (and must not) change the yywrap definition. + my $lfile; + open($lfile,$input) || die "opening $input for reading: $!"; + my $lcode = <$lfile>; + close($lfile); + if ($lcode !~ /\%option\sreentrant/) + { + my $cfile; + open($cfile,$output) || die "opening $output for reading: $!"; + my $ccode = <$cfile>; + close($cfile); + $ccode =~ s/yywrap\(n\)/yywrap()/; + open($cfile,">$output") || die "opening $output for reading: $!"; + print $cfile $ccode; + close($cfile); + } if ($flexflags =~ /\s-b\s/) { my $lexback = "lex.backup"; @@ -84,11 +84,11 @@ if ($? == 0) unlink $lexback; } - exit 0; + exit 0; } else { - exit $? >> 8; + exit $? >> 8; } diff --git a/src/tools/msvc/vcregress.pl b/src/tools/msvc/vcregress.pl index f81611bc3d62371f68f3d4cedf4dd5fc42bdeef9..ef7035068b4e9d45e10ff4b8cc907f4947f5fe66 100644 --- a/src/tools/msvc/vcregress.pl +++ b/src/tools/msvc/vcregress.pl @@ -20,19 +20,19 @@ require 'src/tools/msvc/config.pl' if (-f 'src/tools/msvc/config.pl'); # it should contian lines like: # $ENV{PATH} = "c:/path/to/bison/bin;$ENV{PATH}"; -if ( -e "src/tools/msvc/buildenv.pl") +if (-e "src/tools/msvc/buildenv.pl") { - require "src/tools/msvc/buildenv.pl"; + require "src/tools/msvc/buildenv.pl"; } my $what = shift || ""; if ($what =~ /^(check|installcheck|plcheck|contribcheck|ecpgcheck|isolationcheck)$/i) { - $what = uc $what; + $what = uc $what; } else { - usage(); + usage(); } # use a capital C here because config.pl has $config @@ -48,8 +48,8 @@ $ENV{PATH} = "../../../$Config/libpq;../../$Config/libpq;$ENV{PATH}"; my $schedule = shift; unless ($schedule) { - $schedule = "serial"; - $schedule = "parallel" if ($what eq 'CHECK' || $what =~ /PARALLEL/); + $schedule = "serial"; + $schedule = "parallel" if ($what eq 'CHECK' || $what =~ /PARALLEL/); } my $topdir = getcwd(); @@ -67,12 +67,12 @@ $temp_config = "--temp-config=\"$ENV{TEMP_CONFIG}\"" chdir "src/test/regress"; my %command = ( - CHECK => \&check, - PLCHECK => \&plcheck, - INSTALLCHECK => \&installcheck, - ECPGCHECK => \&ecpgcheck, - CONTRIBCHECK => \&contribcheck, - ISOLATIONCHECK => \&isolationcheck, + CHECK => \&check, + PLCHECK => \&plcheck, + INSTALLCHECK => \&installcheck, + ECPGCHECK => \&ecpgcheck, + CONTRIBCHECK => \&contribcheck, + ISOLATIONCHECK => \&isolationcheck, ); my $proc = $command{$what}; @@ -87,229 +87,229 @@ exit 0; sub installcheck { - my @args = ( - "../../../$Config/pg_regress/pg_regress","--dlpath=.", - "--psqldir=../../../$Config/psql","--schedule=${schedule}_schedule", - "--encoding=SQL_ASCII","--no-locale" - ); - push(@args,$maxconn) if $maxconn; - system(@args); - my $status = $? >>8; - exit $status if $status; + my @args = ( + "../../../$Config/pg_regress/pg_regress","--dlpath=.", + "--psqldir=../../../$Config/psql","--schedule=${schedule}_schedule", + "--encoding=SQL_ASCII","--no-locale" + ); + push(@args,$maxconn) if $maxconn; + system(@args); + my $status = $? >>8; + exit $status if $status; } sub check { - my @args = ( - "../../../$Config/pg_regress/pg_regress","--dlpath=.", - "--psqldir=../../../$Config/psql","--schedule=${schedule}_schedule", - "--encoding=SQL_ASCII","--no-locale", - "--temp-install=./tmp_check","--top-builddir=\"$topdir\"" - ); - push(@args,$maxconn) if $maxconn; - push(@args,$temp_config) if $temp_config; - system(@args); - my $status = $? >>8; - exit $status if $status; + my @args = ( + "../../../$Config/pg_regress/pg_regress","--dlpath=.", + "--psqldir=../../../$Config/psql","--schedule=${schedule}_schedule", + "--encoding=SQL_ASCII","--no-locale", + "--temp-install=./tmp_check","--top-builddir=\"$topdir\"" + ); + push(@args,$maxconn) if $maxconn; + push(@args,$temp_config) if $temp_config; + system(@args); + my $status = $? >>8; + exit $status if $status; } sub ecpgcheck { - chdir $startdir; - system("msbuild ecpg_regression.proj /p:config=$Config"); - my $status = $? >>8; - exit $status if $status; - chdir "$topdir/src/interfaces/ecpg/test"; - $schedule="ecpg"; - my @args = ( - "../../../../$Config/pg_regress_ecpg/pg_regress_ecpg", - "--psqldir=../../../$Config/psql", - "--dbname=regress1,connectdb", - "--create-role=connectuser,connectdb", - "--schedule=${schedule}_schedule", - "--encoding=SQL_ASCII", - "--no-locale", - "--temp-install=./tmp_chk", - "--top-builddir=\"$topdir\"" - ); - push(@args,$maxconn) if $maxconn; - system(@args); - $status = $? >>8; - exit $status if $status; + chdir $startdir; + system("msbuild ecpg_regression.proj /p:config=$Config"); + my $status = $? >>8; + exit $status if $status; + chdir "$topdir/src/interfaces/ecpg/test"; + $schedule="ecpg"; + my @args = ( + "../../../../$Config/pg_regress_ecpg/pg_regress_ecpg", + "--psqldir=../../../$Config/psql", + "--dbname=regress1,connectdb", + "--create-role=connectuser,connectdb", + "--schedule=${schedule}_schedule", + "--encoding=SQL_ASCII", + "--no-locale", + "--temp-install=./tmp_chk", + "--top-builddir=\"$topdir\"" + ); + push(@args,$maxconn) if $maxconn; + system(@args); + $status = $? >>8; + exit $status if $status; } sub isolationcheck { - chdir "../isolation"; - copy("../../../$Config/isolationtester/isolationtester.exe","."); - my @args = ( - "../../../$Config/pg_isolation_regress/pg_isolation_regress", - "--psqldir=../../../$Config/psql", - "--inputdir=.","--schedule=./isolation_schedule" - ); - push(@args,$maxconn) if $maxconn; - system(@args); - my $status = $? >>8; - exit $status if $status; + chdir "../isolation"; + copy("../../../$Config/isolationtester/isolationtester.exe","."); + my @args = ( + "../../../$Config/pg_isolation_regress/pg_isolation_regress", + "--psqldir=../../../$Config/psql", + "--inputdir=.","--schedule=./isolation_schedule" + ); + push(@args,$maxconn) if $maxconn; + system(@args); + my $status = $? >>8; + exit $status if $status; } sub plcheck { - chdir "../../pl"; - - foreach my $pl (glob("*")) - { - next unless -d "$pl/sql" && -d "$pl/expected"; - my $lang = $pl eq 'tcl' ? 'pltcl' : $pl; - next unless -d "../../$Config/$lang"; - $lang = 'plpythonu' if $lang eq 'plpython'; - my @lang_args = ("--load-extension=$lang"); - chdir $pl; - my @tests = fetchTests(); - if ($lang eq 'plperl') - { - - # run both trusted and untrusted perl tests - push(@lang_args, "--load-extension=plperlu"); - - # assume we're using this perl to built postgres - # test if we can run two interpreters in one backend, and if so - # run the trusted/untrusted interaction tests - use Config; - if ($Config{usemultiplicity} eq 'define') - { - push(@tests,'plperl_plperlu'); - } - } - print "============================================================\n"; - print "Checking $lang\n"; - my @args = ( - "../../../$Config/pg_regress/pg_regress", - "--psqldir=../../../$Config/psql", - "--dbname=pl_regression",@lang_args,@tests - ); - system(@args); - my $status = $? >> 8; - exit $status if $status; - chdir ".."; - } - - chdir "../../.."; + chdir "../../pl"; + + foreach my $pl (glob("*")) + { + next unless -d "$pl/sql" && -d "$pl/expected"; + my $lang = $pl eq 'tcl' ? 'pltcl' : $pl; + next unless -d "../../$Config/$lang"; + $lang = 'plpythonu' if $lang eq 'plpython'; + my @lang_args = ("--load-extension=$lang"); + chdir $pl; + my @tests = fetchTests(); + if ($lang eq 'plperl') + { + + # run both trusted and untrusted perl tests + push(@lang_args, "--load-extension=plperlu"); + + # assume we're using this perl to built postgres + # test if we can run two interpreters in one backend, and if so + # run the trusted/untrusted interaction tests + use Config; + if ($Config{usemultiplicity} eq 'define') + { + push(@tests,'plperl_plperlu'); + } + } + print "============================================================\n"; + print "Checking $lang\n"; + my @args = ( + "../../../$Config/pg_regress/pg_regress", + "--psqldir=../../../$Config/psql", + "--dbname=pl_regression",@lang_args,@tests + ); + system(@args); + my $status = $? >> 8; + exit $status if $status; + chdir ".."; + } + + chdir "../../.."; } sub contribcheck { - chdir "../../../contrib"; - my $mstat = 0; - foreach my $module (glob("*")) - { - next if ($module eq 'sepgsql'); - next if ($module eq 'xml2' && !$config->{xml}); - next - unless -d "$module/sql" - &&-d "$module/expected" - &&(-f "$module/GNUmakefile" || -f "$module/Makefile"); - chdir $module; - print "============================================================\n"; - print "Checking $module\n"; - my @tests = fetchTests(); - my @opts = fetchRegressOpts(); - my @args = ( - "../../$Config/pg_regress/pg_regress", - "--psqldir=../../$Config/psql", - "--dbname=contrib_regression",@opts,@tests - ); - system(@args); - my $status = $? >> 8; - $mstat ||= $status; - chdir ".."; - } - exit $mstat if $mstat; + chdir "../../../contrib"; + my $mstat = 0; + foreach my $module (glob("*")) + { + next if ($module eq 'sepgsql'); + next if ($module eq 'xml2' && !$config->{xml}); + next + unless -d "$module/sql" + &&-d "$module/expected" + &&(-f "$module/GNUmakefile" || -f "$module/Makefile"); + chdir $module; + print "============================================================\n"; + print "Checking $module\n"; + my @tests = fetchTests(); + my @opts = fetchRegressOpts(); + my @args = ( + "../../$Config/pg_regress/pg_regress", + "--psqldir=../../$Config/psql", + "--dbname=contrib_regression",@opts,@tests + ); + system(@args); + my $status = $? >> 8; + $mstat ||= $status; + chdir ".."; + } + exit $mstat if $mstat; } sub fetchRegressOpts { - my $handle; - open($handle,"<GNUmakefile") - || open($handle,"<Makefile") - || die "Could not open Makefile"; - local($/) = undef; - my $m = <$handle>; - close($handle); - my @opts; - if ($m =~ /^\s*REGRESS_OPTS\s*=(.*)/m) - { - - # ignore options that use makefile variables - can't handle those - # ignore anything that isn't an option staring with -- - @opts = grep { $_ !~ /\$\(/ && $_ =~ /^--/ } split(/\s+/,$1); - } - if ($m =~ /^\s*ENCODING\s*=\s*(\S+)/m) - { - push @opts, "--encoding=$1"; - } - if ($m =~ /^\s*NO_LOCALE\s*=\s*\S+/m) - { - push @opts, "--no-locale"; - } - return @opts; + my $handle; + open($handle,"<GNUmakefile") + || open($handle,"<Makefile") + || die "Could not open Makefile"; + local($/) = undef; + my $m = <$handle>; + close($handle); + my @opts; + if ($m =~ /^\s*REGRESS_OPTS\s*=(.*)/m) + { + + # ignore options that use makefile variables - can't handle those + # ignore anything that isn't an option staring with -- + @opts = grep { $_ !~ /\$\(/ && $_ =~ /^--/ } split(/\s+/,$1); + } + if ($m =~ /^\s*ENCODING\s*=\s*(\S+)/m) + { + push @opts, "--encoding=$1"; + } + if ($m =~ /^\s*NO_LOCALE\s*=\s*\S+/m) + { + push @opts, "--no-locale"; + } + return @opts; } sub fetchTests { - my $handle; - open($handle,"<GNUmakefile") - || open($handle,"<Makefile") - || die "Could not open Makefile"; - local($/) = undef; - my $m = <$handle>; - close($handle); - my $t = ""; - - $m =~ s/\\[\r\n]*//gs; - if ($m =~ /^REGRESS\s*=\s*(.*)$/gm) - { - $t = $1; - $t =~ s/\s+/ /g; - - if ($m =~ /contrib\/pgcrypto/) - { - - # pgcrypto is special since the tests depend on the - # configuration of the build - - my $cftests = - $config->{openssl} - ?GetTests("OSSL_TESTS",$m) - : GetTests("INT_TESTS",$m); - my $pgptests = - $config->{zlib} - ?GetTests("ZLIB_TST",$m) - : GetTests("ZLIB_OFF_TST",$m); - $t =~ s/\$\(CF_TESTS\)/$cftests/; - $t =~ s/\$\(CF_PGP_TESTS\)/$pgptests/; - } - } - - return split(/\s+/,$t); + my $handle; + open($handle,"<GNUmakefile") + || open($handle,"<Makefile") + || die "Could not open Makefile"; + local($/) = undef; + my $m = <$handle>; + close($handle); + my $t = ""; + + $m =~ s/\\[\r\n]*//gs; + if ($m =~ /^REGRESS\s*=\s*(.*)$/gm) + { + $t = $1; + $t =~ s/\s+/ /g; + + if ($m =~ /contrib\/pgcrypto/) + { + + # pgcrypto is special since the tests depend on the + # configuration of the build + + my $cftests = + $config->{openssl} + ?GetTests("OSSL_TESTS",$m) + : GetTests("INT_TESTS",$m); + my $pgptests = + $config->{zlib} + ?GetTests("ZLIB_TST",$m) + : GetTests("ZLIB_OFF_TST",$m); + $t =~ s/\$\(CF_TESTS\)/$cftests/; + $t =~ s/\$\(CF_PGP_TESTS\)/$pgptests/; + } + } + + return split(/\s+/,$t); } sub GetTests { - my $testname = shift; - my $m = shift; - if ($m =~ /^$testname\s*=\s*(.*)$/gm) - { - return $1; - } - return ""; + my $testname = shift; + my $m = shift; + if ($m =~ /^$testname\s*=\s*(.*)$/gm) + { + return $1; + } + return ""; } sub usage { - print STDERR - "Usage: vcregress.pl ", - "<check|installcheck|plcheck|contribcheck|ecpgcheck> [schedule]\n"; - exit(1); + print STDERR + "Usage: vcregress.pl ", + "<check|installcheck|plcheck|contribcheck|ecpgcheck> [schedule]\n"; + exit(1); }