diff options
author | Bruce Momjian <bruce@momjian.us> | 2001-10-25 05:50:21 +0000 |
---|---|---|
committer | Bruce Momjian <bruce@momjian.us> | 2001-10-25 05:50:21 +0000 |
commit | b81844b1738c584d92330a5ccd0fbd8b603d2886 (patch) | |
tree | 4fae0d4cd26048177fc5cd1a2dd91abc99ba0f99 /src/backend/utils/cache | |
parent | 59da2105d8e6d95345b3b942a2e2aba8cead4838 (diff) | |
download | postgresql-b81844b1738c584d92330a5ccd0fbd8b603d2886.tar.gz postgresql-b81844b1738c584d92330a5ccd0fbd8b603d2886.zip |
pgindent run on all C files. Java run to follow. initdb/regression
tests pass.
Diffstat (limited to 'src/backend/utils/cache')
-rw-r--r-- | src/backend/utils/cache/catcache.c | 30 | ||||
-rw-r--r-- | src/backend/utils/cache/inval.c | 48 | ||||
-rw-r--r-- | src/backend/utils/cache/lsyscache.c | 49 | ||||
-rw-r--r-- | src/backend/utils/cache/relcache.c | 57 | ||||
-rw-r--r-- | src/backend/utils/cache/syscache.c | 8 |
5 files changed, 95 insertions, 97 deletions
diff --git a/src/backend/utils/cache/catcache.c b/src/backend/utils/cache/catcache.c index 77112a694e7..fe347fe5fbd 100644 --- a/src/backend/utils/cache/catcache.c +++ b/src/backend/utils/cache/catcache.c @@ -8,7 +8,7 @@ * * * IDENTIFICATION - * $Header: /cvsroot/pgsql/src/backend/utils/cache/catcache.c,v 1.83 2001/10/06 23:21:44 tgl Exp $ + * $Header: /cvsroot/pgsql/src/backend/utils/cache/catcache.c,v 1.84 2001/10/25 05:49:46 momjian Exp $ * *------------------------------------------------------------------------- */ @@ -30,13 +30,13 @@ #include "utils/syscache.h" -/* #define CACHEDEBUG */ /* turns DEBUG elogs on */ + /* #define CACHEDEBUG *//* turns DEBUG elogs on */ /* * Constants related to size of the catcache. * * NCCBUCKETS should be prime and must be less than 64K (because - * SharedInvalCatcacheMsg crams hash indexes into a uint16 field). In + * SharedInvalCatcacheMsg crams hash indexes into a uint16 field). In * practice it should be a lot less, anyway, to avoid chewing up too much * space on hash bucket headers. * @@ -642,13 +642,13 @@ CatalogCacheFlushRelation(Oid relId) tupRelid = ct->tuple.t_data->t_oid; else { - bool isNull; + bool isNull; tupRelid = DatumGetObjectId( - fastgetattr(&ct->tuple, - cache->cc_reloidattr, - cache->cc_tupdesc, - &isNull)); + fastgetattr(&ct->tuple, + cache->cc_reloidattr, + cache->cc_tupdesc, + &isNull)); Assert(!isNull); } @@ -707,8 +707,8 @@ InitCatCache(int id, oldcxt = MemoryContextSwitchTo(CacheMemoryContext); /* - * if first time through, initialize the cache group header, - * including global LRU list header + * if first time through, initialize the cache group header, including + * global LRU list header */ if (CacheHdr == NULL) { @@ -740,7 +740,7 @@ InitCatCache(int id, cp->cc_relname = relname; cp->cc_indname = indname; cp->cc_reloidattr = reloidattr; - cp->cc_relisshared = false; /* temporary */ + cp->cc_relisshared = false; /* temporary */ cp->cc_tupdesc = (TupleDesc) NULL; cp->cc_ntup = 0; cp->cc_size = NCCBUCKETS; @@ -749,8 +749,8 @@ InitCatCache(int id, cp->cc_key[i] = key[i]; /* - * new cache is initialized as far as we can go for now. - * print some debugging information, if appropriate. + * new cache is initialized as far as we can go for now. print some + * debugging information, if appropriate. */ InitCatCache_DEBUG1; @@ -1105,9 +1105,7 @@ ReleaseCatCache(HeapTuple tuple) && ct->dead #endif ) - { CatCacheRemoveCTup(ct->my_cache, ct); - } } /* @@ -1141,7 +1139,7 @@ ReleaseCatCache(HeapTuple tuple) void PrepareToInvalidateCacheTuple(Relation relation, HeapTuple tuple, - void (*function) (int, Index, ItemPointer, Oid)) + void (*function) (int, Index, ItemPointer, Oid)) { CatCache *ccp; diff --git a/src/backend/utils/cache/inval.c b/src/backend/utils/cache/inval.c index cd01bc9adc2..9ebf5f3bf2f 100644 --- a/src/backend/utils/cache/inval.c +++ b/src/backend/utils/cache/inval.c @@ -22,14 +22,14 @@ * second lives till end of transaction. Finally, we need a third list of * all tuples outdated in the current transaction; if we commit, we send * those invalidation events to all other backends (via the SI message queue) - * so that they can flush obsolete entries from their caches. This list + * so that they can flush obsolete entries from their caches. This list * definitely can't be processed until after we commit, otherwise the other * backends won't see our updated tuples as good. * * We do not need to register EVERY tuple operation in this way, just those - * on tuples in relations that have associated catcaches. We do, however, + * on tuples in relations that have associated catcaches. We do, however, * have to register every operation on every tuple that *could* be in a - * catcache, whether or not it currently is in our cache. Also, if the + * catcache, whether or not it currently is in our cache. Also, if the * tuple is in a relation that has multiple catcaches, we need to register * an invalidation message for each such catcache. catcache.c's * PrepareToInvalidateCacheTuple() routine provides the knowledge of which @@ -56,7 +56,7 @@ * Portions Copyright (c) 1994, Regents of the University of California * * IDENTIFICATION - * $Header: /cvsroot/pgsql/src/backend/utils/cache/inval.c,v 1.45 2001/06/19 19:42:16 tgl Exp $ + * $Header: /cvsroot/pgsql/src/backend/utils/cache/inval.c,v 1.46 2001/10/25 05:49:46 momjian Exp $ * *------------------------------------------------------------------------- */ @@ -74,15 +74,15 @@ /* * To minimize palloc traffic, we keep pending requests in successively- * larger chunks (a slightly more sophisticated version of an expansible - * array). All request types can be stored as SharedInvalidationMessage + * array). All request types can be stored as SharedInvalidationMessage * records. */ typedef struct InvalidationChunk { - struct InvalidationChunk *next; /* list link */ + struct InvalidationChunk *next; /* list link */ int nitems; /* # items currently stored in chunk */ int maxitems; /* size of allocated array in this chunk */ - SharedInvalidationMessage msgs[1]; /* VARIABLE LENGTH ARRAY */ + SharedInvalidationMessage msgs[1]; /* VARIABLE LENGTH ARRAY */ } InvalidationChunk; /* VARIABLE LENGTH STRUCTURE */ typedef struct InvalidationListHeader @@ -148,7 +148,7 @@ AddInvalidationMessage(InvalidationChunk **listHdr, chunk = (InvalidationChunk *) MemoryContextAlloc(TopTransactionContext, sizeof(InvalidationChunk) + - (FIRSTCHUNKSIZE-1) * sizeof(SharedInvalidationMessage)); + (FIRSTCHUNKSIZE - 1) *sizeof(SharedInvalidationMessage)); chunk->nitems = 0; chunk->maxitems = FIRSTCHUNKSIZE; chunk->next = *listHdr; @@ -157,12 +157,12 @@ AddInvalidationMessage(InvalidationChunk **listHdr, else if (chunk->nitems >= chunk->maxitems) { /* Need another chunk; double size of last chunk */ - int chunksize = 2 * chunk->maxitems; + int chunksize = 2 * chunk->maxitems; chunk = (InvalidationChunk *) MemoryContextAlloc(TopTransactionContext, sizeof(InvalidationChunk) + - (chunksize-1) * sizeof(SharedInvalidationMessage)); + (chunksize - 1) *sizeof(SharedInvalidationMessage)); chunk->nitems = 0; chunk->maxitems = chunksize; chunk->next = *listHdr; @@ -279,7 +279,10 @@ DiscardInvalidationMessages(InvalidationListHeader *hdr, bool physicalFree) } else { - /* Assume the storage will go away at xact end, just reset pointers */ + /* + * Assume the storage will go away at xact end, just reset + * pointers + */ hdr->cclist = NULL; hdr->rclist = NULL; } @@ -421,7 +424,7 @@ InvalidateSystemCaches(void) static void PrepareForTupleInvalidation(Relation relation, HeapTuple tuple, void (*CacheIdRegisterFunc) (int, Index, - ItemPointer, Oid), + ItemPointer, Oid), void (*RelationIdRegisterFunc) (Oid, Oid)) { Oid tupleRelId; @@ -460,12 +463,12 @@ PrepareForTupleInvalidation(Relation relation, HeapTuple tuple, * Yes. We need to register a relcache invalidation event for the * relation identified by relationId. * - * KLUGE ALERT: we always send the relcache event with MyDatabaseId, - * even if the rel in question is shared. This essentially means that - * only backends in this same database will react to the relcache flush - * request. This is in fact appropriate, since only those backends could - * see our pg_class or pg_attribute change anyway. It looks a bit ugly - * though. + * KLUGE ALERT: we always send the relcache event with MyDatabaseId, even + * if the rel in question is shared. This essentially means that only + * backends in this same database will react to the relcache flush + * request. This is in fact appropriate, since only those backends + * could see our pg_class or pg_attribute change anyway. It looks a + * bit ugly though. */ (*RelationIdRegisterFunc) (MyDatabaseId, relationId); } @@ -498,7 +501,7 @@ AcceptInvalidationMessages(void) * If isCommit, we must send out the messages in our GlobalInvalidMsgs list * to the shared invalidation message queue. Note that these will be read * not only by other backends, but also by our own backend at the next - * transaction start (via AcceptInvalidationMessages). Therefore, it's okay + * transaction start (via AcceptInvalidationMessages). Therefore, it's okay * to discard any pending LocalInvalidMsgs, since these will be redundant * with the global list. * @@ -538,7 +541,7 @@ AtEOXactInvalidationMessages(bool isCommit) * in a transaction. * * Here, we send no messages to the shared queue, since we don't know yet if - * we will commit. But we do need to locally process the LocalInvalidMsgs + * we will commit. But we do need to locally process the LocalInvalidMsgs * list, so as to flush our caches of any tuples we have outdated in the * current command. * @@ -563,9 +566,10 @@ CommandEndInvalidationMessages(bool isCommit) ProcessInvalidationMessages(&RollbackMsgs, LocalExecuteInvalidationMessage); } + /* - * LocalInvalidMsgs list is not interesting anymore, so flush it - * (for real). Do *not* clear GlobalInvalidMsgs or RollbackMsgs. + * LocalInvalidMsgs list is not interesting anymore, so flush it (for + * real). Do *not* clear GlobalInvalidMsgs or RollbackMsgs. */ DiscardInvalidationMessages(&LocalInvalidMsgs, true); } diff --git a/src/backend/utils/cache/lsyscache.c b/src/backend/utils/cache/lsyscache.c index b788383ba50..b359651b9ba 100644 --- a/src/backend/utils/cache/lsyscache.c +++ b/src/backend/utils/cache/lsyscache.c @@ -7,7 +7,7 @@ * Portions Copyright (c) 1994, Regents of the University of California * * IDENTIFICATION - * $Header: /cvsroot/pgsql/src/backend/utils/cache/lsyscache.c,v 1.58 2001/09/06 02:07:42 tgl Exp $ + * $Header: /cvsroot/pgsql/src/backend/utils/cache/lsyscache.c,v 1.59 2001/10/25 05:49:46 momjian Exp $ * * NOTES * Eventually, the index information should go through here, too. @@ -589,7 +589,6 @@ get_relnatts(Oid relid) else return InvalidAttrNumber; } - #endif /* @@ -723,7 +722,6 @@ get_typalign(Oid typid) else return 'i'; } - #endif char @@ -778,8 +776,8 @@ get_typdefault(Oid typid, Datum *defaultValue) typelem = type->typelem; /* - * typdefault is potentially null, so don't try to access it as a struct - * field. Must do it the hard way with SysCacheGetAttr. + * typdefault is potentially null, so don't try to access it as a + * struct field. Must do it the hard way with SysCacheGetAttr. */ textDefaultVal = SysCacheGetAttr(TYPEOID, typeTuple, @@ -828,6 +826,7 @@ get_typavgwidth(Oid typid, int32 typmod) */ if (typlen > 0) return typlen; + /* * type_maximum_size knows the encoding of typmod for some datatypes; * don't duplicate that knowledge here. @@ -836,16 +835,17 @@ get_typavgwidth(Oid typid, int32 typmod) if (maxwidth > 0) { /* - * For BPCHAR, the max width is also the only width. Otherwise - * we need to guess about the typical data width given the max. - * A sliding scale for percentage of max width seems reasonable. + * For BPCHAR, the max width is also the only width. Otherwise we + * need to guess about the typical data width given the max. A + * sliding scale for percentage of max width seems reasonable. */ if (typid == BPCHAROID) return maxwidth; if (maxwidth <= 32) return maxwidth; /* assume full width */ if (maxwidth < 1000) - return 32 + (maxwidth - 32) / 2; /* assume 50% */ + return 32 + (maxwidth - 32) / 2; /* assume 50% */ + /* * Beyond 1000, assume we're looking at something like * "varchar(10000)" where the limit isn't actually reached often, @@ -853,6 +853,7 @@ get_typavgwidth(Oid typid, int32 typmod) */ return 32 + (1000 - 32) / 2; } + /* * Ooops, we have no idea ... wild guess time. */ @@ -887,7 +888,6 @@ get_typtype(Oid typid) else return '\0'; } - #endif /* ---------- STATISTICS CACHE ---------- */ @@ -909,7 +909,7 @@ get_attavgwidth(Oid relid, AttrNumber attnum) 0, 0); if (HeapTupleIsValid(tp)) { - int32 stawidth = ((Form_pg_statistic) GETSTRUCT(tp))->stawidth; + int32 stawidth = ((Form_pg_statistic) GETSTRUCT(tp))->stawidth; ReleaseSysCache(tp); if (stawidth > 0) @@ -977,14 +977,17 @@ get_attstatsslot(HeapTuple statstuple, if (isnull) elog(ERROR, "get_attstatsslot: stavalues is null"); statarray = DatumGetArrayTypeP(val); + /* - * Do initial examination of the array. This produces a list - * of text Datums --- ie, pointers into the text array value. + * Do initial examination of the array. This produces a list of + * text Datums --- ie, pointers into the text array value. */ deconstruct_array(statarray, false, -1, 'i', values, nvalues); narrayelem = *nvalues; + /* - * We now need to replace each text Datum by its internal equivalent. + * We now need to replace each text Datum by its internal + * equivalent. * * Get the type input proc and typelem for the column datatype. */ @@ -997,9 +1000,10 @@ get_attstatsslot(HeapTuple statstuple, fmgr_info(((Form_pg_type) GETSTRUCT(typeTuple))->typinput, &inputproc); typelem = ((Form_pg_type) GETSTRUCT(typeTuple))->typelem; ReleaseSysCache(typeTuple); + /* - * Do the conversions. The palloc'd array of Datums is reused - * in place. + * Do the conversions. The palloc'd array of Datums is reused in + * place. */ for (j = 0; j < narrayelem; j++) { @@ -1013,6 +1017,7 @@ get_attstatsslot(HeapTuple statstuple, Int32GetDatum(atttypmod)); pfree(strval); } + /* * Free statarray if it's a detoasted copy. */ @@ -1028,10 +1033,11 @@ get_attstatsslot(HeapTuple statstuple, if (isnull) elog(ERROR, "get_attstatsslot: stanumbers is null"); statarray = DatumGetArrayTypeP(val); + /* - * We expect the array to be a 1-D float4 array; verify that. - * We don't need to use deconstruct_array() since the array - * data is just going to look like a C array of float4 values. + * We expect the array to be a 1-D float4 array; verify that. We + * don't need to use deconstruct_array() since the array data is + * just going to look like a C array of float4 values. */ narrayelem = ARR_DIMS(statarray)[0]; if (ARR_NDIM(statarray) != 1 || narrayelem <= 0 || @@ -1040,6 +1046,7 @@ get_attstatsslot(HeapTuple statstuple, *numbers = (float4 *) palloc(narrayelem * sizeof(float4)); memcpy(*numbers, ARR_DATA_PTR(statarray), narrayelem * sizeof(float4)); *nnumbers = narrayelem; + /* * Free statarray if it's a detoasted copy. */ @@ -1057,9 +1064,9 @@ free_attstatsslot(Oid atttype, { if (values) { - if (! get_typbyval(atttype)) + if (!get_typbyval(atttype)) { - int i; + int i; for (i = 0; i < nvalues; i++) pfree(DatumGetPointer(values[i])); diff --git a/src/backend/utils/cache/relcache.c b/src/backend/utils/cache/relcache.c index c56d606076a..1073841f9d1 100644 --- a/src/backend/utils/cache/relcache.c +++ b/src/backend/utils/cache/relcache.c @@ -8,7 +8,7 @@ * * * IDENTIFICATION - * $Header: /cvsroot/pgsql/src/backend/utils/cache/relcache.c,v 1.146 2001/10/06 23:21:44 tgl Exp $ + * $Header: /cvsroot/pgsql/src/backend/utils/cache/relcache.c,v 1.147 2001/10/25 05:49:46 momjian Exp $ * *------------------------------------------------------------------------- */ @@ -226,7 +226,6 @@ static void RelationClearRelation(Relation relation, bool rebuildIt); #ifdef ENABLE_REINDEX_NAILED_RELATIONS static void RelationReloadClassinfo(Relation relation); - #endif /* ENABLE_REINDEX_NAILED_RELATIONS */ static void RelationFlushRelation(Relation relation); static Relation RelationNameCacheGetRelation(const char *relationName); @@ -273,7 +272,6 @@ static List *insert_ordered_oid(List *list, Oid datum); static HeapTuple ScanPgRelation(RelationBuildDescInfo buildinfo) { - /* * If this is bootstrap time (initdb), then we can't use the system * catalog indices, because they may not exist yet. Otherwise, we @@ -333,7 +331,6 @@ scan_pg_rel_seq(RelationBuildDescInfo buildinfo) return_tuple = pg_class_tuple; else { - /* * a satanic bug used to live here: pg_class_tuple used to be * returned here without having the corresponding buffer pinned. @@ -382,7 +379,7 @@ scan_pg_rel_ind(RelationBuildDescInfo buildinfo) default: elog(ERROR, "ScanPgRelation: bad buildinfo"); - return_tuple = NULL;/* keep compiler quiet */ + return_tuple = NULL; /* keep compiler quiet */ } heap_close(pg_class_desc, AccessShareLock); @@ -461,7 +458,6 @@ static void RelationBuildTupleDesc(RelationBuildDescInfo buildinfo, Relation relation) { - /* * If this is bootstrap time (initdb), then we can't use the system * catalog indices, because they may not exist yet. Otherwise, we @@ -649,7 +645,6 @@ build_tupdesc_ind(RelationBuildDescInfo buildinfo, { #ifdef _DROP_COLUMN_HACK__ bool columnDropped = false; - #endif /* _DROP_COLUMN_HACK__ */ atttup = AttributeRelidNumIndexScan(attrel, @@ -767,7 +762,7 @@ RelationBuildRuleLock(Relation relation) */ rulescxt = AllocSetContextCreate(CacheMemoryContext, RelationGetRelationName(relation), - 0, /* minsize */ + 0, /* minsize */ 1024, /* initsize */ 1024); /* maxsize */ relation->rd_rulescxt = rulescxt; @@ -1106,16 +1101,16 @@ RelationInitIndexAccessInfo(Relation relation) amsupport = relation->rd_am->amsupport; /* - * Make the private context to hold index access info. The reason - * we need a context, and not just a couple of pallocs, is so that - * we won't leak any subsidiary info attached to fmgr lookup records. + * Make the private context to hold index access info. The reason we + * need a context, and not just a couple of pallocs, is so that we + * won't leak any subsidiary info attached to fmgr lookup records. * * Context parameters are set on the assumption that it'll probably not * contain much data. */ indexcxt = AllocSetContextCreate(CacheMemoryContext, RelationGetRelationName(relation), - 0, /* minsize */ + 0, /* minsize */ 512, /* initsize */ 1024); /* maxsize */ relation->rd_indexcxt = indexcxt; @@ -1128,7 +1123,7 @@ RelationInitIndexAccessInfo(Relation relation) if (amsupport > 0) { - int nsupport = natts * amsupport; + int nsupport = natts * amsupport; support = (RegProcedure *) MemoryContextAlloc(indexcxt, nsupport * sizeof(RegProcedure)); @@ -1214,8 +1209,9 @@ formrdesc(char *relationName, strcpy(RelationGetPhysicalRelationName(relation), relationName); /* - * It's important to distinguish between shared and non-shared relations, - * even at bootstrap time, to make sure we know where they are stored. + * It's important to distinguish between shared and non-shared + * relations, even at bootstrap time, to make sure we know where they + * are stored. */ relation->rd_rel->relisshared = IsSharedSystemRelationName(relationName); @@ -1267,8 +1263,8 @@ formrdesc(char *relationName, if (!IsBootstrapProcessingMode()) { /* - * This list is incomplete, but it only has to work for the - * set of rels that formrdesc is used for ... + * This list is incomplete, but it only has to work for the set of + * rels that formrdesc is used for ... */ if (strcmp(relationName, RelationRelationName) == 0 || strcmp(relationName, AttributeRelationName) == 0 || @@ -1560,7 +1556,6 @@ RelationReloadClassinfo(Relation relation) return; } - #endif /* ENABLE_REINDEX_NAILED_RELATIONS */ /* @@ -1649,7 +1644,6 @@ RelationClearRelation(Relation relation, bool rebuildIt) } else { - /* * When rebuilding an open relcache entry, must preserve ref count * and myxactonly flag. Also attempt to preserve the tupledesc, @@ -1663,7 +1657,7 @@ RelationClearRelation(Relation relation, bool rebuildIt) RuleLock *old_rules = relation->rd_rules; MemoryContext old_rulescxt = relation->rd_rulescxt; TriggerDesc *old_trigdesc = relation->trigdesc; - BlockNumber old_nblocks = relation->rd_nblocks; + BlockNumber old_nblocks = relation->rd_nblocks; RelationBuildDescInfo buildinfo; buildinfo.infotype = INFO_RELID; @@ -1730,7 +1724,6 @@ RelationFlushRelation(Relation relation) if (relation->rd_myxactonly) { - /* * Local rels should always be rebuilt, not flushed; the relcache * entry must live until RelationPurgeLocalRelation(). @@ -1739,7 +1732,6 @@ RelationFlushRelation(Relation relation) } else { - /* * Nonlocal rels can be dropped from the relcache if not open. */ @@ -1837,7 +1829,6 @@ RelationFlushIndexes(Relation *r, relation->rd_rel->relam == accessMethodId)) RelationFlushRelation(relation); } - #endif @@ -1982,8 +1973,8 @@ RelationBuildLocalRelation(const char *relname, rel->rd_isnailed = true; /* - * create a new tuple descriptor from the one passed in - * (we do this to copy it into the cache context) + * create a new tuple descriptor from the one passed in (we do this to + * copy it into the cache context) */ rel->rd_att = CreateTupleDescCopyConstr(tupDesc); @@ -2003,9 +1994,10 @@ RelationBuildLocalRelation(const char *relname, rel->rd_rel->relchecks = tupDesc->constr->num_check; /* - * Insert relation OID and database/tablespace ID into the right places. - * XXX currently we assume physical tblspace/relnode are same as logical - * dbid/reloid. Probably should pass an extra pair of parameters. + * Insert relation OID and database/tablespace ID into the right + * places. XXX currently we assume physical tblspace/relnode are same + * as logical dbid/reloid. Probably should pass an extra pair of + * parameters. */ rel->rd_rel->relisshared = (dbid == InvalidOid); @@ -2149,7 +2141,6 @@ RelationCacheInitialize(void) void RelationCacheInitializePhase2(void) { - /* * Get the real pg_class tuple for each nailed-in-cache relcache entry * that was made by RelationCacheInitialize(), and replace the phony @@ -2159,7 +2150,6 @@ RelationCacheInitializePhase2(void) */ if (!IsBootstrapProcessingMode()) { - /* * Initialize critical system index relation descriptors, first. * They are to make building relation descriptors fast. @@ -2764,9 +2754,9 @@ init_irels(void) */ indexcxt = AllocSetContextCreate(CacheMemoryContext, RelationGetRelationName(ird), - 0, /* minsize */ - 512, /* initsize */ - 1024); /* maxsize */ + 0, /* minsize */ + 512, /* initsize */ + 1024); /* maxsize */ ird->rd_indexcxt = indexcxt; /* next, read the index strategy map */ @@ -2848,7 +2838,6 @@ write_irels(void) fd = PathNameOpenFile(tempfilename, O_WRONLY | O_CREAT | O_TRUNC | PG_BINARY, 0600); if (fd < 0) { - /* * We used to consider this a fatal error, but we might as well * continue with backend startup ... diff --git a/src/backend/utils/cache/syscache.c b/src/backend/utils/cache/syscache.c index 6020597f2a3..e629bf48e46 100644 --- a/src/backend/utils/cache/syscache.c +++ b/src/backend/utils/cache/syscache.c @@ -8,7 +8,7 @@ * * * IDENTIFICATION - * $Header: /cvsroot/pgsql/src/backend/utils/cache/syscache.c,v 1.65 2001/08/21 16:36:05 tgl Exp $ + * $Header: /cvsroot/pgsql/src/backend/utils/cache/syscache.c,v 1.66 2001/10/25 05:49:46 momjian Exp $ * * NOTES * These routines allow the parser/planner/executor to perform @@ -133,7 +133,7 @@ static struct cachedesc cacheinfo[] = { 0, 0 }}, - {AccessMethodProcedureRelationName, /* AMPROCNUM */ + {AccessMethodProcedureRelationName, /* AMPROCNUM */ AccessMethodProcedureIndex, 0, 2, @@ -365,7 +365,8 @@ static struct cachedesc cacheinfo[] = { }} }; -static CatCache *SysCache[lengthof(cacheinfo)]; +static CatCache *SysCache[ + lengthof(cacheinfo)]; static int SysCacheSize = lengthof(cacheinfo); static bool CacheInitialized = false; @@ -564,7 +565,6 @@ SysCacheGetAttr(int cacheId, HeapTuple tup, AttrNumber attributeNumber, bool *isNull) { - /* * We just need to get the TupleDesc out of the cache entry, and then * we can apply heap_getattr(). We expect that the cache control data |