diff options
author | Tom Lane <tgl@sss.pgh.pa.us> | 2001-08-25 18:52:43 +0000 |
---|---|---|
committer | Tom Lane <tgl@sss.pgh.pa.us> | 2001-08-25 18:52:43 +0000 |
commit | 2589735da08c4e597accb6eab5ae65b6339ee630 (patch) | |
tree | 829f7073292c6b55f86580863837441991638405 /src/backend/access/transam/clog.c | |
parent | 4699d81dc99ef1687e9396b57b0ed10f42699792 (diff) | |
download | postgresql-2589735da08c4e597accb6eab5ae65b6339ee630.tar.gz postgresql-2589735da08c4e597accb6eab5ae65b6339ee630.zip |
Replace implementation of pg_log as a relation accessed through the
buffer manager with 'pg_clog', a specialized access method modeled
on pg_xlog. This simplifies startup (don't need to play games to
open pg_log; among other things, OverrideTransactionSystem goes away),
should improve performance a little, and opens the door to recycling
commit log space by removing no-longer-needed segments of the commit
log. Actual recycling is not there yet, but I felt I should commit
this part separately since it'd still be useful if we chose not to
do transaction ID wraparound.
Diffstat (limited to 'src/backend/access/transam/clog.c')
-rw-r--r-- | src/backend/access/transam/clog.c | 987 |
1 files changed, 987 insertions, 0 deletions
diff --git a/src/backend/access/transam/clog.c b/src/backend/access/transam/clog.c new file mode 100644 index 00000000000..2347edfd095 --- /dev/null +++ b/src/backend/access/transam/clog.c @@ -0,0 +1,987 @@ +/*------------------------------------------------------------------------- + * + * clog.c + * PostgreSQL transaction-commit-log manager + * + * This module replaces the old "pg_log" access code, which treated pg_log + * essentially like a relation, in that it went through the regular buffer + * manager. The problem with that was that there wasn't any good way to + * recycle storage space for transactions so old that they'll never be + * looked up again. Now we use specialized access code so that the commit + * log can be broken into relatively small, independent segments. + * + * Portions Copyright (c) 1996-2001, PostgreSQL Global Development Group + * Portions Copyright (c) 1994, Regents of the University of California + * + * $Header: /cvsroot/pgsql/src/backend/access/transam/clog.c,v 1.1 2001/08/25 18:52:41 tgl Exp $ + * + *------------------------------------------------------------------------- + */ +#include "postgres.h" + +#include <fcntl.h> +#include <dirent.h> +#include <errno.h> +#include <unistd.h> + +#include "access/clog.h" +#include "storage/s_lock.h" +#include "miscadmin.h" + + +/* + * Defines for CLOG page and segment sizes. A page is the same BLCKSZ + * as is used everywhere else in Postgres. The CLOG segment size can be + * chosen somewhat arbitrarily; we make it 1 million transactions by default, + * or 256Kb. + * + * Note: because TransactionIds are 32 bits and wrap around at 0xFFFFFFFF, + * CLOG page numbering also wraps around at 0xFFFFFFFF/CLOG_XACTS_PER_PAGE, + * and CLOG segment numbering at 0xFFFFFFFF/CLOG_XACTS_PER_SEGMENT. We need + * take no explicit notice of that fact in this module, except when comparing + * segment and page numbers in TruncateCLOG (see CLOGPagePrecedes). + */ + +#define CLOG_BLCKSZ BLCKSZ + +/* We need two bits per xact, so four xacts fit in a byte */ +#define CLOG_BITS_PER_XACT 2 +#define CLOG_XACTS_PER_BYTE 4 +#define CLOG_XACTS_PER_PAGE (CLOG_BLCKSZ * CLOG_XACTS_PER_BYTE) +#define CLOG_XACT_BITMASK ((1 << CLOG_BITS_PER_XACT) - 1) + +#define CLOG_XACTS_PER_SEGMENT 0x100000 +#define CLOG_PAGES_PER_SEGMENT (CLOG_XACTS_PER_SEGMENT / CLOG_XACTS_PER_PAGE) + +#define TransactionIdToPage(xid) ((xid) / (TransactionId) CLOG_XACTS_PER_PAGE) +#define TransactionIdToPgIndex(xid) ((xid) % (TransactionId) CLOG_XACTS_PER_PAGE) +#define TransactionIdToByte(xid) (TransactionIdToPgIndex(xid) / CLOG_XACTS_PER_BYTE) +#define TransactionIdToBIndex(xid) ((xid) % (TransactionId) CLOG_XACTS_PER_BYTE) + + +/*---------- + * Shared-memory data structures for CLOG control + * + * We use a simple least-recently-used scheme to manage a pool of page + * buffers for the CLOG. Under ordinary circumstances we expect that write + * traffic will occur mostly to the latest CLOG page (and to the just-prior + * page, soon after a page transition). Read traffic will probably touch + * a larger span of pages, but in any case a fairly small number of page + * buffers should be sufficient. So, we just search the buffers using plain + * linear search; there's no need for a hashtable or anything fancy. + * The management algorithm is straight LRU except that we will never swap + * out the latest page (since we know it's going to be hit again eventually). + * + * We use an overall spinlock to protect the shared data structures, plus + * per-buffer spinlocks that synchronize I/O for each buffer. A process + * that is reading in or writing out a page buffer does not hold the control + * lock, only the per-buffer lock for the buffer it is working on. + * + * To change the page number or state of a buffer, one must normally hold + * the control lock. (The sole exception to this rule is that a writer + * process changes the state from DIRTY to WRITE_IN_PROGRESS while holding + * only the per-buffer lock.) If the buffer's state is neither EMPTY nor + * CLEAN, then there may be processes doing (or waiting to do) I/O on the + * buffer, so the page number may not be changed, and the only allowed state + * transition is to change WRITE_IN_PROGRESS to DIRTY after dirtying the page. + * To do any other state transition involving a buffer with potential I/O + * processes, one must hold both the per-buffer lock and the control lock. + * (Note the control lock must be acquired second; do not wait on a buffer + * lock while holding the control lock.) A process wishing to read a page + * marks the buffer state as READ_IN_PROGRESS, then drops the control lock, + * acquires the per-buffer lock, and rechecks the state before proceeding. + * This recheck takes care of the possibility that someone else already did + * the read, while the early marking prevents someone else from trying to + * read the same page into a different buffer. + * + * Note we are assuming that read and write of the state value is atomic, + * since I/O processes may examine and change the state while not holding + * the control lock. + * + * As with the regular buffer manager, it is possible for another process + * to re-dirty a page that is currently being written out. This is handled + * by setting the page's state from WRITE_IN_PROGRESS to DIRTY. The writing + * process must notice this and not mark the page CLEAN when it's done. + * + * XXX it's probably okay to use a spinlock for the control lock, since + * that lock is only held for very short operations. It'd be nice to use + * some other form of lock for the per-buffer I/O locks, however. + * + * XLOG interactions: this module generates an XLOG record whenever a new + * CLOG page is initialized to zeroes. Other writes of CLOG come from + * recording of transaction commit or abort in xact.c, which generates its + * own XLOG records for these events and will re-perform the status update + * on redo; so we need make no additional XLOG entry here. Also, the XLOG + * is guaranteed flushed through the XLOG commit record before we are called + * to log a commit, so the WAL rule "write xlog before data" is satisfied + * automatically for commits, and we don't really care for aborts. Therefore, + * we don't need to mark XLOG pages with LSN information; we have enough + * synchronization already. + *---------- + */ +#define NUM_CLOG_BUFFERS 8 + +typedef enum +{ + CLOG_PAGE_EMPTY, /* CLOG buffer is not in use */ + CLOG_PAGE_READ_IN_PROGRESS, /* CLOG page is being read in */ + CLOG_PAGE_CLEAN, /* CLOG page is valid and not dirty */ + CLOG_PAGE_DIRTY, /* CLOG page is valid but needs write */ + CLOG_PAGE_WRITE_IN_PROGRESS /* CLOG page is being written out in */ +} ClogPageStatus; + +/* + * Shared-memory state for CLOG. + */ +typedef struct ClogCtlData +{ + /* + * Info for each buffer slot. Page number is undefined when status is + * EMPTY. lru_count is essentially the number of operations since last + * use of this page; the page with highest lru_count is the best candidate + * to replace. + */ + char *page_buffer[NUM_CLOG_BUFFERS]; + ClogPageStatus page_status[NUM_CLOG_BUFFERS]; + int page_number[NUM_CLOG_BUFFERS]; + unsigned int page_lru_count[NUM_CLOG_BUFFERS]; + /* + * latest_page_number is the page number of the current end of the + * CLOG; this is not critical data, since we use it only to avoid + * swapping out the latest page. + */ + int latest_page_number; + + slock_t control_lck; /* Lock for ClogCtlData itself */ + slock_t buffer_lck[NUM_CLOG_BUFFERS]; /* Per-buffer I/O locks */ +} ClogCtlData; + +static ClogCtlData *ClogCtl = NULL; + +/* + * ClogDir is set during CLOGShmemInit and does not change thereafter. + * The value is automatically inherited by backends via fork, and + * doesn't need to be in shared memory. + */ +static char ClogDir[MAXPGPATH]; + +#define ClogFileName(path, seg) \ + snprintf(path, MAXPGPATH, "%s/%04X", ClogDir, seg) + +/* + * Macro to mark a buffer slot "most recently used". + */ +#define ClogRecentlyUsed(slotno) \ + do { \ + int iilru; \ + for (iilru = 0; iilru < NUM_CLOG_BUFFERS; iilru++) \ + ClogCtl->page_lru_count[iilru]++; \ + ClogCtl->page_lru_count[slotno] = 0; \ + } while (0) + + +static int ZeroCLOGPage(int pageno, bool writeXlog); +static int ReadCLOGPage(int pageno); +static void WriteCLOGPage(int slotno); +static void CLOGPhysicalReadPage(int pageno, int slotno); +static void CLOGPhysicalWritePage(int pageno, int slotno); +static int SelectLRUCLOGPage(int pageno); +static bool ScanCLOGDirectory(int cutoffPage, bool doDeletions); +static bool CLOGPagePrecedes(int page1, int page2); +static void WriteZeroPageXlogRec(int pageno); + + +/* + * Record the final state of a transaction in the commit log. + * + * NB: this is a low-level routine and is NOT the preferred entry point + * for most uses; TransactionLogUpdate() in transam.c is the intended caller. + */ +void +TransactionIdSetStatus(TransactionId xid, XidStatus status) +{ + int pageno = TransactionIdToPage(xid); + int byteno = TransactionIdToByte(xid); + int bshift = TransactionIdToBIndex(xid) * CLOG_BITS_PER_XACT; + int slotno; + char *byteptr; + + Assert(status == TRANSACTION_STATUS_COMMITTED || + status == TRANSACTION_STATUS_ABORTED); + + S_LOCK(&(ClogCtl->control_lck)); + + slotno = ReadCLOGPage(pageno); + byteptr = ClogCtl->page_buffer[slotno] + byteno; + + /* Current state should be 0 or target state */ + Assert(((*byteptr >> bshift) & CLOG_XACT_BITMASK) == 0 || + ((*byteptr >> bshift) & CLOG_XACT_BITMASK) == status); + + *byteptr |= (status << bshift); + + ClogCtl->page_status[slotno] = CLOG_PAGE_DIRTY; + + S_UNLOCK(&(ClogCtl->control_lck)); +} + +/* + * Interrogate the state of a transaction in the commit log. + * + * NB: this is a low-level routine and is NOT the preferred entry point + * for most uses; TransactionLogTest() in transam.c is the intended caller. + */ +XidStatus +TransactionIdGetStatus(TransactionId xid) +{ + int pageno = TransactionIdToPage(xid); + int byteno = TransactionIdToByte(xid); + int bshift = TransactionIdToBIndex(xid) * CLOG_BITS_PER_XACT; + int slotno; + char *byteptr; + XidStatus status; + + S_LOCK(&(ClogCtl->control_lck)); + + slotno = ReadCLOGPage(pageno); + byteptr = ClogCtl->page_buffer[slotno] + byteno; + + status = (*byteptr >> bshift) & CLOG_XACT_BITMASK; + + S_UNLOCK(&(ClogCtl->control_lck)); + + return status; +} + + +/* + * Initialization of shared memory for CLOG + */ + +int +CLOGShmemSize(void) +{ + return MAXALIGN(sizeof(ClogCtlData) + CLOG_BLCKSZ * NUM_CLOG_BUFFERS); +} + +void +CLOGShmemInit(void) +{ + bool found; + char *bufptr; + int slotno; + + /* this must agree with space requested by CLOGShmemSize() */ + ClogCtl = (ClogCtlData *) + ShmemInitStruct("CLOG Ctl", + MAXALIGN(sizeof(ClogCtlData) + + CLOG_BLCKSZ * NUM_CLOG_BUFFERS), + &found); + Assert(!found); + + memset(ClogCtl, 0, sizeof(ClogCtlData)); + + S_INIT_LOCK(&(ClogCtl->control_lck)); + + bufptr = ((char *) ClogCtl) + sizeof(ClogCtlData); + + for (slotno = 0; slotno < NUM_CLOG_BUFFERS; slotno++) + { + ClogCtl->page_buffer[slotno] = bufptr; + ClogCtl->page_status[slotno] = CLOG_PAGE_EMPTY; + S_INIT_LOCK(&(ClogCtl->buffer_lck[slotno])); + bufptr += CLOG_BLCKSZ; + } + + /* ClogCtl->latest_page_number will be set later */ + + /* Init CLOG directory path */ + snprintf(ClogDir, MAXPGPATH, "%s/pg_clog", DataDir); +} + +/* + * This func must be called ONCE on system install. It creates + * the initial CLOG segment. (The CLOG directory is assumed to + * have been created by the initdb shell script, and CLOGShmemInit + * must have been called already.) + */ +void +BootStrapCLOG(void) +{ + int slotno; + + S_LOCK(&(ClogCtl->control_lck)); + + /* Create and zero the first page of the commit log */ + slotno = ZeroCLOGPage(0, false); + + /* Make sure it's written out */ + WriteCLOGPage(slotno); + Assert(ClogCtl->page_status[slotno] == CLOG_PAGE_CLEAN); + + S_UNLOCK(&(ClogCtl->control_lck)); +} + +/* + * Initialize (or reinitialize) a page of CLOG to zeroes. + * If writeXlog is TRUE, also emit an XLOG record saying we did this. + * + * The page is not actually written, just set up in shared memory. + * The slot number of the new page is returned. + * + * Control lock must be held at entry, and will be held at exit. + */ +static int +ZeroCLOGPage(int pageno, bool writeXlog) +{ + int slotno; + + /* Find a suitable buffer slot for the page */ + slotno = SelectLRUCLOGPage(pageno); + Assert(ClogCtl->page_status[slotno] == CLOG_PAGE_EMPTY || + ClogCtl->page_status[slotno] == CLOG_PAGE_CLEAN || + ClogCtl->page_number[slotno] == pageno); + + /* Mark the slot as containing this page */ + ClogCtl->page_number[slotno] = pageno; + ClogCtl->page_status[slotno] = CLOG_PAGE_DIRTY; + ClogRecentlyUsed(slotno); + + /* Set the buffer to zeroes */ + MemSet(ClogCtl->page_buffer[slotno], 0, CLOG_BLCKSZ); + + /* Assume this page is now the latest active page */ + ClogCtl->latest_page_number = pageno; + + if (writeXlog) + WriteZeroPageXlogRec(pageno); + + return slotno; +} + +/* + * Find a CLOG page in a shared buffer, reading it in if necessary. + * The page number must correspond to an already-initialized page. + * + * Return value is the shared-buffer slot number now holding the page. + * The buffer's LRU access info is updated. + * + * Control lock must be held at entry, and will be held at exit. + */ +static int +ReadCLOGPage(int pageno) +{ + /* Outer loop handles restart if we lose the buffer to someone else */ + for (;;) + { + int slotno; + + /* See if page already is in memory; if not, pick victim slot */ + slotno = SelectLRUCLOGPage(pageno); + + /* Did we find the page in memory? */ + if (ClogCtl->page_number[slotno] == pageno && + ClogCtl->page_status[slotno] != CLOG_PAGE_EMPTY) + { + /* If page is still being read in, we cannot use it yet */ + if (ClogCtl->page_status[slotno] != CLOG_PAGE_READ_IN_PROGRESS) + { + /* otherwise, it's ready to use */ + ClogRecentlyUsed(slotno); + return slotno; + } + } + else + { + /* We found no match; assert we selected a freeable slot */ + Assert(ClogCtl->page_status[slotno] == CLOG_PAGE_EMPTY || + ClogCtl->page_status[slotno] == CLOG_PAGE_CLEAN); + } + + /* Mark the slot read-busy (no-op if it already was) */ + ClogCtl->page_number[slotno] = pageno; + ClogCtl->page_status[slotno] = CLOG_PAGE_READ_IN_PROGRESS; + + /* + * Temporarily mark page as recently-used to discourage + * SelectLRUCLOGPage from selecting it again for someone else. + */ + ClogCtl->page_lru_count[slotno] = 0; + + /* Release shared lock, grab per-buffer lock instead */ + S_UNLOCK(&(ClogCtl->control_lck)); + S_LOCK(&(ClogCtl->buffer_lck[slotno])); + + /* + * Check to see if someone else already did the read, or took the + * buffer away from us. If so, restart from the top. + */ + if (ClogCtl->page_number[slotno] != pageno || + ClogCtl->page_status[slotno] != CLOG_PAGE_READ_IN_PROGRESS) + { + S_UNLOCK(&(ClogCtl->buffer_lck[slotno])); + S_LOCK(&(ClogCtl->control_lck)); + continue; + } + + /* Okay, do the read */ + CLOGPhysicalReadPage(pageno, slotno); + + /* Re-acquire shared control lock and update page state */ + S_LOCK(&(ClogCtl->control_lck)); + + Assert(ClogCtl->page_number[slotno] == pageno && + ClogCtl->page_status[slotno] == CLOG_PAGE_READ_IN_PROGRESS); + + ClogCtl->page_status[slotno] = CLOG_PAGE_CLEAN; + + S_UNLOCK(&(ClogCtl->buffer_lck[slotno])); + + ClogRecentlyUsed(slotno); + return slotno; + } +} + +/* + * Write a CLOG page from a shared buffer, if necessary. + * Does nothing if the specified slot is not dirty. + * + * NOTE: only one write attempt is made here. Hence, it is possible that + * the page is still dirty at exit (if someone else re-dirtied it during + * the write). However, we *do* attempt a fresh write even if the page + * is already being written; this is for checkpoints. + * + * Control lock must be held at entry, and will be held at exit. + */ +static void +WriteCLOGPage(int slotno) +{ + int pageno; + + /* Do nothing if page does not need writing */ + if (ClogCtl->page_status[slotno] != CLOG_PAGE_DIRTY && + ClogCtl->page_status[slotno] != CLOG_PAGE_WRITE_IN_PROGRESS) + return; + + pageno = ClogCtl->page_number[slotno]; + + /* Release shared lock, grab per-buffer lock instead */ + S_UNLOCK(&(ClogCtl->control_lck)); + S_LOCK(&(ClogCtl->buffer_lck[slotno])); + + /* + * Check to see if someone else already did the write, or took the + * buffer away from us. If so, do nothing. NOTE: we really should + * never see WRITE_IN_PROGRESS here, since that state should only + * occur while the writer is holding the buffer lock. But accept it + * so that we have a recovery path if a writer aborts. + */ + if (ClogCtl->page_number[slotno] != pageno || + (ClogCtl->page_status[slotno] != CLOG_PAGE_DIRTY && + ClogCtl->page_status[slotno] != CLOG_PAGE_WRITE_IN_PROGRESS)) + { + S_UNLOCK(&(ClogCtl->buffer_lck[slotno])); + S_LOCK(&(ClogCtl->control_lck)); + return; + } + + /* + * Mark the slot write-busy. After this point, a transaction status + * update on this page will mark it dirty again. NB: we are assuming + * that read/write of the page status field is atomic, since we change + * the state while not holding control lock. However, we cannot set + * this state any sooner, or we'd possibly fool a previous writer + * into thinking he's successfully dumped the page when he hasn't. + * (Scenario: other writer starts, page is redirtied, we come along and + * set WRITE_IN_PROGRESS again, other writer completes and sets CLEAN + * because redirty info has been lost, then we think it's clean too.) + */ + ClogCtl->page_status[slotno] = CLOG_PAGE_WRITE_IN_PROGRESS; + + /* Okay, do the write */ + CLOGPhysicalWritePage(pageno, slotno); + + /* Re-acquire shared control lock and update page state */ + S_LOCK(&(ClogCtl->control_lck)); + + Assert(ClogCtl->page_number[slotno] == pageno && + (ClogCtl->page_status[slotno] == CLOG_PAGE_WRITE_IN_PROGRESS || + ClogCtl->page_status[slotno] == CLOG_PAGE_DIRTY)); + + /* Cannot set CLEAN if someone re-dirtied page since write started */ + if (ClogCtl->page_status[slotno] == CLOG_PAGE_WRITE_IN_PROGRESS) + ClogCtl->page_status[slotno] = CLOG_PAGE_CLEAN; + + S_UNLOCK(&(ClogCtl->buffer_lck[slotno])); +} + +/* + * Physical read of a (previously existing) page into a buffer slot + * + * For now, assume it's not worth keeping a file pointer open across + * read/write operations. We could cache one virtual file pointer ... + */ +static void +CLOGPhysicalReadPage(int pageno, int slotno) +{ + int segno = pageno / CLOG_PAGES_PER_SEGMENT; + int rpageno = pageno % CLOG_PAGES_PER_SEGMENT; + int offset = rpageno * CLOG_BLCKSZ; + char path[MAXPGPATH]; + int fd; + + ClogFileName(path, segno); + + /* + * In a crash-and-restart situation, it's possible for us to receive + * commands to set the commit status of transactions whose bits are + * in already-truncated segments of the commit log (see notes in + * CLOGPhysicalWritePage). Hence, if we are InRecovery, allow the + * case where the file doesn't exist, and return zeroes instead. + */ + fd = BasicOpenFile(path, O_RDWR | PG_BINARY, S_IRUSR | S_IWUSR); + if (fd < 0) + { + if (errno != ENOENT || !InRecovery) + elog(STOP, "open of %s failed: %m", path); + elog(DEBUG, "clog file %s doesn't exist, reading as zeroes", path); + MemSet(ClogCtl->page_buffer[slotno], 0, CLOG_BLCKSZ); + return; + } + + if (lseek(fd, (off_t) offset, SEEK_SET) < 0) + elog(STOP, "lseek of clog file %u, offset %u failed: %m", + segno, offset); + + errno = 0; + if (read(fd, ClogCtl->page_buffer[slotno], CLOG_BLCKSZ) != CLOG_BLCKSZ) + elog(STOP, "read of clog file %u, offset %u failed: %m", + segno, offset); + + close(fd); +} + +/* + * Physical write of a page from a buffer slot + * + * For now, assume it's not worth keeping a file pointer open across + * read/write operations. We could cache one virtual file pointer ... + */ +static void +CLOGPhysicalWritePage(int pageno, int slotno) +{ + int segno = pageno / CLOG_PAGES_PER_SEGMENT; + int rpageno = pageno % CLOG_PAGES_PER_SEGMENT; + int offset = rpageno * CLOG_BLCKSZ; + char path[MAXPGPATH]; + int fd; + + ClogFileName(path, segno); + + /* + * If the file doesn't already exist, we should create it. It is possible + * for this to need to happen when writing a page that's not first in + * its segment; we assume the OS can cope with that. (Note: it might seem + * that it'd be okay to create files only when ZeroCLOGPage is called for + * the first page of a segment. However, if after a crash and restart + * the REDO logic elects to replay the log from a checkpoint before the + * latest one, then it's possible that we will get commands to set + * transaction status of transactions that have already been truncated + * from the commit log. Easiest way to deal with that is to accept + * references to nonexistent files here and in CLOGPhysicalReadPage.) + */ + fd = BasicOpenFile(path, O_RDWR | PG_BINARY, S_IRUSR | S_IWUSR); + if (fd < 0) + { + if (errno != ENOENT) + elog(STOP, "open of %s failed: %m", path); + fd = BasicOpenFile(path, O_RDWR | O_CREAT | O_EXCL | PG_BINARY, + S_IRUSR | S_IWUSR); + if (fd < 0) + elog(STOP, "creation of file %s failed: %m", path); + } + + if (lseek(fd, (off_t) offset, SEEK_SET) < 0) + elog(STOP, "lseek of clog file %u, offset %u failed: %m", + segno, offset); + + errno = 0; + if (write(fd, ClogCtl->page_buffer[slotno], CLOG_BLCKSZ) != CLOG_BLCKSZ) + { + /* if write didn't set errno, assume problem is no disk space */ + if (errno == 0) + errno = ENOSPC; + elog(STOP, "write of clog file %u, offset %u failed: %m", + segno, offset); + } + + close(fd); +} + +/* + * Select the slot to re-use when we need a free slot. + * + * The target page number is passed because we need to consider the + * possibility that some other process reads in the target page while + * we are doing I/O to free a slot. Hence, check or recheck to see if + * any slot already holds the target page, and return that slot if so. + * Thus, the returned slot is *either* a slot already holding the pageno + * (could be any state except EMPTY), *or* a freeable slot (state EMPTY + * or CLEAN). + * + * Control lock must be held at entry, and will be held at exit. + */ +static int +SelectLRUCLOGPage(int pageno) +{ + /* Outer loop handles restart after I/O */ + for (;;) + { + int slotno; + int bestslot = 0; + unsigned int bestcount = 0; + + /* See if page already has a buffer assigned */ + for (slotno = 0; slotno < NUM_CLOG_BUFFERS; slotno++) + { + if (ClogCtl->page_number[slotno] == pageno && + ClogCtl->page_status[slotno] != CLOG_PAGE_EMPTY) + return slotno; + } + + /* + * If we find any EMPTY slot, just select that one. + * Else locate the least-recently-used slot that isn't the + * latest CLOG page. + */ + for (slotno = 0; slotno < NUM_CLOG_BUFFERS; slotno++) + { + if (ClogCtl->page_status[slotno] == CLOG_PAGE_EMPTY) + return slotno; + if (ClogCtl->page_lru_count[slotno] > bestcount && + ClogCtl->page_number[slotno] != ClogCtl->latest_page_number) + { + bestslot = slotno; + bestcount = ClogCtl->page_lru_count[slotno]; + } + } + + /* + * If the selected page is clean, we're set. + */ + if (ClogCtl->page_status[bestslot] == CLOG_PAGE_CLEAN) + return bestslot; + + /* + * We need to do I/O. Normal case is that we have to write it out, + * but it's possible in the worst case to have selected a read-busy + * page. In that case we use ReadCLOGPage to wait for the read to + * complete. + */ + if (ClogCtl->page_status[bestslot] == CLOG_PAGE_READ_IN_PROGRESS) + (void) ReadCLOGPage(ClogCtl->page_number[bestslot]); + else + WriteCLOGPage(bestslot); + + /* + * Now loop back and try again. This is the easiest way of dealing + * with corner cases such as the victim page being re-dirtied while + * we wrote it. + */ + } +} + +/* + * This must be called ONCE during postmaster or standalone-backend startup, + * after StartupXLOG has initialized ShmemVariableCache->nextXid. + */ +void +StartupCLOG(void) +{ + /* + * Initialize our idea of the latest page number. + */ + ClogCtl->latest_page_number = TransactionIdToPage(ShmemVariableCache->nextXid); +} + +/* + * This must be called ONCE during postmaster or standalone-backend shutdown + */ +void +ShutdownCLOG(void) +{ + int slotno; + + S_LOCK(&(ClogCtl->control_lck)); + + for (slotno = 0; slotno < NUM_CLOG_BUFFERS; slotno++) + { + WriteCLOGPage(slotno); + Assert(ClogCtl->page_status[slotno] == CLOG_PAGE_EMPTY || + ClogCtl->page_status[slotno] == CLOG_PAGE_CLEAN); + } + + S_UNLOCK(&(ClogCtl->control_lck)); +} + +/* + * Perform a checkpoint --- either during shutdown, or on-the-fly + */ +void +CheckPointCLOG(void) +{ + int slotno; + + S_LOCK(&(ClogCtl->control_lck)); + + for (slotno = 0; slotno < NUM_CLOG_BUFFERS; slotno++) + { + WriteCLOGPage(slotno); + /* + * We cannot assert that the slot is clean now, since another + * process might have re-dirtied it already. That's okay. + */ + } + + S_UNLOCK(&(ClogCtl->control_lck)); +} + + +/* + * Make sure that CLOG has room for a newly-allocated XID. + * + * NB: this is called while holding XidGenLock. We want it to be very fast + * most of the time; even when it's not so fast, no actual I/O need happen + * unless we're forced to write out a dirty clog or xlog page to make room + * in shared memory. + */ +void +ExtendCLOG(TransactionId newestXact) +{ + int pageno; + + /* No work except at first XID of a page */ + if (TransactionIdToPgIndex(newestXact) != 0) + return; + + pageno = TransactionIdToPage(newestXact); + + S_LOCK(&(ClogCtl->control_lck)); + + /* Zero the page and make an XLOG entry about it */ + ZeroCLOGPage(pageno, true); + + S_UNLOCK(&(ClogCtl->control_lck)); +} + + +/* + * Remove all CLOG segments before the one holding the passed transaction ID + * + * When this is called, we know that the database logically contains no + * reference to transaction IDs older than oldestXact. However, we must + * not truncate the CLOG until we have performed a checkpoint, to ensure + * that no such references remain on disk either; else a crash just after + * the truncation might leave us with a problem. Since CLOG segments hold + * a large number of transactions, the opportunity to actually remove a + * segment is fairly rare, and so it seems best not to do the checkpoint + * unless we have confirmed that there is a removable segment. Therefore + * we issue the checkpoint command here, not in higher-level code as might + * seem cleaner. + */ +void +TruncateCLOG(TransactionId oldestXact) +{ + int cutoffPage; + int slotno; + + /* + * The cutoff point is the start of the segment containing oldestXact. + */ + oldestXact -= oldestXact % CLOG_XACTS_PER_SEGMENT; + cutoffPage = TransactionIdToPage(oldestXact); + + if (!ScanCLOGDirectory(cutoffPage, false)) + return; /* nothing to remove */ + + /* Perform a CHECKPOINT */ + CreateCheckPoint(false); + + /* + * Scan CLOG shared memory and remove any pages preceding the cutoff + * page, to ensure we won't rewrite them later. (Any dirty pages + * should have been flushed already during the checkpoint, we're + * just being extra careful here.) + */ + S_LOCK(&(ClogCtl->control_lck)); + +restart:; + + for (slotno = 0; slotno < NUM_CLOG_BUFFERS; slotno++) + { + if (ClogCtl->page_status[slotno] == CLOG_PAGE_EMPTY) + continue; + if (!CLOGPagePrecedes(ClogCtl->page_number[slotno], cutoffPage)) + continue; + /* + * If page is CLEAN, just change state to EMPTY (expected case). + */ + if (ClogCtl->page_status[slotno] == CLOG_PAGE_CLEAN) + { + ClogCtl->page_status[slotno] = CLOG_PAGE_EMPTY; + continue; + } + /* + * Hmm, we have (or may have) I/O operations acting on the page, + * so we've got to wait for them to finish and then start again. + * This is the same logic as in SelectLRUCLOGPage. + */ + if (ClogCtl->page_status[slotno] == CLOG_PAGE_READ_IN_PROGRESS) + (void) ReadCLOGPage(ClogCtl->page_number[slotno]); + else + WriteCLOGPage(slotno); + goto restart; + } + + S_UNLOCK(&(ClogCtl->control_lck)); + + /* Now we can remove the old CLOG segment(s) */ + (void) ScanCLOGDirectory(cutoffPage, true); +} + +/* + * TruncateCLOG subroutine: scan CLOG directory for removable segments. + * Actually remove them iff doDeletions is true. Return TRUE iff any + * removable segments were found. Note: no locking is needed. + */ +static bool +ScanCLOGDirectory(int cutoffPage, bool doDeletions) +{ + bool found = false; + DIR *cldir; + struct dirent *clde; + int segno; + int segpage; + char path[MAXPGPATH]; + + cldir = opendir(ClogDir); + if (cldir == NULL) + elog(STOP, "could not open transaction-commit log directory (%s): %m", + ClogDir); + + errno = 0; + while ((clde = readdir(cldir)) != NULL) + { + if (strlen(clde->d_name) == 4 && + strspn(clde->d_name, "0123456789ABCDEF") == 4) + { + segno = (int) strtol(clde->d_name, NULL, 16); + segpage = segno * CLOG_PAGES_PER_SEGMENT; + if (CLOGPagePrecedes(segpage, cutoffPage)) + { + found = true; + if (doDeletions) + { + snprintf(path, MAXPGPATH, "%s/%s", ClogDir, clde->d_name); + unlink(path); + } + } + } + errno = 0; + } + if (errno) + elog(STOP, "could not read transaction-commit log directory (%s): %m", + ClogDir); + closedir(cldir); + + return found; +} + +/* + * Decide which of two CLOG page numbers is "older" for truncation purposes. + * + * We need to use comparison of TransactionIds here in order to do the right + * thing with wraparound XID arithmetic. However, if we are asked about + * page number zero, we don't want to hand InvalidTransactionId to + * TransactionIdPrecedes: it'll get weird about permanent xact IDs. So, + * offset both xids by FirstNormalTransactionId to avoid that. + */ +static bool +CLOGPagePrecedes(int page1, int page2) +{ + TransactionId xid1; + TransactionId xid2; + + xid1 = (TransactionId) page1 * CLOG_XACTS_PER_PAGE; + xid1 += FirstNormalTransactionId; + xid2 = (TransactionId) page2 * CLOG_XACTS_PER_PAGE; + xid2 += FirstNormalTransactionId; + + return TransactionIdPrecedes(xid1, xid2); +} + + +/* + * Write a ZEROPAGE xlog record + * + * Note: xlog record is marked as outside transaction control, since we + * want it to be redone whether the invoking transaction commits or not. + * (Besides which, this is normally done just before entering a transaction.) + */ +static void +WriteZeroPageXlogRec(int pageno) +{ + XLogRecData rdata; + + rdata.buffer = InvalidBuffer; + rdata.data = (char *) (&pageno); + rdata.len = sizeof(int); + rdata.next = NULL; + (void) XLogInsert(RM_CLOG_ID, CLOG_ZEROPAGE | XLOG_NO_TRAN, &rdata); +} + +/* + * CLOG resource manager's routines + */ +void +clog_redo(XLogRecPtr lsn, XLogRecord *record) +{ + uint8 info = record->xl_info & ~XLR_INFO_MASK; + + if (info == CLOG_ZEROPAGE) + { + int pageno; + int slotno; + + memcpy(&pageno, XLogRecGetData(record), sizeof(int)); + + S_LOCK(&(ClogCtl->control_lck)); + + slotno = ZeroCLOGPage(pageno, false); + WriteCLOGPage(slotno); + Assert(ClogCtl->page_status[slotno] == CLOG_PAGE_CLEAN); + + S_UNLOCK(&(ClogCtl->control_lck)); + } +} + +void +clog_undo(XLogRecPtr lsn, XLogRecord *record) +{ +} + +void +clog_desc(char *buf, uint8 xl_info, char *rec) +{ + uint8 info = xl_info & ~XLR_INFO_MASK; + + if (info == CLOG_ZEROPAGE) + { + int pageno; + + memcpy(&pageno, rec, sizeof(int)); + sprintf(buf + strlen(buf), "zeropage: %d", pageno); + } + else + strcat(buf, "UNKNOWN"); +} |