mirror of
https://github.com/postgres/postgres.git
synced 2025-11-28 11:44:57 +03:00
Fix typos in comments.
Backpatch to all supported versions, where applicable, to make backpatching of future fixes go more smoothly. Josh Soref Discussion: https://www.postgresql.org/message-id/CACZqfqCf+5qRztLPgmmosr-B0Ye4srWzzw_mo4c_8_B_mtjmJQ@mail.gmail.com
This commit is contained in:
@@ -860,7 +860,7 @@ WaitEventAdjustWin32(WaitEventSet *set, WaitEvent *event)
|
||||
* reached. At most nevents occurred events are returned.
|
||||
*
|
||||
* If timeout = -1, block until an event occurs; if 0, check sockets for
|
||||
* readiness, but don't block; if > 0, block for at most timeout miliseconds.
|
||||
* readiness, but don't block; if > 0, block for at most timeout milliseconds.
|
||||
*
|
||||
* Returns the number of events occurred, or 0 if the timeout was reached.
|
||||
*
|
||||
|
||||
@@ -501,7 +501,7 @@ shm_mq_sendv(shm_mq_handle *mqh, shm_mq_iovec *iov, int iovcnt, bool nowait)
|
||||
* it will point to a temporary buffer. This mostly avoids data copying in
|
||||
* the hoped-for case where messages are short compared to the buffer size,
|
||||
* while still allowing longer messages. In either case, the return value
|
||||
* remains valid until the next receive operation is perfomed on the queue.
|
||||
* remains valid until the next receive operation is performed on the queue.
|
||||
*
|
||||
* When nowait = false, we'll wait on our process latch when the ring buffer
|
||||
* is empty and we have not yet received a full message. The sender will
|
||||
|
||||
@@ -967,7 +967,7 @@ LogStandbySnapshot(void)
|
||||
* similar. We keep them separate because xl_xact_running_xacts is a
|
||||
* contiguous chunk of memory and never exists fully until it is assembled in
|
||||
* WAL. The inserted records are marked as not being important for durability,
|
||||
* to avoid triggering superflous checkpoint / archiving activity.
|
||||
* to avoid triggering superfluous checkpoint / archiving activity.
|
||||
*/
|
||||
static XLogRecPtr
|
||||
LogCurrentRunningXacts(RunningTransactions CurrRunningXacts)
|
||||
|
||||
@@ -2778,7 +2778,7 @@ GetLockConflicts(const LOCKTAG *locktag, LOCKMODE lockmode)
|
||||
vxids = (VirtualTransactionId *)
|
||||
palloc0(sizeof(VirtualTransactionId) * (MaxBackends + 1));
|
||||
|
||||
/* Compute hash code and partiton lock, and look up conflicting modes. */
|
||||
/* Compute hash code and partition lock, and look up conflicting modes. */
|
||||
hashcode = LockTagHashCode(locktag);
|
||||
partitionLock = LockHashPartitionLock(hashcode);
|
||||
conflictMask = lockMethodTable->conflictTab[lockmode];
|
||||
|
||||
@@ -781,7 +781,7 @@ LWLockAttemptLock(LWLock *lock, LWLockMode mode)
|
||||
return false;
|
||||
}
|
||||
else
|
||||
return true; /* someobdy else has the lock */
|
||||
return true; /* somebody else has the lock */
|
||||
}
|
||||
}
|
||||
pg_unreachable();
|
||||
@@ -953,7 +953,7 @@ LWLockWakeup(LWLock *lock)
|
||||
* that happens before the list unlink happens, the list would end up
|
||||
* being corrupted.
|
||||
*
|
||||
* The barrier pairs with the LWLockWaitListLock() when enqueueing for
|
||||
* The barrier pairs with the LWLockWaitListLock() when enqueuing for
|
||||
* another lock.
|
||||
*/
|
||||
pg_write_barrier();
|
||||
@@ -1029,7 +1029,7 @@ LWLockDequeueSelf(LWLock *lock)
|
||||
|
||||
/*
|
||||
* Can't just remove ourselves from the list, but we need to iterate over
|
||||
* all entries as somebody else could have unqueued us.
|
||||
* all entries as somebody else could have dequeued us.
|
||||
*/
|
||||
proclist_foreach_modify(iter, &lock->waiters, lwWaitLink)
|
||||
{
|
||||
|
||||
@@ -3193,7 +3193,7 @@ ReleasePredicateLocks(bool isCommit)
|
||||
/*
|
||||
* We can't trust XactReadOnly here, because a transaction which started
|
||||
* as READ WRITE can show as READ ONLY later, e.g., within
|
||||
* substransactions. We want to flag a transaction as READ ONLY if it
|
||||
* subtransactions. We want to flag a transaction as READ ONLY if it
|
||||
* commits without writing so that de facto READ ONLY transactions get the
|
||||
* benefit of some RO optimizations, so we will use this local variable to
|
||||
* get some cleanup logic right which is based on whether the transaction
|
||||
|
||||
@@ -1728,7 +1728,7 @@ _fdvec_resize(SMgrRelation reln,
|
||||
else
|
||||
{
|
||||
/*
|
||||
* It doesn't seem worthwile complicating the code by having a more
|
||||
* It doesn't seem worthwhile complicating the code by having a more
|
||||
* aggressive growth strategy here; the number of segments doesn't
|
||||
* grow that fast, and the memory context internally will sometimes
|
||||
* avoid doing an actual reallocation.
|
||||
|
||||
Reference in New Issue
Block a user