mirror of
https://github.com/postgres/postgres.git
synced 2025-11-10 17:42:29 +03:00
Improve hash index bucket split behavior.
Previously, the right to split a bucket was represented by a heavyweight lock on the page number of the primary bucket page. Unfortunately, this meant that every scan needed to take a heavyweight lock on that bucket also, which was bad for concurrency. Instead, use a cleanup lock on the primary bucket page to indicate the right to begin a split, so that scans only need to retain a pin on that page, which is they would have to acquire anyway, and which is also much cheaper. In addition to reducing the locking cost, this also avoids locking out scans and inserts for the entire lifetime of the split: while the new bucket is being populated with copies of the appropriate tuples from the old bucket, scans and inserts can happen in parallel. There are minor concurrency improvements for vacuum operations as well, though the situation there is still far from ideal. This patch also removes the unworldly assumption that a split will never be interrupted. With the new code, a split is done in a series of small steps and the system can pick up where it left off if it is interrupted prior to completion. While this patch does not itself add write-ahead logging for hash indexes, it is clearly a necessary first step, since one of the things that could interrupt a split is the removal of electrical power from the machine performing it. Amit Kapila. I wrote the original design on which this patch is based, and did a good bit of work on the comments and README through multiple rounds of review, but all of the code is Amit's. Also reviewed by Jesper Pedersen, Jeff Janes, and others. Discussion: http://postgr.es/m/CAA4eK1LfzcZYxLoXS874Ad0+S-ZM60U9bwcyiUZx9mHZ-KCWhw@mail.gmail.com
This commit is contained in:
@@ -20,6 +20,8 @@
|
||||
#include "utils/lsyscache.h"
|
||||
#include "utils/rel.h"
|
||||
|
||||
#define CALC_NEW_BUCKET(old_bucket, lowmask) \
|
||||
old_bucket | (lowmask + 1)
|
||||
|
||||
/*
|
||||
* _hash_checkqual -- does the index tuple satisfy the scan conditions?
|
||||
@@ -352,3 +354,95 @@ _hash_binsearch_last(Page page, uint32 hash_value)
|
||||
|
||||
return lower;
|
||||
}
|
||||
|
||||
/*
|
||||
* _hash_get_oldblock_from_newbucket() -- get the block number of a bucket
|
||||
* from which current (new) bucket is being split.
|
||||
*/
|
||||
BlockNumber
|
||||
_hash_get_oldblock_from_newbucket(Relation rel, Bucket new_bucket)
|
||||
{
|
||||
Bucket old_bucket;
|
||||
uint32 mask;
|
||||
Buffer metabuf;
|
||||
HashMetaPage metap;
|
||||
BlockNumber blkno;
|
||||
|
||||
/*
|
||||
* To get the old bucket from the current bucket, we need a mask to modulo
|
||||
* into lower half of table. This mask is stored in meta page as
|
||||
* hashm_lowmask, but here we can't rely on the same, because we need a
|
||||
* value of lowmask that was prevalent at the time when bucket split was
|
||||
* started. Masking the most significant bit of new bucket would give us
|
||||
* old bucket.
|
||||
*/
|
||||
mask = (((uint32) 1) << (fls(new_bucket) - 1)) - 1;
|
||||
old_bucket = new_bucket & mask;
|
||||
|
||||
metabuf = _hash_getbuf(rel, HASH_METAPAGE, HASH_READ, LH_META_PAGE);
|
||||
metap = HashPageGetMeta(BufferGetPage(metabuf));
|
||||
|
||||
blkno = BUCKET_TO_BLKNO(metap, old_bucket);
|
||||
|
||||
_hash_relbuf(rel, metabuf);
|
||||
|
||||
return blkno;
|
||||
}
|
||||
|
||||
/*
|
||||
* _hash_get_newblock_from_oldbucket() -- get the block number of a bucket
|
||||
* that will be generated after split from old bucket.
|
||||
*
|
||||
* This is used to find the new bucket from old bucket based on current table
|
||||
* half. It is mainly required to finish the incomplete splits where we are
|
||||
* sure that not more than one bucket could have split in progress from old
|
||||
* bucket.
|
||||
*/
|
||||
BlockNumber
|
||||
_hash_get_newblock_from_oldbucket(Relation rel, Bucket old_bucket)
|
||||
{
|
||||
Bucket new_bucket;
|
||||
Buffer metabuf;
|
||||
HashMetaPage metap;
|
||||
BlockNumber blkno;
|
||||
|
||||
metabuf = _hash_getbuf(rel, HASH_METAPAGE, HASH_READ, LH_META_PAGE);
|
||||
metap = HashPageGetMeta(BufferGetPage(metabuf));
|
||||
|
||||
new_bucket = _hash_get_newbucket_from_oldbucket(rel, old_bucket,
|
||||
metap->hashm_lowmask,
|
||||
metap->hashm_maxbucket);
|
||||
blkno = BUCKET_TO_BLKNO(metap, new_bucket);
|
||||
|
||||
_hash_relbuf(rel, metabuf);
|
||||
|
||||
return blkno;
|
||||
}
|
||||
|
||||
/*
|
||||
* _hash_get_newbucket_from_oldbucket() -- get the new bucket that will be
|
||||
* generated after split from current (old) bucket.
|
||||
*
|
||||
* This is used to find the new bucket from old bucket. New bucket can be
|
||||
* obtained by OR'ing old bucket with most significant bit of current table
|
||||
* half (lowmask passed in this function can be used to identify msb of
|
||||
* current table half). There could be multiple buckets that could have
|
||||
* been split from current bucket. We need the first such bucket that exists.
|
||||
* Caller must ensure that no more than one split has happened from old
|
||||
* bucket.
|
||||
*/
|
||||
Bucket
|
||||
_hash_get_newbucket_from_oldbucket(Relation rel, Bucket old_bucket,
|
||||
uint32 lowmask, uint32 maxbucket)
|
||||
{
|
||||
Bucket new_bucket;
|
||||
|
||||
new_bucket = CALC_NEW_BUCKET(old_bucket, lowmask);
|
||||
if (new_bucket > maxbucket)
|
||||
{
|
||||
lowmask = lowmask >> 1;
|
||||
new_bucket = CALC_NEW_BUCKET(old_bucket, lowmask);
|
||||
}
|
||||
|
||||
return new_bucket;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user