mirror of
				https://github.com/postgres/postgres.git
				synced 2025-10-29 22:49:41 +03:00 
			
		
		
		
	The new bit indicates whether every tuple on the page is already frozen. It is cleared only when the all-visible bit is cleared, and it can be set only when we vacuum a page and find that every tuple on that page is both visible to every transaction and in no need of any future vacuuming. A future commit will use this new bit to optimize away full-table scans that would otherwise be triggered by XID wraparound considerations. A page which is merely all-visible must still be scanned in that case, but a page which is all-frozen need not be. This commit does not attempt that optimization, although that optimization is the goal here. It seems better to get the basic infrastructure in place first. Per discussion, it's very desirable for pg_upgrade to automatically migrate existing VM forks from the old format to the new format. That, too, will be handled in a follow-on patch. Masahiko Sawada, reviewed by Kyotaro Horiguchi, Fujii Masao, Amit Kapila, Simon Riggs, Andres Freund, and others, and substantially revised by me.
		
			
				
	
	
		
			275 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			275 lines
		
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*-------------------------------------------------------------------------
 | |
|  *
 | |
|  * pgstatapproc.c
 | |
|  *		  Bloat estimation functions
 | |
|  *
 | |
|  * Copyright (c) 2014-2016, PostgreSQL Global Development Group
 | |
|  *
 | |
|  * IDENTIFICATION
 | |
|  *		  contrib/pgstattuple/pgstatapprox.c
 | |
|  *
 | |
|  *-------------------------------------------------------------------------
 | |
|  */
 | |
| #include "postgres.h"
 | |
| 
 | |
| #include "access/visibilitymap.h"
 | |
| #include "access/transam.h"
 | |
| #include "access/xact.h"
 | |
| #include "access/multixact.h"
 | |
| #include "access/htup_details.h"
 | |
| #include "catalog/namespace.h"
 | |
| #include "funcapi.h"
 | |
| #include "miscadmin.h"
 | |
| #include "storage/bufmgr.h"
 | |
| #include "storage/freespace.h"
 | |
| #include "storage/procarray.h"
 | |
| #include "storage/lmgr.h"
 | |
| #include "utils/builtins.h"
 | |
| #include "utils/tqual.h"
 | |
| #include "commands/vacuum.h"
 | |
| 
 | |
| PG_FUNCTION_INFO_V1(pgstattuple_approx);
 | |
| 
 | |
| typedef struct output_type
 | |
| {
 | |
| 	uint64		table_len;
 | |
| 	uint64		scanned_percent;
 | |
| 	uint64		tuple_count;
 | |
| 	uint64		tuple_len;
 | |
| 	double		tuple_percent;
 | |
| 	uint64		dead_tuple_count;
 | |
| 	uint64		dead_tuple_len;
 | |
| 	double		dead_tuple_percent;
 | |
| 	uint64		free_space;
 | |
| 	double		free_percent;
 | |
| } output_type;
 | |
| 
 | |
| #define NUM_OUTPUT_COLUMNS 10
 | |
| 
 | |
| /*
 | |
|  * This function takes an already open relation and scans its pages,
 | |
|  * skipping those that have the corresponding visibility map bit set.
 | |
|  * For pages we skip, we find the free space from the free space map
 | |
|  * and approximate tuple_len on that basis. For the others, we count
 | |
|  * the exact number of dead tuples etc.
 | |
|  *
 | |
|  * This scan is loosely based on vacuumlazy.c:lazy_scan_heap(), but
 | |
|  * we do not try to avoid skipping single pages.
 | |
|  */
 | |
| static void
 | |
| statapprox_heap(Relation rel, output_type *stat)
 | |
| {
 | |
| 	BlockNumber scanned,
 | |
| 				nblocks,
 | |
| 				blkno;
 | |
| 	Buffer		vmbuffer = InvalidBuffer;
 | |
| 	BufferAccessStrategy bstrategy;
 | |
| 	TransactionId OldestXmin;
 | |
| 	uint64		misc_count = 0;
 | |
| 
 | |
| 	OldestXmin = GetOldestXmin(rel, true);
 | |
| 	bstrategy = GetAccessStrategy(BAS_BULKREAD);
 | |
| 
 | |
| 	nblocks = RelationGetNumberOfBlocks(rel);
 | |
| 	scanned = 0;
 | |
| 
 | |
| 	for (blkno = 0; blkno < nblocks; blkno++)
 | |
| 	{
 | |
| 		Buffer		buf;
 | |
| 		Page		page;
 | |
| 		OffsetNumber offnum,
 | |
| 					maxoff;
 | |
| 		Size		freespace;
 | |
| 
 | |
| 		CHECK_FOR_INTERRUPTS();
 | |
| 
 | |
| 		/*
 | |
| 		 * If the page has only visible tuples, then we can find out the free
 | |
| 		 * space from the FSM and move on.
 | |
| 		 */
 | |
| 		if (VM_ALL_VISIBLE(rel, blkno, &vmbuffer))
 | |
| 		{
 | |
| 			freespace = GetRecordedFreeSpace(rel, blkno);
 | |
| 			stat->tuple_len += BLCKSZ - freespace;
 | |
| 			stat->free_space += freespace;
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		buf = ReadBufferExtended(rel, MAIN_FORKNUM, blkno,
 | |
| 								 RBM_NORMAL, bstrategy);
 | |
| 
 | |
| 		LockBuffer(buf, BUFFER_LOCK_SHARE);
 | |
| 
 | |
| 		page = BufferGetPage(buf);
 | |
| 
 | |
| 		/*
 | |
| 		 * It's not safe to call PageGetHeapFreeSpace() on new pages, so we
 | |
| 		 * treat them as being free space for our purposes.
 | |
| 		 */
 | |
| 		if (!PageIsNew(page))
 | |
| 			stat->free_space += PageGetHeapFreeSpace(page);
 | |
| 		else
 | |
| 			stat->free_space += BLCKSZ - SizeOfPageHeaderData;
 | |
| 
 | |
| 		if (PageIsNew(page) || PageIsEmpty(page))
 | |
| 		{
 | |
| 			UnlockReleaseBuffer(buf);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		scanned++;
 | |
| 
 | |
| 		/*
 | |
| 		 * Look at each tuple on the page and decide whether it's live or
 | |
| 		 * dead, then count it and its size. Unlike lazy_scan_heap, we can
 | |
| 		 * afford to ignore problems and special cases.
 | |
| 		 */
 | |
| 		maxoff = PageGetMaxOffsetNumber(page);
 | |
| 
 | |
| 		for (offnum = FirstOffsetNumber;
 | |
| 			 offnum <= maxoff;
 | |
| 			 offnum = OffsetNumberNext(offnum))
 | |
| 		{
 | |
| 			ItemId		itemid;
 | |
| 			HeapTupleData tuple;
 | |
| 
 | |
| 			itemid = PageGetItemId(page, offnum);
 | |
| 
 | |
| 			if (!ItemIdIsUsed(itemid) || ItemIdIsRedirected(itemid) ||
 | |
| 				ItemIdIsDead(itemid))
 | |
| 			{
 | |
| 				continue;
 | |
| 			}
 | |
| 
 | |
| 			Assert(ItemIdIsNormal(itemid));
 | |
| 
 | |
| 			ItemPointerSet(&(tuple.t_self), blkno, offnum);
 | |
| 
 | |
| 			tuple.t_data = (HeapTupleHeader) PageGetItem(page, itemid);
 | |
| 			tuple.t_len = ItemIdGetLength(itemid);
 | |
| 			tuple.t_tableOid = RelationGetRelid(rel);
 | |
| 
 | |
| 			/*
 | |
| 			 * We count live and dead tuples, but we also need to add up
 | |
| 			 * others in order to feed vac_estimate_reltuples.
 | |
| 			 */
 | |
| 			switch (HeapTupleSatisfiesVacuum(&tuple, OldestXmin, buf))
 | |
| 			{
 | |
| 				case HEAPTUPLE_RECENTLY_DEAD:
 | |
| 					misc_count++;
 | |
| 					/* Fall through */
 | |
| 				case HEAPTUPLE_DEAD:
 | |
| 					stat->dead_tuple_len += tuple.t_len;
 | |
| 					stat->dead_tuple_count++;
 | |
| 					break;
 | |
| 				case HEAPTUPLE_LIVE:
 | |
| 					stat->tuple_len += tuple.t_len;
 | |
| 					stat->tuple_count++;
 | |
| 					break;
 | |
| 				case HEAPTUPLE_INSERT_IN_PROGRESS:
 | |
| 				case HEAPTUPLE_DELETE_IN_PROGRESS:
 | |
| 					misc_count++;
 | |
| 					break;
 | |
| 				default:
 | |
| 					elog(ERROR, "unexpected HeapTupleSatisfiesVacuum result");
 | |
| 					break;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		UnlockReleaseBuffer(buf);
 | |
| 	}
 | |
| 
 | |
| 	stat->table_len = (uint64) nblocks *BLCKSZ;
 | |
| 
 | |
| 	stat->tuple_count = vac_estimate_reltuples(rel, false, nblocks, scanned,
 | |
| 											 stat->tuple_count + misc_count);
 | |
| 
 | |
| 	/*
 | |
| 	 * Calculate percentages if the relation has one or more pages.
 | |
| 	 */
 | |
| 	if (nblocks != 0)
 | |
| 	{
 | |
| 		stat->scanned_percent = 100 * scanned / nblocks;
 | |
| 		stat->tuple_percent = 100.0 * stat->tuple_len / stat->table_len;
 | |
| 		stat->dead_tuple_percent = 100.0 * stat->dead_tuple_len / stat->table_len;
 | |
| 		stat->free_percent = 100.0 * stat->free_space / stat->table_len;
 | |
| 	}
 | |
| 
 | |
| 	if (BufferIsValid(vmbuffer))
 | |
| 	{
 | |
| 		ReleaseBuffer(vmbuffer);
 | |
| 		vmbuffer = InvalidBuffer;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Returns estimated live/dead tuple statistics for the given relid.
 | |
|  */
 | |
| Datum
 | |
| pgstattuple_approx(PG_FUNCTION_ARGS)
 | |
| {
 | |
| 	Oid			relid = PG_GETARG_OID(0);
 | |
| 	Relation	rel;
 | |
| 	output_type stat = {0};
 | |
| 	TupleDesc	tupdesc;
 | |
| 	bool		nulls[NUM_OUTPUT_COLUMNS];
 | |
| 	Datum		values[NUM_OUTPUT_COLUMNS];
 | |
| 	HeapTuple	ret;
 | |
| 	int			i = 0;
 | |
| 
 | |
| 	if (!superuser())
 | |
| 		ereport(ERROR,
 | |
| 				(errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
 | |
| 				 (errmsg("must be superuser to use pgstattuple functions"))));
 | |
| 
 | |
| 	if (get_call_result_type(fcinfo, NULL, &tupdesc) != TYPEFUNC_COMPOSITE)
 | |
| 		elog(ERROR, "return type must be a row type");
 | |
| 
 | |
| 	if (tupdesc->natts != NUM_OUTPUT_COLUMNS)
 | |
| 		elog(ERROR, "incorrect number of output arguments");
 | |
| 
 | |
| 	rel = relation_open(relid, AccessShareLock);
 | |
| 
 | |
| 	/*
 | |
| 	 * Reject attempts to read non-local temporary relations; we would be
 | |
| 	 * likely to get wrong data since we have no visibility into the owning
 | |
| 	 * session's local buffers.
 | |
| 	 */
 | |
| 	if (RELATION_IS_OTHER_TEMP(rel))
 | |
| 		ereport(ERROR,
 | |
| 				(errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
 | |
| 				 errmsg("cannot access temporary tables of other sessions")));
 | |
| 
 | |
| 	/*
 | |
| 	 * We support only ordinary relations and materialised views, because we
 | |
| 	 * depend on the visibility map and free space map for our estimates about
 | |
| 	 * unscanned pages.
 | |
| 	 */
 | |
| 	if (!(rel->rd_rel->relkind == RELKIND_RELATION ||
 | |
| 		  rel->rd_rel->relkind == RELKIND_MATVIEW))
 | |
| 		ereport(ERROR,
 | |
| 				(errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
 | |
| 				 errmsg("\"%s\" is not a table or materialized view",
 | |
| 						RelationGetRelationName(rel))));
 | |
| 
 | |
| 	statapprox_heap(rel, &stat);
 | |
| 
 | |
| 	relation_close(rel, AccessShareLock);
 | |
| 
 | |
| 	memset(nulls, 0, sizeof(nulls));
 | |
| 
 | |
| 	values[i++] = Int64GetDatum(stat.table_len);
 | |
| 	values[i++] = Float8GetDatum(stat.scanned_percent);
 | |
| 	values[i++] = Int64GetDatum(stat.tuple_count);
 | |
| 	values[i++] = Int64GetDatum(stat.tuple_len);
 | |
| 	values[i++] = Float8GetDatum(stat.tuple_percent);
 | |
| 	values[i++] = Int64GetDatum(stat.dead_tuple_count);
 | |
| 	values[i++] = Int64GetDatum(stat.dead_tuple_len);
 | |
| 	values[i++] = Float8GetDatum(stat.dead_tuple_percent);
 | |
| 	values[i++] = Int64GetDatum(stat.free_space);
 | |
| 	values[i++] = Float8GetDatum(stat.free_percent);
 | |
| 
 | |
| 	ret = heap_form_tuple(tupdesc, values, nulls);
 | |
| 	return HeapTupleGetDatum(ret);
 | |
| }
 |