PostgreSQL Source Code git master
Loading...
Searching...
No Matches
visibilitymap.c File Reference
#include "postgres.h"
#include "access/heapam_xlog.h"
#include "access/visibilitymap.h"
#include "access/xloginsert.h"
#include "access/xlogutils.h"
#include "miscadmin.h"
#include "port/pg_bitutils.h"
#include "storage/bufmgr.h"
#include "storage/smgr.h"
#include "utils/inval.h"
#include "utils/rel.h"
Include dependency graph for visibilitymap.c:

Go to the source code of this file.

Macros

#define MAPSIZE   (BLCKSZ - MAXALIGN(SizeOfPageHeaderData))
 
#define HEAPBLOCKS_PER_BYTE   (BITS_PER_BYTE / BITS_PER_HEAPBLOCK)
 
#define HEAPBLOCKS_PER_PAGE   (MAPSIZE * HEAPBLOCKS_PER_BYTE)
 
#define HEAPBLK_TO_MAPBLOCK(x)   ((x) / HEAPBLOCKS_PER_PAGE)
 
#define HEAPBLK_TO_MAPBLOCK_LIMIT(x)    (((x) + HEAPBLOCKS_PER_PAGE - 1) / HEAPBLOCKS_PER_PAGE)
 
#define HEAPBLK_TO_MAPBYTE(x)   (((x) % HEAPBLOCKS_PER_PAGE) / HEAPBLOCKS_PER_BYTE)
 
#define HEAPBLK_TO_OFFSET(x)   (((x) % HEAPBLOCKS_PER_BYTE) * BITS_PER_HEAPBLOCK)
 
#define VISIBLE_MASK8   (0x55) /* The lower bit of each bit pair */
 
#define FROZEN_MASK8   (0xaa) /* The upper bit of each bit pair */
 

Functions

static Buffer vm_readbuf (Relation rel, BlockNumber blkno, bool extend)
 
static Buffer vm_extend (Relation rel, BlockNumber vm_nblocks)
 
bool visibilitymap_clear (Relation rel, BlockNumber heapBlk, Buffer vmbuf, uint8 flags)
 
void visibilitymap_pin (Relation rel, BlockNumber heapBlk, Buffer *vmbuf)
 
bool visibilitymap_pin_ok (BlockNumber heapBlk, Buffer vmbuf)
 
void visibilitymap_set (BlockNumber heapBlk, Buffer vmBuf, uint8 flags, const RelFileLocator rlocator)
 
uint8 visibilitymap_get_status (Relation rel, BlockNumber heapBlk, Buffer *vmbuf)
 
void visibilitymap_count (Relation rel, BlockNumber *all_visible, BlockNumber *all_frozen)
 
BlockNumber visibilitymap_prepare_truncate (Relation rel, BlockNumber nheapblocks)
 
BlockNumber visibilitymap_truncation_length (BlockNumber nheapblocks)
 

Macro Definition Documentation

◆ FROZEN_MASK8

#define FROZEN_MASK8   (0xaa) /* The upper bit of each bit pair */

Definition at line 136 of file visibilitymap.c.

◆ HEAPBLK_TO_MAPBLOCK

#define HEAPBLK_TO_MAPBLOCK (   x)    ((x) / HEAPBLOCKS_PER_PAGE)

Definition at line 128 of file visibilitymap.c.

◆ HEAPBLK_TO_MAPBLOCK_LIMIT

#define HEAPBLK_TO_MAPBLOCK_LIMIT (   x)     (((x) + HEAPBLOCKS_PER_PAGE - 1) / HEAPBLOCKS_PER_PAGE)

Definition at line 129 of file visibilitymap.c.

152{
154 int mapByte = HEAPBLK_TO_MAPBYTE(heapBlk);
155 int mapOffset = HEAPBLK_TO_OFFSET(heapBlk);
156 uint8 mask = flags << mapOffset;
157 char *map;
158 bool cleared = false;
159
160 /* Must never clear all_visible bit while leaving all_frozen bit set */
163
164#ifdef TRACE_VISIBILITYMAP
165 elog(DEBUG1, "vm_clear %s %d", RelationGetRelationName(rel), heapBlk);
166#endif
167
169 elog(ERROR, "wrong buffer passed to visibilitymap_clear");
170
173
174 if (map[mapByte] & mask)
175 {
176 map[mapByte] &= ~mask;
177
179 cleared = true;
180 }
181
183
184 return cleared;
185}
186
187/*
188 * visibilitymap_pin - pin a map page for setting a bit
189 *
190 * Setting a bit in the visibility map is a two-phase operation. First, call
191 * visibilitymap_pin, to pin the visibility map page containing the bit for
192 * the heap page. Because that can require I/O to read the map page, you
193 * shouldn't hold a lock on the heap page while doing that. Then, call
194 * visibilitymap_set to actually set the bit.
195 *
196 * On entry, *vmbuf should be InvalidBuffer or a valid buffer returned by
197 * an earlier call to visibilitymap_pin or visibilitymap_get_status on the same
198 * relation. On return, *vmbuf is a valid buffer with the map page containing
199 * the bit for heapBlk.
200 *
201 * If the page doesn't exist in the map file yet, it is extended.
202 */
203void
205{
207
208 /* Reuse the old pinned buffer if possible */
209 if (BufferIsValid(*vmbuf))
210 {
212 return;
213
215 }
216 *vmbuf = vm_readbuf(rel, mapBlock, true);
217}
218
219/*
220 * visibilitymap_pin_ok - do we already have the correct page pinned?
221 *
222 * On entry, vmbuf should be InvalidBuffer or a valid buffer returned by
223 * an earlier call to visibilitymap_pin or visibilitymap_get_status on the same
224 * relation. The return value indicates whether the buffer covers the
225 * given heapBlk.
226 */
227bool
229{
231
233}
234
235/*
236 * Set VM (visibility map) flags in the VM block in vmBuf.
237 *
238 * This function is intended for callers that log VM changes together
239 * with the heap page modifications that rendered the page all-visible.
240 *
241 * vmBuf must be pinned and exclusively locked, and it must cover the VM bits
242 * corresponding to heapBlk.
243 *
244 * In normal operation (not recovery), this must be called inside a critical
245 * section that also applies the necessary heap page changes and, if
246 * applicable, emits WAL.
247 *
248 * The caller is responsible for ensuring consistency between the heap page
249 * and the VM page by holding a pin and exclusive lock on the buffer
250 * containing heapBlk.
251 *
252 * rlocator is used only for debugging messages.
253 */
254void
256 Buffer vmBuf, uint8 flags,
257 const RelFileLocator rlocator)
258{
262 Page page;
263 uint8 *map;
264 uint8 status;
265
266#ifdef TRACE_VISIBILITYMAP
267 elog(DEBUG1, "vm_set flags 0x%02X for %s %d",
268 flags,
270 heapBlk);
271#endif
272
273 /* Call in same critical section where WAL is emitted. */
275
276 /* Flags should be valid. Also never clear bits with this function */
277 Assert((flags & VISIBILITYMAP_VALID_BITS) == flags);
278
279 /* Must never set all_frozen bit without also setting all_visible bit */
281
282 /* Check that we have the right VM page pinned */
284 elog(ERROR, "wrong VM buffer passed to visibilitymap_set");
285
287
288 page = BufferGetPage(vmBuf);
289 map = (uint8 *) PageGetContents(page);
290
291 status = (map[mapByte] >> mapOffset) & VISIBILITYMAP_VALID_BITS;
292 if (flags != status)
293 {
294 map[mapByte] |= (flags << mapOffset);
296 }
297}
298
299/*
300 * visibilitymap_get_status - get status of bits
301 *
302 * Are all tuples on heapBlk visible to all or are marked frozen, according
303 * to the visibility map?
304 *
305 * On entry, *vmbuf should be InvalidBuffer or a valid buffer returned by an
306 * earlier call to visibilitymap_pin or visibilitymap_get_status on the same
307 * relation. On return, *vmbuf is a valid buffer with the map page containing
308 * the bit for heapBlk, or InvalidBuffer. The caller is responsible for
309 * releasing *vmbuf after it's done testing and setting bits.
310 *
311 * NOTE: This function is typically called without a lock on the heap page,
312 * so somebody else could change the bit just after we look at it. In fact,
313 * since we don't lock the visibility map page either, it's even possible that
314 * someone else could have changed the bit just before we look at it, but yet
315 * we might see the old value. It is the caller's responsibility to deal with
316 * all concurrency issues!
317 */
318uint8
320{
324 char *map;
326
327#ifdef TRACE_VISIBILITYMAP
328 elog(DEBUG1, "vm_get_status %s %d", RelationGetRelationName(rel), heapBlk);
329#endif
330
331 /* Reuse the old pinned buffer if possible */
332 if (BufferIsValid(*vmbuf))
333 {
335 {
338 }
339 }
340
341 if (!BufferIsValid(*vmbuf))
342 {
343 *vmbuf = vm_readbuf(rel, mapBlock, false);
344 if (!BufferIsValid(*vmbuf))
345 return (uint8) 0;
346 }
347
349
350 /*
351 * A single byte read is atomic. There could be memory-ordering effects
352 * here, but for performance reasons we make it the caller's job to worry
353 * about that.
354 */
356 return result;
357}
358
359/*
360 * visibilitymap_count - count number of bits set in visibility map
361 *
362 * Note: we ignore the possibility of race conditions when the table is being
363 * extended concurrently with the call. New pages added to the table aren't
364 * going to be marked all-visible or all-frozen, so they won't affect the result.
365 */
366void
367visibilitymap_count(Relation rel, BlockNumber *all_visible, BlockNumber *all_frozen)
368{
371 BlockNumber nfrozen = 0;
372
373 /* all_visible must be specified */
374 Assert(all_visible);
375
376 for (mapBlock = 0;; mapBlock++)
377 {
379 uint64 *map;
380
381 /*
382 * Read till we fall off the end of the map. We assume that any extra
383 * bytes in the last page are zeroed, so we don't bother excluding
384 * them from the count.
385 */
386 mapBuffer = vm_readbuf(rel, mapBlock, false);
388 break;
389
390 /*
391 * We choose not to lock the page, since the result is going to be
392 * immediately stale anyway if anyone is concurrently setting or
393 * clearing bits, and we only really need an approximate value.
394 */
396
397 nvisible += pg_popcount_masked((const char *) map, MAPSIZE, VISIBLE_MASK8);
398 if (all_frozen)
399 nfrozen += pg_popcount_masked((const char *) map, MAPSIZE, FROZEN_MASK8);
400
402 }
403
404 *all_visible = nvisible;
405 if (all_frozen)
406 *all_frozen = nfrozen;
407}
408
409/*
410 * visibilitymap_prepare_truncate -
411 * prepare for truncation of the visibility map
412 *
413 * nheapblocks is the new size of the heap.
414 *
415 * Return the number of blocks of new visibility map.
416 * If it's InvalidBlockNumber, there is nothing to truncate;
417 * otherwise the caller is responsible for calling smgrtruncate()
418 * to truncate the visibility map pages.
419 */
422{
424
425 /* last remaining block, byte, and bit */
429
430#ifdef TRACE_VISIBILITYMAP
431 elog(DEBUG1, "vm_truncate %s %d", RelationGetRelationName(rel), nheapblocks);
432#endif
433
434 /*
435 * If no visibility map has been created yet for this relation, there's
436 * nothing to truncate.
437 */
439 return InvalidBlockNumber;
440
441 /*
442 * Unless the new size is exactly at a visibility map page boundary, the
443 * tail bits in the last remaining map page, representing truncated heap
444 * blocks, need to be cleared. This is not only tidy, but also necessary
445 * because we don't get a chance to clear the bits if the heap is extended
446 * again.
447 */
448 if (truncByte != 0 || truncOffset != 0)
449 {
451 Page page;
452 char *map;
453
455
456 mapBuffer = vm_readbuf(rel, truncBlock, false);
458 {
459 /* nothing to do, the file was already smaller */
460 return InvalidBlockNumber;
461 }
462
463 page = BufferGetPage(mapBuffer);
464 map = PageGetContents(page);
465
467
468 /* NO EREPORT(ERROR) from here till changes are logged */
470
471 /* Clear out the unwanted bytes. */
472 MemSet(&map[truncByte + 1], 0, MAPSIZE - (truncByte + 1));
473
474 /*----
475 * Mask out the unwanted bits of the last remaining byte.
476 *
477 * ((1 << 0) - 1) = 00000000
478 * ((1 << 1) - 1) = 00000001
479 * ...
480 * ((1 << 6) - 1) = 00111111
481 * ((1 << 7) - 1) = 01111111
482 *----
483 */
484 map[truncByte] &= (1 << truncOffset) - 1;
485
486 /*
487 * Truncation of a relation is WAL-logged at a higher-level, and we
488 * will be called at WAL replay. But if checksums are enabled, we need
489 * to still write a WAL record to protect against a torn page, if the
490 * page is flushed to disk before the truncation WAL record. We cannot
491 * use MarkBufferDirtyHint here, because that will not dirty the page
492 * during recovery.
493 */
497
499
501 }
502 else
504
506 {
507 /* nothing to do, the file was already smaller than requested size */
508 return InvalidBlockNumber;
509 }
510
511 return newnblocks;
512}
513
514/*
515 * visibilitymap_truncation_length -
516 * compute truncation length for visibility map
517 *
518 * Given a proposed truncation length for the main fork, compute the
519 * correct truncation length for the visibility map. Should return the
520 * same answer as visibilitymap_prepare_truncate(), but without modifying
521 * anything.
522 */
525{
527}
528
529/*
530 * Read a visibility map page.
531 *
532 * If the page doesn't exist, InvalidBuffer is returned, or if 'extend' is
533 * true, the visibility map file is extended.
534 */
535static Buffer
536vm_readbuf(Relation rel, BlockNumber blkno, bool extend)
537{
538 Buffer buf;
540
541 /*
542 * Caution: re-using this smgr pointer could fail if the relcache entry
543 * gets closed. It's safe as long as we only do smgr-level operations
544 * between here and the last use of the pointer.
545 */
546 reln = RelationGetSmgr(rel);
547
548 /*
549 * If we haven't cached the size of the visibility map fork yet, check it
550 * first.
551 */
552 if (reln->smgr_cached_nblocks[VISIBILITYMAP_FORKNUM] == InvalidBlockNumber)
553 {
556 else
557 reln->smgr_cached_nblocks[VISIBILITYMAP_FORKNUM] = 0;
558 }
559
560 /*
561 * For reading we use ZERO_ON_ERROR mode, and initialize the page if
562 * necessary. It's always safe to clear bits, so it's better to clear
563 * corrupt pages than error out.
564 *
565 * We use the same path below to initialize pages when extending the
566 * relation, as a concurrent extension can end up with vm_extend()
567 * returning an already-initialized page.
568 */
569 if (blkno >= reln->smgr_cached_nblocks[VISIBILITYMAP_FORKNUM])
570 {
571 if (extend)
572 buf = vm_extend(rel, blkno + 1);
573 else
574 return InvalidBuffer;
575 }
576 else
579
580 /*
581 * Initializing the page when needed is trickier than it looks, because of
582 * the possibility of multiple backends doing this concurrently, and our
583 * desire to not uselessly take the buffer lock in the normal path where
584 * the page is OK. We must take the lock to initialize the page, so
585 * recheck page newness after we have the lock, in case someone else
586 * already did it. Also, because we initially check PageIsNew with no
587 * lock, it's possible to fall through and return the buffer while someone
588 * else is still initializing the page (i.e., we might see pd_upper as set
589 * but other page header fields are still zeroes). This is harmless for
590 * callers that will take a buffer lock themselves, but some callers
591 * inspect the page without any lock at all. The latter is OK only so
592 * long as it doesn't depend on the page header having correct contents.
593 * Current usage is safe because PageGetContents() does not require that.
594 */
596 {
601 }
602 return buf;
603}
604
605/*
606 * Ensure that the visibility map fork is at least vm_nblocks long, extending
607 * it if necessary with zeroed pages.
608 */
609static Buffer
611{
612 Buffer buf;
613
619
620 /*
621 * Send a shared-inval message to force other backends to close any smgr
622 * references they may have for this rel, which we are about to change.
623 * This is a useful optimization because it means that backends don't have
624 * to keep checking for creation or extension of the file, which happens
625 * infrequently.
626 */
627 CacheInvalidateSmgr(RelationGetSmgr(rel)->smgr_rlocator);
628
629 return buf;
630}
uint32 BlockNumber
Definition block.h:31
#define InvalidBlockNumber
Definition block.h:33
int Buffer
Definition buf.h:23
#define InvalidBuffer
Definition buf.h:25
BlockNumber BufferGetBlockNumber(Buffer buffer)
Definition bufmgr.c:4446
bool BufferIsLockedByMeInMode(Buffer buffer, BufferLockMode mode)
Definition bufmgr.c:3087
Buffer ExtendBufferedRelTo(BufferManagerRelation bmr, ForkNumber fork, BufferAccessStrategy strategy, uint32 flags, BlockNumber extend_to, ReadBufferMode mode)
Definition bufmgr.c:1040
void ReleaseBuffer(Buffer buffer)
Definition bufmgr.c:5586
void UnlockReleaseBuffer(Buffer buffer)
Definition bufmgr.c:5603
void MarkBufferDirty(Buffer buffer)
Definition bufmgr.c:3147
Buffer ReadBufferExtended(Relation reln, ForkNumber forkNum, BlockNumber blockNum, ReadBufferMode mode, BufferAccessStrategy strategy)
Definition bufmgr.c:926
static Page BufferGetPage(Buffer buffer)
Definition bufmgr.h:468
@ BUFFER_LOCK_EXCLUSIVE
Definition bufmgr.h:222
@ BUFFER_LOCK_UNLOCK
Definition bufmgr.h:207
static void LockBuffer(Buffer buffer, BufferLockMode mode)
Definition bufmgr.h:334
@ EB_CLEAR_SIZE_CACHE
Definition bufmgr.h:90
@ EB_CREATE_FORK_IF_NEEDED
Definition bufmgr.h:84
@ RBM_ZERO_ON_ERROR
Definition bufmgr.h:51
#define BMR_REL(p_rel)
Definition bufmgr.h:114
static bool BufferIsValid(Buffer bufnum)
Definition bufmgr.h:419
void PageInit(Page page, Size pageSize, Size specialSize)
Definition bufpage.c:42
static bool PageIsNew(const PageData *page)
Definition bufpage.h:258
static char * PageGetContents(Page page)
Definition bufpage.h:282
PageData * Page
Definition bufpage.h:81
uint8_t uint8
Definition c.h:622
#define Assert(condition)
Definition c.h:943
uint64_t uint64
Definition c.h:625
uint32_t uint32
Definition c.h:624
#define MemSet(start, val, len)
Definition c.h:1107
uint32 result
#define DEBUG1
Definition elog.h:31
#define ERROR
Definition elog.h:40
#define elog(elevel,...)
Definition elog.h:228
ProcNumber MyProcNumber
Definition globals.c:92
volatile uint32 CritSectionCount
Definition globals.c:45
const char * str
void CacheInvalidateSmgr(RelFileLocatorBackend rlocator)
Definition inval.c:1755
#define START_CRIT_SECTION()
Definition miscadmin.h:152
#define END_CRIT_SECTION()
Definition miscadmin.h:154
static uint64 pg_popcount_masked(const char *buf, int bytes, uint8 mask)
static char buf[DEFAULT_XLOG_SEG_SIZE]
static int fb(int x)
static SMgrRelation RelationGetSmgr(Relation rel)
Definition rel.h:578
#define RelationGetRelationName(relation)
Definition rel.h:550
#define RelationNeedsWAL(relation)
Definition rel.h:639
@ VISIBILITYMAP_FORKNUM
Definition relpath.h:60
@ MAIN_FORKNUM
Definition relpath.h:58
#define relpathbackend(rlocator, backend, forknum)
Definition relpath.h:141
BlockNumber smgrnblocks(SMgrRelation reln, ForkNumber forknum)
Definition smgr.c:819
bool smgrexists(SMgrRelation reln, ForkNumber forknum)
Definition smgr.c:462
#define MAPSIZE
BlockNumber visibilitymap_truncation_length(BlockNumber nheapblocks)
void visibilitymap_set(BlockNumber heapBlk, Buffer vmBuf, uint8 flags, const RelFileLocator rlocator)
#define HEAPBLK_TO_MAPBLOCK_LIMIT(x)
#define FROZEN_MASK8
bool visibilitymap_pin_ok(BlockNumber heapBlk, Buffer vmbuf)
#define VISIBLE_MASK8
#define HEAPBLK_TO_OFFSET(x)
void visibilitymap_pin(Relation rel, BlockNumber heapBlk, Buffer *vmbuf)
uint8 visibilitymap_get_status(Relation rel, BlockNumber heapBlk, Buffer *vmbuf)
static Buffer vm_extend(Relation rel, BlockNumber vm_nblocks)
BlockNumber visibilitymap_prepare_truncate(Relation rel, BlockNumber nheapblocks)
void visibilitymap_count(Relation rel, BlockNumber *all_visible, BlockNumber *all_frozen)
static Buffer vm_readbuf(Relation rel, BlockNumber blkno, bool extend)
#define HEAPBLK_TO_MAPBLOCK(x)
#define HEAPBLK_TO_MAPBYTE(x)
#define VISIBILITYMAP_VALID_BITS
#define VISIBILITYMAP_ALL_FROZEN
#define VISIBILITYMAP_ALL_VISIBLE
#define XLogHintBitIsNeeded()
Definition xlog.h:123
XLogRecPtr log_newpage_buffer(Buffer buffer, bool page_std)
bool InRecovery
Definition xlogutils.c:50

◆ HEAPBLK_TO_MAPBYTE

#define HEAPBLK_TO_MAPBYTE (   x)    (((x) % HEAPBLOCKS_PER_PAGE) / HEAPBLOCKS_PER_BYTE)

Definition at line 131 of file visibilitymap.c.

◆ HEAPBLK_TO_OFFSET

#define HEAPBLK_TO_OFFSET (   x)    (((x) % HEAPBLOCKS_PER_BYTE) * BITS_PER_HEAPBLOCK)

Definition at line 132 of file visibilitymap.c.

◆ HEAPBLOCKS_PER_BYTE

#define HEAPBLOCKS_PER_BYTE   (BITS_PER_BYTE / BITS_PER_HEAPBLOCK)

Definition at line 122 of file visibilitymap.c.

◆ HEAPBLOCKS_PER_PAGE

#define HEAPBLOCKS_PER_PAGE   (MAPSIZE * HEAPBLOCKS_PER_BYTE)

Definition at line 125 of file visibilitymap.c.

◆ MAPSIZE

Definition at line 119 of file visibilitymap.c.

◆ VISIBLE_MASK8

#define VISIBLE_MASK8   (0x55) /* The lower bit of each bit pair */

Definition at line 135 of file visibilitymap.c.

Function Documentation

◆ visibilitymap_clear()

bool visibilitymap_clear ( Relation  rel,
BlockNumber  heapBlk,
Buffer  vmbuf,
uint8  flags 
)

Definition at line 151 of file visibilitymap.c.

152{
154 int mapByte = HEAPBLK_TO_MAPBYTE(heapBlk);
155 int mapOffset = HEAPBLK_TO_OFFSET(heapBlk);
156 uint8 mask = flags << mapOffset;
157 char *map;
158 bool cleared = false;
159
160 /* Must never clear all_visible bit while leaving all_frozen bit set */
163
164#ifdef TRACE_VISIBILITYMAP
165 elog(DEBUG1, "vm_clear %s %d", RelationGetRelationName(rel), heapBlk);
166#endif
167
169 elog(ERROR, "wrong buffer passed to visibilitymap_clear");
170
173
174 if (map[mapByte] & mask)
175 {
176 map[mapByte] &= ~mask;
177
179 cleared = true;
180 }
181
183
184 return cleared;
185}

References Assert, BUFFER_LOCK_EXCLUSIVE, BUFFER_LOCK_UNLOCK, BufferGetBlockNumber(), BufferGetPage(), BufferIsValid(), DEBUG1, elog, ERROR, fb(), HEAPBLK_TO_MAPBLOCK, HEAPBLK_TO_MAPBYTE, HEAPBLK_TO_OFFSET, LockBuffer(), MarkBufferDirty(), PageGetContents(), RelationGetRelationName, VISIBILITYMAP_ALL_VISIBLE, and VISIBILITYMAP_VALID_BITS.

Referenced by heap_delete(), heap_force_common(), heap_insert(), heap_lock_tuple(), heap_lock_updated_tuple_rec(), heap_multi_insert(), heap_page_fix_vm_corruption(), heap_update(), heap_xlog_delete(), heap_xlog_insert(), heap_xlog_lock(), heap_xlog_lock_updated(), heap_xlog_multi_insert(), and heap_xlog_update().

◆ visibilitymap_count()

void visibilitymap_count ( Relation  rel,
BlockNumber all_visible,
BlockNumber all_frozen 
)

Definition at line 367 of file visibilitymap.c.

368{
371 BlockNumber nfrozen = 0;
372
373 /* all_visible must be specified */
374 Assert(all_visible);
375
376 for (mapBlock = 0;; mapBlock++)
377 {
379 uint64 *map;
380
381 /*
382 * Read till we fall off the end of the map. We assume that any extra
383 * bytes in the last page are zeroed, so we don't bother excluding
384 * them from the count.
385 */
386 mapBuffer = vm_readbuf(rel, mapBlock, false);
388 break;
389
390 /*
391 * We choose not to lock the page, since the result is going to be
392 * immediately stale anyway if anyone is concurrently setting or
393 * clearing bits, and we only really need an approximate value.
394 */
396
397 nvisible += pg_popcount_masked((const char *) map, MAPSIZE, VISIBLE_MASK8);
398 if (all_frozen)
399 nfrozen += pg_popcount_masked((const char *) map, MAPSIZE, FROZEN_MASK8);
400
402 }
403
404 *all_visible = nvisible;
405 if (all_frozen)
406 *all_frozen = nfrozen;
407}

References Assert, BufferGetPage(), BufferIsValid(), fb(), FROZEN_MASK8, MAPSIZE, PageGetContents(), pg_popcount_masked(), ReleaseBuffer(), VISIBLE_MASK8, and vm_readbuf().

Referenced by do_analyze_rel(), heap_vacuum_eager_scan_setup(), heap_vacuum_rel(), index_update_stats(), and pg_visibility_map_summary().

◆ visibilitymap_get_status()

uint8 visibilitymap_get_status ( Relation  rel,
BlockNumber  heapBlk,
Buffer vmbuf 
)

Definition at line 319 of file visibilitymap.c.

320{
324 char *map;
326
327#ifdef TRACE_VISIBILITYMAP
328 elog(DEBUG1, "vm_get_status %s %d", RelationGetRelationName(rel), heapBlk);
329#endif
330
331 /* Reuse the old pinned buffer if possible */
332 if (BufferIsValid(*vmbuf))
333 {
335 {
338 }
339 }
340
341 if (!BufferIsValid(*vmbuf))
342 {
343 *vmbuf = vm_readbuf(rel, mapBlock, false);
344 if (!BufferIsValid(*vmbuf))
345 return (uint8) 0;
346 }
347
349
350 /*
351 * A single byte read is atomic. There could be memory-ordering effects
352 * here, but for performance reasons we make it the caller's job to worry
353 * about that.
354 */
356 return result;
357}

References BufferGetBlockNumber(), BufferGetPage(), BufferIsValid(), DEBUG1, elog, fb(), HEAPBLK_TO_MAPBLOCK, HEAPBLK_TO_MAPBYTE, HEAPBLK_TO_OFFSET, InvalidBuffer, PageGetContents(), RelationGetRelationName, ReleaseBuffer(), result, VISIBILITYMAP_VALID_BITS, and vm_readbuf().

Referenced by collect_visibility_data(), find_next_unskippable_block(), heapcheck_read_stream_next_unskippable(), pg_visibility(), pg_visibility_map(), and prune_freeze_setup().

◆ visibilitymap_pin()

◆ visibilitymap_pin_ok()

bool visibilitymap_pin_ok ( BlockNumber  heapBlk,
Buffer  vmbuf 
)

◆ visibilitymap_prepare_truncate()

BlockNumber visibilitymap_prepare_truncate ( Relation  rel,
BlockNumber  nheapblocks 
)

Definition at line 421 of file visibilitymap.c.

422{
424
425 /* last remaining block, byte, and bit */
429
430#ifdef TRACE_VISIBILITYMAP
431 elog(DEBUG1, "vm_truncate %s %d", RelationGetRelationName(rel), nheapblocks);
432#endif
433
434 /*
435 * If no visibility map has been created yet for this relation, there's
436 * nothing to truncate.
437 */
439 return InvalidBlockNumber;
440
441 /*
442 * Unless the new size is exactly at a visibility map page boundary, the
443 * tail bits in the last remaining map page, representing truncated heap
444 * blocks, need to be cleared. This is not only tidy, but also necessary
445 * because we don't get a chance to clear the bits if the heap is extended
446 * again.
447 */
448 if (truncByte != 0 || truncOffset != 0)
449 {
451 Page page;
452 char *map;
453
455
456 mapBuffer = vm_readbuf(rel, truncBlock, false);
458 {
459 /* nothing to do, the file was already smaller */
460 return InvalidBlockNumber;
461 }
462
463 page = BufferGetPage(mapBuffer);
464 map = PageGetContents(page);
465
467
468 /* NO EREPORT(ERROR) from here till changes are logged */
470
471 /* Clear out the unwanted bytes. */
472 MemSet(&map[truncByte + 1], 0, MAPSIZE - (truncByte + 1));
473
474 /*----
475 * Mask out the unwanted bits of the last remaining byte.
476 *
477 * ((1 << 0) - 1) = 00000000
478 * ((1 << 1) - 1) = 00000001
479 * ...
480 * ((1 << 6) - 1) = 00111111
481 * ((1 << 7) - 1) = 01111111
482 *----
483 */
484 map[truncByte] &= (1 << truncOffset) - 1;
485
486 /*
487 * Truncation of a relation is WAL-logged at a higher-level, and we
488 * will be called at WAL replay. But if checksums are enabled, we need
489 * to still write a WAL record to protect against a torn page, if the
490 * page is flushed to disk before the truncation WAL record. We cannot
491 * use MarkBufferDirtyHint here, because that will not dirty the page
492 * during recovery.
493 */
497
499
501 }
502 else
504
506 {
507 /* nothing to do, the file was already smaller than requested size */
508 return InvalidBlockNumber;
509 }
510
511 return newnblocks;
512}

References BUFFER_LOCK_EXCLUSIVE, BufferGetPage(), BufferIsValid(), DEBUG1, elog, END_CRIT_SECTION, fb(), HEAPBLK_TO_MAPBLOCK, HEAPBLK_TO_MAPBYTE, HEAPBLK_TO_OFFSET, InRecovery, InvalidBlockNumber, LockBuffer(), log_newpage_buffer(), MAPSIZE, MarkBufferDirty(), MemSet, PageGetContents(), RelationGetRelationName, RelationGetSmgr(), RelationNeedsWAL, smgrexists(), smgrnblocks(), START_CRIT_SECTION, UnlockReleaseBuffer(), VISIBILITYMAP_FORKNUM, vm_readbuf(), and XLogHintBitIsNeeded.

Referenced by pg_truncate_visibility_map(), RelationTruncate(), and smgr_redo().

◆ visibilitymap_set()

void visibilitymap_set ( BlockNumber  heapBlk,
Buffer  vmBuf,
uint8  flags,
const RelFileLocator  rlocator 
)

Definition at line 255 of file visibilitymap.c.

258{
262 Page page;
263 uint8 *map;
264 uint8 status;
265
266#ifdef TRACE_VISIBILITYMAP
267 elog(DEBUG1, "vm_set flags 0x%02X for %s %d",
268 flags,
270 heapBlk);
271#endif
272
273 /* Call in same critical section where WAL is emitted. */
275
276 /* Flags should be valid. Also never clear bits with this function */
277 Assert((flags & VISIBILITYMAP_VALID_BITS) == flags);
278
279 /* Must never set all_frozen bit without also setting all_visible bit */
281
282 /* Check that we have the right VM page pinned */
284 elog(ERROR, "wrong VM buffer passed to visibilitymap_set");
285
287
288 page = BufferGetPage(vmBuf);
289 map = (uint8 *) PageGetContents(page);
290
291 status = (map[mapByte] >> mapOffset) & VISIBILITYMAP_VALID_BITS;
292 if (flags != status)
293 {
294 map[mapByte] |= (flags << mapOffset);
296 }
297}

References Assert, BUFFER_LOCK_EXCLUSIVE, BufferGetBlockNumber(), BufferGetPage(), BufferIsLockedByMeInMode(), BufferIsValid(), CritSectionCount, DEBUG1, elog, ERROR, fb(), HEAPBLK_TO_MAPBLOCK, HEAPBLK_TO_MAPBYTE, HEAPBLK_TO_OFFSET, InRecovery, MAIN_FORKNUM, MarkBufferDirty(), MyProcNumber, PageGetContents(), relpathbackend, str, VISIBILITYMAP_ALL_FROZEN, and VISIBILITYMAP_VALID_BITS.

Referenced by heap_multi_insert(), heap_page_prune_and_freeze(), heap_xlog_multi_insert(), heap_xlog_prune_freeze(), lazy_scan_new_or_empty(), and lazy_vacuum_heap_page().

◆ visibilitymap_truncation_length()

BlockNumber visibilitymap_truncation_length ( BlockNumber  nheapblocks)

Definition at line 524 of file visibilitymap.c.

525{
527}

References fb(), and HEAPBLK_TO_MAPBLOCK_LIMIT.

Referenced by SummarizeSmgrRecord().

◆ vm_extend()

static Buffer vm_extend ( Relation  rel,
BlockNumber  vm_nblocks 
)
static

Definition at line 610 of file visibilitymap.c.

611{
612 Buffer buf;
613
619
620 /*
621 * Send a shared-inval message to force other backends to close any smgr
622 * references they may have for this rel, which we are about to change.
623 * This is a useful optimization because it means that backends don't have
624 * to keep checking for creation or extension of the file, which happens
625 * infrequently.
626 */
627 CacheInvalidateSmgr(RelationGetSmgr(rel)->smgr_rlocator);
628
629 return buf;
630}

References BMR_REL, buf, CacheInvalidateSmgr(), EB_CLEAR_SIZE_CACHE, EB_CREATE_FORK_IF_NEEDED, ExtendBufferedRelTo(), fb(), RBM_ZERO_ON_ERROR, RelationGetSmgr(), and VISIBILITYMAP_FORKNUM.

Referenced by vm_readbuf().

◆ vm_readbuf()

static Buffer vm_readbuf ( Relation  rel,
BlockNumber  blkno,
bool  extend 
)
static

Definition at line 536 of file visibilitymap.c.

537{
538 Buffer buf;
540
541 /*
542 * Caution: re-using this smgr pointer could fail if the relcache entry
543 * gets closed. It's safe as long as we only do smgr-level operations
544 * between here and the last use of the pointer.
545 */
546 reln = RelationGetSmgr(rel);
547
548 /*
549 * If we haven't cached the size of the visibility map fork yet, check it
550 * first.
551 */
552 if (reln->smgr_cached_nblocks[VISIBILITYMAP_FORKNUM] == InvalidBlockNumber)
553 {
556 else
557 reln->smgr_cached_nblocks[VISIBILITYMAP_FORKNUM] = 0;
558 }
559
560 /*
561 * For reading we use ZERO_ON_ERROR mode, and initialize the page if
562 * necessary. It's always safe to clear bits, so it's better to clear
563 * corrupt pages than error out.
564 *
565 * We use the same path below to initialize pages when extending the
566 * relation, as a concurrent extension can end up with vm_extend()
567 * returning an already-initialized page.
568 */
569 if (blkno >= reln->smgr_cached_nblocks[VISIBILITYMAP_FORKNUM])
570 {
571 if (extend)
572 buf = vm_extend(rel, blkno + 1);
573 else
574 return InvalidBuffer;
575 }
576 else
579
580 /*
581 * Initializing the page when needed is trickier than it looks, because of
582 * the possibility of multiple backends doing this concurrently, and our
583 * desire to not uselessly take the buffer lock in the normal path where
584 * the page is OK. We must take the lock to initialize the page, so
585 * recheck page newness after we have the lock, in case someone else
586 * already did it. Also, because we initially check PageIsNew with no
587 * lock, it's possible to fall through and return the buffer while someone
588 * else is still initializing the page (i.e., we might see pd_upper as set
589 * but other page header fields are still zeroes). This is harmless for
590 * callers that will take a buffer lock themselves, but some callers
591 * inspect the page without any lock at all. The latter is OK only so
592 * long as it doesn't depend on the page header having correct contents.
593 * Current usage is safe because PageGetContents() does not require that.
594 */
596 {
601 }
602 return buf;
603}

References buf, BUFFER_LOCK_EXCLUSIVE, BUFFER_LOCK_UNLOCK, BufferGetPage(), fb(), InvalidBlockNumber, InvalidBuffer, LockBuffer(), PageInit(), PageIsNew(), RBM_ZERO_ON_ERROR, ReadBufferExtended(), RelationGetSmgr(), smgrexists(), smgrnblocks(), VISIBILITYMAP_FORKNUM, and vm_extend().

Referenced by visibilitymap_count(), visibilitymap_get_status(), visibilitymap_pin(), and visibilitymap_prepare_truncate().