summaryrefslogtreecommitdiff
path: root/src/backend/storage/buffer/bufmgr.c
diff options
context:
space:
mode:
authorAndres Freund2025-12-19 18:23:13 +0000
committerAndres Freund2025-12-19 18:23:33 +0000
commit09ae2c8bac8db409a8cd0b8ee438ea7526deb4c3 (patch)
treefe469a798c1cf78e6f25799db5d860881680823e /src/backend/storage/buffer/bufmgr.c
parent80f08a6e6a6c11ac3530c473cfb963f646d0891a (diff)
bufmgr: Optimize & harmonize LockBufHdr(), LWLockWaitListLock()
The main optimization is for LockBufHdr() to delay initializing SpinDelayStatus, similar to what LWLockWaitListLock already did. The initialization is sufficiently expensive & buffer header lock acquisitions are sufficiently frequent, to make it worthwhile to instead have a fastpath (via a likely() branch) that does not initialize the SpinDelayStatus. While LWLockWaitListLock() already the aforementioned optimization, it did not use likely(), and inspection of the assembly shows that this indeed leads to worse code generation (also observed in a microbenchmark). Fix that by adding the likely(). While the LockBufHdr() improvement is a small gain on its own, it mainly is aimed at preventing a regression after a future commit, which requires additional locking to set hint bits. While touching both, also make the comments more similar to each other. Reviewed-by: Heikki Linnakangas <[email protected]> Discussion: https://postgr.es/m/fvfmkr5kk4nyex56ejgxj3uzi63isfxovp2biecb4bspbjrze7@az2pljabhnff
Diffstat (limited to 'src/backend/storage/buffer/bufmgr.c')
-rw-r--r--src/backend/storage/buffer/bufmgr.c36
1 files changed, 27 insertions, 9 deletions
diff --git a/src/backend/storage/buffer/bufmgr.c b/src/backend/storage/buffer/bufmgr.c
index a768fb129ae..eb55102b0d7 100644
--- a/src/backend/storage/buffer/bufmgr.c
+++ b/src/backend/storage/buffer/bufmgr.c
@@ -6358,23 +6358,41 @@ rlocator_comparator(const void *p1, const void *p2)
uint32
LockBufHdr(BufferDesc *desc)
{
- SpinDelayStatus delayStatus;
uint32 old_buf_state;
Assert(!BufferIsLocal(BufferDescriptorGetBuffer(desc)));
- init_local_spin_delay(&delayStatus);
-
while (true)
{
- /* set BM_LOCKED flag */
+ /*
+ * Always try once to acquire the lock directly, without setting up
+ * the spin-delay infrastructure. The work necessary for that shows up
+ * in profiles and is rarely necessary.
+ */
old_buf_state = pg_atomic_fetch_or_u32(&desc->state, BM_LOCKED);
- /* if it wasn't set before we're OK */
- if (!(old_buf_state & BM_LOCKED))
- break;
- perform_spin_delay(&delayStatus);
+ if (likely(!(old_buf_state & BM_LOCKED)))
+ break; /* got lock */
+
+ /* and then spin without atomic operations until lock is released */
+ {
+ SpinDelayStatus delayStatus;
+
+ init_local_spin_delay(&delayStatus);
+
+ while (old_buf_state & BM_LOCKED)
+ {
+ perform_spin_delay(&delayStatus);
+ old_buf_state = pg_atomic_read_u32(&desc->state);
+ }
+ finish_spin_delay(&delayStatus);
+ }
+
+ /*
+ * Retry. The lock might obviously already be re-acquired by the time
+ * we're attempting to get it again.
+ */
}
- finish_spin_delay(&delayStatus);
+
return old_buf_state | BM_LOCKED;
}