Convert extents_t's npages field to use C11-style atomics
In the process, we can do some strength reduction, changing the fetch-adds and fetch-subs to be simple loads followed by stores, since the modifications all occur while holding the mutex.
This commit is contained in:
parent
dafadce622
commit
8adab26972
@ -109,9 +109,12 @@ struct extents_s {
|
||||
/*
|
||||
* Page sum for all extents in heaps.
|
||||
*
|
||||
* Synchronization: atomic.
|
||||
* The synchronization here is a little tricky. Modifications to npages
|
||||
* must hold mtx, but reads need not (though, a reader who sees npages
|
||||
* without holding the mutex can't assume anything about the rest of the
|
||||
* state of the extents_t).
|
||||
*/
|
||||
size_t npages;
|
||||
atomic_zu_t npages;
|
||||
|
||||
/* All stored extents must be in the same state. */
|
||||
extent_state_t state;
|
||||
|
29
src/extent.c
29
src/extent.c
@ -186,7 +186,7 @@ extents_init(tsdn_t *tsdn, extents_t *extents, extent_state_t state,
|
||||
extent_heap_new(&extents->heaps[i]);
|
||||
}
|
||||
extent_list_init(&extents->lru);
|
||||
extents->npages = 0;
|
||||
atomic_store_zu(&extents->npages, 0, ATOMIC_RELAXED);
|
||||
extents->state = state;
|
||||
extents->delay_coalesce = delay_coalesce;
|
||||
return false;
|
||||
@ -199,7 +199,7 @@ extents_state_get(const extents_t *extents) {
|
||||
|
||||
size_t
|
||||
extents_npages_get(extents_t *extents) {
|
||||
return atomic_read_zu(&extents->npages);
|
||||
return atomic_load_zu(&extents->npages, ATOMIC_RELAXED);
|
||||
}
|
||||
|
||||
static void
|
||||
@ -216,7 +216,15 @@ extents_insert_locked(tsdn_t *tsdn, extents_t *extents, extent_t *extent,
|
||||
extent_list_append(&extents->lru, extent);
|
||||
}
|
||||
size_t npages = size >> LG_PAGE;
|
||||
atomic_add_zu(&extents->npages, npages);
|
||||
/*
|
||||
* All modifications to npages hold the mutex (as asserted above), so we
|
||||
* don't need an atomic fetch-add; we can get by with a load followed by
|
||||
* a store.
|
||||
*/
|
||||
size_t cur_extents_npages =
|
||||
atomic_load_zu(&extents->npages, ATOMIC_RELAXED);
|
||||
atomic_store_zu(&extents->npages, cur_extents_npages + npages,
|
||||
ATOMIC_RELAXED);
|
||||
}
|
||||
|
||||
static void
|
||||
@ -233,8 +241,15 @@ extents_remove_locked(tsdn_t *tsdn, extents_t *extents, extent_t *extent,
|
||||
extent_list_remove(&extents->lru, extent);
|
||||
}
|
||||
size_t npages = size >> LG_PAGE;
|
||||
assert(atomic_read_zu(&extents->npages) >= npages);
|
||||
atomic_sub_zu(&extents->npages, size >> LG_PAGE);
|
||||
/*
|
||||
* As in extents_insert_locked, we hold extents->mtx and so don't need
|
||||
* atomic operations for updating extents->npages.
|
||||
*/
|
||||
size_t cur_extents_npages =
|
||||
atomic_load_zu(&extents->npages, ATOMIC_RELAXED);
|
||||
assert(cur_extents_npages >= npages);
|
||||
atomic_store_zu(&extents->npages,
|
||||
cur_extents_npages - (size >> LG_PAGE), ATOMIC_RELAXED);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -299,7 +314,9 @@ extents_evict(tsdn_t *tsdn, arena_t *arena, extent_hooks_t **r_extent_hooks,
|
||||
}
|
||||
/* Check the eviction limit. */
|
||||
size_t npages = extent_size_get(extent) >> LG_PAGE;
|
||||
if (atomic_read_zu(&extents->npages) - npages < npages_min) {
|
||||
size_t extents_npages = atomic_load_zu(&extents->npages,
|
||||
ATOMIC_RELAXED);
|
||||
if (extents_npages - npages < npages_min) {
|
||||
extent = NULL;
|
||||
goto label_return;
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user