Eset: Pull per-pszind data into structs.

We currently have one for stats and one for the data.  The data struct is just a
wrapper around the edata_heap_t, but this will change shortly.
This commit is contained in:
David Goldblatt 2021-07-20 09:26:09 -07:00 committed by David Goldblatt
parent dc0a4b8b2f
commit 252e0942d0
2 changed files with 56 additions and 27 deletions

View File

@ -14,16 +14,28 @@
* there are mutating operations. One exception is the stats counters, which * there are mutating operations. One exception is the stats counters, which
* may be read without any locking. * may be read without any locking.
*/ */
typedef struct eset_bin_s eset_bin_t;
struct eset_bin_s {
edata_heap_t heap;
};
typedef struct eset_bin_stats_s eset_bin_stats_t;
struct eset_bin_stats_s {
atomic_zu_t nextents;
atomic_zu_t nbytes;
};
typedef struct eset_s eset_t; typedef struct eset_s eset_t;
struct eset_s { struct eset_s {
/* Quantized per size class heaps of extents. */
edata_heap_t heaps[SC_NPSIZES + 1];
atomic_zu_t nextents[SC_NPSIZES + 1];
atomic_zu_t nbytes[SC_NPSIZES + 1];
/* Bitmap for which set bits correspond to non-empty heaps. */ /* Bitmap for which set bits correspond to non-empty heaps. */
fb_group_t bitmap[FB_NGROUPS(SC_NPSIZES + 1)]; fb_group_t bitmap[FB_NGROUPS(SC_NPSIZES + 1)];
/* Quantized per size class heaps of extents. */
eset_bin_t bins[SC_NPSIZES + 1];
eset_bin_stats_t bin_stats[SC_NPSIZES + 1];
/* LRU of all extents in heaps. */ /* LRU of all extents in heaps. */
edata_list_inactive_t lru; edata_list_inactive_t lru;

View File

@ -5,14 +5,25 @@
#define ESET_NPSIZES (SC_NPSIZES + 1) #define ESET_NPSIZES (SC_NPSIZES + 1)
static void
eset_bin_init(eset_bin_t *bin) {
edata_heap_new(&bin->heap);
}
static void
eset_bin_stats_init(eset_bin_stats_t *bin_stats) {
atomic_store_zu(&bin_stats->nextents, 0, ATOMIC_RELAXED);
atomic_store_zu(&bin_stats->nbytes, 0, ATOMIC_RELAXED);
}
void void
eset_init(eset_t *eset, extent_state_t state) { eset_init(eset_t *eset, extent_state_t state) {
for (unsigned i = 0; i < ESET_NPSIZES; i++) { for (unsigned i = 0; i < ESET_NPSIZES; i++) {
edata_heap_new(&eset->heaps[i]); eset_bin_init(&eset->bins[i]);
eset_bin_stats_init(&eset->bin_stats[i]);
} }
fb_init(eset->bitmap, ESET_NPSIZES); fb_init(eset->bitmap, ESET_NPSIZES);
edata_list_inactive_init(&eset->lru); edata_list_inactive_init(&eset->lru);
atomic_store_zu(&eset->npages, 0, ATOMIC_RELAXED);
eset->state = state; eset->state = state;
} }
@ -23,28 +34,34 @@ eset_npages_get(eset_t *eset) {
size_t size_t
eset_nextents_get(eset_t *eset, pszind_t pind) { eset_nextents_get(eset_t *eset, pszind_t pind) {
return atomic_load_zu(&eset->nextents[pind], ATOMIC_RELAXED); return atomic_load_zu(&eset->bin_stats[pind].nextents, ATOMIC_RELAXED);
} }
size_t size_t
eset_nbytes_get(eset_t *eset, pszind_t pind) { eset_nbytes_get(eset_t *eset, pszind_t pind) {
return atomic_load_zu(&eset->nbytes[pind], ATOMIC_RELAXED); return atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
} }
static void static void
eset_stats_add(eset_t *eset, pszind_t pind, size_t sz) { eset_stats_add(eset_t *eset, pszind_t pind, size_t sz) {
size_t cur = atomic_load_zu(&eset->nextents[pind], ATOMIC_RELAXED); size_t cur = atomic_load_zu(&eset->bin_stats[pind].nextents,
atomic_store_zu(&eset->nextents[pind], cur + 1, ATOMIC_RELAXED); ATOMIC_RELAXED);
cur = atomic_load_zu(&eset->nbytes[pind], ATOMIC_RELAXED); atomic_store_zu(&eset->bin_stats[pind].nextents, cur + 1,
atomic_store_zu(&eset->nbytes[pind], cur + sz, ATOMIC_RELAXED); ATOMIC_RELAXED);
cur = atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
atomic_store_zu(&eset->bin_stats[pind].nbytes, cur + sz,
ATOMIC_RELAXED);
} }
static void static void
eset_stats_sub(eset_t *eset, pszind_t pind, size_t sz) { eset_stats_sub(eset_t *eset, pszind_t pind, size_t sz) {
size_t cur = atomic_load_zu(&eset->nextents[pind], ATOMIC_RELAXED); size_t cur = atomic_load_zu(&eset->bin_stats[pind].nextents,
atomic_store_zu(&eset->nextents[pind], cur - 1, ATOMIC_RELAXED); ATOMIC_RELAXED);
cur = atomic_load_zu(&eset->nbytes[pind], ATOMIC_RELAXED); atomic_store_zu(&eset->bin_stats[pind].nextents, cur - 1,
atomic_store_zu(&eset->nbytes[pind], cur - sz, ATOMIC_RELAXED); ATOMIC_RELAXED);
cur = atomic_load_zu(&eset->bin_stats[pind].nbytes, ATOMIC_RELAXED);
atomic_store_zu(&eset->bin_stats[pind].nbytes, cur - sz,
ATOMIC_RELAXED);
} }
void void
@ -54,10 +71,10 @@ eset_insert(eset_t *eset, edata_t *edata) {
size_t size = edata_size_get(edata); size_t size = edata_size_get(edata);
size_t psz = sz_psz_quantize_floor(size); size_t psz = sz_psz_quantize_floor(size);
pszind_t pind = sz_psz2ind(psz); pszind_t pind = sz_psz2ind(psz);
if (edata_heap_empty(&eset->heaps[pind])) { if (edata_heap_empty(&eset->bins[pind].heap)) {
fb_set(eset->bitmap, ESET_NPSIZES, (size_t)pind); fb_set(eset->bitmap, ESET_NPSIZES, (size_t)pind);
} }
edata_heap_insert(&eset->heaps[pind], edata); edata_heap_insert(&eset->bins[pind].heap, edata);
if (config_stats) { if (config_stats) {
eset_stats_add(eset, pind, size); eset_stats_add(eset, pind, size);
@ -84,13 +101,13 @@ eset_remove(eset_t *eset, edata_t *edata) {
size_t size = edata_size_get(edata); size_t size = edata_size_get(edata);
size_t psz = sz_psz_quantize_floor(size); size_t psz = sz_psz_quantize_floor(size);
pszind_t pind = sz_psz2ind(psz); pszind_t pind = sz_psz2ind(psz);
edata_heap_remove(&eset->heaps[pind], edata); edata_heap_remove(&eset->bins[pind].heap, edata);
if (config_stats) { if (config_stats) {
eset_stats_sub(eset, pind, size); eset_stats_sub(eset, pind, size);
} }
if (edata_heap_empty(&eset->heaps[pind])) { if (edata_heap_empty(&eset->bins[pind].heap)) {
fb_unset(eset->bitmap, ESET_NPSIZES, (size_t)pind); fb_unset(eset->bitmap, ESET_NPSIZES, (size_t)pind);
} }
edata_list_inactive_remove(&eset->lru, edata); edata_list_inactive_remove(&eset->lru, edata);
@ -125,8 +142,8 @@ eset_fit_alignment(eset_t *eset, size_t min_size, size_t max_size,
i < pind_max; i < pind_max;
i = (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)i + 1)) { i = (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)i + 1)) {
assert(i < SC_NPSIZES); assert(i < SC_NPSIZES);
assert(!edata_heap_empty(&eset->heaps[i])); assert(!edata_heap_empty(&eset->bins[i].heap));
edata_t *edata = edata_heap_first(&eset->heaps[i]); edata_t *edata = edata_heap_first(&eset->bins[i].heap);
uintptr_t base = (uintptr_t)edata_base_get(edata); uintptr_t base = (uintptr_t)edata_base_get(edata);
size_t candidate_size = edata_size_get(edata); size_t candidate_size = edata_size_get(edata);
assert(candidate_size >= min_size); assert(candidate_size >= min_size);
@ -165,16 +182,16 @@ eset_first_fit(eset_t *eset, size_t size, bool exact_only,
pszind_t pind = sz_psz2ind(sz_psz_quantize_ceil(size)); pszind_t pind = sz_psz2ind(sz_psz_quantize_ceil(size));
if (exact_only) { if (exact_only) {
return edata_heap_empty(&eset->heaps[pind]) ? NULL : return edata_heap_empty(&eset->bins[pind].heap) ? NULL :
edata_heap_first(&eset->heaps[pind]); edata_heap_first(&eset->bins[pind].heap);
} }
for (pszind_t i = for (pszind_t i =
(pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)pind); (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)pind);
i < ESET_NPSIZES; i < ESET_NPSIZES;
i = (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)i + 1)) { i = (pszind_t)fb_ffs(eset->bitmap, ESET_NPSIZES, (size_t)i + 1)) {
assert(!edata_heap_empty(&eset->heaps[i])); assert(!edata_heap_empty(&eset->bins[i].heap));
edata_t *edata = edata_heap_first(&eset->heaps[i]); edata_t *edata = edata_heap_first(&eset->bins[i].heap);
assert(edata_size_get(edata) >= size); assert(edata_size_get(edata) >= size);
if (lg_max_fit == SC_PTR_BITS) { if (lg_max_fit == SC_PTR_BITS) {
/* /*