Eset: take opt_lg_max_active_fit as a parameter.

This breaks its dependence on the global.
This commit is contained in:
David Goldblatt 2020-03-14 09:46:09 -07:00 committed by David Goldblatt
parent 883ab327cc
commit 7bb6e2dc0d
3 changed files with 31 additions and 14 deletions

View File

@ -52,6 +52,6 @@ void eset_remove(eset_t *eset, edata_t *edata);
* null if no such item could be found. * null if no such item could be found.
*/ */
edata_t *eset_fit(eset_t *eset, size_t esize, size_t alignment, edata_t *eset_fit(eset_t *eset, size_t esize, size_t alignment,
bool delay_coalesce); unsigned lg_max_fit);
#endif /* JEMALLOC_INTERNAL_ESET_H */ #endif /* JEMALLOC_INTERNAL_ESET_H */

View File

@ -154,9 +154,15 @@ eset_fit_alignment(eset_t *eset, size_t min_size, size_t max_size,
/* /*
* Do first-fit extent selection, i.e. select the oldest/lowest extent that is * Do first-fit extent selection, i.e. select the oldest/lowest extent that is
* large enough. * large enough.
*
* lg_max_fit is the (log of the) maximum ratio between the requested size and
* the returned size that we'll allow. This can reduce fragmentation by
* avoiding reusing and splitting large extents for smaller sizes. In practice,
* it's set to opt_lg_extent_max_active_fit for the dirty eset and SC_PTR_BITS
* for others.
*/ */
static edata_t * static edata_t *
eset_first_fit(eset_t *eset, size_t size, bool delay_coalesce) { eset_first_fit(eset_t *eset, size_t size, unsigned lg_max_fit) {
edata_t *ret = NULL; edata_t *ret = NULL;
pszind_t pind = sz_psz2ind(sz_psz_quantize_ceil(size)); pszind_t pind = sz_psz2ind(sz_psz_quantize_ceil(size));
@ -178,14 +184,15 @@ eset_first_fit(eset_t *eset, size_t size, bool delay_coalesce) {
assert(!edata_heap_empty(&eset->heaps[i])); assert(!edata_heap_empty(&eset->heaps[i]));
edata_t *edata = edata_heap_first(&eset->heaps[i]); edata_t *edata = edata_heap_first(&eset->heaps[i]);
assert(edata_size_get(edata) >= size); assert(edata_size_get(edata) >= size);
if (lg_max_fit == SC_PTR_BITS) {
/* /*
* In order to reduce fragmentation, avoid reusing and splitting * We'll shift by this below, and shifting out all the
* large eset for much smaller sizes. * bits is undefined. Decreasing is safe, since the
* * page size is larger than 1 byte.
* Only do check for dirty eset (delay_coalesce).
*/ */
if (delay_coalesce && lg_max_fit = SC_PTR_BITS - 1;
(sz_pind2sz(i) >> opt_lg_extent_max_active_fit) > size) { }
if ((sz_pind2sz(i) >> lg_max_fit) > size) {
break; break;
} }
if (ret == NULL || edata_snad_comp(edata, ret) < 0) { if (ret == NULL || edata_snad_comp(edata, ret) < 0) {
@ -201,14 +208,14 @@ eset_first_fit(eset_t *eset, size_t size, bool delay_coalesce) {
} }
edata_t * edata_t *
eset_fit(eset_t *eset, size_t esize, size_t alignment, bool delay_coalesce) { eset_fit(eset_t *eset, size_t esize, size_t alignment, unsigned lg_max_fit) {
size_t max_size = esize + PAGE_CEILING(alignment) - PAGE; size_t max_size = esize + PAGE_CEILING(alignment) - PAGE;
/* Beware size_t wrap-around. */ /* Beware size_t wrap-around. */
if (max_size < esize) { if (max_size < esize) {
return NULL; return NULL;
} }
edata_t *edata = eset_first_fit(eset, max_size, delay_coalesce); edata_t *edata = eset_first_fit(eset, max_size, lg_max_fit);
if (alignment > PAGE && edata == NULL) { if (alignment > PAGE && edata == NULL) {
/* /*

View File

@ -398,8 +398,18 @@ extent_recycle_extract(tsdn_t *tsdn, pa_shard_t *shard, ehooks_t *ehooks,
emap_unlock_edata(tsdn, &emap_global, unlock_edata); emap_unlock_edata(tsdn, &emap_global, unlock_edata);
} }
} else { } else {
edata = eset_fit(&ecache->eset, size, alignment, /*
ecache->delay_coalesce); * A large extent might be broken up from its original size to
* some small size to satisfy a small request. When that small
* request is freed, though, it won't merge back with the larger
* extent if delayed coalescing is on. The large extent can
* then no longer satify a request for its original size. To
* limit this effect, when delayed coalescing is enabled, we
* put a cap on how big an extent we can split for a request.
*/
unsigned lg_max_fit = ecache->delay_coalesce
? (unsigned)opt_lg_extent_max_active_fit : SC_PTR_BITS;
edata = eset_fit(&ecache->eset, size, alignment, lg_max_fit);
} }
if (edata == NULL) { if (edata == NULL) {
malloc_mutex_unlock(tsdn, &ecache->mtx); malloc_mutex_unlock(tsdn, &ecache->mtx);