2010-01-17 01:53:50 +08:00
|
|
|
#define JEMALLOC_ARENA_C_
|
2010-02-12 06:45:59 +08:00
|
|
|
#include "jemalloc/internal/jemalloc_internal.h"
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/******************************************************************************/
|
|
|
|
/* Data. */
|
|
|
|
|
|
|
|
ssize_t opt_lg_dirty_mult = LG_DIRTY_MULT_DEFAULT;
|
2015-03-19 09:55:33 +08:00
|
|
|
static ssize_t lg_dirty_mult_default;
|
2012-02-29 08:50:47 +08:00
|
|
|
arena_bin_info_t arena_bin_info[NBINS];
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
size_t map_bias;
|
|
|
|
size_t map_misc_offset;
|
|
|
|
size_t arena_maxrun; /* Max run size for arenas. */
|
2015-09-12 11:50:20 +08:00
|
|
|
size_t large_maxclass; /* Max large size class. */
|
2015-05-05 00:58:36 +08:00
|
|
|
static size_t small_maxrun; /* Max run size used for small size classes. */
|
|
|
|
static bool *small_run_tab; /* Valid small run page multiples. */
|
2014-10-13 13:53:59 +08:00
|
|
|
unsigned nlclasses; /* Number of large size classes. */
|
|
|
|
unsigned nhclasses; /* Number of huge size classes. */
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/******************************************************************************/
|
2014-01-15 08:23:03 +08:00
|
|
|
/*
|
|
|
|
* Function prototypes for static functions that are referenced prior to
|
|
|
|
* definition.
|
|
|
|
*/
|
|
|
|
|
2010-10-01 07:55:08 +08:00
|
|
|
static void arena_purge(arena_t *arena, bool all);
|
2012-10-31 06:42:37 +08:00
|
|
|
static void arena_run_dalloc(arena_t *arena, arena_run_t *run, bool dirty,
|
2015-08-05 01:49:46 +08:00
|
|
|
bool cleaned, bool decommitted);
|
2010-01-17 01:53:50 +08:00
|
|
|
static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk,
|
|
|
|
arena_run_t *run, arena_bin_t *bin);
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk,
|
|
|
|
arena_run_t *run, arena_bin_t *bin);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/******************************************************************************/
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
#define CHUNK_MAP_KEY ((uintptr_t)0x1U)
|
|
|
|
|
|
|
|
JEMALLOC_INLINE_C arena_chunk_map_misc_t *
|
|
|
|
arena_miscelm_key_create(size_t size)
|
|
|
|
{
|
|
|
|
|
2015-08-20 05:12:05 +08:00
|
|
|
return ((arena_chunk_map_misc_t *)(arena_mapbits_size_encode(size) |
|
2015-08-05 01:49:46 +08:00
|
|
|
CHUNK_MAP_KEY));
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE_C bool
|
|
|
|
arena_miscelm_is_key(const arena_chunk_map_misc_t *miscelm)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (((uintptr_t)miscelm & CHUNK_MAP_KEY) != 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
#undef CHUNK_MAP_KEY
|
|
|
|
|
2014-04-07 04:24:16 +08:00
|
|
|
JEMALLOC_INLINE_C size_t
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_miscelm_key_size_get(const arena_chunk_map_misc_t *miscelm)
|
2014-04-07 04:24:16 +08:00
|
|
|
{
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
assert(arena_miscelm_is_key(miscelm));
|
|
|
|
|
2015-08-20 05:12:05 +08:00
|
|
|
return (arena_mapbits_size_decode((uintptr_t)miscelm));
|
2015-08-05 01:49:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE_C size_t
|
|
|
|
arena_miscelm_size_get(arena_chunk_map_misc_t *miscelm)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
size_t pageind, mapbits;
|
|
|
|
|
|
|
|
assert(!arena_miscelm_is_key(miscelm));
|
|
|
|
|
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(miscelm);
|
|
|
|
pageind = arena_miscelm_to_pageind(miscelm);
|
|
|
|
mapbits = arena_mapbits_get(chunk, pageind);
|
2015-08-20 05:12:05 +08:00
|
|
|
return (arena_mapbits_size_decode(mapbits));
|
2014-04-07 04:24:16 +08:00
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C int
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_run_comp(arena_chunk_map_misc_t *a, arena_chunk_map_misc_t *b)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-08-30 04:34:40 +08:00
|
|
|
uintptr_t a_miscelm = (uintptr_t)a;
|
|
|
|
uintptr_t b_miscelm = (uintptr_t)b;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
assert(a != NULL);
|
|
|
|
assert(b != NULL);
|
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
return ((a_miscelm > b_miscelm) - (a_miscelm < b_miscelm));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2010-03-01 07:00:18 +08:00
|
|
|
/* Generate red-black tree functions. */
|
2014-08-30 04:34:40 +08:00
|
|
|
rb_gen(static UNUSED, arena_run_tree_, arena_run_tree_t, arena_chunk_map_misc_t,
|
2014-08-15 05:45:58 +08:00
|
|
|
rb_link, arena_run_comp)
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
static size_t
|
|
|
|
run_quantize(size_t size)
|
|
|
|
{
|
|
|
|
size_t qsize;
|
|
|
|
|
|
|
|
assert(size != 0);
|
|
|
|
assert(size == PAGE_CEILING(size));
|
|
|
|
|
|
|
|
/* Don't change sizes that are valid small run sizes. */
|
|
|
|
if (size <= small_maxrun && small_run_tab[size >> LG_PAGE])
|
|
|
|
return (size);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Round down to the nearest run size that can actually be requested
|
|
|
|
* during normal large allocation. Add large_pad so that cache index
|
|
|
|
* randomization can offset the allocation from the page boundary.
|
|
|
|
*/
|
|
|
|
qsize = index2size(size2index(size - large_pad + 1) - 1) + large_pad;
|
|
|
|
if (qsize <= SMALL_MAXCLASS + large_pad)
|
|
|
|
return (run_quantize(size - large_pad));
|
|
|
|
assert(qsize <= size);
|
|
|
|
return (qsize);
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t
|
|
|
|
run_quantize_next(size_t size)
|
|
|
|
{
|
|
|
|
size_t large_run_size_next;
|
|
|
|
|
|
|
|
assert(size != 0);
|
|
|
|
assert(size == PAGE_CEILING(size));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Return the next quantized size greater than the input size.
|
|
|
|
* Quantized sizes comprise the union of run sizes that back small
|
|
|
|
* region runs, and run sizes that back large regions with no explicit
|
|
|
|
* alignment constraints.
|
|
|
|
*/
|
|
|
|
|
|
|
|
if (size > SMALL_MAXCLASS) {
|
|
|
|
large_run_size_next = PAGE_CEILING(index2size(size2index(size -
|
|
|
|
large_pad) + 1) + large_pad);
|
|
|
|
} else
|
|
|
|
large_run_size_next = SIZE_T_MAX;
|
|
|
|
if (size >= small_maxrun)
|
|
|
|
return (large_run_size_next);
|
|
|
|
|
|
|
|
while (true) {
|
|
|
|
size += PAGE;
|
|
|
|
assert(size <= small_maxrun);
|
|
|
|
if (small_run_tab[size >> LG_PAGE]) {
|
|
|
|
if (large_run_size_next < size)
|
|
|
|
return (large_run_size_next);
|
|
|
|
return (size);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t
|
|
|
|
run_quantize_first(size_t size)
|
|
|
|
{
|
|
|
|
size_t qsize = run_quantize(size);
|
|
|
|
|
|
|
|
if (qsize < size) {
|
|
|
|
/*
|
|
|
|
* Skip a quantization that may have an adequately large run,
|
|
|
|
* because under-sized runs may be mixed in. This only happens
|
|
|
|
* when an unusual size is requested, i.e. for aligned
|
|
|
|
* allocation, and is just one of several places where linear
|
|
|
|
* search would potentially find sufficiently aligned available
|
|
|
|
* memory somewhere lower.
|
|
|
|
*/
|
|
|
|
qsize = run_quantize_next(size);
|
|
|
|
}
|
|
|
|
return (qsize);
|
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C int
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_avail_comp(arena_chunk_map_misc_t *a, arena_chunk_map_misc_t *b)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
int ret;
|
2015-03-07 09:14:05 +08:00
|
|
|
uintptr_t a_miscelm = (uintptr_t)a;
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t a_qsize = run_quantize(arena_miscelm_is_key(a) ?
|
|
|
|
arena_miscelm_key_size_get(a) : arena_miscelm_size_get(a));
|
|
|
|
size_t b_qsize = run_quantize(arena_miscelm_size_get(b));
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-03-07 09:14:05 +08:00
|
|
|
/*
|
2015-05-05 00:58:36 +08:00
|
|
|
* Compare based on quantized size rather than size, in order to sort
|
|
|
|
* equally useful runs only by address.
|
2015-03-07 09:14:05 +08:00
|
|
|
*/
|
2015-05-05 00:58:36 +08:00
|
|
|
ret = (a_qsize > b_qsize) - (a_qsize < b_qsize);
|
2014-08-07 07:10:08 +08:00
|
|
|
if (ret == 0) {
|
2015-08-05 01:49:46 +08:00
|
|
|
if (!arena_miscelm_is_key(a)) {
|
2015-03-07 09:14:05 +08:00
|
|
|
uintptr_t b_miscelm = (uintptr_t)b;
|
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
ret = (a_miscelm > b_miscelm) - (a_miscelm < b_miscelm);
|
2015-03-07 09:14:05 +08:00
|
|
|
} else {
|
2014-08-07 07:43:01 +08:00
|
|
|
/*
|
|
|
|
* Treat keys as if they are lower than anything else.
|
|
|
|
*/
|
2014-08-07 07:10:08 +08:00
|
|
|
ret = -1;
|
2014-08-07 07:43:01 +08:00
|
|
|
}
|
2014-08-07 07:10:08 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2010-03-01 07:00:18 +08:00
|
|
|
/* Generate red-black tree functions. */
|
2014-08-30 04:34:40 +08:00
|
|
|
rb_gen(static UNUSED, arena_avail_tree_, arena_avail_tree_t,
|
|
|
|
arena_chunk_map_misc_t, rb_link, arena_avail_comp)
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-10-31 06:42:37 +08:00
|
|
|
static void
|
|
|
|
arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind,
|
2014-07-22 10:39:20 +08:00
|
|
|
size_t npages)
|
2012-10-31 06:42:37 +08:00
|
|
|
{
|
|
|
|
|
|
|
|
assert(npages == (arena_mapbits_unallocated_size_get(chunk, pageind) >>
|
|
|
|
LG_PAGE));
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_avail_tree_insert(&arena->runs_avail, arena_miscelm_get(chunk,
|
2012-10-31 06:42:37 +08:00
|
|
|
pageind));
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind,
|
2014-07-22 10:39:20 +08:00
|
|
|
size_t npages)
|
2012-10-31 06:42:37 +08:00
|
|
|
{
|
|
|
|
|
|
|
|
assert(npages == (arena_mapbits_unallocated_size_get(chunk, pageind) >>
|
|
|
|
LG_PAGE));
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_avail_tree_remove(&arena->runs_avail, arena_miscelm_get(chunk,
|
2012-10-31 06:42:37 +08:00
|
|
|
pageind));
|
|
|
|
}
|
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
static void
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_run_dirty_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind,
|
2014-08-15 05:45:58 +08:00
|
|
|
size_t npages)
|
|
|
|
{
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_miscelm_get(chunk, pageind);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
assert(npages == (arena_mapbits_unallocated_size_get(chunk, pageind) >>
|
|
|
|
LG_PAGE));
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, pageind) == CHUNK_MAP_DIRTY);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, pageind+npages-1) ==
|
|
|
|
CHUNK_MAP_DIRTY);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
2015-03-11 09:15:40 +08:00
|
|
|
qr_new(&miscelm->rd, rd_link);
|
|
|
|
qr_meld(&arena->runs_dirty, &miscelm->rd, rd_link);
|
2014-08-15 05:45:58 +08:00
|
|
|
arena->ndirty += npages;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_run_dirty_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind,
|
2014-08-15 05:45:58 +08:00
|
|
|
size_t npages)
|
|
|
|
{
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_miscelm_get(chunk, pageind);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
assert(npages == (arena_mapbits_unallocated_size_get(chunk, pageind) >>
|
|
|
|
LG_PAGE));
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, pageind) == CHUNK_MAP_DIRTY);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, pageind+npages-1) ==
|
|
|
|
CHUNK_MAP_DIRTY);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
2015-03-11 09:15:40 +08:00
|
|
|
qr_remove(&miscelm->rd, rd_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
assert(arena->ndirty >= npages);
|
2014-08-15 05:45:58 +08:00
|
|
|
arena->ndirty -= npages;
|
|
|
|
}
|
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
static size_t
|
|
|
|
arena_chunk_dirty_npages(const extent_node_t *node)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (extent_node_size_get(node) >> LG_PAGE);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2015-02-18 17:15:50 +08:00
|
|
|
arena_chunk_cache_maybe_insert(arena_t *arena, extent_node_t *node, bool cache)
|
2015-02-16 10:04:46 +08:00
|
|
|
{
|
|
|
|
|
2015-02-18 17:15:50 +08:00
|
|
|
if (cache) {
|
2015-02-18 14:23:10 +08:00
|
|
|
extent_node_dirty_linkage_init(node);
|
2015-02-18 17:15:50 +08:00
|
|
|
extent_node_dirty_insert(node, &arena->runs_dirty,
|
|
|
|
&arena->chunks_cache);
|
2015-02-16 10:04:46 +08:00
|
|
|
arena->ndirty += arena_chunk_dirty_npages(node);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2015-02-18 17:15:50 +08:00
|
|
|
arena_chunk_cache_maybe_remove(arena_t *arena, extent_node_t *node, bool dirty)
|
2015-02-16 10:04:46 +08:00
|
|
|
{
|
|
|
|
|
|
|
|
if (dirty) {
|
2015-02-18 17:15:50 +08:00
|
|
|
extent_node_dirty_remove(node);
|
2015-02-16 10:04:46 +08:00
|
|
|
assert(arena->ndirty >= arena_chunk_dirty_npages(node));
|
|
|
|
arena->ndirty -= arena_chunk_dirty_npages(node);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C void *
|
2011-03-16 04:59:15 +08:00
|
|
|
arena_run_reg_alloc(arena_run_t *run, arena_bin_info_t *bin_info)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
void *ret;
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
unsigned regind;
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm;
|
|
|
|
void *rpages;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2010-03-14 05:41:58 +08:00
|
|
|
assert(run->nfree > 0);
|
2014-10-04 01:16:09 +08:00
|
|
|
assert(!bitmap_full(run->bitmap, &bin_info->bitmap_info));
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-09-29 16:31:39 +08:00
|
|
|
regind = bitmap_sfu(run->bitmap, &bin_info->bitmap_info);
|
|
|
|
miscelm = arena_run_to_miscelm(run);
|
|
|
|
rpages = arena_miscelm_to_rpages(miscelm);
|
|
|
|
ret = (void *)((uintptr_t)rpages + (uintptr_t)bin_info->reg0_offset +
|
2012-04-06 15:35:09 +08:00
|
|
|
(uintptr_t)(bin_info->reg_interval * regind));
|
2010-03-14 05:41:58 +08:00
|
|
|
run->nfree--;
|
|
|
|
return (ret);
|
2010-02-11 02:37:56 +08:00
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C void
|
2010-03-14 05:41:58 +08:00
|
|
|
arena_run_reg_dalloc(arena_run_t *run, void *ptr)
|
2010-02-11 02:37:56 +08:00
|
|
|
{
|
2011-03-16 04:59:15 +08:00
|
|
|
arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run);
|
2012-05-02 15:30:36 +08:00
|
|
|
size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE;
|
|
|
|
size_t mapbits = arena_mapbits_get(chunk, pageind);
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind = arena_ptr_small_binind_get(ptr, mapbits);
|
2011-03-16 04:59:15 +08:00
|
|
|
arena_bin_info_t *bin_info = &arena_bin_info[binind];
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
unsigned regind = arena_run_regind(run, bin_info, ptr);
|
|
|
|
|
2011-03-16 04:59:15 +08:00
|
|
|
assert(run->nfree < bin_info->nregs);
|
2010-03-14 05:41:58 +08:00
|
|
|
/* Freeing an interior pointer can cause assertion failure. */
|
2014-09-29 16:31:39 +08:00
|
|
|
assert(((uintptr_t)ptr -
|
|
|
|
((uintptr_t)arena_miscelm_to_rpages(arena_run_to_miscelm(run)) +
|
2012-04-06 15:35:09 +08:00
|
|
|
(uintptr_t)bin_info->reg0_offset)) %
|
|
|
|
(uintptr_t)bin_info->reg_interval == 0);
|
2014-09-29 16:31:39 +08:00
|
|
|
assert((uintptr_t)ptr >=
|
|
|
|
(uintptr_t)arena_miscelm_to_rpages(arena_run_to_miscelm(run)) +
|
2011-03-16 04:59:15 +08:00
|
|
|
(uintptr_t)bin_info->reg0_offset);
|
Use bitmaps to track small regions.
The previous free list implementation, which embedded singly linked
lists in available regions, had the unfortunate side effect of causing
many cache misses during thread cache fills. Fix this in two places:
- arena_run_t: Use a new bitmap implementation to track which regions
are available. Furthermore, revert to preferring the
lowest available region (as jemalloc did with its old
bitmap-based approach).
- tcache_t: Move read-only tcache_bin_t metadata into
tcache_bin_info_t, and add a contiguous array of pointers
to tcache_t in order to track cached objects. This
substantially increases the size of tcache_t, but results
in much higher data locality for common tcache operations.
As a side benefit, it is again possible to efficiently
flush the least recently used cached objects, so this
change changes flushing from MRU to LRU.
The new bitmap implementation uses a multi-level summary approach to
make finding the lowest available region very fast. In practice,
bitmaps only have one or two levels, though the implementation is
general enough to handle extremely large bitmaps, mainly so that large
page sizes can still be entertained.
Fix tcache_bin_flush_large() to always flush statistics, in the same way
that tcache_bin_flush_small() was recently fixed.
Use JEMALLOC_DEBUG rather than NDEBUG.
Add dassert(), and use it for debug-only asserts.
2011-03-17 01:30:13 +08:00
|
|
|
/* Freeing an unallocated pointer can cause assertion failure. */
|
2014-09-29 16:31:39 +08:00
|
|
|
assert(bitmap_get(run->bitmap, &bin_info->bitmap_info, regind));
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-09-29 16:31:39 +08:00
|
|
|
bitmap_unset(run->bitmap, &bin_info->bitmap_info, regind);
|
2010-03-14 05:41:58 +08:00
|
|
|
run->nfree++;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C void
|
2013-01-22 12:04:42 +08:00
|
|
|
arena_run_zero(arena_chunk_t *chunk, size_t run_ind, size_t npages)
|
|
|
|
{
|
|
|
|
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED((void *)((uintptr_t)chunk +
|
|
|
|
(run_ind << LG_PAGE)), (npages << LG_PAGE));
|
2013-01-22 12:04:42 +08:00
|
|
|
memset((void *)((uintptr_t)chunk + (run_ind << LG_PAGE)), 0,
|
|
|
|
(npages << LG_PAGE));
|
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C void
|
2013-10-20 14:48:40 +08:00
|
|
|
arena_run_page_mark_zeroed(arena_chunk_t *chunk, size_t run_ind)
|
|
|
|
{
|
|
|
|
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_DEFINED((void *)((uintptr_t)chunk + (run_ind
|
|
|
|
<< LG_PAGE)), PAGE);
|
2013-10-20 14:48:40 +08:00
|
|
|
}
|
|
|
|
|
2014-10-31 07:38:08 +08:00
|
|
|
JEMALLOC_INLINE_C void
|
2013-01-22 12:04:42 +08:00
|
|
|
arena_run_page_validate_zeroed(arena_chunk_t *chunk, size_t run_ind)
|
2010-10-19 08:45:40 +08:00
|
|
|
{
|
2010-10-25 11:08:37 +08:00
|
|
|
size_t i;
|
2012-04-02 22:04:34 +08:00
|
|
|
UNUSED size_t *p = (size_t *)((uintptr_t)chunk + (run_ind << LG_PAGE));
|
2010-10-25 11:08:37 +08:00
|
|
|
|
2013-10-20 14:48:40 +08:00
|
|
|
arena_run_page_mark_zeroed(chunk, run_ind);
|
2012-04-02 22:04:34 +08:00
|
|
|
for (i = 0; i < PAGE / sizeof(size_t); i++)
|
2010-10-19 08:45:40 +08:00
|
|
|
assert(p[i] == 0);
|
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
static void
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_cactive_update(arena_t *arena, size_t add_pages, size_t sub_pages)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
if (config_stats) {
|
2014-08-07 14:38:39 +08:00
|
|
|
ssize_t cactive_diff = CHUNK_CEILING((arena->nactive + add_pages
|
|
|
|
- sub_pages) << LG_PAGE) - CHUNK_CEILING(arena->nactive <<
|
|
|
|
LG_PAGE);
|
2014-01-15 08:23:03 +08:00
|
|
|
if (cactive_diff != 0)
|
|
|
|
stats_cactive_add(cactive_diff);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_run_split_remove(arena_t *arena, arena_chunk_t *chunk, size_t run_ind,
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t flag_dirty, size_t flag_decommitted, size_t need_pages)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
|
|
|
size_t total_pages, rem_pages;
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
assert(flag_dirty == 0 || flag_decommitted == 0);
|
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
total_pages = arena_mapbits_unallocated_size_get(chunk, run_ind) >>
|
|
|
|
LG_PAGE;
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, run_ind+total_pages-1) ==
|
|
|
|
flag_dirty);
|
|
|
|
assert(need_pages <= total_pages);
|
|
|
|
rem_pages = total_pages - need_pages;
|
|
|
|
|
2014-07-22 10:39:20 +08:00
|
|
|
arena_avail_remove(arena, chunk, run_ind, total_pages);
|
2014-08-15 05:45:58 +08:00
|
|
|
if (flag_dirty != 0)
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_run_dirty_remove(arena, chunk, run_ind, total_pages);
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_cactive_update(arena, need_pages, 0);
|
|
|
|
arena->nactive += need_pages;
|
|
|
|
|
|
|
|
/* Keep track of trailing unused pages for later use. */
|
|
|
|
if (rem_pages > 0) {
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t flags = flag_dirty | flag_decommitted;
|
2015-08-12 03:42:33 +08:00
|
|
|
size_t flag_unzeroed_mask = (flags == 0) ? CHUNK_MAP_UNZEROED :
|
|
|
|
0;
|
|
|
|
|
|
|
|
arena_mapbits_unallocated_set(chunk, run_ind+need_pages,
|
|
|
|
(rem_pages << LG_PAGE), flags |
|
|
|
|
(arena_mapbits_unzeroed_get(chunk, run_ind+need_pages) &
|
|
|
|
flag_unzeroed_mask));
|
|
|
|
arena_mapbits_unallocated_set(chunk, run_ind+total_pages-1,
|
|
|
|
(rem_pages << LG_PAGE), flags |
|
|
|
|
(arena_mapbits_unzeroed_get(chunk, run_ind+total_pages-1) &
|
|
|
|
flag_unzeroed_mask));
|
|
|
|
if (flag_dirty != 0) {
|
|
|
|
arena_run_dirty_insert(arena, chunk, run_ind+need_pages,
|
|
|
|
rem_pages);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
2014-07-22 10:39:20 +08:00
|
|
|
arena_avail_insert(arena, chunk, run_ind+need_pages, rem_pages);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
static bool
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_split_large_helper(arena_t *arena, arena_run_t *run, size_t size,
|
|
|
|
bool remove, bool zero)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm;
|
2015-09-04 18:15:28 +08:00
|
|
|
size_t flag_dirty, flag_decommitted, run_ind, need_pages;
|
2015-08-12 03:42:33 +08:00
|
|
|
size_t flag_unzeroed_mask;
|
2012-05-02 15:30:36 +08:00
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run);
|
2014-09-29 16:31:39 +08:00
|
|
|
miscelm = arena_run_to_miscelm(run);
|
|
|
|
run_ind = arena_miscelm_to_pageind(miscelm);
|
2012-05-02 15:30:36 +08:00
|
|
|
flag_dirty = arena_mapbits_dirty_get(chunk, run_ind);
|
2015-08-05 01:49:46 +08:00
|
|
|
flag_decommitted = arena_mapbits_decommitted_get(chunk, run_ind);
|
2012-04-02 22:04:34 +08:00
|
|
|
need_pages = (size >> LG_PAGE);
|
2010-01-17 01:53:50 +08:00
|
|
|
assert(need_pages > 0);
|
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
if (flag_decommitted != 0 && arena->chunk_hooks.commit(chunk, chunksize,
|
|
|
|
run_ind << LG_PAGE, size, arena->ind))
|
2015-08-05 01:49:46 +08:00
|
|
|
return (true);
|
|
|
|
|
2013-10-30 09:17:42 +08:00
|
|
|
if (remove) {
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_split_remove(arena, chunk, run_ind, flag_dirty,
|
2015-08-05 01:49:46 +08:00
|
|
|
flag_decommitted, need_pages);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
2013-10-30 09:17:42 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
if (zero) {
|
2015-08-12 03:42:33 +08:00
|
|
|
if (flag_decommitted != 0) {
|
|
|
|
/* The run is untouched, and therefore zeroed. */
|
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_DEFINED((void
|
|
|
|
*)((uintptr_t)chunk + (run_ind << LG_PAGE)),
|
|
|
|
(need_pages << LG_PAGE));
|
|
|
|
} else if (flag_dirty != 0) {
|
|
|
|
/* The run is dirty, so all pages must be zeroed. */
|
|
|
|
arena_run_zero(chunk, run_ind, need_pages);
|
|
|
|
} else {
|
2013-10-30 09:17:42 +08:00
|
|
|
/*
|
2014-01-15 08:23:03 +08:00
|
|
|
* The run is clean, so some pages may be zeroed (i.e.
|
|
|
|
* never before touched).
|
2013-10-30 09:17:42 +08:00
|
|
|
*/
|
2015-09-04 18:15:28 +08:00
|
|
|
size_t i;
|
2014-01-15 08:23:03 +08:00
|
|
|
for (i = 0; i < need_pages; i++) {
|
|
|
|
if (arena_mapbits_unzeroed_get(chunk, run_ind+i)
|
|
|
|
!= 0)
|
|
|
|
arena_run_zero(chunk, run_ind+i, 1);
|
|
|
|
else if (config_debug) {
|
|
|
|
arena_run_page_validate_zeroed(chunk,
|
|
|
|
run_ind+i);
|
|
|
|
} else {
|
|
|
|
arena_run_page_mark_zeroed(chunk,
|
|
|
|
run_ind+i);
|
2010-03-19 11:36:40 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
}
|
2010-03-19 11:36:40 +08:00
|
|
|
} else {
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED((void *)((uintptr_t)chunk +
|
2013-10-20 14:48:40 +08:00
|
|
|
(run_ind << LG_PAGE)), (need_pages << LG_PAGE));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2014-01-15 08:23:03 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Set the last element first, in case the run only contains one page
|
|
|
|
* (i.e. both statements set the same element).
|
|
|
|
*/
|
2015-08-12 03:42:33 +08:00
|
|
|
flag_unzeroed_mask = (flag_dirty | flag_decommitted) == 0 ?
|
|
|
|
CHUNK_MAP_UNZEROED : 0;
|
|
|
|
arena_mapbits_large_set(chunk, run_ind+need_pages-1, 0, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk,
|
|
|
|
run_ind+need_pages-1)));
|
|
|
|
arena_mapbits_large_set(chunk, run_ind, size, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk, run_ind)));
|
2015-08-05 01:49:46 +08:00
|
|
|
return (false);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
static bool
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_split_large(arena_t *arena, arena_run_t *run, size_t size, bool zero)
|
2013-10-30 09:17:42 +08:00
|
|
|
{
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
return (arena_run_split_large_helper(arena, run, size, true, zero));
|
2013-10-30 09:17:42 +08:00
|
|
|
}
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
static bool
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_init_large(arena_t *arena, arena_run_t *run, size_t size, bool zero)
|
2013-10-30 09:17:42 +08:00
|
|
|
{
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
return (arena_run_split_large_helper(arena, run, size, false, zero));
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
static bool
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_split_small(arena_t *arena, arena_run_t *run, size_t size,
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm;
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t flag_dirty, flag_decommitted, run_ind, need_pages, i;
|
2014-01-15 08:23:03 +08:00
|
|
|
|
|
|
|
assert(binind != BININD_INVALID);
|
|
|
|
|
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run);
|
2014-09-29 16:31:39 +08:00
|
|
|
miscelm = arena_run_to_miscelm(run);
|
|
|
|
run_ind = arena_miscelm_to_pageind(miscelm);
|
2014-01-15 08:23:03 +08:00
|
|
|
flag_dirty = arena_mapbits_dirty_get(chunk, run_ind);
|
2015-08-05 01:49:46 +08:00
|
|
|
flag_decommitted = arena_mapbits_decommitted_get(chunk, run_ind);
|
2014-01-15 08:23:03 +08:00
|
|
|
need_pages = (size >> LG_PAGE);
|
|
|
|
assert(need_pages > 0);
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
if (flag_decommitted != 0 && arena->chunk_hooks.commit(chunk, chunksize,
|
|
|
|
run_ind << LG_PAGE, size, arena->ind))
|
|
|
|
return (true);
|
|
|
|
|
|
|
|
arena_run_split_remove(arena, chunk, run_ind, flag_dirty,
|
|
|
|
flag_decommitted, need_pages);
|
2014-01-15 08:23:03 +08:00
|
|
|
|
2014-10-11 14:01:03 +08:00
|
|
|
for (i = 0; i < need_pages; i++) {
|
2015-08-12 03:42:33 +08:00
|
|
|
size_t flag_unzeroed = arena_mapbits_unzeroed_get(chunk,
|
|
|
|
run_ind+i);
|
|
|
|
arena_mapbits_small_set(chunk, run_ind+i, i, binind,
|
|
|
|
flag_unzeroed);
|
|
|
|
if (config_debug && flag_dirty == 0 && flag_unzeroed == 0)
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_page_validate_zeroed(chunk, run_ind+i);
|
|
|
|
}
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED((void *)((uintptr_t)chunk +
|
2014-01-15 08:23:03 +08:00
|
|
|
(run_ind << LG_PAGE)), (need_pages << LG_PAGE));
|
2015-08-05 01:49:46 +08:00
|
|
|
return (false);
|
2013-10-30 09:17:42 +08:00
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
static arena_chunk_t *
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_chunk_init_spare(arena_t *arena)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
assert(arena->spare != NULL);
|
2010-03-19 11:36:40 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
chunk = arena->spare;
|
|
|
|
arena->spare = NULL;
|
2010-01-25 09:13:07 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
assert(arena_mapbits_allocated_get(chunk, map_bias) == 0);
|
|
|
|
assert(arena_mapbits_allocated_get(chunk, chunk_npages-1) == 0);
|
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk, map_bias) ==
|
2014-10-06 08:54:10 +08:00
|
|
|
arena_maxrun);
|
2014-01-15 08:23:03 +08:00
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk, chunk_npages-1) ==
|
2014-10-06 08:54:10 +08:00
|
|
|
arena_maxrun);
|
2014-01-15 08:23:03 +08:00
|
|
|
assert(arena_mapbits_dirty_get(chunk, map_bias) ==
|
|
|
|
arena_mapbits_dirty_get(chunk, chunk_npages-1));
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
return (chunk);
|
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-02-19 08:40:53 +08:00
|
|
|
static bool
|
|
|
|
arena_chunk_register(arena_t *arena, arena_chunk_t *chunk, bool zero)
|
|
|
|
{
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
/*
|
|
|
|
* The extent node notion of "committed" doesn't directly apply to
|
2015-08-10 07:47:27 +08:00
|
|
|
* arena chunks. Arbitrarily mark them as committed. The commit state
|
|
|
|
* of runs is tracked individually, and upon chunk deallocation the
|
|
|
|
* entire chunk is in a consistent commit state.
|
2015-08-05 01:49:46 +08:00
|
|
|
*/
|
|
|
|
extent_node_init(&chunk->node, arena, chunk, chunksize, zero, true);
|
2015-02-19 08:40:53 +08:00
|
|
|
extent_node_achunk_set(&chunk->node, true);
|
|
|
|
return (chunk_register(chunk, &chunk->node));
|
|
|
|
}
|
|
|
|
|
2014-05-16 13:22:27 +08:00
|
|
|
static arena_chunk_t *
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena_chunk_alloc_internal_hard(arena_t *arena, chunk_hooks_t *chunk_hooks,
|
2015-08-05 01:49:46 +08:00
|
|
|
bool *zero, bool *commit)
|
2014-05-16 13:22:27 +08:00
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
|
|
|
|
chunk = (arena_chunk_t *)chunk_alloc_wrapper(arena, chunk_hooks, NULL,
|
2015-08-05 01:49:46 +08:00
|
|
|
chunksize, chunksize, zero, commit);
|
|
|
|
if (chunk != NULL && !*commit) {
|
|
|
|
/* Commit header. */
|
|
|
|
if (chunk_hooks->commit(chunk, chunksize, 0, map_bias <<
|
|
|
|
LG_PAGE, arena->ind)) {
|
|
|
|
chunk_dalloc_wrapper(arena, chunk_hooks,
|
|
|
|
(void *)chunk, chunksize, *commit);
|
|
|
|
chunk = NULL;
|
|
|
|
}
|
|
|
|
}
|
2015-02-19 08:40:53 +08:00
|
|
|
if (chunk != NULL && arena_chunk_register(arena, chunk, *zero)) {
|
2015-08-05 01:49:46 +08:00
|
|
|
if (!*commit) {
|
|
|
|
/* Undo commit of header. */
|
|
|
|
chunk_hooks->decommit(chunk, chunksize, 0, map_bias <<
|
|
|
|
LG_PAGE, arena->ind);
|
|
|
|
}
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_dalloc_wrapper(arena, chunk_hooks, (void *)chunk,
|
2015-08-05 01:49:46 +08:00
|
|
|
chunksize, *commit);
|
2015-02-19 08:40:53 +08:00
|
|
|
chunk = NULL;
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
}
|
2015-02-19 08:40:53 +08:00
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-02-19 08:40:53 +08:00
|
|
|
return (chunk);
|
|
|
|
}
|
|
|
|
|
|
|
|
static arena_chunk_t *
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_chunk_alloc_internal(arena_t *arena, bool *zero, bool *commit)
|
2015-02-19 08:40:53 +08:00
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_hooks_t chunk_hooks = CHUNK_HOOKS_INITIALIZER;
|
2015-02-19 08:40:53 +08:00
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk = chunk_alloc_cache(arena, &chunk_hooks, NULL, chunksize,
|
|
|
|
chunksize, zero, true);
|
2015-08-05 01:49:46 +08:00
|
|
|
if (chunk != NULL) {
|
|
|
|
if (arena_chunk_register(arena, chunk, *zero)) {
|
|
|
|
chunk_dalloc_cache(arena, &chunk_hooks, chunk,
|
2015-08-10 07:47:27 +08:00
|
|
|
chunksize, true);
|
2015-08-05 01:49:46 +08:00
|
|
|
return (NULL);
|
|
|
|
}
|
|
|
|
*commit = true;
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
}
|
|
|
|
if (chunk == NULL) {
|
|
|
|
chunk = arena_chunk_alloc_internal_hard(arena, &chunk_hooks,
|
2015-08-05 01:49:46 +08:00
|
|
|
zero, commit);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
}
|
2015-02-19 08:40:53 +08:00
|
|
|
|
2014-11-28 03:22:36 +08:00
|
|
|
if (config_stats && chunk != NULL) {
|
2014-05-16 13:22:27 +08:00
|
|
|
arena->stats.mapped += chunksize;
|
2014-11-28 03:22:36 +08:00
|
|
|
arena->stats.metadata_mapped += (map_bias << LG_PAGE);
|
|
|
|
}
|
2014-05-16 13:22:27 +08:00
|
|
|
|
|
|
|
return (chunk);
|
|
|
|
}
|
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
static arena_chunk_t *
|
|
|
|
arena_chunk_init_hard(arena_t *arena)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
2015-08-05 01:49:46 +08:00
|
|
|
bool zero, commit;
|
2015-08-11 14:03:34 +08:00
|
|
|
size_t flag_unzeroed, flag_decommitted, i;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
assert(arena->spare == NULL);
|
2012-10-31 06:42:37 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
zero = false;
|
2015-08-05 01:49:46 +08:00
|
|
|
commit = false;
|
|
|
|
chunk = arena_chunk_alloc_internal(arena, &zero, &commit);
|
2014-01-15 08:23:03 +08:00
|
|
|
if (chunk == NULL)
|
|
|
|
return (NULL);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Initialize the map to contain one maximal free untouched run. Mark
|
2015-08-05 01:49:46 +08:00
|
|
|
* the pages as zeroed if chunk_alloc() returned a zeroed or decommitted
|
|
|
|
* chunk.
|
2014-01-15 08:23:03 +08:00
|
|
|
*/
|
2015-08-11 14:03:34 +08:00
|
|
|
flag_unzeroed = (zero || !commit) ? 0 : CHUNK_MAP_UNZEROED;
|
|
|
|
flag_decommitted = commit ? 0 : CHUNK_MAP_DECOMMITTED;
|
|
|
|
arena_mapbits_unallocated_set(chunk, map_bias, arena_maxrun,
|
|
|
|
flag_unzeroed | flag_decommitted);
|
2014-01-15 08:23:03 +08:00
|
|
|
/*
|
|
|
|
* There is no need to initialize the internal page map entries unless
|
|
|
|
* the chunk is not zeroed.
|
|
|
|
*/
|
2014-10-04 01:16:09 +08:00
|
|
|
if (!zero) {
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED(
|
2014-08-30 04:34:40 +08:00
|
|
|
(void *)arena_bitselm_get(chunk, map_bias+1),
|
|
|
|
(size_t)((uintptr_t) arena_bitselm_get(chunk,
|
|
|
|
chunk_npages-1) - (uintptr_t)arena_bitselm_get(chunk,
|
|
|
|
map_bias+1)));
|
2014-01-15 08:23:03 +08:00
|
|
|
for (i = map_bias+1; i < chunk_npages-1; i++)
|
2015-08-11 14:03:34 +08:00
|
|
|
arena_mapbits_internal_set(chunk, i, flag_unzeroed);
|
2014-01-15 08:23:03 +08:00
|
|
|
} else {
|
2014-08-30 04:34:40 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_DEFINED((void
|
|
|
|
*)arena_bitselm_get(chunk, map_bias+1), (size_t)((uintptr_t)
|
|
|
|
arena_bitselm_get(chunk, chunk_npages-1) -
|
|
|
|
(uintptr_t)arena_bitselm_get(chunk, map_bias+1)));
|
2014-01-15 08:23:03 +08:00
|
|
|
if (config_debug) {
|
|
|
|
for (i = map_bias+1; i < chunk_npages-1; i++) {
|
|
|
|
assert(arena_mapbits_unzeroed_get(chunk, i) ==
|
2015-08-11 14:03:34 +08:00
|
|
|
flag_unzeroed);
|
2012-05-02 15:30:36 +08:00
|
|
|
}
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
}
|
2010-03-19 11:36:40 +08:00
|
|
|
}
|
2014-10-06 08:54:10 +08:00
|
|
|
arena_mapbits_unallocated_set(chunk, chunk_npages-1, arena_maxrun,
|
2015-08-11 14:03:34 +08:00
|
|
|
flag_unzeroed);
|
2014-01-15 08:23:03 +08:00
|
|
|
|
|
|
|
return (chunk);
|
|
|
|
}
|
|
|
|
|
|
|
|
static arena_chunk_t *
|
|
|
|
arena_chunk_alloc(arena_t *arena)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
|
|
|
|
if (arena->spare != NULL)
|
|
|
|
chunk = arena_chunk_init_spare(arena);
|
2014-03-26 13:36:05 +08:00
|
|
|
else {
|
2014-01-15 08:23:03 +08:00
|
|
|
chunk = arena_chunk_init_hard(arena);
|
2014-03-26 13:36:05 +08:00
|
|
|
if (chunk == NULL)
|
|
|
|
return (NULL);
|
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-10-31 06:42:37 +08:00
|
|
|
/* Insert the run into the runs_avail tree. */
|
2014-07-22 10:39:20 +08:00
|
|
|
arena_avail_insert(arena, chunk, map_bias, chunk_npages-map_bias);
|
2012-10-31 06:42:37 +08:00
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
return (chunk);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2014-10-15 13:20:00 +08:00
|
|
|
arena_chunk_dalloc(arena_t *arena, arena_chunk_t *chunk)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(arena_mapbits_allocated_get(chunk, map_bias) == 0);
|
|
|
|
assert(arena_mapbits_allocated_get(chunk, chunk_npages-1) == 0);
|
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk, map_bias) ==
|
|
|
|
arena_maxrun);
|
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk, chunk_npages-1) ==
|
|
|
|
arena_maxrun);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, map_bias) ==
|
|
|
|
arena_mapbits_dirty_get(chunk, chunk_npages-1));
|
2015-08-10 07:47:27 +08:00
|
|
|
assert(arena_mapbits_decommitted_get(chunk, map_bias) ==
|
|
|
|
arena_mapbits_decommitted_get(chunk, chunk_npages-1));
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove run from the runs_avail tree, so that the arena does not use
|
|
|
|
* it.
|
|
|
|
*/
|
|
|
|
arena_avail_remove(arena, chunk, map_bias, chunk_npages-map_bias);
|
|
|
|
|
|
|
|
if (arena->spare != NULL) {
|
|
|
|
arena_chunk_t *spare = arena->spare;
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_hooks_t chunk_hooks = CHUNK_HOOKS_INITIALIZER;
|
2015-08-10 07:47:27 +08:00
|
|
|
bool committed;
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
arena->spare = chunk;
|
|
|
|
if (arena_mapbits_dirty_get(spare, map_bias) != 0) {
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_run_dirty_remove(arena, spare, map_bias,
|
2014-10-15 13:20:00 +08:00
|
|
|
chunk_npages-map_bias);
|
|
|
|
}
|
2015-02-19 08:40:53 +08:00
|
|
|
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
chunk_deregister(spare, &spare->node);
|
2015-02-19 08:40:53 +08:00
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
committed = (arena_mapbits_decommitted_get(spare, map_bias) ==
|
|
|
|
0);
|
|
|
|
if (!committed) {
|
|
|
|
/*
|
|
|
|
* Decommit the header. Mark the chunk as decommitted
|
|
|
|
* even if header decommit fails, since treating a
|
|
|
|
* partially committed chunk as committed has a high
|
|
|
|
* potential for causing later access of decommitted
|
|
|
|
* memory.
|
|
|
|
*/
|
|
|
|
chunk_hooks = chunk_hooks_get(arena);
|
|
|
|
chunk_hooks.decommit(spare, chunksize, 0, map_bias <<
|
|
|
|
LG_PAGE, arena->ind);
|
|
|
|
}
|
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_dalloc_cache(arena, &chunk_hooks, (void *)spare,
|
2015-08-10 07:47:27 +08:00
|
|
|
chunksize, committed);
|
2015-02-19 08:40:53 +08:00
|
|
|
|
2014-11-28 03:22:36 +08:00
|
|
|
if (config_stats) {
|
2014-10-15 13:20:00 +08:00
|
|
|
arena->stats.mapped -= chunksize;
|
2014-11-28 03:22:36 +08:00
|
|
|
arena->stats.metadata_mapped -= (map_bias << LG_PAGE);
|
|
|
|
}
|
2014-10-15 13:20:00 +08:00
|
|
|
} else
|
|
|
|
arena->spare = chunk;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_huge_malloc_stats_update(arena_t *arena, size_t usize)
|
|
|
|
{
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - nlclasses - NBINS;
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
cassert(config_stats);
|
|
|
|
|
|
|
|
arena->stats.nmalloc_huge++;
|
|
|
|
arena->stats.allocated_huge += usize;
|
|
|
|
arena->stats.hstats[index].nmalloc++;
|
|
|
|
arena->stats.hstats[index].curhchunks++;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_huge_malloc_stats_update_undo(arena_t *arena, size_t usize)
|
|
|
|
{
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - nlclasses - NBINS;
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
cassert(config_stats);
|
|
|
|
|
|
|
|
arena->stats.nmalloc_huge--;
|
|
|
|
arena->stats.allocated_huge -= usize;
|
|
|
|
arena->stats.hstats[index].nmalloc--;
|
|
|
|
arena->stats.hstats[index].curhchunks--;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_huge_dalloc_stats_update(arena_t *arena, size_t usize)
|
|
|
|
{
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - nlclasses - NBINS;
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
cassert(config_stats);
|
|
|
|
|
|
|
|
arena->stats.ndalloc_huge++;
|
|
|
|
arena->stats.allocated_huge -= usize;
|
|
|
|
arena->stats.hstats[index].ndalloc++;
|
|
|
|
arena->stats.hstats[index].curhchunks--;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_huge_dalloc_stats_update_undo(arena_t *arena, size_t usize)
|
|
|
|
{
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - nlclasses - NBINS;
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
cassert(config_stats);
|
|
|
|
|
|
|
|
arena->stats.ndalloc_huge--;
|
|
|
|
arena->stats.allocated_huge += usize;
|
|
|
|
arena->stats.hstats[index].ndalloc--;
|
|
|
|
arena->stats.hstats[index].curhchunks++;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_huge_ralloc_stats_update(arena_t *arena, size_t oldsize, size_t usize)
|
|
|
|
{
|
|
|
|
|
|
|
|
arena_huge_dalloc_stats_update(arena, oldsize);
|
|
|
|
arena_huge_malloc_stats_update(arena, usize);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_huge_ralloc_stats_update_undo(arena_t *arena, size_t oldsize,
|
|
|
|
size_t usize)
|
2014-05-16 13:22:27 +08:00
|
|
|
{
|
2014-10-15 13:20:00 +08:00
|
|
|
|
|
|
|
arena_huge_dalloc_stats_update_undo(arena, oldsize);
|
|
|
|
arena_huge_malloc_stats_update_undo(arena, usize);
|
|
|
|
}
|
|
|
|
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
extent_node_t *
|
|
|
|
arena_node_alloc(arena_t *arena)
|
|
|
|
{
|
|
|
|
extent_node_t *node;
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->node_cache_mtx);
|
2015-02-16 08:43:52 +08:00
|
|
|
node = ql_last(&arena->node_cache, ql_link);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
if (node == NULL) {
|
|
|
|
malloc_mutex_unlock(&arena->node_cache_mtx);
|
|
|
|
return (base_alloc(sizeof(extent_node_t)));
|
|
|
|
}
|
2015-02-16 08:43:52 +08:00
|
|
|
ql_tail_remove(&arena->node_cache, extent_node_t, ql_link);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
malloc_mutex_unlock(&arena->node_cache_mtx);
|
|
|
|
return (node);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
arena_node_dalloc(arena_t *arena, extent_node_t *node)
|
|
|
|
{
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->node_cache_mtx);
|
2015-02-16 08:43:52 +08:00
|
|
|
ql_elm_new(node, ql_link);
|
|
|
|
ql_tail_insert(&arena->node_cache, node, ql_link);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
malloc_mutex_unlock(&arena->node_cache_mtx);
|
|
|
|
}
|
|
|
|
|
2015-02-19 08:40:53 +08:00
|
|
|
static void *
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena_chunk_alloc_huge_hard(arena_t *arena, chunk_hooks_t *chunk_hooks,
|
2015-02-19 08:40:53 +08:00
|
|
|
size_t usize, size_t alignment, bool *zero, size_t csize)
|
|
|
|
{
|
|
|
|
void *ret;
|
2015-08-05 01:49:46 +08:00
|
|
|
bool commit = true;
|
2015-02-19 08:40:53 +08:00
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
ret = chunk_alloc_wrapper(arena, chunk_hooks, NULL, csize, alignment,
|
2015-08-05 01:49:46 +08:00
|
|
|
zero, &commit);
|
2015-02-19 08:40:53 +08:00
|
|
|
if (ret == NULL) {
|
|
|
|
/* Revert optimistic stats updates. */
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
if (config_stats) {
|
|
|
|
arena_huge_malloc_stats_update_undo(arena, usize);
|
|
|
|
arena->stats.mapped -= usize;
|
|
|
|
}
|
|
|
|
arena->nactive -= (usize >> LG_PAGE);
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
void *
|
|
|
|
arena_chunk_alloc_huge(arena_t *arena, size_t usize, size_t alignment,
|
|
|
|
bool *zero)
|
|
|
|
{
|
|
|
|
void *ret;
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_hooks_t chunk_hooks = CHUNK_HOOKS_INITIALIZER;
|
2014-10-15 13:20:00 +08:00
|
|
|
size_t csize = CHUNK_CEILING(usize);
|
2014-05-16 13:22:27 +08:00
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-02-19 08:40:53 +08:00
|
|
|
|
|
|
|
/* Optimistically update stats. */
|
2014-10-15 13:20:00 +08:00
|
|
|
if (config_stats) {
|
|
|
|
arena_huge_malloc_stats_update(arena, usize);
|
|
|
|
arena->stats.mapped += usize;
|
|
|
|
}
|
|
|
|
arena->nactive += (usize >> LG_PAGE);
|
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
ret = chunk_alloc_cache(arena, &chunk_hooks, NULL, csize, alignment,
|
|
|
|
zero, true);
|
2015-02-19 08:40:53 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2014-10-15 13:20:00 +08:00
|
|
|
if (ret == NULL) {
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
ret = arena_chunk_alloc_huge_hard(arena, &chunk_hooks, usize,
|
2015-02-19 08:40:53 +08:00
|
|
|
alignment, zero, csize);
|
2014-10-15 13:20:00 +08:00
|
|
|
}
|
|
|
|
|
2015-02-19 08:40:53 +08:00
|
|
|
if (config_stats && ret != NULL)
|
2014-10-15 13:20:00 +08:00
|
|
|
stats_cactive_add(usize);
|
|
|
|
return (ret);
|
2014-05-16 13:22:27 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2014-10-13 13:53:59 +08:00
|
|
|
arena_chunk_dalloc_huge(arena_t *arena, void *chunk, size_t usize)
|
2014-05-16 13:22:27 +08:00
|
|
|
{
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_hooks_t chunk_hooks = CHUNK_HOOKS_INITIALIZER;
|
2015-02-19 08:40:53 +08:00
|
|
|
size_t csize;
|
2014-05-16 13:22:27 +08:00
|
|
|
|
2015-02-19 08:40:53 +08:00
|
|
|
csize = CHUNK_CEILING(usize);
|
2014-05-16 13:22:27 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
if (config_stats) {
|
2014-10-15 13:20:00 +08:00
|
|
|
arena_huge_dalloc_stats_update(arena, usize);
|
2014-10-13 13:53:59 +08:00
|
|
|
arena->stats.mapped -= usize;
|
|
|
|
stats_cactive_sub(usize);
|
2014-05-16 13:22:27 +08:00
|
|
|
}
|
2014-10-13 13:53:59 +08:00
|
|
|
arena->nactive -= (usize >> LG_PAGE);
|
2015-02-19 08:40:53 +08:00
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
chunk_dalloc_cache(arena, &chunk_hooks, chunk, csize, true);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2014-05-16 13:22:27 +08:00
|
|
|
}
|
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
void
|
|
|
|
arena_chunk_ralloc_huge_similar(arena_t *arena, void *chunk, size_t oldsize,
|
|
|
|
size_t usize)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-07-19 05:21:17 +08:00
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
assert(CHUNK_CEILING(oldsize) == CHUNK_CEILING(usize));
|
|
|
|
assert(oldsize != usize);
|
2012-05-11 08:09:17 +08:00
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
if (config_stats)
|
|
|
|
arena_huge_ralloc_stats_update(arena, oldsize, usize);
|
|
|
|
if (oldsize < usize) {
|
|
|
|
size_t udiff = usize - oldsize;
|
|
|
|
arena->nactive += udiff >> LG_PAGE;
|
|
|
|
if (config_stats)
|
|
|
|
stats_cactive_add(udiff);
|
|
|
|
} else {
|
|
|
|
size_t udiff = oldsize - usize;
|
|
|
|
arena->nactive -= udiff >> LG_PAGE;
|
|
|
|
if (config_stats)
|
|
|
|
stats_cactive_sub(udiff);
|
|
|
|
}
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
}
|
2010-04-14 11:53:21 +08:00
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
void
|
|
|
|
arena_chunk_ralloc_huge_shrink(arena_t *arena, void *chunk, size_t oldsize,
|
|
|
|
size_t usize)
|
|
|
|
{
|
|
|
|
size_t udiff = oldsize - usize;
|
|
|
|
size_t cdiff = CHUNK_CEILING(oldsize) - CHUNK_CEILING(usize);
|
2010-03-15 10:43:56 +08:00
|
|
|
|
2014-10-15 13:20:00 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
if (config_stats) {
|
|
|
|
arena_huge_ralloc_stats_update(arena, oldsize, usize);
|
|
|
|
if (cdiff != 0) {
|
|
|
|
arena->stats.mapped -= cdiff;
|
|
|
|
stats_cactive_sub(udiff);
|
2014-08-15 05:45:58 +08:00
|
|
|
}
|
2014-10-15 13:20:00 +08:00
|
|
|
}
|
|
|
|
arena->nactive -= udiff >> LG_PAGE;
|
2015-02-19 08:40:53 +08:00
|
|
|
|
2014-11-18 01:54:49 +08:00
|
|
|
if (cdiff != 0) {
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_hooks_t chunk_hooks = CHUNK_HOOKS_INITIALIZER;
|
2015-02-19 08:40:53 +08:00
|
|
|
void *nchunk = (void *)((uintptr_t)chunk +
|
|
|
|
CHUNK_CEILING(usize));
|
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
chunk_dalloc_cache(arena, &chunk_hooks, nchunk, cdiff, true);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
}
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2015-02-19 08:40:53 +08:00
|
|
|
}
|
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
static bool
|
|
|
|
arena_chunk_ralloc_huge_expand_hard(arena_t *arena, chunk_hooks_t *chunk_hooks,
|
|
|
|
void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk,
|
|
|
|
size_t udiff, size_t cdiff)
|
2015-02-19 08:40:53 +08:00
|
|
|
{
|
|
|
|
bool err;
|
2015-08-05 01:49:46 +08:00
|
|
|
bool commit = true;
|
2015-02-19 08:40:53 +08:00
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
err = (chunk_alloc_wrapper(arena, chunk_hooks, nchunk, cdiff, chunksize,
|
2015-08-05 01:49:46 +08:00
|
|
|
zero, &commit) == NULL);
|
2015-02-19 08:40:53 +08:00
|
|
|
if (err) {
|
|
|
|
/* Revert optimistic stats updates. */
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
if (config_stats) {
|
|
|
|
arena_huge_ralloc_stats_update_undo(arena, oldsize,
|
|
|
|
usize);
|
|
|
|
arena->stats.mapped -= cdiff;
|
|
|
|
}
|
|
|
|
arena->nactive -= (udiff >> LG_PAGE);
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
} else if (chunk_hooks->merge(chunk, CHUNK_CEILING(oldsize), nchunk,
|
|
|
|
cdiff, true, arena->ind)) {
|
2015-08-05 01:49:46 +08:00
|
|
|
chunk_dalloc_arena(arena, chunk_hooks, nchunk, cdiff, *zero,
|
|
|
|
true);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
err = true;
|
2014-11-18 01:54:49 +08:00
|
|
|
}
|
2015-02-19 08:40:53 +08:00
|
|
|
return (err);
|
2014-10-15 13:20:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
|
|
|
arena_chunk_ralloc_huge_expand(arena_t *arena, void *chunk, size_t oldsize,
|
|
|
|
size_t usize, bool *zero)
|
|
|
|
{
|
2015-02-19 08:40:53 +08:00
|
|
|
bool err;
|
2015-08-05 01:49:46 +08:00
|
|
|
chunk_hooks_t chunk_hooks = chunk_hooks_get(arena);
|
2015-02-19 08:40:53 +08:00
|
|
|
void *nchunk = (void *)((uintptr_t)chunk + CHUNK_CEILING(oldsize));
|
2014-10-15 13:20:00 +08:00
|
|
|
size_t udiff = usize - oldsize;
|
|
|
|
size_t cdiff = CHUNK_CEILING(usize) - CHUNK_CEILING(oldsize);
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-02-19 08:40:53 +08:00
|
|
|
|
|
|
|
/* Optimistically update stats. */
|
2014-10-15 13:20:00 +08:00
|
|
|
if (config_stats) {
|
|
|
|
arena_huge_ralloc_stats_update(arena, oldsize, usize);
|
|
|
|
arena->stats.mapped += cdiff;
|
|
|
|
}
|
|
|
|
arena->nactive += (udiff >> LG_PAGE);
|
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
err = (chunk_alloc_cache(arena, &arena->chunk_hooks, nchunk, cdiff,
|
|
|
|
chunksize, zero, true) == NULL);
|
2015-02-19 08:40:53 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
if (err) {
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
err = arena_chunk_ralloc_huge_expand_hard(arena, &chunk_hooks,
|
|
|
|
chunk, oldsize, usize, zero, nchunk, udiff,
|
|
|
|
cdiff);
|
|
|
|
} else if (chunk_hooks.merge(chunk, CHUNK_CEILING(oldsize), nchunk,
|
|
|
|
cdiff, true, arena->ind)) {
|
2015-08-05 01:49:46 +08:00
|
|
|
chunk_dalloc_arena(arena, &chunk_hooks, nchunk, cdiff, *zero,
|
|
|
|
true);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
err = true;
|
2014-10-15 13:20:00 +08:00
|
|
|
}
|
|
|
|
|
2015-02-19 08:40:53 +08:00
|
|
|
if (config_stats && !err)
|
2014-10-15 13:20:00 +08:00
|
|
|
stats_cactive_add(udiff);
|
2015-02-19 08:40:53 +08:00
|
|
|
return (err);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2015-07-16 07:02:21 +08:00
|
|
|
/*
|
|
|
|
* Do first-best-fit run selection, i.e. select the lowest run that best fits.
|
|
|
|
* Run sizes are quantized, so not all candidate runs are necessarily exactly
|
|
|
|
* the same size.
|
|
|
|
*/
|
2010-01-17 01:53:50 +08:00
|
|
|
static arena_run_t *
|
2015-07-16 07:02:21 +08:00
|
|
|
arena_run_first_best_fit(arena_t *arena, size_t size)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
2015-07-16 07:02:21 +08:00
|
|
|
size_t search_size = run_quantize_first(size);
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_chunk_map_misc_t *key = arena_miscelm_key_create(search_size);
|
2015-07-16 07:02:21 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm =
|
|
|
|
arena_avail_tree_nsearch(&arena->runs_avail, key);
|
|
|
|
if (miscelm == NULL)
|
|
|
|
return (NULL);
|
|
|
|
return (&miscelm->run);
|
2015-03-07 11:57:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
|
|
|
arena_run_alloc_large_helper(arena_t *arena, size_t size, bool zero)
|
|
|
|
{
|
2015-07-16 07:02:21 +08:00
|
|
|
arena_run_t *run = arena_run_first_best_fit(arena, s2u(size));
|
2015-08-05 01:49:46 +08:00
|
|
|
if (run != NULL) {
|
|
|
|
if (arena_run_split_large(arena, run, size, zero))
|
|
|
|
run = NULL;
|
|
|
|
}
|
2015-03-07 11:57:36 +08:00
|
|
|
return (run);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
|
|
|
arena_run_alloc_large(arena_t *arena, size_t size, bool zero)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
arena_run_t *run;
|
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
assert(size <= arena_maxrun);
|
2015-05-05 00:58:36 +08:00
|
|
|
assert(size == PAGE_CEILING(size));
|
2014-01-15 08:23:03 +08:00
|
|
|
|
|
|
|
/* Search the arena's chunks for the lowest best fit. */
|
|
|
|
run = arena_run_alloc_large_helper(arena, size, zero);
|
|
|
|
if (run != NULL)
|
|
|
|
return (run);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* No usable runs. Create a new chunk from which to allocate the run.
|
|
|
|
*/
|
|
|
|
chunk = arena_chunk_alloc(arena);
|
|
|
|
if (chunk != NULL) {
|
2014-09-29 16:31:39 +08:00
|
|
|
run = &arena_miscelm_get(chunk, map_bias)->run;
|
2015-08-05 01:49:46 +08:00
|
|
|
if (arena_run_split_large(arena, run, size, zero))
|
|
|
|
run = NULL;
|
2014-01-15 08:23:03 +08:00
|
|
|
return (run);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* arena_chunk_alloc() failed, but another thread may have made
|
|
|
|
* sufficient memory available while this one dropped arena->lock in
|
|
|
|
* arena_chunk_alloc(), so search one more time.
|
|
|
|
*/
|
|
|
|
return (arena_run_alloc_large_helper(arena, size, zero));
|
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
2015-08-20 06:21:32 +08:00
|
|
|
arena_run_alloc_small_helper(arena_t *arena, size_t size, szind_t binind)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2015-07-16 07:02:21 +08:00
|
|
|
arena_run_t *run = arena_run_first_best_fit(arena, size);
|
2015-08-05 01:49:46 +08:00
|
|
|
if (run != NULL) {
|
|
|
|
if (arena_run_split_small(arena, run, size, binind))
|
|
|
|
run = NULL;
|
|
|
|
}
|
2015-03-07 11:57:36 +08:00
|
|
|
return (run);
|
2012-05-11 06:47:24 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
2015-08-20 06:21:32 +08:00
|
|
|
arena_run_alloc_small(arena_t *arena, size_t size, szind_t binind)
|
2012-05-11 06:47:24 +08:00
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
arena_run_t *run;
|
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
assert(size <= arena_maxrun);
|
2015-05-05 00:58:36 +08:00
|
|
|
assert(size == PAGE_CEILING(size));
|
2014-01-15 08:23:03 +08:00
|
|
|
assert(binind != BININD_INVALID);
|
2012-05-11 06:47:24 +08:00
|
|
|
|
|
|
|
/* Search the arena's chunks for the lowest best fit. */
|
2014-01-15 08:23:03 +08:00
|
|
|
run = arena_run_alloc_small_helper(arena, size, binind);
|
2012-05-11 06:47:24 +08:00
|
|
|
if (run != NULL)
|
|
|
|
return (run);
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/*
|
|
|
|
* No usable runs. Create a new chunk from which to allocate the run.
|
|
|
|
*/
|
|
|
|
chunk = arena_chunk_alloc(arena);
|
2010-03-15 10:43:56 +08:00
|
|
|
if (chunk != NULL) {
|
2014-09-29 16:31:39 +08:00
|
|
|
run = &arena_miscelm_get(chunk, map_bias)->run;
|
2015-08-05 01:49:46 +08:00
|
|
|
if (arena_run_split_small(arena, run, size, binind))
|
|
|
|
run = NULL;
|
2010-03-15 10:43:56 +08:00
|
|
|
return (run);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* arena_chunk_alloc() failed, but another thread may have made
|
|
|
|
* sufficient memory available while this one dropped arena->lock in
|
|
|
|
* arena_chunk_alloc(), so search one more time.
|
|
|
|
*/
|
2014-01-15 08:23:03 +08:00
|
|
|
return (arena_run_alloc_small_helper(arena, size, binind));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2015-03-19 09:55:33 +08:00
|
|
|
static bool
|
|
|
|
arena_lg_dirty_mult_valid(ssize_t lg_dirty_mult)
|
|
|
|
{
|
|
|
|
|
2015-03-25 06:59:28 +08:00
|
|
|
return (lg_dirty_mult >= -1 && lg_dirty_mult < (ssize_t)(sizeof(size_t)
|
|
|
|
<< 3));
|
2015-03-19 09:55:33 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
ssize_t
|
|
|
|
arena_lg_dirty_mult_get(arena_t *arena)
|
|
|
|
{
|
|
|
|
ssize_t lg_dirty_mult;
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
lg_dirty_mult = arena->lg_dirty_mult;
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
|
|
|
|
return (lg_dirty_mult);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
|
|
|
arena_lg_dirty_mult_set(arena_t *arena, ssize_t lg_dirty_mult)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (!arena_lg_dirty_mult_valid(lg_dirty_mult))
|
|
|
|
return (true);
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
arena->lg_dirty_mult = lg_dirty_mult;
|
|
|
|
arena_maybe_purge(arena);
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
|
|
|
|
return (false);
|
|
|
|
}
|
|
|
|
|
2015-02-19 08:40:53 +08:00
|
|
|
void
|
2010-03-15 08:36:10 +08:00
|
|
|
arena_maybe_purge(arena_t *arena)
|
|
|
|
{
|
2012-10-31 06:42:37 +08:00
|
|
|
|
|
|
|
/* Don't purge if the option is disabled. */
|
2015-03-19 09:55:33 +08:00
|
|
|
if (arena->lg_dirty_mult < 0)
|
2012-10-31 06:42:37 +08:00
|
|
|
return;
|
2015-06-23 09:50:32 +08:00
|
|
|
/* Don't recursively purge. */
|
|
|
|
if (arena->purging)
|
|
|
|
return;
|
2012-10-31 06:42:37 +08:00
|
|
|
/*
|
2015-06-23 09:50:32 +08:00
|
|
|
* Iterate, since preventing recursive purging could otherwise leave too
|
|
|
|
* many dirty pages.
|
2012-10-31 06:42:37 +08:00
|
|
|
*/
|
2015-06-23 09:50:32 +08:00
|
|
|
while (true) {
|
|
|
|
size_t threshold = (arena->nactive >> arena->lg_dirty_mult);
|
|
|
|
if (threshold < chunk_npages)
|
|
|
|
threshold = chunk_npages;
|
|
|
|
/*
|
|
|
|
* Don't purge unless the number of purgeable pages exceeds the
|
|
|
|
* threshold.
|
|
|
|
*/
|
|
|
|
if (arena->ndirty <= threshold)
|
|
|
|
return;
|
|
|
|
arena_purge(arena, false);
|
|
|
|
}
|
2010-03-15 08:36:10 +08:00
|
|
|
}
|
|
|
|
|
2014-07-22 01:23:36 +08:00
|
|
|
static size_t
|
|
|
|
arena_dirty_count(arena_t *arena)
|
|
|
|
{
|
|
|
|
size_t ndirty = 0;
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t *rdelm;
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *chunkselm;
|
|
|
|
|
2015-03-11 09:15:40 +08:00
|
|
|
for (rdelm = qr_next(&arena->runs_dirty, rd_link),
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm = qr_next(&arena->chunks_cache, cc_link);
|
2015-03-11 09:15:40 +08:00
|
|
|
rdelm != &arena->runs_dirty; rdelm = qr_next(rdelm, rd_link)) {
|
2015-02-16 10:04:46 +08:00
|
|
|
size_t npages;
|
|
|
|
|
2015-03-11 09:29:49 +08:00
|
|
|
if (rdelm == &chunkselm->rd) {
|
2015-02-16 10:04:46 +08:00
|
|
|
npages = extent_node_size_get(chunkselm) >> LG_PAGE;
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm = qr_next(chunkselm, cc_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
} else {
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(
|
|
|
|
rdelm);
|
|
|
|
arena_chunk_map_misc_t *miscelm =
|
|
|
|
arena_rd_to_miscelm(rdelm);
|
|
|
|
size_t pageind = arena_miscelm_to_pageind(miscelm);
|
2015-02-16 10:04:46 +08:00
|
|
|
assert(arena_mapbits_allocated_get(chunk, pageind) ==
|
|
|
|
0);
|
|
|
|
assert(arena_mapbits_large_get(chunk, pageind) == 0);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, pageind) != 0);
|
|
|
|
npages = arena_mapbits_unallocated_size_get(chunk,
|
|
|
|
pageind) >> LG_PAGE;
|
|
|
|
}
|
2014-07-22 01:23:36 +08:00
|
|
|
ndirty += npages;
|
|
|
|
}
|
|
|
|
|
2014-11-01 17:29:10 +08:00
|
|
|
return (ndirty);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static size_t
|
2014-08-15 05:45:58 +08:00
|
|
|
arena_compute_npurge(arena_t *arena, bool all)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
2014-08-15 05:45:58 +08:00
|
|
|
size_t npurge;
|
2010-03-15 08:36:10 +08:00
|
|
|
|
|
|
|
/*
|
2014-01-15 08:23:03 +08:00
|
|
|
* Compute the minimum number of pages that this thread should try to
|
|
|
|
* purge.
|
2010-03-15 08:36:10 +08:00
|
|
|
*/
|
2014-10-04 01:16:09 +08:00
|
|
|
if (!all) {
|
2015-03-19 09:55:33 +08:00
|
|
|
size_t threshold = (arena->nactive >> arena->lg_dirty_mult);
|
2015-02-04 06:16:55 +08:00
|
|
|
threshold = threshold < chunk_npages ? chunk_npages : threshold;
|
2010-03-15 08:36:10 +08:00
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
npurge = arena->ndirty - threshold;
|
2014-01-15 08:23:03 +08:00
|
|
|
} else
|
2014-08-15 05:45:58 +08:00
|
|
|
npurge = arena->ndirty;
|
2012-10-31 06:42:37 +08:00
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
return (npurge);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
|
|
|
|
2014-07-22 09:09:04 +08:00
|
|
|
static size_t
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena_stash_dirty(arena_t *arena, chunk_hooks_t *chunk_hooks, bool all,
|
|
|
|
size_t npurge, arena_runs_dirty_link_t *purge_runs_sentinel,
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *purge_chunks_sentinel)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t *rdelm, *rdelm_next;
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *chunkselm;
|
2014-07-22 09:09:04 +08:00
|
|
|
size_t nstashed = 0;
|
2012-10-31 06:42:37 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
/* Stash at least npurge pages. */
|
2015-03-11 09:15:40 +08:00
|
|
|
for (rdelm = qr_next(&arena->runs_dirty, rd_link),
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm = qr_next(&arena->chunks_cache, cc_link);
|
2015-03-11 09:15:40 +08:00
|
|
|
rdelm != &arena->runs_dirty; rdelm = rdelm_next) {
|
2015-02-16 10:04:46 +08:00
|
|
|
size_t npages;
|
2015-03-11 09:15:40 +08:00
|
|
|
rdelm_next = qr_next(rdelm, rd_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
2015-03-11 09:29:49 +08:00
|
|
|
if (rdelm == &chunkselm->rd) {
|
2015-02-19 08:40:53 +08:00
|
|
|
extent_node_t *chunkselm_next;
|
|
|
|
bool zero;
|
2015-02-16 10:04:46 +08:00
|
|
|
UNUSED void *chunk;
|
|
|
|
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm_next = qr_next(chunkselm, cc_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
/*
|
2015-02-19 08:40:53 +08:00
|
|
|
* Allocate. chunkselm remains valid due to the
|
|
|
|
* dalloc_node=false argument to chunk_alloc_cache().
|
2015-02-16 10:04:46 +08:00
|
|
|
*/
|
|
|
|
zero = false;
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk = chunk_alloc_cache(arena, chunk_hooks,
|
2015-02-19 08:40:53 +08:00
|
|
|
extent_node_addr_get(chunkselm),
|
|
|
|
extent_node_size_get(chunkselm), chunksize, &zero,
|
|
|
|
false);
|
|
|
|
assert(chunk == extent_node_addr_get(chunkselm));
|
|
|
|
assert(zero == extent_node_zeroed_get(chunkselm));
|
|
|
|
extent_node_dirty_insert(chunkselm, purge_runs_sentinel,
|
2015-02-18 17:15:50 +08:00
|
|
|
purge_chunks_sentinel);
|
2015-02-19 08:40:53 +08:00
|
|
|
npages = extent_node_size_get(chunkselm) >> LG_PAGE;
|
2015-02-16 10:04:46 +08:00
|
|
|
chunkselm = chunkselm_next;
|
|
|
|
} else {
|
|
|
|
arena_chunk_t *chunk =
|
2015-03-11 09:15:40 +08:00
|
|
|
(arena_chunk_t *)CHUNK_ADDR2BASE(rdelm);
|
|
|
|
arena_chunk_map_misc_t *miscelm =
|
|
|
|
arena_rd_to_miscelm(rdelm);
|
|
|
|
size_t pageind = arena_miscelm_to_pageind(miscelm);
|
|
|
|
arena_run_t *run = &miscelm->run;
|
2015-02-16 10:04:46 +08:00
|
|
|
size_t run_size =
|
|
|
|
arena_mapbits_unallocated_size_get(chunk, pageind);
|
2014-07-22 09:09:04 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
npages = run_size >> LG_PAGE;
|
2014-07-22 09:09:04 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
assert(pageind + npages <= chunk_npages);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, pageind) ==
|
|
|
|
arena_mapbits_dirty_get(chunk, pageind+npages-1));
|
2014-08-15 05:45:58 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
/*
|
|
|
|
* If purging the spare chunk's run, make it available
|
|
|
|
* prior to allocation.
|
|
|
|
*/
|
|
|
|
if (chunk == arena->spare)
|
|
|
|
arena_chunk_alloc(arena);
|
|
|
|
|
|
|
|
/* Temporarily allocate the free dirty run. */
|
|
|
|
arena_run_split_large(arena, run, run_size, false);
|
2015-02-18 14:25:56 +08:00
|
|
|
/* Stash. */
|
2015-02-16 10:04:46 +08:00
|
|
|
if (false)
|
2015-03-11 09:15:40 +08:00
|
|
|
qr_new(rdelm, rd_link); /* Redundant. */
|
2015-02-16 10:04:46 +08:00
|
|
|
else {
|
2015-03-11 09:15:40 +08:00
|
|
|
assert(qr_next(rdelm, rd_link) == rdelm);
|
|
|
|
assert(qr_prev(rdelm, rd_link) == rdelm);
|
2015-02-16 10:04:46 +08:00
|
|
|
}
|
2015-03-11 09:15:40 +08:00
|
|
|
qr_meld(purge_runs_sentinel, rdelm, rd_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
}
|
2014-07-22 09:09:04 +08:00
|
|
|
|
|
|
|
nstashed += npages;
|
2014-10-04 01:16:09 +08:00
|
|
|
if (!all && nstashed >= npurge)
|
2014-07-22 09:09:04 +08:00
|
|
|
break;
|
2010-03-15 08:36:10 +08:00
|
|
|
}
|
2014-07-22 09:09:04 +08:00
|
|
|
|
|
|
|
return (nstashed);
|
2014-01-15 08:23:03 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static size_t
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena_purge_stashed(arena_t *arena, chunk_hooks_t *chunk_hooks,
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t *purge_runs_sentinel,
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *purge_chunks_sentinel)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
2014-07-22 09:09:04 +08:00
|
|
|
size_t npurged, nmadvise;
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t *rdelm;
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *chunkselm;
|
2010-03-15 08:36:10 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats)
|
|
|
|
nmadvise = 0;
|
2012-10-31 06:42:37 +08:00
|
|
|
npurged = 0;
|
2014-07-22 09:09:04 +08:00
|
|
|
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2015-03-11 09:15:40 +08:00
|
|
|
for (rdelm = qr_next(purge_runs_sentinel, rd_link),
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm = qr_next(purge_chunks_sentinel, cc_link);
|
2015-03-11 09:15:40 +08:00
|
|
|
rdelm != purge_runs_sentinel; rdelm = qr_next(rdelm, rd_link)) {
|
2015-02-16 10:04:46 +08:00
|
|
|
size_t npages;
|
|
|
|
|
2015-03-11 09:29:49 +08:00
|
|
|
if (rdelm == &chunkselm->rd) {
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
/*
|
|
|
|
* Don't actually purge the chunk here because 1)
|
|
|
|
* chunkselm is embedded in the chunk and must remain
|
|
|
|
* valid, and 2) we deallocate the chunk in
|
|
|
|
* arena_unstash_purged(), where it is destroyed,
|
|
|
|
* decommitted, or purged, depending on chunk
|
|
|
|
* deallocation policy.
|
|
|
|
*/
|
2015-02-16 10:04:46 +08:00
|
|
|
size_t size = extent_node_size_get(chunkselm);
|
|
|
|
npages = size >> LG_PAGE;
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm = qr_next(chunkselm, cc_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
} else {
|
2015-08-11 14:03:34 +08:00
|
|
|
size_t pageind, run_size, flag_unzeroed, flags, i;
|
|
|
|
bool decommitted;
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_chunk_t *chunk =
|
|
|
|
(arena_chunk_t *)CHUNK_ADDR2BASE(rdelm);
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm =
|
|
|
|
arena_rd_to_miscelm(rdelm);
|
|
|
|
pageind = arena_miscelm_to_pageind(miscelm);
|
2015-02-16 10:04:46 +08:00
|
|
|
run_size = arena_mapbits_large_size_get(chunk, pageind);
|
|
|
|
npages = run_size >> LG_PAGE;
|
2014-07-22 09:09:04 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
assert(pageind + npages <= chunk_npages);
|
2015-08-10 07:47:27 +08:00
|
|
|
assert(!arena_mapbits_decommitted_get(chunk, pageind));
|
|
|
|
assert(!arena_mapbits_decommitted_get(chunk,
|
|
|
|
pageind+npages-1));
|
2015-08-05 01:49:46 +08:00
|
|
|
decommitted = !chunk_hooks->decommit(chunk, chunksize,
|
|
|
|
pageind << LG_PAGE, npages << LG_PAGE, arena->ind);
|
|
|
|
if (decommitted) {
|
2015-08-11 14:03:34 +08:00
|
|
|
flag_unzeroed = 0;
|
|
|
|
flags = CHUNK_MAP_DECOMMITTED;
|
2015-08-05 01:49:46 +08:00
|
|
|
} else {
|
2015-08-11 14:03:34 +08:00
|
|
|
flag_unzeroed = chunk_purge_wrapper(arena,
|
2015-08-05 01:49:46 +08:00
|
|
|
chunk_hooks, chunk, chunksize, pageind <<
|
2015-08-11 14:03:34 +08:00
|
|
|
LG_PAGE, run_size) ? CHUNK_MAP_UNZEROED : 0;
|
|
|
|
flags = flag_unzeroed;
|
2015-08-05 01:49:46 +08:00
|
|
|
}
|
2015-08-11 14:03:34 +08:00
|
|
|
arena_mapbits_large_set(chunk, pageind+npages-1, 0,
|
|
|
|
flags);
|
|
|
|
arena_mapbits_large_set(chunk, pageind, run_size,
|
|
|
|
flags);
|
2014-07-22 09:09:04 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
/*
|
2015-08-11 14:03:34 +08:00
|
|
|
* Set the unzeroed flag for internal pages, now that
|
2015-03-19 09:55:33 +08:00
|
|
|
* chunk_purge_wrapper() has returned whether the pages
|
|
|
|
* were zeroed as a side effect of purging. This chunk
|
|
|
|
* map modification is safe even though the arena mutex
|
2015-02-16 10:04:46 +08:00
|
|
|
* isn't currently owned by this thread, because the run
|
|
|
|
* is marked as allocated, thus protecting it from being
|
|
|
|
* modified by any other thread. As long as these
|
|
|
|
* writes don't perturb the first and last elements'
|
|
|
|
* CHUNK_MAP_ALLOCATED bits, behavior is well defined.
|
|
|
|
*/
|
2015-08-11 14:03:34 +08:00
|
|
|
for (i = 1; i < npages-1; i++) {
|
|
|
|
arena_mapbits_internal_set(chunk, pageind+i,
|
2015-02-16 10:04:46 +08:00
|
|
|
flag_unzeroed);
|
|
|
|
}
|
2012-10-09 08:56:11 +08:00
|
|
|
}
|
2014-07-22 09:09:04 +08:00
|
|
|
|
2012-10-31 06:42:37 +08:00
|
|
|
npurged += npages;
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats)
|
|
|
|
nmadvise++;
|
2010-03-15 08:36:10 +08:00
|
|
|
}
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
2014-07-22 09:09:04 +08:00
|
|
|
|
|
|
|
if (config_stats) {
|
2012-02-11 12:22:09 +08:00
|
|
|
arena->stats.nmadvise += nmadvise;
|
2014-07-22 09:09:04 +08:00
|
|
|
arena->stats.purged += npurged;
|
|
|
|
}
|
2010-03-15 08:36:10 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
return (npurged);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena_unstash_purged(arena_t *arena, chunk_hooks_t *chunk_hooks,
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t *purge_runs_sentinel,
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *purge_chunks_sentinel)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t *rdelm, *rdelm_next;
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *chunkselm;
|
2014-01-15 08:23:03 +08:00
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
/* Deallocate chunks/runs. */
|
2015-03-11 09:15:40 +08:00
|
|
|
for (rdelm = qr_next(purge_runs_sentinel, rd_link),
|
2015-02-18 17:15:50 +08:00
|
|
|
chunkselm = qr_next(purge_chunks_sentinel, cc_link);
|
2015-03-11 09:15:40 +08:00
|
|
|
rdelm != purge_runs_sentinel; rdelm = rdelm_next) {
|
|
|
|
rdelm_next = qr_next(rdelm, rd_link);
|
2015-03-11 09:29:49 +08:00
|
|
|
if (rdelm == &chunkselm->rd) {
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t *chunkselm_next = qr_next(chunkselm,
|
2015-02-18 17:15:50 +08:00
|
|
|
cc_link);
|
2015-02-18 14:25:56 +08:00
|
|
|
void *addr = extent_node_addr_get(chunkselm);
|
|
|
|
size_t size = extent_node_size_get(chunkselm);
|
2015-02-18 17:15:50 +08:00
|
|
|
bool zeroed = extent_node_zeroed_get(chunkselm);
|
2015-08-10 07:47:27 +08:00
|
|
|
bool committed = extent_node_committed_get(chunkselm);
|
2015-02-18 17:15:50 +08:00
|
|
|
extent_node_dirty_remove(chunkselm);
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_node_dalloc(arena, chunkselm);
|
|
|
|
chunkselm = chunkselm_next;
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
chunk_dalloc_arena(arena, chunk_hooks, addr, size,
|
2015-08-10 07:47:27 +08:00
|
|
|
zeroed, committed);
|
2015-02-16 10:04:46 +08:00
|
|
|
} else {
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_chunk_t *chunk =
|
|
|
|
(arena_chunk_t *)CHUNK_ADDR2BASE(rdelm);
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm =
|
|
|
|
arena_rd_to_miscelm(rdelm);
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t pageind = arena_miscelm_to_pageind(miscelm);
|
|
|
|
bool decommitted = (arena_mapbits_decommitted_get(chunk,
|
|
|
|
pageind) != 0);
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_run_t *run = &miscelm->run;
|
|
|
|
qr_remove(rdelm, rd_link);
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_run_dalloc(arena, run, false, true, decommitted);
|
2015-02-16 10:04:46 +08:00
|
|
|
}
|
2010-03-15 08:36:10 +08:00
|
|
|
}
|
2012-10-31 06:42:37 +08:00
|
|
|
}
|
|
|
|
|
2015-03-19 09:55:33 +08:00
|
|
|
static void
|
2010-10-01 07:55:08 +08:00
|
|
|
arena_purge(arena_t *arena, bool all)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2015-08-05 01:49:46 +08:00
|
|
|
chunk_hooks_t chunk_hooks = chunk_hooks_get(arena);
|
2014-08-15 05:45:58 +08:00
|
|
|
size_t npurge, npurgeable, npurged;
|
2015-03-11 09:15:40 +08:00
|
|
|
arena_runs_dirty_link_t purge_runs_sentinel;
|
2015-02-16 10:04:46 +08:00
|
|
|
extent_node_t purge_chunks_sentinel;
|
2014-07-22 09:09:04 +08:00
|
|
|
|
2015-06-23 09:50:32 +08:00
|
|
|
arena->purging = true;
|
|
|
|
|
2014-11-01 17:29:10 +08:00
|
|
|
/*
|
|
|
|
* Calls to arena_dirty_count() are disabled even for debug builds
|
|
|
|
* because overhead grows nonlinearly as memory usage increases.
|
|
|
|
*/
|
|
|
|
if (false && config_debug) {
|
2014-07-22 10:39:20 +08:00
|
|
|
size_t ndirty = arena_dirty_count(arena);
|
2014-07-22 01:23:36 +08:00
|
|
|
assert(ndirty == arena->ndirty);
|
2010-03-05 13:35:07 +08:00
|
|
|
}
|
2015-03-19 09:55:33 +08:00
|
|
|
assert((arena->nactive >> arena->lg_dirty_mult) < arena->ndirty || all);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats)
|
|
|
|
arena->stats.npurge++;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
npurge = arena_compute_npurge(arena, all);
|
2015-02-16 10:04:46 +08:00
|
|
|
qr_new(&purge_runs_sentinel, rd_link);
|
2015-02-18 14:23:10 +08:00
|
|
|
extent_node_dirty_linkage_init(&purge_chunks_sentinel);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
npurgeable = arena_stash_dirty(arena, &chunk_hooks, all, npurge,
|
|
|
|
&purge_runs_sentinel, &purge_chunks_sentinel);
|
2014-08-15 05:45:58 +08:00
|
|
|
assert(npurgeable >= npurge);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
npurged = arena_purge_stashed(arena, &chunk_hooks, &purge_runs_sentinel,
|
2015-02-16 10:04:46 +08:00
|
|
|
&purge_chunks_sentinel);
|
2014-07-22 09:09:04 +08:00
|
|
|
assert(npurged == npurgeable);
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena_unstash_purged(arena, &chunk_hooks, &purge_runs_sentinel,
|
2015-02-16 10:04:46 +08:00
|
|
|
&purge_chunks_sentinel);
|
2015-06-23 09:50:32 +08:00
|
|
|
|
|
|
|
arena->purging = false;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2010-10-01 07:55:08 +08:00
|
|
|
void
|
|
|
|
arena_purge_all(arena_t *arena)
|
|
|
|
{
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
arena_purge(arena, true);
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
static void
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_run_coalesce(arena_t *arena, arena_chunk_t *chunk, size_t *p_size,
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t *p_run_ind, size_t *p_run_pages, size_t flag_dirty,
|
|
|
|
size_t flag_decommitted)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-01-15 08:23:03 +08:00
|
|
|
size_t size = *p_size;
|
|
|
|
size_t run_ind = *p_run_ind;
|
|
|
|
size_t run_pages = *p_run_pages;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/* Try to coalesce forward. */
|
|
|
|
if (run_ind + run_pages < chunk_npages &&
|
2012-05-02 15:30:36 +08:00
|
|
|
arena_mapbits_allocated_get(chunk, run_ind+run_pages) == 0 &&
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_mapbits_dirty_get(chunk, run_ind+run_pages) == flag_dirty &&
|
|
|
|
arena_mapbits_decommitted_get(chunk, run_ind+run_pages) ==
|
|
|
|
flag_decommitted) {
|
2012-05-02 15:30:36 +08:00
|
|
|
size_t nrun_size = arena_mapbits_unallocated_size_get(chunk,
|
|
|
|
run_ind+run_pages);
|
2012-04-02 22:04:34 +08:00
|
|
|
size_t nrun_pages = nrun_size >> LG_PAGE;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove successor from runs_avail; the coalesced run is
|
|
|
|
* inserted later.
|
|
|
|
*/
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk,
|
|
|
|
run_ind+run_pages+nrun_pages-1) == nrun_size);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk,
|
|
|
|
run_ind+run_pages+nrun_pages-1) == flag_dirty);
|
2015-08-05 01:49:46 +08:00
|
|
|
assert(arena_mapbits_decommitted_get(chunk,
|
|
|
|
run_ind+run_pages+nrun_pages-1) == flag_decommitted);
|
2014-07-22 10:39:20 +08:00
|
|
|
arena_avail_remove(arena, chunk, run_ind+run_pages, nrun_pages);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
/*
|
|
|
|
* If the successor is dirty, remove it from the set of dirty
|
|
|
|
* pages.
|
|
|
|
*/
|
2014-07-19 05:21:17 +08:00
|
|
|
if (flag_dirty != 0) {
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_run_dirty_remove(arena, chunk, run_ind+run_pages,
|
2014-08-15 05:45:58 +08:00
|
|
|
nrun_pages);
|
2014-07-19 05:21:17 +08:00
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
size += nrun_size;
|
2010-10-18 10:56:09 +08:00
|
|
|
run_pages += nrun_pages;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-05-02 15:30:36 +08:00
|
|
|
arena_mapbits_unallocated_size_set(chunk, run_ind, size);
|
|
|
|
arena_mapbits_unallocated_size_set(chunk, run_ind+run_pages-1,
|
|
|
|
size);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Try to coalesce backward. */
|
2014-01-15 08:23:03 +08:00
|
|
|
if (run_ind > map_bias && arena_mapbits_allocated_get(chunk,
|
|
|
|
run_ind-1) == 0 && arena_mapbits_dirty_get(chunk, run_ind-1) ==
|
2015-08-05 01:49:46 +08:00
|
|
|
flag_dirty && arena_mapbits_decommitted_get(chunk, run_ind-1) ==
|
|
|
|
flag_decommitted) {
|
2012-05-02 15:30:36 +08:00
|
|
|
size_t prun_size = arena_mapbits_unallocated_size_get(chunk,
|
|
|
|
run_ind-1);
|
2012-04-02 22:04:34 +08:00
|
|
|
size_t prun_pages = prun_size >> LG_PAGE;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2010-10-18 10:56:09 +08:00
|
|
|
run_ind -= prun_pages;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove predecessor from runs_avail; the coalesced run is
|
|
|
|
* inserted later.
|
|
|
|
*/
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk, run_ind) ==
|
|
|
|
prun_size);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, run_ind) == flag_dirty);
|
2015-08-05 01:49:46 +08:00
|
|
|
assert(arena_mapbits_decommitted_get(chunk, run_ind) ==
|
|
|
|
flag_decommitted);
|
2014-07-22 10:39:20 +08:00
|
|
|
arena_avail_remove(arena, chunk, run_ind, prun_pages);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-02-16 10:04:46 +08:00
|
|
|
/*
|
|
|
|
* If the predecessor is dirty, remove it from the set of dirty
|
|
|
|
* pages.
|
|
|
|
*/
|
|
|
|
if (flag_dirty != 0) {
|
|
|
|
arena_run_dirty_remove(arena, chunk, run_ind,
|
|
|
|
prun_pages);
|
|
|
|
}
|
2014-07-19 05:21:17 +08:00
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
size += prun_size;
|
2010-10-18 10:56:09 +08:00
|
|
|
run_pages += prun_pages;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-05-02 15:30:36 +08:00
|
|
|
arena_mapbits_unallocated_size_set(chunk, run_ind, size);
|
|
|
|
arena_mapbits_unallocated_size_set(chunk, run_ind+run_pages-1,
|
|
|
|
size);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
*p_size = size;
|
|
|
|
*p_run_ind = run_ind;
|
|
|
|
*p_run_pages = run_pages;
|
|
|
|
}
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
static size_t
|
|
|
|
arena_run_size_get(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run,
|
|
|
|
size_t run_ind)
|
2014-01-15 08:23:03 +08:00
|
|
|
{
|
2015-08-05 01:49:46 +08:00
|
|
|
size_t size;
|
2014-01-15 08:23:03 +08:00
|
|
|
|
|
|
|
assert(run_ind >= map_bias);
|
|
|
|
assert(run_ind < chunk_npages);
|
2015-08-05 01:49:46 +08:00
|
|
|
|
2014-01-15 08:23:03 +08:00
|
|
|
if (arena_mapbits_large_get(chunk, run_ind) != 0) {
|
|
|
|
size = arena_mapbits_large_size_get(chunk, run_ind);
|
2015-08-05 01:49:46 +08:00
|
|
|
assert(size == PAGE || arena_mapbits_large_size_get(chunk,
|
2014-01-15 08:23:03 +08:00
|
|
|
run_ind+(size>>LG_PAGE)-1) == 0);
|
|
|
|
} else {
|
2014-10-11 14:01:03 +08:00
|
|
|
arena_bin_info_t *bin_info = &arena_bin_info[run->binind];
|
2014-01-15 08:23:03 +08:00
|
|
|
size = bin_info->run_size;
|
|
|
|
}
|
2015-08-05 01:49:46 +08:00
|
|
|
|
|
|
|
return (size);
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool
|
|
|
|
arena_run_decommit(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run)
|
|
|
|
{
|
|
|
|
arena_chunk_map_misc_t *miscelm = arena_run_to_miscelm(run);
|
|
|
|
size_t run_ind = arena_miscelm_to_pageind(miscelm);
|
|
|
|
size_t offset = run_ind << LG_PAGE;
|
|
|
|
size_t length = arena_run_size_get(arena, chunk, run, run_ind);
|
|
|
|
|
|
|
|
return (arena->chunk_hooks.decommit(chunk, chunksize, offset, length,
|
|
|
|
arena->ind));
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_run_dalloc(arena_t *arena, arena_run_t *run, bool dirty, bool cleaned,
|
|
|
|
bool decommitted)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
|
|
|
arena_chunk_map_misc_t *miscelm;
|
|
|
|
size_t size, run_ind, run_pages, flag_dirty, flag_decommitted;
|
|
|
|
|
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run);
|
|
|
|
miscelm = arena_run_to_miscelm(run);
|
|
|
|
run_ind = arena_miscelm_to_pageind(miscelm);
|
|
|
|
assert(run_ind >= map_bias);
|
|
|
|
assert(run_ind < chunk_npages);
|
|
|
|
size = arena_run_size_get(arena, chunk, run, run_ind);
|
2014-01-15 08:23:03 +08:00
|
|
|
run_pages = (size >> LG_PAGE);
|
|
|
|
arena_cactive_update(arena, 0, run_pages);
|
|
|
|
arena->nactive -= run_pages;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The run is dirty if the caller claims to have dirtied it, as well as
|
|
|
|
* if it was already dirty before being allocated and the caller
|
|
|
|
* doesn't claim to have cleaned it.
|
|
|
|
*/
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, run_ind) ==
|
|
|
|
arena_mapbits_dirty_get(chunk, run_ind+run_pages-1));
|
2015-08-05 01:49:46 +08:00
|
|
|
if (!cleaned && !decommitted && arena_mapbits_dirty_get(chunk, run_ind)
|
|
|
|
!= 0)
|
2014-01-15 08:23:03 +08:00
|
|
|
dirty = true;
|
|
|
|
flag_dirty = dirty ? CHUNK_MAP_DIRTY : 0;
|
2015-08-05 01:49:46 +08:00
|
|
|
flag_decommitted = decommitted ? CHUNK_MAP_DECOMMITTED : 0;
|
2014-01-15 08:23:03 +08:00
|
|
|
|
|
|
|
/* Mark pages as unallocated in the chunk map. */
|
2015-08-05 01:49:46 +08:00
|
|
|
if (dirty || decommitted) {
|
|
|
|
size_t flags = flag_dirty | flag_decommitted;
|
|
|
|
arena_mapbits_unallocated_set(chunk, run_ind, size, flags);
|
2014-01-15 08:23:03 +08:00
|
|
|
arena_mapbits_unallocated_set(chunk, run_ind+run_pages-1, size,
|
2015-08-05 01:49:46 +08:00
|
|
|
flags);
|
2014-01-15 08:23:03 +08:00
|
|
|
} else {
|
|
|
|
arena_mapbits_unallocated_set(chunk, run_ind, size,
|
|
|
|
arena_mapbits_unzeroed_get(chunk, run_ind));
|
|
|
|
arena_mapbits_unallocated_set(chunk, run_ind+run_pages-1, size,
|
|
|
|
arena_mapbits_unzeroed_get(chunk, run_ind+run_pages-1));
|
|
|
|
}
|
|
|
|
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_run_coalesce(arena, chunk, &size, &run_ind, &run_pages,
|
|
|
|
flag_dirty, flag_decommitted);
|
2014-01-15 08:23:03 +08:00
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/* Insert into runs_avail, now that coalescing is complete. */
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(arena_mapbits_unallocated_size_get(chunk, run_ind) ==
|
|
|
|
arena_mapbits_unallocated_size_get(chunk, run_ind+run_pages-1));
|
|
|
|
assert(arena_mapbits_dirty_get(chunk, run_ind) ==
|
|
|
|
arena_mapbits_dirty_get(chunk, run_ind+run_pages-1));
|
2015-08-05 01:49:46 +08:00
|
|
|
assert(arena_mapbits_decommitted_get(chunk, run_ind) ==
|
|
|
|
arena_mapbits_decommitted_get(chunk, run_ind+run_pages-1));
|
2014-07-22 10:39:20 +08:00
|
|
|
arena_avail_insert(arena, chunk, run_ind, run_pages);
|
2010-04-14 11:53:21 +08:00
|
|
|
|
2014-08-15 05:45:58 +08:00
|
|
|
if (dirty)
|
2015-02-16 10:04:46 +08:00
|
|
|
arena_run_dirty_insert(arena, chunk, run_ind, run_pages);
|
2014-07-19 05:21:17 +08:00
|
|
|
|
2012-05-02 15:30:36 +08:00
|
|
|
/* Deallocate chunk if it is now completely unused. */
|
2014-10-06 08:54:10 +08:00
|
|
|
if (size == arena_maxrun) {
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(run_ind == map_bias);
|
2014-10-06 08:54:10 +08:00
|
|
|
assert(run_pages == (arena_maxrun >> LG_PAGE));
|
2014-05-16 13:22:27 +08:00
|
|
|
arena_chunk_dalloc(arena, chunk);
|
2012-05-02 15:30:36 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2010-01-28 10:27:09 +08:00
|
|
|
/*
|
2010-04-14 11:53:21 +08:00
|
|
|
* It is okay to do dirty page processing here even if the chunk was
|
2010-01-28 10:27:09 +08:00
|
|
|
* deallocated above, since in that case it is the spare. Waiting
|
|
|
|
* until after possible chunk deallocation to do dirty processing
|
|
|
|
* allows for an old spare to be fully deallocated, thus decreasing the
|
|
|
|
* chances of spuriously crossing the dirty page purging threshold.
|
|
|
|
*/
|
2010-04-14 11:53:21 +08:00
|
|
|
if (dirty)
|
2010-03-15 08:36:10 +08:00
|
|
|
arena_maybe_purge(arena);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
static void
|
|
|
|
arena_run_dalloc_decommit(arena_t *arena, arena_chunk_t *chunk,
|
|
|
|
arena_run_t *run)
|
|
|
|
{
|
|
|
|
bool committed = arena_run_decommit(arena, chunk, run);
|
|
|
|
|
|
|
|
arena_run_dalloc(arena, run, committed, false, !committed);
|
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
static void
|
|
|
|
arena_run_trim_head(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run,
|
|
|
|
size_t oldsize, size_t newsize)
|
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_run_to_miscelm(run);
|
|
|
|
size_t pageind = arena_miscelm_to_pageind(miscelm);
|
2012-04-02 22:04:34 +08:00
|
|
|
size_t head_npages = (oldsize - newsize) >> LG_PAGE;
|
2012-05-02 15:30:36 +08:00
|
|
|
size_t flag_dirty = arena_mapbits_dirty_get(chunk, pageind);
|
2015-08-12 03:42:33 +08:00
|
|
|
size_t flag_decommitted = arena_mapbits_decommitted_get(chunk, pageind);
|
|
|
|
size_t flag_unzeroed_mask = (flag_dirty | flag_decommitted) == 0 ?
|
|
|
|
CHUNK_MAP_UNZEROED : 0;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
assert(oldsize > newsize);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the chunk map so that arena_run_dalloc() can treat the
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
* leading run as separately allocated. Set the last element of each
|
|
|
|
* run first, in case of single-page runs.
|
2010-01-17 01:53:50 +08:00
|
|
|
*/
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(arena_mapbits_large_size_get(chunk, pageind) == oldsize);
|
2015-08-12 03:42:33 +08:00
|
|
|
arena_mapbits_large_set(chunk, pageind+head_npages-1, 0, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk,
|
|
|
|
pageind+head_npages-1)));
|
|
|
|
arena_mapbits_large_set(chunk, pageind, oldsize-newsize, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk, pageind)));
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_debug) {
|
2012-04-02 22:04:34 +08:00
|
|
|
UNUSED size_t tail_npages = newsize >> LG_PAGE;
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(arena_mapbits_large_size_get(chunk,
|
|
|
|
pageind+head_npages+tail_npages-1) == 0);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk,
|
|
|
|
pageind+head_npages+tail_npages-1) == flag_dirty);
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
}
|
2012-05-11 11:59:39 +08:00
|
|
|
arena_mapbits_large_set(chunk, pageind+head_npages, newsize,
|
2015-08-12 03:42:33 +08:00
|
|
|
flag_dirty | (flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk,
|
|
|
|
pageind+head_npages)));
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-08-12 03:42:33 +08:00
|
|
|
arena_run_dalloc(arena, run, false, false, (flag_decommitted != 0));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_run_trim_tail(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run,
|
|
|
|
size_t oldsize, size_t newsize, bool dirty)
|
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_run_to_miscelm(run);
|
|
|
|
size_t pageind = arena_miscelm_to_pageind(miscelm);
|
2012-04-02 22:04:34 +08:00
|
|
|
size_t head_npages = newsize >> LG_PAGE;
|
2012-05-02 15:30:36 +08:00
|
|
|
size_t flag_dirty = arena_mapbits_dirty_get(chunk, pageind);
|
2015-08-12 03:42:33 +08:00
|
|
|
size_t flag_decommitted = arena_mapbits_decommitted_get(chunk, pageind);
|
|
|
|
size_t flag_unzeroed_mask = (flag_dirty | flag_decommitted) == 0 ?
|
|
|
|
CHUNK_MAP_UNZEROED : 0;
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *tail_miscelm;
|
|
|
|
arena_run_t *tail_run;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
assert(oldsize > newsize);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the chunk map so that arena_run_dalloc() can treat the
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
* trailing run as separately allocated. Set the last element of each
|
|
|
|
* run first, in case of single-page runs.
|
2010-01-17 01:53:50 +08:00
|
|
|
*/
|
2012-05-02 15:30:36 +08:00
|
|
|
assert(arena_mapbits_large_size_get(chunk, pageind) == oldsize);
|
2015-08-12 03:42:33 +08:00
|
|
|
arena_mapbits_large_set(chunk, pageind+head_npages-1, 0, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk,
|
|
|
|
pageind+head_npages-1)));
|
|
|
|
arena_mapbits_large_set(chunk, pageind, newsize, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk, pageind)));
|
2012-05-02 15:30:36 +08:00
|
|
|
|
|
|
|
if (config_debug) {
|
|
|
|
UNUSED size_t tail_npages = (oldsize - newsize) >> LG_PAGE;
|
|
|
|
assert(arena_mapbits_large_size_get(chunk,
|
|
|
|
pageind+head_npages+tail_npages-1) == 0);
|
|
|
|
assert(arena_mapbits_dirty_get(chunk,
|
|
|
|
pageind+head_npages+tail_npages-1) == flag_dirty);
|
|
|
|
}
|
|
|
|
arena_mapbits_large_set(chunk, pageind+head_npages, oldsize-newsize,
|
2015-08-12 03:42:33 +08:00
|
|
|
flag_dirty | (flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk,
|
|
|
|
pageind+head_npages)));
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-09-29 16:31:39 +08:00
|
|
|
tail_miscelm = arena_miscelm_get(chunk, pageind + head_npages);
|
|
|
|
tail_run = &tail_miscelm->run;
|
2015-08-12 03:42:33 +08:00
|
|
|
arena_run_dalloc(arena, tail_run, dirty, false, (flag_decommitted !=
|
|
|
|
0));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
2012-02-14 09:36:52 +08:00
|
|
|
arena_bin_runs_first(arena_bin_t *bin)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_run_tree_first(&bin->runs);
|
2014-09-29 16:31:39 +08:00
|
|
|
if (miscelm != NULL)
|
|
|
|
return (&miscelm->run);
|
2012-02-14 09:36:52 +08:00
|
|
|
|
|
|
|
return (NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_bin_runs_insert(arena_bin_t *bin, arena_run_t *run)
|
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_run_to_miscelm(run);
|
2012-02-14 09:36:52 +08:00
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
assert(arena_run_tree_search(&bin->runs, miscelm) == NULL);
|
2012-02-14 09:36:52 +08:00
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_run_tree_insert(&bin->runs, miscelm);
|
2012-02-14 09:36:52 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_bin_runs_remove(arena_bin_t *bin, arena_run_t *run)
|
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm = arena_run_to_miscelm(run);
|
2012-02-14 09:36:52 +08:00
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
assert(arena_run_tree_search(&bin->runs, miscelm) != NULL);
|
2012-02-14 09:36:52 +08:00
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_run_tree_remove(&bin->runs, miscelm);
|
2012-02-14 09:36:52 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
|
|
|
arena_bin_nonfull_run_tryget(arena_bin_t *bin)
|
|
|
|
{
|
|
|
|
arena_run_t *run = arena_bin_runs_first(bin);
|
|
|
|
if (run != NULL) {
|
|
|
|
arena_bin_runs_remove(bin, run);
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats)
|
|
|
|
bin->stats.reruns++;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2012-02-14 09:36:52 +08:00
|
|
|
return (run);
|
|
|
|
}
|
|
|
|
|
|
|
|
static arena_run_t *
|
|
|
|
arena_bin_nonfull_run_get(arena_t *arena, arena_bin_t *bin)
|
|
|
|
{
|
|
|
|
arena_run_t *run;
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind;
|
2012-02-14 09:36:52 +08:00
|
|
|
arena_bin_info_t *bin_info;
|
|
|
|
|
|
|
|
/* Look for a usable run. */
|
|
|
|
run = arena_bin_nonfull_run_tryget(bin);
|
|
|
|
if (run != NULL)
|
|
|
|
return (run);
|
2010-01-17 01:53:50 +08:00
|
|
|
/* No existing runs have any space available. */
|
|
|
|
|
2011-03-16 04:59:15 +08:00
|
|
|
binind = arena_bin_index(arena, bin);
|
|
|
|
bin_info = &arena_bin_info[binind];
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/* Allocate a new run. */
|
2010-03-15 10:43:56 +08:00
|
|
|
malloc_mutex_unlock(&bin->lock);
|
2010-03-16 13:25:23 +08:00
|
|
|
/******************************/
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
2014-01-15 08:23:03 +08:00
|
|
|
run = arena_run_alloc_small(arena, bin_info->run_size, binind);
|
2010-03-15 10:43:56 +08:00
|
|
|
if (run != NULL) {
|
|
|
|
/* Initialize run internals. */
|
2014-10-11 14:01:03 +08:00
|
|
|
run->binind = binind;
|
2011-03-16 04:59:15 +08:00
|
|
|
run->nfree = bin_info->nregs;
|
2014-09-29 16:31:39 +08:00
|
|
|
bitmap_init(run->bitmap, &bin_info->bitmap_info);
|
2010-03-16 13:25:23 +08:00
|
|
|
}
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
/********************************/
|
|
|
|
malloc_mutex_lock(&bin->lock);
|
|
|
|
if (run != NULL) {
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
|
|
|
bin->stats.nruns++;
|
|
|
|
bin->stats.curruns++;
|
|
|
|
}
|
2010-03-15 10:43:56 +08:00
|
|
|
return (run);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2014-01-15 08:23:03 +08:00
|
|
|
* arena_run_alloc_small() failed, but another thread may have made
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
* sufficient memory available while this one dropped bin->lock above,
|
2010-03-15 10:43:56 +08:00
|
|
|
* so search one more time.
|
|
|
|
*/
|
2012-02-14 09:36:52 +08:00
|
|
|
run = arena_bin_nonfull_run_tryget(bin);
|
|
|
|
if (run != NULL)
|
2010-03-15 10:43:56 +08:00
|
|
|
return (run);
|
|
|
|
|
|
|
|
return (NULL);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2010-03-14 05:41:58 +08:00
|
|
|
/* Re-fill bin->runcur, then call arena_run_reg_alloc(). */
|
2010-01-17 01:53:50 +08:00
|
|
|
static void *
|
|
|
|
arena_bin_malloc_hard(arena_t *arena, arena_bin_t *bin)
|
|
|
|
{
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind;
|
2011-03-16 04:59:15 +08:00
|
|
|
arena_bin_info_t *bin_info;
|
2010-03-15 10:43:56 +08:00
|
|
|
arena_run_t *run;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2011-03-16 04:59:15 +08:00
|
|
|
binind = arena_bin_index(arena, bin);
|
|
|
|
bin_info = &arena_bin_info[binind];
|
2010-03-15 10:43:56 +08:00
|
|
|
bin->runcur = NULL;
|
|
|
|
run = arena_bin_nonfull_run_get(arena, bin);
|
|
|
|
if (bin->runcur != NULL && bin->runcur->nfree > 0) {
|
|
|
|
/*
|
|
|
|
* Another thread updated runcur while this one ran without the
|
|
|
|
* bin lock in arena_bin_nonfull_run_get().
|
|
|
|
*/
|
2015-09-04 18:15:28 +08:00
|
|
|
void *ret;
|
2010-03-15 10:43:56 +08:00
|
|
|
assert(bin->runcur->nfree > 0);
|
2011-03-16 04:59:15 +08:00
|
|
|
ret = arena_run_reg_alloc(bin->runcur, bin_info);
|
2010-03-15 10:43:56 +08:00
|
|
|
if (run != NULL) {
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
arena_chunk_t *chunk;
|
|
|
|
|
|
|
|
/*
|
2014-01-15 08:23:03 +08:00
|
|
|
* arena_run_alloc_small() may have allocated run, or
|
|
|
|
* it may have pulled run from the bin's run tree.
|
|
|
|
* Therefore it is unsafe to make any assumptions about
|
|
|
|
* how run has previously been used, and
|
|
|
|
* arena_bin_lower_run() must be called, as if a region
|
|
|
|
* were just deallocated from the run.
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
*/
|
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run);
|
2011-03-16 04:59:15 +08:00
|
|
|
if (run->nfree == bin_info->nregs)
|
2010-10-18 11:57:30 +08:00
|
|
|
arena_dalloc_bin_run(arena, chunk, run, bin);
|
|
|
|
else
|
|
|
|
arena_bin_lower_run(arena, chunk, run, bin);
|
2010-03-15 10:43:56 +08:00
|
|
|
}
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (run == NULL)
|
2010-01-17 01:53:50 +08:00
|
|
|
return (NULL);
|
2010-03-15 10:43:56 +08:00
|
|
|
|
|
|
|
bin->runcur = run;
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
assert(bin->runcur->nfree > 0);
|
|
|
|
|
2011-03-16 04:59:15 +08:00
|
|
|
return (arena_run_reg_alloc(bin->runcur, bin_info));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2015-08-20 06:21:32 +08:00
|
|
|
arena_tcache_fill_small(arena_t *arena, tcache_bin_t *tbin, szind_t binind,
|
2012-02-11 12:22:09 +08:00
|
|
|
uint64_t prof_accumbytes)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
unsigned i, nfill;
|
|
|
|
arena_bin_t *bin;
|
|
|
|
|
|
|
|
assert(tbin->ncached == 0);
|
|
|
|
|
2013-02-07 03:59:30 +08:00
|
|
|
if (config_prof && arena_prof_accum(arena, prof_accumbytes))
|
|
|
|
prof_idump();
|
2010-03-16 13:25:23 +08:00
|
|
|
bin = &arena->bins[binind];
|
|
|
|
malloc_mutex_lock(&bin->lock);
|
2011-03-21 15:18:17 +08:00
|
|
|
for (i = 0, nfill = (tcache_bin_info[binind].ncached_max >>
|
|
|
|
tbin->lg_fill_div); i < nfill; i++) {
|
2015-09-04 18:15:28 +08:00
|
|
|
arena_run_t *run;
|
|
|
|
void *ptr;
|
2010-01-17 01:53:50 +08:00
|
|
|
if ((run = bin->runcur) != NULL && run->nfree > 0)
|
2011-03-16 04:59:15 +08:00
|
|
|
ptr = arena_run_reg_alloc(run, &arena_bin_info[binind]);
|
2010-01-17 01:53:50 +08:00
|
|
|
else
|
|
|
|
ptr = arena_bin_malloc_hard(arena, bin);
|
2014-10-06 04:05:10 +08:00
|
|
|
if (ptr == NULL) {
|
|
|
|
/*
|
|
|
|
* OOM. tbin->avail isn't yet filled down to its first
|
|
|
|
* element, so the successful allocations (if any) must
|
|
|
|
* be moved to the base of tbin->avail before bailing
|
|
|
|
* out.
|
|
|
|
*/
|
|
|
|
if (i > 0) {
|
|
|
|
memmove(tbin->avail, &tbin->avail[nfill - i],
|
|
|
|
i * sizeof(void *));
|
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
break;
|
2014-10-06 04:05:10 +08:00
|
|
|
}
|
2014-12-09 05:12:41 +08:00
|
|
|
if (config_fill && unlikely(opt_junk_alloc)) {
|
2012-04-06 15:35:09 +08:00
|
|
|
arena_alloc_junk_small(ptr, &arena_bin_info[binind],
|
|
|
|
true);
|
|
|
|
}
|
2011-03-19 01:53:15 +08:00
|
|
|
/* Insert such that low regions get used first. */
|
|
|
|
tbin->avail[nfill - 1 - i] = ptr;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
|
|
|
bin->stats.nmalloc += i;
|
|
|
|
bin->stats.nrequests += tbin->tstats.nrequests;
|
2014-10-13 13:53:59 +08:00
|
|
|
bin->stats.curregs += i;
|
2012-02-11 12:22:09 +08:00
|
|
|
bin->stats.nfills++;
|
|
|
|
tbin->tstats.nrequests = 0;
|
|
|
|
}
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_unlock(&bin->lock);
|
2010-01-17 01:53:50 +08:00
|
|
|
tbin->ncached = i;
|
|
|
|
}
|
|
|
|
|
2012-04-06 15:35:09 +08:00
|
|
|
void
|
|
|
|
arena_alloc_junk_small(void *ptr, arena_bin_info_t *bin_info, bool zero)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (zero) {
|
|
|
|
size_t redzone_size = bin_info->redzone_size;
|
|
|
|
memset((void *)((uintptr_t)ptr - redzone_size), 0xa5,
|
|
|
|
redzone_size);
|
|
|
|
memset((void *)((uintptr_t)ptr + bin_info->reg_size), 0xa5,
|
|
|
|
redzone_size);
|
|
|
|
} else {
|
|
|
|
memset((void *)((uintptr_t)ptr - bin_info->redzone_size), 0xa5,
|
|
|
|
bin_info->reg_interval);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-12-18 07:14:36 +08:00
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_redzone_corruption
|
|
|
|
#define arena_redzone_corruption JEMALLOC_N(arena_redzone_corruption_impl)
|
|
|
|
#endif
|
|
|
|
static void
|
|
|
|
arena_redzone_corruption(void *ptr, size_t usize, bool after,
|
|
|
|
size_t offset, uint8_t byte)
|
|
|
|
{
|
|
|
|
|
2015-07-24 04:56:25 +08:00
|
|
|
malloc_printf("<jemalloc>: Corrupt redzone %zu byte%s %s %p "
|
|
|
|
"(size %zu), byte=%#x\n", offset, (offset == 1) ? "" : "s",
|
2013-12-18 07:14:36 +08:00
|
|
|
after ? "after" : "before", ptr, usize, byte);
|
|
|
|
}
|
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_redzone_corruption
|
2014-01-08 08:47:56 +08:00
|
|
|
#define arena_redzone_corruption JEMALLOC_N(arena_redzone_corruption)
|
|
|
|
arena_redzone_corruption_t *arena_redzone_corruption =
|
|
|
|
JEMALLOC_N(arena_redzone_corruption_impl);
|
2013-12-18 07:14:36 +08:00
|
|
|
#endif
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_redzones_validate(void *ptr, arena_bin_info_t *bin_info, bool reset)
|
2012-04-06 15:35:09 +08:00
|
|
|
{
|
|
|
|
bool error = false;
|
|
|
|
|
2014-12-09 05:12:41 +08:00
|
|
|
if (opt_junk_alloc) {
|
2015-09-04 18:15:28 +08:00
|
|
|
size_t size = bin_info->reg_size;
|
|
|
|
size_t redzone_size = bin_info->redzone_size;
|
|
|
|
size_t i;
|
|
|
|
|
2014-12-09 05:12:41 +08:00
|
|
|
for (i = 1; i <= redzone_size; i++) {
|
|
|
|
uint8_t *byte = (uint8_t *)((uintptr_t)ptr - i);
|
|
|
|
if (*byte != 0xa5) {
|
|
|
|
error = true;
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_redzone_corruption(ptr, size, false, i,
|
|
|
|
*byte);
|
2014-12-09 05:12:41 +08:00
|
|
|
if (reset)
|
|
|
|
*byte = 0xa5;
|
|
|
|
}
|
2012-04-06 15:35:09 +08:00
|
|
|
}
|
2014-12-09 05:12:41 +08:00
|
|
|
for (i = 0; i < redzone_size; i++) {
|
|
|
|
uint8_t *byte = (uint8_t *)((uintptr_t)ptr + size + i);
|
|
|
|
if (*byte != 0xa5) {
|
|
|
|
error = true;
|
2015-08-05 01:49:46 +08:00
|
|
|
arena_redzone_corruption(ptr, size, true, i,
|
|
|
|
*byte);
|
2014-12-09 05:12:41 +08:00
|
|
|
if (reset)
|
|
|
|
*byte = 0xa5;
|
|
|
|
}
|
2012-04-06 15:35:09 +08:00
|
|
|
}
|
|
|
|
}
|
2014-12-09 05:12:41 +08:00
|
|
|
|
2012-04-06 15:35:09 +08:00
|
|
|
if (opt_abort && error)
|
|
|
|
abort();
|
2013-12-18 07:14:36 +08:00
|
|
|
}
|
|
|
|
|
2014-01-08 08:47:56 +08:00
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_dalloc_junk_small
|
|
|
|
#define arena_dalloc_junk_small JEMALLOC_N(arena_dalloc_junk_small_impl)
|
|
|
|
#endif
|
2013-12-18 07:14:36 +08:00
|
|
|
void
|
|
|
|
arena_dalloc_junk_small(void *ptr, arena_bin_info_t *bin_info)
|
|
|
|
{
|
|
|
|
size_t redzone_size = bin_info->redzone_size;
|
2012-04-06 15:35:09 +08:00
|
|
|
|
2013-12-18 07:14:36 +08:00
|
|
|
arena_redzones_validate(ptr, bin_info, false);
|
2012-04-06 15:35:09 +08:00
|
|
|
memset((void *)((uintptr_t)ptr - redzone_size), 0x5a,
|
|
|
|
bin_info->reg_interval);
|
|
|
|
}
|
2014-01-08 08:47:56 +08:00
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_dalloc_junk_small
|
|
|
|
#define arena_dalloc_junk_small JEMALLOC_N(arena_dalloc_junk_small)
|
|
|
|
arena_dalloc_junk_small_t *arena_dalloc_junk_small =
|
|
|
|
JEMALLOC_N(arena_dalloc_junk_small_impl);
|
|
|
|
#endif
|
2012-04-06 15:35:09 +08:00
|
|
|
|
2013-12-18 07:14:36 +08:00
|
|
|
void
|
|
|
|
arena_quarantine_junk_small(void *ptr, size_t usize)
|
|
|
|
{
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind;
|
2013-12-18 07:14:36 +08:00
|
|
|
arena_bin_info_t *bin_info;
|
|
|
|
cassert(config_fill);
|
2014-12-09 05:12:41 +08:00
|
|
|
assert(opt_junk_free);
|
2013-12-18 07:14:36 +08:00
|
|
|
assert(opt_quarantine);
|
|
|
|
assert(usize <= SMALL_MAXCLASS);
|
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
binind = size2index(usize);
|
2013-12-18 07:14:36 +08:00
|
|
|
bin_info = &arena_bin_info[binind];
|
|
|
|
arena_redzones_validate(ptr, bin_info, true);
|
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
void *
|
|
|
|
arena_malloc_small(arena_t *arena, size_t size, bool zero)
|
|
|
|
{
|
|
|
|
void *ret;
|
|
|
|
arena_bin_t *bin;
|
|
|
|
arena_run_t *run;
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
binind = size2index(size);
|
2012-02-29 08:50:47 +08:00
|
|
|
assert(binind < NBINS);
|
2010-01-17 01:53:50 +08:00
|
|
|
bin = &arena->bins[binind];
|
2014-10-06 08:54:10 +08:00
|
|
|
size = index2size(binind);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_lock(&bin->lock);
|
2010-01-17 01:53:50 +08:00
|
|
|
if ((run = bin->runcur) != NULL && run->nfree > 0)
|
2011-03-16 04:59:15 +08:00
|
|
|
ret = arena_run_reg_alloc(run, &arena_bin_info[binind]);
|
2010-01-17 01:53:50 +08:00
|
|
|
else
|
|
|
|
ret = arena_bin_malloc_hard(arena, bin);
|
|
|
|
|
|
|
|
if (ret == NULL) {
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_unlock(&bin->lock);
|
2010-01-17 01:53:50 +08:00
|
|
|
return (NULL);
|
|
|
|
}
|
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
|
|
|
bin->stats.nmalloc++;
|
|
|
|
bin->stats.nrequests++;
|
2014-10-13 13:53:59 +08:00
|
|
|
bin->stats.curregs++;
|
2012-02-11 12:22:09 +08:00
|
|
|
}
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_unlock(&bin->lock);
|
2014-10-04 01:16:09 +08:00
|
|
|
if (config_prof && !isthreaded && arena_prof_accum(arena, size))
|
2013-02-07 03:59:30 +08:00
|
|
|
prof_idump();
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-10-04 01:16:09 +08:00
|
|
|
if (!zero) {
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_fill) {
|
2014-12-09 05:12:41 +08:00
|
|
|
if (unlikely(opt_junk_alloc)) {
|
2012-04-06 15:35:09 +08:00
|
|
|
arena_alloc_junk_small(ret,
|
|
|
|
&arena_bin_info[binind], false);
|
2014-09-12 07:20:44 +08:00
|
|
|
} else if (unlikely(opt_zero))
|
2012-02-11 12:22:09 +08:00
|
|
|
memset(ret, 0, size);
|
|
|
|
}
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED(ret, size);
|
2012-04-06 15:35:09 +08:00
|
|
|
} else {
|
2014-12-09 05:12:41 +08:00
|
|
|
if (config_fill && unlikely(opt_junk_alloc)) {
|
2012-04-06 15:35:09 +08:00
|
|
|
arena_alloc_junk_small(ret, &arena_bin_info[binind],
|
|
|
|
true);
|
|
|
|
}
|
2014-04-16 07:35:08 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED(ret, size);
|
2010-01-17 01:53:50 +08:00
|
|
|
memset(ret, 0, size);
|
2012-04-06 15:35:09 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
void *
|
|
|
|
arena_malloc_large(arena_t *arena, size_t size, bool zero)
|
|
|
|
{
|
|
|
|
void *ret;
|
2014-10-06 08:54:10 +08:00
|
|
|
size_t usize;
|
2015-05-05 00:58:36 +08:00
|
|
|
uintptr_t random_offset;
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_run_t *run;
|
|
|
|
arena_chunk_map_misc_t *miscelm;
|
2013-02-07 03:59:30 +08:00
|
|
|
UNUSED bool idump;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/* Large allocation. */
|
2014-10-06 08:54:10 +08:00
|
|
|
usize = s2u(size);
|
2010-01-17 01:53:50 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-05-05 00:58:36 +08:00
|
|
|
if (config_cache_oblivious) {
|
2015-07-08 00:32:05 +08:00
|
|
|
uint64_t r;
|
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
/*
|
|
|
|
* Compute a uniformly distributed offset within the first page
|
|
|
|
* that is a multiple of the cacheline size, e.g. [0 .. 63) * 64
|
|
|
|
* for 4 KiB pages and 64-byte cachelines.
|
|
|
|
*/
|
|
|
|
prng64(r, LG_PAGE - LG_CACHELINE, arena->offset_state,
|
2015-06-23 09:50:32 +08:00
|
|
|
UINT64_C(6364136223846793009),
|
|
|
|
UINT64_C(1442695040888963409));
|
2015-05-05 00:58:36 +08:00
|
|
|
random_offset = ((uintptr_t)r) << LG_CACHELINE;
|
|
|
|
} else
|
|
|
|
random_offset = 0;
|
|
|
|
run = arena_run_alloc_large(arena, usize + large_pad, zero);
|
2014-09-29 16:31:39 +08:00
|
|
|
if (run == NULL) {
|
2010-01-17 01:53:50 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
return (NULL);
|
|
|
|
}
|
2014-09-29 16:31:39 +08:00
|
|
|
miscelm = arena_run_to_miscelm(run);
|
2015-05-05 00:58:36 +08:00
|
|
|
ret = (void *)((uintptr_t)arena_miscelm_to_rpages(miscelm) +
|
|
|
|
random_offset);
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - NBINS;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
arena->stats.nmalloc_large++;
|
|
|
|
arena->stats.nrequests_large++;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.allocated_large += usize;
|
|
|
|
arena->stats.lstats[index].nmalloc++;
|
|
|
|
arena->stats.lstats[index].nrequests++;
|
|
|
|
arena->stats.lstats[index].curruns++;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_prof)
|
2014-10-06 08:54:10 +08:00
|
|
|
idump = arena_prof_accum_locked(arena, usize);
|
2010-01-17 01:53:50 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2013-02-07 03:59:30 +08:00
|
|
|
if (config_prof && idump)
|
|
|
|
prof_idump();
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-10-04 01:16:09 +08:00
|
|
|
if (!zero) {
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_fill) {
|
2014-12-09 05:12:41 +08:00
|
|
|
if (unlikely(opt_junk_alloc))
|
2014-10-06 08:54:10 +08:00
|
|
|
memset(ret, 0xa5, usize);
|
2014-09-12 07:20:44 +08:00
|
|
|
else if (unlikely(opt_zero))
|
2014-10-06 08:54:10 +08:00
|
|
|
memset(ret, 0, usize);
|
2012-02-11 12:22:09 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Only handles large allocations that require more than page alignment. */
|
2015-02-13 06:06:37 +08:00
|
|
|
static void *
|
2015-07-24 08:13:18 +08:00
|
|
|
arena_palloc_large(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment,
|
2015-02-13 06:06:37 +08:00
|
|
|
bool zero)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
void *ret;
|
2012-04-12 09:13:45 +08:00
|
|
|
size_t alloc_size, leadsize, trailsize;
|
|
|
|
arena_run_t *run;
|
2010-01-17 01:53:50 +08:00
|
|
|
arena_chunk_t *chunk;
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *miscelm;
|
|
|
|
void *rpages;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-07-24 08:13:18 +08:00
|
|
|
assert(usize == PAGE_CEILING(usize));
|
2010-10-21 08:39:18 +08:00
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
arena = arena_choose(tsd, arena);
|
|
|
|
if (unlikely(arena == NULL))
|
|
|
|
return (NULL);
|
|
|
|
|
2010-10-21 08:39:18 +08:00
|
|
|
alignment = PAGE_CEILING(alignment);
|
2015-07-24 08:13:18 +08:00
|
|
|
alloc_size = usize + large_pad + alignment - PAGE;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
2014-01-15 08:23:03 +08:00
|
|
|
run = arena_run_alloc_large(arena, alloc_size, false);
|
2012-04-12 09:13:45 +08:00
|
|
|
if (run == NULL) {
|
2010-01-17 01:53:50 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
return (NULL);
|
|
|
|
}
|
2012-04-12 09:13:45 +08:00
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run);
|
2014-09-29 16:31:39 +08:00
|
|
|
miscelm = arena_run_to_miscelm(run);
|
|
|
|
rpages = arena_miscelm_to_rpages(miscelm);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-09-29 16:31:39 +08:00
|
|
|
leadsize = ALIGNMENT_CEILING((uintptr_t)rpages, alignment) -
|
|
|
|
(uintptr_t)rpages;
|
2015-07-24 08:13:18 +08:00
|
|
|
assert(alloc_size >= leadsize + usize);
|
|
|
|
trailsize = alloc_size - leadsize - usize - large_pad;
|
2012-04-12 09:13:45 +08:00
|
|
|
if (leadsize != 0) {
|
2014-09-29 16:31:39 +08:00
|
|
|
arena_chunk_map_misc_t *head_miscelm = miscelm;
|
|
|
|
arena_run_t *head_run = run;
|
|
|
|
|
|
|
|
miscelm = arena_miscelm_get(chunk,
|
|
|
|
arena_miscelm_to_pageind(head_miscelm) + (leadsize >>
|
|
|
|
LG_PAGE));
|
|
|
|
run = &miscelm->run;
|
|
|
|
|
|
|
|
arena_run_trim_head(arena, chunk, head_run, alloc_size,
|
|
|
|
alloc_size - leadsize);
|
2012-04-12 09:13:45 +08:00
|
|
|
}
|
|
|
|
if (trailsize != 0) {
|
2015-07-24 08:13:18 +08:00
|
|
|
arena_run_trim_tail(arena, chunk, run, usize + large_pad +
|
|
|
|
trailsize, usize + large_pad, false);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2015-08-05 01:49:46 +08:00
|
|
|
if (arena_run_init_large(arena, run, usize + large_pad, zero)) {
|
|
|
|
size_t run_ind =
|
|
|
|
arena_miscelm_to_pageind(arena_run_to_miscelm(run));
|
2015-08-10 07:47:27 +08:00
|
|
|
bool dirty = (arena_mapbits_dirty_get(chunk, run_ind) != 0);
|
|
|
|
bool decommitted = (arena_mapbits_decommitted_get(chunk,
|
|
|
|
run_ind) != 0);
|
2015-08-05 01:49:46 +08:00
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
assert(decommitted); /* Cause of OOM. */
|
|
|
|
arena_run_dalloc(arena, run, dirty, false, decommitted);
|
2015-08-05 01:49:46 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
return (NULL);
|
|
|
|
}
|
2014-09-29 16:31:39 +08:00
|
|
|
ret = arena_miscelm_to_rpages(miscelm);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - NBINS;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
arena->stats.nmalloc_large++;
|
|
|
|
arena->stats.nrequests_large++;
|
2015-07-24 08:13:18 +08:00
|
|
|
arena->stats.allocated_large += usize;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.lstats[index].nmalloc++;
|
|
|
|
arena->stats.lstats[index].nrequests++;
|
|
|
|
arena->stats.lstats[index].curruns++;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
|
2014-10-04 01:16:09 +08:00
|
|
|
if (config_fill && !zero) {
|
2014-12-09 05:12:41 +08:00
|
|
|
if (unlikely(opt_junk_alloc))
|
2015-07-24 08:13:18 +08:00
|
|
|
memset(ret, 0xa5, usize);
|
2014-09-12 07:20:44 +08:00
|
|
|
else if (unlikely(opt_zero))
|
2015-07-24 08:13:18 +08:00
|
|
|
memset(ret, 0, usize);
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
void *
|
|
|
|
arena_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment,
|
|
|
|
bool zero, tcache_t *tcache)
|
|
|
|
{
|
|
|
|
void *ret;
|
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
if (usize <= SMALL_MAXCLASS && (alignment < PAGE || (alignment == PAGE
|
2015-05-20 08:42:31 +08:00
|
|
|
&& (usize & PAGE_MASK) == 0))) {
|
|
|
|
/* Small; alignment doesn't require special run placement. */
|
2015-02-13 06:06:37 +08:00
|
|
|
ret = arena_malloc(tsd, arena, usize, zero, tcache);
|
2015-09-12 11:50:20 +08:00
|
|
|
} else if (usize <= large_maxclass && alignment <= PAGE) {
|
2015-05-20 08:42:31 +08:00
|
|
|
/*
|
|
|
|
* Large; alignment doesn't require special run placement.
|
|
|
|
* However, the cached pointer may be at a random offset from
|
|
|
|
* the base of the run, so do some bit manipulation to retrieve
|
|
|
|
* the base.
|
|
|
|
*/
|
|
|
|
ret = arena_malloc(tsd, arena, usize, zero, tcache);
|
|
|
|
if (config_cache_oblivious)
|
|
|
|
ret = (void *)((uintptr_t)ret & ~PAGE_MASK);
|
|
|
|
} else {
|
2015-09-12 11:50:20 +08:00
|
|
|
if (likely(usize <= large_maxclass)) {
|
2015-02-13 06:06:37 +08:00
|
|
|
ret = arena_palloc_large(tsd, arena, usize, alignment,
|
|
|
|
zero);
|
|
|
|
} else if (likely(alignment <= chunksize))
|
|
|
|
ret = huge_malloc(tsd, arena, usize, zero, tcache);
|
|
|
|
else {
|
|
|
|
ret = huge_palloc(tsd, arena, usize, alignment, zero,
|
|
|
|
tcache);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2010-04-01 07:45:04 +08:00
|
|
|
void
|
|
|
|
arena_prof_promoted(const void *ptr, size_t size)
|
|
|
|
{
|
|
|
|
arena_chunk_t *chunk;
|
2014-10-06 08:54:10 +08:00
|
|
|
size_t pageind;
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind;
|
2010-04-01 07:45:04 +08:00
|
|
|
|
2012-04-19 04:38:40 +08:00
|
|
|
cassert(config_prof);
|
2010-04-01 07:45:04 +08:00
|
|
|
assert(ptr != NULL);
|
|
|
|
assert(CHUNK_ADDR2BASE(ptr) != ptr);
|
2014-10-06 08:54:10 +08:00
|
|
|
assert(isalloc(ptr, false) == LARGE_MINCLASS);
|
|
|
|
assert(isalloc(ptr, true) == LARGE_MINCLASS);
|
2012-02-29 08:50:47 +08:00
|
|
|
assert(size <= SMALL_MAXCLASS);
|
2010-04-01 07:45:04 +08:00
|
|
|
|
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(ptr);
|
2012-04-02 22:04:34 +08:00
|
|
|
pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE;
|
2014-10-06 08:54:10 +08:00
|
|
|
binind = size2index(size);
|
2012-02-29 08:50:47 +08:00
|
|
|
assert(binind < NBINS);
|
2012-05-02 15:30:36 +08:00
|
|
|
arena_mapbits_large_binind_set(chunk, pageind, binind);
|
2010-04-01 07:45:04 +08:00
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
assert(isalloc(ptr, false) == LARGE_MINCLASS);
|
2012-04-06 15:35:09 +08:00
|
|
|
assert(isalloc(ptr, true) == size);
|
2010-04-01 07:45:04 +08:00
|
|
|
}
|
2010-02-11 02:37:56 +08:00
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
static void
|
2010-10-18 15:04:44 +08:00
|
|
|
arena_dissociate_bin_run(arena_chunk_t *chunk, arena_run_t *run,
|
2010-01-17 01:53:50 +08:00
|
|
|
arena_bin_t *bin)
|
|
|
|
{
|
|
|
|
|
2010-03-19 11:36:40 +08:00
|
|
|
/* Dissociate run from bin. */
|
2010-01-17 01:53:50 +08:00
|
|
|
if (run == bin->runcur)
|
|
|
|
bin->runcur = NULL;
|
2011-03-16 04:59:15 +08:00
|
|
|
else {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind = arena_bin_index(extent_node_arena_get(
|
2015-02-16 10:04:46 +08:00
|
|
|
&chunk->node), bin);
|
2011-03-16 04:59:15 +08:00
|
|
|
arena_bin_info_t *bin_info = &arena_bin_info[binind];
|
|
|
|
|
|
|
|
if (bin_info->nregs != 1) {
|
|
|
|
/*
|
|
|
|
* This block's conditional is necessary because if the
|
|
|
|
* run only contains one region, then it never gets
|
|
|
|
* inserted into the non-full runs tree.
|
|
|
|
*/
|
2012-02-14 09:36:52 +08:00
|
|
|
arena_bin_runs_remove(bin, run);
|
2011-03-16 04:59:15 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2010-10-18 15:04:44 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run,
|
|
|
|
arena_bin_t *bin)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(run != bin->runcur);
|
2014-09-29 16:31:39 +08:00
|
|
|
assert(arena_run_tree_search(&bin->runs, arena_run_to_miscelm(run)) ==
|
|
|
|
NULL);
|
2010-03-14 12:32:56 +08:00
|
|
|
|
2010-03-15 10:43:56 +08:00
|
|
|
malloc_mutex_unlock(&bin->lock);
|
|
|
|
/******************************/
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-08-10 07:47:27 +08:00
|
|
|
arena_run_dalloc_decommit(arena, chunk, run);
|
2010-03-14 12:32:56 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2010-03-15 10:43:56 +08:00
|
|
|
/****************************/
|
|
|
|
malloc_mutex_lock(&bin->lock);
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats)
|
|
|
|
bin->stats.curruns--;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
static void
|
|
|
|
arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run,
|
|
|
|
arena_bin_t *bin)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
|
|
|
|
2010-10-18 11:57:30 +08:00
|
|
|
/*
|
2012-02-14 09:36:52 +08:00
|
|
|
* Make sure that if bin->runcur is non-NULL, it refers to the lowest
|
|
|
|
* non-full run. It is okay to NULL runcur out rather than proactively
|
|
|
|
* keeping it pointing at the lowest non-full run.
|
2010-10-18 11:57:30 +08:00
|
|
|
*/
|
2012-02-14 09:36:52 +08:00
|
|
|
if ((uintptr_t)run < (uintptr_t)bin->runcur) {
|
2010-10-18 11:57:30 +08:00
|
|
|
/* Switch runcur. */
|
2012-02-14 09:36:52 +08:00
|
|
|
if (bin->runcur->nfree > 0)
|
|
|
|
arena_bin_runs_insert(bin, bin->runcur);
|
2010-10-18 11:57:30 +08:00
|
|
|
bin->runcur = run;
|
2012-02-14 09:36:52 +08:00
|
|
|
if (config_stats)
|
|
|
|
bin->stats.reruns++;
|
|
|
|
} else
|
|
|
|
arena_bin_runs_insert(bin, run);
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
}
|
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
static void
|
|
|
|
arena_dalloc_bin_locked_impl(arena_t *arena, arena_chunk_t *chunk, void *ptr,
|
|
|
|
arena_chunk_map_bits_t *bitselm, bool junked)
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
size_t pageind, rpages_ind;
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
arena_run_t *run;
|
|
|
|
arena_bin_t *bin;
|
2012-04-25 05:22:02 +08:00
|
|
|
arena_bin_info_t *bin_info;
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t binind;
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
|
2012-04-02 22:04:34 +08:00
|
|
|
pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE;
|
2014-09-29 16:31:39 +08:00
|
|
|
rpages_ind = pageind - arena_mapbits_small_runind_get(chunk, pageind);
|
|
|
|
run = &arena_miscelm_get(chunk, rpages_ind)->run;
|
2014-10-11 14:01:03 +08:00
|
|
|
binind = run->binind;
|
|
|
|
bin = &arena->bins[binind];
|
2012-04-25 05:22:02 +08:00
|
|
|
bin_info = &arena_bin_info[binind];
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
|
2014-12-09 05:12:41 +08:00
|
|
|
if (!junked && config_fill && unlikely(opt_junk_free))
|
2012-04-06 15:35:09 +08:00
|
|
|
arena_dalloc_junk_small(ptr, bin_info);
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
|
|
|
|
arena_run_reg_dalloc(run, ptr);
|
2011-03-16 04:59:15 +08:00
|
|
|
if (run->nfree == bin_info->nregs) {
|
2010-10-18 15:04:44 +08:00
|
|
|
arena_dissociate_bin_run(chunk, run, bin);
|
2010-10-18 11:57:30 +08:00
|
|
|
arena_dalloc_bin_run(arena, chunk, run, bin);
|
2010-10-18 15:04:44 +08:00
|
|
|
} else if (run->nfree == 1 && run != bin->runcur)
|
2010-10-18 11:57:30 +08:00
|
|
|
arena_bin_lower_run(arena, chunk, run, bin);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
|
|
|
bin->stats.ndalloc++;
|
2014-10-13 13:53:59 +08:00
|
|
|
bin->stats.curregs--;
|
2012-02-11 12:22:09 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
void
|
|
|
|
arena_dalloc_bin_junked_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr,
|
|
|
|
arena_chunk_map_bits_t *bitselm)
|
|
|
|
{
|
|
|
|
|
|
|
|
arena_dalloc_bin_locked_impl(arena, chunk, ptr, bitselm, true);
|
|
|
|
}
|
|
|
|
|
2012-05-02 15:30:36 +08:00
|
|
|
void
|
|
|
|
arena_dalloc_bin(arena_t *arena, arena_chunk_t *chunk, void *ptr,
|
2014-08-30 04:34:40 +08:00
|
|
|
size_t pageind, arena_chunk_map_bits_t *bitselm)
|
2012-05-02 15:30:36 +08:00
|
|
|
{
|
|
|
|
arena_run_t *run;
|
|
|
|
arena_bin_t *bin;
|
2014-09-29 16:31:39 +08:00
|
|
|
size_t rpages_ind;
|
2012-05-02 15:30:36 +08:00
|
|
|
|
2014-09-29 16:31:39 +08:00
|
|
|
rpages_ind = pageind - arena_mapbits_small_runind_get(chunk, pageind);
|
|
|
|
run = &arena_miscelm_get(chunk, rpages_ind)->run;
|
2014-10-11 14:01:03 +08:00
|
|
|
bin = &arena->bins[run->binind];
|
2012-05-02 15:30:36 +08:00
|
|
|
malloc_mutex_lock(&bin->lock);
|
2014-10-10 08:54:06 +08:00
|
|
|
arena_dalloc_bin_locked_impl(arena, chunk, ptr, bitselm, false);
|
2012-05-02 15:30:36 +08:00
|
|
|
malloc_mutex_unlock(&bin->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
arena_dalloc_small(arena_t *arena, arena_chunk_t *chunk, void *ptr,
|
|
|
|
size_t pageind)
|
|
|
|
{
|
2014-08-30 04:34:40 +08:00
|
|
|
arena_chunk_map_bits_t *bitselm;
|
2012-05-02 15:30:36 +08:00
|
|
|
|
|
|
|
if (config_debug) {
|
2012-05-03 07:11:03 +08:00
|
|
|
/* arena_ptr_small_binind_get() does extra sanity checking. */
|
|
|
|
assert(arena_ptr_small_binind_get(ptr, arena_mapbits_get(chunk,
|
|
|
|
pageind)) != BININD_INVALID);
|
2012-05-02 15:30:36 +08:00
|
|
|
}
|
2014-08-30 04:34:40 +08:00
|
|
|
bitselm = arena_bitselm_get(chunk, pageind);
|
|
|
|
arena_dalloc_bin(arena, chunk, ptr, pageind, bitselm);
|
2012-05-02 15:30:36 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-01-08 08:47:56 +08:00
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_dalloc_junk_large
|
|
|
|
#define arena_dalloc_junk_large JEMALLOC_N(arena_dalloc_junk_large_impl)
|
|
|
|
#endif
|
2014-10-10 08:54:06 +08:00
|
|
|
void
|
2014-01-08 08:47:56 +08:00
|
|
|
arena_dalloc_junk_large(void *ptr, size_t usize)
|
|
|
|
{
|
|
|
|
|
2014-12-09 05:12:41 +08:00
|
|
|
if (config_fill && unlikely(opt_junk_free))
|
2014-01-08 08:47:56 +08:00
|
|
|
memset(ptr, 0x5a, usize);
|
|
|
|
}
|
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_dalloc_junk_large
|
|
|
|
#define arena_dalloc_junk_large JEMALLOC_N(arena_dalloc_junk_large)
|
|
|
|
arena_dalloc_junk_large_t *arena_dalloc_junk_large =
|
|
|
|
JEMALLOC_N(arena_dalloc_junk_large_impl);
|
|
|
|
#endif
|
|
|
|
|
2015-09-21 00:58:10 +08:00
|
|
|
static void
|
2014-10-10 08:54:06 +08:00
|
|
|
arena_dalloc_large_locked_impl(arena_t *arena, arena_chunk_t *chunk,
|
|
|
|
void *ptr, bool junked)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE;
|
|
|
|
arena_chunk_map_misc_t *miscelm = arena_miscelm_get(chunk, pageind);
|
|
|
|
arena_run_t *run = &miscelm->run;
|
2010-01-31 19:57:29 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_fill || config_stats) {
|
2015-05-05 00:58:36 +08:00
|
|
|
size_t usize = arena_mapbits_large_size_get(chunk, pageind) -
|
|
|
|
large_pad;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
if (!junked)
|
|
|
|
arena_dalloc_junk_large(ptr, usize);
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t index = size2index(usize) - NBINS;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
arena->stats.ndalloc_large++;
|
2014-01-08 08:47:56 +08:00
|
|
|
arena->stats.allocated_large -= usize;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.lstats[index].ndalloc++;
|
|
|
|
arena->stats.lstats[index].curruns--;
|
2012-02-11 12:22:09 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2015-08-10 07:47:27 +08:00
|
|
|
arena_run_dalloc_decommit(arena, chunk, run);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2014-10-10 08:54:06 +08:00
|
|
|
void
|
|
|
|
arena_dalloc_large_junked_locked(arena_t *arena, arena_chunk_t *chunk,
|
|
|
|
void *ptr)
|
|
|
|
{
|
|
|
|
|
|
|
|
arena_dalloc_large_locked_impl(arena, chunk, ptr, true);
|
|
|
|
}
|
|
|
|
|
2012-05-02 15:30:36 +08:00
|
|
|
void
|
|
|
|
arena_dalloc_large(arena_t *arena, arena_chunk_t *chunk, void *ptr)
|
|
|
|
{
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
2014-10-10 08:54:06 +08:00
|
|
|
arena_dalloc_large_locked_impl(arena, chunk, ptr, false);
|
2012-05-02 15:30:36 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
}
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
static void
|
|
|
|
arena_ralloc_large_shrink(arena_t *arena, arena_chunk_t *chunk, void *ptr,
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
size_t oldsize, size_t size)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2014-09-29 16:31:39 +08:00
|
|
|
size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE;
|
|
|
|
arena_chunk_map_misc_t *miscelm = arena_miscelm_get(chunk, pageind);
|
|
|
|
arena_run_t *run = &miscelm->run;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
assert(size < oldsize);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Shrink the run, and make trailing pages available for other
|
|
|
|
* allocations.
|
|
|
|
*/
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-05-05 00:58:36 +08:00
|
|
|
arena_run_trim_tail(arena, chunk, run, oldsize + large_pad, size +
|
|
|
|
large_pad, true);
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t oldindex = size2index(oldsize) - NBINS;
|
|
|
|
szind_t index = size2index(size) - NBINS;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
arena->stats.ndalloc_large++;
|
|
|
|
arena->stats.allocated_large -= oldsize;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.lstats[oldindex].ndalloc++;
|
|
|
|
arena->stats.lstats[oldindex].curruns--;
|
2012-02-11 12:22:09 +08:00
|
|
|
|
|
|
|
arena->stats.nmalloc_large++;
|
|
|
|
arena->stats.nrequests_large++;
|
|
|
|
arena->stats.allocated_large += size;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.lstats[index].nmalloc++;
|
|
|
|
arena->stats.lstats[index].nrequests++;
|
|
|
|
arena->stats.lstats[index].curruns++;
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool
|
|
|
|
arena_ralloc_large_grow(arena_t *arena, arena_chunk_t *chunk, void *ptr,
|
2015-09-12 07:18:53 +08:00
|
|
|
size_t oldsize, size_t usize_min, size_t usize_max, bool zero)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2012-04-02 22:04:34 +08:00
|
|
|
size_t pageind = ((uintptr_t)ptr - (uintptr_t)chunk) >> LG_PAGE;
|
2015-08-07 14:34:12 +08:00
|
|
|
size_t npages = (oldsize + large_pad) >> LG_PAGE;
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
size_t followsize;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
assert(oldsize == arena_mapbits_large_size_get(chunk, pageind) -
|
|
|
|
large_pad);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/* Try to extend the run. */
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
2015-09-12 07:18:53 +08:00
|
|
|
if (pageind+npages >= chunk_npages || arena_mapbits_allocated_get(chunk,
|
|
|
|
pageind+npages) != 0)
|
|
|
|
goto label_fail;
|
|
|
|
followsize = arena_mapbits_unallocated_size_get(chunk, pageind+npages);
|
|
|
|
if (oldsize + followsize >= usize_min) {
|
2010-01-17 01:53:50 +08:00
|
|
|
/*
|
|
|
|
* The next run is available and sufficiently large. Split the
|
|
|
|
* following run, then merge the first part with the existing
|
|
|
|
* allocation.
|
|
|
|
*/
|
2014-12-18 00:46:35 +08:00
|
|
|
arena_run_t *run;
|
2015-09-12 07:18:53 +08:00
|
|
|
size_t usize, splitsize, size, flag_dirty, flag_unzeroed_mask;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
usize = usize_max;
|
2014-10-06 08:54:10 +08:00
|
|
|
while (oldsize + followsize < usize)
|
|
|
|
usize = index2size(size2index(usize)-1);
|
|
|
|
assert(usize >= usize_min);
|
2015-09-12 07:18:53 +08:00
|
|
|
assert(usize >= oldsize);
|
2015-08-07 14:34:12 +08:00
|
|
|
splitsize = usize - oldsize;
|
2015-09-12 07:18:53 +08:00
|
|
|
if (splitsize == 0)
|
|
|
|
goto label_fail;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2014-12-18 00:46:35 +08:00
|
|
|
run = &arena_miscelm_get(chunk, pageind+npages)->run;
|
2015-09-12 07:18:53 +08:00
|
|
|
if (arena_run_split_large(arena, run, splitsize, zero))
|
|
|
|
goto label_fail;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-09-25 07:38:45 +08:00
|
|
|
if (config_cache_oblivious && zero) {
|
|
|
|
/*
|
|
|
|
* Zero the trailing bytes of the original allocation's
|
|
|
|
* last page, since they are in an indeterminate state.
|
|
|
|
*/
|
|
|
|
assert(PAGE_CEILING(oldsize) == oldsize);
|
|
|
|
memset((void *)((uintptr_t)ptr + oldsize), 0,
|
|
|
|
PAGE_CEILING((uintptr_t)ptr) - (uintptr_t)ptr);
|
|
|
|
}
|
|
|
|
|
2010-10-18 15:04:44 +08:00
|
|
|
size = oldsize + splitsize;
|
2015-08-07 14:34:12 +08:00
|
|
|
npages = (size + large_pad) >> LG_PAGE;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
/*
|
|
|
|
* Mark the extended run as dirty if either portion of the run
|
|
|
|
* was dirty before allocation. This is rather pedantic,
|
|
|
|
* because there's not actually any sequence of events that
|
|
|
|
* could cause the resulting run to be passed to
|
|
|
|
* arena_run_dalloc() with the dirty argument set to false
|
|
|
|
* (which is when dirty flag consistency would really matter).
|
|
|
|
*/
|
2012-05-02 15:30:36 +08:00
|
|
|
flag_dirty = arena_mapbits_dirty_get(chunk, pageind) |
|
|
|
|
arena_mapbits_dirty_get(chunk, pageind+npages-1);
|
2015-08-12 03:42:33 +08:00
|
|
|
flag_unzeroed_mask = flag_dirty == 0 ? CHUNK_MAP_UNZEROED : 0;
|
2015-08-07 14:34:12 +08:00
|
|
|
arena_mapbits_large_set(chunk, pageind, size + large_pad,
|
2015-08-12 03:42:33 +08:00
|
|
|
flag_dirty | (flag_unzeroed_mask &
|
|
|
|
arena_mapbits_unzeroed_get(chunk, pageind)));
|
|
|
|
arena_mapbits_large_set(chunk, pageind+npages-1, 0, flag_dirty |
|
|
|
|
(flag_unzeroed_mask & arena_mapbits_unzeroed_get(chunk,
|
|
|
|
pageind+npages-1)));
|
2010-01-31 19:49:35 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
2015-08-20 06:21:32 +08:00
|
|
|
szind_t oldindex = size2index(oldsize) - NBINS;
|
|
|
|
szind_t index = size2index(size) - NBINS;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
arena->stats.ndalloc_large++;
|
|
|
|
arena->stats.allocated_large -= oldsize;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.lstats[oldindex].ndalloc++;
|
|
|
|
arena->stats.lstats[oldindex].curruns--;
|
2012-02-11 12:22:09 +08:00
|
|
|
|
|
|
|
arena->stats.nmalloc_large++;
|
|
|
|
arena->stats.nrequests_large++;
|
|
|
|
arena->stats.allocated_large += size;
|
2014-10-06 08:54:10 +08:00
|
|
|
arena->stats.lstats[index].nmalloc++;
|
|
|
|
arena->stats.lstats[index].nrequests++;
|
|
|
|
arena->stats.lstats[index].curruns++;
|
Fix numerous arena bugs.
In arena_ralloc_large_grow(), update the map element for the end of the
newly grown run, rather than the interior map element that was the
beginning of the appended run. This is a long-standing bug, and it had
the potential to cause massive corruption, but triggering it required
roughly the following sequence of events:
1) Large in-place growing realloc(), with left-over space in the run
that followed the large object.
2) Allocation of the remainder run left over from (1).
3) Deallocation of the remainder run *before* deallocation of the
large run, with unfortunate interior map state left over from
previous run allocation/deallocation activity, such that one or
more pages of allocated memory would be treated as part of the
remainder run during run coalescing.
In summary, this was a bad bug, but it was difficult to trigger.
In arena_bin_malloc_hard(), if another thread wins the race to allocate
a bin run, dispose of the spare run via arena_bin_lower_run() rather
than arena_run_dalloc(), since the run has already been prepared for use
as a bin run. This bug has existed since March 14, 2010:
e00572b384c81bd2aba57fac32f7077a34388915
mmap()/munmap() without arena->lock or bin->lock.
Fix bugs in arena_dalloc_bin_run(), arena_trim_head(),
arena_trim_tail(), and arena_ralloc_large_grow() that could cause the
CHUNK_MAP_UNZEROED map bit to become corrupted. These are all
long-standing bugs, but the chances of them actually causing problems
was much lower before the CHUNK_MAP_ZEROED --> CHUNK_MAP_UNZEROED
conversion.
Fix a large run statistics regression in arena_ralloc_large_grow() that
was introduced on September 17, 2010:
8e3c3c61b5bb676a705450708e7e79698cdc9e0c
Add {,r,s,d}allocm().
Add debug code to validate that supposedly pre-zeroed memory really is.
2010-10-18 08:51:37 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
return (false);
|
|
|
|
}
|
2015-09-12 07:18:53 +08:00
|
|
|
label_fail:
|
2010-01-17 01:53:50 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
return (true);
|
|
|
|
}
|
|
|
|
|
2014-01-08 08:47:56 +08:00
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_ralloc_junk_large
|
|
|
|
#define arena_ralloc_junk_large JEMALLOC_N(arena_ralloc_junk_large_impl)
|
|
|
|
#endif
|
|
|
|
static void
|
|
|
|
arena_ralloc_junk_large(void *ptr, size_t old_usize, size_t usize)
|
|
|
|
{
|
|
|
|
|
2014-12-09 05:12:41 +08:00
|
|
|
if (config_fill && unlikely(opt_junk_free)) {
|
2014-01-08 08:47:56 +08:00
|
|
|
memset((void *)((uintptr_t)ptr + usize), 0x5a,
|
|
|
|
old_usize - usize);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#ifdef JEMALLOC_JET
|
|
|
|
#undef arena_ralloc_junk_large
|
|
|
|
#define arena_ralloc_junk_large JEMALLOC_N(arena_ralloc_junk_large)
|
|
|
|
arena_ralloc_junk_large_t *arena_ralloc_junk_large =
|
|
|
|
JEMALLOC_N(arena_ralloc_junk_large_impl);
|
|
|
|
#endif
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/*
|
|
|
|
* Try to resize a large allocation, in order to avoid copying. This will
|
|
|
|
* always fail if growing an object, and the following run is already in use.
|
|
|
|
*/
|
|
|
|
static bool
|
2015-09-12 07:18:53 +08:00
|
|
|
arena_ralloc_large(void *ptr, size_t oldsize, size_t usize_min,
|
|
|
|
size_t usize_max, bool zero)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2015-09-12 07:18:53 +08:00
|
|
|
arena_chunk_t *chunk;
|
|
|
|
arena_t *arena;
|
2014-10-06 08:54:10 +08:00
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
if (oldsize == usize_max) {
|
|
|
|
/* Current size class is compatible and maximal. */
|
2010-01-17 01:53:50 +08:00
|
|
|
return (false);
|
2015-09-12 07:18:53 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(ptr);
|
|
|
|
arena = extent_node_arena_get(&chunk->node);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
if (oldsize < usize_max) {
|
|
|
|
bool ret = arena_ralloc_large_grow(arena, chunk, ptr, oldsize,
|
|
|
|
usize_min, usize_max, zero);
|
|
|
|
if (config_fill && !ret && !zero) {
|
|
|
|
if (unlikely(opt_junk_alloc)) {
|
|
|
|
memset((void *)((uintptr_t)ptr + oldsize), 0xa5,
|
|
|
|
isalloc(ptr, config_prof) - oldsize);
|
|
|
|
} else if (unlikely(opt_zero)) {
|
|
|
|
memset((void *)((uintptr_t)ptr + oldsize), 0,
|
|
|
|
isalloc(ptr, config_prof) - oldsize);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
}
|
2015-09-12 07:18:53 +08:00
|
|
|
return (ret);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2015-09-12 07:18:53 +08:00
|
|
|
|
|
|
|
assert(oldsize > usize_max);
|
|
|
|
/* Fill before shrinking in order avoid a race. */
|
|
|
|
arena_ralloc_junk_large(ptr, oldsize, usize_max);
|
|
|
|
arena_ralloc_large_shrink(arena, chunk, ptr, oldsize, usize_max);
|
|
|
|
return (false);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2014-01-13 07:05:44 +08:00
|
|
|
bool
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
arena_ralloc_no_move(void *ptr, size_t oldsize, size_t size, size_t extra,
|
|
|
|
bool zero)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2015-09-12 07:18:53 +08:00
|
|
|
size_t usize_min, usize_max;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
usize_min = s2u(size);
|
|
|
|
usize_max = s2u(size + extra);
|
2015-09-12 11:50:20 +08:00
|
|
|
if (likely(oldsize <= large_maxclass && usize_min <= large_maxclass)) {
|
2015-02-13 06:06:37 +08:00
|
|
|
/*
|
|
|
|
* Avoid moving the allocation if the size class can be left the
|
|
|
|
* same.
|
|
|
|
*/
|
2015-09-12 07:18:53 +08:00
|
|
|
if (oldsize <= SMALL_MAXCLASS) {
|
|
|
|
assert(arena_bin_info[size2index(oldsize)].reg_size ==
|
|
|
|
oldsize);
|
|
|
|
if ((usize_max <= SMALL_MAXCLASS &&
|
|
|
|
size2index(usize_max) == size2index(oldsize)) ||
|
|
|
|
(size <= oldsize && usize_max >= oldsize))
|
|
|
|
return (false);
|
|
|
|
} else {
|
|
|
|
if (usize_max > SMALL_MAXCLASS) {
|
|
|
|
if (!arena_ralloc_large(ptr, oldsize, usize_min,
|
|
|
|
usize_max, zero))
|
2014-01-13 07:05:44 +08:00
|
|
|
return (false);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
/* Reallocation would require a move. */
|
|
|
|
return (true);
|
2015-09-12 07:18:53 +08:00
|
|
|
} else {
|
|
|
|
return (huge_ralloc_no_move(ptr, oldsize, usize_min, usize_max,
|
|
|
|
zero));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void *
|
|
|
|
arena_ralloc_move_helper(tsd_t *tsd, arena_t *arena, size_t usize,
|
|
|
|
size_t alignment, bool zero, tcache_t *tcache)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (alignment == 0)
|
|
|
|
return (arena_malloc(tsd, arena, usize, zero, tcache));
|
|
|
|
usize = sa2u(usize, alignment);
|
|
|
|
if (usize == 0)
|
|
|
|
return (NULL);
|
|
|
|
return (ipalloct(tsd, usize, alignment, zero, tcache, arena));
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void *
|
2014-09-23 12:09:23 +08:00
|
|
|
arena_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t size,
|
2015-09-12 07:18:53 +08:00
|
|
|
size_t alignment, bool zero, tcache_t *tcache)
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
{
|
|
|
|
void *ret;
|
2015-09-12 07:18:53 +08:00
|
|
|
size_t usize;
|
|
|
|
|
|
|
|
usize = s2u(size);
|
|
|
|
if (usize == 0)
|
|
|
|
return (NULL);
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
|
2015-09-12 11:50:20 +08:00
|
|
|
if (likely(usize <= large_maxclass)) {
|
2015-02-13 06:06:37 +08:00
|
|
|
size_t copysize;
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
/* Try to avoid moving the allocation. */
|
2015-09-12 07:18:53 +08:00
|
|
|
if (!arena_ralloc_no_move(ptr, oldsize, usize, 0, zero))
|
2015-02-13 06:06:37 +08:00
|
|
|
return (ptr);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
/*
|
|
|
|
* size and oldsize are different enough that we need to move
|
|
|
|
* the object. In that case, fall back to allocating new space
|
|
|
|
* and copying.
|
|
|
|
*/
|
2015-09-12 07:18:53 +08:00
|
|
|
ret = arena_ralloc_move_helper(tsd, arena, usize, alignment,
|
|
|
|
zero, tcache);
|
|
|
|
if (ret == NULL)
|
|
|
|
return (NULL);
|
Add {,r,s,d}allocm().
Add allocm(), rallocm(), sallocm(), and dallocm(), which are a
functional superset of malloc(), calloc(), posix_memalign(),
malloc_usable_size(), and free().
2010-09-18 06:46:18 +08:00
|
|
|
|
2015-02-13 06:06:37 +08:00
|
|
|
/*
|
|
|
|
* Junk/zero-filling were already done by
|
|
|
|
* ipalloc()/arena_malloc().
|
|
|
|
*/
|
|
|
|
|
2015-09-12 07:18:53 +08:00
|
|
|
copysize = (usize < oldsize) ? usize : oldsize;
|
2015-02-13 06:06:37 +08:00
|
|
|
JEMALLOC_VALGRIND_MAKE_MEM_UNDEFINED(ret, copysize);
|
|
|
|
memcpy(ret, ptr, copysize);
|
|
|
|
isqalloc(tsd, ptr, oldsize, tcache);
|
|
|
|
} else {
|
2015-09-12 07:18:53 +08:00
|
|
|
ret = huge_ralloc(tsd, arena, ptr, oldsize, usize, alignment,
|
|
|
|
zero, tcache);
|
2015-02-13 06:06:37 +08:00
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2012-10-12 04:53:15 +08:00
|
|
|
dss_prec_t
|
|
|
|
arena_dss_prec_get(arena_t *arena)
|
|
|
|
{
|
|
|
|
dss_prec_t ret;
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
ret = arena->dss_prec;
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
return (ret);
|
|
|
|
}
|
|
|
|
|
2014-04-16 03:09:48 +08:00
|
|
|
bool
|
2012-10-12 04:53:15 +08:00
|
|
|
arena_dss_prec_set(arena_t *arena, dss_prec_t dss_prec)
|
|
|
|
{
|
|
|
|
|
2014-10-04 01:16:09 +08:00
|
|
|
if (!have_dss)
|
2014-04-16 03:09:48 +08:00
|
|
|
return (dss_prec != dss_prec_disabled);
|
2012-10-12 04:53:15 +08:00
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
arena->dss_prec = dss_prec;
|
|
|
|
malloc_mutex_unlock(&arena->lock);
|
2014-04-16 03:09:48 +08:00
|
|
|
return (false);
|
2012-10-12 04:53:15 +08:00
|
|
|
}
|
|
|
|
|
2015-03-19 09:55:33 +08:00
|
|
|
ssize_t
|
|
|
|
arena_lg_dirty_mult_default_get(void)
|
|
|
|
{
|
|
|
|
|
|
|
|
return ((ssize_t)atomic_read_z((size_t *)&lg_dirty_mult_default));
|
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
|
|
|
arena_lg_dirty_mult_default_set(ssize_t lg_dirty_mult)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (!arena_lg_dirty_mult_valid(lg_dirty_mult))
|
|
|
|
return (true);
|
|
|
|
atomic_write_z((size_t *)&lg_dirty_mult_default, (size_t)lg_dirty_mult);
|
|
|
|
return (false);
|
|
|
|
}
|
|
|
|
|
2012-10-12 04:53:15 +08:00
|
|
|
void
|
2015-03-25 07:36:12 +08:00
|
|
|
arena_stats_merge(arena_t *arena, const char **dss, ssize_t *lg_dirty_mult,
|
|
|
|
size_t *nactive, size_t *ndirty, arena_stats_t *astats,
|
|
|
|
malloc_bin_stats_t *bstats, malloc_large_stats_t *lstats,
|
|
|
|
malloc_huge_stats_t *hstats)
|
2012-10-12 04:53:15 +08:00
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
malloc_mutex_lock(&arena->lock);
|
|
|
|
*dss = dss_prec_names[arena->dss_prec];
|
2015-03-25 07:36:12 +08:00
|
|
|
*lg_dirty_mult = arena->lg_dirty_mult;
|
2012-10-12 04:53:15 +08:00
|
|
|
*nactive += arena->nactive;
|
|
|
|
*ndirty += arena->ndirty;
|
|
|
|
|
|
|
|
astats->mapped += arena->stats.mapped;
|
|
|
|
astats->npurge += arena->stats.npurge;
|
|
|
|
astats->nmadvise += arena->stats.nmadvise;
|
|
|
|
astats->purged += arena->stats.purged;
|
2014-11-28 03:22:36 +08:00
|
|
|
astats->metadata_mapped += arena->stats.metadata_mapped;
|
|
|
|
astats->metadata_allocated += arena_metadata_allocated_get(arena);
|
2012-10-12 04:53:15 +08:00
|
|
|
astats->allocated_large += arena->stats.allocated_large;
|
|
|
|
astats->nmalloc_large += arena->stats.nmalloc_large;
|
|
|
|
astats->ndalloc_large += arena->stats.ndalloc_large;
|
|
|
|
astats->nrequests_large += arena->stats.nrequests_large;
|
2014-05-16 13:22:27 +08:00
|
|
|
astats->allocated_huge += arena->stats.allocated_huge;
|
|
|
|
astats->nmalloc_huge += arena->stats.nmalloc_huge;
|
|
|
|
astats->ndalloc_huge += arena->stats.ndalloc_huge;
|
2012-10-12 04:53:15 +08:00
|
|
|
|
|
|
|
for (i = 0; i < nlclasses; i++) {
|
|
|
|
lstats[i].nmalloc += arena->stats.lstats[i].nmalloc;
|
|
|
|
lstats[i].ndalloc += arena->stats.lstats[i].ndalloc;
|
|
|
|
lstats[i].nrequests += arena->stats.lstats[i].nrequests;
|
|
|
|
lstats[i].curruns += arena->stats.lstats[i].curruns;
|
|
|
|
}
|
2014-10-13 13:53:59 +08:00
|
|
|
|
|
|
|
for (i = 0; i < nhclasses; i++) {
|
|
|
|
hstats[i].nmalloc += arena->stats.hstats[i].nmalloc;
|
|
|
|
hstats[i].ndalloc += arena->stats.hstats[i].ndalloc;
|
|
|
|
hstats[i].curhchunks += arena->stats.hstats[i].curhchunks;
|
|
|
|
}
|
2012-10-12 04:53:15 +08:00
|
|
|
malloc_mutex_unlock(&arena->lock);
|
|
|
|
|
|
|
|
for (i = 0; i < NBINS; i++) {
|
|
|
|
arena_bin_t *bin = &arena->bins[i];
|
|
|
|
|
|
|
|
malloc_mutex_lock(&bin->lock);
|
|
|
|
bstats[i].nmalloc += bin->stats.nmalloc;
|
|
|
|
bstats[i].ndalloc += bin->stats.ndalloc;
|
|
|
|
bstats[i].nrequests += bin->stats.nrequests;
|
2014-10-13 13:53:59 +08:00
|
|
|
bstats[i].curregs += bin->stats.curregs;
|
2012-10-12 04:53:15 +08:00
|
|
|
if (config_tcache) {
|
|
|
|
bstats[i].nfills += bin->stats.nfills;
|
|
|
|
bstats[i].nflushes += bin->stats.nflushes;
|
|
|
|
}
|
|
|
|
bstats[i].nruns += bin->stats.nruns;
|
|
|
|
bstats[i].reruns += bin->stats.reruns;
|
|
|
|
bstats[i].curruns += bin->stats.curruns;
|
|
|
|
malloc_mutex_unlock(&bin->lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
arena_t *
|
|
|
|
arena_new(unsigned ind)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
arena_t *arena;
|
2010-01-17 01:53:50 +08:00
|
|
|
unsigned i;
|
|
|
|
arena_bin_t *bin;
|
|
|
|
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
/*
|
2014-10-13 13:53:59 +08:00
|
|
|
* Allocate arena, arena->lstats, and arena->hstats contiguously, mainly
|
|
|
|
* because there is no way to clean up if base_alloc() OOMs.
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
*/
|
|
|
|
if (config_stats) {
|
|
|
|
arena = (arena_t *)base_alloc(CACHELINE_CEILING(sizeof(arena_t))
|
2014-10-13 13:53:59 +08:00
|
|
|
+ QUANTUM_CEILING(nlclasses * sizeof(malloc_large_stats_t) +
|
|
|
|
nhclasses) * sizeof(malloc_huge_stats_t));
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
} else
|
|
|
|
arena = (arena_t *)base_alloc(sizeof(arena_t));
|
|
|
|
if (arena == NULL)
|
|
|
|
return (NULL);
|
|
|
|
|
2010-02-11 02:37:56 +08:00
|
|
|
arena->ind = ind;
|
2011-03-19 04:41:33 +08:00
|
|
|
arena->nthreads = 0;
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
if (malloc_mutex_init(&arena->lock))
|
|
|
|
return (NULL);
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats) {
|
|
|
|
memset(&arena->stats, 0, sizeof(arena_stats_t));
|
2014-11-18 02:31:59 +08:00
|
|
|
arena->stats.lstats = (malloc_large_stats_t *)((uintptr_t)arena
|
|
|
|
+ CACHELINE_CEILING(sizeof(arena_t)));
|
2012-02-11 12:22:09 +08:00
|
|
|
memset(arena->stats.lstats, 0, nlclasses *
|
|
|
|
sizeof(malloc_large_stats_t));
|
2014-11-18 02:31:59 +08:00
|
|
|
arena->stats.hstats = (malloc_huge_stats_t *)((uintptr_t)arena
|
|
|
|
+ CACHELINE_CEILING(sizeof(arena_t)) +
|
2014-10-13 13:53:59 +08:00
|
|
|
QUANTUM_CEILING(nlclasses * sizeof(malloc_large_stats_t)));
|
|
|
|
memset(arena->stats.hstats, 0, nhclasses *
|
|
|
|
sizeof(malloc_huge_stats_t));
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_tcache)
|
|
|
|
ql_new(&arena->tcache_ql);
|
|
|
|
}
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_prof)
|
|
|
|
arena->prof_accumbytes = 0;
|
2010-02-12 05:19:21 +08:00
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
if (config_cache_oblivious) {
|
|
|
|
/*
|
|
|
|
* A nondeterministic seed based on the address of arena reduces
|
|
|
|
* the likelihood of lockstep non-uniform cache index
|
|
|
|
* utilization among identical concurrent processes, but at the
|
|
|
|
* cost of test repeatability. For debug builds, instead use a
|
|
|
|
* deterministic seed.
|
|
|
|
*/
|
|
|
|
arena->offset_state = config_debug ? ind :
|
|
|
|
(uint64_t)(uintptr_t)arena;
|
|
|
|
}
|
|
|
|
|
2012-10-12 04:53:15 +08:00
|
|
|
arena->dss_prec = chunk_dss_prec_get();
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
arena->spare = NULL;
|
|
|
|
|
2015-03-19 09:55:33 +08:00
|
|
|
arena->lg_dirty_mult = arena_lg_dirty_mult_default_get();
|
2015-06-23 09:50:32 +08:00
|
|
|
arena->purging = false;
|
2010-01-17 01:53:50 +08:00
|
|
|
arena->nactive = 0;
|
|
|
|
arena->ndirty = 0;
|
|
|
|
|
2012-10-31 06:42:37 +08:00
|
|
|
arena_avail_tree_new(&arena->runs_avail);
|
2015-02-16 10:04:46 +08:00
|
|
|
qr_new(&arena->runs_dirty, rd_link);
|
2015-02-18 17:15:50 +08:00
|
|
|
qr_new(&arena->chunks_cache, cc_link);
|
2015-02-16 10:04:46 +08:00
|
|
|
|
|
|
|
ql_new(&arena->huge);
|
|
|
|
if (malloc_mutex_init(&arena->huge_mtx))
|
|
|
|
return (NULL);
|
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
extent_tree_szad_new(&arena->chunks_szad_cached);
|
|
|
|
extent_tree_ad_new(&arena->chunks_ad_cached);
|
|
|
|
extent_tree_szad_new(&arena->chunks_szad_retained);
|
|
|
|
extent_tree_ad_new(&arena->chunks_ad_retained);
|
2015-02-16 10:04:46 +08:00
|
|
|
if (malloc_mutex_init(&arena->chunks_mtx))
|
|
|
|
return (NULL);
|
|
|
|
ql_new(&arena->node_cache);
|
|
|
|
if (malloc_mutex_init(&arena->node_cache_mtx))
|
|
|
|
return (NULL);
|
|
|
|
|
Generalize chunk management hooks.
Add the "arena.<i>.chunk_hooks" mallctl, which replaces and expands on
the "arena.<i>.chunk.{alloc,dalloc,purge}" mallctls. The chunk hooks
allow control over chunk allocation/deallocation, decommit/commit,
purging, and splitting/merging, such that the application can rely on
jemalloc's internal chunk caching and retaining functionality, yet
implement a variety of chunk management mechanisms and policies.
Merge the chunks_[sz]ad_{mmap,dss} red-black trees into
chunks_[sz]ad_retained. This slightly reduces how hard jemalloc tries
to honor the dss precedence setting; prior to this change the precedence
setting was also consulted when recycling chunks.
Fix chunk purging. Don't purge chunks in arena_purge_stashed(); instead
deallocate them in arena_unstash_purged(), so that the dirty memory
linkage remains valid until after the last time it is used.
This resolves #176 and #201.
2015-07-28 23:28:19 +08:00
|
|
|
arena->chunk_hooks = chunk_hooks_default;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
|
|
|
/* Initialize bins. */
|
2012-02-29 08:50:47 +08:00
|
|
|
for (i = 0; i < NBINS; i++) {
|
2010-01-17 01:53:50 +08:00
|
|
|
bin = &arena->bins[i];
|
2010-03-14 12:32:56 +08:00
|
|
|
if (malloc_mutex_init(&bin->lock))
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
return (NULL);
|
2010-01-17 01:53:50 +08:00
|
|
|
bin->runcur = NULL;
|
|
|
|
arena_run_tree_new(&bin->runs);
|
2012-02-11 12:22:09 +08:00
|
|
|
if (config_stats)
|
|
|
|
memset(&bin->stats, 0, sizeof(malloc_bin_stats_t));
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
Refactor/fix arenas manipulation.
Abstract arenas access to use arena_get() (or a0get() where appropriate)
rather than directly reading e.g. arenas[ind]. Prior to the addition of
the arenas.extend mallctl, the worst possible outcome of directly
accessing arenas was a stale read, but arenas.extend may allocate and
assign a new array to arenas.
Add a tsd-based arenas_cache, which amortizes arenas reads. This
introduces some subtle bootstrapping issues, with tsd_boot() now being
split into tsd_boot[01]() to support tsd wrapper allocation
bootstrapping, as well as an arenas_cache_bypass tsd variable which
dynamically terminates allocation of arenas_cache itself.
Promote a0malloc(), a0calloc(), and a0free() to be generally useful for
internal allocation, and use them in several places (more may be
appropriate).
Abstract arena->nthreads management and fix a missing decrement during
thread destruction (recent tsd refactoring left arenas_cleanup()
unused).
Change arena_choose() to propagate OOM, and handle OOM in all callers.
This is important for providing consistent allocation behavior when the
MALLOCX_ARENA() flag is being used. Prior to this fix, it was possible
for an OOM to result in allocation silently allocating from a different
arena than the one specified.
2014-10-08 14:14:57 +08:00
|
|
|
return (arena);
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
|
|
|
|
2011-03-16 04:59:15 +08:00
|
|
|
/*
|
|
|
|
* Calculate bin_info->run_size such that it meets the following constraints:
|
|
|
|
*
|
2014-10-06 08:54:10 +08:00
|
|
|
* *) bin_info->run_size <= arena_maxrun
|
2011-03-23 00:00:56 +08:00
|
|
|
* *) bin_info->nregs <= RUN_MAXREGS
|
2011-03-16 04:59:15 +08:00
|
|
|
*
|
2014-09-29 16:31:39 +08:00
|
|
|
* bin_info->nregs and bin_info->reg0_offset are also calculated here, since
|
|
|
|
* these settings are all interdependent.
|
2011-03-16 04:59:15 +08:00
|
|
|
*/
|
2014-09-29 16:31:39 +08:00
|
|
|
static void
|
|
|
|
bin_info_run_size_calc(arena_bin_info_t *bin_info)
|
2011-03-16 04:59:15 +08:00
|
|
|
{
|
2012-04-06 15:35:09 +08:00
|
|
|
size_t pad_size;
|
2014-09-29 16:31:39 +08:00
|
|
|
size_t try_run_size, perfect_run_size, actual_run_size;
|
|
|
|
uint32_t try_nregs, perfect_nregs, actual_nregs;
|
2011-03-16 04:59:15 +08:00
|
|
|
|
2012-04-06 15:35:09 +08:00
|
|
|
/*
|
|
|
|
* Determine redzone size based on minimum alignment and minimum
|
|
|
|
* redzone size. Add padding to the end of the run if it is needed to
|
|
|
|
* align the regions. The padding allows each redzone to be half the
|
|
|
|
* minimum alignment; without the padding, each redzone would have to
|
|
|
|
* be twice as large in order to maintain alignment.
|
|
|
|
*/
|
2014-09-12 07:20:44 +08:00
|
|
|
if (config_fill && unlikely(opt_redzone)) {
|
2014-08-30 04:34:40 +08:00
|
|
|
size_t align_min = ZU(1) << (jemalloc_ffs(bin_info->reg_size) -
|
|
|
|
1);
|
2012-04-06 15:35:09 +08:00
|
|
|
if (align_min <= REDZONE_MINSIZE) {
|
|
|
|
bin_info->redzone_size = REDZONE_MINSIZE;
|
|
|
|
pad_size = 0;
|
|
|
|
} else {
|
|
|
|
bin_info->redzone_size = align_min >> 1;
|
|
|
|
pad_size = bin_info->redzone_size;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
bin_info->redzone_size = 0;
|
|
|
|
pad_size = 0;
|
|
|
|
}
|
|
|
|
bin_info->reg_interval = bin_info->reg_size +
|
|
|
|
(bin_info->redzone_size << 1);
|
|
|
|
|
2011-03-16 04:59:15 +08:00
|
|
|
/*
|
2014-09-29 16:31:39 +08:00
|
|
|
* Compute run size under ideal conditions (no redzones, no limit on run
|
|
|
|
* size).
|
2011-03-16 04:59:15 +08:00
|
|
|
*/
|
2014-09-29 16:31:39 +08:00
|
|
|
try_run_size = PAGE;
|
|
|
|
try_nregs = try_run_size / bin_info->reg_size;
|
2011-03-16 04:59:15 +08:00
|
|
|
do {
|
2014-09-29 16:31:39 +08:00
|
|
|
perfect_run_size = try_run_size;
|
|
|
|
perfect_nregs = try_nregs;
|
2011-03-16 04:59:15 +08:00
|
|
|
|
2012-04-02 22:04:34 +08:00
|
|
|
try_run_size += PAGE;
|
2014-09-29 16:31:39 +08:00
|
|
|
try_nregs = try_run_size / bin_info->reg_size;
|
|
|
|
} while (perfect_run_size != perfect_nregs * bin_info->reg_size);
|
|
|
|
assert(perfect_nregs <= RUN_MAXREGS);
|
|
|
|
|
|
|
|
actual_run_size = perfect_run_size;
|
|
|
|
actual_nregs = (actual_run_size - pad_size) / bin_info->reg_interval;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Redzones can require enough padding that not even a single region can
|
|
|
|
* fit within the number of pages that would normally be dedicated to a
|
|
|
|
* run for this size class. Increase the run size until at least one
|
|
|
|
* region fits.
|
|
|
|
*/
|
|
|
|
while (actual_nregs == 0) {
|
|
|
|
assert(config_fill && unlikely(opt_redzone));
|
|
|
|
|
|
|
|
actual_run_size += PAGE;
|
|
|
|
actual_nregs = (actual_run_size - pad_size) /
|
|
|
|
bin_info->reg_interval;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Make sure that the run will fit within an arena chunk.
|
|
|
|
*/
|
2014-10-06 08:54:10 +08:00
|
|
|
while (actual_run_size > arena_maxrun) {
|
2014-09-29 16:31:39 +08:00
|
|
|
actual_run_size -= PAGE;
|
|
|
|
actual_nregs = (actual_run_size - pad_size) /
|
|
|
|
bin_info->reg_interval;
|
|
|
|
}
|
|
|
|
assert(actual_nregs > 0);
|
2015-03-07 09:14:05 +08:00
|
|
|
assert(actual_run_size == s2u(actual_run_size));
|
2011-03-16 04:59:15 +08:00
|
|
|
|
|
|
|
/* Copy final settings. */
|
2014-09-29 16:31:39 +08:00
|
|
|
bin_info->run_size = actual_run_size;
|
|
|
|
bin_info->nregs = actual_nregs;
|
|
|
|
bin_info->reg0_offset = actual_run_size - (actual_nregs *
|
|
|
|
bin_info->reg_interval) - pad_size + bin_info->redzone_size;
|
2012-04-06 15:35:09 +08:00
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
if (actual_run_size > small_maxrun)
|
|
|
|
small_maxrun = actual_run_size;
|
|
|
|
|
2012-04-06 15:35:09 +08:00
|
|
|
assert(bin_info->reg0_offset - bin_info->redzone_size + (bin_info->nregs
|
|
|
|
* bin_info->reg_interval) + pad_size == bin_info->run_size);
|
2011-03-16 04:59:15 +08:00
|
|
|
}
|
|
|
|
|
2012-02-29 08:50:47 +08:00
|
|
|
static void
|
2011-03-16 04:59:15 +08:00
|
|
|
bin_info_init(void)
|
|
|
|
{
|
|
|
|
arena_bin_info_t *bin_info;
|
2012-02-29 08:50:47 +08:00
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
#define BIN_INFO_INIT_bin_yes(index, size) \
|
2014-05-29 07:11:55 +08:00
|
|
|
bin_info = &arena_bin_info[index]; \
|
2012-02-29 08:50:47 +08:00
|
|
|
bin_info->reg_size = size; \
|
2014-09-29 16:31:39 +08:00
|
|
|
bin_info_run_size_calc(bin_info); \
|
2012-02-29 08:50:47 +08:00
|
|
|
bitmap_info_init(&bin_info->bitmap_info, bin_info->nregs);
|
2014-05-29 07:11:55 +08:00
|
|
|
#define BIN_INFO_INIT_bin_no(index, size)
|
|
|
|
#define SC(index, lg_grp, lg_delta, ndelta, bin, lg_delta_lookup) \
|
|
|
|
BIN_INFO_INIT_bin_##bin(index, (ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta))
|
2012-02-29 08:50:47 +08:00
|
|
|
SIZE_CLASSES
|
2014-05-29 07:11:55 +08:00
|
|
|
#undef BIN_INFO_INIT_bin_yes
|
|
|
|
#undef BIN_INFO_INIT_bin_no
|
|
|
|
#undef SC
|
2011-03-16 04:59:15 +08:00
|
|
|
}
|
|
|
|
|
2015-05-05 00:58:36 +08:00
|
|
|
static bool
|
|
|
|
small_run_size_init(void)
|
|
|
|
{
|
|
|
|
|
|
|
|
assert(small_maxrun != 0);
|
|
|
|
|
|
|
|
small_run_tab = (bool *)base_alloc(sizeof(bool) * (small_maxrun >>
|
|
|
|
LG_PAGE));
|
|
|
|
if (small_run_tab == NULL)
|
|
|
|
return (true);
|
|
|
|
|
|
|
|
#define TAB_INIT_bin_yes(index, size) { \
|
|
|
|
arena_bin_info_t *bin_info = &arena_bin_info[index]; \
|
|
|
|
small_run_tab[bin_info->run_size >> LG_PAGE] = true; \
|
|
|
|
}
|
|
|
|
#define TAB_INIT_bin_no(index, size)
|
|
|
|
#define SC(index, lg_grp, lg_delta, ndelta, bin, lg_delta_lookup) \
|
|
|
|
TAB_INIT_bin_##bin(index, (ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta))
|
|
|
|
SIZE_CLASSES
|
|
|
|
#undef TAB_INIT_bin_yes
|
|
|
|
#undef TAB_INIT_bin_no
|
|
|
|
#undef SC
|
|
|
|
|
|
|
|
return (false);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
2010-01-30 06:30:41 +08:00
|
|
|
arena_boot(void)
|
2010-01-17 01:53:50 +08:00
|
|
|
{
|
2010-10-02 08:35:43 +08:00
|
|
|
unsigned i;
|
2010-01-17 01:53:50 +08:00
|
|
|
|
2015-03-19 09:55:33 +08:00
|
|
|
arena_lg_dirty_mult_default_set(opt_lg_dirty_mult);
|
|
|
|
|
2010-01-17 01:53:50 +08:00
|
|
|
/*
|
|
|
|
* Compute the header size such that it is large enough to contain the
|
2010-10-02 08:35:43 +08:00
|
|
|
* page map. The page map is biased to omit entries for the header
|
|
|
|
* itself, so some iteration is necessary to compute the map bias.
|
|
|
|
*
|
|
|
|
* 1) Compute safe header_size and map_bias values that include enough
|
|
|
|
* space for an unbiased page map.
|
|
|
|
* 2) Refine map_bias based on (1) to omit the header pages in the page
|
|
|
|
* map. The resulting map_bias may be one too small.
|
|
|
|
* 3) Refine map_bias based on (2). The result will be >= the result
|
|
|
|
* from (2), and will always be correct.
|
2010-01-17 01:53:50 +08:00
|
|
|
*/
|
2010-10-02 08:35:43 +08:00
|
|
|
map_bias = 0;
|
|
|
|
for (i = 0; i < 3; i++) {
|
2015-09-04 18:15:28 +08:00
|
|
|
size_t header_size = offsetof(arena_chunk_t, map_bits) +
|
2014-08-30 04:34:40 +08:00
|
|
|
((sizeof(arena_chunk_map_bits_t) +
|
|
|
|
sizeof(arena_chunk_map_misc_t)) * (chunk_npages-map_bias));
|
2014-09-29 16:31:39 +08:00
|
|
|
map_bias = (header_size + PAGE_MASK) >> LG_PAGE;
|
2010-10-02 08:35:43 +08:00
|
|
|
}
|
|
|
|
assert(map_bias > 0);
|
|
|
|
|
2014-08-30 04:34:40 +08:00
|
|
|
map_misc_offset = offsetof(arena_chunk_t, map_bits) +
|
|
|
|
sizeof(arena_chunk_map_bits_t) * (chunk_npages-map_bias);
|
|
|
|
|
2014-10-06 08:54:10 +08:00
|
|
|
arena_maxrun = chunksize - (map_bias << LG_PAGE);
|
2014-10-10 08:54:06 +08:00
|
|
|
assert(arena_maxrun > 0);
|
2015-09-12 11:50:20 +08:00
|
|
|
large_maxclass = index2size(size2index(chunksize)-1);
|
|
|
|
if (large_maxclass > arena_maxrun) {
|
2014-10-06 08:54:10 +08:00
|
|
|
/*
|
|
|
|
* For small chunk sizes it's possible for there to be fewer
|
|
|
|
* non-header pages available than are necessary to serve the
|
|
|
|
* size classes just below chunksize.
|
|
|
|
*/
|
2015-09-12 11:50:20 +08:00
|
|
|
large_maxclass = arena_maxrun;
|
2014-10-06 08:54:10 +08:00
|
|
|
}
|
2015-09-12 11:50:20 +08:00
|
|
|
assert(large_maxclass > 0);
|
|
|
|
nlclasses = size2index(large_maxclass) - size2index(SMALL_MAXCLASS);
|
2014-10-13 13:53:59 +08:00
|
|
|
nhclasses = NSIZES - nlclasses - NBINS;
|
2010-01-30 06:30:41 +08:00
|
|
|
|
2012-02-29 08:50:47 +08:00
|
|
|
bin_info_init();
|
2015-05-05 00:58:36 +08:00
|
|
|
return (small_run_size_init());
|
2010-01-17 01:53:50 +08:00
|
|
|
}
|
2012-03-14 07:31:41 +08:00
|
|
|
|
|
|
|
void
|
|
|
|
arena_prefork(arena_t *arena)
|
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
malloc_mutex_prefork(&arena->lock);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
malloc_mutex_prefork(&arena->huge_mtx);
|
|
|
|
malloc_mutex_prefork(&arena->chunks_mtx);
|
|
|
|
malloc_mutex_prefork(&arena->node_cache_mtx);
|
2012-03-14 07:31:41 +08:00
|
|
|
for (i = 0; i < NBINS; i++)
|
|
|
|
malloc_mutex_prefork(&arena->bins[i].lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
arena_postfork_parent(arena_t *arena)
|
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
for (i = 0; i < NBINS; i++)
|
|
|
|
malloc_mutex_postfork_parent(&arena->bins[i].lock);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
malloc_mutex_postfork_parent(&arena->node_cache_mtx);
|
|
|
|
malloc_mutex_postfork_parent(&arena->chunks_mtx);
|
|
|
|
malloc_mutex_postfork_parent(&arena->huge_mtx);
|
2012-03-14 07:31:41 +08:00
|
|
|
malloc_mutex_postfork_parent(&arena->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
arena_postfork_child(arena_t *arena)
|
|
|
|
{
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
for (i = 0; i < NBINS; i++)
|
|
|
|
malloc_mutex_postfork_child(&arena->bins[i].lock);
|
Move centralized chunk management into arenas.
Migrate all centralized data structures related to huge allocations and
recyclable chunks into arena_t, so that each arena can manage huge
allocations and recyclable virtual memory completely independently of
other arenas.
Add chunk node caching to arenas, in order to avoid contention on the
base allocator.
Use chunks_rtree to look up huge allocations rather than a red-black
tree. Maintain a per arena unsorted list of huge allocations (which
will be needed to enumerate huge allocations during arena reset).
Remove the --enable-ivsalloc option, make ivsalloc() always available,
and use it for size queries if --enable-debug is enabled. The only
practical implications to this removal are that 1) ivsalloc() is now
always available during live debugging (and the underlying radix tree is
available during core-based debugging), and 2) size query validation can
no longer be enabled independent of --enable-debug.
Remove the stats.chunks.{current,total,high} mallctls, and replace their
underlying statistics with simpler atomically updated counters used
exclusively for gdump triggering. These statistics are no longer very
useful because each arena manages chunks independently, and per arena
statistics provide similar information.
Simplify chunk synchronization code, now that base chunk allocation
cannot cause recursive lock acquisition.
2015-02-12 04:24:27 +08:00
|
|
|
malloc_mutex_postfork_child(&arena->node_cache_mtx);
|
|
|
|
malloc_mutex_postfork_child(&arena->chunks_mtx);
|
|
|
|
malloc_mutex_postfork_child(&arena->huge_mtx);
|
2012-03-14 07:31:41 +08:00
|
|
|
malloc_mutex_postfork_child(&arena->lock);
|
|
|
|
}
|