Fix a run initialization race condition.
Initialize small run header before dropping arena->lock, arena_chunk_purge() relies on valid small run headers during run iteration. Add some assertions.
This commit is contained in:
parent
f00bb7f132
commit
e69bee01de
@ -548,6 +548,7 @@ arena_chunk_purge(arena_t *arena, arena_chunk_t *chunk)
|
|||||||
|
|
||||||
npages = (mapelm->bits & CHUNK_MAP_PG_MASK) >>
|
npages = (mapelm->bits & CHUNK_MAP_PG_MASK) >>
|
||||||
CHUNK_MAP_PG_SHIFT;
|
CHUNK_MAP_PG_SHIFT;
|
||||||
|
assert(pageind + npages <= chunk_npages);
|
||||||
for (i = 0; i < npages; i++) {
|
for (i = 0; i < npages; i++) {
|
||||||
if (chunk->map[pageind + i].bits &
|
if (chunk->map[pageind + i].bits &
|
||||||
CHUNK_MAP_DIRTY) {
|
CHUNK_MAP_DIRTY) {
|
||||||
@ -588,13 +589,15 @@ arena_chunk_purge(arena_t *arena, arena_chunk_t *chunk)
|
|||||||
>> CHUNK_MAP_PG_SHIFT;
|
>> CHUNK_MAP_PG_SHIFT;
|
||||||
} else {
|
} else {
|
||||||
arena_run_t *run = (arena_run_t *)((uintptr_t)
|
arena_run_t *run = (arena_run_t *)((uintptr_t)
|
||||||
chunk + (uintptr_t)((pageind -
|
chunk + (uintptr_t)(pageind << PAGE_SHIFT));
|
||||||
((mapelm->bits & CHUNK_MAP_PG_MASK) >>
|
|
||||||
CHUNK_MAP_PG_SHIFT)) << PAGE_SHIFT));
|
assert((mapelm->bits & CHUNK_MAP_PG_MASK) == 0);
|
||||||
|
assert(run->magic == ARENA_RUN_MAGIC);
|
||||||
pageind += run->bin->run_size >> PAGE_SHIFT;
|
pageind += run->bin->run_size >> PAGE_SHIFT;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
assert(pageind == chunk_npages);
|
||||||
|
|
||||||
#ifdef JEMALLOC_DEBUG
|
#ifdef JEMALLOC_DEBUG
|
||||||
ndirty = chunk->ndirty;
|
ndirty = chunk->ndirty;
|
||||||
@ -618,6 +621,7 @@ arena_chunk_purge(arena_t *arena, arena_chunk_t *chunk)
|
|||||||
size_t npages = (mapelm->bits & CHUNK_MAP_PG_MASK) >>
|
size_t npages = (mapelm->bits & CHUNK_MAP_PG_MASK) >>
|
||||||
CHUNK_MAP_PG_SHIFT;
|
CHUNK_MAP_PG_SHIFT;
|
||||||
|
|
||||||
|
assert(pageind + npages <= chunk_npages);
|
||||||
for (i = 0; i < npages;) {
|
for (i = 0; i < npages;) {
|
||||||
if (chunk->map[pageind + i].bits & CHUNK_MAP_DIRTY) {
|
if (chunk->map[pageind + i].bits & CHUNK_MAP_DIRTY) {
|
||||||
chunk->map[pageind + i].bits ^= CHUNK_MAP_DIRTY;
|
chunk->map[pageind + i].bits ^= CHUNK_MAP_DIRTY;
|
||||||
@ -963,20 +967,24 @@ arena_bin_nonfull_run_get(arena_t *arena, arena_bin_t *bin)
|
|||||||
|
|
||||||
/* Allocate a new run. */
|
/* Allocate a new run. */
|
||||||
malloc_mutex_unlock(&bin->lock);
|
malloc_mutex_unlock(&bin->lock);
|
||||||
|
/******************************/
|
||||||
malloc_mutex_lock(&arena->lock);
|
malloc_mutex_lock(&arena->lock);
|
||||||
run = arena_run_alloc(arena, bin->run_size, false, false);
|
run = arena_run_alloc(arena, bin->run_size, false, false);
|
||||||
malloc_mutex_unlock(&arena->lock);
|
|
||||||
malloc_mutex_lock(&bin->lock);
|
|
||||||
if (run != NULL) {
|
if (run != NULL) {
|
||||||
/* Initialize run internals. */
|
/* Initialize run internals. */
|
||||||
run->bin = bin;
|
run->bin = bin;
|
||||||
run->avail = NULL;
|
run->avail = NULL;
|
||||||
run->next = (void *)(((uintptr_t)run) + (uintptr_t)bin->reg0_offset);
|
run->next = (void *)(((uintptr_t)run) +
|
||||||
|
(uintptr_t)bin->reg0_offset);
|
||||||
run->nfree = bin->nregs;
|
run->nfree = bin->nregs;
|
||||||
#ifdef JEMALLOC_DEBUG
|
#ifdef JEMALLOC_DEBUG
|
||||||
run->magic = ARENA_RUN_MAGIC;
|
run->magic = ARENA_RUN_MAGIC;
|
||||||
#endif
|
#endif
|
||||||
|
}
|
||||||
|
malloc_mutex_unlock(&arena->lock);
|
||||||
|
/********************************/
|
||||||
|
malloc_mutex_lock(&bin->lock);
|
||||||
|
if (run != NULL) {
|
||||||
#ifdef JEMALLOC_STATS
|
#ifdef JEMALLOC_STATS
|
||||||
bin->stats.nruns++;
|
bin->stats.nruns++;
|
||||||
bin->stats.curruns++;
|
bin->stats.curruns++;
|
||||||
@ -1082,13 +1090,13 @@ arena_tcache_fill(arena_t *arena, tcache_bin_t *tbin, size_t binind
|
|||||||
|
|
||||||
assert(tbin->ncached == 0);
|
assert(tbin->ncached == 0);
|
||||||
|
|
||||||
bin = &arena->bins[binind];
|
|
||||||
malloc_mutex_lock(&bin->lock);
|
|
||||||
#ifdef JEMALLOC_PROF
|
#ifdef JEMALLOC_PROF
|
||||||
malloc_mutex_lock(&arena->lock);
|
malloc_mutex_lock(&arena->lock);
|
||||||
arena_prof_accum(arena, prof_accumbytes);
|
arena_prof_accum(arena, prof_accumbytes);
|
||||||
malloc_mutex_unlock(&arena->lock);
|
malloc_mutex_unlock(&arena->lock);
|
||||||
#endif
|
#endif
|
||||||
|
bin = &arena->bins[binind];
|
||||||
|
malloc_mutex_lock(&bin->lock);
|
||||||
for (i = 0, nfill = (tbin->ncached_max >> 1); i < nfill; i++) {
|
for (i = 0, nfill = (tbin->ncached_max >> 1); i < nfill; i++) {
|
||||||
if ((run = bin->runcur) != NULL && run->nfree > 0)
|
if ((run = bin->runcur) != NULL && run->nfree > 0)
|
||||||
ptr = arena_run_reg_alloc(run, bin);
|
ptr = arena_run_reg_alloc(run, bin);
|
||||||
|
@ -62,20 +62,21 @@ tcache_bin_flush(tcache_bin_t *tbin, size_t binind, unsigned rem
|
|||||||
arena_t *arena = chunk->arena;
|
arena_t *arena = chunk->arena;
|
||||||
arena_bin_t *bin = &arena->bins[binind];
|
arena_bin_t *bin = &arena->bins[binind];
|
||||||
|
|
||||||
malloc_mutex_lock(&bin->lock);
|
#ifdef JEMALLOC_PROF
|
||||||
#if (defined(JEMALLOC_STATS) || defined(JEMALLOC_PROF))
|
|
||||||
if (arena == tcache->arena) {
|
if (arena == tcache->arena) {
|
||||||
# ifdef JEMALLOC_PROF
|
|
||||||
malloc_mutex_lock(&arena->lock);
|
malloc_mutex_lock(&arena->lock);
|
||||||
arena_prof_accum(arena, tcache->prof_accumbytes);
|
arena_prof_accum(arena, tcache->prof_accumbytes);
|
||||||
malloc_mutex_unlock(&arena->lock);
|
malloc_mutex_unlock(&arena->lock);
|
||||||
tcache->prof_accumbytes = 0;
|
tcache->prof_accumbytes = 0;
|
||||||
# endif
|
}
|
||||||
# ifdef JEMALLOC_STATS
|
#endif
|
||||||
|
|
||||||
|
malloc_mutex_lock(&bin->lock);
|
||||||
|
#ifdef JEMALLOC_STATS
|
||||||
|
if (arena == tcache->arena) {
|
||||||
bin->stats.nflushes++;
|
bin->stats.nflushes++;
|
||||||
bin->stats.nrequests += tbin->tstats.nrequests;
|
bin->stats.nrequests += tbin->tstats.nrequests;
|
||||||
tbin->tstats.nrequests = 0;
|
tbin->tstats.nrequests = 0;
|
||||||
# endif
|
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
deferred = NULL;
|
deferred = NULL;
|
||||||
|
Loading…
Reference in New Issue
Block a user