2020-01-28 05:55:46 +08:00
|
|
|
#ifndef JEMALLOC_INTERNAL_EMAP_H
|
|
|
|
#define JEMALLOC_INTERNAL_EMAP_H
|
|
|
|
|
|
|
|
#include "jemalloc/internal/mutex_pool.h"
|
|
|
|
#include "jemalloc/internal/rtree.h"
|
|
|
|
|
|
|
|
typedef struct emap_s emap_t;
|
|
|
|
struct emap_s {
|
|
|
|
rtree_t rtree;
|
|
|
|
/* Keyed by the address of the edata_t being protected. */
|
|
|
|
mutex_pool_t mtx_pool;
|
|
|
|
};
|
|
|
|
|
|
|
|
extern emap_t emap_global;
|
|
|
|
|
|
|
|
bool emap_init(emap_t *emap);
|
|
|
|
|
|
|
|
void emap_lock_edata(tsdn_t *tsdn, emap_t *emap, edata_t *edata);
|
|
|
|
void emap_unlock_edata(tsdn_t *tsdn, emap_t *emap, edata_t *edata);
|
|
|
|
|
|
|
|
void emap_lock_edata2(tsdn_t *tsdn, emap_t *emap, edata_t *edata1,
|
|
|
|
edata_t *edata2);
|
|
|
|
void emap_unlock_edata2(tsdn_t *tsdn, emap_t *emap, edata_t *edata1,
|
|
|
|
edata_t *edata2);
|
|
|
|
|
|
|
|
edata_t *emap_lock_edata_from_addr(tsdn_t *tsdn, emap_t *emap,
|
|
|
|
rtree_ctx_t *rtree_ctx, void *addr, bool inactive_only);
|
|
|
|
|
|
|
|
bool emap_rtree_leaf_elms_lookup(tsdn_t *tsdn, emap_t *emap,
|
|
|
|
rtree_ctx_t *rtree_ctx, const edata_t *edata, bool dependent,
|
|
|
|
bool init_missing, rtree_leaf_elm_t **r_elm_a, rtree_leaf_elm_t **r_elm_b);
|
|
|
|
|
2020-01-31 04:31:19 +08:00
|
|
|
/* Only temporarily public; this will be internal eventually. */
|
|
|
|
void emap_rtree_write_acquired(tsdn_t *tsdn, emap_t *emap,
|
|
|
|
rtree_leaf_elm_t *elm_a, rtree_leaf_elm_t *elm_b, edata_t *edata,
|
|
|
|
szind_t szind, bool slab);
|
|
|
|
|
2020-01-31 04:40:07 +08:00
|
|
|
/*
|
|
|
|
* Associate the given edata with its beginning and end address, setting the
|
|
|
|
* szind and slab info appropriately.
|
|
|
|
* Returns true on error (i.e. resource exhaustion).
|
|
|
|
*/
|
|
|
|
bool emap_register_boundary(tsdn_t *tsdn, emap_t *emap, rtree_ctx_t *rtree_ctx,
|
|
|
|
edata_t *edata, szind_t szind, bool slab);
|
|
|
|
|
2020-01-31 05:32:38 +08:00
|
|
|
/*
|
|
|
|
* Does the same thing, but with the interior of the range, for slab
|
|
|
|
* allocations.
|
|
|
|
*
|
|
|
|
* You might wonder why we don't just have a single emap_register function that
|
|
|
|
* does both depending on the value of 'slab'. The answer is twofold:
|
|
|
|
* - As a practical matter, in places like the extract->split->commit pathway,
|
|
|
|
* we defer the interior operation until we're sure that the commit won't fail
|
|
|
|
* (but we have to register the split boundaries there).
|
|
|
|
* - In general, we're trying to move to a world where the page-specific
|
|
|
|
* allocator doesn't know as much about how the pages it allocates will be
|
|
|
|
* used, and passing a 'slab' parameter everywhere makes that more
|
|
|
|
* complicated.
|
|
|
|
*
|
|
|
|
* Unlike the boundary version, this function can't fail; this is because slabs
|
|
|
|
* can't get big enough to touch a new page that neither of the boundaries
|
|
|
|
* touched, so no allocation is necessary to fill the interior once the boundary
|
|
|
|
* has been touched.
|
|
|
|
*/
|
|
|
|
void emap_register_interior(tsdn_t *tsdn, emap_t *emap, rtree_ctx_t *rtree_ctx,
|
|
|
|
edata_t *edata, szind_t szind);
|
|
|
|
|
2020-01-31 06:55:36 +08:00
|
|
|
void emap_deregister_boundary(tsdn_t *tsdn, emap_t *emap,
|
|
|
|
rtree_ctx_t *rtree_ctx, edata_t *edata);
|
2020-01-31 07:02:52 +08:00
|
|
|
void emap_deregister_interior(tsdn_t *tsdn, emap_t *emap,
|
|
|
|
rtree_ctx_t *rtree_ctx, edata_t *edata);
|
2020-01-31 06:55:36 +08:00
|
|
|
|
2020-01-28 05:55:46 +08:00
|
|
|
#endif /* JEMALLOC_INTERNAL_EMAP_H */
|