2017-01-11 10:06:31 +08:00
|
|
|
#ifndef JEMALLOC_INTERNAL_RTREE_INLINES_H
|
|
|
|
#define JEMALLOC_INTERNAL_RTREE_INLINES_H
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2017-04-20 06:09:01 +08:00
|
|
|
#include "jemalloc/internal/size_classes.h"
|
2017-04-18 07:35:04 +08:00
|
|
|
#include "jemalloc/internal/spin.h"
|
|
|
|
|
2017-02-04 12:21:56 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE uintptr_t
|
2017-02-07 05:17:12 +08:00
|
|
|
rtree_leafkey(uintptr_t key) {
|
2017-02-04 12:21:56 +08:00
|
|
|
unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3);
|
2017-02-07 05:17:12 +08:00
|
|
|
unsigned cumbits = (rtree_levels[RTREE_HEIGHT-1].cumbits -
|
|
|
|
rtree_levels[RTREE_HEIGHT-1].bits);
|
2017-02-04 12:21:56 +08:00
|
|
|
unsigned maskbits = ptrbits - cumbits;
|
|
|
|
uintptr_t mask = ~((ZU(1) << maskbits) - 1);
|
|
|
|
return (key & mask);
|
2010-09-06 01:35:13 +08:00
|
|
|
}
|
|
|
|
|
2017-04-15 02:05:38 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE size_t
|
|
|
|
rtree_cache_direct_map(uintptr_t key) {
|
|
|
|
unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3);
|
|
|
|
unsigned cumbits = (rtree_levels[RTREE_HEIGHT-1].cumbits -
|
|
|
|
rtree_levels[RTREE_HEIGHT-1].bits);
|
|
|
|
unsigned maskbits = ptrbits - cumbits;
|
|
|
|
return (size_t)((key >> maskbits) & (RTREE_CTX_NCACHE - 1));
|
|
|
|
}
|
|
|
|
|
2016-03-24 07:14:41 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE uintptr_t
|
2017-02-07 05:17:12 +08:00
|
|
|
rtree_subkey(uintptr_t key, unsigned level) {
|
2017-02-04 12:21:56 +08:00
|
|
|
unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3);
|
2017-02-07 05:17:12 +08:00
|
|
|
unsigned cumbits = rtree_levels[level].cumbits;
|
2017-02-04 12:21:56 +08:00
|
|
|
unsigned shiftbits = ptrbits - cumbits;
|
2017-02-07 05:17:12 +08:00
|
|
|
unsigned maskbits = rtree_levels[level].bits;
|
2017-02-10 04:31:11 +08:00
|
|
|
uintptr_t mask = (ZU(1) << maskbits) - 1;
|
2017-02-04 12:21:56 +08:00
|
|
|
return ((key >> shiftbits) & mask);
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2017-03-17 08:57:52 +08:00
|
|
|
/*
|
|
|
|
* Atomic getters.
|
|
|
|
*
|
|
|
|
* dependent: Reading a value on behalf of a pointer to a valid allocation
|
|
|
|
* is guaranteed to be a clean read even without synchronization,
|
|
|
|
* because the rtree update became visible in memory before the
|
|
|
|
* pointer came into existence.
|
|
|
|
* !dependent: An arbitrary read, e.g. on behalf of ivsalloc(), may not be
|
|
|
|
* dependent on a previous rtree write, which means a stale read
|
|
|
|
* could result if synchronization were omitted here.
|
|
|
|
*/
|
2017-03-21 07:38:21 +08:00
|
|
|
# ifdef RTREE_LEAF_COMPACT
|
|
|
|
JEMALLOC_ALWAYS_INLINE uintptr_t
|
|
|
|
rtree_leaf_elm_bits_read(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
bool dependent) {
|
2017-03-21 07:38:21 +08:00
|
|
|
return (uintptr_t)atomic_load_p(&elm->le_bits, dependent
|
2017-05-03 12:42:33 +08:00
|
|
|
? ATOMIC_RELAXED : ATOMIC_ACQUIRE);
|
2017-03-21 07:38:21 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE extent_t *
|
|
|
|
rtree_leaf_elm_bits_extent_get(uintptr_t bits) {
|
2017-05-16 05:23:51 +08:00
|
|
|
/* Restore sign-extended high bits, mask slab bit. */
|
2017-03-21 07:38:21 +08:00
|
|
|
return (extent_t *)((uintptr_t)((intptr_t)(bits << RTREE_NHIB) >>
|
2017-05-16 05:23:51 +08:00
|
|
|
RTREE_NHIB) & ~((uintptr_t)0x1));
|
2017-03-21 07:38:21 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE szind_t
|
|
|
|
rtree_leaf_elm_bits_szind_get(uintptr_t bits) {
|
|
|
|
return (szind_t)(bits >> LG_VADDR);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
|
|
|
rtree_leaf_elm_bits_slab_get(uintptr_t bits) {
|
|
|
|
return (bool)(bits & (uintptr_t)0x1);
|
|
|
|
}
|
2017-05-16 05:23:51 +08:00
|
|
|
|
2017-03-21 07:38:21 +08:00
|
|
|
# endif
|
|
|
|
|
2016-03-24 12:09:28 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE extent_t *
|
2017-03-17 08:57:52 +08:00
|
|
|
rtree_leaf_elm_extent_read(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
bool dependent) {
|
2017-03-21 07:38:21 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
2017-05-16 05:23:51 +08:00
|
|
|
uintptr_t bits = rtree_leaf_elm_bits_read(tsdn, rtree, elm, dependent);
|
2017-03-21 07:38:21 +08:00
|
|
|
return rtree_leaf_elm_bits_extent_get(bits);
|
|
|
|
#else
|
2017-03-17 08:57:52 +08:00
|
|
|
extent_t *extent = (extent_t *)atomic_load_p(&elm->le_extent, dependent
|
|
|
|
? ATOMIC_RELAXED : ATOMIC_ACQUIRE);
|
2017-01-20 10:15:45 +08:00
|
|
|
return extent;
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
|
|
|
|
2017-03-17 08:57:52 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE szind_t
|
|
|
|
rtree_leaf_elm_szind_read(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
bool dependent) {
|
2017-03-21 07:38:21 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
2017-05-16 05:23:51 +08:00
|
|
|
uintptr_t bits = rtree_leaf_elm_bits_read(tsdn, rtree, elm, dependent);
|
2017-03-21 07:38:21 +08:00
|
|
|
return rtree_leaf_elm_bits_szind_get(bits);
|
|
|
|
#else
|
2017-03-17 08:57:52 +08:00
|
|
|
return (szind_t)atomic_load_u(&elm->le_szind, dependent ? ATOMIC_RELAXED
|
|
|
|
: ATOMIC_ACQUIRE);
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
|
|
|
rtree_leaf_elm_slab_read(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
bool dependent) {
|
2017-03-21 07:38:21 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
2017-05-16 05:23:51 +08:00
|
|
|
uintptr_t bits = rtree_leaf_elm_bits_read(tsdn, rtree, elm, dependent);
|
2017-03-21 07:38:21 +08:00
|
|
|
return rtree_leaf_elm_bits_slab_get(bits);
|
|
|
|
#else
|
2017-03-17 08:57:52 +08:00
|
|
|
return atomic_load_b(&elm->le_slab, dependent ? ATOMIC_RELAXED :
|
|
|
|
ATOMIC_ACQUIRE);
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-05-16 05:23:51 +08:00
|
|
|
rtree_leaf_elm_extent_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
|
|
|
extent_t *extent) {
|
2017-03-21 07:38:21 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
2017-05-16 05:23:51 +08:00
|
|
|
uintptr_t old_bits = rtree_leaf_elm_bits_read(tsdn, rtree, elm, true);
|
2017-03-21 07:38:21 +08:00
|
|
|
uintptr_t bits = ((uintptr_t)rtree_leaf_elm_bits_szind_get(old_bits) <<
|
|
|
|
LG_VADDR) | ((uintptr_t)extent & (((uintptr_t)0x1 << LG_VADDR) - 1))
|
2017-05-16 05:23:51 +08:00
|
|
|
| ((uintptr_t)rtree_leaf_elm_bits_slab_get(old_bits));
|
2017-03-21 07:38:21 +08:00
|
|
|
atomic_store_p(&elm->le_bits, (void *)bits, ATOMIC_RELEASE);
|
|
|
|
#else
|
2017-03-17 08:57:52 +08:00
|
|
|
atomic_store_p(&elm->le_extent, extent, ATOMIC_RELEASE);
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-03-17 08:57:52 +08:00
|
|
|
rtree_leaf_elm_szind_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
szind_t szind) {
|
2017-03-17 08:57:52 +08:00
|
|
|
assert(szind <= NSIZES);
|
|
|
|
|
2017-03-21 07:38:21 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
|
|
|
uintptr_t old_bits = rtree_leaf_elm_bits_read(tsdn, rtree, elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
true);
|
2017-03-21 07:38:21 +08:00
|
|
|
uintptr_t bits = ((uintptr_t)szind << LG_VADDR) |
|
|
|
|
((uintptr_t)rtree_leaf_elm_bits_extent_get(old_bits) &
|
|
|
|
(((uintptr_t)0x1 << LG_VADDR) - 1)) |
|
2017-05-16 05:23:51 +08:00
|
|
|
((uintptr_t)rtree_leaf_elm_bits_slab_get(old_bits));
|
2017-03-21 07:38:21 +08:00
|
|
|
atomic_store_p(&elm->le_bits, (void *)bits, ATOMIC_RELEASE);
|
|
|
|
#else
|
2017-03-17 08:57:52 +08:00
|
|
|
atomic_store_u(&elm->le_szind, szind, ATOMIC_RELEASE);
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-03-17 08:57:52 +08:00
|
|
|
rtree_leaf_elm_slab_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
bool slab) {
|
2017-03-21 07:38:21 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
|
|
|
uintptr_t old_bits = rtree_leaf_elm_bits_read(tsdn, rtree, elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
true);
|
2017-03-21 07:38:21 +08:00
|
|
|
uintptr_t bits = ((uintptr_t)rtree_leaf_elm_bits_szind_get(old_bits) <<
|
|
|
|
LG_VADDR) | ((uintptr_t)rtree_leaf_elm_bits_extent_get(old_bits) &
|
2017-05-16 05:23:51 +08:00
|
|
|
(((uintptr_t)0x1 << LG_VADDR) - 1)) | ((uintptr_t)slab);
|
2017-03-21 07:38:21 +08:00
|
|
|
atomic_store_p(&elm->le_bits, (void *)bits, ATOMIC_RELEASE);
|
|
|
|
#else
|
2017-03-17 08:57:52 +08:00
|
|
|
atomic_store_b(&elm->le_slab, slab, ATOMIC_RELEASE);
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-03-17 08:57:52 +08:00
|
|
|
rtree_leaf_elm_write(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm,
|
2017-05-16 05:23:51 +08:00
|
|
|
extent_t *extent, szind_t szind, bool slab) {
|
2017-03-27 19:08:51 +08:00
|
|
|
#ifdef RTREE_LEAF_COMPACT
|
2017-03-21 07:38:21 +08:00
|
|
|
uintptr_t bits = ((uintptr_t)szind << LG_VADDR) |
|
|
|
|
((uintptr_t)extent & (((uintptr_t)0x1 << LG_VADDR) - 1)) |
|
2017-05-16 05:23:51 +08:00
|
|
|
((uintptr_t)slab);
|
2017-03-21 07:38:21 +08:00
|
|
|
atomic_store_p(&elm->le_bits, (void *)bits, ATOMIC_RELEASE);
|
|
|
|
#else
|
2017-05-16 05:23:51 +08:00
|
|
|
rtree_leaf_elm_slab_write(tsdn, rtree, elm, slab);
|
|
|
|
rtree_leaf_elm_szind_write(tsdn, rtree, elm, szind);
|
2017-03-17 08:57:52 +08:00
|
|
|
/*
|
|
|
|
* Write extent last, since the element is atomically considered valid
|
|
|
|
* as soon as the extent field is non-NULL.
|
|
|
|
*/
|
2017-05-16 05:23:51 +08:00
|
|
|
rtree_leaf_elm_extent_write(tsdn, rtree, elm, extent);
|
2017-03-21 07:38:21 +08:00
|
|
|
#endif
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-03-27 19:08:51 +08:00
|
|
|
rtree_leaf_elm_szind_slab_update(tsdn_t *tsdn, rtree_t *rtree,
|
|
|
|
rtree_leaf_elm_t *elm, szind_t szind, bool slab) {
|
|
|
|
assert(!slab || szind < NBINS);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The caller implicitly assures that it is the only writer to the szind
|
|
|
|
* and slab fields, and that the extent field cannot currently change.
|
|
|
|
*/
|
2017-05-16 05:23:51 +08:00
|
|
|
rtree_leaf_elm_slab_write(tsdn, rtree, elm, slab);
|
|
|
|
rtree_leaf_elm_szind_write(tsdn, rtree, elm, szind);
|
2017-03-27 19:08:51 +08:00
|
|
|
}
|
|
|
|
|
2017-03-17 00:46:42 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_leaf_elm_t *
|
|
|
|
rtree_leaf_elm_lookup(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
2017-01-16 08:56:30 +08:00
|
|
|
uintptr_t key, bool dependent, bool init_missing) {
|
2017-02-05 18:50:59 +08:00
|
|
|
assert(key != 0);
|
2016-03-28 18:06:35 +08:00
|
|
|
assert(!dependent || !init_missing);
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2017-04-15 02:05:38 +08:00
|
|
|
size_t slot = rtree_cache_direct_map(key);
|
2017-02-05 18:50:59 +08:00
|
|
|
uintptr_t leafkey = rtree_leafkey(key);
|
2017-03-28 12:50:38 +08:00
|
|
|
assert(leafkey != RTREE_LEAFKEY_INVALID);
|
|
|
|
|
2017-04-15 02:05:38 +08:00
|
|
|
/* Fast path: L1 direct mapped cache. */
|
|
|
|
if (likely(rtree_ctx->cache[slot].leafkey == leafkey)) {
|
|
|
|
rtree_leaf_elm_t *leaf = rtree_ctx->cache[slot].leaf;
|
|
|
|
assert(leaf != NULL);
|
|
|
|
uintptr_t subkey = rtree_subkey(key, RTREE_HEIGHT-1);
|
|
|
|
return &leaf[subkey];
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Search the L2 LRU cache. On hit, swap the matching element into the
|
|
|
|
* slot in L1 cache, and move the position in L2 up by 1.
|
|
|
|
*/
|
|
|
|
#define RTREE_CACHE_CHECK_L2(i) do { \
|
|
|
|
if (likely(rtree_ctx->l2_cache[i].leafkey == leafkey)) { \
|
|
|
|
rtree_leaf_elm_t *leaf = rtree_ctx->l2_cache[i].leaf; \
|
2017-03-29 08:14:43 +08:00
|
|
|
assert(leaf != NULL); \
|
|
|
|
if (i > 0) { \
|
2017-03-28 02:48:39 +08:00
|
|
|
/* Bubble up by one. */ \
|
2017-04-15 02:05:38 +08:00
|
|
|
rtree_ctx->l2_cache[i].leafkey = \
|
|
|
|
rtree_ctx->l2_cache[i - 1].leafkey; \
|
|
|
|
rtree_ctx->l2_cache[i].leaf = \
|
|
|
|
rtree_ctx->l2_cache[i - 1].leaf; \
|
|
|
|
rtree_ctx->l2_cache[i - 1].leafkey = \
|
|
|
|
rtree_ctx->cache[slot].leafkey; \
|
|
|
|
rtree_ctx->l2_cache[i - 1].leaf = \
|
|
|
|
rtree_ctx->cache[slot].leaf; \
|
|
|
|
} else { \
|
|
|
|
rtree_ctx->l2_cache[0].leafkey = \
|
|
|
|
rtree_ctx->cache[slot].leafkey; \
|
|
|
|
rtree_ctx->l2_cache[0].leaf = \
|
|
|
|
rtree_ctx->cache[slot].leaf; \
|
2017-02-04 12:21:56 +08:00
|
|
|
} \
|
2017-04-15 02:05:38 +08:00
|
|
|
rtree_ctx->cache[slot].leafkey = leafkey; \
|
|
|
|
rtree_ctx->cache[slot].leaf = leaf; \
|
2017-03-29 08:14:43 +08:00
|
|
|
uintptr_t subkey = rtree_subkey(key, RTREE_HEIGHT-1); \
|
|
|
|
return &leaf[subkey]; \
|
2017-02-05 18:50:59 +08:00
|
|
|
} \
|
2017-02-04 12:21:56 +08:00
|
|
|
} while (0)
|
2017-03-28 02:48:39 +08:00
|
|
|
/* Check the first cache entry. */
|
2017-04-15 02:05:38 +08:00
|
|
|
RTREE_CACHE_CHECK_L2(0);
|
|
|
|
/* Search the remaining cache elements. */
|
|
|
|
for (unsigned i = 1; i < RTREE_CTX_NCACHE_L2; i++) {
|
|
|
|
RTREE_CACHE_CHECK_L2(i);
|
2017-02-05 18:50:59 +08:00
|
|
|
}
|
2017-04-15 02:05:38 +08:00
|
|
|
#undef RTREE_CACHE_CHECK_L2
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2017-03-17 00:46:42 +08:00
|
|
|
return rtree_leaf_elm_lookup_hard(tsdn, rtree, rtree_ctx, key,
|
|
|
|
dependent, init_missing);
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline bool
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_write(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key,
|
2017-03-17 08:57:52 +08:00
|
|
|
extent_t *extent, szind_t szind, bool slab) {
|
|
|
|
/* Use rtree_clear() to set the extent to NULL. */
|
|
|
|
assert(extent != NULL);
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2017-03-17 08:57:52 +08:00
|
|
|
rtree_leaf_elm_t *elm = rtree_leaf_elm_lookup(tsdn, rtree, rtree_ctx,
|
|
|
|
key, false, true);
|
2017-01-16 08:56:30 +08:00
|
|
|
if (elm == NULL) {
|
2017-01-20 10:15:45 +08:00
|
|
|
return true;
|
2017-01-16 08:56:30 +08:00
|
|
|
}
|
2017-03-17 08:57:52 +08:00
|
|
|
|
2017-05-16 05:23:51 +08:00
|
|
|
assert(rtree_leaf_elm_extent_read(tsdn, rtree, elm, false) == NULL);
|
|
|
|
rtree_leaf_elm_write(tsdn, rtree, elm, extent, szind, slab);
|
2016-03-28 18:06:35 +08:00
|
|
|
|
2017-01-20 10:15:45 +08:00
|
|
|
return false;
|
2016-03-28 18:06:35 +08:00
|
|
|
}
|
|
|
|
|
2017-03-17 08:57:52 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_leaf_elm_t *
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key,
|
2017-01-16 08:56:30 +08:00
|
|
|
bool dependent) {
|
2017-03-17 08:57:52 +08:00
|
|
|
rtree_leaf_elm_t *elm = rtree_leaf_elm_lookup(tsdn, rtree, rtree_ctx,
|
|
|
|
key, dependent, false);
|
|
|
|
if (!dependent && elm == NULL) {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
assert(elm != NULL);
|
|
|
|
return elm;
|
|
|
|
}
|
2016-03-28 18:06:35 +08:00
|
|
|
|
2017-03-17 08:57:52 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE extent_t *
|
|
|
|
rtree_extent_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, bool dependent) {
|
|
|
|
rtree_leaf_elm_t *elm = rtree_read(tsdn, rtree, rtree_ctx, key,
|
|
|
|
dependent);
|
2017-02-04 12:17:47 +08:00
|
|
|
if (!dependent && elm == NULL) {
|
2017-01-20 10:15:45 +08:00
|
|
|
return NULL;
|
2017-01-16 08:56:30 +08:00
|
|
|
}
|
2017-05-16 05:23:51 +08:00
|
|
|
return rtree_leaf_elm_extent_read(tsdn, rtree, elm, dependent);
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE szind_t
|
|
|
|
rtree_szind_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, bool dependent) {
|
|
|
|
rtree_leaf_elm_t *elm = rtree_read(tsdn, rtree, rtree_ctx, key,
|
|
|
|
dependent);
|
|
|
|
if (!dependent && elm == NULL) {
|
|
|
|
return NSIZES;
|
|
|
|
}
|
2017-05-16 05:23:51 +08:00
|
|
|
return rtree_leaf_elm_szind_read(tsdn, rtree, elm, dependent);
|
2017-03-17 08:57:52 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* rtree_slab_read() is intentionally omitted because slab is always read in
|
|
|
|
* conjunction with szind, which makes rtree_szind_slab_read() a better choice.
|
|
|
|
*/
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
|
|
|
rtree_extent_szind_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, bool dependent, extent_t **r_extent, szind_t *r_szind) {
|
|
|
|
rtree_leaf_elm_t *elm = rtree_read(tsdn, rtree, rtree_ctx, key,
|
|
|
|
dependent);
|
|
|
|
if (!dependent && elm == NULL) {
|
|
|
|
return true;
|
|
|
|
}
|
2017-05-16 05:23:51 +08:00
|
|
|
*r_extent = rtree_leaf_elm_extent_read(tsdn, rtree, elm, dependent);
|
|
|
|
*r_szind = rtree_leaf_elm_szind_read(tsdn, rtree, elm, dependent);
|
2017-03-17 08:57:52 +08:00
|
|
|
return false;
|
|
|
|
}
|
2016-03-28 18:06:35 +08:00
|
|
|
|
2017-03-17 08:57:52 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
|
|
|
rtree_szind_slab_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, bool dependent, szind_t *r_szind, bool *r_slab) {
|
|
|
|
rtree_leaf_elm_t *elm = rtree_read(tsdn, rtree, rtree_ctx, key,
|
|
|
|
dependent);
|
|
|
|
if (!dependent && elm == NULL) {
|
|
|
|
return true;
|
|
|
|
}
|
2017-05-16 05:23:51 +08:00
|
|
|
*r_szind = rtree_leaf_elm_szind_read(tsdn, rtree, elm, dependent);
|
|
|
|
*r_slab = rtree_leaf_elm_slab_read(tsdn, rtree, elm, dependent);
|
2017-03-17 08:57:52 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-03-27 19:08:51 +08:00
|
|
|
rtree_szind_slab_update(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, szind_t szind, bool slab) {
|
|
|
|
assert(!slab || szind < NBINS);
|
|
|
|
|
|
|
|
rtree_leaf_elm_t *elm = rtree_read(tsdn, rtree, rtree_ctx, key, true);
|
|
|
|
rtree_leaf_elm_szind_slab_update(tsdn, rtree, elm, szind, slab);
|
|
|
|
}
|
|
|
|
|
2017-04-22 00:37:34 +08:00
|
|
|
static inline void
|
2017-01-16 08:56:30 +08:00
|
|
|
rtree_clear(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key) {
|
2017-03-27 19:10:11 +08:00
|
|
|
rtree_leaf_elm_t *elm = rtree_read(tsdn, rtree, rtree_ctx, key, true);
|
2017-05-16 05:23:51 +08:00
|
|
|
assert(rtree_leaf_elm_extent_read(tsdn, rtree, elm, false) !=
|
2017-03-27 19:10:11 +08:00
|
|
|
NULL);
|
2017-05-16 05:23:51 +08:00
|
|
|
rtree_leaf_elm_write(tsdn, rtree, elm, NULL, NSIZES, false);
|
2010-09-06 01:35:13 +08:00
|
|
|
}
|
|
|
|
|
2017-01-11 10:06:31 +08:00
|
|
|
#endif /* JEMALLOC_INTERNAL_RTREE_INLINES_H */
|