2017-01-11 10:06:31 +08:00
|
|
|
#ifndef JEMALLOC_INTERNAL_RTREE_INLINES_H
|
|
|
|
#define JEMALLOC_INTERNAL_RTREE_INLINES_H
|
2010-09-06 01:35:13 +08:00
|
|
|
|
|
|
|
#ifndef JEMALLOC_ENABLE_INLINE
|
2016-06-03 09:43:10 +08:00
|
|
|
unsigned rtree_start_level(const rtree_t *rtree, uintptr_t key);
|
|
|
|
unsigned rtree_ctx_start_level(const rtree_t *rtree,
|
|
|
|
const rtree_ctx_t *rtree_ctx, uintptr_t key);
|
2015-01-31 14:54:08 +08:00
|
|
|
uintptr_t rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level);
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
bool rtree_node_valid(rtree_elm_t *node);
|
|
|
|
rtree_elm_t *rtree_child_tryread(rtree_elm_t *elm, bool dependent);
|
2016-04-16 15:36:11 +08:00
|
|
|
rtree_elm_t *rtree_child_read(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *elm,
|
2016-03-23 08:54:35 +08:00
|
|
|
unsigned level, bool dependent);
|
2016-03-28 18:06:35 +08:00
|
|
|
extent_t *rtree_elm_read(rtree_elm_t *elm, bool dependent);
|
|
|
|
void rtree_elm_write(rtree_elm_t *elm, const extent_t *extent);
|
|
|
|
rtree_elm_t *rtree_subtree_tryread(rtree_t *rtree, unsigned level,
|
2016-03-23 08:54:35 +08:00
|
|
|
bool dependent);
|
2016-04-16 15:36:11 +08:00
|
|
|
rtree_elm_t *rtree_subtree_read(tsdn_t *tsdn, rtree_t *rtree,
|
|
|
|
unsigned level, bool dependent);
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_elm_t *rtree_elm_lookup(tsdn_t *tsdn, rtree_t *rtree,
|
|
|
|
rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent, bool init_missing);
|
|
|
|
|
|
|
|
bool rtree_write(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, const extent_t *extent);
|
|
|
|
extent_t *rtree_read(tsdn_t *tsdn, rtree_t *rtree,
|
|
|
|
rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent);
|
|
|
|
rtree_elm_t *rtree_elm_acquire(tsdn_t *tsdn, rtree_t *rtree,
|
|
|
|
rtree_ctx_t *rtree_ctx, uintptr_t key, bool dependent, bool init_missing);
|
2016-04-18 03:55:10 +08:00
|
|
|
extent_t *rtree_elm_read_acquired(tsdn_t *tsdn, const rtree_t *rtree,
|
|
|
|
rtree_elm_t *elm);
|
|
|
|
void rtree_elm_write_acquired(tsdn_t *tsdn, const rtree_t *rtree,
|
|
|
|
rtree_elm_t *elm, const extent_t *extent);
|
|
|
|
void rtree_elm_release(tsdn_t *tsdn, const rtree_t *rtree, rtree_elm_t *elm);
|
2016-06-03 09:43:10 +08:00
|
|
|
void rtree_clear(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key);
|
2010-09-06 01:35:13 +08:00
|
|
|
#endif
|
|
|
|
|
2011-03-19 08:56:14 +08:00
|
|
|
#if (defined(JEMALLOC_ENABLE_INLINE) || defined(JEMALLOC_RTREE_C_))
|
2016-03-24 07:14:41 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE unsigned
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_start_level(const rtree_t *rtree, uintptr_t key)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
|
|
|
unsigned start_level;
|
|
|
|
|
|
|
|
if (unlikely(key == 0))
|
|
|
|
return (rtree->height - 1);
|
|
|
|
|
2016-06-03 09:43:10 +08:00
|
|
|
start_level = rtree->start_level[(lg_floor(key) + 1) >>
|
|
|
|
LG_RTREE_BITS_PER_LEVEL];
|
|
|
|
assert(start_level < rtree->height);
|
|
|
|
return (start_level);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE unsigned
|
|
|
|
rtree_ctx_start_level(const rtree_t *rtree, const rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key)
|
|
|
|
{
|
|
|
|
unsigned start_level;
|
|
|
|
uintptr_t key_diff;
|
|
|
|
|
|
|
|
/* Compute the difference between old and new lookup keys. */
|
|
|
|
key_diff = key ^ rtree_ctx->key;
|
|
|
|
assert(key_diff != 0); /* Handled in rtree_elm_lookup(). */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Compute the last traversal path element at which the keys' paths
|
|
|
|
* are the same.
|
|
|
|
*/
|
|
|
|
start_level = rtree->start_level[(lg_floor(key_diff) + 1) >>
|
2015-01-31 14:54:08 +08:00
|
|
|
LG_RTREE_BITS_PER_LEVEL];
|
|
|
|
assert(start_level < rtree->height);
|
|
|
|
return (start_level);
|
2010-09-06 01:35:13 +08:00
|
|
|
}
|
|
|
|
|
2016-03-24 07:14:41 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE uintptr_t
|
2015-01-31 14:54:08 +08:00
|
|
|
rtree_subkey(rtree_t *rtree, uintptr_t key, unsigned level)
|
|
|
|
{
|
|
|
|
return ((key >> ((ZU(1) << (LG_SIZEOF_PTR+3)) -
|
|
|
|
rtree->levels[level].cumbits)) & ((ZU(1) <<
|
|
|
|
rtree->levels[level].bits) - 1));
|
|
|
|
}
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2016-03-24 07:14:41 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE bool
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_node_valid(rtree_elm_t *node)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-11-01 07:23:33 +08:00
|
|
|
return ((uintptr_t)node != (uintptr_t)0);
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_elm_t *
|
|
|
|
rtree_child_tryread(rtree_elm_t *elm, bool dependent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_t *child;
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
/* Double-checked read (first read may be stale). */
|
2015-01-31 14:54:08 +08:00
|
|
|
child = elm->child;
|
2016-03-23 08:54:35 +08:00
|
|
|
if (!dependent && !rtree_node_valid(child))
|
jemalloc cpp new/delete bindings
Adds cpp bindings for jemalloc, along with necessary autoconf settings.
This is mostly to add sized deallocation support, which can't be added
from C directly. Sized deallocation is ~10% microbench improvement.
* Import ax_cxx_compile_stdcxx.m4 from the autoconf repo, seems like the
easiest way to get c++14 detection.
* Adds various other changes, like CXXFLAGS, to configure.ac.
* Adds new rules to Makefile.in for src/jemalloc-cpp.cpp, and a basic
unittest.
* Both new and delete are overridden, to ensure jemalloc is used for
both.
* TODO future enhancement of avoiding extra PLT thunks for new and
delete - sdallocx and malloc are publicly exported jemalloc symbols,
using an alias would link them directly. Unfortunately, was having
trouble getting it to play nice with jemalloc's namespace support.
Testing:
Tested gcc 4.8, gcc 5, gcc 5.2, clang 4.0. Only gcc >= 5 has sized
deallocation support, verified that the rest build correctly.
Tested mac osx and Centos.
Tested --with-jemalloc-prefix and --without-export.
This resolves #202.
2016-10-24 06:56:30 +08:00
|
|
|
child = (rtree_elm_t *)atomic_read_p(&elm->pun);
|
2016-03-23 08:54:35 +08:00
|
|
|
assert(!dependent || child != NULL);
|
2015-01-31 14:54:08 +08:00
|
|
|
return (child);
|
|
|
|
}
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_elm_t *
|
2016-04-16 15:36:11 +08:00
|
|
|
rtree_child_read(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *elm, unsigned level,
|
2016-03-23 08:54:35 +08:00
|
|
|
bool dependent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_t *child;
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-03-23 08:54:35 +08:00
|
|
|
child = rtree_child_tryread(elm, dependent);
|
|
|
|
if (!dependent && unlikely(!rtree_node_valid(child)))
|
2016-04-16 15:36:11 +08:00
|
|
|
child = rtree_child_read_hard(tsdn, rtree, elm, level);
|
2016-03-23 08:54:35 +08:00
|
|
|
assert(!dependent || child != NULL);
|
2015-01-31 14:54:08 +08:00
|
|
|
return (child);
|
|
|
|
}
|
|
|
|
|
2016-03-24 12:09:28 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE extent_t *
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_read(rtree_elm_t *elm, bool dependent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
extent_t *extent;
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2015-05-16 08:02:30 +08:00
|
|
|
if (dependent) {
|
|
|
|
/*
|
2016-03-28 18:06:35 +08:00
|
|
|
* Reading a value on behalf of a pointer to a valid allocation
|
|
|
|
* is guaranteed to be a clean read even without
|
|
|
|
* synchronization, because the rtree update became visible in
|
|
|
|
* memory before the pointer came into existence.
|
2015-05-16 08:02:30 +08:00
|
|
|
*/
|
2016-03-28 18:06:35 +08:00
|
|
|
extent = elm->extent;
|
2015-05-16 08:02:30 +08:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* An arbitrary read, e.g. on behalf of ivsalloc(), may not be
|
|
|
|
* dependent on a previous rtree write, which means a stale read
|
|
|
|
* could result if synchronization were omitted here.
|
|
|
|
*/
|
2016-03-28 18:06:35 +08:00
|
|
|
extent = (extent_t *)atomic_read_p(&elm->pun);
|
2015-05-16 08:02:30 +08:00
|
|
|
}
|
2016-03-28 18:06:35 +08:00
|
|
|
|
|
|
|
/* Mask the lock bit. */
|
|
|
|
extent = (extent_t *)((uintptr_t)extent & ~((uintptr_t)0x1));
|
|
|
|
|
|
|
|
return (extent);
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE void
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_write(rtree_elm_t *elm, const extent_t *extent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
atomic_write_p(&elm->pun, extent);
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_elm_t *
|
2016-03-23 08:54:35 +08:00
|
|
|
rtree_subtree_tryread(rtree_t *rtree, unsigned level, bool dependent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_t *subtree;
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
/* Double-checked read (first read may be stale). */
|
2015-01-31 14:54:08 +08:00
|
|
|
subtree = rtree->levels[level].subtree;
|
jemalloc cpp new/delete bindings
Adds cpp bindings for jemalloc, along with necessary autoconf settings.
This is mostly to add sized deallocation support, which can't be added
from C directly. Sized deallocation is ~10% microbench improvement.
* Import ax_cxx_compile_stdcxx.m4 from the autoconf repo, seems like the
easiest way to get c++14 detection.
* Adds various other changes, like CXXFLAGS, to configure.ac.
* Adds new rules to Makefile.in for src/jemalloc-cpp.cpp, and a basic
unittest.
* Both new and delete are overridden, to ensure jemalloc is used for
both.
* TODO future enhancement of avoiding extra PLT thunks for new and
delete - sdallocx and malloc are publicly exported jemalloc symbols,
using an alias would link them directly. Unfortunately, was having
trouble getting it to play nice with jemalloc's namespace support.
Testing:
Tested gcc 4.8, gcc 5, gcc 5.2, clang 4.0. Only gcc >= 5 has sized
deallocation support, verified that the rest build correctly.
Tested mac osx and Centos.
Tested --with-jemalloc-prefix and --without-export.
This resolves #202.
2016-10-24 06:56:30 +08:00
|
|
|
if (!dependent && unlikely(!rtree_node_valid(subtree))) {
|
|
|
|
subtree = (rtree_elm_t *)atomic_read_p(
|
|
|
|
&rtree->levels[level].subtree_pun);
|
|
|
|
}
|
2016-03-23 08:54:35 +08:00
|
|
|
assert(!dependent || subtree != NULL);
|
2015-01-31 14:54:08 +08:00
|
|
|
return (subtree);
|
|
|
|
}
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_elm_t *
|
2016-04-16 15:36:11 +08:00
|
|
|
rtree_subtree_read(tsdn_t *tsdn, rtree_t *rtree, unsigned level, bool dependent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_t *subtree;
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-03-23 08:54:35 +08:00
|
|
|
subtree = rtree_subtree_tryread(rtree, level, dependent);
|
|
|
|
if (!dependent && unlikely(!rtree_node_valid(subtree)))
|
2016-04-16 15:36:11 +08:00
|
|
|
subtree = rtree_subtree_read_hard(tsdn, rtree, level);
|
2016-03-23 08:54:35 +08:00
|
|
|
assert(!dependent || subtree != NULL);
|
2015-01-31 14:54:08 +08:00
|
|
|
return (subtree);
|
|
|
|
}
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
JEMALLOC_ALWAYS_INLINE rtree_elm_t *
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_elm_lookup(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, bool dependent, bool init_missing)
|
2010-09-06 01:35:13 +08:00
|
|
|
{
|
|
|
|
uintptr_t subkey;
|
2016-03-23 08:54:35 +08:00
|
|
|
unsigned start_level;
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_t *node;
|
|
|
|
|
|
|
|
assert(!dependent || !init_missing);
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-06-03 09:43:10 +08:00
|
|
|
if (dependent || init_missing) {
|
|
|
|
if (likely(rtree_ctx->valid)) {
|
|
|
|
if (key == rtree_ctx->key)
|
|
|
|
return (rtree_ctx->elms[rtree->height]);
|
|
|
|
else {
|
|
|
|
unsigned no_ctx_start_level =
|
|
|
|
rtree_start_level(rtree, key);
|
|
|
|
unsigned ctx_start_level;
|
|
|
|
|
|
|
|
if (likely(no_ctx_start_level <=
|
|
|
|
rtree_ctx->start_level && (ctx_start_level =
|
|
|
|
rtree_ctx_start_level(rtree, rtree_ctx,
|
|
|
|
key)) >= rtree_ctx->start_level)) {
|
|
|
|
start_level = ctx_start_level;
|
|
|
|
node = rtree_ctx->elms[ctx_start_level];
|
|
|
|
} else {
|
|
|
|
start_level = no_ctx_start_level;
|
|
|
|
node = init_missing ?
|
|
|
|
rtree_subtree_read(tsdn, rtree,
|
|
|
|
no_ctx_start_level, dependent) :
|
|
|
|
rtree_subtree_tryread(rtree,
|
|
|
|
no_ctx_start_level, dependent);
|
|
|
|
rtree_ctx->start_level =
|
|
|
|
no_ctx_start_level;
|
|
|
|
rtree_ctx->elms[no_ctx_start_level] =
|
|
|
|
node;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
unsigned no_ctx_start_level = rtree_start_level(rtree,
|
|
|
|
key);
|
|
|
|
|
|
|
|
start_level = no_ctx_start_level;
|
|
|
|
node = init_missing ? rtree_subtree_read(tsdn, rtree,
|
|
|
|
no_ctx_start_level, dependent) :
|
|
|
|
rtree_subtree_tryread(rtree, no_ctx_start_level,
|
|
|
|
dependent);
|
|
|
|
rtree_ctx->valid = true;
|
|
|
|
rtree_ctx->start_level = no_ctx_start_level;
|
|
|
|
rtree_ctx->elms[no_ctx_start_level] = node;
|
|
|
|
}
|
|
|
|
rtree_ctx->key = key;
|
|
|
|
} else {
|
|
|
|
start_level = rtree_start_level(rtree, key);
|
|
|
|
node = init_missing ? rtree_subtree_read(tsdn, rtree,
|
|
|
|
start_level, dependent) : rtree_subtree_tryread(rtree,
|
|
|
|
start_level, dependent);
|
|
|
|
}
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-03-23 08:54:35 +08:00
|
|
|
#define RTREE_GET_BIAS (RTREE_HEIGHT_MAX - rtree->height)
|
|
|
|
switch (start_level + RTREE_GET_BIAS) {
|
|
|
|
#define RTREE_GET_SUBTREE(level) \
|
|
|
|
case level: \
|
|
|
|
assert(level < (RTREE_HEIGHT_MAX-1)); \
|
2016-06-03 09:43:10 +08:00
|
|
|
if (!dependent && unlikely(!rtree_node_valid(node))) { \
|
|
|
|
if (init_missing) \
|
|
|
|
rtree_ctx->valid = false; \
|
2016-03-23 08:54:35 +08:00
|
|
|
return (NULL); \
|
2016-06-03 09:43:10 +08:00
|
|
|
} \
|
2016-03-23 08:54:35 +08:00
|
|
|
subkey = rtree_subkey(rtree, key, level - \
|
|
|
|
RTREE_GET_BIAS); \
|
2016-04-16 15:36:11 +08:00
|
|
|
node = init_missing ? rtree_child_read(tsdn, rtree, \
|
2016-03-28 18:06:35 +08:00
|
|
|
&node[subkey], level - RTREE_GET_BIAS, dependent) : \
|
|
|
|
rtree_child_tryread(&node[subkey], dependent); \
|
2016-06-03 09:43:10 +08:00
|
|
|
if (dependent || init_missing) { \
|
|
|
|
rtree_ctx->elms[level - RTREE_GET_BIAS + 1] = \
|
|
|
|
node; \
|
|
|
|
} \
|
2016-03-23 08:54:35 +08:00
|
|
|
/* Fall through. */
|
|
|
|
#define RTREE_GET_LEAF(level) \
|
|
|
|
case level: \
|
|
|
|
assert(level == (RTREE_HEIGHT_MAX-1)); \
|
2016-06-03 09:43:10 +08:00
|
|
|
if (!dependent && unlikely(!rtree_node_valid(node))) { \
|
|
|
|
if (init_missing) \
|
|
|
|
rtree_ctx->valid = false; \
|
2016-03-23 08:54:35 +08:00
|
|
|
return (NULL); \
|
2016-06-03 09:43:10 +08:00
|
|
|
} \
|
2016-03-23 08:54:35 +08:00
|
|
|
subkey = rtree_subkey(rtree, key, level - \
|
|
|
|
RTREE_GET_BIAS); \
|
|
|
|
/* \
|
|
|
|
* node is a leaf, so it contains values rather than \
|
|
|
|
* child pointers. \
|
|
|
|
*/ \
|
2016-06-03 09:43:10 +08:00
|
|
|
node = &node[subkey]; \
|
|
|
|
if (dependent || init_missing) { \
|
|
|
|
rtree_ctx->elms[level - RTREE_GET_BIAS + 1] = \
|
|
|
|
node; \
|
|
|
|
} \
|
|
|
|
return (node);
|
2016-03-23 08:54:35 +08:00
|
|
|
#if RTREE_HEIGHT_MAX > 1
|
|
|
|
RTREE_GET_SUBTREE(0)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 2
|
|
|
|
RTREE_GET_SUBTREE(1)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 3
|
|
|
|
RTREE_GET_SUBTREE(2)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 4
|
|
|
|
RTREE_GET_SUBTREE(3)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 5
|
|
|
|
RTREE_GET_SUBTREE(4)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 6
|
|
|
|
RTREE_GET_SUBTREE(5)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 7
|
|
|
|
RTREE_GET_SUBTREE(6)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 8
|
|
|
|
RTREE_GET_SUBTREE(7)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 9
|
|
|
|
RTREE_GET_SUBTREE(8)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 10
|
|
|
|
RTREE_GET_SUBTREE(9)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 11
|
|
|
|
RTREE_GET_SUBTREE(10)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 12
|
|
|
|
RTREE_GET_SUBTREE(11)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 13
|
|
|
|
RTREE_GET_SUBTREE(12)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 14
|
|
|
|
RTREE_GET_SUBTREE(13)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 15
|
|
|
|
RTREE_GET_SUBTREE(14)
|
|
|
|
#endif
|
|
|
|
#if RTREE_HEIGHT_MAX > 16
|
|
|
|
# error Unsupported RTREE_HEIGHT_MAX
|
|
|
|
#endif
|
|
|
|
RTREE_GET_LEAF(RTREE_HEIGHT_MAX-1)
|
|
|
|
#undef RTREE_GET_SUBTREE
|
|
|
|
#undef RTREE_GET_LEAF
|
|
|
|
default: not_reached();
|
2010-09-06 01:35:13 +08:00
|
|
|
}
|
2016-03-24 07:14:41 +08:00
|
|
|
#undef RTREE_GET_BIAS
|
2015-01-31 14:54:08 +08:00
|
|
|
not_reached();
|
|
|
|
}
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2015-01-31 14:54:08 +08:00
|
|
|
JEMALLOC_INLINE bool
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_write(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key,
|
|
|
|
const extent_t *extent)
|
2015-01-31 14:54:08 +08:00
|
|
|
{
|
2016-03-28 18:06:35 +08:00
|
|
|
rtree_elm_t *elm;
|
2010-09-06 01:35:13 +08:00
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
assert(extent != NULL); /* Use rtree_clear() for this case. */
|
|
|
|
assert(((uintptr_t)extent & (uintptr_t)0x1) == (uintptr_t)0x0);
|
2015-01-31 14:54:08 +08:00
|
|
|
|
2016-06-03 09:43:10 +08:00
|
|
|
elm = rtree_elm_lookup(tsdn, rtree, rtree_ctx, key, false, true);
|
2016-03-28 18:06:35 +08:00
|
|
|
if (elm == NULL)
|
2015-01-31 14:54:08 +08:00
|
|
|
return (true);
|
2016-03-28 18:06:35 +08:00
|
|
|
assert(rtree_elm_read(elm, false) == NULL);
|
|
|
|
rtree_elm_write(elm, extent);
|
|
|
|
|
|
|
|
return (false);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_ALWAYS_INLINE extent_t *
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_read(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key,
|
|
|
|
bool dependent)
|
2016-03-28 18:06:35 +08:00
|
|
|
{
|
|
|
|
rtree_elm_t *elm;
|
|
|
|
|
2016-06-03 09:43:10 +08:00
|
|
|
elm = rtree_elm_lookup(tsdn, rtree, rtree_ctx, key, dependent, false);
|
2016-03-28 18:06:35 +08:00
|
|
|
if (elm == NULL)
|
|
|
|
return (NULL);
|
|
|
|
|
|
|
|
return (rtree_elm_read(elm, dependent));
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE rtree_elm_t *
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_elm_acquire(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx,
|
|
|
|
uintptr_t key, bool dependent, bool init_missing)
|
2016-03-28 18:06:35 +08:00
|
|
|
{
|
|
|
|
rtree_elm_t *elm;
|
|
|
|
|
2016-06-03 09:43:10 +08:00
|
|
|
elm = rtree_elm_lookup(tsdn, rtree, rtree_ctx, key, dependent,
|
|
|
|
init_missing);
|
2016-03-28 18:06:35 +08:00
|
|
|
if (!dependent && elm == NULL)
|
|
|
|
return (NULL);
|
|
|
|
{
|
|
|
|
extent_t *extent;
|
|
|
|
void *s;
|
|
|
|
|
|
|
|
do {
|
|
|
|
extent = rtree_elm_read(elm, false);
|
|
|
|
/* The least significant bit serves as a lock. */
|
|
|
|
s = (void *)((uintptr_t)extent | (uintptr_t)0x1);
|
|
|
|
} while (atomic_cas_p(&elm->pun, (void *)extent, s));
|
2015-01-31 14:54:08 +08:00
|
|
|
}
|
2016-03-28 18:06:35 +08:00
|
|
|
|
2016-04-18 03:55:10 +08:00
|
|
|
if (config_debug)
|
|
|
|
rtree_elm_witness_acquire(tsdn, rtree, key, elm);
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
return (elm);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE extent_t *
|
2016-04-18 03:55:10 +08:00
|
|
|
rtree_elm_read_acquired(tsdn_t *tsdn, const rtree_t *rtree, rtree_elm_t *elm)
|
2016-03-28 18:06:35 +08:00
|
|
|
{
|
|
|
|
extent_t *extent;
|
|
|
|
|
|
|
|
assert(((uintptr_t)elm->pun & (uintptr_t)0x1) == (uintptr_t)0x1);
|
|
|
|
extent = (extent_t *)((uintptr_t)elm->pun & ~((uintptr_t)0x1));
|
|
|
|
assert(((uintptr_t)extent & (uintptr_t)0x1) == (uintptr_t)0x0);
|
|
|
|
|
2016-04-18 03:55:10 +08:00
|
|
|
if (config_debug)
|
|
|
|
rtree_elm_witness_access(tsdn, rtree, elm);
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
return (extent);
|
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE void
|
2016-04-18 03:55:10 +08:00
|
|
|
rtree_elm_write_acquired(tsdn_t *tsdn, const rtree_t *rtree, rtree_elm_t *elm,
|
|
|
|
const extent_t *extent)
|
2016-03-28 18:06:35 +08:00
|
|
|
{
|
|
|
|
assert(((uintptr_t)extent & (uintptr_t)0x1) == (uintptr_t)0x0);
|
|
|
|
assert(((uintptr_t)elm->pun & (uintptr_t)0x1) == (uintptr_t)0x1);
|
2016-04-18 03:55:10 +08:00
|
|
|
|
|
|
|
if (config_debug)
|
|
|
|
rtree_elm_witness_access(tsdn, rtree, elm);
|
|
|
|
|
2016-03-28 18:06:35 +08:00
|
|
|
elm->pun = (void *)((uintptr_t)extent | (uintptr_t)0x1);
|
2016-04-18 03:55:10 +08:00
|
|
|
assert(rtree_elm_read_acquired(tsdn, rtree, elm) == extent);
|
2016-03-28 18:06:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE void
|
2016-04-18 03:55:10 +08:00
|
|
|
rtree_elm_release(tsdn_t *tsdn, const rtree_t *rtree, rtree_elm_t *elm)
|
2016-03-28 18:06:35 +08:00
|
|
|
{
|
2016-04-18 03:55:10 +08:00
|
|
|
rtree_elm_write(elm, rtree_elm_read_acquired(tsdn, rtree, elm));
|
|
|
|
if (config_debug)
|
|
|
|
rtree_elm_witness_release(tsdn, rtree, elm);
|
2016-03-28 18:06:35 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
JEMALLOC_INLINE void
|
2016-06-03 09:43:10 +08:00
|
|
|
rtree_clear(tsdn_t *tsdn, rtree_t *rtree, rtree_ctx_t *rtree_ctx, uintptr_t key)
|
2016-03-28 18:06:35 +08:00
|
|
|
{
|
|
|
|
rtree_elm_t *elm;
|
|
|
|
|
2016-06-03 09:43:10 +08:00
|
|
|
elm = rtree_elm_acquire(tsdn, rtree, rtree_ctx, key, true, false);
|
2016-04-18 03:55:10 +08:00
|
|
|
rtree_elm_write_acquired(tsdn, rtree, elm, NULL);
|
|
|
|
rtree_elm_release(tsdn, rtree, elm);
|
2010-09-06 01:35:13 +08:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2017-01-11 10:06:31 +08:00
|
|
|
#endif /* JEMALLOC_INTERNAL_RTREE_INLINES_H */
|