server-skynet-source-3rd-je.../test/unit/rtree.c

300 lines
7.3 KiB
C
Raw Normal View History

2014-01-03 08:08:28 +08:00
#include "test/jemalloc_test.h"
rtree_node_alloc_t *rtree_node_alloc_orig;
rtree_node_dalloc_t *rtree_node_dalloc_orig;
rtree_t *test_rtree;
static rtree_elm_t *
rtree_node_alloc_intercept(tsdn_t *tsdn, rtree_t *rtree, size_t nelms)
{
rtree_elm_t *node;
if (rtree != test_rtree)
return rtree_node_alloc_orig(tsdn, rtree, nelms);
malloc_mutex_unlock(tsdn, &rtree->init_lock);
node = (rtree_elm_t *)calloc(nelms, sizeof(rtree_elm_t));
assert_ptr_not_null(node, "Unexpected calloc() failure");
malloc_mutex_lock(tsdn, &rtree->init_lock);
return (node);
}
static void
rtree_node_dalloc_intercept(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *node)
{
if (rtree != test_rtree) {
rtree_node_dalloc_orig(tsdn, rtree, node);
return;
}
free(node);
}
TEST_BEGIN(test_rtree_read_empty)
2014-01-03 08:08:28 +08:00
{
tsdn_t *tsdn;
2014-01-03 08:08:28 +08:00
unsigned i;
tsdn = tsdn_fetch();
2014-01-03 08:08:28 +08:00
for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) {
rtree_t rtree;
rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER;
test_rtree = &rtree;
assert_false(rtree_new(&rtree, i),
"Unexpected rtree_new() failure");
assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx, 0, false),
"rtree_read() should return NULL for empty tree");
rtree_delete(tsdn, &rtree);
test_rtree = NULL;
2014-01-03 08:08:28 +08:00
}
}
TEST_END
#define NTHREADS 8
#define MAX_NBITS 18
#define NITERS 1000
#define SEED 42
typedef struct {
unsigned nbits;
rtree_t rtree;
uint32_t seed;
} thd_start_arg_t;
static void *
thd_start(void *varg)
{
thd_start_arg_t *arg = (thd_start_arg_t *)varg;
rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER;
sfmt_t *sfmt;
extent_t *extent;
tsdn_t *tsdn;
unsigned i;
sfmt = init_gen_rand(arg->seed);
extent = (extent_t *)malloc(sizeof(extent));
assert_ptr_not_null(extent, "Unexpected malloc() failure");
tsdn = tsdn_fetch();
for (i = 0; i < NITERS; i++) {
uintptr_t key = (uintptr_t)gen_rand64(sfmt);
if (i % 2 == 0) {
rtree_elm_t *elm;
elm = rtree_elm_acquire(tsdn, &arg->rtree, &rtree_ctx,
key, false, true);
assert_ptr_not_null(elm,
"Unexpected rtree_elm_acquire() failure");
rtree_elm_write_acquired(tsdn, &arg->rtree, elm,
extent);
2016-04-18 03:55:10 +08:00
rtree_elm_release(tsdn, &arg->rtree, elm);
elm = rtree_elm_acquire(tsdn, &arg->rtree, &rtree_ctx,
key, true, false);
assert_ptr_not_null(elm,
"Unexpected rtree_elm_acquire() failure");
2016-04-18 03:55:10 +08:00
rtree_elm_read_acquired(tsdn, &arg->rtree, elm);
rtree_elm_release(tsdn, &arg->rtree, elm);
} else
rtree_read(tsdn, &arg->rtree, &rtree_ctx, key, false);
}
free(extent);
fini_gen_rand(sfmt);
return (NULL);
}
TEST_BEGIN(test_rtree_concurrent)
{
thd_start_arg_t arg;
thd_t thds[NTHREADS];
sfmt_t *sfmt;
tsdn_t *tsdn;
unsigned i, j;
sfmt = init_gen_rand(SEED);
tsdn = tsdn_fetch();
for (i = 1; i < MAX_NBITS; i++) {
arg.nbits = i;
test_rtree = &arg.rtree;
assert_false(rtree_new(&arg.rtree, arg.nbits),
"Unexpected rtree_new() failure");
arg.seed = gen_rand32(sfmt);
for (j = 0; j < NTHREADS; j++)
thd_create(&thds[j], thd_start, (void *)&arg);
for (j = 0; j < NTHREADS; j++)
thd_join(thds[j], NULL);
rtree_delete(tsdn, &arg.rtree);
test_rtree = NULL;
}
fini_gen_rand(sfmt);
}
TEST_END
#undef NTHREADS
#undef MAX_NBITS
#undef NITERS
#undef SEED
2014-01-03 08:08:28 +08:00
TEST_BEGIN(test_rtree_extrema)
{
unsigned i;
2016-03-24 12:09:28 +08:00
extent_t extent_a, extent_b;
tsdn_t *tsdn;
tsdn = tsdn_fetch();
2014-01-03 08:08:28 +08:00
for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) {
rtree_t rtree;
rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER;
test_rtree = &rtree;
assert_false(rtree_new(&rtree, i),
"Unexpected rtree_new() failure");
2014-01-03 08:08:28 +08:00
assert_false(rtree_write(tsdn, &rtree, &rtree_ctx, 0,
&extent_a), "Unexpected rtree_write() failure, i=%u", i);
assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx, 0, true),
&extent_a,
"rtree_read() should return previously set value, i=%u", i);
2014-01-03 08:08:28 +08:00
assert_false(rtree_write(tsdn, &rtree, &rtree_ctx,
~((uintptr_t)0), &extent_b),
"Unexpected rtree_write() failure, i=%u", i);
assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx,
~((uintptr_t)0), true), &extent_b,
"rtree_read() should return previously set value, i=%u", i);
2014-01-03 08:08:28 +08:00
rtree_delete(tsdn, &rtree);
test_rtree = NULL;
2014-01-03 08:08:28 +08:00
}
}
TEST_END
TEST_BEGIN(test_rtree_bits)
{
tsdn_t *tsdn;
2014-01-03 08:08:28 +08:00
unsigned i, j, k;
tsdn = tsdn_fetch();
2014-01-03 08:08:28 +08:00
for (i = 1; i < (sizeof(uintptr_t) << 3); i++) {
uintptr_t keys[] = {0, 1,
(((uintptr_t)1) << (sizeof(uintptr_t)*8-i)) - 1};
2016-03-24 12:09:28 +08:00
extent_t extent;
rtree_t rtree;
rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER;
test_rtree = &rtree;
assert_false(rtree_new(&rtree, i),
"Unexpected rtree_new() failure");
2014-01-03 08:08:28 +08:00
for (j = 0; j < sizeof(keys)/sizeof(uintptr_t); j++) {
assert_false(rtree_write(tsdn, &rtree, &rtree_ctx,
keys[j], &extent),
"Unexpected rtree_write() failure");
2014-01-03 08:08:28 +08:00
for (k = 0; k < sizeof(keys)/sizeof(uintptr_t); k++) {
assert_ptr_eq(rtree_read(tsdn, &rtree,
&rtree_ctx, keys[k], true), &extent,
"rtree_read() should return previously set "
"value and ignore insignificant key bits; "
"i=%u, j=%u, k=%u, set key=%#"FMTxPTR", "
"get key=%#"FMTxPTR, i, j, k, keys[j],
keys[k]);
2014-01-03 08:08:28 +08:00
}
assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx,
(((uintptr_t)1) << (sizeof(uintptr_t)*8-i)), false),
2014-01-03 08:08:28 +08:00
"Only leftmost rtree leaf should be set; "
"i=%u, j=%u", i, j);
rtree_clear(tsdn, &rtree, &rtree_ctx, keys[j]);
2014-01-03 08:08:28 +08:00
}
rtree_delete(tsdn, &rtree);
test_rtree = NULL;
2014-01-03 08:08:28 +08:00
}
}
TEST_END
TEST_BEGIN(test_rtree_random)
{
unsigned i;
sfmt_t *sfmt;
tsdn_t *tsdn;
#define NSET 16
2014-01-03 08:08:28 +08:00
#define SEED 42
sfmt = init_gen_rand(SEED);
tsdn = tsdn_fetch();
2014-01-03 08:08:28 +08:00
for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) {
uintptr_t keys[NSET];
2016-03-24 12:09:28 +08:00
extent_t extent;
2014-01-03 08:08:28 +08:00
unsigned j;
rtree_t rtree;
rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER;
rtree_elm_t *elm;
test_rtree = &rtree;
assert_false(rtree_new(&rtree, i),
"Unexpected rtree_new() failure");
2014-01-03 08:08:28 +08:00
for (j = 0; j < NSET; j++) {
keys[j] = (uintptr_t)gen_rand64(sfmt);
elm = rtree_elm_acquire(tsdn, &rtree, &rtree_ctx,
keys[j], false, true);
assert_ptr_not_null(elm,
"Unexpected rtree_elm_acquire() failure");
2016-04-18 03:55:10 +08:00
rtree_elm_write_acquired(tsdn, &rtree, elm, &extent);
rtree_elm_release(tsdn, &rtree, elm);
assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx,
keys[j], true), &extent,
"rtree_read() should return previously set value");
2014-01-03 08:08:28 +08:00
}
for (j = 0; j < NSET; j++) {
assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx,
keys[j], true), &extent,
"rtree_read() should return previously set value, "
"j=%u", j);
2014-01-03 08:08:28 +08:00
}
for (j = 0; j < NSET; j++) {
rtree_clear(tsdn, &rtree, &rtree_ctx, keys[j]);
assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx,
keys[j], true),
"rtree_read() should return previously set value");
2014-01-03 08:08:28 +08:00
}
for (j = 0; j < NSET; j++) {
assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx,
keys[j], true),
"rtree_read() should return previously set value");
2014-01-03 08:08:28 +08:00
}
rtree_delete(tsdn, &rtree);
test_rtree = NULL;
2014-01-03 08:08:28 +08:00
}
fini_gen_rand(sfmt);
#undef NSET
#undef SEED
}
TEST_END
int
main(void)
{
rtree_node_alloc_orig = rtree_node_alloc;
rtree_node_alloc = rtree_node_alloc_intercept;
rtree_node_dalloc_orig = rtree_node_dalloc;
rtree_node_dalloc = rtree_node_dalloc_intercept;
test_rtree = NULL;
2014-01-03 08:08:28 +08:00
return (test(
test_rtree_read_empty,
test_rtree_concurrent,
2014-01-03 08:08:28 +08:00
test_rtree_extrema,
test_rtree_bits,
test_rtree_random));
}