#include "test/jemalloc_test.h" rtree_node_alloc_t *rtree_node_alloc_orig; rtree_node_dalloc_t *rtree_node_dalloc_orig; rtree_t *test_rtree; static rtree_elm_t * rtree_node_alloc_intercept(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { rtree_elm_t *node; if (rtree != test_rtree) return rtree_node_alloc_orig(tsdn, rtree, nelms); node = (rtree_elm_t *)calloc(nelms, sizeof(rtree_elm_t)); assert_ptr_not_null(node, "Unexpected calloc() failure"); return (node); } static void rtree_node_dalloc_intercept(tsdn_t *tsdn, rtree_t *rtree, rtree_elm_t *node) { if (rtree != test_rtree) { rtree_node_dalloc_orig(tsdn, rtree, node); return; } free(node); } TEST_BEGIN(test_rtree_read_empty) { tsdn_t *tsdn; unsigned i; tsdn = tsdn_fetch(); for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { rtree_t rtree; rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER; test_rtree = &rtree; assert_false(rtree_new(&rtree, i), "Unexpected rtree_new() failure"); assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx, 0, false), "rtree_read() should return NULL for empty tree"); rtree_delete(tsdn, &rtree); test_rtree = NULL; } } TEST_END #define NTHREADS 8 #define MAX_NBITS 18 #define NITERS 1000 #define SEED 42 typedef struct { unsigned nbits; rtree_t rtree; uint32_t seed; } thd_start_arg_t; static void * thd_start(void *varg) { thd_start_arg_t *arg = (thd_start_arg_t *)varg; rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER; sfmt_t *sfmt; extent_t *extent; tsdn_t *tsdn; unsigned i; sfmt = init_gen_rand(arg->seed); extent = (extent_t *)malloc(sizeof(extent)); assert_ptr_not_null(extent, "Unexpected malloc() failure"); tsdn = tsdn_fetch(); for (i = 0; i < NITERS; i++) { uintptr_t key = (uintptr_t)gen_rand64(sfmt); if (i % 2 == 0) { rtree_elm_t *elm; elm = rtree_elm_acquire(tsdn, &arg->rtree, &rtree_ctx, key, false, true); assert_ptr_not_null(elm, "Unexpected rtree_elm_acquire() failure"); rtree_elm_write_acquired(tsdn, &arg->rtree, elm, extent); rtree_elm_release(tsdn, &arg->rtree, elm); elm = rtree_elm_acquire(tsdn, &arg->rtree, &rtree_ctx, key, true, false); assert_ptr_not_null(elm, "Unexpected rtree_elm_acquire() failure"); rtree_elm_read_acquired(tsdn, &arg->rtree, elm); rtree_elm_release(tsdn, &arg->rtree, elm); } else rtree_read(tsdn, &arg->rtree, &rtree_ctx, key, false); } free(extent); fini_gen_rand(sfmt); return (NULL); } TEST_BEGIN(test_rtree_concurrent) { thd_start_arg_t arg; thd_t thds[NTHREADS]; sfmt_t *sfmt; tsdn_t *tsdn; unsigned i, j; sfmt = init_gen_rand(SEED); tsdn = tsdn_fetch(); for (i = 1; i < MAX_NBITS; i++) { arg.nbits = i; test_rtree = &arg.rtree; assert_false(rtree_new(&arg.rtree, arg.nbits), "Unexpected rtree_new() failure"); arg.seed = gen_rand32(sfmt); for (j = 0; j < NTHREADS; j++) thd_create(&thds[j], thd_start, (void *)&arg); for (j = 0; j < NTHREADS; j++) thd_join(thds[j], NULL); rtree_delete(tsdn, &arg.rtree); test_rtree = NULL; } fini_gen_rand(sfmt); } TEST_END #undef NTHREADS #undef MAX_NBITS #undef NITERS #undef SEED TEST_BEGIN(test_rtree_extrema) { unsigned i; extent_t extent_a, extent_b; tsdn_t *tsdn; tsdn = tsdn_fetch(); for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { rtree_t rtree; rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER; test_rtree = &rtree; assert_false(rtree_new(&rtree, i), "Unexpected rtree_new() failure"); assert_false(rtree_write(tsdn, &rtree, &rtree_ctx, 0, &extent_a), "Unexpected rtree_write() failure, i=%u", i); assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx, 0, true), &extent_a, "rtree_read() should return previously set value, i=%u", i); assert_false(rtree_write(tsdn, &rtree, &rtree_ctx, ~((uintptr_t)0), &extent_b), "Unexpected rtree_write() failure, i=%u", i); assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx, ~((uintptr_t)0), true), &extent_b, "rtree_read() should return previously set value, i=%u", i); rtree_delete(tsdn, &rtree); test_rtree = NULL; } } TEST_END TEST_BEGIN(test_rtree_bits) { tsdn_t *tsdn; unsigned i, j, k; tsdn = tsdn_fetch(); for (i = 1; i < (sizeof(uintptr_t) << 3); i++) { uintptr_t keys[] = {0, 1, (((uintptr_t)1) << (sizeof(uintptr_t)*8-i)) - 1}; extent_t extent; rtree_t rtree; rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER; test_rtree = &rtree; assert_false(rtree_new(&rtree, i), "Unexpected rtree_new() failure"); for (j = 0; j < sizeof(keys)/sizeof(uintptr_t); j++) { assert_false(rtree_write(tsdn, &rtree, &rtree_ctx, keys[j], &extent), "Unexpected rtree_write() failure"); for (k = 0; k < sizeof(keys)/sizeof(uintptr_t); k++) { assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx, keys[k], true), &extent, "rtree_read() should return previously set " "value and ignore insignificant key bits; " "i=%u, j=%u, k=%u, set key=%#"FMTxPTR", " "get key=%#"FMTxPTR, i, j, k, keys[j], keys[k]); } assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx, (((uintptr_t)1) << (sizeof(uintptr_t)*8-i)), false), "Only leftmost rtree leaf should be set; " "i=%u, j=%u", i, j); rtree_clear(tsdn, &rtree, &rtree_ctx, keys[j]); } rtree_delete(tsdn, &rtree); test_rtree = NULL; } } TEST_END TEST_BEGIN(test_rtree_random) { unsigned i; sfmt_t *sfmt; tsdn_t *tsdn; #define NSET 16 #define SEED 42 sfmt = init_gen_rand(SEED); tsdn = tsdn_fetch(); for (i = 1; i <= (sizeof(uintptr_t) << 3); i++) { uintptr_t keys[NSET]; extent_t extent; unsigned j; rtree_t rtree; rtree_ctx_t rtree_ctx = RTREE_CTX_INITIALIZER; rtree_elm_t *elm; test_rtree = &rtree; assert_false(rtree_new(&rtree, i), "Unexpected rtree_new() failure"); for (j = 0; j < NSET; j++) { keys[j] = (uintptr_t)gen_rand64(sfmt); elm = rtree_elm_acquire(tsdn, &rtree, &rtree_ctx, keys[j], false, true); assert_ptr_not_null(elm, "Unexpected rtree_elm_acquire() failure"); rtree_elm_write_acquired(tsdn, &rtree, elm, &extent); rtree_elm_release(tsdn, &rtree, elm); assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx, keys[j], true), &extent, "rtree_read() should return previously set value"); } for (j = 0; j < NSET; j++) { assert_ptr_eq(rtree_read(tsdn, &rtree, &rtree_ctx, keys[j], true), &extent, "rtree_read() should return previously set value, " "j=%u", j); } for (j = 0; j < NSET; j++) { rtree_clear(tsdn, &rtree, &rtree_ctx, keys[j]); assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx, keys[j], true), "rtree_read() should return previously set value"); } for (j = 0; j < NSET; j++) { assert_ptr_null(rtree_read(tsdn, &rtree, &rtree_ctx, keys[j], true), "rtree_read() should return previously set value"); } rtree_delete(tsdn, &rtree); test_rtree = NULL; } fini_gen_rand(sfmt); #undef NSET #undef SEED } TEST_END int main(void) { rtree_node_alloc_orig = rtree_node_alloc; rtree_node_alloc = rtree_node_alloc_intercept; rtree_node_dalloc_orig = rtree_node_dalloc; rtree_node_dalloc = rtree_node_dalloc_intercept; test_rtree = NULL; return (test( test_rtree_read_empty, test_rtree_concurrent, test_rtree_extrema, test_rtree_bits, test_rtree_random)); }