2016-12-23 06:39:10 +08:00
|
|
|
#include "test/jemalloc_test.h"
|
|
|
|
|
2017-01-07 03:22:08 +08:00
|
|
|
#include "test/extent_hooks.h"
|
2016-12-23 06:39:10 +08:00
|
|
|
|
|
|
|
static extent_hooks_t hooks_null = {
|
|
|
|
extent_alloc_hook,
|
|
|
|
NULL, /* dalloc */
|
|
|
|
NULL, /* commit */
|
|
|
|
NULL, /* decommit */
|
|
|
|
NULL, /* purge_lazy */
|
|
|
|
NULL, /* purge_forced */
|
|
|
|
NULL, /* split */
|
|
|
|
NULL /* merge */
|
|
|
|
};
|
|
|
|
|
2017-01-07 03:22:08 +08:00
|
|
|
static extent_hooks_t hooks_not_null = {
|
|
|
|
extent_alloc_hook,
|
|
|
|
extent_dalloc_hook,
|
|
|
|
NULL, /* commit */
|
|
|
|
extent_decommit_hook,
|
|
|
|
extent_purge_lazy_hook,
|
|
|
|
extent_purge_forced_hook,
|
|
|
|
NULL, /* split */
|
|
|
|
NULL /* merge */
|
|
|
|
};
|
2016-12-23 06:39:10 +08:00
|
|
|
|
|
|
|
TEST_BEGIN(test_base_hooks_default)
|
|
|
|
{
|
|
|
|
tsdn_t *tsdn;
|
|
|
|
base_t *base;
|
|
|
|
size_t allocated0, allocated1, resident, mapped;
|
|
|
|
|
|
|
|
tsdn = tsdn_fetch();
|
|
|
|
base = base_new(tsdn, 0, (extent_hooks_t *)&extent_hooks_default);
|
|
|
|
|
|
|
|
base_stats_get(tsdn, base, &allocated0, &resident, &mapped);
|
|
|
|
assert_zu_ge(allocated0, sizeof(base_t),
|
|
|
|
"Base header should count as allocated");
|
|
|
|
|
|
|
|
assert_ptr_not_null(base_alloc(tsdn, base, 42, 1),
|
|
|
|
"Unexpected base_alloc() failure");
|
|
|
|
|
|
|
|
base_stats_get(tsdn, base, &allocated1, &resident, &mapped);
|
|
|
|
assert_zu_ge(allocated1 - allocated0, 42,
|
|
|
|
"At least 42 bytes were allocated by base_alloc()");
|
|
|
|
|
|
|
|
base_delete(base);
|
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
TEST_BEGIN(test_base_hooks_null)
|
|
|
|
{
|
2017-01-07 03:22:08 +08:00
|
|
|
extent_hooks_t hooks_orig;
|
2016-12-23 06:39:10 +08:00
|
|
|
tsdn_t *tsdn;
|
|
|
|
base_t *base;
|
|
|
|
size_t allocated0, allocated1, resident, mapped;
|
|
|
|
|
2017-01-07 03:22:08 +08:00
|
|
|
extent_hooks_prep();
|
|
|
|
try_dalloc = false;
|
|
|
|
try_decommit = false;
|
|
|
|
try_purge_lazy = false;
|
|
|
|
try_purge_forced = false;
|
|
|
|
memcpy(&hooks_orig, &hooks, sizeof(extent_hooks_t));
|
|
|
|
memcpy(&hooks, &hooks_null, sizeof(extent_hooks_t));
|
|
|
|
|
2016-12-23 06:39:10 +08:00
|
|
|
tsdn = tsdn_fetch();
|
2017-01-07 03:22:08 +08:00
|
|
|
base = base_new(tsdn, 0, &hooks);
|
2016-12-23 06:39:10 +08:00
|
|
|
assert_ptr_not_null(base, "Unexpected base_new() failure");
|
|
|
|
|
|
|
|
base_stats_get(tsdn, base, &allocated0, &resident, &mapped);
|
|
|
|
assert_zu_ge(allocated0, sizeof(base_t),
|
|
|
|
"Base header should count as allocated");
|
|
|
|
|
|
|
|
assert_ptr_not_null(base_alloc(tsdn, base, 42, 1),
|
|
|
|
"Unexpected base_alloc() failure");
|
|
|
|
|
|
|
|
base_stats_get(tsdn, base, &allocated1, &resident, &mapped);
|
|
|
|
assert_zu_ge(allocated1 - allocated0, 42,
|
|
|
|
"At least 42 bytes were allocated by base_alloc()");
|
|
|
|
|
|
|
|
base_delete(base);
|
2017-01-07 03:22:08 +08:00
|
|
|
|
|
|
|
memcpy(&hooks, &hooks_orig, sizeof(extent_hooks_t));
|
2016-12-23 06:39:10 +08:00
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
TEST_BEGIN(test_base_hooks_not_null)
|
|
|
|
{
|
2017-01-07 03:22:08 +08:00
|
|
|
extent_hooks_t hooks_orig;
|
2016-12-23 06:39:10 +08:00
|
|
|
tsdn_t *tsdn;
|
|
|
|
base_t *base;
|
|
|
|
void *p, *q, *r, *r_exp;
|
|
|
|
|
2017-01-07 03:22:08 +08:00
|
|
|
extent_hooks_prep();
|
|
|
|
try_dalloc = false;
|
|
|
|
try_decommit = false;
|
|
|
|
try_purge_lazy = false;
|
|
|
|
try_purge_forced = false;
|
|
|
|
memcpy(&hooks_orig, &hooks, sizeof(extent_hooks_t));
|
|
|
|
memcpy(&hooks, &hooks_not_null, sizeof(extent_hooks_t));
|
|
|
|
|
2016-12-23 06:39:10 +08:00
|
|
|
tsdn = tsdn_fetch();
|
|
|
|
did_alloc = false;
|
2017-01-07 03:22:08 +08:00
|
|
|
base = base_new(tsdn, 0, &hooks);
|
2016-12-23 06:39:10 +08:00
|
|
|
assert_ptr_not_null(base, "Unexpected base_new() failure");
|
2017-01-07 03:22:08 +08:00
|
|
|
assert_true(did_alloc, "Expected alloc");
|
2016-12-23 06:39:10 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Check for tight packing at specified alignment under simple
|
|
|
|
* conditions.
|
|
|
|
*/
|
|
|
|
{
|
|
|
|
const size_t alignments[] = {
|
|
|
|
1,
|
|
|
|
QUANTUM,
|
|
|
|
QUANTUM << 1,
|
|
|
|
CACHELINE,
|
|
|
|
CACHELINE << 1,
|
|
|
|
};
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
for (i = 0; i < sizeof(alignments) / sizeof(size_t); i++) {
|
|
|
|
size_t alignment = alignments[i];
|
|
|
|
size_t align_ceil = ALIGNMENT_CEILING(alignment,
|
|
|
|
QUANTUM);
|
|
|
|
p = base_alloc(tsdn, base, 1, alignment);
|
|
|
|
assert_ptr_not_null(p,
|
|
|
|
"Unexpected base_alloc() failure");
|
|
|
|
assert_ptr_eq(p,
|
|
|
|
(void *)(ALIGNMENT_CEILING((uintptr_t)p,
|
|
|
|
alignment)), "Expected quantum alignment");
|
|
|
|
q = base_alloc(tsdn, base, alignment, alignment);
|
|
|
|
assert_ptr_not_null(q,
|
|
|
|
"Unexpected base_alloc() failure");
|
|
|
|
assert_ptr_eq((void *)((uintptr_t)p + align_ceil), q,
|
|
|
|
"Minimal allocation should take up %zu bytes",
|
|
|
|
align_ceil);
|
|
|
|
r = base_alloc(tsdn, base, 1, alignment);
|
|
|
|
assert_ptr_not_null(r,
|
|
|
|
"Unexpected base_alloc() failure");
|
|
|
|
assert_ptr_eq((void *)((uintptr_t)q + align_ceil), r,
|
|
|
|
"Minimal allocation should take up %zu bytes",
|
|
|
|
align_ceil);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate an object that cannot fit in the first block, then verify
|
|
|
|
* that the first block's remaining space is considered for subsequent
|
|
|
|
* allocation.
|
|
|
|
*/
|
|
|
|
assert_zu_ge(extent_size_get(&base->blocks->extent), QUANTUM,
|
|
|
|
"Remainder insufficient for test");
|
|
|
|
/* Use up all but one quantum of block. */
|
|
|
|
while (extent_size_get(&base->blocks->extent) > QUANTUM) {
|
|
|
|
p = base_alloc(tsdn, base, QUANTUM, QUANTUM);
|
|
|
|
assert_ptr_not_null(p, "Unexpected base_alloc() failure");
|
|
|
|
}
|
|
|
|
r_exp = extent_addr_get(&base->blocks->extent);
|
|
|
|
assert_zu_eq(base->extent_sn_next, 1, "One extant block expected");
|
|
|
|
q = base_alloc(tsdn, base, QUANTUM + 1, QUANTUM);
|
|
|
|
assert_ptr_not_null(q, "Unexpected base_alloc() failure");
|
|
|
|
assert_ptr_ne(q, r_exp, "Expected allocation from new block");
|
|
|
|
assert_zu_eq(base->extent_sn_next, 2, "Two extant blocks expected");
|
|
|
|
r = base_alloc(tsdn, base, QUANTUM, QUANTUM);
|
|
|
|
assert_ptr_not_null(r, "Unexpected base_alloc() failure");
|
|
|
|
assert_ptr_eq(r, r_exp, "Expected allocation from first block");
|
|
|
|
assert_zu_eq(base->extent_sn_next, 2, "Two extant blocks expected");
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check for proper alignment support when normal blocks are too small.
|
|
|
|
*/
|
|
|
|
{
|
|
|
|
const size_t alignments[] = {
|
|
|
|
HUGEPAGE,
|
|
|
|
HUGEPAGE << 1
|
|
|
|
};
|
|
|
|
unsigned i;
|
|
|
|
|
|
|
|
for (i = 0; i < sizeof(alignments) / sizeof(size_t); i++) {
|
|
|
|
size_t alignment = alignments[i];
|
|
|
|
p = base_alloc(tsdn, base, QUANTUM, alignment);
|
|
|
|
assert_ptr_not_null(p,
|
|
|
|
"Unexpected base_alloc() failure");
|
|
|
|
assert_ptr_eq(p,
|
|
|
|
(void *)(ALIGNMENT_CEILING((uintptr_t)p,
|
|
|
|
alignment)), "Expected %zu-byte alignment",
|
|
|
|
alignment);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-01-07 03:22:08 +08:00
|
|
|
called_dalloc = called_decommit = called_purge_lazy =
|
|
|
|
called_purge_forced = false;
|
2016-12-23 06:39:10 +08:00
|
|
|
base_delete(base);
|
2017-01-07 03:22:08 +08:00
|
|
|
assert_true(called_dalloc, "Expected dalloc call");
|
|
|
|
assert_true(called_decommit, "Expected decommit call");
|
|
|
|
assert_true(called_purge_lazy, "Expected purge_lazy call");
|
|
|
|
assert_true(called_purge_forced, "Expected purge_forced call");
|
|
|
|
|
|
|
|
try_dalloc = true;
|
|
|
|
try_decommit = true;
|
|
|
|
try_purge_lazy = true;
|
|
|
|
try_purge_forced = true;
|
|
|
|
memcpy(&hooks, &hooks_orig, sizeof(extent_hooks_t));
|
2016-12-23 06:39:10 +08:00
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
int
|
|
|
|
main(void)
|
|
|
|
{
|
|
|
|
return (test(
|
|
|
|
test_base_hooks_default,
|
|
|
|
test_base_hooks_null,
|
|
|
|
test_base_hooks_not_null));
|
|
|
|
}
|