2014-01-08 08:47:56 +08:00
|
|
|
#include "test/jemalloc_test.h"
|
|
|
|
|
|
|
|
#ifdef JEMALLOC_FILL
|
|
|
|
const char *malloc_conf =
|
|
|
|
"abort:false,junk:true,zero:false,redzone:true,quarantine:0";
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static arena_dalloc_junk_small_t *arena_dalloc_junk_small_orig;
|
|
|
|
static arena_dalloc_junk_large_t *arena_dalloc_junk_large_orig;
|
|
|
|
static huge_dalloc_junk_t *huge_dalloc_junk_orig;
|
|
|
|
static void *most_recently_junked;
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_dalloc_junk_small_intercept(void *ptr, arena_bin_info_t *bin_info)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
arena_dalloc_junk_small_orig(ptr, bin_info);
|
|
|
|
for (i = 0; i < bin_info->reg_size; i++) {
|
|
|
|
assert_c_eq(((char *)ptr)[i], 0x5a,
|
|
|
|
"Missing junk fill for byte %zu/%zu of deallocated region",
|
|
|
|
i, bin_info->reg_size);
|
|
|
|
}
|
|
|
|
most_recently_junked = ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_dalloc_junk_large_intercept(void *ptr, size_t usize)
|
|
|
|
{
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
arena_dalloc_junk_large_orig(ptr, usize);
|
|
|
|
for (i = 0; i < usize; i++) {
|
|
|
|
assert_c_eq(((char *)ptr)[i], 0x5a,
|
|
|
|
"Missing junk fill for byte %zu/%zu of deallocated region",
|
|
|
|
i, usize);
|
|
|
|
}
|
|
|
|
most_recently_junked = ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
huge_dalloc_junk_intercept(void *ptr, size_t usize)
|
|
|
|
{
|
|
|
|
|
|
|
|
huge_dalloc_junk_orig(ptr, usize);
|
|
|
|
/*
|
|
|
|
* The conditions under which junk filling actually occurs are nuanced
|
|
|
|
* enough that it doesn't make sense to duplicate the decision logic in
|
|
|
|
* test code, so don't actually check that the region is junk-filled.
|
|
|
|
*/
|
|
|
|
most_recently_junked = ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
test_junk(size_t sz_min, size_t sz_max)
|
|
|
|
{
|
|
|
|
char *s;
|
|
|
|
size_t sz_prev, sz, i;
|
|
|
|
|
|
|
|
arena_dalloc_junk_small_orig = arena_dalloc_junk_small;
|
|
|
|
arena_dalloc_junk_small = arena_dalloc_junk_small_intercept;
|
|
|
|
arena_dalloc_junk_large_orig = arena_dalloc_junk_large;
|
|
|
|
arena_dalloc_junk_large = arena_dalloc_junk_large_intercept;
|
|
|
|
huge_dalloc_junk_orig = huge_dalloc_junk;
|
|
|
|
huge_dalloc_junk = huge_dalloc_junk_intercept;
|
|
|
|
|
|
|
|
sz_prev = 0;
|
|
|
|
s = (char *)mallocx(sz_min, 0);
|
|
|
|
assert_ptr_not_null((void *)s, "Unexpected mallocx() failure");
|
|
|
|
|
|
|
|
for (sz = sallocx(s, 0); sz <= sz_max;
|
|
|
|
sz_prev = sz, sz = sallocx(s, 0)) {
|
|
|
|
if (sz_prev > 0) {
|
|
|
|
assert_c_eq(s[0], 'a',
|
|
|
|
"Previously allocated byte %zu/%zu is corrupted",
|
|
|
|
0, sz_prev);
|
|
|
|
assert_c_eq(s[sz_prev-1], 'a',
|
|
|
|
"Previously allocated byte %zu/%zu is corrupted",
|
|
|
|
sz_prev-1, sz_prev);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = sz_prev; i < sz; i++) {
|
|
|
|
assert_c_eq(s[i], 0xa5,
|
|
|
|
"Newly allocated byte %zu/%zu isn't junk-filled",
|
|
|
|
i, sz);
|
|
|
|
s[i] = 'a';
|
|
|
|
}
|
|
|
|
|
|
|
|
if (xallocx(s, sz+1, 0, 0) == sz) {
|
|
|
|
void *junked = (void *)s;
|
|
|
|
|
|
|
|
s = (char *)rallocx(s, sz+1, 0);
|
|
|
|
assert_ptr_not_null((void *)s,
|
|
|
|
"Unexpected rallocx() failure");
|
2014-02-26 03:58:50 +08:00
|
|
|
if (!config_mremap || sz+1 <= arena_maxclass) {
|
|
|
|
assert_ptr_eq(most_recently_junked, junked,
|
|
|
|
"Expected region of size %zu to be "
|
|
|
|
"junk-filled",
|
|
|
|
sz);
|
|
|
|
}
|
2014-01-08 08:47:56 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
dallocx(s, 0);
|
|
|
|
assert_ptr_eq(most_recently_junked, (void *)s,
|
|
|
|
"Expected region of size %zu to be junk-filled", sz);
|
|
|
|
|
|
|
|
arena_dalloc_junk_small = arena_dalloc_junk_small_orig;
|
|
|
|
arena_dalloc_junk_large = arena_dalloc_junk_large_orig;
|
|
|
|
huge_dalloc_junk = huge_dalloc_junk_orig;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_BEGIN(test_junk_small)
|
|
|
|
{
|
|
|
|
|
|
|
|
test_skip_if(!config_fill);
|
|
|
|
test_junk(1, SMALL_MAXCLASS-1);
|
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
TEST_BEGIN(test_junk_large)
|
|
|
|
{
|
|
|
|
|
|
|
|
test_skip_if(!config_fill);
|
|
|
|
test_junk(SMALL_MAXCLASS+1, arena_maxclass);
|
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
TEST_BEGIN(test_junk_huge)
|
|
|
|
{
|
|
|
|
|
|
|
|
test_skip_if(!config_fill);
|
|
|
|
test_junk(arena_maxclass+1, chunksize*2);
|
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
arena_ralloc_junk_large_t *arena_ralloc_junk_large_orig;
|
|
|
|
static void *most_recently_trimmed;
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_ralloc_junk_large_intercept(void *ptr, size_t old_usize, size_t usize)
|
|
|
|
{
|
|
|
|
|
|
|
|
arena_ralloc_junk_large_orig(ptr, old_usize, usize);
|
|
|
|
assert_zu_eq(old_usize, arena_maxclass, "Unexpected old_usize");
|
|
|
|
assert_zu_eq(usize, arena_maxclass-PAGE, "Unexpected usize");
|
|
|
|
most_recently_trimmed = ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_BEGIN(test_junk_large_ralloc_shrink)
|
|
|
|
{
|
|
|
|
void *p1, *p2;
|
|
|
|
|
|
|
|
p1 = mallocx(arena_maxclass, 0);
|
|
|
|
assert_ptr_not_null(p1, "Unexpected mallocx() failure");
|
|
|
|
|
|
|
|
arena_ralloc_junk_large_orig = arena_ralloc_junk_large;
|
|
|
|
arena_ralloc_junk_large = arena_ralloc_junk_large_intercept;
|
|
|
|
|
|
|
|
p2 = rallocx(p1, arena_maxclass-PAGE, 0);
|
|
|
|
assert_ptr_eq(p1, p2, "Unexpected move during shrink");
|
|
|
|
|
|
|
|
arena_ralloc_junk_large = arena_ralloc_junk_large_orig;
|
|
|
|
|
|
|
|
assert_ptr_eq(most_recently_trimmed, p1,
|
|
|
|
"Expected trimmed portion of region to be junk-filled");
|
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
static bool detected_redzone_corruption;
|
|
|
|
|
|
|
|
static void
|
|
|
|
arena_redzone_corruption_replacement(void *ptr, size_t usize, bool after,
|
|
|
|
size_t offset, uint8_t byte)
|
|
|
|
{
|
|
|
|
|
|
|
|
detected_redzone_corruption = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_BEGIN(test_junk_redzone)
|
|
|
|
{
|
|
|
|
char *s;
|
|
|
|
arena_redzone_corruption_t *arena_redzone_corruption_orig;
|
|
|
|
|
|
|
|
test_skip_if(!config_fill);
|
|
|
|
|
|
|
|
arena_redzone_corruption_orig = arena_redzone_corruption;
|
|
|
|
arena_redzone_corruption = arena_redzone_corruption_replacement;
|
|
|
|
|
|
|
|
/* Test underflow. */
|
|
|
|
detected_redzone_corruption = false;
|
|
|
|
s = (char *)mallocx(1, 0);
|
|
|
|
assert_ptr_not_null((void *)s, "Unexpected mallocx() failure");
|
|
|
|
s[-1] = 0xbb;
|
|
|
|
dallocx(s, 0);
|
|
|
|
assert_true(detected_redzone_corruption,
|
|
|
|
"Did not detect redzone corruption");
|
|
|
|
|
|
|
|
/* Test overflow. */
|
|
|
|
detected_redzone_corruption = false;
|
|
|
|
s = (char *)mallocx(1, 0);
|
|
|
|
assert_ptr_not_null((void *)s, "Unexpected mallocx() failure");
|
|
|
|
s[sallocx(s, 0)] = 0xbb;
|
|
|
|
dallocx(s, 0);
|
|
|
|
assert_true(detected_redzone_corruption,
|
|
|
|
"Did not detect redzone corruption");
|
|
|
|
|
|
|
|
arena_redzone_corruption = arena_redzone_corruption_orig;
|
|
|
|
}
|
|
|
|
TEST_END
|
|
|
|
|
|
|
|
int
|
|
|
|
main(void)
|
|
|
|
{
|
|
|
|
|
|
|
|
return (test(
|
|
|
|
test_junk_small,
|
|
|
|
test_junk_large,
|
|
|
|
test_junk_huge,
|
|
|
|
test_junk_large_ralloc_shrink,
|
|
|
|
test_junk_redzone));
|
|
|
|
}
|