server-skynet-source-3rd-je.../test/unit/prof_recent.c

679 lines
18 KiB
C
Raw Permalink Normal View History

2019-12-19 05:38:14 +08:00
#include "test/jemalloc_test.h"
#include "jemalloc/internal/prof_recent.h"
/* As specified in the shell script */
#define OPT_ALLOC_MAX 3
2019-12-19 05:38:14 +08:00
/* Invariant before and after every test (when config_prof is on) */
static void
confirm_prof_setup() {
2019-12-19 05:38:14 +08:00
/* Options */
assert_true(opt_prof, "opt_prof not on");
assert_true(opt_prof_active, "opt_prof_active not on");
assert_zd_eq(opt_prof_recent_alloc_max, OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"opt_prof_recent_alloc_max not set correctly");
/* Dynamics */
assert_true(prof_active_state, "prof_active not on");
assert_zd_eq(prof_recent_alloc_max_ctl_read(), OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"prof_recent_alloc_max not set correctly");
}
TEST_BEGIN(test_confirm_setup) {
test_skip_if(!config_prof);
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
}
TEST_END
TEST_BEGIN(test_prof_recent_off) {
test_skip_if(config_prof);
const ssize_t past_ref = 0, future_ref = 0;
const size_t len_ref = sizeof(ssize_t);
ssize_t past = past_ref, future = future_ref;
size_t len = len_ref;
#define ASSERT_SHOULD_FAIL(opt, a, b, c, d) do { \
assert_d_eq(mallctl("experimental.prof_recent." opt, a, b, c, \
2019-12-19 05:38:14 +08:00
d), ENOENT, "Should return ENOENT when config_prof is off");\
assert_zd_eq(past, past_ref, "output was touched"); \
assert_zu_eq(len, len_ref, "output length was touched"); \
assert_zd_eq(future, future_ref, "input was touched"); \
2019-12-19 05:38:14 +08:00
} while (0)
ASSERT_SHOULD_FAIL("alloc_max", NULL, NULL, NULL, 0);
ASSERT_SHOULD_FAIL("alloc_max", &past, &len, NULL, 0);
ASSERT_SHOULD_FAIL("alloc_max", NULL, NULL, &future, len);
ASSERT_SHOULD_FAIL("alloc_max", &past, &len, &future, len);
#undef ASSERT_SHOULD_FAIL
}
TEST_END
TEST_BEGIN(test_prof_recent_on) {
test_skip_if(!config_prof);
ssize_t past, future;
size_t len = sizeof(ssize_t);
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, NULL, 0), 0, "no-op mallctl should be allowed");
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
&past, &len, NULL, 0), 0, "Read error");
expect_zd_eq(past, OPT_ALLOC_MAX, "Wrong read result");
2019-12-19 05:38:14 +08:00
future = OPT_ALLOC_MAX + 1;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, len), 0, "Write error");
future = -1;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
&past, &len, &future, len), 0, "Read/write error");
expect_zd_eq(past, OPT_ALLOC_MAX + 1, "Wrong read result");
2019-12-19 05:38:14 +08:00
future = -2;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
&past, &len, &future, len), EINVAL,
"Invalid write should return EINVAL");
expect_zd_eq(past, OPT_ALLOC_MAX + 1,
2019-12-19 05:38:14 +08:00
"Output should not be touched given invalid write");
future = OPT_ALLOC_MAX;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
&past, &len, &future, len), 0, "Read/write error");
expect_zd_eq(past, -1, "Wrong read result");
2019-12-19 05:38:14 +08:00
future = OPT_ALLOC_MAX + 2;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
&past, &len, &future, len * 2), EINVAL,
"Invalid write should return EINVAL");
expect_zd_eq(past, -1,
2019-12-19 05:38:14 +08:00
"Output should not be touched given invalid write");
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
}
TEST_END
/* Reproducible sequence of request sizes */
#define NTH_REQ_SIZE(n) ((n) * 97 + 101)
static void
2020-04-07 07:19:53 +08:00
confirm_malloc(void *p) {
assert_ptr_not_null(p, "malloc failed unexpectedly");
edata_t *e = emap_edata_lookup(TSDN_NULL, &arena_emap_global, p);
assert_ptr_not_null(e, "NULL edata for living pointer");
2020-06-18 00:57:54 +08:00
prof_recent_t *n = edata_prof_recent_alloc_get_no_lock_test(e);
assert_ptr_not_null(n, "Record in edata should not be NULL");
expect_ptr_not_null(n->alloc_tctx,
2019-12-19 05:38:14 +08:00
"alloc_tctx in record should not be NULL");
2020-06-18 00:57:54 +08:00
expect_ptr_eq(e, prof_recent_alloc_edata_get_no_lock_test(n),
2019-12-19 05:38:14 +08:00
"edata pointer in record is not correct");
expect_ptr_null(n->dalloc_tctx, "dalloc_tctx in record should be NULL");
2019-12-19 05:38:14 +08:00
}
static void
2020-04-07 07:19:53 +08:00
confirm_record_size(prof_recent_t *n, unsigned kth) {
expect_zu_eq(n->size, NTH_REQ_SIZE(kth),
"Recorded allocation size is wrong");
2019-12-19 05:38:14 +08:00
}
static void
2020-04-07 07:19:53 +08:00
confirm_record_living(prof_recent_t *n) {
expect_ptr_not_null(n->alloc_tctx,
2019-12-19 05:38:14 +08:00
"alloc_tctx in record should not be NULL");
2020-06-18 00:57:54 +08:00
edata_t *edata = prof_recent_alloc_edata_get_no_lock_test(n);
assert_ptr_not_null(edata,
2019-12-19 05:38:14 +08:00
"Recorded edata should not be NULL for living pointer");
2020-06-18 00:57:54 +08:00
expect_ptr_eq(n, edata_prof_recent_alloc_get_no_lock_test(edata),
2019-12-19 05:38:14 +08:00
"Record in edata is not correct");
expect_ptr_null(n->dalloc_tctx, "dalloc_tctx in record should be NULL");
2019-12-19 05:38:14 +08:00
}
static void
2020-04-07 07:19:53 +08:00
confirm_record_released(prof_recent_t *n) {
expect_ptr_not_null(n->alloc_tctx,
2019-12-19 05:38:14 +08:00
"alloc_tctx in record should not be NULL");
2020-06-18 00:57:54 +08:00
expect_ptr_null(prof_recent_alloc_edata_get_no_lock_test(n),
2019-12-19 05:38:14 +08:00
"Recorded edata should be NULL for released pointer");
expect_ptr_not_null(n->dalloc_tctx,
2019-12-19 05:38:14 +08:00
"dalloc_tctx in record should not be NULL for released pointer");
}
TEST_BEGIN(test_prof_recent_alloc) {
test_skip_if(!config_prof);
bool b;
unsigned i, c;
size_t req_size;
void *p;
prof_recent_t *n;
ssize_t future;
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
/*
* First batch of 2 * OPT_ALLOC_MAX allocations. After the
* (OPT_ALLOC_MAX - 1)'th allocation the recorded allocations should
* always be the last OPT_ALLOC_MAX allocations coming from here.
*/
for (i = 0; i < 2 * OPT_ALLOC_MAX; ++i) {
req_size = NTH_REQ_SIZE(i);
p = malloc(req_size);
2020-04-07 07:19:53 +08:00
confirm_malloc(p);
2019-12-19 05:38:14 +08:00
if (i < OPT_ALLOC_MAX - 1) {
assert_false(ql_empty(&prof_recent_alloc_list),
2019-12-19 05:38:14 +08:00
"Empty recent allocation");
free(p);
/*
* The recorded allocations may still include some
* other allocations before the test run started,
* so keep allocating without checking anything.
*/
continue;
}
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2019-12-19 05:38:14 +08:00
++c;
2020-04-07 07:19:53 +08:00
confirm_record_size(n, i + c - OPT_ALLOC_MAX);
2019-12-19 05:38:14 +08:00
if (c == OPT_ALLOC_MAX) {
2020-04-07 07:19:53 +08:00
confirm_record_living(n);
2019-12-19 05:38:14 +08:00
} else {
2020-04-07 07:19:53 +08:00
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
}
}
assert_u_eq(c, OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
free(p);
}
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
b = false;
assert_d_eq(mallctl("prof.active", NULL, NULL, &b, sizeof(bool)), 0,
2019-12-19 05:38:14 +08:00
"mallctl for turning off prof_active failed");
/*
* Second batch of OPT_ALLOC_MAX allocations. Since prof_active is
* turned off, this batch shouldn't be recorded.
*/
for (; i < 3 * OPT_ALLOC_MAX; ++i) {
req_size = NTH_REQ_SIZE(i);
p = malloc(req_size);
assert_ptr_not_null(p, "malloc failed unexpectedly");
2019-12-19 05:38:14 +08:00
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2020-04-07 07:19:53 +08:00
confirm_record_size(n, c + OPT_ALLOC_MAX);
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
++c;
}
assert_u_eq(c, OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
free(p);
}
b = true;
assert_d_eq(mallctl("prof.active", NULL, NULL, &b, sizeof(bool)), 0,
2019-12-19 05:38:14 +08:00
"mallctl for turning on prof_active failed");
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
/*
* Third batch of OPT_ALLOC_MAX allocations. Since prof_active is
* turned back on, they should be recorded, and in the list of recorded
* allocations they should follow the first batch rather than the
* second batch.
*/
for (; i < 4 * OPT_ALLOC_MAX; ++i) {
req_size = NTH_REQ_SIZE(i);
p = malloc(req_size);
2020-04-07 07:19:53 +08:00
confirm_malloc(p);
2019-12-19 05:38:14 +08:00
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2019-12-19 05:38:14 +08:00
++c;
2020-04-07 07:19:53 +08:00
confirm_record_size(n,
2019-12-19 05:38:14 +08:00
/* Is the allocation from the third batch? */
i + c - OPT_ALLOC_MAX >= 3 * OPT_ALLOC_MAX ?
/* If yes, then it's just recorded. */
i + c - OPT_ALLOC_MAX :
/*
* Otherwise, it should come from the first batch
* instead of the second batch.
*/
i + c - 2 * OPT_ALLOC_MAX);
if (c == OPT_ALLOC_MAX) {
2020-04-07 07:19:53 +08:00
confirm_record_living(n);
2019-12-19 05:38:14 +08:00
} else {
2020-04-07 07:19:53 +08:00
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
}
}
assert_u_eq(c, OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
free(p);
}
/* Increasing the limit shouldn't alter the list of records. */
future = OPT_ALLOC_MAX + 1;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2020-04-07 07:19:53 +08:00
confirm_record_size(n, c + 3 * OPT_ALLOC_MAX);
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
++c;
}
assert_u_eq(c, OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
/*
* Decreasing the limit shouldn't alter the list of records as long as
* the new limit is still no less than the length of the list.
*/
future = OPT_ALLOC_MAX;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2020-04-07 07:19:53 +08:00
confirm_record_size(n, c + 3 * OPT_ALLOC_MAX);
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
++c;
}
assert_u_eq(c, OPT_ALLOC_MAX,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
/*
* Decreasing the limit should shorten the list of records if the new
* limit is less than the length of the list.
*/
future = OPT_ALLOC_MAX - 1;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2019-12-19 05:38:14 +08:00
++c;
2020-04-07 07:19:53 +08:00
confirm_record_size(n, c + 3 * OPT_ALLOC_MAX);
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
}
assert_u_eq(c, OPT_ALLOC_MAX - 1,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
/* Setting to unlimited shouldn't alter the list of records. */
future = -1;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
c = 0;
ql_foreach(n, &prof_recent_alloc_list, link) {
2019-12-19 05:38:14 +08:00
++c;
2020-04-07 07:19:53 +08:00
confirm_record_size(n, c + 3 * OPT_ALLOC_MAX);
confirm_record_released(n);
2019-12-19 05:38:14 +08:00
}
assert_u_eq(c, OPT_ALLOC_MAX - 1,
2019-12-19 05:38:14 +08:00
"Incorrect total number of allocations");
/* Downshift to only one record. */
future = 1;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
assert_false(ql_empty(&prof_recent_alloc_list), "Recent list is empty");
n = ql_first(&prof_recent_alloc_list);
2020-04-07 07:19:53 +08:00
confirm_record_size(n, 4 * OPT_ALLOC_MAX - 1);
confirm_record_released(n);
n = ql_next(&prof_recent_alloc_list, n, link);
assert_ptr_null(n, "Recent list should only contain one record");
2019-12-19 05:38:14 +08:00
/* Completely turn off. */
future = 0;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
assert_true(ql_empty(&prof_recent_alloc_list),
"Recent list should be empty");
2019-12-19 05:38:14 +08:00
/* Restore the settings. */
future = OPT_ALLOC_MAX;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
2019-12-19 05:38:14 +08:00
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
assert_true(ql_empty(&prof_recent_alloc_list),
"Recent list should be empty");
2019-12-19 05:38:14 +08:00
confirm_prof_setup();
2019-12-19 05:38:14 +08:00
}
TEST_END
#undef NTH_REQ_SIZE
#define DUMP_OUT_SIZE 4096
static char dump_out[DUMP_OUT_SIZE];
static size_t dump_out_len = 0;
static void
test_dump_write_cb(void *not_used, const char *str) {
size_t len = strlen(str);
assert(dump_out_len + len < DUMP_OUT_SIZE);
memcpy(dump_out + dump_out_len, str, len + 1);
dump_out_len += len;
}
static void
call_dump() {
static void *in[2] = {test_dump_write_cb, NULL};
dump_out_len = 0;
assert_d_eq(mallctl("experimental.prof_recent.alloc_dump",
NULL, NULL, in, sizeof(in)), 0, "Dump mallctl raised error");
}
typedef struct {
size_t size;
size_t usize;
bool released;
} confirm_record_t;
#define DUMP_ERROR "Dump output is wrong"
static void
confirm_record(const char *template, const confirm_record_t *records,
const size_t n_records) {
static const char *types[2] = {"alloc", "dalloc"};
static char buf[64];
/*
* The template string would be in the form of:
* "{...,\"recent_alloc\":[]}",
* and dump_out would be in the form of:
* "{...,\"recent_alloc\":[...]}".
* Using "- 2" serves to cut right before the ending "]}".
*/
assert_d_eq(memcmp(dump_out, template, strlen(template) - 2), 0,
DUMP_ERROR);
assert_d_eq(memcmp(dump_out + strlen(dump_out) - 2,
template + strlen(template) - 2, 2), 0, DUMP_ERROR);
const char *start = dump_out + strlen(template) - 2;
const char *end = dump_out + strlen(dump_out) - 2;
const confirm_record_t *record;
for (record = records; record < records + n_records; ++record) {
#define ASSERT_CHAR(c) do { \
assert_true(start < end, DUMP_ERROR); \
assert_c_eq(*start++, c, DUMP_ERROR); \
} while (0)
#define ASSERT_STR(s) do { \
const size_t len = strlen(s); \
assert_true(start + len <= end, DUMP_ERROR); \
assert_d_eq(memcmp(start, s, len), 0, DUMP_ERROR); \
start += len; \
} while (0)
#define ASSERT_FORMATTED_STR(s, ...) do { \
malloc_snprintf(buf, sizeof(buf), s, __VA_ARGS__); \
ASSERT_STR(buf); \
} while (0)
if (record != records) {
ASSERT_CHAR(',');
}
ASSERT_CHAR('{');
ASSERT_STR("\"size\"");
ASSERT_CHAR(':');
ASSERT_FORMATTED_STR("%zu", record->size);
ASSERT_CHAR(',');
ASSERT_STR("\"usize\"");
ASSERT_CHAR(':');
ASSERT_FORMATTED_STR("%zu", record->usize);
ASSERT_CHAR(',');
ASSERT_STR("\"released\"");
ASSERT_CHAR(':');
ASSERT_STR(record->released ? "true" : "false");
ASSERT_CHAR(',');
const char **type = types;
while (true) {
ASSERT_FORMATTED_STR("\"%s_thread_uid\"", *type);
ASSERT_CHAR(':');
while (isdigit(*start)) {
++start;
}
ASSERT_CHAR(',');
if (opt_prof_sys_thread_name) {
ASSERT_FORMATTED_STR("\"%s_thread_name\"",
*type);
ASSERT_CHAR(':');
ASSERT_CHAR('"');
while (*start != '"') {
++start;
}
ASSERT_CHAR('"');
ASSERT_CHAR(',');
}
ASSERT_FORMATTED_STR("\"%s_time\"", *type);
ASSERT_CHAR(':');
while (isdigit(*start)) {
++start;
}
ASSERT_CHAR(',');
ASSERT_FORMATTED_STR("\"%s_trace\"", *type);
ASSERT_CHAR(':');
ASSERT_CHAR('[');
while (isdigit(*start) || *start == 'x' ||
(*start >= 'a' && *start <= 'f') ||
*start == '\"' || *start == ',') {
++start;
}
ASSERT_CHAR(']');
if (strcmp(*type, "dalloc") == 0) {
break;
}
assert(strcmp(*type, "alloc") == 0);
if (!record->released) {
break;
}
ASSERT_CHAR(',');
++type;
}
ASSERT_CHAR('}');
#undef ASSERT_FORMATTED_STR
#undef ASSERT_STR
#undef ASSERT_CHAR
}
assert_ptr_eq(record, records + n_records, DUMP_ERROR);
assert_ptr_eq(start, end, DUMP_ERROR);
}
TEST_BEGIN(test_prof_recent_alloc_dump) {
test_skip_if(!config_prof);
confirm_prof_setup();
ssize_t future;
void *p, *q;
confirm_record_t records[2];
assert_zu_eq(lg_prof_sample, (size_t)0,
"lg_prof_sample not set correctly");
future = 0;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
call_dump();
expect_str_eq(dump_out, "{\"sample_interval\":1,"
"\"recent_alloc_max\":0,\"recent_alloc\":[]}", DUMP_ERROR);
future = 2;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
call_dump();
const char *template = "{\"sample_interval\":1,"
"\"recent_alloc_max\":2,\"recent_alloc\":[]}";
expect_str_eq(dump_out, template, DUMP_ERROR);
p = malloc(7);
call_dump();
records[0].size = 7;
records[0].usize = sz_s2u(7);
records[0].released = false;
confirm_record(template, records, 1);
q = mallocx(17, MALLOCX_ALIGN(128));
call_dump();
records[1].size = 17;
records[1].usize = sz_sa2u(17, 128);
records[1].released = false;
confirm_record(template, records, 2);
free(q);
call_dump();
records[1].released = true;
confirm_record(template, records, 2);
free(p);
call_dump();
records[0].released = true;
confirm_record(template, records, 2);
future = OPT_ALLOC_MAX;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
NULL, NULL, &future, sizeof(ssize_t)), 0, "Write error");
confirm_prof_setup();
}
TEST_END
#undef DUMP_ERROR
#undef DUMP_OUT_SIZE
#define N_THREADS 8
#define N_PTRS 512
#define N_CTLS 8
#define N_ITERS 2048
#define STRESS_ALLOC_MAX 4096
typedef struct {
thd_t thd;
size_t id;
void *ptrs[N_PTRS];
size_t count;
} thd_data_t;
static thd_data_t thd_data[N_THREADS];
static ssize_t test_max;
static void
test_write_cb(void *cbopaque, const char *str) {
sleep_ns(1000 * 1000);
}
static void *
f_thread(void *arg) {
const size_t thd_id = *(size_t *)arg;
thd_data_t *data_p = thd_data + thd_id;
assert(data_p->id == thd_id);
data_p->count = 0;
uint64_t rand = (uint64_t)thd_id;
tsd_t *tsd = tsd_fetch();
assert(test_max > 1);
ssize_t last_max = -1;
for (int i = 0; i < N_ITERS; i++) {
rand = prng_range_u64(&rand, N_PTRS + N_CTLS * 5);
assert(data_p->count <= N_PTRS);
if (rand < data_p->count) {
assert(data_p->count > 0);
if (rand != data_p->count - 1) {
assert(data_p->count > 1);
void *temp = data_p->ptrs[rand];
data_p->ptrs[rand] =
data_p->ptrs[data_p->count - 1];
data_p->ptrs[data_p->count - 1] = temp;
}
free(data_p->ptrs[--data_p->count]);
} else if (rand < N_PTRS) {
assert(data_p->count < N_PTRS);
data_p->ptrs[data_p->count++] = malloc(1);
} else if (rand % 5 == 0) {
prof_recent_alloc_dump(tsd, test_write_cb, NULL);
} else if (rand % 5 == 1) {
last_max = prof_recent_alloc_max_ctl_read();
} else if (rand % 5 == 2) {
last_max =
prof_recent_alloc_max_ctl_write(tsd, test_max * 2);
} else if (rand % 5 == 3) {
last_max =
prof_recent_alloc_max_ctl_write(tsd, test_max);
} else {
assert(rand % 5 == 4);
last_max =
prof_recent_alloc_max_ctl_write(tsd, test_max / 2);
}
assert_zd_ge(last_max, -1, "Illegal last-N max");
}
while (data_p->count > 0) {
free(data_p->ptrs[--data_p->count]);
}
return NULL;
}
TEST_BEGIN(test_prof_recent_stress) {
test_skip_if(!config_prof);
confirm_prof_setup();
test_max = OPT_ALLOC_MAX;
for (size_t i = 0; i < N_THREADS; i++) {
thd_data_t *data_p = thd_data + i;
data_p->id = i;
thd_create(&data_p->thd, &f_thread, &data_p->id);
}
for (size_t i = 0; i < N_THREADS; i++) {
thd_data_t *data_p = thd_data + i;
thd_join(data_p->thd, NULL);
}
test_max = STRESS_ALLOC_MAX;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
NULL, NULL, &test_max, sizeof(ssize_t)), 0, "Write error");
for (size_t i = 0; i < N_THREADS; i++) {
thd_data_t *data_p = thd_data + i;
data_p->id = i;
thd_create(&data_p->thd, &f_thread, &data_p->id);
}
for (size_t i = 0; i < N_THREADS; i++) {
thd_data_t *data_p = thd_data + i;
thd_join(data_p->thd, NULL);
}
test_max = OPT_ALLOC_MAX;
assert_d_eq(mallctl("experimental.prof_recent.alloc_max",
NULL, NULL, &test_max, sizeof(ssize_t)), 0, "Write error");
confirm_prof_setup();
}
TEST_END
#undef STRESS_ALLOC_MAX
#undef N_ITERS
#undef N_PTRS
#undef N_THREADS
2019-12-19 05:38:14 +08:00
int
main(void) {
return test(
test_confirm_setup,
test_prof_recent_off,
test_prof_recent_on,
test_prof_recent_alloc,
test_prof_recent_alloc_dump,
test_prof_recent_stress);
2019-12-19 05:38:14 +08:00
}