/external/jemalloc/test/unit/ |
D | stats.c | 10 assert_d_eq(mallctl("stats.cactive", (void *)&cactive, &sz, NULL, 0), in TEST_BEGIN() 14 assert_d_eq(mallctl("stats.allocated", (void *)&allocated, &sz, NULL, in TEST_BEGIN() 16 assert_d_eq(mallctl("stats.active", (void *)&active, &sz, NULL, 0), in TEST_BEGIN() 18 assert_d_eq(mallctl("stats.resident", (void *)&resident, &sz, NULL, 0), in TEST_BEGIN() 20 assert_d_eq(mallctl("stats.mapped", (void *)&mapped, &sz, NULL, 0), in TEST_BEGIN() 48 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), in TEST_BEGIN() 52 assert_d_eq(mallctl("stats.arenas.0.huge.allocated", (void *)&allocated, in TEST_BEGIN() 55 assert_d_eq(mallctl("stats.arenas.0.huge.nmalloc", (void *)&nmalloc, in TEST_BEGIN() 57 assert_d_eq(mallctl("stats.arenas.0.huge.ndalloc", (void *)&ndalloc, in TEST_BEGIN() 59 assert_d_eq(mallctl("stats.arenas.0.huge.nrequests", (void *)&nrequests, in TEST_BEGIN() [all …]
|
D | mallctl.c | 8 assert_d_eq(mallctl("no_such_name", NULL, NULL, NULL, 0), ENOENT, in TEST_BEGIN() 11 assert_d_eq(mallctl("version", NULL, NULL, "0.0.0", strlen("0.0.0")), in TEST_BEGIN() 15 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 18 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 23 assert_d_eq(mallctl("epoch", (void *)&epoch, &sz, NULL, 0), EINVAL, in TEST_BEGIN() 26 assert_d_eq(mallctl("epoch", (void *)&epoch, &sz, NULL, 0), EINVAL, in TEST_BEGIN() 85 assert_d_eq(mallctl("epoch", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 90 assert_d_eq(mallctl("epoch", (void *)&old_epoch, &sz, NULL, 0), 0, in TEST_BEGIN() 95 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&new_epoch, in TEST_BEGIN() 100 assert_d_eq(mallctl("epoch", (void *)&old_epoch, &sz, in TEST_BEGIN() [all …]
|
D | decay.c | 43 assert_d_eq(mallctl("arenas.hchunk.0.size", (void *)&huge0, &sz, NULL, in TEST_BEGIN() 45 assert_d_eq(mallctl("arenas.lrun.0.size", (void *)&large0, &sz, NULL, in TEST_BEGIN() 178 assert_d_eq(mallctl("tcache.create", (void *)&tcache_ind, &sz, in TEST_BEGIN() 195 assert_d_eq(mallctl("tcache.flush", NULL, NULL, in TEST_BEGIN() 231 assert_d_eq(mallctl("arenas.tcache_max", (void *)&tcache_max, in TEST_BEGIN() 236 assert_d_eq(mallctl("arenas.lrun.0.size", (void *)&large, &sz, in TEST_BEGIN() 240 assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 242 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 245 assert_d_eq(mallctl("stats.arenas.0.npurge", (void *)&npurge0, &sz, in TEST_BEGIN() 266 assert_d_eq(mallctl("arena.0.decay", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() [all …]
|
D | prof_reset.c | 23 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in set_prof_active() 33 assert_d_eq(mallctl("prof.lg_sample", (void *)&lg_prof_sample, &sz, in get_lg_prof_sample() 42 assert_d_eq(mallctl("prof.reset", NULL, NULL, in do_prof_reset() 58 assert_d_eq(mallctl("opt.lg_prof_sample", (void *)&lg_prof_sample_orig, in TEST_BEGIN() 70 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 126 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in TEST_BEGIN() 131 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 133 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in TEST_BEGIN() 163 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), in thd_start() 169 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in thd_start()
|
D | prof_thread_name.c | 15 assert_d_eq(mallctl("thread.prof.name", (void *)&thread_name_old, &sz, in mallctl_thread_name_get_impl() 30 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in mallctl_thread_name_set_impl() 50 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in TEST_BEGIN() 57 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in TEST_BEGIN() 68 assert_d_eq(mallctl("thread.prof.name", in TEST_BEGIN() 94 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in thd_start()
|
D | run_quantize.c | 16 assert_d_eq(mallctl("arenas.nbins", (void *)&nbins, &sz, NULL, 0), 0, in TEST_BEGIN() 50 assert_d_eq(mallctl("config.cache_oblivious", (void *)&cache_oblivious, in TEST_BEGIN() 54 assert_d_eq(mallctl("arenas.nlruns", (void *)&nlruns, &sz, NULL, 0), 0, in TEST_BEGIN() 105 assert_d_eq(mallctl("arenas.nbins", (void *)&nbins, &sz, NULL, 0), 0, in TEST_BEGIN() 109 assert_d_eq(mallctl("arenas.nlruns", (void *)&nlruns, &sz, NULL, 0), 0, in TEST_BEGIN()
|
D | prof_gdump.c | 31 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in TEST_BEGIN() 49 assert_d_eq(mallctl("prof.gdump", (void *)&gdump_old, &sz, in TEST_BEGIN() 60 assert_d_eq(mallctl("prof.gdump", (void *)&gdump_old, &sz, in TEST_BEGIN()
|
/external/jemalloc_new/test/unit/ |
D | stats.c | 8 assert_d_eq(mallctl("stats.allocated", (void *)&allocated, &sz, NULL, in TEST_BEGIN() 10 assert_d_eq(mallctl("stats.active", (void *)&active, &sz, NULL, 0), in TEST_BEGIN() 12 assert_d_eq(mallctl("stats.resident", (void *)&resident, &sz, NULL, 0), in TEST_BEGIN() 14 assert_d_eq(mallctl("stats.mapped", (void *)&mapped, &sz, NULL, 0), in TEST_BEGIN() 39 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), in TEST_BEGIN() 43 assert_d_eq(mallctl("stats.arenas.0.large.allocated", in TEST_BEGIN() 47 assert_d_eq(mallctl("stats.arenas.0.large.nmalloc", (void *)&nmalloc, in TEST_BEGIN() 49 assert_d_eq(mallctl("stats.arenas.0.large.ndalloc", (void *)&ndalloc, in TEST_BEGIN() 51 assert_d_eq(mallctl("stats.arenas.0.large.nrequests", in TEST_BEGIN() 85 assert_d_eq(mallctl("thread.tcache.flush", NULL, NULL, NULL, 0), in TEST_BEGIN() [all …]
|
D | mallctl.c | 9 assert_d_eq(mallctl("no_such_name", NULL, NULL, NULL, 0), ENOENT, in TEST_BEGIN() 12 assert_d_eq(mallctl("version", NULL, NULL, "0.0.0", strlen("0.0.0")), in TEST_BEGIN() 16 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 19 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 24 assert_d_eq(mallctl("epoch", (void *)&epoch, &sz, NULL, 0), EINVAL, in TEST_BEGIN() 27 assert_d_eq(mallctl("epoch", (void *)&epoch, &sz, NULL, 0), EINVAL, in TEST_BEGIN() 83 assert_d_eq(mallctl("epoch", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 88 assert_d_eq(mallctl("epoch", (void *)&old_epoch, &sz, NULL, 0), 0, in TEST_BEGIN() 93 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&new_epoch, in TEST_BEGIN() 98 assert_d_eq(mallctl("epoch", (void *)&old_epoch, &sz, in TEST_BEGIN() [all …]
|
D | background_thread_enable.c | 19 assert_d_eq(mallctl("arenas.create", &id, &sz_u, NULL, 0), 0, in TEST_BEGIN() 25 assert_d_eq(mallctl("background_thread", NULL, NULL, &enable, sz_b), 0, in TEST_BEGIN() 28 assert_d_eq(mallctl("background_thread", NULL, NULL, &enable, sz_b), 0, in TEST_BEGIN() 39 assert_d_eq(mallctl("opt.max_background_threads", in TEST_BEGIN() 42 assert_d_eq(mallctl("max_background_threads", &maxt, &sz_m, NULL, 0), 0, in TEST_BEGIN() 48 assert_d_eq(mallctl("max_background_threads", NULL, NULL, &maxt, sz_m), in TEST_BEGIN() 55 assert_d_eq(mallctl("arenas.create", &id, &sz_u, NULL, 0), 0, in TEST_BEGIN() 61 assert_d_eq(mallctl("background_thread", NULL, NULL, &enable, sz_b), 0, in TEST_BEGIN() 66 assert_d_eq(mallctl("max_background_threads", NULL, NULL, &maxt, sz_m), in TEST_BEGIN() 71 assert_d_eq(mallctl("max_background_threads", NULL, NULL, &maxt, sz_m), in TEST_BEGIN()
|
D | prof_reset.c | 15 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in set_prof_active() 24 assert_d_eq(mallctl("prof.lg_sample", (void *)&lg_prof_sample, &sz, in get_lg_prof_sample() 32 assert_d_eq(mallctl("prof.reset", NULL, NULL, in do_prof_reset() 47 assert_d_eq(mallctl("opt.lg_prof_sample", (void *)&lg_prof_sample_orig, in TEST_BEGIN() 59 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 112 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in TEST_BEGIN() 117 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 119 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in TEST_BEGIN() 148 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), in thd_start() 154 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in thd_start()
|
D | prof_thread_name.c | 10 assert_d_eq(mallctl("thread.prof.name", (void *)&thread_name_old, &sz, in mallctl_thread_name_get_impl() 23 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in mallctl_thread_name_set_impl() 42 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in TEST_BEGIN() 49 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in TEST_BEGIN() 60 assert_d_eq(mallctl("thread.prof.name", in TEST_BEGIN() 85 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in thd_start()
|
D | background_thread.c | 11 assert_d_eq(mallctl("background_thread", (void *)&e0, &sz, in test_switch_background_thread_ctl() 30 assert_d_eq(mallctl("background_thread", (void *)&e0, &sz, in test_repeat_background_thread_ctl() 49 assert_d_eq(mallctl("opt.background_thread", (void *)&e0, &sz, in TEST_BEGIN() 51 assert_d_eq(mallctl("background_thread", (void *)&e1, &sz, in TEST_BEGIN()
|
D | decay.c | 16 int ret = mallctl("background_thread", (void *)&enabled, &sz, NULL,0); in check_background_thread_enabled() 42 assert_d_eq(mallctl("arenas.create", (void *)&arena_ind, &sz, NULL, 0), in do_arena_create() 78 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), in do_epoch() 190 assert_d_eq(mallctl("arenas.lextent.0.size", (void *)&large0, &sz, NULL, in TEST_BEGIN() 199 assert_d_eq(mallctl("thread.arena", (void *)&old_arena_ind, in TEST_BEGIN() 335 assert_d_eq(mallctl("arenas.tcache_max", (void *)&tcache_max, in TEST_BEGIN() 339 assert_d_eq(mallctl("tcache.create", (void *)&tcache_ind, &sz, in TEST_BEGIN() 356 assert_d_eq(mallctl("tcache.flush", NULL, NULL, in TEST_BEGIN() 440 assert_d_eq(mallctl("arenas.tcache_max", (void *)&tcache_max, &sz, NULL, in TEST_BEGIN() 494 assert_d_eq(mallctl("arenas.lextent.0.size", (void *)&large0, &sz, NULL, in TEST_BEGIN() [all …]
|
D | prof_gdump.c | 25 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in TEST_BEGIN() 43 assert_d_eq(mallctl("prof.gdump", (void *)&gdump_old, &sz, in TEST_BEGIN() 54 assert_d_eq(mallctl("prof.gdump", (void *)&gdump_old, &sz, in TEST_BEGIN()
|
/external/jemalloc_new/test/integration/ |
D | thread_tcache_enabled.c | 7 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, NULL, in thd_start() 12 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 18 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 23 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 28 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 33 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 39 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 45 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 51 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 57 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start()
|
D | allocated.c | 20 if ((err = mallctl("thread.allocated", (void *)&a0, &sz, NULL, 0))) { in thd_start() 28 if ((err = mallctl("thread.allocatedp", (void *)&ap0, &sz, NULL, 0))) { in thd_start() 40 if ((err = mallctl("thread.deallocated", (void *)&d0, &sz, NULL, 0))) { in thd_start() 48 if ((err = mallctl("thread.deallocatedp", (void *)&dp0, &sz, NULL, in thd_start() 64 mallctl("thread.allocated", (void *)&a1, &sz, NULL, 0); in thd_start() 66 mallctl("thread.allocatedp", (void *)&ap1, &sz, NULL, 0); in thd_start() 81 mallctl("thread.deallocated", (void *)&d1, &sz, NULL, 0); in thd_start() 83 mallctl("thread.deallocatedp", (void *)&dp1, &sz, NULL, 0); in thd_start()
|
D | extent.c | 9 int ret = mallctl("background_thread", (void *)&enabled, &sz, NULL,0); in check_background_thread_enabled() 30 assert_d_eq(mallctl("arenas.lextent.0.size", (void *)&large0, &sz, NULL, in test_extent_body() 32 assert_d_eq(mallctl("arenas.lextent.1.size", (void *)&large1, &sz, NULL, in test_extent_body() 34 assert_d_eq(mallctl("arenas.lextent.2.size", (void *)&large2, &sz, NULL, in test_extent_body() 113 assert_d_eq(mallctl("opt.narenas", (void *)&narenas, &sz, NULL, 0), in test_manual_hook_auto_arena() 149 assert_d_eq(mallctl("arenas.create", (void *)&arena_ind, &sz, NULL, 0), in test_manual_hook_body() 235 assert_d_eq(mallctl("arenas.create", (void *)&arena_ind, &sz, in TEST_BEGIN()
|
D | thread_arena.c | 18 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, in thd_start() 27 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, NULL, in thd_start() 59 assert_d_eq(mallctl("arenas.create", (void *)&arena_ind, &sz, NULL, 0), in TEST_BEGIN() 63 if ((err = mallctl("thread.arena", (void *)&old_arena_ind, &size, in TEST_BEGIN()
|
/external/jemalloc/test/integration/ |
D | thread_tcache_enabled.c | 19 if ((err = mallctl("thread.tcache.enabled", (void *)&e0, &sz, NULL, in thd_start() 31 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 37 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 42 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 47 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 52 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 58 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 64 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 70 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 76 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start()
|
D | allocated.c | 21 if ((err = mallctl("thread.allocated", (void *)&a0, &sz, NULL, 0))) { in thd_start() 28 if ((err = mallctl("thread.allocatedp", (void *)&ap0, &sz, NULL, 0))) { in thd_start() 39 if ((err = mallctl("thread.deallocated", (void *)&d0, &sz, NULL, 0))) { in thd_start() 46 if ((err = mallctl("thread.deallocatedp", (void *)&dp0, &sz, NULL, in thd_start() 61 mallctl("thread.allocated", (void *)&a1, &sz, NULL, 0); in thd_start() 63 mallctl("thread.allocatedp", (void *)&ap1, &sz, NULL, 0); in thd_start() 78 mallctl("thread.deallocated", (void *)&d1, &sz, NULL, 0); in thd_start() 80 mallctl("thread.deallocatedp", (void *)&dp1, &sz, NULL, 0); in thd_start()
|
D | thread_arena.c | 19 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, in thd_start() 28 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, NULL, in thd_start() 54 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, NULL, in TEST_BEGIN()
|
D | chunk.c | 140 assert_d_eq(mallctl("arenas.extend", (void *)&arena_ind, &sz, NULL, 0), in TEST_BEGIN() 168 assert_d_eq(mallctl("arenas.lrun.0.size", (void *)&large0, &sz, NULL, in TEST_BEGIN() 170 assert_d_eq(mallctl("arenas.lrun.1.size", (void *)&large1, &sz, NULL, in TEST_BEGIN() 174 assert_d_eq(mallctl("arenas.hchunk.0.size", (void *)&huge0, &sz, NULL, in TEST_BEGIN() 176 assert_d_eq(mallctl("arenas.hchunk.1.size", (void *)&huge1, &sz, NULL, in TEST_BEGIN() 178 assert_d_eq(mallctl("arenas.hchunk.2.size", (void *)&huge2, &sz, NULL, in TEST_BEGIN()
|
D | iterate.c | 39 mallctl("thread.tcache.flush", NULL, NULL, NULL, 0); in TEST_BEGIN() 69 mallctl("thread.tcache.flush", NULL, NULL, NULL, 0); in TEST_BEGIN() 101 mallctl("thread.tcache.flush", NULL, NULL, NULL, 0); in TEST_BEGIN()
|
/external/jemalloc/ |
D | ChangeLog | 101 - Add the arena.<i>.reset mallctl, which makes it possible to discard all of 154 mallctl API impact. This is an alternative to the existing ratio-based 282 visible to custom functions set via the "arena.<i>.chunk_hooks" mallctl. 285 with interposed resets (triggered via the "prof.reset" mallctl). This bug 343 via the "arena.<i>.chunk_hooks" mallctl. 361 mallctl), so that it is possible to separately enable junk filling for 368 - Add the "stats.resident" mallctl, which reports the upper limit of 373 - Add the "prof.gdump" mallctl, which makes it possible to toggle the gdump 404 the "config.cache_oblivious" mallctl. 426 - Change the "arenas.nlruns" mallctl type from size_t to unsigned. [all …]
|