/external/jemalloc/test/unit/ |
D | stats.c | 10 assert_d_eq(mallctl("stats.cactive", (void *)&cactive, &sz, NULL, 0), in TEST_BEGIN() 14 assert_d_eq(mallctl("stats.allocated", (void *)&allocated, &sz, NULL, in TEST_BEGIN() 16 assert_d_eq(mallctl("stats.active", (void *)&active, &sz, NULL, 0), in TEST_BEGIN() 18 assert_d_eq(mallctl("stats.resident", (void *)&resident, &sz, NULL, 0), in TEST_BEGIN() 20 assert_d_eq(mallctl("stats.mapped", (void *)&mapped, &sz, NULL, 0), in TEST_BEGIN() 48 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), in TEST_BEGIN() 52 assert_d_eq(mallctl("stats.arenas.0.huge.allocated", (void *)&allocated, in TEST_BEGIN() 55 assert_d_eq(mallctl("stats.arenas.0.huge.nmalloc", (void *)&nmalloc, in TEST_BEGIN() 57 assert_d_eq(mallctl("stats.arenas.0.huge.ndalloc", (void *)&ndalloc, in TEST_BEGIN() 59 assert_d_eq(mallctl("stats.arenas.0.huge.nrequests", (void *)&nrequests, in TEST_BEGIN() [all …]
|
D | mallctl.c | 8 assert_d_eq(mallctl("no_such_name", NULL, NULL, NULL, 0), ENOENT, in TEST_BEGIN() 11 assert_d_eq(mallctl("version", NULL, NULL, "0.0.0", strlen("0.0.0")), in TEST_BEGIN() 15 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 18 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 23 assert_d_eq(mallctl("epoch", (void *)&epoch, &sz, NULL, 0), EINVAL, in TEST_BEGIN() 26 assert_d_eq(mallctl("epoch", (void *)&epoch, &sz, NULL, 0), EINVAL, in TEST_BEGIN() 85 assert_d_eq(mallctl("epoch", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 90 assert_d_eq(mallctl("epoch", (void *)&old_epoch, &sz, NULL, 0), 0, in TEST_BEGIN() 95 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&new_epoch, in TEST_BEGIN() 100 assert_d_eq(mallctl("epoch", (void *)&old_epoch, &sz, in TEST_BEGIN() [all …]
|
D | decay.c | 43 assert_d_eq(mallctl("arenas.hchunk.0.size", (void *)&huge0, &sz, NULL, in TEST_BEGIN() 45 assert_d_eq(mallctl("arenas.lrun.0.size", (void *)&large0, &sz, NULL, in TEST_BEGIN() 178 assert_d_eq(mallctl("tcache.create", (void *)&tcache_ind, &sz, in TEST_BEGIN() 195 assert_d_eq(mallctl("tcache.flush", NULL, NULL, in TEST_BEGIN() 231 assert_d_eq(mallctl("arenas.tcache_max", (void *)&tcache_max, in TEST_BEGIN() 236 assert_d_eq(mallctl("arenas.lrun.0.size", (void *)&large, &sz, in TEST_BEGIN() 240 assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 242 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, in TEST_BEGIN() 245 assert_d_eq(mallctl("stats.arenas.0.npurge", (void *)&npurge0, &sz, in TEST_BEGIN() 266 assert_d_eq(mallctl("arena.0.decay", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() [all …]
|
D | prof_reset.c | 23 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in set_prof_active() 33 assert_d_eq(mallctl("prof.lg_sample", (void *)&lg_prof_sample, &sz, in get_lg_prof_sample() 42 assert_d_eq(mallctl("prof.reset", NULL, NULL, in do_prof_reset() 58 assert_d_eq(mallctl("opt.lg_prof_sample", (void *)&lg_prof_sample_orig, in TEST_BEGIN() 70 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 126 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in TEST_BEGIN() 131 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in TEST_BEGIN() 133 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in TEST_BEGIN() 163 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), in thd_start() 169 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in thd_start()
|
D | prof_thread_name.c | 15 assert_d_eq(mallctl("thread.prof.name", (void *)&thread_name_old, &sz, in mallctl_thread_name_get_impl() 30 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in mallctl_thread_name_set_impl() 50 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in TEST_BEGIN() 57 assert_d_eq(mallctl("thread.prof.name", NULL, NULL, in TEST_BEGIN() 68 assert_d_eq(mallctl("thread.prof.name", in TEST_BEGIN() 94 assert_d_eq(mallctl("prof.reset", NULL, NULL, NULL, 0), 0, in thd_start()
|
D | run_quantize.c | 16 assert_d_eq(mallctl("arenas.nbins", (void *)&nbins, &sz, NULL, 0), 0, in TEST_BEGIN() 50 assert_d_eq(mallctl("config.cache_oblivious", (void *)&cache_oblivious, in TEST_BEGIN() 54 assert_d_eq(mallctl("arenas.nlruns", (void *)&nlruns, &sz, NULL, 0), 0, in TEST_BEGIN() 105 assert_d_eq(mallctl("arenas.nbins", (void *)&nbins, &sz, NULL, 0), 0, in TEST_BEGIN() 109 assert_d_eq(mallctl("arenas.nlruns", (void *)&nlruns, &sz, NULL, 0), 0, in TEST_BEGIN()
|
D | prof_gdump.c | 31 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in TEST_BEGIN() 49 assert_d_eq(mallctl("prof.gdump", (void *)&gdump_old, &sz, in TEST_BEGIN() 60 assert_d_eq(mallctl("prof.gdump", (void *)&gdump_old, &sz, in TEST_BEGIN()
|
D | prof_accum.c | 44 assert_d_eq(mallctl("prof.dump", NULL, NULL, NULL, 0), in thd_start() 71 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in TEST_BEGIN()
|
D | prof_active.c | 15 assert_d_eq(mallctl(name, (void *)&old, &sz, NULL, 0), 0, in mallctl_bool_get() 29 assert_d_eq(mallctl(name, (void *)&old, &sz, (void *)&val_new, in mallctl_bool_set()
|
D | arena_reset.c | 14 assert_d_eq(mallctl(cmd, (void *)&ret, &z, NULL, 0), 0, in get_nsizes_impl() 95 assert_d_eq(mallctl("arenas.extend", (void *)&arena_ind, &sz, NULL, 0), in TEST_BEGIN()
|
D | pack.c | 34 assert_d_eq(mallctl("arenas.nbins", (void *)&nbins, &sz, NULL, 0), 0, in binind_compute() 114 assert_d_eq(mallctl("arenas.extend", (void *)&arena_ind, &sz, NULL, 0), in arenas_extend_mallctl()
|
D | prof_idump.c | 32 assert_d_eq(mallctl("prof.active", NULL, NULL, (void *)&active, in TEST_BEGIN()
|
/external/jemalloc/test/integration/ |
D | thread_tcache_enabled.c | 19 if ((err = mallctl("thread.tcache.enabled", (void *)&e0, &sz, NULL, in thd_start() 31 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 37 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 42 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 47 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 52 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 58 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 64 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 70 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start() 76 assert_d_eq(mallctl("thread.tcache.enabled", (void *)&e0, &sz, in thd_start()
|
D | allocated.c | 21 if ((err = mallctl("thread.allocated", (void *)&a0, &sz, NULL, 0))) { in thd_start() 28 if ((err = mallctl("thread.allocatedp", (void *)&ap0, &sz, NULL, 0))) { in thd_start() 39 if ((err = mallctl("thread.deallocated", (void *)&d0, &sz, NULL, 0))) { in thd_start() 46 if ((err = mallctl("thread.deallocatedp", (void *)&dp0, &sz, NULL, in thd_start() 61 mallctl("thread.allocated", (void *)&a1, &sz, NULL, 0); in thd_start() 63 mallctl("thread.allocatedp", (void *)&ap1, &sz, NULL, 0); in thd_start() 78 mallctl("thread.deallocated", (void *)&d1, &sz, NULL, 0); in thd_start() 80 mallctl("thread.deallocatedp", (void *)&dp1, &sz, NULL, 0); in thd_start()
|
D | thread_arena.c | 19 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, in thd_start() 28 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, NULL, in thd_start() 54 if ((err = mallctl("thread.arena", (void *)&arena_ind, &size, NULL, in TEST_BEGIN()
|
D | chunk.c | 140 assert_d_eq(mallctl("arenas.extend", (void *)&arena_ind, &sz, NULL, 0), in TEST_BEGIN() 168 assert_d_eq(mallctl("arenas.lrun.0.size", (void *)&large0, &sz, NULL, in TEST_BEGIN() 170 assert_d_eq(mallctl("arenas.lrun.1.size", (void *)&large1, &sz, NULL, in TEST_BEGIN() 174 assert_d_eq(mallctl("arenas.hchunk.0.size", (void *)&huge0, &sz, NULL, in TEST_BEGIN() 176 assert_d_eq(mallctl("arenas.hchunk.1.size", (void *)&huge1, &sz, NULL, in TEST_BEGIN() 178 assert_d_eq(mallctl("arenas.hchunk.2.size", (void *)&huge2, &sz, NULL, in TEST_BEGIN()
|
D | iterate.c | 39 mallctl("thread.tcache.flush", NULL, NULL, NULL, 0); in TEST_BEGIN() 69 mallctl("thread.tcache.flush", NULL, NULL, NULL, 0); in TEST_BEGIN() 101 mallctl("thread.tcache.flush", NULL, NULL, NULL, 0); in TEST_BEGIN()
|
D | mallocx.c | 14 assert_d_eq(mallctl(cmd, (void *)&ret, &z, NULL, 0), 0, in get_nsizes_impl() 62 assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0, in purge()
|
D | overflow.c | 11 assert_d_eq(mallctl("arenas.nhchunks", (void *)&nhchunks, &sz, NULL, 0), in TEST_BEGIN()
|
/external/jemalloc/ |
D | ChangeLog | 101 - Add the arena.<i>.reset mallctl, which makes it possible to discard all of 154 mallctl API impact. This is an alternative to the existing ratio-based 282 visible to custom functions set via the "arena.<i>.chunk_hooks" mallctl. 285 with interposed resets (triggered via the "prof.reset" mallctl). This bug 343 via the "arena.<i>.chunk_hooks" mallctl. 361 mallctl), so that it is possible to separately enable junk filling for 368 - Add the "stats.resident" mallctl, which reports the upper limit of 373 - Add the "prof.gdump" mallctl, which makes it possible to toggle the gdump 404 the "config.cache_oblivious" mallctl. 426 - Change the "arenas.nlruns" mallctl type from size_t to unsigned. [all …]
|
/external/jemalloc/include/jemalloc/internal/ |
D | public_symbols.txt | 16 mallctl:je_mallctl
|
D | public_namespace.h | 16 #define je_mallctl JEMALLOC_N(mallctl)
|
/external/jemalloc/doc/ |
D | stylesheet.xsl | 7 <xsl:template match="mallctl">
|
/external/jemalloc/include/jemalloc/ |
D | jemalloc_mangle.h | 27 # define mallctl je_mallctl macro
|
D | jemalloc_mangle_jet.h | 27 # define mallctl jet_mallctl macro
|