/external/jemalloc_new/include/jemalloc/internal/ |
D | arena_externs.h | 19 void arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, 22 void arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, 26 void arena_extents_dirty_dalloc(tsdn_t *tsdn, arena_t *arena, 31 extent_t *arena_extent_alloc_large(tsdn_t *tsdn, arena_t *arena, 33 void arena_extent_dalloc_large_prep(tsdn_t *tsdn, arena_t *arena, 35 void arena_extent_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena, 37 void arena_extent_ralloc_large_expand(tsdn_t *tsdn, arena_t *arena, 39 ssize_t arena_dirty_decay_ms_get(arena_t *arena); 40 bool arena_dirty_decay_ms_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_ms); 41 ssize_t arena_muzzy_decay_ms_get(arena_t *arena); [all …]
|
D | extent_externs.h | 15 extent_t *extent_alloc(tsdn_t *tsdn, arena_t *arena); 16 void extent_dalloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent); 18 extent_hooks_t *extent_hooks_get(arena_t *arena); 19 extent_hooks_t *extent_hooks_set(tsd_t *tsd, arena_t *arena, 34 extent_t *extents_alloc(tsdn_t *tsdn, arena_t *arena, 38 void extents_dalloc(tsdn_t *tsdn, arena_t *arena, 40 extent_t *extents_evict(tsdn_t *tsdn, arena_t *arena, 45 extent_t *extent_alloc_wrapper(tsdn_t *tsdn, arena_t *arena, 48 void extent_dalloc_gap(tsdn_t *tsdn, arena_t *arena, extent_t *extent); 49 void extent_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena, [all …]
|
D | jemalloc_internal_inlines_b.h | 7 static inline arena_t * 8 arena_choose_impl(tsd_t *tsd, arena_t *arena, bool internal) { in arena_choose_impl() 9 arena_t *ret; in arena_choose_impl() 61 static inline arena_t * 62 arena_choose(tsd_t *tsd, arena_t *arena) { in arena_choose() 66 static inline arena_t * 67 arena_ichoose(tsd_t *tsd, arena_t *arena) { in arena_ichoose() 72 arena_is_auto(arena_t *arena) { in arena_is_auto()
|
D | arena_inlines_a.h | 5 arena_ind_get(const arena_t *arena) { in arena_ind_get() 10 arena_internal_add(arena_t *arena, size_t size) { in arena_internal_add() 15 arena_internal_sub(arena_t *arena, size_t size) { in arena_internal_sub() 20 arena_internal_get(arena_t *arena) { in arena_internal_get() 25 arena_prof_accum(tsdn_t *tsdn, arena_t *arena, uint64_t accumbytes) { in arena_prof_accum() 38 arena_t *oldarena = tsd_arena_get(tsd); in percpu_arena_update() 44 arena_t *newarena = arena_get(tsd_tsdn(tsd), newind, true); in percpu_arena_update()
|
D | tcache_externs.h | 32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 39 arena_t *arena); 42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
|
D | jemalloc_internal_externs.h | 39 void arena_set(unsigned ind, arena_t *arena); 41 arena_t *arena_init(tsdn_t *tsdn, unsigned ind, extent_hooks_t *extent_hooks); 43 arena_t *arena_choose_hard(tsd_t *tsd, bool internal);
|
D | large_externs.h | 4 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 5 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, 9 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent, size_t usize,
|
D | jemalloc_internal_inlines_a.h | 82 static inline arena_t * 84 arena_t *ret; in arena_get() 88 ret = (arena_t *)atomic_load_p(&arenas[ind], ATOMIC_ACQUIRE); in arena_get() 149 pre_reentrancy(tsd_t *tsd, arena_t *arena) { in pre_reentrancy()
|
D | jemalloc_internal_inlines_c.h | 26 JEMALLOC_ALWAYS_INLINE arena_t * 42 bool is_internal, arena_t *arena, bool slow_path) { in iallocztm() 68 tcache_t *tcache, bool is_internal, arena_t *arena) { in ipallocztm() 88 tcache_t *tcache, arena_t *arena) { in ipalloct() 137 arena_t *arena) { in iralloct_realign() 174 bool zero, tcache_t *tcache, arena_t *arena) { in iralloct()
|
D | tsd.h | 73 O(iarena, arena_t *, arena_t *) \ 74 O(arena, arena_t *, arena_t *) \
|
D | background_thread_inlines.h | 16 arena_background_thread_info_get(arena_t *arena) { in arena_background_thread_info_get() 44 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, in arena_background_thread_inactivity_check()
|
D | arena_inlines_b.h | 61 arena_decay_ticks(tsdn_t *tsdn, arena_t *arena, unsigned nticks) { in arena_decay_ticks() 79 arena_decay_tick(tsdn_t *tsdn, arena_t *arena) { in arena_decay_tick() 87 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero, in arena_malloc() 108 JEMALLOC_ALWAYS_INLINE arena_t *
|
D | extent_inlines.h | 37 static inline arena_t * 49 return (arena_t *)atomic_load_p(&arenas[arena_ind], ATOMIC_ACQUIRE); in extent_arena_get() 180 extent_arena_set(extent_t *extent, arena_t *arena) { in extent_arena_set() 303 extent_init(extent_t *extent, arena_t *arena, void *addr, size_t size, in extent_init()
|
D | tcache_structs.h | 40 arena_t *arena;
|
D | extent_dss.h | 20 void *extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr,
|
D | arena_types.h | 21 typedef struct arena_s arena_t; typedef
|
D | background_thread_externs.h | 16 void background_thread_interval_check(tsdn_t *tsdn, arena_t *arena,
|
D | tcache_inlines.h | 42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small() 106 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_large()
|
/external/jemalloc_new/src/ |
D | arena.c | 51 static void arena_decay_to_limit(tsdn_t *tsdn, arena_t *arena, 54 static bool arena_decay_dirty(tsdn_t *tsdn, arena_t *arena, 56 static void arena_dalloc_bin_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab, 58 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab, 64 arena_basic_stats_merge(UNUSED tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_basic_stats_merge() 77 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_stats_merge() 209 arena_extents_dirty_dalloc(tsdn_t *tsdn, arena_t *arena, in arena_extents_dirty_dalloc() 278 arena_nactive_add(arena_t *arena, size_t add_pages) { in arena_nactive_add() 283 arena_nactive_sub(arena_t *arena, size_t sub_pages) { in arena_nactive_sub() 289 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update() [all …]
|
D | extent.c | 34 static bool extent_commit_impl(tsdn_t *tsdn, arena_t *arena, 43 static bool extent_purge_lazy_impl(tsdn_t *tsdn, arena_t *arena, 50 static bool extent_purge_forced_impl(tsdn_t *tsdn, arena_t *arena, 58 static extent_t *extent_split_impl(tsdn_t *tsdn, arena_t *arena, 67 static bool extent_merge_impl(tsdn_t *tsdn, arena_t *arena, 109 static extent_t *extent_recycle(tsdn_t *tsdn, arena_t *arena, 113 static extent_t *extent_try_coalesce(tsdn_t *tsdn, arena_t *arena, 116 static void extent_record(tsdn_t *tsdn, arena_t *arena, 177 extent_alloc(tsdn_t *tsdn, arena_t *arena) { in extent_alloc() 190 extent_dalloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent) { in extent_dalloc() [all …]
|
D | android_je_mallinfo.c | 17 static size_t accumulate_large_allocs(arena_t* arena) { in accumulate_large_allocs() 34 static size_t accumulate_small_allocs(arena_t* arena) { in accumulate_small_allocs() 57 arena_t* arena = atomic_load_p(&arenas[i], ATOMIC_ACQUIRE); in je_mallinfo() 85 arena_t* arena = atomic_load_p(&arenas[aidx], ATOMIC_ACQUIRE); in je_mallinfo_arena_info() 102 arena_t* arena = atomic_load_p(&arenas[aidx], ATOMIC_ACQUIRE); in je_mallinfo_bin_info()
|
D | large.c | 14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc() 21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc() 105 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_shrink() 141 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_expand() 272 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, in large_ralloc_move_helper() 281 large_ralloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent, size_t usize, in large_ralloc() 317 large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent, in large_dalloc_prep_impl() 338 large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent) { in large_dalloc_finish_impl() 355 arena_t *arena = extent_arena_get(extent); in large_dalloc()
|
D | tcache.c | 92 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, in tcache_alloc_small_hard() 115 arena_t *arena = tcache->arena; in tcache_bin_flush_small() 127 arena_t *bin_arena = extent_arena_get(extent); in tcache_bin_flush_small() 206 arena_t *arena = tcache->arena; in tcache_bin_flush_large() 218 arena_t *locked_arena = extent_arena_get(extent); in tcache_bin_flush_large() 301 tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_arena_associate() 323 arena_t *arena = tcache->arena; in tcache_arena_dissociate() 349 tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_arena_reassociate() 427 arena_t *arena; in tsd_tcache_data_init() 547 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_stats_merge()
|
D | jemalloc.c | 88 static arena_t *a0; /* arenas[0]; read-only after initialization. */ 291 arena_set(unsigned ind, arena_t *arena) { in arena_set() 311 static arena_t * 313 arena_t *arena; in arena_init_locked() 357 arena_t * 359 arena_t *arena; in arena_init() 372 arena_t *arena = arena_get(tsd_tsdn(tsd), ind, false); in arena_bind() 384 arena_t *oldarena, *newarena; in arena_migrate() 395 arena_t *arena; in arena_unbind() 484 arena_t * [all …]
|
D | ctl.c | 707 ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_t *ctl_arena, arena_t *arena) { in ctl_arena_stats_amerge() 874 ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, ctl_arena_t *ctl_sdarena, in ctl_arena_refresh() 928 VARIABLE_ARRAY(arena_t *, tarenas, ctl_arenas->narenas); in ctl_refresh() 1696 arena_t *oldarena; in CTL_RO_CONFIG_GEN() 1708 arena_t *newarena; in CTL_RO_CONFIG_GEN() 1953 VARIABLE_ARRAY(arena_t *, tarenas, narenas); in arena_i_decay() 1972 arena_t *tarena; in arena_i_decay() 2023 arena_t **arena) { in arena_i_reset_destroy_helper() 2079 arena_t *arena; in arena_i_reset_ctl() 2099 arena_t *arena; in arena_i_destroy_ctl() [all …]
|