Home
last modified time | relevance | path

Searched refs:tsdn_t (Results 1 – 25 of 46) sorted by relevance

12

/external/jemalloc_new/include/jemalloc/internal/
Darena_externs.h19 void arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena,
22 void arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads,
26 void arena_extents_dirty_dalloc(tsdn_t *tsdn, arena_t *arena,
31 extent_t *arena_extent_alloc_large(tsdn_t *tsdn, arena_t *arena,
33 void arena_extent_dalloc_large_prep(tsdn_t *tsdn, arena_t *arena,
35 void arena_extent_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena,
37 void arena_extent_ralloc_large_expand(tsdn_t *tsdn, arena_t *arena,
40 bool arena_dirty_decay_ms_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_ms);
42 bool arena_muzzy_decay_ms_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_ms);
43 void arena_decay(tsdn_t *tsdn, arena_t *arena, bool is_background_thread,
[all …]
Dextent_externs.h15 extent_t *extent_alloc(tsdn_t *tsdn, arena_t *arena);
16 void extent_dalloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent);
30 bool extents_init(tsdn_t *tsdn, extents_t *extents, extent_state_t state,
34 extent_t *extents_alloc(tsdn_t *tsdn, arena_t *arena,
38 void extents_dalloc(tsdn_t *tsdn, arena_t *arena,
40 extent_t *extents_evict(tsdn_t *tsdn, arena_t *arena,
42 void extents_prefork(tsdn_t *tsdn, extents_t *extents);
43 void extents_postfork_parent(tsdn_t *tsdn, extents_t *extents);
44 void extents_postfork_child(tsdn_t *tsdn, extents_t *extents);
45 extent_t *extent_alloc_wrapper(tsdn_t *tsdn, arena_t *arena,
[all …]
Dprof_externs.h46 void prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize,
59 typedef bool (prof_dump_header_t)(tsdn_t *, bool, const prof_cnt_t *);
65 bool prof_accum_init(tsdn_t *tsdn, prof_accum_t *prof_accum);
66 void prof_idump(tsdn_t *tsdn);
68 void prof_gdump(tsdn_t *tsdn);
73 bool prof_active_get(tsdn_t *tsdn);
74 bool prof_active_set(tsdn_t *tsdn, bool active);
79 bool prof_thread_active_init_get(tsdn_t *tsdn);
80 bool prof_thread_active_init_set(tsdn_t *tsdn, bool active_init);
81 bool prof_gdump_get(tsdn_t *tsdn);
[all …]
Dlarge_externs.h4 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero);
5 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment,
7 bool large_ralloc_no_move(tsdn_t *tsdn, extent_t *extent, size_t usize_min,
9 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent, size_t usize,
18 void large_dalloc_prep_junked_locked(tsdn_t *tsdn, extent_t *extent);
19 void large_dalloc_finish(tsdn_t *tsdn, extent_t *extent);
20 void large_dalloc(tsdn_t *tsdn, extent_t *extent);
21 size_t large_salloc(tsdn_t *tsdn, const extent_t *extent);
22 prof_tctx_t *large_prof_tctx_get(tsdn_t *tsdn, const extent_t *extent);
23 void large_prof_tctx_set(tsdn_t *tsdn, extent_t *extent, prof_tctx_t *tctx);
[all …]
Dbase_externs.h8 base_t *base_new(tsdn_t *tsdn, unsigned ind, extent_hooks_t *extent_hooks);
9 void base_delete(tsdn_t *tsdn, base_t *base);
13 void *base_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment);
14 extent_t *base_alloc_extent(tsdn_t *tsdn, base_t *base);
15 void base_stats_get(tsdn_t *tsdn, base_t *base, size_t *allocated,
17 void base_prefork(tsdn_t *tsdn, base_t *base);
18 void base_postfork_parent(tsdn_t *tsdn, base_t *base);
19 void base_postfork_child(tsdn_t *tsdn, base_t *base);
20 bool base_boot(tsdn_t *tsdn);
Dbackground_thread_externs.h16 void background_thread_interval_check(tsdn_t *tsdn, arena_t *arena,
18 void background_thread_prefork0(tsdn_t *tsdn);
19 void background_thread_prefork1(tsdn_t *tsdn);
20 void background_thread_postfork_parent(tsdn_t *tsdn);
21 void background_thread_postfork_child(tsdn_t *tsdn);
22 bool background_thread_stats_read(tsdn_t *tsdn,
24 void background_thread_ctl_init(tsdn_t *tsdn);
31 bool background_thread_boot1(tsdn_t *tsdn);
Dtcache_externs.h30 size_t tcache_salloc(tsdn_t *tsdn, const void *ptr);
32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
38 void tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache,
42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
46 bool tcache_boot(tsdn_t *tsdn);
47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
48 void tcache_prefork(tsdn_t *tsdn);
49 void tcache_postfork_parent(tsdn_t *tsdn);
50 void tcache_postfork_child(tsdn_t *tsdn);
Drtree.h114 typedef rtree_node_elm_t *(rtree_node_alloc_t)(tsdn_t *, rtree_t *, size_t);
117 typedef rtree_leaf_elm_t *(rtree_leaf_alloc_t)(tsdn_t *, rtree_t *, size_t);
120 typedef void (rtree_node_dalloc_t)(tsdn_t *, rtree_t *, rtree_node_elm_t *);
123 typedef void (rtree_leaf_dalloc_t)(tsdn_t *, rtree_t *, rtree_leaf_elm_t *);
126 void rtree_delete(tsdn_t *tsdn, rtree_t *rtree);
128 rtree_leaf_elm_t *rtree_leaf_elm_lookup_hard(tsdn_t *tsdn, rtree_t *rtree,
173 rtree_leaf_elm_bits_read(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *elm, in rtree_leaf_elm_bits_read()
211 rtree_leaf_elm_extent_read(UNUSED tsdn_t *tsdn, UNUSED rtree_t *rtree, in rtree_leaf_elm_extent_read()
224 rtree_leaf_elm_szind_read(UNUSED tsdn_t *tsdn, UNUSED rtree_t *rtree, in rtree_leaf_elm_szind_read()
236 rtree_leaf_elm_slab_read(UNUSED tsdn_t *tsdn, UNUSED rtree_t *rtree, in rtree_leaf_elm_slab_read()
[all …]
Darena_stats.h98 arena_stats_init(UNUSED tsdn_t *tsdn, arena_stats_t *arena_stats) { in arena_stats_init()
115 arena_stats_lock(tsdn_t *tsdn, arena_stats_t *arena_stats) { in arena_stats_lock()
122 arena_stats_unlock(tsdn_t *tsdn, arena_stats_t *arena_stats) { in arena_stats_unlock()
129 arena_stats_read_u64(tsdn_t *tsdn, arena_stats_t *arena_stats, in arena_stats_read_u64()
140 arena_stats_add_u64(tsdn_t *tsdn, arena_stats_t *arena_stats, in arena_stats_add_u64()
151 arena_stats_sub_u64(tsdn_t *tsdn, arena_stats_t *arena_stats, in arena_stats_sub_u64()
179 arena_stats_read_zu(tsdn_t *tsdn, arena_stats_t *arena_stats, atomic_zu_t *p) { in arena_stats_read_zu()
189 arena_stats_add_zu(tsdn_t *tsdn, arena_stats_t *arena_stats, atomic_zu_t *p, in arena_stats_add_zu()
201 arena_stats_sub_zu(tsdn_t *tsdn, arena_stats_t *arena_stats, atomic_zu_t *p, in arena_stats_sub_zu()
221 arena_stats_large_nrequests_add(tsdn_t *tsdn, arena_stats_t *arena_stats, in arena_stats_large_nrequests_add()
[all …]
Dmutex.h131 void malloc_mutex_prefork(tsdn_t *tsdn, malloc_mutex_t *mutex);
132 void malloc_mutex_postfork_parent(tsdn_t *tsdn, malloc_mutex_t *mutex);
133 void malloc_mutex_postfork_child(tsdn_t *tsdn, malloc_mutex_t *mutex);
135 void malloc_mutex_prof_data_reset(tsdn_t *tsdn, malloc_mutex_t *mutex);
150 mutex_owner_stats_update(tsdn_t *tsdn, malloc_mutex_t *mutex) { in mutex_owner_stats_update()
163 malloc_mutex_trylock(tsdn_t *tsdn, malloc_mutex_t *mutex) { in malloc_mutex_trylock()
201 malloc_mutex_lock(tsdn_t *tsdn, malloc_mutex_t *mutex) { in malloc_mutex_lock()
213 malloc_mutex_unlock(tsdn_t *tsdn, malloc_mutex_t *mutex) { in malloc_mutex_unlock()
221 malloc_mutex_assert_owner(tsdn_t *tsdn, malloc_mutex_t *mutex) { in malloc_mutex_assert_owner()
226 malloc_mutex_assert_not_owner(tsdn_t *tsdn, malloc_mutex_t *mutex) { in malloc_mutex_assert_not_owner()
[all …]
Djemalloc_internal_inlines_c.h27 iaalloc(tsdn_t *tsdn, const void *ptr) { in iaalloc()
34 isalloc(tsdn_t *tsdn, const void *ptr) { in isalloc()
41 iallocztm(tsdn_t *tsdn, size_t size, szind_t ind, bool zero, tcache_t *tcache, in iallocztm()
67 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipallocztm()
87 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipalloct()
99 ivsalloc(tsdn_t *tsdn, const void *ptr) { in ivsalloc()
104 idalloctm(tsdn_t *tsdn, void *ptr, tcache_t *tcache, alloc_ctx_t *alloc_ctx, in idalloctm()
127 isdalloct(tsdn_t *tsdn, void *ptr, size_t size, tcache_t *tcache, in isdalloct()
135 iralloct_realign(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t size, in iralloct_realign()
173 iralloct(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t size, size_t alignment, in iralloct()
[all …]
Dmutex_pool.h27 mutex_pool_assert_not_held(tsdn_t *tsdn, mutex_pool_t *pool) { in mutex_pool_assert_not_held()
41 mutex_pool_lock(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key) { in mutex_pool_lock()
49 mutex_pool_unlock(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key) { in mutex_pool_unlock()
57 mutex_pool_lock2(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key1, in mutex_pool_lock2()
75 mutex_pool_unlock2(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key1, in mutex_pool_unlock2()
90 mutex_pool_assert_owner(tsdn_t *tsdn, mutex_pool_t *pool, uintptr_t key) { in mutex_pool_assert_owner()
Darena_inlines_b.h12 arena_prof_tctx_get(tsdn_t *tsdn, const void *ptr, alloc_ctx_t *alloc_ctx) { in arena_prof_tctx_get()
31 arena_prof_tctx_set(tsdn_t *tsdn, const void *ptr, UNUSED size_t usize, in arena_prof_tctx_set()
50 arena_prof_tctx_reset(tsdn_t *tsdn, const void *ptr, UNUSED prof_tctx_t *tctx) { in arena_prof_tctx_reset()
61 arena_decay_ticks(tsdn_t *tsdn, arena_t *arena, unsigned nticks) { in arena_decay_ticks()
79 arena_decay_tick(tsdn_t *tsdn, arena_t *arena) { in arena_decay_tick()
87 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero, in arena_malloc()
109 arena_aalloc(tsdn_t *tsdn, const void *ptr) { in arena_aalloc()
114 arena_salloc(tsdn_t *tsdn, const void *ptr) { in arena_salloc()
128 arena_vsalloc(tsdn_t *tsdn, const void *ptr) { in arena_vsalloc()
161 arena_dalloc_no_tcache(tsdn_t *tsdn, void *ptr) { in arena_dalloc_no_tcache()
[all …]
Dtsd.h135 #define TSDN_NULL ((tsdn_t *)0)
136 JEMALLOC_ALWAYS_INLINE tsdn_t *
138 return (tsdn_t *)tsd; in tsd_tsdn()
142 tsdn_null(const tsdn_t *tsdn) { in tsdn_null()
147 tsdn_tsd(tsdn_t *tsdn) { in tsdn_tsd()
209 tsdn_##n##p_get(tsdn_t *tsdn) { \
299 JEMALLOC_ALWAYS_INLINE tsdn_t *
314 tsdn_rtree_ctx(tsdn_t *tsdn, rtree_ctx_t *fallback) { in tsdn_rtree_ctx()
Dbin.h85 void bin_prefork(tsdn_t *tsdn, bin_t *bin);
86 void bin_postfork_parent(tsdn_t *tsdn, bin_t *bin);
87 void bin_postfork_child(tsdn_t *tsdn, bin_t *bin);
91 bin_stats_merge(tsdn_t *tsdn, bin_stats_t *dst_bin_stats, bin_t *bin) { in bin_stats_merge()
Dctl.h30 const ctl_named_node_t *(*index)(tsdn_t *, const size_t *, size_t,
100 void ctl_prefork(tsdn_t *tsdn);
101 void ctl_postfork_parent(tsdn_t *tsdn);
102 void ctl_postfork_child(tsdn_t *tsdn);
/external/jemalloc_new/src/
Dlarge.c14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc()
21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc()
104 large_ralloc_no_move_shrink(tsdn_t *tsdn, extent_t *extent, size_t usize) { in large_ralloc_no_move_shrink()
139 large_ralloc_no_move_expand(tsdn_t *tsdn, extent_t *extent, size_t usize, in large_ralloc_no_move_expand()
227 large_ralloc_no_move(tsdn_t *tsdn, extent_t *extent, size_t usize_min, in large_ralloc_no_move()
272 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, in large_ralloc_move_helper()
281 large_ralloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent, size_t usize, in large_ralloc()
317 large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent, in large_dalloc_prep_impl()
338 large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent) { in large_dalloc_finish_impl()
344 large_dalloc_prep_junked_locked(tsdn_t *tsdn, extent_t *extent) { in large_dalloc_prep_junked_locked()
[all …]
Dextent.c34 static bool extent_commit_impl(tsdn_t *tsdn, arena_t *arena,
43 static bool extent_purge_lazy_impl(tsdn_t *tsdn, arena_t *arena,
50 static bool extent_purge_forced_impl(tsdn_t *tsdn, arena_t *arena,
58 static extent_t *extent_split_impl(tsdn_t *tsdn, arena_t *arena,
67 static bool extent_merge_impl(tsdn_t *tsdn, arena_t *arena,
108 static void extent_deregister(tsdn_t *tsdn, extent_t *extent);
109 static extent_t *extent_recycle(tsdn_t *tsdn, arena_t *arena,
113 static extent_t *extent_try_coalesce(tsdn_t *tsdn, arena_t *arena,
116 static void extent_record(tsdn_t *tsdn, arena_t *arena,
132 extent_rtree_leaf_elm_try_lock(tsdn_t *tsdn, rtree_leaf_elm_t *elm, in extent_rtree_leaf_elm_try_lock()
[all …]
Darena.c51 static void arena_decay_to_limit(tsdn_t *tsdn, arena_t *arena,
54 static bool arena_decay_dirty(tsdn_t *tsdn, arena_t *arena,
56 static void arena_dalloc_bin_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab,
58 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab,
64 arena_basic_stats_merge(UNUSED tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_basic_stats_merge()
77 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_stats_merge()
209 arena_extents_dirty_dalloc(tsdn_t *tsdn, arena_t *arena, in arena_extents_dirty_dalloc()
289 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update()
305 arena_large_dalloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_dalloc_stats_update()
321 arena_large_ralloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t oldusize, in arena_large_ralloc_stats_update()
[all …]
Drtree.c31 rtree_node_alloc_impl(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { in rtree_node_alloc_impl()
38 rtree_node_dalloc_impl(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *node) { in rtree_node_dalloc_impl()
46 rtree_leaf_alloc_impl(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { in rtree_leaf_alloc_impl()
53 rtree_leaf_dalloc_impl(tsdn_t *tsdn, rtree_t *rtree, rtree_leaf_elm_t *leaf) { in rtree_leaf_dalloc_impl()
63 rtree_delete_subtree(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *subtree, in rtree_delete_subtree()
94 rtree_delete(tsdn_t *tsdn, rtree_t *rtree) { in rtree_delete()
102 rtree_node_init(tsdn_t *tsdn, rtree_t *rtree, unsigned level, in rtree_node_init()
129 rtree_leaf_init(tsdn_t *tsdn, rtree_t *rtree, atomic_p_t *elmp) { in rtree_leaf_init()
181 rtree_child_node_read(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *elm, in rtree_child_node_read()
210 rtree_child_leaf_read(tsdn_t *tsdn, rtree_t *rtree, rtree_node_elm_t *elm, in rtree_child_leaf_read()
[all …]
Dbackground_thread.c59 void background_thread_interval_check(tsdn_t *tsdn, arena_t *arena, in background_thread_create()
61 void background_thread_prefork0(tsdn_t *tsdn) NOT_REACHED in background_thread_create()
62 void background_thread_prefork1(tsdn_t *tsdn) NOT_REACHED in background_thread_create()
63 void background_thread_postfork_parent(tsdn_t *tsdn) NOT_REACHED in background_thread_create()
64 void background_thread_postfork_child(tsdn_t *tsdn) NOT_REACHED in background_thread_create()
65 bool background_thread_stats_read(tsdn_t *tsdn, in background_thread_create()
67 void background_thread_ctl_init(tsdn_t *tsdn) NOT_REACHED in background_thread_create()
74 background_thread_info_init(tsdn_t *tsdn, background_thread_info_t *info) {
118 arena_decay_compute_purge_interval_impl(tsdn_t *tsdn, arena_decay_t *decay,
204 arena_decay_compute_purge_interval(tsdn_t *tsdn, arena_t *arena) {
[all …]
Dbase.c32 base_map(tsdn_t *tsdn, extent_hooks_t *extent_hooks, unsigned ind, size_t size) { in base_map()
55 base_unmap(tsdn_t *tsdn, extent_hooks_t *extent_hooks, unsigned ind, void *addr, in base_unmap()
144 base_auto_thp_switch(tsdn_t *tsdn, base_t *base) { in base_auto_thp_switch()
248 base_block_alloc(tsdn_t *tsdn, base_t *base, extent_hooks_t *extent_hooks, in base_block_alloc()
308 base_extent_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment) { in base_extent_alloc()
350 base_new(tsdn_t *tsdn, unsigned ind, extent_hooks_t *extent_hooks) { in base_new()
396 base_delete(tsdn_t *tsdn, base_t *base) { in base_delete()
421 base_alloc_impl(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment, in base_alloc_impl()
464 base_alloc(tsdn_t *tsdn, base_t *base, size_t size, size_t alignment) { in base_alloc()
469 base_alloc_extent(tsdn_t *tsdn, base_t *base) { in base_alloc_extent()
[all …]
Dprof.c140 static bool prof_tctx_should_destroy(tsdn_t *tsdn, prof_tctx_t *tctx);
142 static bool prof_tdata_should_destroy(tsdn_t *tsdn, prof_tdata_t *tdata,
146 static char *prof_thread_name_alloc(tsdn_t *tsdn, const char *thread_name);
241 prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, in prof_malloc_sample_object()
553 prof_gctx_create(tsdn_t *tsdn, prof_bt_t *bt) { in prof_gctx_create()
614 prof_tctx_should_destroy(tsdn_t *tsdn, prof_tctx_t *tctx) { in prof_tctx_should_destroy()
926 tsdn_t *tsdn; in prof_tdata_count()
1056 prof_tctx_merge_tdata(tsdn_t *tsdn, prof_tctx_t *tctx, prof_tdata_t *tdata) { in prof_tctx_merge_tdata()
1087 prof_tctx_merge_gctx(tsdn_t *tsdn, prof_tctx_t *tctx, prof_gctx_t *gctx) { in prof_tctx_merge_gctx()
1100 tsdn_t *tsdn = (tsdn_t *)arg; in prof_tctx_merge_iter()
[all …]
Dbin.c38 bin_prefork(tsdn_t *tsdn, bin_t *bin) { in bin_prefork()
43 bin_postfork_parent(tsdn_t *tsdn, bin_t *bin) { in bin_postfork_parent()
48 bin_postfork_child(tsdn_t *tsdn, bin_t *bin) { in bin_postfork_child()
/external/jemalloc_new/test/unit/
Drtree.c14 rtree_node_alloc_intercept(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { in rtree_node_alloc_intercept()
30 rtree_node_dalloc_intercept(tsdn_t *tsdn, rtree_t *rtree, in rtree_node_dalloc_intercept()
41 rtree_leaf_alloc_intercept(tsdn_t *tsdn, rtree_t *rtree, size_t nelms) { in rtree_leaf_alloc_intercept()
57 rtree_leaf_dalloc_intercept(tsdn_t *tsdn, rtree_t *rtree, in rtree_leaf_dalloc_intercept()
68 tsdn_t *tsdn; in TEST_BEGIN()
94 tsdn_t *tsdn = tsdn_fetch(); in TEST_BEGIN()
122 tsdn_t *tsdn = tsdn_fetch(); in TEST_BEGIN()
162 tsdn_t *tsdn = tsdn_fetch(); in TEST_BEGIN()

12