/external/jemalloc_new/test/unit/ |
D | atomic.c | 26 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 30 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 31 atomic_store_##ta(&atom, val2, ATOMIC_RELAXED); \ 32 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 36 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 37 val = atomic_exchange_##ta(&atom, val2, ATOMIC_RELAXED); \ 39 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ 46 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \ 51 &expected, val3, ATOMIC_RELAXED, ATOMIC_RELAXED); \ 57 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \ [all …]
|
D | prng.c | 9 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32() 11 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32() 16 atomic_store_u32(&sb, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32() 21 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32() 27 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32() 30 atomic_store_u32(&sb, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32() 82 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu() 84 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu() 89 atomic_store_zu(&sb, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu() 94 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu() [all …]
|
D | log.c | 33 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN() 45 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN() 87 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN() 137 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN() 159 atomic_store_b(&log_init_done, false, ATOMIC_RELAXED); in TEST_BEGIN()
|
D | retained.c | 107 atomic_store_u(&epoch, 0, ATOMIC_RELAXED); in TEST_BEGIN()
|
/external/jemalloc_new/include/jemalloc/internal/ |
D | arena_stats.h | 132 return atomic_load_u64(p, ATOMIC_RELAXED); in arena_stats_read_u64() 143 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED); in arena_stats_add_u64() 154 UNUSED uint64_t r = atomic_fetch_sub_u64(p, x, ATOMIC_RELAXED); in arena_stats_sub_u64() 171 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED); in arena_stats_accum_u64() 172 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED); in arena_stats_accum_u64() 181 return atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_read_zu() 184 return atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_read_zu() 192 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED); in arena_stats_add_zu() 195 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_add_zu() 196 atomic_store_zu(p, cur + x, ATOMIC_RELAXED); in arena_stats_add_zu() [all …]
|
D | prng.h | 73 state0 = atomic_load_u32(state, ATOMIC_RELAXED); in prng_lg_range_u32() 79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prng_lg_range_u32() 82 atomic_store_u32(state, state1, ATOMIC_RELAXED); in prng_lg_range_u32() 110 state0 = atomic_load_zu(state, ATOMIC_RELAXED); in prng_lg_range_zu() 116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prng_lg_range_zu() 119 atomic_store_zu(state, state1, ATOMIC_RELAXED); in prng_lg_range_zu()
|
D | prof_inlines_a.h | 20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); in prof_accum_add() 29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prof_accum_add() 56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); in prof_accum_cancel() 61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prof_accum_cancel()
|
D | arena_inlines_a.h | 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_add() 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_sub() 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED); in arena_internal_get()
|
D | mutex.h | 191 ATOMIC_RELAXED); in malloc_mutex_prof_merge() 193 &data->n_waiting_thds, ATOMIC_RELAXED); in malloc_mutex_prof_merge() 195 ATOMIC_RELAXED); in malloc_mutex_prof_merge() 245 atomic_store_u32(&data->n_waiting_thds, 0, ATOMIC_RELAXED); in malloc_mutex_prof_read()
|
D | background_thread_inlines.h | 6 return atomic_load_b(&background_thread_enabled_state, ATOMIC_RELAXED); in background_thread_enabled() 12 atomic_store_b(&background_thread_enabled_state, state, ATOMIC_RELAXED); in background_thread_enabled_set()
|
D | atomic.h | 41 #define ATOMIC_RELAXED atomic_memory_order_relaxed macro
|
D | rtree.h | 176 ? ATOMIC_RELAXED : ATOMIC_ACQUIRE); in rtree_leaf_elm_bits_read() 218 ? ATOMIC_RELAXED : ATOMIC_ACQUIRE); in rtree_leaf_elm_extent_read() 230 return (szind_t)atomic_load_u(&elm->le_szind, dependent ? ATOMIC_RELAXED in rtree_leaf_elm_szind_read() 242 return atomic_load_b(&elm->le_slab, dependent ? ATOMIC_RELAXED : in rtree_leaf_elm_slab_read()
|
D | log.h | 68 ATOMIC_RELAXED); \
|
/external/jemalloc_new/src/ |
D | extent_dss.c | 77 true, ATOMIC_ACQ_REL, ATOMIC_RELAXED)) { in extent_dss_extending_start() 86 assert(atomic_load_b(&dss_extending, ATOMIC_RELAXED)); in extent_dss_extending_finish() 265 atomic_store_b(&dss_extending, false, ATOMIC_RELAXED); in extent_dss_boot() 266 atomic_store_b(&dss_exhausted, dss_base == (void *)-1, ATOMIC_RELAXED); in extent_dss_boot() 267 atomic_store_p(&dss_max, dss_base, ATOMIC_RELAXED); in extent_dss_boot()
|
D | rtree.c | 70 ATOMIC_RELAXED); in rtree_delete_subtree() 80 ATOMIC_RELAXED); in rtree_delete_subtree() 109 rtree_node_elm_t *node = atomic_load_p(elmp, ATOMIC_RELAXED); in rtree_node_init() 135 rtree_leaf_elm_t *leaf = atomic_load_p(elmp, ATOMIC_RELAXED); in rtree_leaf_init() 170 ATOMIC_RELAXED); in rtree_child_node_tryread() 199 ATOMIC_RELAXED); in rtree_child_leaf_tryread()
|
D | log.c | 66 ATOMIC_RELAXED); in log_var_update_state() 72 LOG_INITIALIZED_NOT_ENABLED, ATOMIC_RELAXED); in log_var_update_state()
|
D | arena.c | 71 *nactive += atomic_load_zu(&arena->nactive, ATOMIC_RELAXED); in arena_basic_stats_merge() 121 (((atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) + in arena_stats_merge() 154 atomic_store_zu(&astats->tcache_bytes, 0, ATOMIC_RELAXED); in arena_stats_merge() 279 atomic_fetch_add_zu(&arena->nactive, add_pages, ATOMIC_RELAXED); in arena_nactive_add() 284 assert(atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) >= sub_pages); in arena_nactive_sub() 285 atomic_fetch_sub_zu(&arena->nactive, sub_pages, ATOMIC_RELAXED); in arena_nactive_sub() 420 return atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED); in arena_decay_ms_read() 425 atomic_store_zd(&decay->time_ms, decay_ms, ATOMIC_RELAXED); in arena_decay_ms_write() 1033 atomic_store_zu(&arena->nactive, 0, ATOMIC_RELAXED); in arena_reset() 1692 return atomic_load_zd(&dirty_decay_ms_default, ATOMIC_RELAXED); in arena_dirty_decay_ms_default_get() [all …]
|
D | mutex.c | 75 ATOMIC_RELAXED) + 1; in malloc_mutex_lock_slow() 78 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); in malloc_mutex_lock_slow() 86 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); in malloc_mutex_lock_slow()
|
D | ctl.c | 571 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED); in ctl_accum_arena_stats_u64() 572 uint64_t cur_src = atomic_load_u64(src, ATOMIC_RELAXED); in ctl_accum_arena_stats_u64() 573 atomic_store_u64(dst, cur_dst + cur_src, ATOMIC_RELAXED); in ctl_accum_arena_stats_u64() 583 return atomic_load_u64(p, ATOMIC_RELAXED); in ctl_arena_stats_read_u64() 591 size_t cur_dst = atomic_load_zu(dst, ATOMIC_RELAXED); in accum_atomic_zu() 592 size_t cur_src = atomic_load_zu(src, ATOMIC_RELAXED); in accum_atomic_zu() 593 atomic_store_zu(dst, cur_dst + cur_src, ATOMIC_RELAXED); in accum_atomic_zu() 797 &astats->astats.internal, ATOMIC_RELAXED) == 0); in ctl_arena_stats_sdmerge() 814 ATOMIC_RELAXED) == 0); in ctl_arena_stats_sdmerge() 954 ATOMIC_RELAXED); in ctl_refresh() [all …]
|
D | extent.c | 292 atomic_store_zu(&extents->npages, 0, ATOMIC_RELAXED); in extents_init() 305 return atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_npages_get() 329 atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_insert_locked() 331 ATOMIC_RELAXED); in extents_insert_locked() 354 atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_remove_locked() 357 cur_extents_npages - (size >> LG_PAGE), ATOMIC_RELAXED); in extents_remove_locked() 568 ATOMIC_RELAXED); in extents_evict() 726 ATOMIC_RELAXED) + nadd; in extent_gdump_add() 727 size_t high = atomic_load_zu(&highpages, ATOMIC_RELAXED); in extent_gdump_add() 729 &highpages, &high, cur, ATOMIC_RELAXED, ATOMIC_RELAXED)) { in extent_gdump_add() [all …]
|
D | background_thread.c | 126 ssize_t decay_time = atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED); 666 ssize_t decay_time = atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED);
|
D | base.c | 365 atomic_store_p(&base->extent_hooks, extent_hooks, ATOMIC_RELAXED); in base_new()
|
D | prof.c | 542 unsigned ngctxs = atomic_fetch_add_u(&cum_gctxs, 1, ATOMIC_RELAXED); in prof_gctx_mutex_choose() 1772 atomic_store_u64(&prof_accum->accumbytes, 0, ATOMIC_RELAXED); in prof_accum_init()
|