Home
last modified time | relevance | path

Searched refs:ATOMIC_RELAXED (Results 1 – 23 of 23) sorted by relevance

/external/jemalloc_new/test/unit/
Datomic.c26 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
30 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
31 atomic_store_##ta(&atom, val2, ATOMIC_RELAXED); \
32 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
36 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
37 val = atomic_exchange_##ta(&atom, val2, ATOMIC_RELAXED); \
39 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
46 atomic_store_##ta(&atom, val1, ATOMIC_RELAXED); \
51 &expected, val3, ATOMIC_RELAXED, ATOMIC_RELAXED); \
57 val = atomic_load_##ta(&atom, ATOMIC_RELAXED); \
[all …]
Dprng.c9 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32()
11 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32()
16 atomic_store_u32(&sb, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32()
21 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32()
27 atomic_store_u32(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32()
30 atomic_store_u32(&sb, 42, ATOMIC_RELAXED); in test_prng_lg_range_u32()
82 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu()
84 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu()
89 atomic_store_zu(&sb, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu()
94 atomic_store_zu(&sa, 42, ATOMIC_RELAXED); in test_prng_lg_range_zu()
[all …]
Dlog.c33 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN()
45 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN()
87 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN()
137 atomic_store_b(&log_init_done, true, ATOMIC_RELAXED); in TEST_BEGIN()
159 atomic_store_b(&log_init_done, false, ATOMIC_RELAXED); in TEST_BEGIN()
Dretained.c107 atomic_store_u(&epoch, 0, ATOMIC_RELAXED); in TEST_BEGIN()
/external/jemalloc_new/include/jemalloc/internal/
Darena_stats.h132 return atomic_load_u64(p, ATOMIC_RELAXED); in arena_stats_read_u64()
143 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED); in arena_stats_add_u64()
154 UNUSED uint64_t r = atomic_fetch_sub_u64(p, x, ATOMIC_RELAXED); in arena_stats_sub_u64()
171 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED); in arena_stats_accum_u64()
172 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED); in arena_stats_accum_u64()
181 return atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_read_zu()
184 return atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_read_zu()
192 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED); in arena_stats_add_zu()
195 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_add_zu()
196 atomic_store_zu(p, cur + x, ATOMIC_RELAXED); in arena_stats_add_zu()
[all …]
Dprng.h73 state0 = atomic_load_u32(state, ATOMIC_RELAXED); in prng_lg_range_u32()
79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prng_lg_range_u32()
82 atomic_store_u32(state, state1, ATOMIC_RELAXED); in prng_lg_range_u32()
110 state0 = atomic_load_zu(state, ATOMIC_RELAXED); in prng_lg_range_zu()
116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prng_lg_range_zu()
119 atomic_store_zu(state, state1, ATOMIC_RELAXED); in prng_lg_range_zu()
Dprof_inlines_a.h20 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); in prof_accum_add()
29 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prof_accum_add()
56 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); in prof_accum_cancel()
61 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prof_accum_cancel()
Darena_inlines_a.h11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_add()
16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_sub()
21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED); in arena_internal_get()
Dmutex.h191 ATOMIC_RELAXED); in malloc_mutex_prof_merge()
193 &data->n_waiting_thds, ATOMIC_RELAXED); in malloc_mutex_prof_merge()
195 ATOMIC_RELAXED); in malloc_mutex_prof_merge()
245 atomic_store_u32(&data->n_waiting_thds, 0, ATOMIC_RELAXED); in malloc_mutex_prof_read()
Dbackground_thread_inlines.h6 return atomic_load_b(&background_thread_enabled_state, ATOMIC_RELAXED); in background_thread_enabled()
12 atomic_store_b(&background_thread_enabled_state, state, ATOMIC_RELAXED); in background_thread_enabled_set()
Datomic.h41 #define ATOMIC_RELAXED atomic_memory_order_relaxed macro
Drtree.h176 ? ATOMIC_RELAXED : ATOMIC_ACQUIRE); in rtree_leaf_elm_bits_read()
218 ? ATOMIC_RELAXED : ATOMIC_ACQUIRE); in rtree_leaf_elm_extent_read()
230 return (szind_t)atomic_load_u(&elm->le_szind, dependent ? ATOMIC_RELAXED in rtree_leaf_elm_szind_read()
242 return atomic_load_b(&elm->le_slab, dependent ? ATOMIC_RELAXED : in rtree_leaf_elm_slab_read()
Dlog.h68 ATOMIC_RELAXED); \
/external/jemalloc_new/src/
Dextent_dss.c77 true, ATOMIC_ACQ_REL, ATOMIC_RELAXED)) { in extent_dss_extending_start()
86 assert(atomic_load_b(&dss_extending, ATOMIC_RELAXED)); in extent_dss_extending_finish()
265 atomic_store_b(&dss_extending, false, ATOMIC_RELAXED); in extent_dss_boot()
266 atomic_store_b(&dss_exhausted, dss_base == (void *)-1, ATOMIC_RELAXED); in extent_dss_boot()
267 atomic_store_p(&dss_max, dss_base, ATOMIC_RELAXED); in extent_dss_boot()
Drtree.c70 ATOMIC_RELAXED); in rtree_delete_subtree()
80 ATOMIC_RELAXED); in rtree_delete_subtree()
109 rtree_node_elm_t *node = atomic_load_p(elmp, ATOMIC_RELAXED); in rtree_node_init()
135 rtree_leaf_elm_t *leaf = atomic_load_p(elmp, ATOMIC_RELAXED); in rtree_leaf_init()
170 ATOMIC_RELAXED); in rtree_child_node_tryread()
199 ATOMIC_RELAXED); in rtree_child_leaf_tryread()
Dlog.c66 ATOMIC_RELAXED); in log_var_update_state()
72 LOG_INITIALIZED_NOT_ENABLED, ATOMIC_RELAXED); in log_var_update_state()
Darena.c71 *nactive += atomic_load_zu(&arena->nactive, ATOMIC_RELAXED); in arena_basic_stats_merge()
121 (((atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) + in arena_stats_merge()
154 atomic_store_zu(&astats->tcache_bytes, 0, ATOMIC_RELAXED); in arena_stats_merge()
279 atomic_fetch_add_zu(&arena->nactive, add_pages, ATOMIC_RELAXED); in arena_nactive_add()
284 assert(atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) >= sub_pages); in arena_nactive_sub()
285 atomic_fetch_sub_zu(&arena->nactive, sub_pages, ATOMIC_RELAXED); in arena_nactive_sub()
420 return atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED); in arena_decay_ms_read()
425 atomic_store_zd(&decay->time_ms, decay_ms, ATOMIC_RELAXED); in arena_decay_ms_write()
1033 atomic_store_zu(&arena->nactive, 0, ATOMIC_RELAXED); in arena_reset()
1692 return atomic_load_zd(&dirty_decay_ms_default, ATOMIC_RELAXED); in arena_dirty_decay_ms_default_get()
[all …]
Dmutex.c75 ATOMIC_RELAXED) + 1; in malloc_mutex_lock_slow()
78 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); in malloc_mutex_lock_slow()
86 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); in malloc_mutex_lock_slow()
Dctl.c571 uint64_t cur_dst = atomic_load_u64(dst, ATOMIC_RELAXED); in ctl_accum_arena_stats_u64()
572 uint64_t cur_src = atomic_load_u64(src, ATOMIC_RELAXED); in ctl_accum_arena_stats_u64()
573 atomic_store_u64(dst, cur_dst + cur_src, ATOMIC_RELAXED); in ctl_accum_arena_stats_u64()
583 return atomic_load_u64(p, ATOMIC_RELAXED); in ctl_arena_stats_read_u64()
591 size_t cur_dst = atomic_load_zu(dst, ATOMIC_RELAXED); in accum_atomic_zu()
592 size_t cur_src = atomic_load_zu(src, ATOMIC_RELAXED); in accum_atomic_zu()
593 atomic_store_zu(dst, cur_dst + cur_src, ATOMIC_RELAXED); in accum_atomic_zu()
797 &astats->astats.internal, ATOMIC_RELAXED) == 0); in ctl_arena_stats_sdmerge()
814 ATOMIC_RELAXED) == 0); in ctl_arena_stats_sdmerge()
954 ATOMIC_RELAXED); in ctl_refresh()
[all …]
Dextent.c292 atomic_store_zu(&extents->npages, 0, ATOMIC_RELAXED); in extents_init()
305 return atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_npages_get()
329 atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_insert_locked()
331 ATOMIC_RELAXED); in extents_insert_locked()
354 atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_remove_locked()
357 cur_extents_npages - (size >> LG_PAGE), ATOMIC_RELAXED); in extents_remove_locked()
568 ATOMIC_RELAXED); in extents_evict()
726 ATOMIC_RELAXED) + nadd; in extent_gdump_add()
727 size_t high = atomic_load_zu(&highpages, ATOMIC_RELAXED); in extent_gdump_add()
729 &highpages, &high, cur, ATOMIC_RELAXED, ATOMIC_RELAXED)) { in extent_gdump_add()
[all …]
Dbackground_thread.c126 ssize_t decay_time = atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED);
666 ssize_t decay_time = atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED);
Dbase.c365 atomic_store_p(&base->extent_hooks, extent_hooks, ATOMIC_RELAXED); in base_new()
Dprof.c542 unsigned ngctxs = atomic_fetch_add_u(&cum_gctxs, 1, ATOMIC_RELAXED); in prof_gctx_mutex_choose()
1772 atomic_store_u64(&prof_accum->accumbytes, 0, ATOMIC_RELAXED); in prof_accum_init()