/external/jemalloc/src/ |
D | arena.c | 40 static void arena_purge_to_limit(arena_t *arena, size_t ndirty_limit); 41 static void arena_run_dalloc(arena_t *arena, arena_run_t *run, bool dirty, 43 static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, 45 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, 203 arena_runs_avail_get(arena_t *arena, szind_t ind) in arena_runs_avail_get() argument 209 return (&arena->runs_avail[ind - runs_avail_bias]); in arena_runs_avail_get() 213 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, in arena_avail_insert() argument 220 arena_run_tree_insert(arena_runs_avail_get(arena, ind), in arena_avail_insert() 225 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, in arena_avail_remove() argument 232 arena_run_tree_remove(arena_runs_avail_get(arena, ind), in arena_avail_remove() [all …]
|
D | chunk.c | 52 static void chunk_record(arena_t *arena, chunk_hooks_t *chunk_hooks, 59 chunk_hooks_get_locked(arena_t *arena) in chunk_hooks_get_locked() argument 62 return (arena->chunk_hooks); in chunk_hooks_get_locked() 66 chunk_hooks_get(arena_t *arena) in chunk_hooks_get() argument 70 malloc_mutex_lock(&arena->chunks_mtx); in chunk_hooks_get() 71 chunk_hooks = chunk_hooks_get_locked(arena); in chunk_hooks_get() 72 malloc_mutex_unlock(&arena->chunks_mtx); in chunk_hooks_get() 78 chunk_hooks_set(arena_t *arena, const chunk_hooks_t *chunk_hooks) in chunk_hooks_set() argument 82 malloc_mutex_lock(&arena->chunks_mtx); in chunk_hooks_set() 83 old_chunk_hooks = arena->chunk_hooks; in chunk_hooks_set() [all …]
|
D | huge.c | 34 huge_malloc(tsd_t *tsd, arena_t *arena, size_t usize, bool zero, in huge_malloc() argument 40 return (huge_palloc(tsd, arena, usize, chunksize, zero, tcache)); in huge_malloc() 44 huge_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, in huge_palloc() argument 61 CACHELINE, false, tcache, true, arena); in huge_palloc() 77 arena = a0get(); in huge_palloc() 79 arena = arena_choose(tsd, arena); in huge_palloc() 82 if (unlikely(arena == NULL) || (ret = arena_chunk_alloc_huge(arena, in huge_palloc() 88 extent_node_init(node, arena, ret, usize, is_zeroed, true); in huge_palloc() 91 arena_chunk_dalloc_huge(arena, ret, usize); in huge_palloc() 97 malloc_mutex_lock(&arena->huge_mtx); in huge_palloc() [all …]
|
D | tcache.c | 73 tcache_alloc_small_hard(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small_hard() argument 78 arena_tcache_fill_small(tsd, arena, tbin, binind, config_prof ? in tcache_alloc_small_hard() 91 arena_t *arena; in tcache_bin_flush_small() local 99 arena = arena_choose(tsd, NULL); in tcache_bin_flush_small() 100 assert(arena != NULL); in tcache_bin_flush_small() 108 if (config_prof && bin_arena == arena) { in tcache_bin_flush_small() 109 if (arena_prof_accum(arena, tcache->prof_accumbytes)) in tcache_bin_flush_small() 115 if (config_stats && bin_arena == arena) { in tcache_bin_flush_small() 153 arena_bin_t *bin = &arena->bins[binind]; in tcache_bin_flush_small() 172 arena_t *arena; in tcache_bin_flush_large() local [all …]
|
D | jemalloc.c | 387 arena_set(unsigned ind, arena_t *arena) in arena_set() argument 390 atomic_write_p((void **)&arenas[ind], arena); in arena_set() 418 arena_t *arena; in arena_init_locked() local 430 arena = arena_get(ind, false); in arena_init_locked() 431 if (arena != NULL) { in arena_init_locked() 433 return (arena); in arena_init_locked() 437 arena = arena_new(ind); in arena_init_locked() 438 arena_set(ind, arena); in arena_init_locked() 439 return (arena); in arena_init_locked() 445 arena_t *arena; in arena_init() local [all …]
|
D | ctl.c | 55 arena_t *arena); 58 static void ctl_arena_refresh(arena_t *arena, unsigned i); 495 {NAME("arena"), CHILD(indexed, arena)}, 557 ctl_arena_stats_amerge(ctl_arena_stats_t *cstats, arena_t *arena) in ctl_arena_stats_amerge() argument 562 arena_stats_merge(arena, &cstats->nthreads, &cstats->dss, in ctl_arena_stats_amerge() 575 arena_basic_stats_merge(arena, &cstats->nthreads, &cstats->dss, in ctl_arena_stats_amerge() 652 ctl_arena_refresh(arena_t *arena, unsigned i) in ctl_arena_refresh() argument 658 ctl_arena_stats_amerge(astats, arena); in ctl_arena_refresh() 1652 arena_t *arena = arena_get(arena_ind, false); in arena_i_dss_ctl() local 1653 if (arena == NULL || (dss_prec != dss_prec_limit && in arena_i_dss_ctl() [all …]
|
/external/jemalloc/include/jemalloc/internal/ |
D | arena.h | 510 void arena_chunk_cache_maybe_insert(arena_t *arena, extent_node_t *node, 512 void arena_chunk_cache_maybe_remove(arena_t *arena, extent_node_t *node, 514 extent_node_t *arena_node_alloc(arena_t *arena); 515 void arena_node_dalloc(arena_t *arena, extent_node_t *node); 516 void *arena_chunk_alloc_huge(arena_t *arena, size_t usize, size_t alignment, 518 void arena_chunk_dalloc_huge(arena_t *arena, void *chunk, size_t usize); 519 void arena_chunk_ralloc_huge_similar(arena_t *arena, void *chunk, 521 void arena_chunk_ralloc_huge_shrink(arena_t *arena, void *chunk, 523 bool arena_chunk_ralloc_huge_expand(arena_t *arena, void *chunk, 525 ssize_t arena_lg_dirty_mult_get(arena_t *arena); [all …]
|
D | chunk.h | 55 chunk_hooks_t chunk_hooks_get(arena_t *arena); 56 chunk_hooks_t chunk_hooks_set(arena_t *arena, 62 void *chunk_alloc_cache(arena_t *arena, chunk_hooks_t *chunk_hooks, 65 void *chunk_alloc_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks, 67 void chunk_dalloc_cache(arena_t *arena, chunk_hooks_t *chunk_hooks, 69 void chunk_dalloc_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks, 71 bool chunk_purge_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks,
|
D | tcache.h | 147 void *tcache_alloc_small_hard(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 153 void tcache_arena_associate(tcache_t *tcache, arena_t *arena); 156 void tcache_arena_dissociate(tcache_t *tcache, arena_t *arena); 158 tcache_t *tcache_create(tsd_t *tsd, arena_t *arena); 161 void tcache_stats_merge(tcache_t *tcache, arena_t *arena); 178 void *tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 180 void *tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 294 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_small() argument 308 arena = arena_choose(tsd, arena); in tcache_alloc_small() 309 if (unlikely(arena == NULL)) in tcache_alloc_small() [all …]
|
D | jemalloc_internal.h | 545 arena_t *arena_choose(tsd_t *tsd, arena_t *arena); 783 arena_choose(tsd_t *tsd, arena_t *arena) in arena_choose() argument 787 if (arena != NULL) in arena_choose() 788 return (arena); in arena_choose() 868 tcache_t *tcache, bool is_metadata, arena_t *arena, bool slow_path); 870 arena_t *arena); 873 arena_t *arena); 876 tcache_t *tcache, bool is_metadata, arena_t *arena); 878 tcache_t *tcache, arena_t *arena); 892 arena_t *arena); [all …]
|
D | extent.h | 80 void extent_node_arena_set(extent_node_t *node, arena_t *arena); 87 void extent_node_init(extent_node_t *node, arena_t *arena, void *addr, 147 extent_node_arena_set(extent_node_t *node, arena_t *arena) in extent_node_arena_set() argument 150 node->en_arena = arena; in extent_node_arena_set() 196 extent_node_init(extent_node_t *node, arena_t *arena, void *addr, size_t size, in extent_node_init() argument 200 extent_node_arena_set(node, arena); in extent_node_init()
|
D | jemalloc_internal.h.in | 170 * jemalloc can conceptually be broken into components (arena, tcache, etc.), 193 * a: arena 374 #include "jemalloc/internal/arena.h" 405 #include "jemalloc/internal/arena.h" 409 #include "jemalloc/internal/arena.h" 497 #include "jemalloc/internal/arena.h" 545 arena_t *arena_choose(tsd_t *tsd, arena_t *arena); 781 /* Choose an arena based on a per-thread value. */ 783 arena_choose(tsd_t *tsd, arena_t *arena) argument 787 if (arena != NULL) [all …]
|
D | huge.h | 12 void *huge_malloc(tsd_t *tsd, arena_t *arena, size_t usize, bool zero, 14 void *huge_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, 18 void *huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize,
|
/external/jemalloc/test/unit/ |
D | stats.c | 77 unsigned arena; in TEST_BEGIN() local 85 arena = 0; in TEST_BEGIN() 86 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeof(arena)), in TEST_BEGIN() 144 unsigned arena; in TEST_BEGIN() local 152 arena = 0; in TEST_BEGIN() 153 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeof(arena)), in TEST_BEGIN() 193 unsigned arena; in TEST_BEGIN() local 199 arena = 0; in TEST_BEGIN() 200 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeof(arena)), in TEST_BEGIN() 237 unsigned arena; in TEST_BEGIN() local [all …]
|
/external/valgrind/coregrind/ |
D | pub_core_mallocfree.h | 94 int arena; /* total space allocated from system */ member 106 extern void* VG_(arena_malloc) ( ArenaId arena, const HChar* cc, SizeT nbytes ); 107 extern void VG_(arena_free) ( ArenaId arena, void* ptr ); 108 extern void* VG_(arena_calloc) ( ArenaId arena, const HChar* cc, 110 extern void* VG_(arena_realloc) ( ArenaId arena, const HChar* cc,
|
/external/valgrind/memcheck/tests/solaris/ |
D | name_service_door.stderr.exp-amd64 | 18 Address 0x........ is 32 bytes before an unallocated block of size 4,194,048 in arena "client" 22 Address 0x........ is 24 bytes before an unallocated block of size 4,194,048 in arena "client" 26 Address 0x........ is 16 bytes before an unallocated block of size 4,194,048 in arena "client" 30 Address 0x........ is 8 bytes before an unallocated block of size 4,194,048 in arena "client" 34 Address 0x........ is 0 bytes inside an unallocated block of size 4,194,048 in arena "client" 38 Address 0x........ is 8 bytes inside an unallocated block of size 4,194,048 in arena "client" 42 Address 0x........ is 24 bytes after a block of size 16 in arena "client" 52 …Address 0x........ is 4,194,033 bytes inside an unallocated block of size 4,194,048 in arena "clie… 125 Address 0x........ is 20 bytes inside an unallocated block of size 4,193,840 in arena "client" 129 Address 0x........ is 20 bytes inside an unallocated block of size 4,193,840 in arena "client" [all …]
|
D | name_service_door.stderr.exp-x86 | 18 Address 0x........ is 8 bytes before an unallocated block of size 4,194,128 in arena "client" 22 Address 0x........ is 0 bytes inside an unallocated block of size 4,194,128 in arena "client" 26 Address 0x........ is 8 bytes inside an unallocated block of size 4,194,128 in arena "client" 30 Address 0x........ is 16 bytes inside an unallocated block of size 4,194,128 in arena "client" 34 Address 0x........ is 24 bytes inside an unallocated block of size 4,194,128 in arena "client" 38 Address 0x........ is 32 bytes inside an unallocated block of size 4,194,128 in arena "client" 42 Address 0x........ is 16 bytes before an unallocated block of size 4,194,128 in arena "client" 115 Address 0x........ is 36 bytes inside an unallocated block of size 4,193,936 in arena "client" 119 Address 0x........ is 36 bytes inside an unallocated block of size 4,193,936 in arena "client" 123 Address 0x........ is 36 bytes inside an unallocated block of size 4,193,936 in arena "client"
|
/external/valgrind/VEX/unused/ |
D | arena.h | 40 extern void *Arena_alloc (T arena, long nbytes, 42 extern void *Arena_calloc(T arena, long count, 44 extern void Arena_free (T arena);
|
/external/libchrome/crypto/third_party/nss/ |
D | secsign.cc | 54 SECStatus DerSignData(PLArenaPool *arena, in DerSignData() argument 60 return SEC_DerSignData(arena, result, input->data, input->len, key, in DerSignData() 115 rv = SECOID_SetAlgorithmID(arena, &sd.signatureAlgorithm, algo_id, 0); in DerSignData() 127 arena, result, &sd, SEC_ASN1_GET(CERT_SignedDataTemplate)); in DerSignData()
|
/external/libchrome/crypto/ |
D | signature_verifier_nss.cc | 93 PLArenaPool* arena = PORT_NewArena(DER_DEFAULT_CHUNKSIZE); in VerifyInit() local 94 if (!arena) { in VerifyInit() 105 rv = SEC_QuickDERDecodeItem(arena, &sig_alg_id, in VerifyInit() 110 PORT_FreeArena(arena, PR_TRUE); in VerifyInit() 123 PORT_FreeArena(arena, PR_TRUE); // Done with sig_alg_id. in VerifyInit()
|
D | nss_key_util.cc | 91 ScopedPLArenaPool arena(PORT_NewArena(DER_DEFAULT_CHUNKSIZE)); in ImportNSSKeyFromPrivateKeyInfo() local 92 DCHECK(arena); in ImportNSSKeyFromPrivateKeyInfo() 101 SEC_QuickDERDecodeItem(arena.get(), &der_private_key_info, in ImportNSSKeyFromPrivateKeyInfo()
|
/external/valgrind/memcheck/tests/ |
D | sbfragment.c | 22 …rintf("%10d int arena; /* non-mmapped space allocated from system */\n", mallinfo_result.arena); in stats() 88 if (mallinfo_result.arena + mallinfo_result.hblkhd > 3 * bigsize) in main() 90 (unsigned long) mallinfo_result.arena in main()
|
D | mallinfo.c | 21 printf("arena = %d\n", mi.arena); /* non-mmapped space allocated from system */ in check() 65 if (mi.arena < mi.fordblks + mi.uordblks) in check() 68 if (mi.arena/5*4 > mi.fordblks + mi.uordblks) in check()
|
/external/regex-re2/util/ |
D | arena.h | 98 re2::UnsafeArena *arena) { in new() argument 99 return reinterpret_cast<char*>(arena->GetMemory(size, 1)); in new()
|
/external/libchrome/base/trace_event/ |
D | malloc_dump_provider.cc | 67 DCHECK_GE(info.arena + info.hblkhd, info.uordblks); in OnMemoryDump() 72 total_virtual_size = info.arena + info.hblkhd; in OnMemoryDump()
|