/external/jemalloc_new/test/unit/ |
D | slab.c | 9 const bin_info_t *bin_info = &bin_infos[binind]; in TEST_BEGIN() local 10 extent_init(&slab, NULL, mallocx(bin_info->slab_size, in TEST_BEGIN() 11 MALLOCX_LG_ALIGN(LG_PAGE)), bin_info->slab_size, true, in TEST_BEGIN() 15 for (regind = 0; regind < bin_info->nregs; regind++) { in TEST_BEGIN() 17 (bin_info->reg_size * regind)); in TEST_BEGIN() 21 bin_info->reg_size); in TEST_BEGIN()
|
D | junk.c | 18 arena_dalloc_junk_small_intercept(void *ptr, const bin_info_t *bin_info) { in arena_dalloc_junk_small_intercept() argument 21 arena_dalloc_junk_small_orig(ptr, bin_info); in arena_dalloc_junk_small_intercept() 22 for (i = 0; i < bin_info->reg_size; i++) { in arena_dalloc_junk_small_intercept() 25 i, bin_info->reg_size); in arena_dalloc_junk_small_intercept()
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_bfc_allocator_test.cc | 449 const BFCAllocator::BinDebugInfo& bin_info = bin_infos[i]; in TestBinDebugInfo() local 456 EXPECT_EQ(bin_info.total_bytes_in_use, allocated_size); in TestBinDebugInfo() 457 EXPECT_EQ(bin_info.total_bytes_in_bin, allocated_size); in TestBinDebugInfo() 458 EXPECT_EQ(bin_info.total_requested_bytes_in_use, requested_size); in TestBinDebugInfo() 459 EXPECT_EQ(bin_info.total_chunks_in_use, 2); in TestBinDebugInfo() 460 EXPECT_EQ(bin_info.total_chunks_in_bin, 2); in TestBinDebugInfo() 462 EXPECT_EQ(bin_info.total_bytes_in_use, 0); in TestBinDebugInfo() 463 EXPECT_EQ(bin_info.total_requested_bytes_in_use, 0); in TestBinDebugInfo() 464 EXPECT_EQ(bin_info.total_chunks_in_use, 0); in TestBinDebugInfo() 466 EXPECT_GT(bin_info.total_bytes_in_bin, 0); in TestBinDebugInfo() [all …]
|
/external/jemalloc_new/include/jemalloc/internal/ |
D | tcache_inlines.h | 171 cache_bin_info_t *bin_info; in tcache_dalloc_small() local 180 bin_info = &tcache_bin_info[binind]; in tcache_dalloc_small() 181 if (unlikely(bin->ncached == bin_info->ncached_max)) { in tcache_dalloc_small() 183 (bin_info->ncached_max >> 1)); in tcache_dalloc_small() 185 assert(bin->ncached < bin_info->ncached_max); in tcache_dalloc_small() 196 cache_bin_info_t *bin_info; in tcache_dalloc_large() local 206 bin_info = &tcache_bin_info[binind]; in tcache_dalloc_large() 207 if (unlikely(bin->ncached == bin_info->ncached_max)) { in tcache_dalloc_large() 209 (bin_info->ncached_max >> 1), tcache); in tcache_dalloc_large() 211 assert(bin->ncached < bin_info->ncached_max); in tcache_dalloc_large()
|
D | arena_externs.h | 49 void arena_alloc_junk_small(void *ptr, const bin_info_t *bin_info,
|
/external/jemalloc/src/ |
D | android_je_iterate.c | 131 const arena_bin_info_t *bin_info; in je_iterate_small() local 137 bin_info = &arena_bin_info[binind]; in je_iterate_small() 139 ptr = (uintptr_t)rpages + bin_info->reg0_offset; in je_iterate_small() 141 for (regind = 0; regind < bin_info->nregs; regind++) { in je_iterate_small() 142 if (bitmap_get(run->bitmap, &bin_info->bitmap_info, regind)) { in je_iterate_small() 143 callback(ptr, bin_info->reg_size, arg); in je_iterate_small() 145 ptr += bin_info->reg_interval; in je_iterate_small()
|
D | arena.c | 280 arena_run_reg_alloc(arena_run_t *run, arena_bin_info_t *bin_info) in arena_run_reg_alloc() argument 288 assert(!bitmap_full(run->bitmap, &bin_info->bitmap_info)); in arena_run_reg_alloc() 290 regind = (unsigned)bitmap_sfu(run->bitmap, &bin_info->bitmap_info); in arena_run_reg_alloc() 293 ret = (void *)((uintptr_t)rpages + (uintptr_t)bin_info->reg0_offset + in arena_run_reg_alloc() 294 (uintptr_t)(bin_info->reg_interval * regind)); in arena_run_reg_alloc() 306 arena_bin_info_t *bin_info = &arena_bin_info[binind]; in arena_run_reg_dalloc() local 307 size_t regind = arena_run_regind(run, bin_info, ptr); in arena_run_reg_dalloc() 309 assert(run->nfree < bin_info->nregs); in arena_run_reg_dalloc() 313 (uintptr_t)bin_info->reg0_offset)) % in arena_run_reg_dalloc() 314 (uintptr_t)bin_info->reg_interval == 0); in arena_run_reg_dalloc() [all …]
|
/external/jemalloc_new/src/ |
D | android_je_iterate.c | 49 const bin_info_t* bin_info = &bin_infos[binind]; in je_iterate() local 53 size_t bin_size = bin_info->reg_size; in je_iterate() 56 for (size_t bit = (ptr - first_ptr) / bin_size; bit < bin_info->bitmap_info.nbits; bit++) { in je_iterate() 57 if (bitmap_get(slab_data->bitmap, &bin_info->bitmap_info, bit)) { in je_iterate()
|
D | arena.c | 224 arena_slab_reg_alloc(extent_t *slab, const bin_info_t *bin_info) { in arena_slab_reg_alloc() argument 230 assert(!bitmap_full(slab_data->bitmap, &bin_info->bitmap_info)); in arena_slab_reg_alloc() 232 regind = bitmap_sfu(slab_data->bitmap, &bin_info->bitmap_info); in arena_slab_reg_alloc() 234 (uintptr_t)(bin_info->reg_size * regind)); in arena_slab_reg_alloc() 266 const bin_info_t *bin_info = &bin_infos[binind]; in arena_slab_reg_dalloc() local 269 assert(extent_nfree_get(slab) < bin_info->nregs); in arena_slab_reg_dalloc() 271 assert(bitmap_get(slab_data->bitmap, &bin_info->bitmap_info, regind)); in arena_slab_reg_dalloc() 273 bitmap_unset(slab_data->bitmap, &bin_info->bitmap_info, regind); in arena_slab_reg_dalloc() 1093 extent_hooks_t **r_extent_hooks, const bin_info_t *bin_info, in arena_slab_alloc_hard() argument 1104 bin_info->slab_size, 0, PAGE, true, szind, &zero, &commit); in arena_slab_alloc_hard() [all …]
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | bfc_allocator.cc | 629 const BinDebugInfo& bin_info = bin_infos[bin_num]; in DumpMemoryLog() local 631 bin_info.total_chunks_in_bin - bin_info.total_chunks_in_use); in DumpMemoryLog() 634 << "): \tTotal Chunks: " << bin_info.total_chunks_in_bin in DumpMemoryLog() 635 << ", Chunks in use: " << bin_info.total_chunks_in_use << ". " in DumpMemoryLog() 636 << strings::HumanReadableNumBytes(bin_info.total_bytes_in_bin) in DumpMemoryLog() 638 << strings::HumanReadableNumBytes(bin_info.total_bytes_in_use) in DumpMemoryLog() 641 bin_info.total_requested_bytes_in_use) in DumpMemoryLog() 709 BinDebugInfo& bin_info = bin_infos[bin_num]; in get_bin_debug_info() local 710 bin_info.total_bytes_in_bin += c->size; in get_bin_debug_info() 711 bin_info.total_chunks_in_bin++; in get_bin_debug_info() [all …]
|
/external/jemalloc/test/unit/ |
D | junk.c | 26 arena_dalloc_junk_small_intercept(void *ptr, arena_bin_info_t *bin_info) in arena_dalloc_junk_small_intercept() argument 30 arena_dalloc_junk_small_orig(ptr, bin_info); in arena_dalloc_junk_small_intercept() 31 for (i = 0; i < bin_info->reg_size; i++) { in arena_dalloc_junk_small_intercept() 34 i, bin_info->reg_size); in arena_dalloc_junk_small_intercept()
|
/external/jemalloc/include/jemalloc/internal/ |
D | arena.h | 569 void arena_alloc_junk_small(void *ptr, arena_bin_info_t *bin_info, 578 void arena_dalloc_junk_small(void *ptr, arena_bin_info_t *bin_info); 697 size_t arena_run_regind(arena_run_t *run, arena_bin_info_t *bin_info, 1121 arena_bin_info_t *bin_info; in arena_ptr_small_binind_get() local 1142 bin_info = &arena_bin_info[actual_binind]; in arena_ptr_small_binind_get() 1145 (uintptr_t)bin_info->reg0_offset)) % bin_info->reg_interval in arena_ptr_small_binind_get() 1163 arena_run_regind(arena_run_t *run, arena_bin_info_t *bin_info, const void *ptr) in arena_run_regind() argument 1174 (uintptr_t)bin_info->reg0_offset); in arena_run_regind() 1181 bin_info->reg0_offset); in arena_run_regind() 1184 interval = bin_info->reg_interval; in arena_run_regind() [all …]
|
/external/webp/src/enc/ |
D | histogram_enc.c | 658 } bin_info[BIN_SIZE]; in HistogramCombineEntropyBin() local 662 bin_info[idx].first = -1; in HistogramCombineEntropyBin() 663 bin_info[idx].num_combine_failures = 0; in HistogramCombineEntropyBin() 672 first = bin_info[bin_id].first; in HistogramCombineEntropyBin() 674 bin_info[bin_id].first = idx; in HistogramCombineEntropyBin() 698 bin_info[bin_id].num_combine_failures >= max_combine_failures) { in HistogramCombineEntropyBin() 704 ++bin_info[bin_id].num_combine_failures; in HistogramCombineEntropyBin()
|