Home
last modified time | relevance | path

Searched refs:bin_info (Results 1 – 13 of 13) sorted by relevance

/external/jemalloc_new/test/unit/
Dslab.c9 const bin_info_t *bin_info = &bin_infos[binind]; in TEST_BEGIN() local
10 extent_init(&slab, NULL, mallocx(bin_info->slab_size, in TEST_BEGIN()
11 MALLOCX_LG_ALIGN(LG_PAGE)), bin_info->slab_size, true, in TEST_BEGIN()
15 for (regind = 0; regind < bin_info->nregs; regind++) { in TEST_BEGIN()
17 (bin_info->reg_size * regind)); in TEST_BEGIN()
21 bin_info->reg_size); in TEST_BEGIN()
Djunk.c18 arena_dalloc_junk_small_intercept(void *ptr, const bin_info_t *bin_info) { in arena_dalloc_junk_small_intercept() argument
21 arena_dalloc_junk_small_orig(ptr, bin_info); in arena_dalloc_junk_small_intercept()
22 for (i = 0; i < bin_info->reg_size; i++) { in arena_dalloc_junk_small_intercept()
25 i, bin_info->reg_size); in arena_dalloc_junk_small_intercept()
/external/tensorflow/tensorflow/core/common_runtime/gpu/
Dgpu_bfc_allocator_test.cc449 const BFCAllocator::BinDebugInfo& bin_info = bin_infos[i]; in TestBinDebugInfo() local
456 EXPECT_EQ(bin_info.total_bytes_in_use, allocated_size); in TestBinDebugInfo()
457 EXPECT_EQ(bin_info.total_bytes_in_bin, allocated_size); in TestBinDebugInfo()
458 EXPECT_EQ(bin_info.total_requested_bytes_in_use, requested_size); in TestBinDebugInfo()
459 EXPECT_EQ(bin_info.total_chunks_in_use, 2); in TestBinDebugInfo()
460 EXPECT_EQ(bin_info.total_chunks_in_bin, 2); in TestBinDebugInfo()
462 EXPECT_EQ(bin_info.total_bytes_in_use, 0); in TestBinDebugInfo()
463 EXPECT_EQ(bin_info.total_requested_bytes_in_use, 0); in TestBinDebugInfo()
464 EXPECT_EQ(bin_info.total_chunks_in_use, 0); in TestBinDebugInfo()
466 EXPECT_GT(bin_info.total_bytes_in_bin, 0); in TestBinDebugInfo()
[all …]
/external/jemalloc_new/include/jemalloc/internal/
Dtcache_inlines.h171 cache_bin_info_t *bin_info; in tcache_dalloc_small() local
180 bin_info = &tcache_bin_info[binind]; in tcache_dalloc_small()
181 if (unlikely(bin->ncached == bin_info->ncached_max)) { in tcache_dalloc_small()
183 (bin_info->ncached_max >> 1)); in tcache_dalloc_small()
185 assert(bin->ncached < bin_info->ncached_max); in tcache_dalloc_small()
196 cache_bin_info_t *bin_info; in tcache_dalloc_large() local
206 bin_info = &tcache_bin_info[binind]; in tcache_dalloc_large()
207 if (unlikely(bin->ncached == bin_info->ncached_max)) { in tcache_dalloc_large()
209 (bin_info->ncached_max >> 1), tcache); in tcache_dalloc_large()
211 assert(bin->ncached < bin_info->ncached_max); in tcache_dalloc_large()
Darena_externs.h49 void arena_alloc_junk_small(void *ptr, const bin_info_t *bin_info,
/external/jemalloc/src/
Dandroid_je_iterate.c131 const arena_bin_info_t *bin_info; in je_iterate_small() local
137 bin_info = &arena_bin_info[binind]; in je_iterate_small()
139 ptr = (uintptr_t)rpages + bin_info->reg0_offset; in je_iterate_small()
141 for (regind = 0; regind < bin_info->nregs; regind++) { in je_iterate_small()
142 if (bitmap_get(run->bitmap, &bin_info->bitmap_info, regind)) { in je_iterate_small()
143 callback(ptr, bin_info->reg_size, arg); in je_iterate_small()
145 ptr += bin_info->reg_interval; in je_iterate_small()
Darena.c280 arena_run_reg_alloc(arena_run_t *run, arena_bin_info_t *bin_info) in arena_run_reg_alloc() argument
288 assert(!bitmap_full(run->bitmap, &bin_info->bitmap_info)); in arena_run_reg_alloc()
290 regind = (unsigned)bitmap_sfu(run->bitmap, &bin_info->bitmap_info); in arena_run_reg_alloc()
293 ret = (void *)((uintptr_t)rpages + (uintptr_t)bin_info->reg0_offset + in arena_run_reg_alloc()
294 (uintptr_t)(bin_info->reg_interval * regind)); in arena_run_reg_alloc()
306 arena_bin_info_t *bin_info = &arena_bin_info[binind]; in arena_run_reg_dalloc() local
307 size_t regind = arena_run_regind(run, bin_info, ptr); in arena_run_reg_dalloc()
309 assert(run->nfree < bin_info->nregs); in arena_run_reg_dalloc()
313 (uintptr_t)bin_info->reg0_offset)) % in arena_run_reg_dalloc()
314 (uintptr_t)bin_info->reg_interval == 0); in arena_run_reg_dalloc()
[all …]
/external/jemalloc_new/src/
Dandroid_je_iterate.c49 const bin_info_t* bin_info = &bin_infos[binind]; in je_iterate() local
53 size_t bin_size = bin_info->reg_size; in je_iterate()
56 for (size_t bit = (ptr - first_ptr) / bin_size; bit < bin_info->bitmap_info.nbits; bit++) { in je_iterate()
57 if (bitmap_get(slab_data->bitmap, &bin_info->bitmap_info, bit)) { in je_iterate()
Darena.c224 arena_slab_reg_alloc(extent_t *slab, const bin_info_t *bin_info) { in arena_slab_reg_alloc() argument
230 assert(!bitmap_full(slab_data->bitmap, &bin_info->bitmap_info)); in arena_slab_reg_alloc()
232 regind = bitmap_sfu(slab_data->bitmap, &bin_info->bitmap_info); in arena_slab_reg_alloc()
234 (uintptr_t)(bin_info->reg_size * regind)); in arena_slab_reg_alloc()
266 const bin_info_t *bin_info = &bin_infos[binind]; in arena_slab_reg_dalloc() local
269 assert(extent_nfree_get(slab) < bin_info->nregs); in arena_slab_reg_dalloc()
271 assert(bitmap_get(slab_data->bitmap, &bin_info->bitmap_info, regind)); in arena_slab_reg_dalloc()
273 bitmap_unset(slab_data->bitmap, &bin_info->bitmap_info, regind); in arena_slab_reg_dalloc()
1093 extent_hooks_t **r_extent_hooks, const bin_info_t *bin_info, in arena_slab_alloc_hard() argument
1104 bin_info->slab_size, 0, PAGE, true, szind, &zero, &commit); in arena_slab_alloc_hard()
[all …]
/external/tensorflow/tensorflow/core/common_runtime/
Dbfc_allocator.cc629 const BinDebugInfo& bin_info = bin_infos[bin_num]; in DumpMemoryLog() local
631 bin_info.total_chunks_in_bin - bin_info.total_chunks_in_use); in DumpMemoryLog()
634 << "): \tTotal Chunks: " << bin_info.total_chunks_in_bin in DumpMemoryLog()
635 << ", Chunks in use: " << bin_info.total_chunks_in_use << ". " in DumpMemoryLog()
636 << strings::HumanReadableNumBytes(bin_info.total_bytes_in_bin) in DumpMemoryLog()
638 << strings::HumanReadableNumBytes(bin_info.total_bytes_in_use) in DumpMemoryLog()
641 bin_info.total_requested_bytes_in_use) in DumpMemoryLog()
709 BinDebugInfo& bin_info = bin_infos[bin_num]; in get_bin_debug_info() local
710 bin_info.total_bytes_in_bin += c->size; in get_bin_debug_info()
711 bin_info.total_chunks_in_bin++; in get_bin_debug_info()
[all …]
/external/jemalloc/test/unit/
Djunk.c26 arena_dalloc_junk_small_intercept(void *ptr, arena_bin_info_t *bin_info) in arena_dalloc_junk_small_intercept() argument
30 arena_dalloc_junk_small_orig(ptr, bin_info); in arena_dalloc_junk_small_intercept()
31 for (i = 0; i < bin_info->reg_size; i++) { in arena_dalloc_junk_small_intercept()
34 i, bin_info->reg_size); in arena_dalloc_junk_small_intercept()
/external/jemalloc/include/jemalloc/internal/
Darena.h569 void arena_alloc_junk_small(void *ptr, arena_bin_info_t *bin_info,
578 void arena_dalloc_junk_small(void *ptr, arena_bin_info_t *bin_info);
697 size_t arena_run_regind(arena_run_t *run, arena_bin_info_t *bin_info,
1121 arena_bin_info_t *bin_info; in arena_ptr_small_binind_get() local
1142 bin_info = &arena_bin_info[actual_binind]; in arena_ptr_small_binind_get()
1145 (uintptr_t)bin_info->reg0_offset)) % bin_info->reg_interval in arena_ptr_small_binind_get()
1163 arena_run_regind(arena_run_t *run, arena_bin_info_t *bin_info, const void *ptr) in arena_run_regind() argument
1174 (uintptr_t)bin_info->reg0_offset); in arena_run_regind()
1181 bin_info->reg0_offset); in arena_run_regind()
1184 interval = bin_info->reg_interval; in arena_run_regind()
[all …]
/external/webp/src/enc/
Dhistogram_enc.c658 } bin_info[BIN_SIZE]; in HistogramCombineEntropyBin() local
662 bin_info[idx].first = -1; in HistogramCombineEntropyBin()
663 bin_info[idx].num_combine_failures = 0; in HistogramCombineEntropyBin()
672 first = bin_info[bin_id].first; in HistogramCombineEntropyBin()
674 bin_info[bin_id].first = idx; in HistogramCombineEntropyBin()
698 bin_info[bin_id].num_combine_failures >= max_combine_failures) { in HistogramCombineEntropyBin()
704 ++bin_info[bin_id].num_combine_failures; in HistogramCombineEntropyBin()