/external/jemalloc/src/ |
D | android_je_mallinfo.c | 26 if (arenas[i] != NULL) { in je_mallinfo() 27 malloc_mutex_lock(TSDN_NULL, &arenas[i]->lock); in je_mallinfo() 28 mi.hblkhd += arenas[i]->stats.mapped; in je_mallinfo() 29 mi.uordblks += arenas[i]->stats.allocated_large; in je_mallinfo() 30 mi.uordblks += arenas[i]->stats.allocated_huge; in je_mallinfo() 31 malloc_mutex_unlock(TSDN_NULL, &arenas[i]->lock); in je_mallinfo() 34 arena_bin_t* bin = &arenas[i]->bins[j]; in je_mallinfo() 62 if (arenas[aidx] != NULL) { in __mallinfo_arena_info() 63 malloc_mutex_lock(TSDN_NULL, &arenas[aidx]->lock); in __mallinfo_arena_info() 64 mi.hblkhd = arenas[aidx]->stats.mapped; in __mallinfo_arena_info() [all …]
|
D | ctl.c | 502 {NAME("arenas"), CHILD(named, arenas)}, 661 ctl_arena_stats_t *astats = &ctl_stats.arenas[i]; in ctl_arena_refresh() 662 ctl_arena_stats_t *sstats = &ctl_stats.arenas[ctl_stats.narenas]; in ctl_arena_refresh() 686 memcpy(astats, ctl_stats.arenas, (ctl_stats.narenas + 1) * in ctl_grow() 703 a0dalloc(ctl_stats.arenas); in ctl_grow() 704 ctl_stats.arenas = astats; in ctl_grow() 720 ctl_arena_clear(&ctl_stats.arenas[ctl_stats.narenas]); in ctl_refresh() 728 ctl_stats.arenas[i].initialized = initialized; in ctl_refresh() 738 ctl_stats.arenas[ctl_stats.narenas].allocated_small + in ctl_refresh() 739 ctl_stats.arenas[ctl_stats.narenas].astats.allocated_large + in ctl_refresh() [all …]
|
D | jemalloc.c | 64 arena_t **arenas; variable 412 atomic_write_p((void **)&arenas[ind], arena); in arena_set() 1384 arenas = &a0; 1385 memset(arenas, 0, sizeof(arena_t *) * narenas_auto); 1471 arenas = (arena_t **)base_alloc(tsdn, sizeof(arena_t *) * 1473 if (arenas == NULL)
|
D | stats.c | 731 arenas.lg_dirty_mult, ",") in stats_general_print() 734 OPT_WRITE_SSIZE_T_MUTABLE(decay_time, arenas.decay_time, ",") in stats_general_print()
|
/external/python/cpython2/Objects/ |
D | obmalloc.c | 521 static struct arena_object* arenas = NULL; variable 580 if (numarenas > PY_SIZE_MAX / sizeof(*arenas)) in new_arena() 583 nbytes = numarenas * sizeof(*arenas); in new_arena() 584 arenaobj = (struct arena_object *)realloc(arenas, nbytes); in new_arena() 587 arenas = arenaobj; in new_arena() 600 arenas[i].address = 0; /* mark as unassociated */ in new_arena() 601 arenas[i].nextarena = i < numarenas - 1 ? in new_arena() 602 &arenas[i+1] : NULL; in new_arena() 606 unused_arena_objects = &arenas[maxarenas]; in new_arena() 739 (uptr)(P) - arenas[arenaindex_temp].address < (uptr)ARENA_SIZE && \ [all …]
|
/external/jemalloc/ |
D | ChangeLog | 60 - Fix stats.arenas.<i>.nthreads accounting. (@interwq) 103 - Add the stats.retained and stats.arenas.<i>.retained statistics. (@jasone) 161 + arenas.decay_time 162 + stats.arenas.<i>.decay_time 198 - Refactor arenas array. In addition to fixing a fork-related deadlock, this 212 - Fix stats.arenas.<i>.{dss,lg_dirty_mult,decay_time,pactive,pdirty} for 344 - Refactor huge allocation to be managed by arenas, so that arenas now 348 + The "stats.arenas.<i>.huge.allocated", "stats.arenas.<i>.huge.nmalloc", 349 "stats.arenas.<i>.huge.ndalloc", and "stats.arenas.<i>.huge.nrequests" 351 + The "arenas.nhchunks", "arenas.hchunk.<i>.size", [all …]
|
D | Android.bp | 28 // The total number of arenas will be less than or equal to this number. 29 // The number of arenas will be calculated as 2 * the number of cpus
|
/external/jemalloc/include/jemalloc/internal/ |
D | ctl.h | 66 ctl_arena_stats_t *arenas; /* (narenas + 1) elements. */ member
|
D | jemalloc_internal.h.in | 456 /* Number of arenas used for automatic multiplexing of threads and arenas. */ 461 * arenas array are necessarily used; arenas are created lazily as needed. 463 extern arena_t **arenas; 954 ret = arenas[ind]; 956 ret = atomic_read_p((void *)&arenas[ind]);
|
D | jemalloc_internal.h | 463 extern arena_t **arenas; 954 ret = arenas[ind]; in arena_get() 956 ret = atomic_read_p((void *)&arenas[ind]); in arena_get()
|
D | private_symbols.txt | 119 arenas
|
D | private_namespace.h | 119 #define arenas JEMALLOC_N(arenas) macro
|
D | jemalloc_internal_defs.h.in | 209 /* TLS is used to map arenas and magazine caches to threads. */
|
D | private_unnamespace.h | 119 #undef arenas
|
/external/python/cpython2/Misc/ |
D | README.valgrind | 62 The memory pymalloc manages itself is in one or more "arenas",
|
D | NEWS | 2065 - Issue #20494: Ensure that free()d memory arenas are really released on POSIX 10476 in order to allow pymalloc to free more arenas. Python may give back
|
D | HISTORY | 1079 again. Prior to Python 2.5, arenas (256KB chunks of memory) were never
|
/external/protobuf/src/google/protobuf/ |
D | descriptor.proto | 368 // Enables the use of arenas for the proto messages in this file. This applies
|
/external/python/cpython2/Doc/whatsnew/ |
D | 2.5.rst | 2136 was applied. Python 2.4 allocated small objects in 256K-sized arenas, but never 2137 freed arenas. With this patch, Python will free arenas when they're empty. The
|
/external/protobuf/ |
D | CHANGES.txt | 566 message classes work with arenas. This does not change the existing API
|
/external/protobuf/util/ |
D | CHANGES.txt | 566 message classes work with arenas. This does not change the existing API
|
/external/valgrind/ |
D | NEWS | 1585 308711 give more info about aspacemgr and arenas in out_of_memory
|
/external/jline/src/src/test/resources/jline/example/ |
D | english.gz |
|