/external/jemalloc_new/include/jemalloc/internal/ |
D | sz.h | 64 size_t delta_inverse_mask = ZU(-1) << lg_delta; in sz_psz2ind() 66 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in sz_psz2ind() 80 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in sz_pind2sz_compute() 83 size_t grp_size = ((ZU(1) << (LG_PAGE + in sz_pind2sz_compute() 117 size_t delta = ZU(1) << lg_delta; in sz_psz2u() 130 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { in sz_size2index_compute() 145 size_t delta_inverse_mask = ZU(-1) << lg_delta; in sz_size2index_compute() 147 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in sz_size2index_compute() 177 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); in sz_index2size_compute() 183 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - in sz_index2size_compute() [all …]
|
D | bitmap.h | 19 #define BITMAP_MAXBITS (ZU(1) << LG_BITMAP_MAXBITS) 202 return !(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_get() 216 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set() 217 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set() 229 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set() 230 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set() 268 size_t sib_base = bit + (ZU(1) << lg_bits_per_group); in bitmap_ffu() 343 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) == 0); in bitmap_unset() 344 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_unset() 357 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) in bitmap_unset() [all …]
|
D | jemalloc_internal_macros.h | 15 #define ZU(z) ((size_t)z) macro 20 #define KZU(z) ZU(z##ULL)
|
D | rtree.h | 133 unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3); in rtree_leafkey() 137 uintptr_t mask = ~((ZU(1) << maskbits) - 1); in rtree_leafkey() 143 unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3); in rtree_cache_direct_map() 152 unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3); in rtree_subkey() 156 uintptr_t mask = (ZU(1) << maskbits) - 1; in rtree_subkey()
|
/external/jemalloc/include/jemalloc/internal/ |
D | bitmap.h | 6 #define BITMAP_MAXBITS (ZU(1) << LG_BITMAP_MAXBITS) 15 #define BITMAP_GROUP_NBITS (ZU(1) << LG_BITMAP_GROUP_NBITS) 162 return (!(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK)))); in bitmap_get() 177 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set() 178 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set() 190 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set() 191 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set() 246 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) == 0); in bitmap_unset() 247 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_unset() 260 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) in bitmap_unset() [all …]
|
D | jemalloc_internal.h | 214 (ZU(1) << (flags & MALLOCX_LG_ALIGN_MASK)) 603 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in psz2ind() 616 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in pind2sz_compute() 619 size_t grp_size = ((ZU(1) << (LG_PAGE + in pind2sz_compute() 657 size_t delta = ZU(1) << lg_delta; in psz2u() 671 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { in size2index_compute() 688 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in size2index_compute() 723 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); in index2size_compute() 728 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - in index2size_compute() 732 size_t grp_size = ((ZU(1) << (LG_QUANTUM + in index2size_compute() [all …]
|
D | jemalloc_internal_macros.h | 41 #define ZU(z) ((size_t)z) macro 46 #define KZU(z) ZU(z##ULL)
|
/external/jemalloc/test/unit/ |
D | prng.c | 84 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 86 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 91 rb = prng_lg_range_zu(&sb, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 96 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 97 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 102 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 103 for (lg_range = (ZU(1) << (3 + LG_SIZEOF_PTR)) - 1; lg_range > 0; in test_prng_lg_range_zu() 109 assert_zu_eq(rb, (ra >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - in test_prng_lg_range_zu()
|
D | size_classes.c | 148 assert_u_eq(size2index(ZU(PTRDIFF_MAX)+1), NSIZES, in TEST_BEGIN() 155 assert_zu_eq(s2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN() 162 assert_u_eq(psz2ind(ZU(PTRDIFF_MAX)+1), NPSIZES, in TEST_BEGIN() 169 assert_zu_eq(psz2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN()
|
D | ckh.c | 38 "ckh_count() should return %zu, but it returned %zu", ZU(0), in TEST_BEGIN() 110 #define NITEMS ZU(1000) in TEST_BEGIN() 200 ZU(0), ckh_count(&ckh)); in TEST_BEGIN()
|
/external/jemalloc/test/integration/ |
D | rallocx.c | 57 #define MAXSZ ZU(12 * 1024 * 1024) in TEST_BEGIN() 163 #define MAX_ALIGN (ZU(1) << 25) in TEST_BEGIN() 165 align = ZU(1); in TEST_BEGIN() 190 #define MAX_VALIDATE (ZU(1) << 22) in TEST_BEGIN() 201 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN() 235 assert_ptr_null(rallocx(p, ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN() 236 "Expected OOM for rallocx(p, size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN() 241 assert_ptr_null(rallocx(p, 1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN() 243 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
|
D | mallocx.c | 75 assert_ptr_null(mallocx(ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN() 76 "Expected OOM for mallocx(size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN() 81 assert_ptr_null(mallocx(1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN() 83 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
|
/external/jemalloc_new/test/integration/ |
D | rallocx.c | 50 #define MAXSZ ZU(12 * 1024 * 1024) in TEST_BEGIN() 153 #define MAX_ALIGN (ZU(1) << 25) in TEST_BEGIN() 155 align = ZU(1); in TEST_BEGIN() 179 #define MAX_VALIDATE (ZU(1) << 22) in TEST_BEGIN() 190 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN() 223 assert_ptr_null(rallocx(p, ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN() 224 "Expected OOM for rallocx(p, size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN() 229 assert_ptr_null(rallocx(p, 1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN() 231 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
|
D | mallocx.c | 62 assert_ptr_null(mallocx(ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN() 63 "Expected OOM for mallocx(size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN() 68 assert_ptr_null(mallocx(1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN() 70 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
|
/external/jemalloc_new/test/unit/ |
D | prng.c | 83 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 85 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 90 rb = prng_lg_range_zu(&sb, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 95 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 96 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 101 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 102 for (lg_range = (ZU(1) << (3 + LG_SIZEOF_PTR)) - 1; lg_range > 0; in test_prng_lg_range_zu() 108 assert_zu_eq(rb, (ra >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - in test_prng_lg_range_zu()
|
D | size_classes.c | 147 assert_u_eq(sz_size2index(ZU(PTRDIFF_MAX)+1), NSIZES, in TEST_BEGIN() 154 assert_zu_eq(sz_s2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN() 161 assert_u_eq(sz_psz2ind(ZU(PTRDIFF_MAX)+1), NPSIZES, in TEST_BEGIN() 169 assert_zu_eq(sz_psz2u(ZU(PTRDIFF_MAX)+1), max_psz, in TEST_BEGIN()
|
D | ckh.c | 36 "ckh_count() should return %zu, but it returned %zu", ZU(0), in TEST_BEGIN() 107 #define NITEMS ZU(1000) in TEST_BEGIN() 199 ZU(0), ckh_count(&ckh)); in TEST_BEGIN()
|
/external/jemalloc/src/ |
D | ckh.c | 58 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_bucket_search() 80 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch() 86 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch() 104 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_try_bucket_insert() 106 ((i + offset) & ((ZU(1) << LG_CKH_BUCKET_CELLS) - 1))]; in ckh_try_bucket_insert() 161 tbucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_evict_reloc_insert() 163 tbucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) in ckh_evict_reloc_insert() 205 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert() 210 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert() 382 (ZU(1) << lg_mincells) < mincells; in ckh_new() [all …]
|
D | quarantine.c | 28 size = offsetof(quarantine_t, objs) + ((ZU(1) << lg_maxobjs) * in quarantine_init() 74 if (quarantine->first + quarantine->curobjs <= (ZU(1) << in quarantine_grow() 81 size_t ncopy_a = (ZU(1) << quarantine->lg_maxobjs) - in quarantine_grow() 104 quarantine->first = (quarantine->first + 1) & ((ZU(1) << in quarantine_drain_one() 139 if (quarantine->curobjs == (ZU(1) << quarantine->lg_maxobjs)) in quarantine() 142 assert(quarantine->curobjs < (ZU(1) << quarantine->lg_maxobjs)); in quarantine() 146 ((ZU(1) << quarantine->lg_maxobjs) - 1); in quarantine()
|
D | rtree.c | 18 assert(RTREE_HEIGHT_MAX == ((ZU(1) << (LG_SIZEOF_PTR+3)) / in rtree_new() 71 nchildren = ZU(1) << rtree->levels[level].bits; in rtree_delete_subtree() 111 node = rtree->alloc(ZU(1) << rtree->levels[level].bits); in rtree_node_init()
|
D | bitmap.c | 15 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init() 79 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
|
/external/jemalloc_new/src/ |
D | ckh.c | 67 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_bucket_search() 89 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch() 96 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch() 113 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_try_bucket_insert() 115 ((i + offset) & ((ZU(1) << LG_CKH_BUCKET_CELLS) - 1))]; in ckh_try_bucket_insert() 169 tbucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_evict_reloc_insert() 171 tbucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) in ckh_evict_reloc_insert() 213 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert() 219 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert() 389 (ZU(1) << lg_mincells) < mincells; in ckh_new() [all …]
|
D | sz.c | 7 (((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta))), 21 ((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta)),
|
D | bitmap.c | 17 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init() 88 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
|
/external/ImageMagick/PerlMagick/t/reference/filter/ |
D | Level.miff | 340 #FHg4A_5XS:XS:XS:ZU<ZU<\S:_PDbWAXZ2_nf���������������������������������������������������������… 347 &DX72F/WR9XS:ZU<ZU<ZU<\S:bPD_XAU_5Xqd����������������������������������������������������…
|