Home
last modified time | relevance | path

Searched refs:ZU (Results 1 – 25 of 244) sorted by relevance

12345678910

/external/jemalloc_new/include/jemalloc/internal/
Dsz.h64 size_t delta_inverse_mask = ZU(-1) << lg_delta; in sz_psz2ind()
66 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in sz_psz2ind()
80 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in sz_pind2sz_compute()
83 size_t grp_size = ((ZU(1) << (LG_PAGE + in sz_pind2sz_compute()
117 size_t delta = ZU(1) << lg_delta; in sz_psz2u()
130 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { in sz_size2index_compute()
145 size_t delta_inverse_mask = ZU(-1) << lg_delta; in sz_size2index_compute()
147 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in sz_size2index_compute()
177 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); in sz_index2size_compute()
183 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - in sz_index2size_compute()
[all …]
Dbitmap.h19 #define BITMAP_MAXBITS (ZU(1) << LG_BITMAP_MAXBITS)
202 return !(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_get()
216 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set()
217 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set()
229 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set()
230 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set()
268 size_t sib_base = bit + (ZU(1) << lg_bits_per_group); in bitmap_ffu()
343 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) == 0); in bitmap_unset()
344 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_unset()
357 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) in bitmap_unset()
[all …]
Djemalloc_internal_macros.h15 #define ZU(z) ((size_t)z) macro
20 #define KZU(z) ZU(z##ULL)
Drtree.h133 unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3); in rtree_leafkey()
137 uintptr_t mask = ~((ZU(1) << maskbits) - 1); in rtree_leafkey()
143 unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3); in rtree_cache_direct_map()
152 unsigned ptrbits = ZU(1) << (LG_SIZEOF_PTR+3); in rtree_subkey()
156 uintptr_t mask = (ZU(1) << maskbits) - 1; in rtree_subkey()
/external/jemalloc/include/jemalloc/internal/
Dbitmap.h6 #define BITMAP_MAXBITS (ZU(1) << LG_BITMAP_MAXBITS)
15 #define BITMAP_GROUP_NBITS (ZU(1) << LG_BITMAP_GROUP_NBITS)
162 return (!(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK)))); in bitmap_get()
177 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set()
178 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set()
190 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set()
191 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set()
246 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) == 0); in bitmap_unset()
247 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_unset()
260 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) in bitmap_unset()
[all …]
Djemalloc_internal.h214 (ZU(1) << (flags & MALLOCX_LG_ALIGN_MASK))
603 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in psz2ind()
616 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in pind2sz_compute()
619 size_t grp_size = ((ZU(1) << (LG_PAGE + in pind2sz_compute()
657 size_t delta = ZU(1) << lg_delta; in psz2u()
671 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { in size2index_compute()
688 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in size2index_compute()
723 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); in index2size_compute()
728 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - in index2size_compute()
732 size_t grp_size = ((ZU(1) << (LG_QUANTUM + in index2size_compute()
[all …]
Djemalloc_internal_macros.h41 #define ZU(z) ((size_t)z) macro
46 #define KZU(z) ZU(z##ULL)
/external/jemalloc/test/unit/
Dprng.c84 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
86 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
91 rb = prng_lg_range_zu(&sb, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
96 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
97 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
102 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
103 for (lg_range = (ZU(1) << (3 + LG_SIZEOF_PTR)) - 1; lg_range > 0; in test_prng_lg_range_zu()
109 assert_zu_eq(rb, (ra >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - in test_prng_lg_range_zu()
Dsize_classes.c148 assert_u_eq(size2index(ZU(PTRDIFF_MAX)+1), NSIZES, in TEST_BEGIN()
155 assert_zu_eq(s2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN()
162 assert_u_eq(psz2ind(ZU(PTRDIFF_MAX)+1), NPSIZES, in TEST_BEGIN()
169 assert_zu_eq(psz2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN()
Dckh.c38 "ckh_count() should return %zu, but it returned %zu", ZU(0), in TEST_BEGIN()
110 #define NITEMS ZU(1000) in TEST_BEGIN()
200 ZU(0), ckh_count(&ckh)); in TEST_BEGIN()
/external/jemalloc/test/integration/
Drallocx.c57 #define MAXSZ ZU(12 * 1024 * 1024) in TEST_BEGIN()
163 #define MAX_ALIGN (ZU(1) << 25) in TEST_BEGIN()
165 align = ZU(1); in TEST_BEGIN()
190 #define MAX_VALIDATE (ZU(1) << 22) in TEST_BEGIN()
201 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN()
235 assert_ptr_null(rallocx(p, ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN()
236 "Expected OOM for rallocx(p, size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
241 assert_ptr_null(rallocx(p, 1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN()
243 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
Dmallocx.c75 assert_ptr_null(mallocx(ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN()
76 "Expected OOM for mallocx(size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
81 assert_ptr_null(mallocx(1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN()
83 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
/external/jemalloc_new/test/integration/
Drallocx.c50 #define MAXSZ ZU(12 * 1024 * 1024) in TEST_BEGIN()
153 #define MAX_ALIGN (ZU(1) << 25) in TEST_BEGIN()
155 align = ZU(1); in TEST_BEGIN()
179 #define MAX_VALIDATE (ZU(1) << 22) in TEST_BEGIN()
190 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN()
223 assert_ptr_null(rallocx(p, ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN()
224 "Expected OOM for rallocx(p, size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
229 assert_ptr_null(rallocx(p, 1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN()
231 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
Dmallocx.c62 assert_ptr_null(mallocx(ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN()
63 "Expected OOM for mallocx(size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
68 assert_ptr_null(mallocx(1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN()
70 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
/external/jemalloc_new/test/unit/
Dprng.c83 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
85 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
90 rb = prng_lg_range_zu(&sb, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
95 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
96 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
101 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu()
102 for (lg_range = (ZU(1) << (3 + LG_SIZEOF_PTR)) - 1; lg_range > 0; in test_prng_lg_range_zu()
108 assert_zu_eq(rb, (ra >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - in test_prng_lg_range_zu()
Dsize_classes.c147 assert_u_eq(sz_size2index(ZU(PTRDIFF_MAX)+1), NSIZES, in TEST_BEGIN()
154 assert_zu_eq(sz_s2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN()
161 assert_u_eq(sz_psz2ind(ZU(PTRDIFF_MAX)+1), NPSIZES, in TEST_BEGIN()
169 assert_zu_eq(sz_psz2u(ZU(PTRDIFF_MAX)+1), max_psz, in TEST_BEGIN()
Dckh.c36 "ckh_count() should return %zu, but it returned %zu", ZU(0), in TEST_BEGIN()
107 #define NITEMS ZU(1000) in TEST_BEGIN()
199 ZU(0), ckh_count(&ckh)); in TEST_BEGIN()
/external/jemalloc/src/
Dckh.c58 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_bucket_search()
80 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch()
86 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch()
104 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_try_bucket_insert()
106 ((i + offset) & ((ZU(1) << LG_CKH_BUCKET_CELLS) - 1))]; in ckh_try_bucket_insert()
161 tbucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_evict_reloc_insert()
163 tbucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) in ckh_evict_reloc_insert()
205 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert()
210 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert()
382 (ZU(1) << lg_mincells) < mincells; in ckh_new()
[all …]
Dquarantine.c28 size = offsetof(quarantine_t, objs) + ((ZU(1) << lg_maxobjs) * in quarantine_init()
74 if (quarantine->first + quarantine->curobjs <= (ZU(1) << in quarantine_grow()
81 size_t ncopy_a = (ZU(1) << quarantine->lg_maxobjs) - in quarantine_grow()
104 quarantine->first = (quarantine->first + 1) & ((ZU(1) << in quarantine_drain_one()
139 if (quarantine->curobjs == (ZU(1) << quarantine->lg_maxobjs)) in quarantine()
142 assert(quarantine->curobjs < (ZU(1) << quarantine->lg_maxobjs)); in quarantine()
146 ((ZU(1) << quarantine->lg_maxobjs) - 1); in quarantine()
Drtree.c18 assert(RTREE_HEIGHT_MAX == ((ZU(1) << (LG_SIZEOF_PTR+3)) / in rtree_new()
71 nchildren = ZU(1) << rtree->levels[level].bits; in rtree_delete_subtree()
111 node = rtree->alloc(ZU(1) << rtree->levels[level].bits); in rtree_node_init()
Dbitmap.c15 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
79 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
/external/jemalloc_new/src/
Dckh.c67 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_bucket_search()
89 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch()
96 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch()
113 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_try_bucket_insert()
115 ((i + offset) & ((ZU(1) << LG_CKH_BUCKET_CELLS) - 1))]; in ckh_try_bucket_insert()
169 tbucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_evict_reloc_insert()
171 tbucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) in ckh_evict_reloc_insert()
213 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert()
219 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert()
389 (ZU(1) << lg_mincells) < mincells; in ckh_new()
[all …]
Dsz.c7 (((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta))),
21 ((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta)),
Dbitmap.c17 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
88 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
/external/ImageMagick/PerlMagick/t/reference/filter/
DLevel.miff340 #FHg4A_5��XS:XS:XS:ZU<ZU<\S:_PDbWAXZ2_nf���������������������������������������������������������…
347 ��&DX72F/��WR9XS:ZU<ZU<ZU<\S:bPD_XAU_5Xqd����������������������������������������������������…

12345678910