/external/jemalloc/include/jemalloc/internal/ |
jemalloc_internal_macros.h | 41 #define ZU(z) ((size_t)z) 46 #define KZU(z) ZU(z##ULL)
|
bitmap.h | 6 #define BITMAP_MAXBITS (ZU(1) << LG_BITMAP_MAXBITS) 15 #define BITMAP_GROUP_NBITS (ZU(1) << LG_BITMAP_GROUP_NBITS) 162 return (!(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK)))); 177 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); 178 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); 190 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); 191 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); 246 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) == 0); 247 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); 260 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK)) [all...] |
quarantine.h | 8 #define JEMALLOC_VALGRIND_QUARANTINE_DEFAULT (ZU(1) << 24)
|
jemalloc_internal.h | 214 (ZU(1) << (flags & MALLOCX_LG_ALIGN_MASK)) 603 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); 616 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); 619 size_t grp_size = ((ZU(1) << (LG_PAGE + 657 size_t delta = ZU(1) << lg_delta; 671 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { 688 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); 723 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); 728 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 732 size_t grp_size = ((ZU(1) << (LG_QUANTUM [all...] |
prng.h | 130 assert(lg_range <= ZU(1) << (3 + LG_SIZEOF_PTR)); 143 ret = state1 >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - lg_range);
|
rtree.h | 151 return ((key >> ((ZU(1) << (LG_SIZEOF_PTR+3)) - 152 rtree->levels[level].cumbits)) & ((ZU(1) <<
|
arena.h | 4 #define LARGE_MINCLASS (ZU(1) << LG_LARGE_MINCLASS) [all...] |
/external/jemalloc/test/integration/ |
rallocx.c | 37 0, "Unexpected mallctlbymib([\"%s\", %zu], ...) failure", cmd, ind); 57 #define MAXSZ ZU(12 * 1024 * 1024) 67 "Unexpected rallocx() error for size=%zu-->%zu", 71 "Expected size to be at least: %zu", szs[j-1]+1); 78 "Unexpected rallocx() error for size=%zu-->%zu", 82 "Expected size=%zu, got size=%zu", szs[j-1], tsz); 104 test_fail("Allocation at %p (len=%zu) contains %#x [all...] |
mallocx.c | 41 0, "Unexpected mallctlbymib([\"%s\", %zu], ...) failure", cmd, ind); 75 assert_ptr_null(mallocx(ZU(PTRDIFF_MAX)+1, 0), 76 "Expected OOM for mallocx(size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); 81 assert_ptr_null(mallocx(1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), 83 ZU(PTRDIFF_MAX)+1); 106 "Expected OOM during series of calls to mallocx(size=%zu, 0)", 188 "nallocx() error for alignment=%zu, " 189 "size=%zu (%#zx)", alignment, sz, sz); 193 "mallocx() error for alignment=%zu, " 194 "size=%zu (%#zx)", alignment, sz, sz) [all...] |
/external/jemalloc/test/unit/ |
prng.c | 84 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 86 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 91 rb = prng_lg_range_zu(&sb, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 96 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 97 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 102 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); 103 for (lg_range = (ZU(1) << (3 + LG_SIZEOF_PTR)) - 1; lg_range > 0; 109 assert_zu_eq(rb, (ra >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) -
|
size_classes.c | 39 "size_class=%zu (%#zx)", index, size_class, size_class); 42 "size_class=%zu (%#zx)", index, size_class, size_class); 46 " size_class=%zu --> index=%u --> size_class=%zu", index, 51 " size_class=%zu --> index=%u --> size_class=%zu", index, 96 "size_class=%zu (%#zx)", pind, size_class, size_class); 99 "size_class=%zu (%#zx)", pind, size_class, size_class); 103 " size_class=%zu --> pind=%u --> size_class=%zu", pind [all...] |
quarantine.c | 24 #define SZ ZU(256) 32 "SZ=%zu does not precisely equal a size class", SZ); 51 "i=%zu, j=%zu", i, j);
|
zero.c | 23 "Previously allocated byte %zu/%zu is corrupted", 24 ZU(0), sz_prev); 26 "Previously allocated byte %zu/%zu is corrupted", 32 "Newly allocated byte %zu/%zu isn't zero-filled",
|
ckh.c | 38 "ckh_count() should return %zu, but it returned %zu", ZU(0), 45 "ckh_count() should return %zu, but it returned %zu", i+1, 67 assert_ptr_eq((void *)ks, (void *)k.s, "Key mismatch, i=%zu", 69 assert_ptr_eq((void *)vs, (void *)v.s, "Value mismatch, i=%zu", 93 assert_ptr_eq((void *)ks, (void *)k.s, "Key mismatch, i=%zu", 95 assert_ptr_eq((void *)vs, (void *)v.s, "Value mismatch, i=%zu", 99 "ckh_count() should return %zu, but it returned %zu" [all...] |
junk.c | 33 "Missing junk fill for byte %zu/%zu of deallocated region", 48 "Missing junk fill for byte %zu/%zu of deallocated region", 92 "Previously allocated byte %zu/%zu is corrupted", 93 ZU(0), sz_prev); 95 "Previously allocated byte %zu/%zu is corrupted", 102 "Newly allocated byte %zu/%zu isn't [all...] |
pack.c | 15 #define SZ (ZU(1) << (LG_PAGE - 2)) 149 "Unexpected mallocx(%zu, MALLOCX_ARENA(%u) |" 150 " MALLOCX_TCACHE_NONE) failure, run=%zu, reg=%zu", 190 "Unexpected refill discrepancy, run=%zu, reg=%zu\n",
|
/external/jemalloc/src/ |
quarantine.c | 28 size = offsetof(quarantine_t, objs) + ((ZU(1) << lg_maxobjs) * 74 if (quarantine->first + quarantine->curobjs <= (ZU(1) << 81 size_t ncopy_a = (ZU(1) << quarantine->lg_maxobjs) - 104 quarantine->first = (quarantine->first + 1) & ((ZU(1) << 139 if (quarantine->curobjs == (ZU(1) << quarantine->lg_maxobjs)) 142 assert(quarantine->curobjs < (ZU(1) << quarantine->lg_maxobjs)); 146 ((ZU(1) << quarantine->lg_maxobjs) - 1);
|
ckh.c | 58 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { 80 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); 86 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); 104 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { 106 ((i + offset) & ((ZU(1) << LG_CKH_BUCKET_CELLS) - 1))]; 161 tbucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); 163 tbucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) 205 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); 210 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); 382 (ZU(1) << lg_mincells) < mincells [all...] |
rtree.c | 18 assert(RTREE_HEIGHT_MAX == ((ZU(1) << (LG_SIZEOF_PTR+3)) / 71 nchildren = ZU(1) << rtree->levels[level].bits; 111 node = rtree->alloc(ZU(1) << rtree->levels[level].bits);
|
bitmap.c | 15 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); 79 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS));
|
tcache.c | 559 if (opt_lg_tcache_max < 0 || (ZU(1) << opt_lg_tcache_max) < SMALL_MAXCLASS) 561 else if ((ZU(1) << opt_lg_tcache_max) > large_maxclass) 564 tcache_maxclass = (ZU(1) << opt_lg_tcache_max);
|
jemalloc.c | 42 size_t opt_quarantine = ZU(0); 95 (((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta))), 108 ((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta)), [all...] |
chunk.c | 783 chunksize = (ZU(1) << opt_lg_chunk); 790 if (rtree_new(&chunks_rtree, (unsigned)((ZU(1) << (LG_SIZEOF_PTR+3)) -
|
/toolchain/binutils/binutils-2.25/gas/testsuite/gas/aarch64/ |
fp_cvt_int.s | 87 do_fcvt ZU, \fbits, \reg 99 do_fcvt ZU, \fbits, \reg
|
/frameworks/base/media/tests/contents/media_api/music/ |
test_amr_ietf.amr | 358 }K??????+4 ?<B[??Ax??T?f?'??9?5r??f9?<Dm?f?y????;??_??B?jR????vp< ~??i??)b?J??tI??A??Y?x?<D?%O? ?U?IM?j?? ??t?[p??#>/?<?g?0?U?(??%??}???L1ph#?$P<?y?'?a??3?p??c?iF?+8+m?0<>??Q??=+?|-???vx?ol??c0<D?W?????Iz?q?q?m?agd??`* <D^?=?????5f?????[]?Q????<????KU??p0~?!?i???????`<"$'???????{?$t8????&?D?<?B??)s????G2
??^[0????<;{'??{??n???<???[l_'?m???<?7{/?K??????o?X )??A#??b?<<]?>X???K1v?D?7???z/u7?>?+?<RpO???0???m??lz??Rj??AZ < |