/external/chromium_org/third_party/tcmalloc/chromium/src/tests/ |
low_level_alloc_unittest.cc | 77 // If use_new_arena is true, use a fresh arena, and then delete it. 87 LowLevelAlloc::Arena *arena = 0; local 90 arena = LowLevelAlloc::NewArena(flags, LowLevelAlloc::DefaultArena()); 104 arena == 0 106 : LowLevelAlloc::AllocWithArena(block_desc.len, arena)); 142 CHECK(LowLevelAlloc::DeleteArena(arena)); 177 CHECK_NE(allocates, 0); // default arena calls hooks 185 CHECK_GT(allocates, 5000); // arena calls hooks 188 CHECK_EQ(allocates, 0); // arena doesn't call hook [all...] |
/external/chromium_org/third_party/tcmalloc/vendor/src/tests/ |
low_level_alloc_unittest.cc | 77 // If use_new_arena is true, use a fresh arena, and then delete it. 87 LowLevelAlloc::Arena *arena = 0; local 90 arena = LowLevelAlloc::NewArena(flags, LowLevelAlloc::DefaultArena()); 104 arena == 0 106 : LowLevelAlloc::AllocWithArena(block_desc.len, arena)); 142 CHECK(LowLevelAlloc::DeleteArena(arena)); 177 CHECK_NE(allocates, 0); // default arena calls hooks 185 CHECK_GT(allocates, 5000); // arena calls hooks 188 CHECK_EQ(allocates, 0); // arena doesn't call hook [all...] |
/system/core/libcutils/tests/memset_mips/ |
test_memset.c | 71 * Using a big arena means that memset will most likely miss in the cache 76 char arena[ARENASIZE+8]; /* Allow space for guard words */ variable 78 char arena[ARENASIZE]; variable 123 d = &arena[offset]; 127 for (p = &arena[offset-4]; p < (void *)&arena[offset+n+4]; p = (void *)((uint32_t)p + size)) { 139 for (p = &arena[offset-4]; p < (void *)&arena[offset+n+4]; p = (void *)((uint32_t)p + size)) { 147 if (p < (void *)&arena[offset]) { 151 else if (p < (void *)&arena[offset+n]) [all...] |
/art/compiler/dex/ |
growable_array.h | 69 static void* operator new(size_t size, ArenaAllocator* arena) { 70 return arena->Alloc(sizeof(GrowableArray::Iterator), ArenaAllocator::kAllocGrowableArray); 79 GrowableArray(ArenaAllocator* arena, size_t init_length, OatListKind kind = kGrowableArrayMisc) 80 : arena_(arena), 155 static void* operator new(size_t size, ArenaAllocator* arena) { 156 return arena->Alloc(sizeof(GrowableArray<T>), ArenaAllocator::kAllocGrowableArray);
|
arena_bit_vector.h | 69 static void* operator new(size_t size, ArenaAllocator* arena) { 70 return arena->Alloc(sizeof(ArenaBitVector::Iterator), 82 ArenaBitVector(ArenaAllocator* arena, unsigned int start_bits, bool expandable, 86 static void* operator new(size_t size, ArenaAllocator* arena) { 87 return arena->Alloc(sizeof(ArenaBitVector), ArenaAllocator::kAllocGrowableBitMap);
|
frontend.cc | 173 cu.mir_graph.reset(new MIRGraph(&cu, &cu.arena)); 226 cu.cg.reset(PortableCodeGenerator(&cu, cu.mir_graph.get(), &cu.arena, llvm_compilation_unit)); 231 cu.cg.reset(ArmCodeGenerator(&cu, cu.mir_graph.get(), &cu.arena)); 234 cu.cg.reset(MipsCodeGenerator(&cu, cu.mir_graph.get(), &cu.arena)); 237 cu.cg.reset(X86CodeGenerator(&cu, cu.mir_graph.get(), &cu.arena)); 257 if (cu.arena.BytesAllocated() > (5 * 1024 *1024)) { 258 MemStats mem_stats(cu.arena); 264 LOG(INFO) << "MEMINFO " << cu.arena.BytesAllocated() << " " << cu.mir_graph->GetNumBlocks()
|
/external/valgrind/main/memcheck/tests/ |
sbfragment.c | 22 printf("%10d int arena; /* non-mmapped space allocated from system */\n", mallinfo_result.arena); 79 // We consider that an arena of up to 3 times more than bigsize is ok. 84 // Under valgrind, hblkhd is 0 : all the space is in arena. 88 if (mallinfo_result.arena + mallinfo_result.hblkhd > 3 * bigsize) 90 (unsigned long) mallinfo_result.arena
|
mallinfo.c | 21 printf("arena = %d\n", mi.arena); /* non-mmapped space allocated from system */ 64 // arena should be reasonably close to fordblks + uordblks 65 if (mi.arena < mi.fordblks + mi.uordblks) 68 if (mi.arena/5*4 > mi.fordblks + mi.uordblks)
|
/external/chromium_org/net/cert/ |
x509_util_nss_unittest.cc | 37 crypto::ScopedPLArenaPool arena(PORT_NewArena(DER_DEFAULT_CHUNKSIZE)); 47 SECStatus rv = SEC_ASN1DecodeItem(arena.get(), &sd, 55 arena.get(), 91 // IA5Encode and arena allocate SECItem. 92 PLArenaPool* arena = PORT_NewArena(DER_DEFAULT_CHUNKSIZE); local 93 SECItem* expected = SEC_ASN1EncodeItem(arena, 102 SECStatus ok = SEC_StringToOID(arena, &ob_cert_oid, 139 PORT_FreeArena(arena, PR_FALSE);
|
x509_util_nss.cc | 146 // |arena| is used to encode the cert. 147 PLArenaPool* arena = cert->arena; local 153 SECStatus rv = SECOID_SetAlgorithmID(arena, &cert->signature, algo_id, 0); 173 rv = DerSignData(arena, &result, &der, key, algo_id); 211 // |arena| is the allocation pool to use. 214 CERTName* CreateCertNameFromEncoded(PLArenaPool* arena, 216 if (!arena) 219 ScopedCERTName name(PORT_ArenaZNew(arena, CERTName)); 229 arena, name.get(), SEC_ASN1_GET(CERT_NameTemplate), &item) [all...] |
x509_util_nss.h | 64 // DistinguishedName items. All objects are created in a given arena. 66 // |arena| is the arena used for all allocations. 72 PLArenaPool* arena,
|
/libcore/luni/src/main/java/java/util/concurrent/ |
Exchanger.java | 106 * form of elimination arena, that spreads out this contention by 110 * across threads, but instead give threads arena indices that 120 * Implementing an effective arena requires allocating a bunch of 135 * The arena starts out with only one used slot. We expand the 136 * effective arena size by tracking collisions; i.e., failed CASes 144 * collided at each slot within the current arena bound, it tries 145 * to expand the arena size by one. We track collisions within 151 * The effective arena size is reduced (when there is more than 153 * decrement the arena size on expiration. The value of "a while" 158 * switching on multiprocessors is extremely slow/wasteful. Arena 310 private volatile Node[] arena; field in class:Exchanger [all...] |
/prebuilts/python/darwin-x86/2.7.5/include/python2.7/ |
asdl.h | 30 asdl_seq *asdl_seq_new(int size, PyArena *arena); 31 asdl_int_seq *asdl_int_seq_new(int size, PyArena *arena);
|
/prebuilts/python/linux-x86/2.7.5/include/python2.7/ |
asdl.h | 30 asdl_seq *asdl_seq_new(int size, PyArena *arena); 31 asdl_int_seq *asdl_int_seq_new(int size, PyArena *arena);
|
/external/chromium_org/crypto/ |
signature_verifier_nss.cc | 92 PLArenaPool* arena = PORT_NewArena(DER_DEFAULT_CHUNKSIZE); 93 if (!arena) { 104 rv = SEC_QuickDERDecodeItem(arena, &sig_alg_id, 109 PORT_FreeArena(arena, PR_TRUE); 122 PORT_FreeArena(arena, PR_TRUE); // Done with sig_alg_id.
|
/external/chromium_org/third_party/npapi/npspy/extern/nspr/ |
plarenas.h | 44 ** Allocate an arena pool as specified by the parameters. 59 ** Destroy an arena pool previously allocated by PL_AllocArenaPool(). 61 ** This function may fail if the arena is not empty and the caller 70 ** Initialize an arena pool with the given name for debugging and metering, 71 ** with a minimum size per arena of size bytes.
|
/external/chromium/chrome/browser/importer/ |
nss_decryptor_system_nss.cc | 144 pk11Decrypt(PK11SlotInfo *slot, PLArenaPool *arena, 160 PORT_ArenaAlloc(arena, paddedResult.len)); 186 PLArenaPool *arena = 0; local 188 arena = PORT_NewArena(SEC_ASN1_DEFAULT_ARENA_SIZE); 189 if (!arena) { rv = SECFailure; goto loser; } 193 rv = SEC_QuickDERDecodeItem(arena, &sdrResult, g_template, data); 206 rv = pk11Decrypt(slot, arena, type, key, params, 229 rv = pk11Decrypt(slot, arena, type, testKey, params, 263 if (arena) PORT_FreeArena(arena, PR_TRUE) [all...] |
/external/chromium_org/chrome/utility/importer/ |
nss_decryptor_system_nss.cc | 147 pk11Decrypt(PK11SlotInfo *slot, PLArenaPool *arena, 163 PORT_ArenaAlloc(arena, paddedResult.len)); 189 PLArenaPool *arena = 0; local 191 arena = PORT_NewArena(SEC_ASN1_DEFAULT_ARENA_SIZE); 192 if (!arena) { rv = SECFailure; goto loser; } 196 rv = SEC_QuickDERDecodeItem(arena, &sdrResult, g_template, data); 209 rv = pk11Decrypt(slot, arena, type, key, params, 232 rv = pk11Decrypt(slot, arena, type, testKey, params, 266 if (arena) PORT_FreeArena(arena, PR_TRUE) [all...] |
/dalvik/vm/compiler/ |
Utility.cpp | 23 /* Allocate the initial memory block for arena-based allocation */ 42 /* Arena-based malloc for compilation tasks */ 58 * See if there are previously allocated arena blocks before the last 68 /* Time to allocate a new arena */ 72 ALOGE("Arena allocation failure"); 82 ALOGI("Total arena pages for JIT: %d", numArenaBlocks); 89 /* Reclaim all the arena blocks allocated so far */ 236 * the code cache, arena size, and work queue length, and various JIT stats. 247 ALOGD("Compiler arena uses %d blocks (%d bytes each)", 268 * memory is allocated from the compiler arena [all...] |
CompilerUtility.h | 22 /* Each arena page has some overhead, so take a few bytes off 8k */ 25 /* Allocate the initial memory block for arena-based allocation */
|
/external/chromium_org/crypto/third_party/nss/ |
secsign.cc | 54 SECStatus DerSignData(PLArenaPool *arena, 60 return SEC_DerSignData(arena, result, input->data, input->len, key, 115 rv = SECOID_SetAlgorithmID(arena, &sd.signatureAlgorithm, algo_id, 0); 127 arena, result, &sd, SEC_ASN1_GET(CERT_SignedDataTemplate));
|
/external/chromium_org/third_party/mesa/src/src/gallium/drivers/radeon/ |
AMDILDeviceInfo.h | 47 ArenaSegment = 0xD, // Use for Arena UAV per pointer 12-1023. 54 // uav(Arena on HD5XXX/HD6XXX and Raw on HD4XXX). 60 ArenaVectors = 0x17, // Flag to specify if vector loads from arena work. 65 ArenaUAV = 0x1C, // Flag to specify that arena uav is supported.
|
/external/chromium_org/tools/deep_memory_profiler/tests/data/ |
heap.01234.0002.heap | 19 unhooked-arena 64128 64128
|
/external/mesa3d/src/gallium/drivers/radeon/ |
AMDILDeviceInfo.h | 47 ArenaSegment = 0xD, // Use for Arena UAV per pointer 12-1023. 54 // uav(Arena on HD5XXX/HD6XXX and Raw on HD4XXX). 60 ArenaVectors = 0x17, // Flag to specify if vector loads from arena work. 65 ArenaUAV = 0x1C, // Flag to specify that arena uav is supported.
|
/external/chromium_org/third_party/leveldatabase/src/db/ |
skiplist.h | 30 #include "util/arena.h" 35 class Arena; 44 // and will allocate memory using "*arena". Objects allocated in the arena 46 explicit SkipList(Comparator cmp, Arena* arena); 99 Arena* const arena_; // Arena used for allocations of nodes 322 SkipList<Key,Comparator>::SkipList(Comparator cmp, Arena* arena) [all...] |