/art/compiler/utils/ |
arena_bit_vector.cc | 25 explicit ArenaBitVectorAllocator(ArenaAlloc* arena) : arena_(arena) {} 34 static void* operator new(size_t size, ArenaAlloc* arena) { 35 return arena->Alloc(sizeof(ArenaBitVectorAllocator), kArenaAllocGrowableBitMap); 44 ArenaBitVector::ArenaBitVector(ArenaAllocator* arena, unsigned int start_bits, 47 new (arena) ArenaBitVectorAllocator<ArenaAllocator>(arena)), kind_(kind) { 51 ArenaBitVector::ArenaBitVector(ScopedArenaAllocator* arena, unsigned int start_bits, 54 new (arena) ArenaBitVectorAllocator<ScopedArenaAllocator>(arena)), kind_(kind) [all...] |
arena_bit_vector.h | 51 * A BitVector implementation that uses Arena allocation. 55 ArenaBitVector(ArenaAllocator* arena, uint32_t start_bits, bool expandable, 57 ArenaBitVector(ScopedArenaAllocator* arena, uint32_t start_bits, bool expandable, 61 static void* operator new(size_t size, ArenaAllocator* arena) { 62 return arena->Alloc(sizeof(ArenaBitVector), kArenaAllocGrowableBitMap); 64 static void* operator new(size_t size, ScopedArenaAllocator* arena) { 65 return arena->Alloc(sizeof(ArenaBitVector), kArenaAllocGrowableBitMap);
|
arena_allocator_test.cc | 25 ArenaAllocator arena(&pool); 26 ArenaBitVector bv(&arena, 10, true);
|
arena_allocator.cc | 32 constexpr size_t Arena::kDefaultSize; 83 void ArenaAllocatorStatsImpl<kCount>::Dump(std::ostream& os, const Arena* first, 88 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) { 89 malloc_bytes += arena->Size(); 90 lost_bytes += arena->RemainingSpace(); 93 // The lost_bytes_adjustment is used to make up for the fact that the current arena 157 auto* arena = free_arenas_; local [all...] |
/external/chromium_org/third_party/tcmalloc/chromium/src/base/ |
low_level_alloc.cc | 71 LowLevelAlloc::Arena *arena; // pointer to parent arena member in struct:low_level_alloc_internal::AllocList::Header 181 // Arena implementation 183 struct LowLevelAlloc::Arena { 184 Arena() : mu(SpinLock::LINKER_INITIALIZED) {} // does nothing; for static init 185 explicit Arena(int) : pagesize(0) {} // set pagesize to zero explicitly 200 // The default arena, which is used when 0 is passed instead of an Arena 202 static struct LowLevelAlloc::Arena default_arena 383 LowLevelAlloc::Arena *arena = a->header.arena; local 420 LowLevelAlloc::Arena *arena = f->header.arena; local [all...] |
low_level_alloc.h | 45 struct Arena; // an arena from which memory may be allocated 48 // that have been newly allocated from the specific arena. 55 static void *AllocWithArena(size_t request, Arena *arena) 61 // Free() since that call to Alloc(). The space is returned to the arena 70 // Create a new arena. 71 // The root metadata for the new arena is allocated in the 87 // as well as generatating an arena that provides async-signal-safe 90 static Arena *NewArena(int32 flags, Arena *meta_data_arena) [all...] |
/external/chromium_org/third_party/tcmalloc/vendor/src/base/ |
low_level_alloc.cc | 71 LowLevelAlloc::Arena *arena; // pointer to parent arena member in struct:low_level_alloc_internal::AllocList::Header 181 // Arena implementation 183 struct LowLevelAlloc::Arena { 184 Arena() : mu(SpinLock::LINKER_INITIALIZED) {} // does nothing; for static init 185 explicit Arena(int) : pagesize(0) {} // set pagesize to zero explicitly 200 // The default arena, which is used when 0 is passed instead of an Arena 202 static struct LowLevelAlloc::Arena default_arena 383 LowLevelAlloc::Arena *arena = a->header.arena; local 420 LowLevelAlloc::Arena *arena = f->header.arena; local [all...] |
low_level_alloc.h | 45 struct Arena; // an arena from which memory may be allocated 48 // that have been newly allocated from the specific arena. 55 static void *AllocWithArena(size_t request, Arena *arena) 61 // Free() since that call to Alloc(). The space is returned to the arena 70 // Create a new arena. 71 // The root metadata for the new arena is allocated in the 87 // as well as generatating an arena that provides async-signal-safe 90 static Arena *NewArena(int32 flags, Arena *meta_data_arena) [all...] |
/external/jemalloc/src/ |
tcache.c | 72 arena_tcache_fill_small(tcache->arena, tbin, binind, 93 /* Lock the arena bin associated with the first object. */ 96 arena_t *arena = chunk->arena; local 97 arena_bin_t *bin = &arena->bins[binind]; 99 if (config_prof && arena == tcache->arena) { 100 if (arena_prof_accum(arena, tcache->prof_accumbytes)) 106 if (config_stats && arena == tcache->arena) { 177 arena_t *arena = chunk->arena; local 225 arena_t *arena = tcache->arena; local 368 arena_t *arena = tcache->arena; local 381 arena_t *arena = tcache->arena; local 397 arena_t *arena = chunk->arena; local 405 arena_t *arena = chunk->arena; local [all...] |
arena.c | 53 static void arena_purge(arena_t *arena, bool all); 54 static void arena_run_dalloc(arena_t *arena, arena_run_t *run, bool dirty, 56 static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, 58 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, 230 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, 242 arena_chunk_dirty_remove(&arena->chunks_dirty, chunk); 252 arena->ndirty += npages; 256 arena_chunk_dirty_insert(&arena->chunks_dirty, chunk); 258 arena_avail_tree_insert(&arena->runs_avail, arena_mapp_get(chunk, 263 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind 2210 arena_t *arena; local [all...] |
huge.c | 16 huge_malloc(arena_t *arena, size_t size, bool zero) 19 return (huge_palloc(arena, size, chunksize, zero)); 23 huge_palloc(arena_t *arena, size_t size, size_t alignment, bool zero) 48 arena = choose_arena(arena); 49 ret = arena_chunk_alloc_huge(arena, csize, alignment, &is_zeroed); 58 node->arena = arena; 93 huge_ralloc(arena_t *arena, void *ptr, size_t oldsize, size_t size, 109 ret = huge_palloc(arena, size + extra, alignment, zero) [all...] |
/art/compiler/dex/ |
backend.h | 51 explicit Backend(ArenaAllocator* arena) : arena_(arena) {}
|
/external/chromium_org/third_party/WebKit/Source/platform/ |
PODFreeListArenaTest.cpp | 64 int getFreeListSize(const PassRefPtr<PODFreeListArena<TestClass1> > arena) const 66 return arena->getFreeListSizeForTesting(); 70 // Make sure the arena can successfully allocate from more than one 75 RefPtr<PODFreeListArena<TestClass1> > arena = PODFreeListArena<TestClass1>::create(allocator); local 78 arena->allocateObject(); 82 // Make sure the arena frees all allocated regions during destruction. 87 RefPtr<PODFreeListArena<TestClass1> > arena = PODFreeListArena<TestClass1>::create(allocator); local 89 arena->allocateObject(); 95 // Make sure the arena runs constructors of the objects allocated within. 98 RefPtr<PODFreeListArena<TestClass1> > arena = PODFreeListArena<TestClass1>::create() local 112 RefPtr<PODFreeListArena<TestClass1> > arena = PODFreeListArena<TestClass1>::create(); local 141 RefPtr<PODFreeListArena<TestClass1> > arena = PODFreeListArena<TestClass1>::create(); local 155 RefPtr<PODFreeListArena<TestClass2> > arena = PODFreeListArena<TestClass2>::create(); local [all...] |
PODArenaTest.cpp | 62 // Make sure the arena can successfully allocate from more than one 67 RefPtr<PODArena> arena = PODArena::create(allocator); local 70 arena->allocateObject<TestClass1>(); 74 // Make sure the arena frees all allocated regions during destruction. 79 RefPtr<PODArena> arena = PODArena::create(allocator); local 81 arena->allocateObject<TestClass1>(); 87 // Make sure the arena runs constructors of the objects allocated within. 90 RefPtr<PODArena> arena = PODArena::create(); local 92 TestClass1* tc1 = arena->allocateObject<TestClass1>(); 97 TestClass2* tc2 = arena->allocateObject<TestClass2>() [all...] |
/external/chromium_org/net/third_party/nss/ssl/bodge/ |
secitem_array.c | 28 SECITEM_AllocArray(PLArenaPool *arena, SECItemArray *array, unsigned int len) 33 if (arena != NULL) { 34 mark = PORT_ArenaMark(arena); 38 if (arena != NULL) { 39 result = PORT_ArenaZAlloc(arena, sizeof(SECItemArray)); 53 if (arena != NULL) { 54 result->items = PORT_ArenaZNewArray(arena, SECItem, len); 66 PORT_ArenaUnmark(arena, mark); 71 if ( arena != NULL ) { 73 PORT_ArenaRelease(arena, mark) [all...] |
/external/chromium_org/third_party/leveldatabase/src/util/ |
arena_test.cc | 5 #include "util/arena.h" 15 Arena arena; local 20 Arena arena; local 33 // Our arena disallows size 0 allocations. 38 r = arena.AllocateAligned(s); 40 r = arena.Allocate(s); 49 ASSERT_GE(arena.MemoryUsage(), bytes); 51 ASSERT_LE(arena.MemoryUsage(), bytes * 1.10) [all...] |
/external/jemalloc/include/jemalloc/internal/ |
huge.h | 12 void *huge_malloc(arena_t *arena, size_t size, bool zero); 13 void *huge_palloc(arena_t *arena, size_t size, size_t alignment, bool zero); 16 void *huge_ralloc(arena_t *arena, void *ptr, size_t oldsize, size_t size,
|
extent.h | 27 /* Arena from which this extent came, if any */ 28 arena_t *arena; member in struct:extent_node_s
|
arena.h | 37 * The minimum ratio of active:dirty pages per arena is computed as: 62 * Such antics aren't warranted for per arena data structures, but 81 * purging thread is operating without the arena lock held. 162 /* Arena chunk header. */ 164 /* Arena that owns the chunk. */ 165 arena_t *arena; member in struct:arena_chunk_s 167 /* Linkage for tree of arena chunks that contain dirty runs. */ 209 * than one per arena), but mainly to avoid false cacheline sharing. 277 * locked. Run allocation/deallocation are protected by the arena lock, 303 /* This arena's index within the arenas array. * [all...] |
/art/compiler/optimizing/ |
find_loops_test.cc | 46 ArenaPool arena; local 47 HGraph* graph = TestCode(data, &arena); 58 ArenaPool arena; local 59 HGraph* graph = TestCode(data, &arena); 73 ArenaPool arena; local 74 HGraph* graph = TestCode(data, &arena); 89 ArenaPool arena; local 90 HGraph* graph = TestCode(data, &arena); 103 ArenaPool arena; local 104 HGraph* graph = TestCode(data, &arena); 148 ArenaPool arena; local 175 ArenaPool arena; local 199 ArenaPool arena; local 224 ArenaPool arena; local 250 ArenaPool arena; local 275 ArenaPool arena; local 306 ArenaPool arena; local 335 ArenaPool arena; local 349 ArenaPool arena; local [all...] |
stack_map_test.cc | 36 ArenaAllocator arena(&pool); 37 StackMapStream<size_t> stream(&arena); 39 ArenaBitVector sp_mask(&arena, 0, false); 45 void* memory = arena.Alloc(size, kArenaAllocMisc); 73 ArenaAllocator arena(&pool); 74 StackMapStream<size_t> stream(&arena); 76 ArenaBitVector sp_mask1(&arena, 0, true); 85 ArenaBitVector sp_mask2(&arena, 0, true); 92 void* memory = arena.Alloc(size, kArenaAllocMisc);
|
/external/jemalloc/test/unit/ |
stats.c | 99 unsigned arena; local 107 arena = 0; 108 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeof(arena)), 116 assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0, 163 unsigned arena; local 171 arena = 0; 172 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeof(arena)) 212 unsigned arena; local 256 unsigned arena; local 329 unsigned arena; local [all...] |
/external/chromium_org/net/third_party/mozilla_security_manager/ |
nsKeygenHandler.cpp | 110 PLArenaPool *arena = NULL; local 179 arena = PORT_NewArena(DER_DEFAULT_CHUNKSIZE); 180 if (!arena) { 187 sec_rv = DER_Encode(arena, &spkiItem, CERTSubjectPublicKeyInfoTemplate, 200 sec_rv = DER_Encode(arena, &pkacItem, CERTPublicKeyAndChallengeTemplate, 209 sec_rv = SEC_DerSignData(arena, &signedItem, pkacItem.data, pkacItem.len, 250 if (arena) { 251 PORT_FreeArena(arena, PR_TRUE);
|
/external/chromium_org/third_party/re2/util/ |
arena.h | 16 // #include "lib/arena-inl.h" 18 // UnsafeArena arena(1000); 19 // Foo* foo = new (AllocateInArena, &arena) Foo; 89 // Operators for allocation on the arena 90 // Syntax: new (AllocateInArena, arena) MyClass; 98 re2::UnsafeArena *arena) { 99 return reinterpret_cast<char*>(arena->GetMemory(size, 1));
|
/external/chromium_org/crypto/third_party/nss/ |
secsign.cc | 54 SECStatus DerSignData(PLArenaPool *arena, 60 return SEC_DerSignData(arena, result, input->data, input->len, key, 115 rv = SECOID_SetAlgorithmID(arena, &sd.signatureAlgorithm, algo_id, 0); 127 arena, result, &sd, SEC_ASN1_GET(CERT_SignedDataTemplate));
|