/external/chromium_org/third_party/tcmalloc/chromium/src/base/ |
low_level_alloc.h | 45 struct Arena; // an arena from which memory may be allocated 48 // that have been newly allocated from the specific arena. 55 static void *AllocWithArena(size_t request, Arena *arena) 61 // Free() since that call to Alloc(). The space is returned to the arena 70 // Create a new arena. 71 // The root metadata for the new arena is allocated in the 87 // as well as generatating an arena that provides async-signal-safe 90 static Arena *NewArena(int32 flags, Arena *meta_data_arena) [all...] |
low_level_alloc.cc | 71 LowLevelAlloc::Arena *arena; // pointer to parent arena member in struct:low_level_alloc_internal::AllocList::Header 181 // Arena implementation 183 struct LowLevelAlloc::Arena { 184 Arena() : mu(SpinLock::LINKER_INITIALIZED) {} // does nothing; for static init 185 explicit Arena(int) : pagesize(0) {} // set pagesize to zero explicitly 200 // The default arena, which is used when 0 is passed instead of an Arena 202 static struct LowLevelAlloc::Arena default_arena 383 LowLevelAlloc::Arena *arena = a->header.arena; local 420 LowLevelAlloc::Arena *arena = f->header.arena; local [all...] |
/external/chromium_org/third_party/tcmalloc/vendor/src/base/ |
low_level_alloc.h | 45 struct Arena; // an arena from which memory may be allocated 48 // that have been newly allocated from the specific arena. 55 static void *AllocWithArena(size_t request, Arena *arena) 61 // Free() since that call to Alloc(). The space is returned to the arena 70 // Create a new arena. 71 // The root metadata for the new arena is allocated in the 87 // as well as generatating an arena that provides async-signal-safe 90 static Arena *NewArena(int32 flags, Arena *meta_data_arena) [all...] |
low_level_alloc.cc | 71 LowLevelAlloc::Arena *arena; // pointer to parent arena member in struct:low_level_alloc_internal::AllocList::Header 181 // Arena implementation 183 struct LowLevelAlloc::Arena { 184 Arena() : mu(SpinLock::LINKER_INITIALIZED) {} // does nothing; for static init 185 explicit Arena(int) : pagesize(0) {} // set pagesize to zero explicitly 200 // The default arena, which is used when 0 is passed instead of an Arena 202 static struct LowLevelAlloc::Arena default_arena 383 LowLevelAlloc::Arena *arena = a->header.arena; local 420 LowLevelAlloc::Arena *arena = f->header.arena; local [all...] |
/external/chromium_org/third_party/leveldatabase/src/util/ |
arena.h | 15 class Arena { 17 Arena(); 18 ~Arena(); 27 // by the arena (including space allocated but not yet used for user 48 Arena(const Arena&); 49 void operator=(const Arena&); 52 inline char* Arena::Allocate(size_t bytes) {
|
arena.cc | 5 #include "util/arena.h" 12 Arena::Arena() { 18 Arena::~Arena() { 24 char* Arena::AllocateFallback(size_t bytes) { 42 char* Arena::AllocateAligned(size_t bytes) { 61 char* Arena::AllocateNewBlock(size_t block_bytes) {
|
arena_test.cc | 5 #include "util/arena.h" 15 Arena arena; local 20 Arena arena; local 33 // Our arena disallows size 0 allocations. 38 r = arena.AllocateAligned(s); 40 r = arena.Allocate(s); 49 ASSERT_GE(arena.MemoryUsage(), bytes); 51 ASSERT_LE(arena.MemoryUsage(), bytes * 1.10) [all...] |
/art/compiler/utils/ |
arena_allocator.h | 31 class Arena; 77 void Dump(std::ostream& os, const Arena* first, ssize_t lost_bytes_adjustment) const { 93 void Dump(std::ostream& os, const Arena* first, ssize_t lost_bytes_adjustment) const; 105 class Arena { 108 explicit Arena(size_t size = kDefaultSize); 109 ~Arena(); 136 Arena* next_; 142 DISALLOW_COPY_AND_ASSIGN(Arena); 149 Arena* AllocArena(size_t size) LOCKS_EXCLUDED(lock_); 150 void FreeArenaChain(Arena* first) LOCKS_EXCLUDED(lock_) [all...] |
arena_allocator.cc | 32 constexpr size_t Arena::kDefaultSize; 83 void ArenaAllocatorStatsImpl<kCount>::Dump(std::ostream& os, const Arena* first, 88 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) { 89 malloc_bytes += arena->Size(); 90 lost_bytes += arena->RemainingSpace(); 93 // The lost_bytes_adjustment is used to make up for the fact that the current arena 157 auto* arena = free_arenas_; local [all...] |
scoped_arena_allocator.h | 86 Arena* bottom_arena_; 87 Arena* top_arena_; 131 Arena* mark_arena_;
|
scoped_arena_allocator.cc | 58 size_t allocation_size = std::max(Arena::kDefaultSize, rounded_bytes); 65 Arena* tail = top_arena_->next_; 84 // Update how many bytes we have allocated into the arena so that the arena pool knows how 86 // zero-initialized, the Arena may be reused by ArenaAllocator which does guarantee this.
|
/external/clang/lib/Analysis/ |
ThreadSafetyCommon.cpp | 151 return new (Arena) til::Literal(cast<Expr>(S)); 161 return new (Arena) til::Undefined(S); 186 return new (Arena) til::LiteralPtr(VD); 203 E = new (Arena) til::SApply(E); 204 return new (Arena) til::Project(E, ME->getMemberDecl()); 214 E = new (Arena) til::Apply(E, A); 216 return new (Arena) til::Call(E, CE); 239 return new (Arena) til::Undefined(UO); 248 return new (Arena) 251 return new (Arena) [all...] |
ThreadSafetyTIL.cpp | 53 Predecessors.reserveCheck(1, Arena); 57 Ph->values().reserveCheck(1, Arena); 65 Predecessors.reserve(NumPreds, Arena); 68 Ph->values().reserve(NumPreds, Arena);
|
/external/clang/include/clang/Analysis/Analyses/ |
ThreadSafetyTraverse.h | 133 Container(CopyReducerBase &S, unsigned N) : Elems(S.Arena, N) {} 141 CopyReducerBase(MemRegionRef A) : Arena(A) {} 144 MemRegionRef Arena; 165 return new (Arena) Undefined(Orig); 168 return new (Arena) Wildcard(Orig); 172 return new (Arena) Literal(Orig); 176 return new (Arena) LiteralT<T>(Orig); 179 return new (Arena) LiteralPtr(Orig); 183 return new (Arena) Function(Orig, Nvd, E0); 186 return new (Arena) SFunction(Orig, Nvd, E0) [all...] |
ThreadSafetyCommon.h | 238 : Arena(A), SelfVar(nullptr), Scfg(nullptr), CurrentBB(nullptr), 241 SelfVar = new (Arena) til::Variable(nullptr); 365 til::MemRegionRef Arena;
|
/external/chromium_org/third_party/leveldatabase/src/db/ |
memtable.h | 12 #include "util/arena.h" 81 Arena arena_;
|
skiplist.h | 30 #include "util/arena.h" 35 class Arena; 44 // and will allocate memory using "*arena". Objects allocated in the arena 46 explicit SkipList(Comparator cmp, Arena* arena); 99 Arena* const arena_; // Arena used for allocations of nodes 322 SkipList<Key,Comparator>::SkipList(Comparator cmp, Arena* arena) [all...] |
skiplist_test.cc | 8 #include "util/arena.h" 32 Arena arena; local 34 SkipList<Key, Comparator> list(cmp, &arena); 52 Arena arena; local 54 SkipList<Key, Comparator> list(cmp, &arena); 207 Arena arena_;
|
/external/valgrind/main/coregrind/ |
m_mallocfree.c | 70 // Each arena has a sorted array of superblocks, which expands 80 red zone bytes (depends on Arena.rz_szB, but >= sizeof(void*)) 82 red zone bytes (depends on Arena.rz_szB, but >= sizeof(void*)) 90 excess red zone bytes (if Arena.rz_szB > sizeof(void*)) 92 excess red zone bytes (if Arena.rz_szB > sizeof(void*)) 174 // of the same arena can be reclaimed or till a new superblock is needed 175 // in any arena. 191 // An arena. 'freelist' is a circular, doubly-linked list. 'rz_szB' is 211 // superblocks in the arena. If this array is expanded, which 247 Arena; 510 ArenaId arena = a -vg_arena; local 799 Arena* arena = arenaId_to_ArenaP(aid); local 1447 Arena *arena; local [all...] |
/external/chromium_org/third_party/tcmalloc/chromium/src/ |
type_profiler_map.cc | 51 LowLevelAlloc::Arena* g_type_profiler_map_memory = NULL;
|
memory_region_map.h | 250 // We allocate STL objects in our own arena. 296 // Arena used for our allocations in regions_. 297 static LowLevelAlloc::Arena* arena_;
|
/prebuilts/python/darwin-x86/2.7.5/lib/python2.7/multiprocessing/ |
heap.py | 57 class Arena(object): 63 self.name = 'pym-%d-%d' % (os.getpid(), Arena._counter.next()) 79 class Arena(object): 120 arena = Arena(length) 121 self._arenas.append(arena) 122 return (arena, 0, length) 130 (arena, start, stop) = block 131 del self._start_to_block[(arena, start)] 132 del self._stop_to_block[(arena, stop) [all...] |
/prebuilts/python/linux-x86/2.7.5/lib/python2.7/multiprocessing/ |
heap.py | 57 class Arena(object): 63 self.name = 'pym-%d-%d' % (os.getpid(), Arena._counter.next()) 79 class Arena(object): 120 arena = Arena(length) 121 self._arenas.append(arena) 122 return (arena, 0, length) 130 (arena, start, stop) = block 131 del self._start_to_block[(arena, start)] 132 del self._stop_to_block[(arena, stop) [all...] |
/external/chromium_org/third_party/tcmalloc/chromium/src/tests/ |
low_level_alloc_unittest.cc | 77 // If use_new_arena is true, use a fresh arena, and then delete it. 87 LowLevelAlloc::Arena *arena = 0; local 90 arena = LowLevelAlloc::NewArena(flags, LowLevelAlloc::DefaultArena()); 104 arena == 0 106 : LowLevelAlloc::AllocWithArena(block_desc.len, arena)); 142 CHECK(LowLevelAlloc::DeleteArena(arena)); 177 CHECK_NE(allocates, 0); // default arena calls hooks 185 CHECK_GT(allocates, 5000); // arena calls hooks 188 CHECK_EQ(allocates, 0); // arena doesn't call hook [all...] |
/external/chromium_org/third_party/tcmalloc/vendor/src/ |
memory_region_map.h | 226 // We allocate STL objects in our own arena. 272 // Arena used for our allocations in regions_. 273 static LowLevelAlloc::Arena* arena_;
|