HomeSort by relevance Sort by last modified time
    Searched refs:Arena (Results 1 - 25 of 34) sorted by null

1 2

  /external/chromium_org/third_party/tcmalloc/chromium/src/base/
low_level_alloc.h 45 struct Arena; // an arena from which memory may be allocated
48 // that have been newly allocated from the specific arena.
55 static void *AllocWithArena(size_t request, Arena *arena)
61 // Free() since that call to Alloc(). The space is returned to the arena
70 // Create a new arena.
71 // The root metadata for the new arena is allocated in the
87 // as well as generatating an arena that provides async-signal-safe
90 static Arena *NewArena(int32 flags, Arena *meta_data_arena)
    [all...]
low_level_alloc.cc 71 LowLevelAlloc::Arena *arena; // pointer to parent arena member in struct:low_level_alloc_internal::AllocList::Header
181 // Arena implementation
183 struct LowLevelAlloc::Arena {
184 Arena() : mu(SpinLock::LINKER_INITIALIZED) {} // does nothing; for static init
185 explicit Arena(int) : pagesize(0) {} // set pagesize to zero explicitly
200 // The default arena, which is used when 0 is passed instead of an Arena
202 static struct LowLevelAlloc::Arena default_arena
383 LowLevelAlloc::Arena *arena = a->header.arena; local
420 LowLevelAlloc::Arena *arena = f->header.arena; local
    [all...]
  /external/chromium_org/third_party/tcmalloc/vendor/src/base/
low_level_alloc.h 45 struct Arena; // an arena from which memory may be allocated
48 // that have been newly allocated from the specific arena.
55 static void *AllocWithArena(size_t request, Arena *arena)
61 // Free() since that call to Alloc(). The space is returned to the arena
70 // Create a new arena.
71 // The root metadata for the new arena is allocated in the
87 // as well as generatating an arena that provides async-signal-safe
90 static Arena *NewArena(int32 flags, Arena *meta_data_arena)
    [all...]
low_level_alloc.cc 71 LowLevelAlloc::Arena *arena; // pointer to parent arena member in struct:low_level_alloc_internal::AllocList::Header
181 // Arena implementation
183 struct LowLevelAlloc::Arena {
184 Arena() : mu(SpinLock::LINKER_INITIALIZED) {} // does nothing; for static init
185 explicit Arena(int) : pagesize(0) {} // set pagesize to zero explicitly
200 // The default arena, which is used when 0 is passed instead of an Arena
202 static struct LowLevelAlloc::Arena default_arena
383 LowLevelAlloc::Arena *arena = a->header.arena; local
420 LowLevelAlloc::Arena *arena = f->header.arena; local
    [all...]
  /external/chromium_org/third_party/leveldatabase/src/util/
arena.h 15 class Arena {
17 Arena();
18 ~Arena();
27 // by the arena (including space allocated but not yet used for user
48 Arena(const Arena&);
49 void operator=(const Arena&);
52 inline char* Arena::Allocate(size_t bytes) {
arena.cc 5 #include "util/arena.h"
12 Arena::Arena() {
18 Arena::~Arena() {
24 char* Arena::AllocateFallback(size_t bytes) {
42 char* Arena::AllocateAligned(size_t bytes) {
61 char* Arena::AllocateNewBlock(size_t block_bytes) {
arena_test.cc 5 #include "util/arena.h"
15 Arena arena; local
20 Arena arena; local
33 // Our arena disallows size 0 allocations.
38 r = arena.AllocateAligned(s);
40 r = arena.Allocate(s);
49 ASSERT_GE(arena.MemoryUsage(), bytes);
51 ASSERT_LE(arena.MemoryUsage(), bytes * 1.10)
    [all...]
  /art/compiler/utils/
arena_allocator.h 31 class Arena;
77 void Dump(std::ostream& os, const Arena* first, ssize_t lost_bytes_adjustment) const {
93 void Dump(std::ostream& os, const Arena* first, ssize_t lost_bytes_adjustment) const;
105 class Arena {
108 explicit Arena(size_t size = kDefaultSize);
109 ~Arena();
136 Arena* next_;
142 DISALLOW_COPY_AND_ASSIGN(Arena);
149 Arena* AllocArena(size_t size) LOCKS_EXCLUDED(lock_);
150 void FreeArenaChain(Arena* first) LOCKS_EXCLUDED(lock_)
    [all...]
arena_allocator.cc 32 constexpr size_t Arena::kDefaultSize;
83 void ArenaAllocatorStatsImpl<kCount>::Dump(std::ostream& os, const Arena* first,
88 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) {
89 malloc_bytes += arena->Size();
90 lost_bytes += arena->RemainingSpace();
93 // The lost_bytes_adjustment is used to make up for the fact that the current arena
157 auto* arena = free_arenas_; local
    [all...]
scoped_arena_allocator.h 86 Arena* bottom_arena_;
87 Arena* top_arena_;
131 Arena* mark_arena_;
scoped_arena_allocator.cc 58 size_t allocation_size = std::max(Arena::kDefaultSize, rounded_bytes);
65 Arena* tail = top_arena_->next_;
84 // Update how many bytes we have allocated into the arena so that the arena pool knows how
86 // zero-initialized, the Arena may be reused by ArenaAllocator which does guarantee this.
  /external/clang/lib/Analysis/
ThreadSafetyCommon.cpp 151 return new (Arena) til::Literal(cast<Expr>(S));
161 return new (Arena) til::Undefined(S);
186 return new (Arena) til::LiteralPtr(VD);
203 E = new (Arena) til::SApply(E);
204 return new (Arena) til::Project(E, ME->getMemberDecl());
214 E = new (Arena) til::Apply(E, A);
216 return new (Arena) til::Call(E, CE);
239 return new (Arena) til::Undefined(UO);
248 return new (Arena)
251 return new (Arena)
    [all...]
ThreadSafetyTIL.cpp 53 Predecessors.reserveCheck(1, Arena);
57 Ph->values().reserveCheck(1, Arena);
65 Predecessors.reserve(NumPreds, Arena);
68 Ph->values().reserve(NumPreds, Arena);
  /external/clang/include/clang/Analysis/Analyses/
ThreadSafetyTraverse.h 133 Container(CopyReducerBase &S, unsigned N) : Elems(S.Arena, N) {}
141 CopyReducerBase(MemRegionRef A) : Arena(A) {}
144 MemRegionRef Arena;
165 return new (Arena) Undefined(Orig);
168 return new (Arena) Wildcard(Orig);
172 return new (Arena) Literal(Orig);
176 return new (Arena) LiteralT<T>(Orig);
179 return new (Arena) LiteralPtr(Orig);
183 return new (Arena) Function(Orig, Nvd, E0);
186 return new (Arena) SFunction(Orig, Nvd, E0)
    [all...]
ThreadSafetyCommon.h 238 : Arena(A), SelfVar(nullptr), Scfg(nullptr), CurrentBB(nullptr),
241 SelfVar = new (Arena) til::Variable(nullptr);
365 til::MemRegionRef Arena;
  /external/chromium_org/third_party/leveldatabase/src/db/
memtable.h 12 #include "util/arena.h"
81 Arena arena_;
skiplist.h 30 #include "util/arena.h"
35 class Arena;
44 // and will allocate memory using "*arena". Objects allocated in the arena
46 explicit SkipList(Comparator cmp, Arena* arena);
99 Arena* const arena_; // Arena used for allocations of nodes
322 SkipList<Key,Comparator>::SkipList(Comparator cmp, Arena* arena)
    [all...]
skiplist_test.cc 8 #include "util/arena.h"
32 Arena arena; local
34 SkipList<Key, Comparator> list(cmp, &arena);
52 Arena arena; local
54 SkipList<Key, Comparator> list(cmp, &arena);
207 Arena arena_;
  /external/valgrind/main/coregrind/
m_mallocfree.c 70 // Each arena has a sorted array of superblocks, which expands
80 red zone bytes (depends on Arena.rz_szB, but >= sizeof(void*))
82 red zone bytes (depends on Arena.rz_szB, but >= sizeof(void*))
90 excess red zone bytes (if Arena.rz_szB > sizeof(void*))
92 excess red zone bytes (if Arena.rz_szB > sizeof(void*))
174 // of the same arena can be reclaimed or till a new superblock is needed
175 // in any arena.
191 // An arena. 'freelist' is a circular, doubly-linked list. 'rz_szB' is
211 // superblocks in the arena. If this array is expanded, which
247 Arena;
510 ArenaId arena = a -vg_arena; local
799 Arena* arena = arenaId_to_ArenaP(aid); local
1447 Arena *arena; local
    [all...]
  /external/chromium_org/third_party/tcmalloc/chromium/src/
type_profiler_map.cc 51 LowLevelAlloc::Arena* g_type_profiler_map_memory = NULL;
memory_region_map.h 250 // We allocate STL objects in our own arena.
296 // Arena used for our allocations in regions_.
297 static LowLevelAlloc::Arena* arena_;
  /prebuilts/python/darwin-x86/2.7.5/lib/python2.7/multiprocessing/
heap.py 57 class Arena(object):
63 self.name = 'pym-%d-%d' % (os.getpid(), Arena._counter.next())
79 class Arena(object):
120 arena = Arena(length)
121 self._arenas.append(arena)
122 return (arena, 0, length)
130 (arena, start, stop) = block
131 del self._start_to_block[(arena, start)]
132 del self._stop_to_block[(arena, stop)
    [all...]
  /prebuilts/python/linux-x86/2.7.5/lib/python2.7/multiprocessing/
heap.py 57 class Arena(object):
63 self.name = 'pym-%d-%d' % (os.getpid(), Arena._counter.next())
79 class Arena(object):
120 arena = Arena(length)
121 self._arenas.append(arena)
122 return (arena, 0, length)
130 (arena, start, stop) = block
131 del self._start_to_block[(arena, start)]
132 del self._stop_to_block[(arena, stop)
    [all...]
  /external/chromium_org/third_party/tcmalloc/chromium/src/tests/
low_level_alloc_unittest.cc 77 // If use_new_arena is true, use a fresh arena, and then delete it.
87 LowLevelAlloc::Arena *arena = 0; local
90 arena = LowLevelAlloc::NewArena(flags, LowLevelAlloc::DefaultArena());
104 arena == 0
106 : LowLevelAlloc::AllocWithArena(block_desc.len, arena));
142 CHECK(LowLevelAlloc::DeleteArena(arena));
177 CHECK_NE(allocates, 0); // default arena calls hooks
185 CHECK_GT(allocates, 5000); // arena calls hooks
188 CHECK_EQ(allocates, 0); // arena doesn't call hook
    [all...]
  /external/chromium_org/third_party/tcmalloc/vendor/src/
memory_region_map.h 226 // We allocate STL objects in our own arena.
272 // Arena used for our allocations in regions_.
273 static LowLevelAlloc::Arena* arena_;

Completed in 1275 milliseconds

1 2