/art/runtime/gc/space/ |
rosalloc_space-inl.h | 33 size_t* bytes_allocated, size_t* usable_size, 50 DCHECK(bytes_allocated != nullptr); 51 *bytes_allocated = rosalloc_bytes_allocated; 67 size_t* bytes_allocated) { 68 DCHECK(bytes_allocated != nullptr); 70 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated));
|
dlmalloc_space-inl.h | 29 size_t* bytes_allocated, 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 56 size_t* bytes_allocated, 66 DCHECK(bytes_allocated != nullptr); 67 *bytes_allocated = allocation_size;
|
bump_pointer_space-inl.h | 28 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, 34 *bytes_allocated = num_bytes; 44 size_t* bytes_allocated, 55 *bytes_allocated = num_bytes;
|
memory_tool_malloc_space.h | 34 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 37 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 39 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
memory_tool_malloc_space-inl.h | 32 size_t bytes_allocated, size_t usable_size, 37 *bytes_allocated_out = bytes_allocated; 63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is 90 size_t bytes_allocated; local 94 &bytes_allocated, &usable_size, 102 bytes_allocated, usable_size, 119 size_t bytes_allocated; local 123 &bytes_allocated, &usable_size, &bytes_tl_bulk_allocated); 130 bytes_allocated, usable_size, 147 size_t bytes_allocated; local [all...] |
rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 78 size_t* bytes_allocated, size_t* usable_size, 81 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size, 91 size_t* bytes_allocated); [all...] |
region_space-inl.h | 27 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, 31 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, 36 size_t* bytes_allocated, 40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 44 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, 52 bytes_allocated, 61 bytes_allocated, 69 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 82 obj = AllocLarge<kForEvac>(num_bytes, bytes_allocated, usable_size, 91 inline mirror::Object* RegionSpace::Region::Alloc(size_t num_bytes, size_t* bytes_allocated, [all...] |
dlmalloc_space.h | 50 virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 55 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 87 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, 147 mirror::Object* AllocWithoutGrowthLocked(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
space_test.h | 71 size_t* bytes_allocated, 79 bytes_allocated, 91 size_t* bytes_allocated, 97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size, 192 size_t bytes_allocated = 0; local 195 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr, 198 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr, 206 EXPECT_EQ(bytes_allocated, allocation_size); 290 size_t bytes_allocated = 0; local 293 large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullptr [all...] |
bump_pointer_space.h | 53 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 56 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
malloc_space.h | 57 size_t* bytes_allocated, size_t* usable_size, 60 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
zygote_space.h | 48 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
space.h | 202 // succeeds, the output parameter bytes_allocated will be set to the 208 // 1) equal to bytes_allocated if it's not a thread local allocation, 209 // 2) greater than bytes_allocated if it's a thread local 214 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 218 virtual mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 222 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
|
dlmalloc_space.cc | 130 size_t* bytes_allocated, size_t* usable_size, 139 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 267 size_t bytes_allocated = 0; local 268 mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated); 269 return bytes_allocated;
|
rosalloc_space.cc | 156 size_t* bytes_allocated, size_t* usable_size, 165 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size, 291 size_t bytes_allocated = 0; local 292 InspectAllRosAlloc(art::gc::allocator::RosAlloc::BytesAllocatedCallback, &bytes_allocated, false); 293 return bytes_allocated;
|
region_space.h | 50 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 54 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 59 ALWAYS_INLINE mirror::Object* AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, 65 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, 67 void FreeLarge(mirror::Object* large_obj, size_t bytes_allocated) REQUIRES(!region_lock_); 290 ALWAYS_INLINE mirror::Object* Alloc(size_t num_bytes, size_t* bytes_allocated,
|
large_object_space.cc | 54 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated, 136 size_t* bytes_allocated, size_t* usable_size, 149 DCHECK(bytes_allocated != nullptr); 156 *bytes_allocated = allocation_size; 472 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 506 DCHECK(bytes_allocated != nullptr); 507 *bytes_allocated = allocation_size;
|
/art/runtime/gc/ |
heap-inl.h | 78 size_t bytes_allocated; local 92 bytes_allocated = byte_count; 93 usable_size = bytes_allocated; 98 (obj = rosalloc_space_->AllocThreadLocal(self, byte_count, &bytes_allocated)) != nullptr && 105 usable_size = bytes_allocated; 111 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, 120 &bytes_allocated, 137 DCHECK_GT(bytes_allocated, 0u); 173 thread_stats->allocated_bytes += bytes_allocated; 176 global_stats->allocated_bytes += bytes_allocated; [all...] |
/art/runtime/gc/allocator/ |
rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size, 79 size_t* bytes_allocated) { 80 DCHECK(bytes_allocated != nullptr); 97 *bytes_allocated = bracket_size;
|
dlmalloc.cc | 89 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local 90 *bytes_allocated += used_bytes + sizeof(size_t);
|
/external/v8/src/heap/ |
scavenge-job.h | 40 void ScheduleIdleTaskIfNeeded(Heap* heap, int bytes_allocated);
|
scavenge-job.cc | 90 void ScavengeJob::ScheduleIdleTaskIfNeeded(Heap* heap, int bytes_allocated) { 91 bytes_allocated_since_the_last_task_ += bytes_allocated;
|
/art/runtime/gc/accounting/ |
mod_union_table_test.cc | 53 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 55 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated)); 60 EXPECT_GE(bytes_allocated, size); 82 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 83 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated,
|
/art/test/099-vmdebug/src/ |
Main.java | 183 String bytes_allocated = VMDebug.getRuntimeStat("art.gc.bytes-allocated"); local 192 checkNumber(bytes_allocated); 210 String bytes_allocated = map.get("art.gc.bytes-allocated"); local 219 checkNumber(bytes_allocated);
|
/cts/tests/tests/os/src/android/os/cts/ |
DebugTest.java | 262 String bytes_allocated = Debug.getRuntimeStat("art.gc.bytes-allocated"); local 271 checkNumber(bytes_allocated); 289 String bytes_allocated = map.get("art.gc.bytes-allocated"); local 298 checkNumber(bytes_allocated);
|