/art/runtime/gc/space/ |
dlmalloc_space-inl.h | 30 size_t* usable_size, 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 45 inline size_t DlMallocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { 48 if (usable_size != nullptr) { 49 *usable_size = size; 57 size_t* usable_size, 65 size_t allocation_size = AllocationSizeNonvirtual(result, usable_size);
|
memory_tool_malloc_space-inl.h | 32 size_t bytes_allocated, size_t usable_size, 49 *usable_size_out = usable_size - 2 * kMemoryToolRedZoneBytes; 63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is 65 // At the moment, this fits RosAlloc (no management data in a slot, usable_size == alloc_size) 66 // and DlMalloc (allocation_size = (usable_size == num_bytes) + 4, 4 is management) 68 usable_size - (num_bytes + kMemoryToolRedZoneBytes)); 91 size_t usable_size; local 94 &bytes_allocated, &usable_size, 102 bytes_allocated, usable_size, 120 size_t usable_size; local 148 size_t usable_size; local 201 size_t usable_size; local [all...] |
bump_pointer_space-inl.h | 29 size_t* usable_size, 35 if (usable_size != nullptr) { 36 *usable_size = num_bytes; 45 size_t* usable_size, 59 if (UNLIKELY(usable_size != nullptr)) { 60 *usable_size = num_bytes;
|
memory_tool_malloc_space.h | 35 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 38 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE; 40 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 43 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
|
rosalloc_space.h | 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 53 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE { 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 58 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 63 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 64 return AllocationSizeNonvirtual<true>(obj, usable_size); 72 size_t* usable_size, size_t* bytes_tl_bulk_allocated) { 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 78 size_t* bytes_allocated, size_t* usable_size, [all...] |
dlmalloc_space.h | 51 size_t* usable_size, 56 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 62 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 63 return AllocationSizeNonvirtual(obj, usable_size); 88 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 92 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size); 148 size_t* usable_size,
|
rosalloc_space-inl.h | 33 size_t* bytes_allocated, size_t* usable_size, 53 if (usable_size != nullptr) { 54 *usable_size = rosalloc_usable_size;
|
region_space-inl.h | 28 size_t* usable_size, 31 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, 37 size_t* usable_size, 40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 45 size_t* usable_size, 53 usable_size, 62 usable_size, 69 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 82 obj = AllocLarge<kForEvac>(num_bytes, bytes_allocated, usable_size, 92 size_t* usable_size, [all...] |
bump_pointer_space.h | 54 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE; 57 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 64 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE 66 return AllocationSizeNonvirtual(obj, usable_size); 78 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
|
zygote_space.h | 49 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE; 51 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
|
malloc_space.h | 57 size_t* bytes_allocated, size_t* usable_size, 61 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0; 62 // Return the storage space required by obj. If usable_size isn't null then it is set to the 64 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
|
large_object_space.cc | 55 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 59 usable_size, bytes_tl_bulk_allocated); 66 if (usable_size != nullptr) { 67 *usable_size = num_bytes; // Since we have redzones, shrink the usable size. 72 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 73 return LargeObjectMapSpace::AllocationSize(ObjectWithRedzone(obj), usable_size); 136 size_t* bytes_allocated, size_t* usable_size, 157 if (usable_size != nullptr) { 158 *usable_size = allocation_size; 202 size_t LargeObjectMapSpace::AllocationSize(mirror::Object* obj, size_t* usable_size) { [all...] |
large_object_space.h | 137 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) REQUIRES(!lock_); 139 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 172 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE 175 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
|
space.h | 215 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0; 219 size_t* usable_size, 222 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 226 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
|
space_create_test.cc | 315 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local 321 &usable_size, 326 EXPECT_EQ(usable_size, computed_usable_size); 336 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local 341 &usable_size, 346 EXPECT_EQ(usable_size, computed_usable_size);
|
region_space.h | 51 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 55 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 60 size_t* usable_size, 65 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, 70 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE 72 return AllocationSizeNonvirtual(obj, usable_size); 74 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) 291 size_t* usable_size,
|
bump_pointer_space.cc | 222 size_t BumpPointerSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { 224 if (usable_size != nullptr) { 225 *usable_size = RoundUp(num_bytes, kAlignment);
|
rosalloc_space.cc | 156 size_t* bytes_allocated, size_t* usable_size, 165 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size, 378 size_t RosAllocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { 404 if (usable_size != nullptr) { 405 *usable_size = size_by_size;
|
space_test.h | 72 size_t* usable_size, 80 usable_size, 92 size_t* usable_size, 97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size,
|
/art/runtime/gc/ |
heap-inl.h | 79 size_t usable_size; local 93 usable_size = bytes_allocated; 94 pre_fence_visitor(obj, usable_size); 105 usable_size = bytes_allocated; 106 pre_fence_visitor(obj, usable_size); 112 &usable_size, &bytes_tl_bulk_allocated); 121 &usable_size, 138 DCHECK_GT(usable_size, 0u); 155 pre_fence_visitor(obj, usable_size); 166 CHECK_LE(obj->SizeOf(), usable_size); [all...] |
/bionic/libc/malloc_debug/ |
malloc_debug.h | 55 size_t usable_size; member in struct:Header
|
FreeTrackData.cpp | 49 for (size_t i = 0; i < header->usable_size; i++) { 71 size_t bytes = header->usable_size;
|
malloc_debug.cpp | 143 header->usable_size = g_dispatch->malloc_usable_size(orig_pointer); 144 if (header->usable_size == 0) { 148 header->usable_size -= g_debug->pointer_offset() + 161 header->usable_size = header->real_size(); 275 return header->usable_size; 376 bytes = header->usable_size; 538 if (real_size < header->usable_size) { 545 // change the header usable_size and reset the rear guard. 546 header->usable_size = header->real_size(); 561 prev_size = header->usable_size; [all...] |
/art/runtime/gc/allocator/ |
rosalloc-inl.h | 32 size_t* usable_size, 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size,
|
/art/runtime/mirror/ |
string-inl.h | 48 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 68 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 104 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 135 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
|