/art/runtime/gc/space/ |
dlmalloc_space-inl.h | 30 size_t* usable_size, 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 45 inline size_t DlMallocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { 48 if (usable_size != nullptr) { 49 *usable_size = size; 57 size_t* usable_size, 65 size_t allocation_size = AllocationSizeNonvirtual(result, usable_size);
|
memory_tool_malloc_space-inl.h | 32 size_t bytes_allocated, size_t usable_size, 49 *usable_size_out = usable_size - 2 * kMemoryToolRedZoneBytes; 63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is 65 // At the moment, this fits RosAlloc (no management data in a slot, usable_size == alloc_size) 66 // and DlMalloc (allocation_size = (usable_size == num_bytes) + 4, 4 is management) 68 usable_size - (num_bytes + kMemoryToolRedZoneBytes)); 91 size_t usable_size; local 94 &bytes_allocated, &usable_size, 102 bytes_allocated, usable_size, 120 size_t usable_size; local 148 size_t usable_size; local 201 size_t usable_size; local [all...] |
bump_pointer_space-inl.h | 28 size_t* usable_size, 34 if (usable_size != nullptr) { 35 *usable_size = num_bytes; 44 size_t* usable_size, 58 if (UNLIKELY(usable_size != nullptr)) { 59 *usable_size = num_bytes; 89 inline size_t BumpPointerSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) 92 if (usable_size != nullptr) { 93 *usable_size = RoundUp(num_bytes, kAlignment);
|
memory_tool_malloc_space.h | 35 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 38 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE; 40 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 43 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
|
rosalloc_space-inl.h | 31 inline size_t RosAllocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { 57 if (usable_size != nullptr) { 58 *usable_size = size_by_size; 65 size_t* bytes_allocated, size_t* usable_size, 85 if (usable_size != nullptr) { 86 *usable_size = rosalloc_usable_size;
|
rosalloc_space.h | 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 53 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE { 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 58 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 63 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 64 return AllocationSizeNonvirtual<true>(obj, usable_size); 72 size_t* usable_size, size_t* bytes_tl_bulk_allocated) { 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 78 size_t* bytes_allocated, size_t* usable_size, [all...] |
dlmalloc_space.h | 51 size_t* usable_size, 56 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 62 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 63 return AllocationSizeNonvirtual(obj, usable_size); 88 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 92 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size); 148 size_t* usable_size,
|
region_space-inl.h | 28 size_t* usable_size, 31 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, 37 size_t* usable_size, 40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 45 size_t* usable_size, 52 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 56 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 65 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 68 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 85 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated) [all...] |
bump_pointer_space.h | 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE; 53 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 60 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE 62 return AllocationSizeNonvirtual(obj, usable_size); 74 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
|
zygote_space.h | 49 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE; 51 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
|
malloc_space.h | 57 size_t* bytes_allocated, size_t* usable_size, 61 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0; 62 // Return the storage space required by obj. If usable_size isn't null then it is set to the 64 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
|
large_object_space.cc | 52 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 56 usable_size, bytes_tl_bulk_allocated); 63 if (usable_size != nullptr) { 64 *usable_size = num_bytes; // Since we have redzones, shrink the usable size. 69 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 70 return LargeObjectMapSpace::AllocationSize(ObjectWithRedzone(obj), usable_size); 133 size_t* bytes_allocated, size_t* usable_size, 154 if (usable_size != nullptr) { 155 *usable_size = allocation_size; 199 size_t LargeObjectMapSpace::AllocationSize(mirror::Object* obj, size_t* usable_size) { [all...] |
large_object_space.h | 137 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) REQUIRES(!lock_); 139 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 172 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE 175 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
|
space.h | 216 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0; 220 size_t* usable_size, 223 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 227 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
|
space_create_test.cc | 315 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local 321 &usable_size, 326 EXPECT_EQ(usable_size, computed_usable_size); 336 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local 341 &usable_size, 346 EXPECT_EQ(usable_size, computed_usable_size);
|
region_space.h | 46 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated) 55 size_t* usable_size, 60 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, 65 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE 67 return AllocationSizeNonvirtual(obj, usable_size); 69 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) 299 size_t* usable_size,
|
space_test.h | 72 size_t* usable_size, 80 usable_size, 92 size_t* usable_size, 97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size,
|
/art/runtime/gc/ |
heap-inl.h | 78 size_t usable_size; local 92 usable_size = bytes_allocated; 93 pre_fence_visitor(obj, usable_size); 104 usable_size = bytes_allocated; 105 pre_fence_visitor(obj, usable_size); 111 &usable_size, &bytes_tl_bulk_allocated); 120 &usable_size, 137 DCHECK_GT(usable_size, 0u); 154 pre_fence_visitor(obj, usable_size); 160 CHECK_LE(obj->SizeOf(), usable_size); [all...] |
/bionic/libc/malloc_debug/ |
malloc_debug.h | 55 size_t usable_size; member in struct:Header
|
FreeTrackData.cpp | 49 for (size_t i = 0; i < header->usable_size; i++) { 71 size_t bytes = header->usable_size;
|
malloc_debug.cpp | 143 header->usable_size = g_dispatch->malloc_usable_size(orig_pointer); 144 if (header->usable_size == 0) { 148 header->usable_size -= g_debug->pointer_offset() + 161 header->usable_size = header->real_size(); 275 return header->usable_size; 376 bytes = header->usable_size; 538 if (real_size < header->usable_size) { 545 // change the header usable_size and reset the rear guard. 546 header->usable_size = header->real_size(); 561 prev_size = header->usable_size; [all...] |
/art/runtime/gc/allocator/ |
rosalloc-inl.h | 32 size_t* usable_size, 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size,
|
/art/runtime/mirror/ |
string-inl.h | 49 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 69 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 105 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 136 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
|
array-inl.h | 106 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const 130 void operator()(ObjPtr<Object> obj, size_t usable_size) const 135 int32_t length = (usable_size - header_size_) >> component_size_shift_;
|
/bionic/libc/malloc_debug/tests/ |
malloc_debug_unit_tests.cpp | 221 size_t usable_size = debug_malloc_usable_size(pointer); local 222 memset(pointer, 0, usable_size); 225 for (size_t i = 0; i < usable_size; i++) { 238 size_t usable_size = debug_malloc_usable_size(pointer); local 239 memset(pointer, 0, usable_size); 245 for (size_t i = 30; i < usable_size; i++) { 258 size_t usable_size = debug_malloc_usable_size(pointer); local 259 memset(pointer, 0, usable_size); 265 for (size_t i = 30; i < usable_size; i++) { 1230 size_t usable_size = debug_malloc_usable_size(pointer); local [all...] |