HomeSort by relevance Sort by last modified time
    Searched refs:usable_size (Results 1 - 25 of 38) sorted by null

1 2

  /art/runtime/gc/space/
dlmalloc_space-inl.h 30 size_t* usable_size,
35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
45 inline size_t DlMallocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) {
48 if (usable_size != nullptr) {
49 *usable_size = size;
57 size_t* usable_size,
65 size_t allocation_size = AllocationSizeNonvirtual(result, usable_size);
memory_tool_malloc_space-inl.h 32 size_t bytes_allocated, size_t usable_size,
49 *usable_size_out = usable_size - 2 * kMemoryToolRedZoneBytes;
63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is
65 // At the moment, this fits RosAlloc (no management data in a slot, usable_size == alloc_size)
66 // and DlMalloc (allocation_size = (usable_size == num_bytes) + 4, 4 is management)
68 usable_size - (num_bytes + kMemoryToolRedZoneBytes));
91 size_t usable_size; local
94 &bytes_allocated, &usable_size,
102 bytes_allocated, usable_size,
120 size_t usable_size; local
148 size_t usable_size; local
201 size_t usable_size; local
    [all...]
bump_pointer_space-inl.h 28 size_t* usable_size,
34 if (usable_size != nullptr) {
35 *usable_size = num_bytes;
44 size_t* usable_size,
58 if (UNLIKELY(usable_size != nullptr)) {
59 *usable_size = num_bytes;
89 inline size_t BumpPointerSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
92 if (usable_size != nullptr) {
93 *usable_size = RoundUp(num_bytes, kAlignment);
memory_tool_malloc_space.h 35 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
38 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE;
40 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
43 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
rosalloc_space-inl.h 31 inline size_t RosAllocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) {
57 if (usable_size != nullptr) {
58 *usable_size = size_by_size;
65 size_t* bytes_allocated, size_t* usable_size,
85 if (usable_size != nullptr) {
86 *usable_size = rosalloc_usable_size;
rosalloc_space.h 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
53 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE {
54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size,
58 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size,
63 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE {
64 return AllocationSizeNonvirtual<true>(obj, usable_size);
72 size_t* usable_size, size_t* bytes_tl_bulk_allocated) {
74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size,
78 size_t* bytes_allocated, size_t* usable_size,
    [all...]
dlmalloc_space.h 51 size_t* usable_size,
56 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size,
62 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE {
63 return AllocationSizeNonvirtual(obj, usable_size);
88 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
92 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size);
148 size_t* usable_size,
region_space-inl.h 28 size_t* usable_size,
31 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size,
37 size_t* usable_size,
40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
45 size_t* usable_size,
52 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size,
56 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size,
65 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size,
68 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size,
85 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated)
    [all...]
bump_pointer_space.h 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE;
53 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
60 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE
62 return AllocationSizeNonvirtual(obj, usable_size);
74 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
zygote_space.h 49 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE;
51 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
malloc_space.h 57 size_t* bytes_allocated, size_t* usable_size,
61 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0;
62 // Return the storage space required by obj. If usable_size isn't null then it is set to the
64 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
large_object_space.cc 52 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
56 usable_size, bytes_tl_bulk_allocated);
63 if (usable_size != nullptr) {
64 *usable_size = num_bytes; // Since we have redzones, shrink the usable size.
69 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE {
70 return LargeObjectMapSpace::AllocationSize(ObjectWithRedzone(obj), usable_size);
133 size_t* bytes_allocated, size_t* usable_size,
154 if (usable_size != nullptr) {
155 *usable_size = allocation_size;
199 size_t LargeObjectMapSpace::AllocationSize(mirror::Object* obj, size_t* usable_size) {
    [all...]
large_object_space.h 137 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) REQUIRES(!lock_);
139 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
172 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE
175 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
space.h 216 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0;
220 size_t* usable_size,
223 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
227 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
space_create_test.cc 315 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local
321 &usable_size,
326 EXPECT_EQ(usable_size, computed_usable_size);
336 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local
341 &usable_size,
346 EXPECT_EQ(usable_size, computed_usable_size);
region_space.h 46 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
50 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
55 size_t* usable_size,
60 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size,
65 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE
67 return AllocationSizeNonvirtual(obj, usable_size);
69 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
299 size_t* usable_size,
space_test.h 72 size_t* usable_size,
80 usable_size,
92 size_t* usable_size,
97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size,
  /art/runtime/gc/
heap-inl.h 78 size_t usable_size; local
92 usable_size = bytes_allocated;
93 pre_fence_visitor(obj, usable_size);
104 usable_size = bytes_allocated;
105 pre_fence_visitor(obj, usable_size);
111 &usable_size, &bytes_tl_bulk_allocated);
120 &usable_size,
137 DCHECK_GT(usable_size, 0u);
154 pre_fence_visitor(obj, usable_size);
160 CHECK_LE(obj->SizeOf(), usable_size);
    [all...]
  /bionic/libc/malloc_debug/
malloc_debug.h 55 size_t usable_size; member in struct:Header
FreeTrackData.cpp 49 for (size_t i = 0; i < header->usable_size; i++) {
71 size_t bytes = header->usable_size;
malloc_debug.cpp 143 header->usable_size = g_dispatch->malloc_usable_size(orig_pointer);
144 if (header->usable_size == 0) {
148 header->usable_size -= g_debug->pointer_offset() +
161 header->usable_size = header->real_size();
275 return header->usable_size;
376 bytes = header->usable_size;
538 if (real_size < header->usable_size) {
545 // change the header usable_size and reset the rear guard.
546 header->usable_size = header->real_size();
561 prev_size = header->usable_size;
    [all...]
  /art/runtime/gc/allocator/
rosalloc-inl.h 32 size_t* usable_size,
35 return AllocLargeObject(self, size, bytes_allocated, usable_size,
40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size,
  /art/runtime/mirror/
string-inl.h 49 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
69 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
105 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
136 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
array-inl.h 106 void operator()(ObjPtr<Object> obj, size_t usable_size ATTRIBUTE_UNUSED) const
130 void operator()(ObjPtr<Object> obj, size_t usable_size) const
135 int32_t length = (usable_size - header_size_) >> component_size_shift_;
  /bionic/libc/malloc_debug/tests/
malloc_debug_unit_tests.cpp 221 size_t usable_size = debug_malloc_usable_size(pointer); local
222 memset(pointer, 0, usable_size);
225 for (size_t i = 0; i < usable_size; i++) {
238 size_t usable_size = debug_malloc_usable_size(pointer); local
239 memset(pointer, 0, usable_size);
245 for (size_t i = 30; i < usable_size; i++) {
258 size_t usable_size = debug_malloc_usable_size(pointer); local
259 memset(pointer, 0, usable_size);
265 for (size_t i = 30; i < usable_size; i++) {
1230 size_t usable_size = debug_malloc_usable_size(pointer); local
    [all...]

Completed in 627 milliseconds

1 2