HomeSort by relevance Sort by last modified time
    Searched refs:usable_size (Results 1 - 25 of 37) sorted by null

1 2

  /art/runtime/gc/space/
dlmalloc_space-inl.h 30 size_t* usable_size,
35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
45 inline size_t DlMallocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) {
48 if (usable_size != nullptr) {
49 *usable_size = size;
57 size_t* usable_size,
65 size_t allocation_size = AllocationSizeNonvirtual(result, usable_size);
memory_tool_malloc_space-inl.h 32 size_t bytes_allocated, size_t usable_size,
49 *usable_size_out = usable_size - 2 * kMemoryToolRedZoneBytes;
63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is
65 // At the moment, this fits RosAlloc (no management data in a slot, usable_size == alloc_size)
66 // and DlMalloc (allocation_size = (usable_size == num_bytes) + 4, 4 is management)
68 usable_size - (num_bytes + kMemoryToolRedZoneBytes));
91 size_t usable_size; local
94 &bytes_allocated, &usable_size,
102 bytes_allocated, usable_size,
120 size_t usable_size; local
148 size_t usable_size; local
201 size_t usable_size; local
    [all...]
bump_pointer_space-inl.h 28 size_t* usable_size,
34 if (usable_size != nullptr) {
35 *usable_size = num_bytes;
44 size_t* usable_size,
58 if (UNLIKELY(usable_size != nullptr)) {
59 *usable_size = num_bytes;
89 inline size_t BumpPointerSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
92 if (usable_size != nullptr) {
93 *usable_size = RoundUp(num_bytes, kAlignment);
region_space-inl.h 27 size_t* usable_size,
30 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size,
36 size_t* usable_size,
39 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
44 size_t* usable_size,
51 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size,
55 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size,
64 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size,
67 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size,
84 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated)
    [all...]
memory_tool_malloc_space.h 35 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
38 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE;
40 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
43 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
rosalloc_space-inl.h 31 inline size_t RosAllocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) {
57 if (usable_size != nullptr) {
58 *usable_size = size_by_size;
65 size_t* bytes_allocated, size_t* usable_size,
85 if (usable_size != nullptr) {
86 *usable_size = rosalloc_usable_size;
rosalloc_space.h 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
53 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE {
54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size,
58 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size,
63 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE {
64 return AllocationSizeNonvirtual<true>(obj, usable_size);
72 size_t* usable_size, size_t* bytes_tl_bulk_allocated) {
74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size,
78 size_t* bytes_allocated, size_t* usable_size,
    [all...]
dlmalloc_space.h 51 size_t* usable_size,
56 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size,
62 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE {
63 return AllocationSizeNonvirtual(obj, usable_size);
88 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
92 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size);
148 size_t* usable_size,
bump_pointer_space.h 50 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE;
53 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
60 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE
62 return AllocationSizeNonvirtual(obj, usable_size);
74 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
zygote_space.h 49 size_t* usable_size, size_t* bytes_tl_bulk_allocated) OVERRIDE;
51 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE;
malloc_space.h 57 size_t* bytes_allocated, size_t* usable_size,
61 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0;
62 // Return the storage space required by obj. If usable_size isn't null then it is set to the
64 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
large_object_space.cc 53 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
57 usable_size, bytes_tl_bulk_allocated);
64 if (usable_size != nullptr) {
65 *usable_size = num_bytes; // Since we have redzones, shrink the usable size.
70 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE {
71 return LargeObjectMapSpace::AllocationSize(ObjectWithRedzone(obj), usable_size);
134 size_t* bytes_allocated, size_t* usable_size,
164 if (usable_size != nullptr) {
165 *usable_size = allocation_size;
209 size_t LargeObjectMapSpace::AllocationSize(mirror::Object* obj, size_t* usable_size) {
    [all...]
large_object_space.h 133 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) REQUIRES(!lock_);
135 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
166 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE
169 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
space.h 216 size_t* usable_size, size_t* bytes_tl_bulk_allocated) = 0;
220 size_t* usable_size,
223 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
227 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) = 0;
space_create_test.cc 315 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local
321 &usable_size,
326 EXPECT_EQ(usable_size, computed_usable_size);
336 size_t allocation_size, usable_size, bytes_tl_bulk_allocated; local
341 &usable_size,
346 EXPECT_EQ(usable_size, computed_usable_size);
region_space.h 45 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
49 size_t* usable_size, size_t* bytes_tl_bulk_allocated)
54 size_t* usable_size,
59 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size,
64 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE
66 return AllocationSizeNonvirtual(obj, usable_size);
68 size_t AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size)
276 size_t* usable_size,
space_test.h 72 size_t* usable_size,
80 usable_size,
92 size_t* usable_size,
97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size,
  /art/runtime/gc/
heap-inl.h 71 size_t usable_size; local
89 usable_size = bytes_allocated;
90 pre_fence_visitor(obj, usable_size);
103 usable_size = bytes_allocated;
104 pre_fence_visitor(obj, usable_size);
110 &usable_size, &bytes_tl_bulk_allocated);
119 &usable_size,
136 DCHECK_GT(usable_size, 0u);
156 pre_fence_visitor(obj, usable_size);
162 CHECK_LE(obj->SizeOf(), usable_size);
    [all...]
  /bionic/libc/malloc_debug/
malloc_debug.h 55 size_t usable_size; member in struct:Header
malloc_debug.cpp 143 header->usable_size = g_dispatch->malloc_usable_size(orig_pointer);
144 if (header->usable_size == 0) {
148 header->usable_size -= g_debug->pointer_offset() +
161 header->usable_size = header->real_size();
278 return header->usable_size;
360 bytes = header->usable_size;
493 if (real_size < header->usable_size) {
500 // change the header usable_size and reset the rear guard.
501 header->usable_size = header->real_size();
515 prev_size = header->usable_size;
    [all...]
FreeTrackData.cpp 52 for (size_t i = 0; i < header->usable_size; i++) {
76 size_t bytes = header->usable_size;
  /art/runtime/gc/allocator/
rosalloc-inl.h 32 size_t* usable_size,
35 return AllocLargeObject(self, size, bytes_allocated, usable_size,
40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size,
  /art/runtime/mirror/
string-inl.h 46 void operator()(Object* obj, size_t usable_size ATTRIBUTE_UNUSED) const
65 void operator()(Object* obj, size_t usable_size ATTRIBUTE_UNUSED) const
92 void operator()(Object* obj, size_t usable_size ATTRIBUTE_UNUSED) const
115 void operator()(Object* obj, size_t usable_size ATTRIBUTE_UNUSED) const
array-inl.h 103 void operator()(Object* obj, size_t usable_size ATTRIBUTE_UNUSED) const
127 void operator()(Object* obj, size_t usable_size) const
132 int32_t length = (usable_size - header_size_) >> component_size_shift_;
  /bionic/libc/malloc_debug/tests/
malloc_debug_unit_tests.cpp 216 size_t usable_size = debug_malloc_usable_size(pointer); local
217 memset(pointer, 0, usable_size);
220 for (size_t i = 0; i < usable_size; i++) {
233 size_t usable_size = debug_malloc_usable_size(pointer); local
234 memset(pointer, 0, usable_size);
240 for (size_t i = 30; i < usable_size; i++) {
253 size_t usable_size = debug_malloc_usable_size(pointer); local
254 memset(pointer, 0, usable_size);
260 for (size_t i = 30; i < usable_size; i++) {
1225 size_t usable_size = debug_malloc_usable_size(pointer); local
    [all...]

Completed in 188 milliseconds

1 2