Home | History | Annotate | Download | only in gc

Lines Matching refs:alloc_size

184                                            size_t alloc_size, size_t* bytes_allocated,
187 UNLIKELY(IsOutOfMemoryOnAllocation<kGrow>(allocator_type, alloc_size))) {
194 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment);
195 ret = bump_pointer_space_->AllocNonvirtual(alloc_size);
197 *bytes_allocated = alloc_size;
198 *usable_size = alloc_size;
205 ret = rosalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size);
208 ret = rosalloc_space_->AllocNonvirtual(self, alloc_size, bytes_allocated, usable_size);
215 ret = dlmalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size);
218 ret = dlmalloc_space_->AllocNonvirtual(self, alloc_size, bytes_allocated, usable_size);
223 ret = non_moving_space_->Alloc(self, alloc_size, bytes_allocated, usable_size);
227 ret = large_object_space_->Alloc(self, alloc_size, bytes_allocated, usable_size);
235 DCHECK_ALIGNED(alloc_size, space::BumpPointerSpace::kAlignment);
236 if (UNLIKELY(self->TlabSize() < alloc_size)) {
237 const size_t new_tlab_size = alloc_size + kDefaultTLABSize;
251 ret = self->AllocTlab(alloc_size);
253 *usable_size = alloc_size;
291 inline bool Heap::IsOutOfMemoryOnAllocation(AllocatorType allocator_type, size_t alloc_size) {
292 size_t new_footprint = num_bytes_allocated_.LoadSequentiallyConsistent() + alloc_size;
303 << PrettySize(new_footprint) << " for a " << PrettySize(alloc_size) << " allocation";