Home | History | Annotate | Download | only in src

Lines Matching refs:size_in_bytes

982 HeapObject* PagedSpace::SlowMCAllocateRaw(int size_in_bytes) {
1004 return AllocateLinearly(&mc_forwarding_info_, size_in_bytes);
1743 void FreeListNode::set_size(Heap* heap, int size_in_bytes) {
1744 ASSERT(size_in_bytes > 0);
1745 ASSERT(IsAligned(size_in_bytes, kPointerSize));
1754 if (size_in_bytes > ByteArray::kHeaderSize) {
1758 this_as_byte_array->set_length(ByteArray::LengthFor(size_in_bytes));
1759 } else if (size_in_bytes == kPointerSize) {
1761 } else if (size_in_bytes == 2 * kPointerSize) {
1766 // We would like to ASSERT(Size() == size_in_bytes) but this would fail during
1825 int OldSpaceFreeList::Free(Address start, int size_in_bytes) {
1827 Isolate::Current()->memory_allocator()->ZapBlock(start, size_in_bytes);
1830 node->set_size(heap_, size_in_bytes);
1836 return size_in_bytes;
1842 if (size_in_bytes < kMinBlockSize) {
1843 return size_in_bytes;
1847 int index = size_in_bytes >> kPointerSizeLog2;
1850 available_ += size_in_bytes;
1856 MaybeObject* OldSpaceFreeList::Allocate(int size_in_bytes, int* wasted_bytes) {
1857 ASSERT(0 < size_in_bytes);
1858 ASSERT(size_in_bytes <= kMaxBlockSize);
1859 ASSERT(IsAligned(size_in_bytes, kPointerSize));
1862 int index = size_in_bytes >> kPointerSizeLog2;
1869 available_ -= size_in_bytes;
1889 size_in_bytes);
1915 available_ -= size_in_bytes + rem_bytes;
1925 available_ -= size_in_bytes;
2158 int size_in_bytes = static_cast<int>(PageAllocationLimit(last_in_use) -
2162 if (size_in_bytes > 0) {
2165 accounting_stats_.AllocateBytes(size_in_bytes);
2166 DeallocateBlock(start, size_in_bytes, add_to_freelist);
2168 heap()->CreateFillerObjectAt(start, size_in_bytes);
2186 int size_in_bytes = static_cast<int>(PageAllocationLimit(p) -
2192 accounting_stats_.AllocateBytes(size_in_bytes);
2193 DeallocateBlock(start, size_in_bytes, add_to_freelist);
2195 heap()->CreateFillerObjectAt(start, size_in_bytes);
2249 HeapObject* OldSpace::SlowAllocateRaw(int size_in_bytes) {
2252 // should succeed (size_in_bytes should not be greater than a page's
2256 return AllocateInNextPage(current_page, size_in_bytes);
2264 MaybeObject* maybe = free_list_.Allocate(size_in_bytes, &wasted_bytes);
2267 accounting_stats_.AllocateBytes(size_in_bytes);
2278 p->SetAllocationWatermark(obj->address() + size_in_bytes);
2296 return AllocateInNextPage(current_page, size_in_bytes);
2334 int size_in_bytes) {
2340 return AllocateLinearly(&allocation_info_, size_in_bytes);
2345 int size_in_bytes,
2347 Free(start, size_in_bytes, add_to_freelist);
2553 HeapObject* FixedSpace::SlowAllocateRaw(int size_in_bytes) {
2554 ASSERT_EQ(object_size_in_bytes_, size_in_bytes);
2560 return AllocateInNextPage(current_page, size_in_bytes);
2570 accounting_stats_.AllocateBytes(size_in_bytes);
2580 p->SetAllocationWatermark(obj->address() + size_in_bytes);
2598 return AllocateInNextPage(current_page, size_in_bytes);
2610 int size_in_bytes) {
2613 ASSERT_EQ(object_size_in_bytes_, size_in_bytes);
2619 return AllocateLinearly(&allocation_info_, size_in_bytes);
2624 int size_in_bytes,
2630 ASSERT(size_in_bytes % size == 0);
2631 Address end = start + size_in_bytes;
2726 LargeObjectChunk* LargeObjectChunk::New(int size_in_bytes,
2728 size_t requested = ChunkSizeFor(size_in_bytes);
2761 int LargeObjectChunk::ChunkSizeFor(int size_in_bytes) {
2764 size_in_bytes += (Page::kPageSize - os_alignment);
2766 return size_in_bytes + Page::kObjectStartOffset;
2875 MaybeObject* LargeObjectSpace::AllocateRawCode(int size_in_bytes) {
2876 ASSERT(0 < size_in_bytes);
2877 return AllocateRawInternal(size_in_bytes,
2878 size_in_bytes,
2883 MaybeObject* LargeObjectSpace::AllocateRawFixedArray(int size_in_bytes) {
2884 ASSERT(0 < size_in_bytes);
2885 return AllocateRawInternal(size_in_bytes,
2886 size_in_bytes,
2891 MaybeObject* LargeObjectSpace::AllocateRaw(int size_in_bytes) {
2892 ASSERT(0 < size_in_bytes);
2893 return AllocateRawInternal(size_in_bytes,
2894 size_in_bytes,