Home | History | Annotate | Download | only in heap
      1 // Copyright 2011 the V8 project authors. All rights reserved.
      2 // Use of this source code is governed by a BSD-style license that can be
      3 // found in the LICENSE file.
      4 
      5 #include "src/heap/spaces.h"
      6 
      7 #include <utility>
      8 
      9 #include "src/base/bits.h"
     10 #include "src/base/macros.h"
     11 #include "src/base/platform/semaphore.h"
     12 #include "src/base/template-utils.h"
     13 #include "src/counters.h"
     14 #include "src/heap/array-buffer-tracker.h"
     15 #include "src/heap/concurrent-marking.h"
     16 #include "src/heap/gc-tracer.h"
     17 #include "src/heap/heap-controller.h"
     18 #include "src/heap/incremental-marking.h"
     19 #include "src/heap/mark-compact.h"
     20 #include "src/heap/remembered-set.h"
     21 #include "src/heap/slot-set.h"
     22 #include "src/heap/sweeper.h"
     23 #include "src/msan.h"
     24 #include "src/objects-inl.h"
     25 #include "src/objects/js-array-buffer-inl.h"
     26 #include "src/objects/js-array-inl.h"
     27 #include "src/snapshot/snapshot.h"
     28 #include "src/v8.h"
     29 #include "src/vm-state-inl.h"
     30 
     31 namespace v8 {
     32 namespace internal {
     33 
     34 // ----------------------------------------------------------------------------
     35 // HeapObjectIterator
     36 
     37 HeapObjectIterator::HeapObjectIterator(PagedSpace* space)
     38     : cur_addr_(kNullAddress),
     39       cur_end_(kNullAddress),
     40       space_(space),
     41       page_range_(space->first_page(), nullptr),
     42       current_page_(page_range_.begin()) {}
     43 
     44 HeapObjectIterator::HeapObjectIterator(Page* page)
     45     : cur_addr_(kNullAddress),
     46       cur_end_(kNullAddress),
     47       space_(reinterpret_cast<PagedSpace*>(page->owner())),
     48       page_range_(page),
     49       current_page_(page_range_.begin()) {
     50 #ifdef DEBUG
     51   Space* owner = page->owner();
     52   DCHECK(owner == page->heap()->old_space() ||
     53          owner == page->heap()->map_space() ||
     54          owner == page->heap()->code_space() ||
     55          owner == page->heap()->read_only_space());
     56 #endif  // DEBUG
     57 }
     58 
     59 // We have hit the end of the page and should advance to the next block of
     60 // objects.  This happens at the end of the page.
     61 bool HeapObjectIterator::AdvanceToNextPage() {
     62   DCHECK_EQ(cur_addr_, cur_end_);
     63   if (current_page_ == page_range_.end()) return false;
     64   Page* cur_page = *(current_page_++);
     65   Heap* heap = space_->heap();
     66 
     67   heap->mark_compact_collector()->sweeper()->EnsurePageIsIterable(cur_page);
     68 #ifdef ENABLE_MINOR_MC
     69   if (cur_page->IsFlagSet(Page::SWEEP_TO_ITERATE))
     70     heap->minor_mark_compact_collector()->MakeIterable(
     71         cur_page, MarkingTreatmentMode::CLEAR,
     72         FreeSpaceTreatmentMode::IGNORE_FREE_SPACE);
     73 #else
     74   DCHECK(!cur_page->IsFlagSet(Page::SWEEP_TO_ITERATE));
     75 #endif  // ENABLE_MINOR_MC
     76   cur_addr_ = cur_page->area_start();
     77   cur_end_ = cur_page->area_end();
     78   DCHECK(cur_page->SweepingDone());
     79   return true;
     80 }
     81 
     82 PauseAllocationObserversScope::PauseAllocationObserversScope(Heap* heap)
     83     : heap_(heap) {
     84   DCHECK_EQ(heap->gc_state(), Heap::NOT_IN_GC);
     85 
     86   for (SpaceIterator it(heap_); it.has_next();) {
     87     it.next()->PauseAllocationObservers();
     88   }
     89 }
     90 
     91 PauseAllocationObserversScope::~PauseAllocationObserversScope() {
     92   for (SpaceIterator it(heap_); it.has_next();) {
     93     it.next()->ResumeAllocationObservers();
     94   }
     95 }
     96 
     97 // -----------------------------------------------------------------------------
     98 // CodeRange
     99 
    100 static base::LazyInstance<CodeRangeAddressHint>::type code_range_address_hint =
    101     LAZY_INSTANCE_INITIALIZER;
    102 
    103 CodeRange::CodeRange(Isolate* isolate, size_t requested)
    104     : isolate_(isolate),
    105       free_list_(0),
    106       allocation_list_(0),
    107       current_allocation_block_index_(0),
    108       requested_code_range_size_(0) {
    109   DCHECK(!virtual_memory_.IsReserved());
    110 
    111   if (requested == 0) {
    112     // When a target requires the code range feature, we put all code objects
    113     // in a kMaximalCodeRangeSize range of virtual address space, so that
    114     // they can call each other with near calls.
    115     if (kRequiresCodeRange) {
    116       requested = kMaximalCodeRangeSize;
    117     } else {
    118       return;
    119     }
    120   }
    121 
    122   if (requested <= kMinimumCodeRangeSize) {
    123     requested = kMinimumCodeRangeSize;
    124   }
    125 
    126   const size_t reserved_area =
    127       kReservedCodeRangePages * MemoryAllocator::GetCommitPageSize();
    128   if (requested < (kMaximalCodeRangeSize - reserved_area))
    129     requested += reserved_area;
    130 
    131   DCHECK(!kRequiresCodeRange || requested <= kMaximalCodeRangeSize);
    132 
    133   requested_code_range_size_ = requested;
    134 
    135   VirtualMemory reservation;
    136   void* hint = code_range_address_hint.Pointer()->GetAddressHint(requested);
    137   if (!AlignedAllocVirtualMemory(
    138           requested, Max(kCodeRangeAreaAlignment, AllocatePageSize()), hint,
    139           &reservation)) {
    140     V8::FatalProcessOutOfMemory(isolate,
    141                                 "CodeRange setup: allocate virtual memory");
    142   }
    143 
    144   // We are sure that we have mapped a block of requested addresses.
    145   DCHECK_GE(reservation.size(), requested);
    146   Address base = reservation.address();
    147 
    148   // On some platforms, specifically Win64, we need to reserve some pages at
    149   // the beginning of an executable space.
    150   if (reserved_area > 0) {
    151     if (!reservation.SetPermissions(base, reserved_area,
    152                                     PageAllocator::kReadWrite))
    153       V8::FatalProcessOutOfMemory(isolate, "CodeRange setup: set permissions");
    154 
    155     base += reserved_area;
    156   }
    157   Address aligned_base = ::RoundUp(base, MemoryChunk::kAlignment);
    158   size_t size = reservation.size() - (aligned_base - base) - reserved_area;
    159   allocation_list_.emplace_back(aligned_base, size);
    160   current_allocation_block_index_ = 0;
    161 
    162   LOG(isolate_,
    163       NewEvent("CodeRange", reinterpret_cast<void*>(reservation.address()),
    164                requested));
    165   virtual_memory_.TakeControl(&reservation);
    166 }
    167 
    168 CodeRange::~CodeRange() {
    169   if (virtual_memory_.IsReserved()) {
    170     Address addr = start();
    171     virtual_memory_.Free();
    172     code_range_address_hint.Pointer()->NotifyFreedCodeRange(
    173         reinterpret_cast<void*>(addr), requested_code_range_size_);
    174   }
    175 }
    176 
    177 bool CodeRange::CompareFreeBlockAddress(const FreeBlock& left,
    178                                         const FreeBlock& right) {
    179   return left.start < right.start;
    180 }
    181 
    182 
    183 bool CodeRange::GetNextAllocationBlock(size_t requested) {
    184   for (current_allocation_block_index_++;
    185        current_allocation_block_index_ < allocation_list_.size();
    186        current_allocation_block_index_++) {
    187     if (requested <= allocation_list_[current_allocation_block_index_].size) {
    188       return true;  // Found a large enough allocation block.
    189     }
    190   }
    191 
    192   // Sort and merge the free blocks on the free list and the allocation list.
    193   free_list_.insert(free_list_.end(), allocation_list_.begin(),
    194                     allocation_list_.end());
    195   allocation_list_.clear();
    196   std::sort(free_list_.begin(), free_list_.end(), &CompareFreeBlockAddress);
    197   for (size_t i = 0; i < free_list_.size();) {
    198     FreeBlock merged = free_list_[i];
    199     i++;
    200     // Add adjacent free blocks to the current merged block.
    201     while (i < free_list_.size() &&
    202            free_list_[i].start == merged.start + merged.size) {
    203       merged.size += free_list_[i].size;
    204       i++;
    205     }
    206     if (merged.size > 0) {
    207       allocation_list_.push_back(merged);
    208     }
    209   }
    210   free_list_.clear();
    211 
    212   for (current_allocation_block_index_ = 0;
    213        current_allocation_block_index_ < allocation_list_.size();
    214        current_allocation_block_index_++) {
    215     if (requested <= allocation_list_[current_allocation_block_index_].size) {
    216       return true;  // Found a large enough allocation block.
    217     }
    218   }
    219   current_allocation_block_index_ = 0;
    220   // Code range is full or too fragmented.
    221   return false;
    222 }
    223 
    224 
    225 Address CodeRange::AllocateRawMemory(const size_t requested_size,
    226                                      const size_t commit_size,
    227                                      size_t* allocated) {
    228   // requested_size includes the header and two guard regions, while commit_size
    229   // only includes the header.
    230   DCHECK_LE(commit_size,
    231             requested_size - 2 * MemoryAllocator::CodePageGuardSize());
    232   FreeBlock current;
    233   if (!ReserveBlock(requested_size, &current)) {
    234     *allocated = 0;
    235     return kNullAddress;
    236   }
    237   *allocated = current.size;
    238   DCHECK(IsAddressAligned(current.start, MemoryChunk::kAlignment));
    239   if (!isolate_->heap()->memory_allocator()->CommitExecutableMemory(
    240           &virtual_memory_, current.start, commit_size, *allocated)) {
    241     *allocated = 0;
    242     ReleaseBlock(&current);
    243     return kNullAddress;
    244   }
    245   return current.start;
    246 }
    247 
    248 void CodeRange::FreeRawMemory(Address address, size_t length) {
    249   DCHECK(IsAddressAligned(address, MemoryChunk::kAlignment));
    250   base::LockGuard<base::Mutex> guard(&code_range_mutex_);
    251   free_list_.emplace_back(address, length);
    252   virtual_memory_.SetPermissions(address, length, PageAllocator::kNoAccess);
    253 }
    254 
    255 bool CodeRange::ReserveBlock(const size_t requested_size, FreeBlock* block) {
    256   base::LockGuard<base::Mutex> guard(&code_range_mutex_);
    257   DCHECK(allocation_list_.empty() ||
    258          current_allocation_block_index_ < allocation_list_.size());
    259   if (allocation_list_.empty() ||
    260       requested_size > allocation_list_[current_allocation_block_index_].size) {
    261     // Find an allocation block large enough.
    262     if (!GetNextAllocationBlock(requested_size)) return false;
    263   }
    264   // Commit the requested memory at the start of the current allocation block.
    265   size_t aligned_requested = ::RoundUp(requested_size, MemoryChunk::kAlignment);
    266   *block = allocation_list_[current_allocation_block_index_];
    267   // Don't leave a small free block, useless for a large object or chunk.
    268   if (aligned_requested < (block->size - Page::kPageSize)) {
    269     block->size = aligned_requested;
    270   }
    271   DCHECK(IsAddressAligned(block->start, MemoryChunk::kAlignment));
    272   allocation_list_[current_allocation_block_index_].start += block->size;
    273   allocation_list_[current_allocation_block_index_].size -= block->size;
    274   return true;
    275 }
    276 
    277 
    278 void CodeRange::ReleaseBlock(const FreeBlock* block) {
    279   base::LockGuard<base::Mutex> guard(&code_range_mutex_);
    280   free_list_.push_back(*block);
    281 }
    282 
    283 void* CodeRangeAddressHint::GetAddressHint(size_t code_range_size) {
    284   base::LockGuard<base::Mutex> guard(&mutex_);
    285   auto it = recently_freed_.find(code_range_size);
    286   if (it == recently_freed_.end() || it->second.empty()) {
    287     return GetRandomMmapAddr();
    288   }
    289   void* result = it->second.back();
    290   it->second.pop_back();
    291   return result;
    292 }
    293 
    294 void CodeRangeAddressHint::NotifyFreedCodeRange(void* code_range_start,
    295                                                 size_t code_range_size) {
    296   base::LockGuard<base::Mutex> guard(&mutex_);
    297   recently_freed_[code_range_size].push_back(code_range_start);
    298 }
    299 
    300 // -----------------------------------------------------------------------------
    301 // MemoryAllocator
    302 //
    303 
    304 MemoryAllocator::MemoryAllocator(Isolate* isolate, size_t capacity,
    305                                  size_t code_range_size)
    306     : isolate_(isolate),
    307       code_range_(nullptr),
    308       capacity_(RoundUp(capacity, Page::kPageSize)),
    309       size_(0),
    310       size_executable_(0),
    311       lowest_ever_allocated_(static_cast<Address>(-1ll)),
    312       highest_ever_allocated_(kNullAddress),
    313       unmapper_(isolate->heap(), this) {
    314   code_range_ = new CodeRange(isolate_, code_range_size);
    315 }
    316 
    317 
    318 void MemoryAllocator::TearDown() {
    319   unmapper()->TearDown();
    320 
    321   // Check that spaces were torn down before MemoryAllocator.
    322   DCHECK_EQ(size_, 0u);
    323   // TODO(gc) this will be true again when we fix FreeMemory.
    324   // DCHECK_EQ(0, size_executable_);
    325   capacity_ = 0;
    326 
    327   if (last_chunk_.IsReserved()) {
    328     last_chunk_.Free();
    329   }
    330 
    331   delete code_range_;
    332   code_range_ = nullptr;
    333 }
    334 
    335 class MemoryAllocator::Unmapper::UnmapFreeMemoryTask : public CancelableTask {
    336  public:
    337   explicit UnmapFreeMemoryTask(Isolate* isolate, Unmapper* unmapper)
    338       : CancelableTask(isolate),
    339         unmapper_(unmapper),
    340         tracer_(isolate->heap()->tracer()) {}
    341 
    342  private:
    343   void RunInternal() override {
    344     TRACE_BACKGROUND_GC(tracer_,
    345                         GCTracer::BackgroundScope::BACKGROUND_UNMAPPER);
    346     unmapper_->PerformFreeMemoryOnQueuedChunks<FreeMode::kUncommitPooled>();
    347     unmapper_->active_unmapping_tasks_--;
    348     unmapper_->pending_unmapping_tasks_semaphore_.Signal();
    349     if (FLAG_trace_unmapper) {
    350       PrintIsolate(unmapper_->heap_->isolate(),
    351                    "UnmapFreeMemoryTask Done: id=%" PRIu64 "\n", id());
    352     }
    353   }
    354 
    355   Unmapper* const unmapper_;
    356   GCTracer* const tracer_;
    357   DISALLOW_COPY_AND_ASSIGN(UnmapFreeMemoryTask);
    358 };
    359 
    360 void MemoryAllocator::Unmapper::FreeQueuedChunks() {
    361   if (!heap_->IsTearingDown() && FLAG_concurrent_sweeping) {
    362     if (!MakeRoomForNewTasks()) {
    363       // kMaxUnmapperTasks are already running. Avoid creating any more.
    364       if (FLAG_trace_unmapper) {
    365         PrintIsolate(heap_->isolate(),
    366                      "Unmapper::FreeQueuedChunks: reached task limit (%d)\n",
    367                      kMaxUnmapperTasks);
    368       }
    369       return;
    370     }
    371     auto task = base::make_unique<UnmapFreeMemoryTask>(heap_->isolate(), this);
    372     if (FLAG_trace_unmapper) {
    373       PrintIsolate(heap_->isolate(),
    374                    "Unmapper::FreeQueuedChunks: new task id=%" PRIu64 "\n",
    375                    task->id());
    376     }
    377     DCHECK_LT(pending_unmapping_tasks_, kMaxUnmapperTasks);
    378     DCHECK_LE(active_unmapping_tasks_, pending_unmapping_tasks_);
    379     DCHECK_GE(active_unmapping_tasks_, 0);
    380     active_unmapping_tasks_++;
    381     task_ids_[pending_unmapping_tasks_++] = task->id();
    382     V8::GetCurrentPlatform()->CallOnWorkerThread(std::move(task));
    383   } else {
    384     PerformFreeMemoryOnQueuedChunks<FreeMode::kUncommitPooled>();
    385   }
    386 }
    387 
    388 void MemoryAllocator::Unmapper::CancelAndWaitForPendingTasks() {
    389   for (int i = 0; i < pending_unmapping_tasks_; i++) {
    390     if (heap_->isolate()->cancelable_task_manager()->TryAbort(task_ids_[i]) !=
    391         CancelableTaskManager::kTaskAborted) {
    392       pending_unmapping_tasks_semaphore_.Wait();
    393     }
    394   }
    395   pending_unmapping_tasks_ = 0;
    396   active_unmapping_tasks_ = 0;
    397 
    398   if (FLAG_trace_unmapper) {
    399     PrintIsolate(
    400         heap_->isolate(),
    401         "Unmapper::CancelAndWaitForPendingTasks: no tasks remaining\n");
    402   }
    403 }
    404 
    405 void MemoryAllocator::Unmapper::PrepareForMarkCompact() {
    406   CancelAndWaitForPendingTasks();
    407   // Free non-regular chunks because they cannot be re-used.
    408   PerformFreeMemoryOnQueuedNonRegularChunks();
    409 }
    410 
    411 void MemoryAllocator::Unmapper::EnsureUnmappingCompleted() {
    412   CancelAndWaitForPendingTasks();
    413   PerformFreeMemoryOnQueuedChunks<FreeMode::kReleasePooled>();
    414 }
    415 
    416 bool MemoryAllocator::Unmapper::MakeRoomForNewTasks() {
    417   DCHECK_LE(pending_unmapping_tasks_, kMaxUnmapperTasks);
    418 
    419   if (active_unmapping_tasks_ == 0 && pending_unmapping_tasks_ > 0) {
    420     // All previous unmapping tasks have been run to completion.
    421     // Finalize those tasks to make room for new ones.
    422     CancelAndWaitForPendingTasks();
    423   }
    424   return pending_unmapping_tasks_ != kMaxUnmapperTasks;
    425 }
    426 
    427 void MemoryAllocator::Unmapper::PerformFreeMemoryOnQueuedNonRegularChunks() {
    428   MemoryChunk* chunk = nullptr;
    429   while ((chunk = GetMemoryChunkSafe<kNonRegular>()) != nullptr) {
    430     allocator_->PerformFreeMemory(chunk);
    431   }
    432 }
    433 
    434 template <MemoryAllocator::Unmapper::FreeMode mode>
    435 void MemoryAllocator::Unmapper::PerformFreeMemoryOnQueuedChunks() {
    436   MemoryChunk* chunk = nullptr;
    437   if (FLAG_trace_unmapper) {
    438     PrintIsolate(
    439         heap_->isolate(),
    440         "Unmapper::PerformFreeMemoryOnQueuedChunks: %d queued chunks\n",
    441         NumberOfChunks());
    442   }
    443   // Regular chunks.
    444   while ((chunk = GetMemoryChunkSafe<kRegular>()) != nullptr) {
    445     bool pooled = chunk->IsFlagSet(MemoryChunk::POOLED);
    446     allocator_->PerformFreeMemory(chunk);
    447     if (pooled) AddMemoryChunkSafe<kPooled>(chunk);
    448   }
    449   if (mode == MemoryAllocator::Unmapper::FreeMode::kReleasePooled) {
    450     // The previous loop uncommitted any pages marked as pooled and added them
    451     // to the pooled list. In case of kReleasePooled we need to free them
    452     // though.
    453     while ((chunk = GetMemoryChunkSafe<kPooled>()) != nullptr) {
    454       allocator_->Free<MemoryAllocator::kAlreadyPooled>(chunk);
    455     }
    456   }
    457   PerformFreeMemoryOnQueuedNonRegularChunks();
    458 }
    459 
    460 void MemoryAllocator::Unmapper::TearDown() {
    461   CHECK_EQ(0, pending_unmapping_tasks_);
    462   PerformFreeMemoryOnQueuedChunks<FreeMode::kReleasePooled>();
    463   for (int i = 0; i < kNumberOfChunkQueues; i++) {
    464     DCHECK(chunks_[i].empty());
    465   }
    466 }
    467 
    468 int MemoryAllocator::Unmapper::NumberOfChunks() {
    469   base::LockGuard<base::Mutex> guard(&mutex_);
    470   size_t result = 0;
    471   for (int i = 0; i < kNumberOfChunkQueues; i++) {
    472     result += chunks_[i].size();
    473   }
    474   return static_cast<int>(result);
    475 }
    476 
    477 size_t MemoryAllocator::Unmapper::CommittedBufferedMemory() {
    478   base::LockGuard<base::Mutex> guard(&mutex_);
    479 
    480   size_t sum = 0;
    481   // kPooled chunks are already uncommited. We only have to account for
    482   // kRegular and kNonRegular chunks.
    483   for (auto& chunk : chunks_[kRegular]) {
    484     sum += chunk->size();
    485   }
    486   for (auto& chunk : chunks_[kNonRegular]) {
    487     sum += chunk->size();
    488   }
    489   return sum;
    490 }
    491 
    492 bool MemoryAllocator::CommitMemory(Address base, size_t size) {
    493   if (!SetPermissions(base, size, PageAllocator::kReadWrite)) {
    494     return false;
    495   }
    496   UpdateAllocatedSpaceLimits(base, base + size);
    497   return true;
    498 }
    499 
    500 void MemoryAllocator::FreeMemory(VirtualMemory* reservation,
    501                                  Executability executable) {
    502   // TODO(gc) make code_range part of memory allocator?
    503   // Code which is part of the code-range does not have its own VirtualMemory.
    504   DCHECK(code_range() == nullptr ||
    505          !code_range()->contains(reservation->address()));
    506   DCHECK(executable == NOT_EXECUTABLE || !code_range()->valid() ||
    507          reservation->size() <= Page::kPageSize);
    508 
    509   reservation->Free();
    510 }
    511 
    512 
    513 void MemoryAllocator::FreeMemory(Address base, size_t size,
    514                                  Executability executable) {
    515   // TODO(gc) make code_range part of memory allocator?
    516   if (code_range() != nullptr && code_range()->contains(base)) {
    517     DCHECK(executable == EXECUTABLE);
    518     code_range()->FreeRawMemory(base, size);
    519   } else {
    520     DCHECK(executable == NOT_EXECUTABLE || !code_range()->valid());
    521     CHECK(FreePages(reinterpret_cast<void*>(base), size));
    522   }
    523 }
    524 
    525 Address MemoryAllocator::ReserveAlignedMemory(size_t size, size_t alignment,
    526                                               void* hint,
    527                                               VirtualMemory* controller) {
    528   VirtualMemory reservation;
    529   if (!AlignedAllocVirtualMemory(size, alignment, hint, &reservation)) {
    530     return kNullAddress;
    531   }
    532 
    533   Address result = reservation.address();
    534   size_ += reservation.size();
    535   controller->TakeControl(&reservation);
    536   return result;
    537 }
    538 
    539 Address MemoryAllocator::AllocateAlignedMemory(
    540     size_t reserve_size, size_t commit_size, size_t alignment,
    541     Executability executable, void* hint, VirtualMemory* controller) {
    542   DCHECK(commit_size <= reserve_size);
    543   VirtualMemory reservation;
    544   Address base =
    545       ReserveAlignedMemory(reserve_size, alignment, hint, &reservation);
    546   if (base == kNullAddress) return kNullAddress;
    547 
    548   if (executable == EXECUTABLE) {
    549     if (!CommitExecutableMemory(&reservation, base, commit_size,
    550                                 reserve_size)) {
    551       base = kNullAddress;
    552     }
    553   } else {
    554     if (reservation.SetPermissions(base, commit_size,
    555                                    PageAllocator::kReadWrite)) {
    556       UpdateAllocatedSpaceLimits(base, base + commit_size);
    557     } else {
    558       base = kNullAddress;
    559     }
    560   }
    561 
    562   if (base == kNullAddress) {
    563     // Failed to commit the body. Free the mapping and any partially committed
    564     // regions inside it.
    565     reservation.Free();
    566     size_ -= reserve_size;
    567     return kNullAddress;
    568   }
    569 
    570   controller->TakeControl(&reservation);
    571   return base;
    572 }
    573 
    574 Heap* MemoryChunk::synchronized_heap() {
    575   return reinterpret_cast<Heap*>(
    576       base::Acquire_Load(reinterpret_cast<base::AtomicWord*>(&heap_)));
    577 }
    578 
    579 void MemoryChunk::InitializationMemoryFence() {
    580   base::SeqCst_MemoryFence();
    581 #ifdef THREAD_SANITIZER
    582   // Since TSAN does not process memory fences, we use the following annotation
    583   // to tell TSAN that there is no data race when emitting a
    584   // InitializationMemoryFence. Note that the other thread still needs to
    585   // perform MemoryChunk::synchronized_heap().
    586   base::Release_Store(reinterpret_cast<base::AtomicWord*>(&heap_),
    587                       reinterpret_cast<base::AtomicWord>(heap_));
    588 #endif
    589 }
    590 
    591 void MemoryChunk::SetReadAndExecutable() {
    592   DCHECK(IsFlagSet(MemoryChunk::IS_EXECUTABLE));
    593   DCHECK(owner()->identity() == CODE_SPACE || owner()->identity() == LO_SPACE);
    594   // Decrementing the write_unprotect_counter_ and changing the page
    595   // protection mode has to be atomic.
    596   base::LockGuard<base::Mutex> guard(page_protection_change_mutex_);
    597   if (write_unprotect_counter_ == 0) {
    598     // This is a corner case that may happen when we have a
    599     // CodeSpaceMemoryModificationScope open and this page was newly
    600     // added.
    601     return;
    602   }
    603   write_unprotect_counter_--;
    604   DCHECK_LT(write_unprotect_counter_, kMaxWriteUnprotectCounter);
    605   if (write_unprotect_counter_ == 0) {
    606     Address protect_start =
    607         address() + MemoryAllocator::CodePageAreaStartOffset();
    608     size_t page_size = MemoryAllocator::GetCommitPageSize();
    609     DCHECK(IsAddressAligned(protect_start, page_size));
    610     size_t protect_size = RoundUp(area_size(), page_size);
    611     CHECK(SetPermissions(protect_start, protect_size,
    612                          PageAllocator::kReadExecute));
    613   }
    614 }
    615 
    616 void MemoryChunk::SetReadAndWritable() {
    617   DCHECK(IsFlagSet(MemoryChunk::IS_EXECUTABLE));
    618   DCHECK(owner()->identity() == CODE_SPACE || owner()->identity() == LO_SPACE);
    619   // Incrementing the write_unprotect_counter_ and changing the page
    620   // protection mode has to be atomic.
    621   base::LockGuard<base::Mutex> guard(page_protection_change_mutex_);
    622   write_unprotect_counter_++;
    623   DCHECK_LE(write_unprotect_counter_, kMaxWriteUnprotectCounter);
    624   if (write_unprotect_counter_ == 1) {
    625     Address unprotect_start =
    626         address() + MemoryAllocator::CodePageAreaStartOffset();
    627     size_t page_size = MemoryAllocator::GetCommitPageSize();
    628     DCHECK(IsAddressAligned(unprotect_start, page_size));
    629     size_t unprotect_size = RoundUp(area_size(), page_size);
    630     CHECK(SetPermissions(unprotect_start, unprotect_size,
    631                          PageAllocator::kReadWrite));
    632   }
    633 }
    634 
    635 MemoryChunk* MemoryChunk::Initialize(Heap* heap, Address base, size_t size,
    636                                      Address area_start, Address area_end,
    637                                      Executability executable, Space* owner,
    638                                      VirtualMemory* reservation) {
    639   MemoryChunk* chunk = FromAddress(base);
    640 
    641   DCHECK(base == chunk->address());
    642 
    643   chunk->heap_ = heap;
    644   chunk->size_ = size;
    645   chunk->area_start_ = area_start;
    646   chunk->area_end_ = area_end;
    647   chunk->flags_ = Flags(NO_FLAGS);
    648   chunk->set_owner(owner);
    649   chunk->InitializeReservedMemory();
    650   base::AsAtomicPointer::Release_Store(&chunk->slot_set_[OLD_TO_NEW], nullptr);
    651   base::AsAtomicPointer::Release_Store(&chunk->slot_set_[OLD_TO_OLD], nullptr);
    652   base::AsAtomicPointer::Release_Store(&chunk->typed_slot_set_[OLD_TO_NEW],
    653                                        nullptr);
    654   base::AsAtomicPointer::Release_Store(&chunk->typed_slot_set_[OLD_TO_OLD],
    655                                        nullptr);
    656   chunk->invalidated_slots_ = nullptr;
    657   chunk->skip_list_ = nullptr;
    658   chunk->progress_bar_ = 0;
    659   chunk->high_water_mark_ = static_cast<intptr_t>(area_start - base);
    660   chunk->set_concurrent_sweeping_state(kSweepingDone);
    661   chunk->page_protection_change_mutex_ = new base::Mutex();
    662   chunk->write_unprotect_counter_ = 0;
    663   chunk->mutex_ = new base::Mutex();
    664   chunk->allocated_bytes_ = chunk->area_size();
    665   chunk->wasted_memory_ = 0;
    666   chunk->young_generation_bitmap_ = nullptr;
    667   chunk->local_tracker_ = nullptr;
    668 
    669   chunk->external_backing_store_bytes_[ExternalBackingStoreType::kArrayBuffer] =
    670       0;
    671   chunk->external_backing_store_bytes_
    672       [ExternalBackingStoreType::kExternalString] = 0;
    673 
    674   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
    675     chunk->categories_[i] = nullptr;
    676   }
    677 
    678   if (owner->identity() == RO_SPACE) {
    679     heap->incremental_marking()
    680         ->non_atomic_marking_state()
    681         ->bitmap(chunk)
    682         ->MarkAllBits();
    683   } else {
    684     heap->incremental_marking()->non_atomic_marking_state()->ClearLiveness(
    685         chunk);
    686   }
    687 
    688   DCHECK_EQ(kFlagsOffset, OFFSET_OF(MemoryChunk, flags_));
    689 
    690   if (executable == EXECUTABLE) {
    691     chunk->SetFlag(IS_EXECUTABLE);
    692     if (heap->write_protect_code_memory()) {
    693       chunk->write_unprotect_counter_ =
    694           heap->code_space_memory_modification_scope_depth();
    695     } else {
    696       size_t page_size = MemoryAllocator::GetCommitPageSize();
    697       DCHECK(IsAddressAligned(area_start, page_size));
    698       size_t area_size = RoundUp(area_end - area_start, page_size);
    699       CHECK(SetPermissions(area_start, area_size,
    700                            PageAllocator::kReadWriteExecute));
    701     }
    702   }
    703 
    704   if (reservation != nullptr) {
    705     chunk->reservation_.TakeControl(reservation);
    706   }
    707 
    708   return chunk;
    709 }
    710 
    711 Page* PagedSpace::InitializePage(MemoryChunk* chunk, Executability executable) {
    712   Page* page = static_cast<Page*>(chunk);
    713   DCHECK_GE(Page::kAllocatableMemory, page->area_size());
    714   // Make sure that categories are initialized before freeing the area.
    715   page->ResetAllocatedBytes();
    716   page->SetOldGenerationPageFlags(heap()->incremental_marking()->IsMarking());
    717   page->AllocateFreeListCategories();
    718   page->InitializeFreeListCategories();
    719   page->list_node().Initialize();
    720   page->InitializationMemoryFence();
    721   return page;
    722 }
    723 
    724 Page* SemiSpace::InitializePage(MemoryChunk* chunk, Executability executable) {
    725   DCHECK_EQ(executable, Executability::NOT_EXECUTABLE);
    726   bool in_to_space = (id() != kFromSpace);
    727   chunk->SetFlag(in_to_space ? MemoryChunk::IN_TO_SPACE
    728                              : MemoryChunk::IN_FROM_SPACE);
    729   DCHECK(!chunk->IsFlagSet(in_to_space ? MemoryChunk::IN_FROM_SPACE
    730                                        : MemoryChunk::IN_TO_SPACE));
    731   Page* page = static_cast<Page*>(chunk);
    732   page->SetYoungGenerationPageFlags(heap()->incremental_marking()->IsMarking());
    733   page->AllocateLocalTracker();
    734   page->list_node().Initialize();
    735 #ifdef ENABLE_MINOR_MC
    736   if (FLAG_minor_mc) {
    737     page->AllocateYoungGenerationBitmap();
    738     heap()
    739         ->minor_mark_compact_collector()
    740         ->non_atomic_marking_state()
    741         ->ClearLiveness(page);
    742   }
    743 #endif  // ENABLE_MINOR_MC
    744   page->InitializationMemoryFence();
    745   return page;
    746 }
    747 
    748 LargePage* LargePage::Initialize(Heap* heap, MemoryChunk* chunk,
    749                                  Executability executable) {
    750   if (executable && chunk->size() > LargePage::kMaxCodePageSize) {
    751     STATIC_ASSERT(LargePage::kMaxCodePageSize <= TypedSlotSet::kMaxOffset);
    752     FATAL("Code page is too large.");
    753   }
    754 
    755   MSAN_ALLOCATED_UNINITIALIZED_MEMORY(chunk->area_start(), chunk->area_size());
    756 
    757   LargePage* page = static_cast<LargePage*>(chunk);
    758   page->list_node().Initialize();
    759   return page;
    760 }
    761 
    762 void Page::AllocateFreeListCategories() {
    763   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
    764     categories_[i] = new FreeListCategory(
    765         reinterpret_cast<PagedSpace*>(owner())->free_list(), this);
    766   }
    767 }
    768 
    769 void Page::InitializeFreeListCategories() {
    770   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
    771     categories_[i]->Initialize(static_cast<FreeListCategoryType>(i));
    772   }
    773 }
    774 
    775 void Page::ReleaseFreeListCategories() {
    776   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
    777     if (categories_[i] != nullptr) {
    778       delete categories_[i];
    779       categories_[i] = nullptr;
    780     }
    781   }
    782 }
    783 
    784 Page* Page::ConvertNewToOld(Page* old_page) {
    785   DCHECK(old_page);
    786   DCHECK(old_page->InNewSpace());
    787   OldSpace* old_space = old_page->heap()->old_space();
    788   old_page->set_owner(old_space);
    789   old_page->SetFlags(0, static_cast<uintptr_t>(~0));
    790   Page* new_page = old_space->InitializePage(old_page, NOT_EXECUTABLE);
    791   old_space->AddPage(new_page);
    792   return new_page;
    793 }
    794 
    795 size_t MemoryChunk::CommittedPhysicalMemory() {
    796   if (!base::OS::HasLazyCommits() || owner()->identity() == LO_SPACE)
    797     return size();
    798   return high_water_mark_;
    799 }
    800 
    801 bool MemoryChunk::IsPagedSpace() const {
    802   return owner()->identity() != LO_SPACE;
    803 }
    804 
    805 bool MemoryChunk::InOldSpace() const {
    806   return owner()->identity() == OLD_SPACE;
    807 }
    808 
    809 bool MemoryChunk::InLargeObjectSpace() const {
    810   return owner()->identity() == LO_SPACE;
    811 }
    812 
    813 MemoryChunk* MemoryAllocator::AllocateChunk(size_t reserve_area_size,
    814                                             size_t commit_area_size,
    815                                             Executability executable,
    816                                             Space* owner) {
    817   DCHECK_LE(commit_area_size, reserve_area_size);
    818 
    819   size_t chunk_size;
    820   Heap* heap = isolate_->heap();
    821   Address base = kNullAddress;
    822   VirtualMemory reservation;
    823   Address area_start = kNullAddress;
    824   Address area_end = kNullAddress;
    825   void* address_hint =
    826       AlignedAddress(heap->GetRandomMmapAddr(), MemoryChunk::kAlignment);
    827 
    828   //
    829   // MemoryChunk layout:
    830   //
    831   //             Executable
    832   // +----------------------------+<- base aligned with MemoryChunk::kAlignment
    833   // |           Header           |
    834   // +----------------------------+<- base + CodePageGuardStartOffset
    835   // |           Guard            |
    836   // +----------------------------+<- area_start_
    837   // |           Area             |
    838   // +----------------------------+<- area_end_ (area_start + commit_area_size)
    839   // |   Committed but not used   |
    840   // +----------------------------+<- aligned at OS page boundary
    841   // | Reserved but not committed |
    842   // +----------------------------+<- aligned at OS page boundary
    843   // |           Guard            |
    844   // +----------------------------+<- base + chunk_size
    845   //
    846   //           Non-executable
    847   // +----------------------------+<- base aligned with MemoryChunk::kAlignment
    848   // |          Header            |
    849   // +----------------------------+<- area_start_ (base + kObjectStartOffset)
    850   // |           Area             |
    851   // +----------------------------+<- area_end_ (area_start + commit_area_size)
    852   // |  Committed but not used    |
    853   // +----------------------------+<- aligned at OS page boundary
    854   // | Reserved but not committed |
    855   // +----------------------------+<- base + chunk_size
    856   //
    857 
    858   if (executable == EXECUTABLE) {
    859     chunk_size = ::RoundUp(
    860         CodePageAreaStartOffset() + reserve_area_size + CodePageGuardSize(),
    861         GetCommitPageSize());
    862 
    863     // Size of header (not executable) plus area (executable).
    864     size_t commit_size = ::RoundUp(
    865         CodePageGuardStartOffset() + commit_area_size, GetCommitPageSize());
    866 // Allocate executable memory either from code range or from the OS.
    867 #ifdef V8_TARGET_ARCH_MIPS64
    868     // Use code range only for large object space on mips64 to keep address
    869     // range within 256-MB memory region.
    870     if (code_range()->valid() && reserve_area_size > CodePageAreaSize()) {
    871 #else
    872     if (code_range()->valid()) {
    873 #endif
    874       base =
    875           code_range()->AllocateRawMemory(chunk_size, commit_size, &chunk_size);
    876       DCHECK(IsAligned(base, MemoryChunk::kAlignment));
    877       if (base == kNullAddress) return nullptr;
    878       size_ += chunk_size;
    879       // Update executable memory size.
    880       size_executable_ += chunk_size;
    881     } else {
    882       base = AllocateAlignedMemory(chunk_size, commit_size,
    883                                    MemoryChunk::kAlignment, executable,
    884                                    address_hint, &reservation);
    885       if (base == kNullAddress) return nullptr;
    886       // Update executable memory size.
    887       size_executable_ += reservation.size();
    888     }
    889 
    890     if (Heap::ShouldZapGarbage()) {
    891       ZapBlock(base, CodePageGuardStartOffset(), kZapValue);
    892       ZapBlock(base + CodePageAreaStartOffset(), commit_area_size, kZapValue);
    893     }
    894 
    895     area_start = base + CodePageAreaStartOffset();
    896     area_end = area_start + commit_area_size;
    897   } else {
    898     chunk_size = ::RoundUp(MemoryChunk::kObjectStartOffset + reserve_area_size,
    899                            GetCommitPageSize());
    900     size_t commit_size =
    901         ::RoundUp(MemoryChunk::kObjectStartOffset + commit_area_size,
    902                   GetCommitPageSize());
    903     base =
    904         AllocateAlignedMemory(chunk_size, commit_size, MemoryChunk::kAlignment,
    905                               executable, address_hint, &reservation);
    906 
    907     if (base == kNullAddress) return nullptr;
    908 
    909     if (Heap::ShouldZapGarbage()) {
    910       ZapBlock(base, Page::kObjectStartOffset + commit_area_size, kZapValue);
    911     }
    912 
    913     area_start = base + Page::kObjectStartOffset;
    914     area_end = area_start + commit_area_size;
    915   }
    916 
    917   // Use chunk_size for statistics and callbacks because we assume that they
    918   // treat reserved but not-yet committed memory regions of chunks as allocated.
    919   isolate_->counters()->memory_allocated()->Increment(
    920       static_cast<int>(chunk_size));
    921 
    922   LOG(isolate_,
    923       NewEvent("MemoryChunk", reinterpret_cast<void*>(base), chunk_size));
    924 
    925   // We cannot use the last chunk in the address space because we would
    926   // overflow when comparing top and limit if this chunk is used for a
    927   // linear allocation area.
    928   if ((base + chunk_size) == 0u) {
    929     CHECK(!last_chunk_.IsReserved());
    930     last_chunk_.TakeControl(&reservation);
    931     UncommitBlock(last_chunk_.address(), last_chunk_.size());
    932     size_ -= chunk_size;
    933     if (executable == EXECUTABLE) {
    934       size_executable_ -= chunk_size;
    935     }
    936     CHECK(last_chunk_.IsReserved());
    937     return AllocateChunk(reserve_area_size, commit_area_size, executable,
    938                          owner);
    939   }
    940 
    941   MemoryChunk* chunk =
    942       MemoryChunk::Initialize(heap, base, chunk_size, area_start, area_end,
    943                               executable, owner, &reservation);
    944 
    945   if (chunk->executable()) RegisterExecutableMemoryChunk(chunk);
    946   return chunk;
    947 }
    948 
    949 void MemoryChunk::SetOldGenerationPageFlags(bool is_marking) {
    950   if (is_marking) {
    951     SetFlag(MemoryChunk::POINTERS_TO_HERE_ARE_INTERESTING);
    952     SetFlag(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING);
    953     SetFlag(MemoryChunk::INCREMENTAL_MARKING);
    954   } else {
    955     ClearFlag(MemoryChunk::POINTERS_TO_HERE_ARE_INTERESTING);
    956     SetFlag(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING);
    957     ClearFlag(MemoryChunk::INCREMENTAL_MARKING);
    958   }
    959 }
    960 
    961 void MemoryChunk::SetYoungGenerationPageFlags(bool is_marking) {
    962   SetFlag(MemoryChunk::POINTERS_TO_HERE_ARE_INTERESTING);
    963   if (is_marking) {
    964     SetFlag(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING);
    965     SetFlag(MemoryChunk::INCREMENTAL_MARKING);
    966   } else {
    967     ClearFlag(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING);
    968     ClearFlag(MemoryChunk::INCREMENTAL_MARKING);
    969   }
    970 }
    971 
    972 void Page::ResetAllocatedBytes() { allocated_bytes_ = area_size(); }
    973 
    974 void Page::AllocateLocalTracker() {
    975   DCHECK_NULL(local_tracker_);
    976   local_tracker_ = new LocalArrayBufferTracker(this);
    977 }
    978 
    979 bool Page::contains_array_buffers() {
    980   return local_tracker_ != nullptr && !local_tracker_->IsEmpty();
    981 }
    982 
    983 void Page::ResetFreeListStatistics() {
    984   wasted_memory_ = 0;
    985 }
    986 
    987 size_t Page::AvailableInFreeList() {
    988   size_t sum = 0;
    989   ForAllFreeListCategories([&sum](FreeListCategory* category) {
    990     sum += category->available();
    991   });
    992   return sum;
    993 }
    994 
    995 #ifdef DEBUG
    996 namespace {
    997 // Skips filler starting from the given filler until the end address.
    998 // Returns the first address after the skipped fillers.
    999 Address SkipFillers(HeapObject* filler, Address end) {
   1000   Address addr = filler->address();
   1001   while (addr < end) {
   1002     filler = HeapObject::FromAddress(addr);
   1003     CHECK(filler->IsFiller());
   1004     addr = filler->address() + filler->Size();
   1005   }
   1006   return addr;
   1007 }
   1008 }  // anonymous namespace
   1009 #endif  // DEBUG
   1010 
   1011 size_t Page::ShrinkToHighWaterMark() {
   1012   // Shrinking only makes sense outside of the CodeRange, where we don't care
   1013   // about address space fragmentation.
   1014   VirtualMemory* reservation = reserved_memory();
   1015   if (!reservation->IsReserved()) return 0;
   1016 
   1017   // Shrink pages to high water mark. The water mark points either to a filler
   1018   // or the area_end.
   1019   HeapObject* filler = HeapObject::FromAddress(HighWaterMark());
   1020   if (filler->address() == area_end()) return 0;
   1021   CHECK(filler->IsFiller());
   1022   // Ensure that no objects were allocated in [filler, area_end) region.
   1023   DCHECK_EQ(area_end(), SkipFillers(filler, area_end()));
   1024   // Ensure that no objects will be allocated on this page.
   1025   DCHECK_EQ(0u, AvailableInFreeList());
   1026 
   1027   size_t unused = RoundDown(static_cast<size_t>(area_end() - filler->address()),
   1028                             MemoryAllocator::GetCommitPageSize());
   1029   if (unused > 0) {
   1030     DCHECK_EQ(0u, unused % MemoryAllocator::GetCommitPageSize());
   1031     if (FLAG_trace_gc_verbose) {
   1032       PrintIsolate(heap()->isolate(), "Shrinking page %p: end %p -> %p\n",
   1033                    reinterpret_cast<void*>(this),
   1034                    reinterpret_cast<void*>(area_end()),
   1035                    reinterpret_cast<void*>(area_end() - unused));
   1036     }
   1037     heap()->CreateFillerObjectAt(
   1038         filler->address(),
   1039         static_cast<int>(area_end() - filler->address() - unused),
   1040         ClearRecordedSlots::kNo);
   1041     heap()->memory_allocator()->PartialFreeMemory(
   1042         this, address() + size() - unused, unused, area_end() - unused);
   1043     if (filler->address() != area_end()) {
   1044       CHECK(filler->IsFiller());
   1045       CHECK_EQ(filler->address() + filler->Size(), area_end());
   1046     }
   1047   }
   1048   return unused;
   1049 }
   1050 
   1051 void Page::CreateBlackArea(Address start, Address end) {
   1052   DCHECK(heap()->incremental_marking()->black_allocation());
   1053   DCHECK_EQ(Page::FromAddress(start), this);
   1054   DCHECK_NE(start, end);
   1055   DCHECK_EQ(Page::FromAddress(end - 1), this);
   1056   IncrementalMarking::MarkingState* marking_state =
   1057       heap()->incremental_marking()->marking_state();
   1058   marking_state->bitmap(this)->SetRange(AddressToMarkbitIndex(start),
   1059                                         AddressToMarkbitIndex(end));
   1060   marking_state->IncrementLiveBytes(this, static_cast<intptr_t>(end - start));
   1061 }
   1062 
   1063 void Page::DestroyBlackArea(Address start, Address end) {
   1064   DCHECK(heap()->incremental_marking()->black_allocation());
   1065   DCHECK_EQ(Page::FromAddress(start), this);
   1066   DCHECK_NE(start, end);
   1067   DCHECK_EQ(Page::FromAddress(end - 1), this);
   1068   IncrementalMarking::MarkingState* marking_state =
   1069       heap()->incremental_marking()->marking_state();
   1070   marking_state->bitmap(this)->ClearRange(AddressToMarkbitIndex(start),
   1071                                           AddressToMarkbitIndex(end));
   1072   marking_state->IncrementLiveBytes(this, -static_cast<intptr_t>(end - start));
   1073 }
   1074 
   1075 void MemoryAllocator::PartialFreeMemory(MemoryChunk* chunk, Address start_free,
   1076                                         size_t bytes_to_free,
   1077                                         Address new_area_end) {
   1078   VirtualMemory* reservation = chunk->reserved_memory();
   1079   DCHECK(reservation->IsReserved());
   1080   chunk->size_ -= bytes_to_free;
   1081   chunk->area_end_ = new_area_end;
   1082   if (chunk->IsFlagSet(MemoryChunk::IS_EXECUTABLE)) {
   1083     // Add guard page at the end.
   1084     size_t page_size = GetCommitPageSize();
   1085     DCHECK_EQ(0, chunk->area_end_ % static_cast<Address>(page_size));
   1086     DCHECK_EQ(chunk->address() + chunk->size(),
   1087               chunk->area_end() + CodePageGuardSize());
   1088     reservation->SetPermissions(chunk->area_end_, page_size,
   1089                                 PageAllocator::kNoAccess);
   1090   }
   1091   // On e.g. Windows, a reservation may be larger than a page and releasing
   1092   // partially starting at |start_free| will also release the potentially
   1093   // unused part behind the current page.
   1094   const size_t released_bytes = reservation->Release(start_free);
   1095   DCHECK_GE(size_, released_bytes);
   1096   size_ -= released_bytes;
   1097   isolate_->counters()->memory_allocated()->Decrement(
   1098       static_cast<int>(released_bytes));
   1099 }
   1100 
   1101 void MemoryAllocator::PreFreeMemory(MemoryChunk* chunk) {
   1102   DCHECK(!chunk->IsFlagSet(MemoryChunk::PRE_FREED));
   1103   LOG(isolate_, DeleteEvent("MemoryChunk", chunk));
   1104 
   1105   isolate_->heap()->RememberUnmappedPage(reinterpret_cast<Address>(chunk),
   1106                                          chunk->IsEvacuationCandidate());
   1107 
   1108   VirtualMemory* reservation = chunk->reserved_memory();
   1109   const size_t size =
   1110       reservation->IsReserved() ? reservation->size() : chunk->size();
   1111   DCHECK_GE(size_, static_cast<size_t>(size));
   1112   size_ -= size;
   1113   isolate_->counters()->memory_allocated()->Decrement(static_cast<int>(size));
   1114   if (chunk->executable() == EXECUTABLE) {
   1115     DCHECK_GE(size_executable_, size);
   1116     size_executable_ -= size;
   1117   }
   1118 
   1119   chunk->SetFlag(MemoryChunk::PRE_FREED);
   1120 
   1121   if (chunk->executable()) UnregisterExecutableMemoryChunk(chunk);
   1122 }
   1123 
   1124 
   1125 void MemoryAllocator::PerformFreeMemory(MemoryChunk* chunk) {
   1126   DCHECK(chunk->IsFlagSet(MemoryChunk::PRE_FREED));
   1127   chunk->ReleaseAllocatedMemory();
   1128 
   1129   VirtualMemory* reservation = chunk->reserved_memory();
   1130   if (chunk->IsFlagSet(MemoryChunk::POOLED)) {
   1131     UncommitBlock(reinterpret_cast<Address>(chunk), MemoryChunk::kPageSize);
   1132   } else {
   1133     if (reservation->IsReserved()) {
   1134       FreeMemory(reservation, chunk->executable());
   1135     } else {
   1136       FreeMemory(chunk->address(), chunk->size(), chunk->executable());
   1137     }
   1138   }
   1139 }
   1140 
   1141 template <MemoryAllocator::FreeMode mode>
   1142 void MemoryAllocator::Free(MemoryChunk* chunk) {
   1143   switch (mode) {
   1144     case kFull:
   1145       PreFreeMemory(chunk);
   1146       PerformFreeMemory(chunk);
   1147       break;
   1148     case kAlreadyPooled:
   1149       // Pooled pages cannot be touched anymore as their memory is uncommitted.
   1150       FreeMemory(chunk->address(), static_cast<size_t>(MemoryChunk::kPageSize),
   1151                  Executability::NOT_EXECUTABLE);
   1152       break;
   1153     case kPooledAndQueue:
   1154       DCHECK_EQ(chunk->size(), static_cast<size_t>(MemoryChunk::kPageSize));
   1155       DCHECK_EQ(chunk->executable(), NOT_EXECUTABLE);
   1156       chunk->SetFlag(MemoryChunk::POOLED);
   1157       V8_FALLTHROUGH;
   1158     case kPreFreeAndQueue:
   1159       PreFreeMemory(chunk);
   1160       // The chunks added to this queue will be freed by a concurrent thread.
   1161       unmapper()->AddMemoryChunkSafe(chunk);
   1162       break;
   1163   }
   1164 }
   1165 
   1166 template void MemoryAllocator::Free<MemoryAllocator::kFull>(MemoryChunk* chunk);
   1167 
   1168 template void MemoryAllocator::Free<MemoryAllocator::kAlreadyPooled>(
   1169     MemoryChunk* chunk);
   1170 
   1171 template void MemoryAllocator::Free<MemoryAllocator::kPreFreeAndQueue>(
   1172     MemoryChunk* chunk);
   1173 
   1174 template void MemoryAllocator::Free<MemoryAllocator::kPooledAndQueue>(
   1175     MemoryChunk* chunk);
   1176 
   1177 template <MemoryAllocator::AllocationMode alloc_mode, typename SpaceType>
   1178 Page* MemoryAllocator::AllocatePage(size_t size, SpaceType* owner,
   1179                                     Executability executable) {
   1180   MemoryChunk* chunk = nullptr;
   1181   if (alloc_mode == kPooled) {
   1182     DCHECK_EQ(size, static_cast<size_t>(MemoryChunk::kAllocatableMemory));
   1183     DCHECK_EQ(executable, NOT_EXECUTABLE);
   1184     chunk = AllocatePagePooled(owner);
   1185   }
   1186   if (chunk == nullptr) {
   1187     chunk = AllocateChunk(size, size, executable, owner);
   1188   }
   1189   if (chunk == nullptr) return nullptr;
   1190   return owner->InitializePage(chunk, executable);
   1191 }
   1192 
   1193 template Page*
   1194 MemoryAllocator::AllocatePage<MemoryAllocator::kRegular, PagedSpace>(
   1195     size_t size, PagedSpace* owner, Executability executable);
   1196 template Page*
   1197 MemoryAllocator::AllocatePage<MemoryAllocator::kRegular, SemiSpace>(
   1198     size_t size, SemiSpace* owner, Executability executable);
   1199 template Page*
   1200 MemoryAllocator::AllocatePage<MemoryAllocator::kPooled, SemiSpace>(
   1201     size_t size, SemiSpace* owner, Executability executable);
   1202 
   1203 LargePage* MemoryAllocator::AllocateLargePage(size_t size,
   1204                                               LargeObjectSpace* owner,
   1205                                               Executability executable) {
   1206   MemoryChunk* chunk = AllocateChunk(size, size, executable, owner);
   1207   if (chunk == nullptr) return nullptr;
   1208   return LargePage::Initialize(isolate_->heap(), chunk, executable);
   1209 }
   1210 
   1211 template <typename SpaceType>
   1212 MemoryChunk* MemoryAllocator::AllocatePagePooled(SpaceType* owner) {
   1213   MemoryChunk* chunk = unmapper()->TryGetPooledMemoryChunkSafe();
   1214   if (chunk == nullptr) return nullptr;
   1215   const int size = MemoryChunk::kPageSize;
   1216   const Address start = reinterpret_cast<Address>(chunk);
   1217   const Address area_start = start + MemoryChunk::kObjectStartOffset;
   1218   const Address area_end = start + size;
   1219   if (!CommitBlock(start, size)) {
   1220     return nullptr;
   1221   }
   1222   VirtualMemory reservation(start, size);
   1223   MemoryChunk::Initialize(isolate_->heap(), start, size, area_start, area_end,
   1224                           NOT_EXECUTABLE, owner, &reservation);
   1225   size_ += size;
   1226   return chunk;
   1227 }
   1228 
   1229 bool MemoryAllocator::CommitBlock(Address start, size_t size) {
   1230   if (!CommitMemory(start, size)) return false;
   1231 
   1232   if (Heap::ShouldZapGarbage()) {
   1233     ZapBlock(start, size, kZapValue);
   1234   }
   1235 
   1236   isolate_->counters()->memory_allocated()->Increment(static_cast<int>(size));
   1237   return true;
   1238 }
   1239 
   1240 
   1241 bool MemoryAllocator::UncommitBlock(Address start, size_t size) {
   1242   if (!SetPermissions(start, size, PageAllocator::kNoAccess)) return false;
   1243   isolate_->counters()->memory_allocated()->Decrement(static_cast<int>(size));
   1244   return true;
   1245 }
   1246 
   1247 void MemoryAllocator::ZapBlock(Address start, size_t size,
   1248                                uintptr_t zap_value) {
   1249   DCHECK_EQ(start % kPointerSize, 0);
   1250   DCHECK_EQ(size % kPointerSize, 0);
   1251   for (size_t s = 0; s + kPointerSize <= size; s += kPointerSize) {
   1252     Memory<Address>(start + s) = static_cast<Address>(zap_value);
   1253   }
   1254 }
   1255 
   1256 size_t MemoryAllocator::CodePageGuardStartOffset() {
   1257   // We are guarding code pages: the first OS page after the header
   1258   // will be protected as non-writable.
   1259   return ::RoundUp(Page::kObjectStartOffset, GetCommitPageSize());
   1260 }
   1261 
   1262 size_t MemoryAllocator::CodePageGuardSize() { return GetCommitPageSize(); }
   1263 
   1264 size_t MemoryAllocator::CodePageAreaStartOffset() {
   1265   // We are guarding code pages: the first OS page after the header
   1266   // will be protected as non-writable.
   1267   return CodePageGuardStartOffset() + CodePageGuardSize();
   1268 }
   1269 
   1270 size_t MemoryAllocator::CodePageAreaEndOffset() {
   1271   // We are guarding code pages: the last OS page will be protected as
   1272   // non-writable.
   1273   return Page::kPageSize - static_cast<int>(GetCommitPageSize());
   1274 }
   1275 
   1276 intptr_t MemoryAllocator::GetCommitPageSize() {
   1277   if (FLAG_v8_os_page_size != 0) {
   1278     DCHECK(base::bits::IsPowerOfTwo(FLAG_v8_os_page_size));
   1279     return FLAG_v8_os_page_size * KB;
   1280   } else {
   1281     return CommitPageSize();
   1282   }
   1283 }
   1284 
   1285 bool MemoryAllocator::CommitExecutableMemory(VirtualMemory* vm, Address start,
   1286                                              size_t commit_size,
   1287                                              size_t reserved_size) {
   1288   const size_t page_size = GetCommitPageSize();
   1289   // All addresses and sizes must be aligned to the commit page size.
   1290   DCHECK(IsAddressAligned(start, page_size));
   1291   DCHECK_EQ(0, commit_size % page_size);
   1292   DCHECK_EQ(0, reserved_size % page_size);
   1293   const size_t guard_size = CodePageGuardSize();
   1294   const size_t pre_guard_offset = CodePageGuardStartOffset();
   1295   const size_t code_area_offset = CodePageAreaStartOffset();
   1296   // reserved_size includes two guard regions, commit_size does not.
   1297   DCHECK_LE(commit_size, reserved_size - 2 * guard_size);
   1298   const Address pre_guard_page = start + pre_guard_offset;
   1299   const Address code_area = start + code_area_offset;
   1300   const Address post_guard_page = start + reserved_size - guard_size;
   1301   // Commit the non-executable header, from start to pre-code guard page.
   1302   if (vm->SetPermissions(start, pre_guard_offset, PageAllocator::kReadWrite)) {
   1303     // Create the pre-code guard page, following the header.
   1304     if (vm->SetPermissions(pre_guard_page, page_size,
   1305                            PageAllocator::kNoAccess)) {
   1306       // Commit the executable code body.
   1307       if (vm->SetPermissions(code_area, commit_size - pre_guard_offset,
   1308                              PageAllocator::kReadWrite)) {
   1309         // Create the post-code guard page.
   1310         if (vm->SetPermissions(post_guard_page, page_size,
   1311                                PageAllocator::kNoAccess)) {
   1312           UpdateAllocatedSpaceLimits(start, code_area + commit_size);
   1313           return true;
   1314         }
   1315         vm->SetPermissions(code_area, commit_size, PageAllocator::kNoAccess);
   1316       }
   1317     }
   1318     vm->SetPermissions(start, pre_guard_offset, PageAllocator::kNoAccess);
   1319   }
   1320   return false;
   1321 }
   1322 
   1323 
   1324 // -----------------------------------------------------------------------------
   1325 // MemoryChunk implementation
   1326 
   1327 void MemoryChunk::ReleaseAllocatedMemory() {
   1328   if (skip_list_ != nullptr) {
   1329     delete skip_list_;
   1330     skip_list_ = nullptr;
   1331   }
   1332   if (mutex_ != nullptr) {
   1333     delete mutex_;
   1334     mutex_ = nullptr;
   1335   }
   1336   if (page_protection_change_mutex_ != nullptr) {
   1337     delete page_protection_change_mutex_;
   1338     page_protection_change_mutex_ = nullptr;
   1339   }
   1340   ReleaseSlotSet<OLD_TO_NEW>();
   1341   ReleaseSlotSet<OLD_TO_OLD>();
   1342   ReleaseTypedSlotSet<OLD_TO_NEW>();
   1343   ReleaseTypedSlotSet<OLD_TO_OLD>();
   1344   ReleaseInvalidatedSlots();
   1345   if (local_tracker_ != nullptr) ReleaseLocalTracker();
   1346   if (young_generation_bitmap_ != nullptr) ReleaseYoungGenerationBitmap();
   1347 
   1348   if (IsPagedSpace()) {
   1349     Page* page = static_cast<Page*>(this);
   1350     page->ReleaseFreeListCategories();
   1351   }
   1352 }
   1353 
   1354 static SlotSet* AllocateAndInitializeSlotSet(size_t size, Address page_start) {
   1355   size_t pages = (size + Page::kPageSize - 1) / Page::kPageSize;
   1356   DCHECK_LT(0, pages);
   1357   SlotSet* slot_set = new SlotSet[pages];
   1358   for (size_t i = 0; i < pages; i++) {
   1359     slot_set[i].SetPageStart(page_start + i * Page::kPageSize);
   1360   }
   1361   return slot_set;
   1362 }
   1363 
   1364 template SlotSet* MemoryChunk::AllocateSlotSet<OLD_TO_NEW>();
   1365 template SlotSet* MemoryChunk::AllocateSlotSet<OLD_TO_OLD>();
   1366 
   1367 template <RememberedSetType type>
   1368 SlotSet* MemoryChunk::AllocateSlotSet() {
   1369   SlotSet* slot_set = AllocateAndInitializeSlotSet(size_, address());
   1370   SlotSet* old_slot_set = base::AsAtomicPointer::Release_CompareAndSwap(
   1371       &slot_set_[type], nullptr, slot_set);
   1372   if (old_slot_set != nullptr) {
   1373     delete[] slot_set;
   1374     slot_set = old_slot_set;
   1375   }
   1376   DCHECK(slot_set);
   1377   return slot_set;
   1378 }
   1379 
   1380 template void MemoryChunk::ReleaseSlotSet<OLD_TO_NEW>();
   1381 template void MemoryChunk::ReleaseSlotSet<OLD_TO_OLD>();
   1382 
   1383 template <RememberedSetType type>
   1384 void MemoryChunk::ReleaseSlotSet() {
   1385   SlotSet* slot_set = slot_set_[type];
   1386   if (slot_set) {
   1387     slot_set_[type] = nullptr;
   1388     delete[] slot_set;
   1389   }
   1390 }
   1391 
   1392 template TypedSlotSet* MemoryChunk::AllocateTypedSlotSet<OLD_TO_NEW>();
   1393 template TypedSlotSet* MemoryChunk::AllocateTypedSlotSet<OLD_TO_OLD>();
   1394 
   1395 template <RememberedSetType type>
   1396 TypedSlotSet* MemoryChunk::AllocateTypedSlotSet() {
   1397   TypedSlotSet* typed_slot_set = new TypedSlotSet(address());
   1398   TypedSlotSet* old_value = base::AsAtomicPointer::Release_CompareAndSwap(
   1399       &typed_slot_set_[type], nullptr, typed_slot_set);
   1400   if (old_value != nullptr) {
   1401     delete typed_slot_set;
   1402     typed_slot_set = old_value;
   1403   }
   1404   DCHECK(typed_slot_set);
   1405   return typed_slot_set;
   1406 }
   1407 
   1408 template void MemoryChunk::ReleaseTypedSlotSet<OLD_TO_NEW>();
   1409 template void MemoryChunk::ReleaseTypedSlotSet<OLD_TO_OLD>();
   1410 
   1411 template <RememberedSetType type>
   1412 void MemoryChunk::ReleaseTypedSlotSet() {
   1413   TypedSlotSet* typed_slot_set = typed_slot_set_[type];
   1414   if (typed_slot_set) {
   1415     typed_slot_set_[type] = nullptr;
   1416     delete typed_slot_set;
   1417   }
   1418 }
   1419 
   1420 InvalidatedSlots* MemoryChunk::AllocateInvalidatedSlots() {
   1421   DCHECK_NULL(invalidated_slots_);
   1422   invalidated_slots_ = new InvalidatedSlots();
   1423   return invalidated_slots_;
   1424 }
   1425 
   1426 void MemoryChunk::ReleaseInvalidatedSlots() {
   1427   if (invalidated_slots_) {
   1428     delete invalidated_slots_;
   1429     invalidated_slots_ = nullptr;
   1430   }
   1431 }
   1432 
   1433 void MemoryChunk::RegisterObjectWithInvalidatedSlots(HeapObject* object,
   1434                                                      int size) {
   1435   if (!ShouldSkipEvacuationSlotRecording()) {
   1436     if (invalidated_slots() == nullptr) {
   1437       AllocateInvalidatedSlots();
   1438     }
   1439     int old_size = (*invalidated_slots())[object];
   1440     (*invalidated_slots())[object] = std::max(old_size, size);
   1441   }
   1442 }
   1443 
   1444 void MemoryChunk::MoveObjectWithInvalidatedSlots(HeapObject* old_start,
   1445                                                  HeapObject* new_start) {
   1446   DCHECK_LT(old_start, new_start);
   1447   DCHECK_EQ(MemoryChunk::FromHeapObject(old_start),
   1448             MemoryChunk::FromHeapObject(new_start));
   1449   if (!ShouldSkipEvacuationSlotRecording() && invalidated_slots()) {
   1450     auto it = invalidated_slots()->find(old_start);
   1451     if (it != invalidated_slots()->end()) {
   1452       int old_size = it->second;
   1453       int delta = static_cast<int>(new_start->address() - old_start->address());
   1454       invalidated_slots()->erase(it);
   1455       (*invalidated_slots())[new_start] = old_size - delta;
   1456     }
   1457   }
   1458 }
   1459 
   1460 void MemoryChunk::ReleaseLocalTracker() {
   1461   DCHECK_NOT_NULL(local_tracker_);
   1462   delete local_tracker_;
   1463   local_tracker_ = nullptr;
   1464 }
   1465 
   1466 void MemoryChunk::AllocateYoungGenerationBitmap() {
   1467   DCHECK_NULL(young_generation_bitmap_);
   1468   young_generation_bitmap_ = static_cast<Bitmap*>(calloc(1, Bitmap::kSize));
   1469 }
   1470 
   1471 void MemoryChunk::ReleaseYoungGenerationBitmap() {
   1472   DCHECK_NOT_NULL(young_generation_bitmap_);
   1473   free(young_generation_bitmap_);
   1474   young_generation_bitmap_ = nullptr;
   1475 }
   1476 
   1477 void MemoryChunk::IncrementExternalBackingStoreBytes(
   1478     ExternalBackingStoreType type, size_t amount) {
   1479   external_backing_store_bytes_[type] += amount;
   1480   owner()->IncrementExternalBackingStoreBytes(type, amount);
   1481 }
   1482 
   1483 void MemoryChunk::DecrementExternalBackingStoreBytes(
   1484     ExternalBackingStoreType type, size_t amount) {
   1485   DCHECK_GE(external_backing_store_bytes_[type], amount);
   1486   external_backing_store_bytes_[type] -= amount;
   1487   owner()->DecrementExternalBackingStoreBytes(type, amount);
   1488 }
   1489 
   1490 // -----------------------------------------------------------------------------
   1491 // PagedSpace implementation
   1492 
   1493 void Space::AddAllocationObserver(AllocationObserver* observer) {
   1494   allocation_observers_.push_back(observer);
   1495   StartNextInlineAllocationStep();
   1496 }
   1497 
   1498 void Space::RemoveAllocationObserver(AllocationObserver* observer) {
   1499   auto it = std::find(allocation_observers_.begin(),
   1500                       allocation_observers_.end(), observer);
   1501   DCHECK(allocation_observers_.end() != it);
   1502   allocation_observers_.erase(it);
   1503   StartNextInlineAllocationStep();
   1504 }
   1505 
   1506 void Space::PauseAllocationObservers() { allocation_observers_paused_ = true; }
   1507 
   1508 void Space::ResumeAllocationObservers() {
   1509   allocation_observers_paused_ = false;
   1510 }
   1511 
   1512 void Space::AllocationStep(int bytes_since_last, Address soon_object,
   1513                            int size) {
   1514   if (!AllocationObserversActive()) {
   1515     return;
   1516   }
   1517 
   1518   DCHECK(!heap()->allocation_step_in_progress());
   1519   heap()->set_allocation_step_in_progress(true);
   1520   heap()->CreateFillerObjectAt(soon_object, size, ClearRecordedSlots::kNo);
   1521   for (AllocationObserver* observer : allocation_observers_) {
   1522     observer->AllocationStep(bytes_since_last, soon_object, size);
   1523   }
   1524   heap()->set_allocation_step_in_progress(false);
   1525 }
   1526 
   1527 intptr_t Space::GetNextInlineAllocationStepSize() {
   1528   intptr_t next_step = 0;
   1529   for (AllocationObserver* observer : allocation_observers_) {
   1530     next_step = next_step ? Min(next_step, observer->bytes_to_next_step())
   1531                           : observer->bytes_to_next_step();
   1532   }
   1533   DCHECK(allocation_observers_.size() == 0 || next_step > 0);
   1534   return next_step;
   1535 }
   1536 
   1537 PagedSpace::PagedSpace(Heap* heap, AllocationSpace space,
   1538                        Executability executable)
   1539     : SpaceWithLinearArea(heap, space), executable_(executable) {
   1540   area_size_ = MemoryAllocator::PageAreaSize(space);
   1541   accounting_stats_.Clear();
   1542 }
   1543 
   1544 void PagedSpace::TearDown() {
   1545   while (!memory_chunk_list_.Empty()) {
   1546     MemoryChunk* chunk = memory_chunk_list_.front();
   1547     memory_chunk_list_.Remove(chunk);
   1548     heap()->memory_allocator()->Free<MemoryAllocator::kFull>(chunk);
   1549   }
   1550   accounting_stats_.Clear();
   1551 }
   1552 
   1553 void PagedSpace::RefillFreeList() {
   1554   // Any PagedSpace might invoke RefillFreeList. We filter all but our old
   1555   // generation spaces out.
   1556   if (identity() != OLD_SPACE && identity() != CODE_SPACE &&
   1557       identity() != MAP_SPACE && identity() != RO_SPACE) {
   1558     return;
   1559   }
   1560   MarkCompactCollector* collector = heap()->mark_compact_collector();
   1561   size_t added = 0;
   1562   {
   1563     Page* p = nullptr;
   1564     while ((p = collector->sweeper()->GetSweptPageSafe(this)) != nullptr) {
   1565       // Only during compaction pages can actually change ownership. This is
   1566       // safe because there exists no other competing action on the page links
   1567       // during compaction.
   1568       if (is_local()) {
   1569         DCHECK_NE(this, p->owner());
   1570         PagedSpace* owner = reinterpret_cast<PagedSpace*>(p->owner());
   1571         base::LockGuard<base::Mutex> guard(owner->mutex());
   1572         owner->RefineAllocatedBytesAfterSweeping(p);
   1573         owner->RemovePage(p);
   1574         added += AddPage(p);
   1575       } else {
   1576         base::LockGuard<base::Mutex> guard(mutex());
   1577         DCHECK_EQ(this, p->owner());
   1578         RefineAllocatedBytesAfterSweeping(p);
   1579         added += RelinkFreeListCategories(p);
   1580       }
   1581       added += p->wasted_memory();
   1582       if (is_local() && (added > kCompactionMemoryWanted)) break;
   1583     }
   1584   }
   1585 }
   1586 
   1587 void PagedSpace::MergeCompactionSpace(CompactionSpace* other) {
   1588   base::LockGuard<base::Mutex> guard(mutex());
   1589 
   1590   DCHECK(identity() == other->identity());
   1591   // Unmerged fields:
   1592   //   area_size_
   1593   other->FreeLinearAllocationArea();
   1594 
   1595   // The linear allocation area of {other} should be destroyed now.
   1596   DCHECK_EQ(kNullAddress, other->top());
   1597   DCHECK_EQ(kNullAddress, other->limit());
   1598 
   1599   // Move over pages.
   1600   for (auto it = other->begin(); it != other->end();) {
   1601     Page* p = *(it++);
   1602     // Relinking requires the category to be unlinked.
   1603     other->RemovePage(p);
   1604     AddPage(p);
   1605     DCHECK_EQ(p->AvailableInFreeList(),
   1606               p->AvailableInFreeListFromAllocatedBytes());
   1607   }
   1608   DCHECK_EQ(0u, other->Size());
   1609   DCHECK_EQ(0u, other->Capacity());
   1610 }
   1611 
   1612 
   1613 size_t PagedSpace::CommittedPhysicalMemory() {
   1614   if (!base::OS::HasLazyCommits()) return CommittedMemory();
   1615   MemoryChunk::UpdateHighWaterMark(allocation_info_.top());
   1616   size_t size = 0;
   1617   for (Page* page : *this) {
   1618     size += page->CommittedPhysicalMemory();
   1619   }
   1620   return size;
   1621 }
   1622 
   1623 bool PagedSpace::ContainsSlow(Address addr) {
   1624   Page* p = Page::FromAddress(addr);
   1625   for (Page* page : *this) {
   1626     if (page == p) return true;
   1627   }
   1628   return false;
   1629 }
   1630 
   1631 void PagedSpace::RefineAllocatedBytesAfterSweeping(Page* page) {
   1632   CHECK(page->SweepingDone());
   1633   auto marking_state =
   1634       heap()->incremental_marking()->non_atomic_marking_state();
   1635   // The live_byte on the page was accounted in the space allocated
   1636   // bytes counter. After sweeping allocated_bytes() contains the
   1637   // accurate live byte count on the page.
   1638   size_t old_counter = marking_state->live_bytes(page);
   1639   size_t new_counter = page->allocated_bytes();
   1640   DCHECK_GE(old_counter, new_counter);
   1641   if (old_counter > new_counter) {
   1642     DecreaseAllocatedBytes(old_counter - new_counter, page);
   1643     // Give the heap a chance to adjust counters in response to the
   1644     // more precise and smaller old generation size.
   1645     heap()->NotifyRefinedOldGenerationSize(old_counter - new_counter);
   1646   }
   1647   marking_state->SetLiveBytes(page, 0);
   1648 }
   1649 
   1650 Page* PagedSpace::RemovePageSafe(int size_in_bytes) {
   1651   base::LockGuard<base::Mutex> guard(mutex());
   1652   // Check for pages that still contain free list entries. Bail out for smaller
   1653   // categories.
   1654   const int minimum_category =
   1655       static_cast<int>(FreeList::SelectFreeListCategoryType(size_in_bytes));
   1656   Page* page = free_list()->GetPageForCategoryType(kHuge);
   1657   if (!page && static_cast<int>(kLarge) >= minimum_category)
   1658     page = free_list()->GetPageForCategoryType(kLarge);
   1659   if (!page && static_cast<int>(kMedium) >= minimum_category)
   1660     page = free_list()->GetPageForCategoryType(kMedium);
   1661   if (!page && static_cast<int>(kSmall) >= minimum_category)
   1662     page = free_list()->GetPageForCategoryType(kSmall);
   1663   if (!page && static_cast<int>(kTiny) >= minimum_category)
   1664     page = free_list()->GetPageForCategoryType(kTiny);
   1665   if (!page && static_cast<int>(kTiniest) >= minimum_category)
   1666     page = free_list()->GetPageForCategoryType(kTiniest);
   1667   if (!page) return nullptr;
   1668   RemovePage(page);
   1669   return page;
   1670 }
   1671 
   1672 size_t PagedSpace::AddPage(Page* page) {
   1673   CHECK(page->SweepingDone());
   1674   page->set_owner(this);
   1675   memory_chunk_list_.PushBack(page);
   1676   AccountCommitted(page->size());
   1677   IncreaseCapacity(page->area_size());
   1678   IncreaseAllocatedBytes(page->allocated_bytes(), page);
   1679   for (size_t i = 0; i < ExternalBackingStoreType::kNumTypes; i++) {
   1680     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   1681     IncrementExternalBackingStoreBytes(t, page->ExternalBackingStoreBytes(t));
   1682   }
   1683   return RelinkFreeListCategories(page);
   1684 }
   1685 
   1686 void PagedSpace::RemovePage(Page* page) {
   1687   CHECK(page->SweepingDone());
   1688   memory_chunk_list_.Remove(page);
   1689   UnlinkFreeListCategories(page);
   1690   DecreaseAllocatedBytes(page->allocated_bytes(), page);
   1691   DecreaseCapacity(page->area_size());
   1692   AccountUncommitted(page->size());
   1693   for (size_t i = 0; i < ExternalBackingStoreType::kNumTypes; i++) {
   1694     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   1695     DecrementExternalBackingStoreBytes(t, page->ExternalBackingStoreBytes(t));
   1696   }
   1697 }
   1698 
   1699 size_t PagedSpace::ShrinkPageToHighWaterMark(Page* page) {
   1700   size_t unused = page->ShrinkToHighWaterMark();
   1701   accounting_stats_.DecreaseCapacity(static_cast<intptr_t>(unused));
   1702   AccountUncommitted(unused);
   1703   return unused;
   1704 }
   1705 
   1706 void PagedSpace::ResetFreeList() {
   1707   for (Page* page : *this) {
   1708     free_list_.EvictFreeListItems(page);
   1709   }
   1710   DCHECK(free_list_.IsEmpty());
   1711 }
   1712 
   1713 void PagedSpace::ShrinkImmortalImmovablePages() {
   1714   DCHECK(!heap()->deserialization_complete());
   1715   MemoryChunk::UpdateHighWaterMark(allocation_info_.top());
   1716   FreeLinearAllocationArea();
   1717   ResetFreeList();
   1718   for (Page* page : *this) {
   1719     DCHECK(page->IsFlagSet(Page::NEVER_EVACUATE));
   1720     ShrinkPageToHighWaterMark(page);
   1721   }
   1722 }
   1723 
   1724 bool PagedSpace::Expand() {
   1725   // Always lock against the main space as we can only adjust capacity and
   1726   // pages concurrently for the main paged space.
   1727   base::LockGuard<base::Mutex> guard(heap()->paged_space(identity())->mutex());
   1728 
   1729   const int size = AreaSize();
   1730 
   1731   if (!heap()->CanExpandOldGeneration(size)) return false;
   1732 
   1733   Page* page =
   1734       heap()->memory_allocator()->AllocatePage(size, this, executable());
   1735   if (page == nullptr) return false;
   1736   // Pages created during bootstrapping may contain immortal immovable objects.
   1737   if (!heap()->deserialization_complete()) page->MarkNeverEvacuate();
   1738   AddPage(page);
   1739   Free(page->area_start(), page->area_size(),
   1740        SpaceAccountingMode::kSpaceAccounted);
   1741   return true;
   1742 }
   1743 
   1744 
   1745 int PagedSpace::CountTotalPages() {
   1746   int count = 0;
   1747   for (Page* page : *this) {
   1748     count++;
   1749     USE(page);
   1750   }
   1751   return count;
   1752 }
   1753 
   1754 
   1755 void PagedSpace::ResetFreeListStatistics() {
   1756   for (Page* page : *this) {
   1757     page->ResetFreeListStatistics();
   1758   }
   1759 }
   1760 
   1761 void PagedSpace::SetLinearAllocationArea(Address top, Address limit) {
   1762   SetTopAndLimit(top, limit);
   1763   if (top != kNullAddress && top != limit &&
   1764       heap()->incremental_marking()->black_allocation()) {
   1765     Page::FromAllocationAreaAddress(top)->CreateBlackArea(top, limit);
   1766   }
   1767 }
   1768 
   1769 void PagedSpace::DecreaseLimit(Address new_limit) {
   1770   Address old_limit = limit();
   1771   DCHECK_LE(top(), new_limit);
   1772   DCHECK_GE(old_limit, new_limit);
   1773   if (new_limit != old_limit) {
   1774     SetTopAndLimit(top(), new_limit);
   1775     Free(new_limit, old_limit - new_limit,
   1776          SpaceAccountingMode::kSpaceAccounted);
   1777     if (heap()->incremental_marking()->black_allocation()) {
   1778       Page::FromAllocationAreaAddress(new_limit)->DestroyBlackArea(new_limit,
   1779                                                                    old_limit);
   1780     }
   1781   }
   1782 }
   1783 
   1784 Address SpaceWithLinearArea::ComputeLimit(Address start, Address end,
   1785                                           size_t min_size) {
   1786   DCHECK_GE(end - start, min_size);
   1787 
   1788   if (heap()->inline_allocation_disabled()) {
   1789     // Fit the requested area exactly.
   1790     return start + min_size;
   1791   } else if (SupportsInlineAllocation() && AllocationObserversActive()) {
   1792     // Generated code may allocate inline from the linear allocation area for.
   1793     // To make sure we can observe these allocations, we use a lower limit.
   1794     size_t step = GetNextInlineAllocationStepSize();
   1795 
   1796     // TODO(ofrobots): there is subtle difference between old space and new
   1797     // space here. Any way to avoid it? `step - 1` makes more sense as we would
   1798     // like to sample the object that straddles the `start + step` boundary.
   1799     // Rounding down further would introduce a small statistical error in
   1800     // sampling. However, presently PagedSpace requires limit to be aligned.
   1801     size_t rounded_step;
   1802     if (identity() == NEW_SPACE) {
   1803       DCHECK_GE(step, 1);
   1804       rounded_step = step - 1;
   1805     } else {
   1806       rounded_step = RoundSizeDownToObjectAlignment(static_cast<int>(step));
   1807     }
   1808     return Min(static_cast<Address>(start + min_size + rounded_step), end);
   1809   } else {
   1810     // The entire node can be used as the linear allocation area.
   1811     return end;
   1812   }
   1813 }
   1814 
   1815 void PagedSpace::MarkLinearAllocationAreaBlack() {
   1816   DCHECK(heap()->incremental_marking()->black_allocation());
   1817   Address current_top = top();
   1818   Address current_limit = limit();
   1819   if (current_top != kNullAddress && current_top != current_limit) {
   1820     Page::FromAllocationAreaAddress(current_top)
   1821         ->CreateBlackArea(current_top, current_limit);
   1822   }
   1823 }
   1824 
   1825 void PagedSpace::UnmarkLinearAllocationArea() {
   1826   Address current_top = top();
   1827   Address current_limit = limit();
   1828   if (current_top != kNullAddress && current_top != current_limit) {
   1829     Page::FromAllocationAreaAddress(current_top)
   1830         ->DestroyBlackArea(current_top, current_limit);
   1831   }
   1832 }
   1833 
   1834 void PagedSpace::FreeLinearAllocationArea() {
   1835   // Mark the old linear allocation area with a free space map so it can be
   1836   // skipped when scanning the heap.
   1837   Address current_top = top();
   1838   Address current_limit = limit();
   1839   if (current_top == kNullAddress) {
   1840     DCHECK_EQ(kNullAddress, current_limit);
   1841     return;
   1842   }
   1843 
   1844   if (heap()->incremental_marking()->black_allocation()) {
   1845     Page* page = Page::FromAllocationAreaAddress(current_top);
   1846 
   1847     // Clear the bits in the unused black area.
   1848     if (current_top != current_limit) {
   1849       IncrementalMarking::MarkingState* marking_state =
   1850           heap()->incremental_marking()->marking_state();
   1851       marking_state->bitmap(page)->ClearRange(
   1852           page->AddressToMarkbitIndex(current_top),
   1853           page->AddressToMarkbitIndex(current_limit));
   1854       marking_state->IncrementLiveBytes(
   1855           page, -static_cast<int>(current_limit - current_top));
   1856     }
   1857   }
   1858 
   1859   InlineAllocationStep(current_top, kNullAddress, kNullAddress, 0);
   1860   SetTopAndLimit(kNullAddress, kNullAddress);
   1861   DCHECK_GE(current_limit, current_top);
   1862 
   1863   // The code page of the linear allocation area needs to be unprotected
   1864   // because we are going to write a filler into that memory area below.
   1865   if (identity() == CODE_SPACE) {
   1866     heap()->UnprotectAndRegisterMemoryChunk(
   1867         MemoryChunk::FromAddress(current_top));
   1868   }
   1869   Free(current_top, current_limit - current_top,
   1870        SpaceAccountingMode::kSpaceAccounted);
   1871 }
   1872 
   1873 void PagedSpace::ReleasePage(Page* page) {
   1874   DCHECK_EQ(
   1875       0, heap()->incremental_marking()->non_atomic_marking_state()->live_bytes(
   1876              page));
   1877   DCHECK_EQ(page->owner(), this);
   1878 
   1879   free_list_.EvictFreeListItems(page);
   1880   DCHECK(!free_list_.ContainsPageFreeListItems(page));
   1881 
   1882   if (Page::FromAllocationAreaAddress(allocation_info_.top()) == page) {
   1883     DCHECK(!top_on_previous_step_);
   1884     allocation_info_.Reset(kNullAddress, kNullAddress);
   1885   }
   1886 
   1887   AccountUncommitted(page->size());
   1888   accounting_stats_.DecreaseCapacity(page->area_size());
   1889   heap()->memory_allocator()->Free<MemoryAllocator::kPreFreeAndQueue>(page);
   1890 }
   1891 
   1892 void PagedSpace::SetReadAndExecutable() {
   1893   DCHECK(identity() == CODE_SPACE);
   1894   for (Page* page : *this) {
   1895     CHECK(heap()->memory_allocator()->IsMemoryChunkExecutable(page));
   1896     page->SetReadAndExecutable();
   1897   }
   1898 }
   1899 
   1900 void PagedSpace::SetReadAndWritable() {
   1901   DCHECK(identity() == CODE_SPACE);
   1902   for (Page* page : *this) {
   1903     CHECK(heap()->memory_allocator()->IsMemoryChunkExecutable(page));
   1904     page->SetReadAndWritable();
   1905   }
   1906 }
   1907 
   1908 std::unique_ptr<ObjectIterator> PagedSpace::GetObjectIterator() {
   1909   return std::unique_ptr<ObjectIterator>(new HeapObjectIterator(this));
   1910 }
   1911 
   1912 bool PagedSpace::RefillLinearAllocationAreaFromFreeList(size_t size_in_bytes) {
   1913   DCHECK(IsAligned(size_in_bytes, kPointerSize));
   1914   DCHECK_LE(top(), limit());
   1915 #ifdef DEBUG
   1916   if (top() != limit()) {
   1917     DCHECK_EQ(Page::FromAddress(top()), Page::FromAddress(limit() - 1));
   1918   }
   1919 #endif
   1920   // Don't free list allocate if there is linear space available.
   1921   DCHECK_LT(static_cast<size_t>(limit() - top()), size_in_bytes);
   1922 
   1923   // Mark the old linear allocation area with a free space map so it can be
   1924   // skipped when scanning the heap.  This also puts it back in the free list
   1925   // if it is big enough.
   1926   FreeLinearAllocationArea();
   1927 
   1928   if (!is_local()) {
   1929     heap()->StartIncrementalMarkingIfAllocationLimitIsReached(
   1930         heap()->GCFlagsForIncrementalMarking(),
   1931         kGCCallbackScheduleIdleGarbageCollection);
   1932   }
   1933 
   1934   size_t new_node_size = 0;
   1935   FreeSpace* new_node = free_list_.Allocate(size_in_bytes, &new_node_size);
   1936   if (new_node == nullptr) return false;
   1937 
   1938   DCHECK_GE(new_node_size, size_in_bytes);
   1939 
   1940   // The old-space-step might have finished sweeping and restarted marking.
   1941   // Verify that it did not turn the page of the new node into an evacuation
   1942   // candidate.
   1943   DCHECK(!MarkCompactCollector::IsOnEvacuationCandidate(new_node));
   1944 
   1945   // Memory in the linear allocation area is counted as allocated.  We may free
   1946   // a little of this again immediately - see below.
   1947   Page* page = Page::FromAddress(new_node->address());
   1948   IncreaseAllocatedBytes(new_node_size, page);
   1949 
   1950   Address start = new_node->address();
   1951   Address end = new_node->address() + new_node_size;
   1952   Address limit = ComputeLimit(start, end, size_in_bytes);
   1953   DCHECK_LE(limit, end);
   1954   DCHECK_LE(size_in_bytes, limit - start);
   1955   if (limit != end) {
   1956     if (identity() == CODE_SPACE) {
   1957       heap()->UnprotectAndRegisterMemoryChunk(page);
   1958     }
   1959     Free(limit, end - limit, SpaceAccountingMode::kSpaceAccounted);
   1960   }
   1961   SetLinearAllocationArea(start, limit);
   1962 
   1963   return true;
   1964 }
   1965 
   1966 #ifdef DEBUG
   1967 void PagedSpace::Print() {}
   1968 #endif
   1969 
   1970 #ifdef VERIFY_HEAP
   1971 void PagedSpace::Verify(Isolate* isolate, ObjectVisitor* visitor) {
   1972   bool allocation_pointer_found_in_space =
   1973       (allocation_info_.top() == allocation_info_.limit());
   1974   size_t external_space_bytes[kNumTypes];
   1975   size_t external_page_bytes[kNumTypes];
   1976 
   1977   for (int i = 0; i < kNumTypes; i++) {
   1978     external_space_bytes[static_cast<ExternalBackingStoreType>(i)] = 0;
   1979   }
   1980 
   1981   for (Page* page : *this) {
   1982     CHECK(page->owner() == this);
   1983 
   1984     for (int i = 0; i < kNumTypes; i++) {
   1985       external_page_bytes[static_cast<ExternalBackingStoreType>(i)] = 0;
   1986     }
   1987 
   1988     if (page == Page::FromAllocationAreaAddress(allocation_info_.top())) {
   1989       allocation_pointer_found_in_space = true;
   1990     }
   1991     CHECK(page->SweepingDone());
   1992     HeapObjectIterator it(page);
   1993     Address end_of_previous_object = page->area_start();
   1994     Address top = page->area_end();
   1995 
   1996     for (HeapObject* object = it.Next(); object != nullptr;
   1997          object = it.Next()) {
   1998       CHECK(end_of_previous_object <= object->address());
   1999 
   2000       // The first word should be a map, and we expect all map pointers to
   2001       // be in map space.
   2002       Map* map = object->map();
   2003       CHECK(map->IsMap());
   2004       CHECK(heap()->map_space()->Contains(map) ||
   2005             heap()->read_only_space()->Contains(map));
   2006 
   2007       // Perform space-specific object verification.
   2008       VerifyObject(object);
   2009 
   2010       // The object itself should look OK.
   2011       object->ObjectVerify(isolate);
   2012 
   2013       if (!FLAG_verify_heap_skip_remembered_set) {
   2014         heap()->VerifyRememberedSetFor(object);
   2015       }
   2016 
   2017       // All the interior pointers should be contained in the heap.
   2018       int size = object->Size();
   2019       object->IterateBody(map, size, visitor);
   2020       CHECK(object->address() + size <= top);
   2021       end_of_previous_object = object->address() + size;
   2022 
   2023       if (object->IsExternalString()) {
   2024         ExternalString* external_string = ExternalString::cast(object);
   2025         size_t size = external_string->ExternalPayloadSize();
   2026         external_page_bytes[ExternalBackingStoreType::kExternalString] += size;
   2027       } else if (object->IsJSArrayBuffer()) {
   2028         JSArrayBuffer* array_buffer = JSArrayBuffer::cast(object);
   2029         if (ArrayBufferTracker::IsTracked(array_buffer)) {
   2030           size_t size = NumberToSize(array_buffer->byte_length());
   2031           external_page_bytes[ExternalBackingStoreType::kArrayBuffer] += size;
   2032         }
   2033       }
   2034     }
   2035     for (int i = 0; i < kNumTypes; i++) {
   2036       ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2037       CHECK_EQ(external_page_bytes[t], page->ExternalBackingStoreBytes(t));
   2038       external_space_bytes[t] += external_page_bytes[t];
   2039     }
   2040   }
   2041   for (int i = 0; i < kNumTypes; i++) {
   2042     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2043     CHECK_EQ(external_space_bytes[t], ExternalBackingStoreBytes(t));
   2044   }
   2045   CHECK(allocation_pointer_found_in_space);
   2046 #ifdef DEBUG
   2047   VerifyCountersAfterSweeping();
   2048 #endif
   2049 }
   2050 
   2051 void PagedSpace::VerifyLiveBytes() {
   2052   IncrementalMarking::MarkingState* marking_state =
   2053       heap()->incremental_marking()->marking_state();
   2054   for (Page* page : *this) {
   2055     CHECK(page->SweepingDone());
   2056     HeapObjectIterator it(page);
   2057     int black_size = 0;
   2058     for (HeapObject* object = it.Next(); object != nullptr;
   2059          object = it.Next()) {
   2060       // All the interior pointers should be contained in the heap.
   2061       if (marking_state->IsBlack(object)) {
   2062         black_size += object->Size();
   2063       }
   2064     }
   2065     CHECK_LE(black_size, marking_state->live_bytes(page));
   2066   }
   2067 }
   2068 #endif  // VERIFY_HEAP
   2069 
   2070 #ifdef DEBUG
   2071 void PagedSpace::VerifyCountersAfterSweeping() {
   2072   size_t total_capacity = 0;
   2073   size_t total_allocated = 0;
   2074   for (Page* page : *this) {
   2075     DCHECK(page->SweepingDone());
   2076     total_capacity += page->area_size();
   2077     HeapObjectIterator it(page);
   2078     size_t real_allocated = 0;
   2079     for (HeapObject* object = it.Next(); object != nullptr;
   2080          object = it.Next()) {
   2081       if (!object->IsFiller()) {
   2082         real_allocated += object->Size();
   2083       }
   2084     }
   2085     total_allocated += page->allocated_bytes();
   2086     // The real size can be smaller than the accounted size if array trimming,
   2087     // object slack tracking happened after sweeping.
   2088     DCHECK_LE(real_allocated, accounting_stats_.AllocatedOnPage(page));
   2089     DCHECK_EQ(page->allocated_bytes(), accounting_stats_.AllocatedOnPage(page));
   2090   }
   2091   DCHECK_EQ(total_capacity, accounting_stats_.Capacity());
   2092   DCHECK_EQ(total_allocated, accounting_stats_.Size());
   2093 }
   2094 
   2095 void PagedSpace::VerifyCountersBeforeConcurrentSweeping() {
   2096   // We need to refine the counters on pages that are already swept and have
   2097   // not been moved over to the actual space. Otherwise, the AccountingStats
   2098   // are just an over approximation.
   2099   RefillFreeList();
   2100 
   2101   size_t total_capacity = 0;
   2102   size_t total_allocated = 0;
   2103   auto marking_state =
   2104       heap()->incremental_marking()->non_atomic_marking_state();
   2105   for (Page* page : *this) {
   2106     size_t page_allocated =
   2107         page->SweepingDone()
   2108             ? page->allocated_bytes()
   2109             : static_cast<size_t>(marking_state->live_bytes(page));
   2110     total_capacity += page->area_size();
   2111     total_allocated += page_allocated;
   2112     DCHECK_EQ(page_allocated, accounting_stats_.AllocatedOnPage(page));
   2113   }
   2114   DCHECK_EQ(total_capacity, accounting_stats_.Capacity());
   2115   DCHECK_EQ(total_allocated, accounting_stats_.Size());
   2116 }
   2117 #endif
   2118 
   2119 // -----------------------------------------------------------------------------
   2120 // NewSpace implementation
   2121 
   2122 NewSpace::NewSpace(Heap* heap, size_t initial_semispace_capacity,
   2123                    size_t max_semispace_capacity)
   2124     : SpaceWithLinearArea(heap, NEW_SPACE),
   2125       to_space_(heap, kToSpace),
   2126       from_space_(heap, kFromSpace),
   2127       reservation_() {
   2128   DCHECK(initial_semispace_capacity <= max_semispace_capacity);
   2129   DCHECK(
   2130       base::bits::IsPowerOfTwo(static_cast<uint32_t>(max_semispace_capacity)));
   2131 
   2132   to_space_.SetUp(initial_semispace_capacity, max_semispace_capacity);
   2133   from_space_.SetUp(initial_semispace_capacity, max_semispace_capacity);
   2134   if (!to_space_.Commit()) {
   2135     V8::FatalProcessOutOfMemory(heap->isolate(), "New space setup");
   2136   }
   2137   DCHECK(!from_space_.is_committed());  // No need to use memory yet.
   2138   ResetLinearAllocationArea();
   2139 }
   2140 
   2141 void NewSpace::TearDown() {
   2142   allocation_info_.Reset(kNullAddress, kNullAddress);
   2143 
   2144   to_space_.TearDown();
   2145   from_space_.TearDown();
   2146 }
   2147 
   2148 void NewSpace::Flip() { SemiSpace::Swap(&from_space_, &to_space_); }
   2149 
   2150 
   2151 void NewSpace::Grow() {
   2152   // Double the semispace size but only up to maximum capacity.
   2153   DCHECK(TotalCapacity() < MaximumCapacity());
   2154   size_t new_capacity =
   2155       Min(MaximumCapacity(),
   2156           static_cast<size_t>(FLAG_semi_space_growth_factor) * TotalCapacity());
   2157   if (to_space_.GrowTo(new_capacity)) {
   2158     // Only grow from space if we managed to grow to-space.
   2159     if (!from_space_.GrowTo(new_capacity)) {
   2160       // If we managed to grow to-space but couldn't grow from-space,
   2161       // attempt to shrink to-space.
   2162       if (!to_space_.ShrinkTo(from_space_.current_capacity())) {
   2163         // We are in an inconsistent state because we could not
   2164         // commit/uncommit memory from new space.
   2165         FATAL("inconsistent state");
   2166       }
   2167     }
   2168   }
   2169   DCHECK_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_);
   2170 }
   2171 
   2172 
   2173 void NewSpace::Shrink() {
   2174   size_t new_capacity = Max(InitialTotalCapacity(), 2 * Size());
   2175   size_t rounded_new_capacity = ::RoundUp(new_capacity, Page::kPageSize);
   2176   if (rounded_new_capacity < TotalCapacity() &&
   2177       to_space_.ShrinkTo(rounded_new_capacity)) {
   2178     // Only shrink from-space if we managed to shrink to-space.
   2179     from_space_.Reset();
   2180     if (!from_space_.ShrinkTo(rounded_new_capacity)) {
   2181       // If we managed to shrink to-space but couldn't shrink from
   2182       // space, attempt to grow to-space again.
   2183       if (!to_space_.GrowTo(from_space_.current_capacity())) {
   2184         // We are in an inconsistent state because we could not
   2185         // commit/uncommit memory from new space.
   2186         FATAL("inconsistent state");
   2187       }
   2188     }
   2189   }
   2190   DCHECK_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_);
   2191 }
   2192 
   2193 bool NewSpace::Rebalance() {
   2194   // Order here is important to make use of the page pool.
   2195   return to_space_.EnsureCurrentCapacity() &&
   2196          from_space_.EnsureCurrentCapacity();
   2197 }
   2198 
   2199 bool SemiSpace::EnsureCurrentCapacity() {
   2200   if (is_committed()) {
   2201     const int expected_pages =
   2202         static_cast<int>(current_capacity_ / Page::kPageSize);
   2203     MemoryChunk* current_page = first_page();
   2204     int actual_pages = 0;
   2205 
   2206     // First iterate through the pages list until expected pages if so many
   2207     // pages exist.
   2208     while (current_page != nullptr && actual_pages < expected_pages) {
   2209       actual_pages++;
   2210       current_page = current_page->list_node().next();
   2211     }
   2212 
   2213     // Free all overallocated pages which are behind current_page.
   2214     while (current_page) {
   2215       MemoryChunk* next_current = current_page->list_node().next();
   2216       memory_chunk_list_.Remove(current_page);
   2217       // Clear new space flags to avoid this page being treated as a new
   2218       // space page that is potentially being swept.
   2219       current_page->SetFlags(0, Page::kIsInNewSpaceMask);
   2220       heap()->memory_allocator()->Free<MemoryAllocator::kPooledAndQueue>(
   2221           current_page);
   2222       current_page = next_current;
   2223     }
   2224 
   2225     // Add more pages if we have less than expected_pages.
   2226     IncrementalMarking::NonAtomicMarkingState* marking_state =
   2227         heap()->incremental_marking()->non_atomic_marking_state();
   2228     while (actual_pages < expected_pages) {
   2229       actual_pages++;
   2230       current_page =
   2231           heap()->memory_allocator()->AllocatePage<MemoryAllocator::kPooled>(
   2232               Page::kAllocatableMemory, this, NOT_EXECUTABLE);
   2233       if (current_page == nullptr) return false;
   2234       DCHECK_NOT_NULL(current_page);
   2235       memory_chunk_list_.PushBack(current_page);
   2236       marking_state->ClearLiveness(current_page);
   2237       current_page->SetFlags(first_page()->GetFlags(),
   2238                              static_cast<uintptr_t>(Page::kCopyAllFlags));
   2239       heap()->CreateFillerObjectAt(current_page->area_start(),
   2240                                    static_cast<int>(current_page->area_size()),
   2241                                    ClearRecordedSlots::kNo);
   2242     }
   2243   }
   2244   return true;
   2245 }
   2246 
   2247 LinearAllocationArea LocalAllocationBuffer::Close() {
   2248   if (IsValid()) {
   2249     heap_->CreateFillerObjectAt(
   2250         allocation_info_.top(),
   2251         static_cast<int>(allocation_info_.limit() - allocation_info_.top()),
   2252         ClearRecordedSlots::kNo);
   2253     const LinearAllocationArea old_info = allocation_info_;
   2254     allocation_info_ = LinearAllocationArea(kNullAddress, kNullAddress);
   2255     return old_info;
   2256   }
   2257   return LinearAllocationArea(kNullAddress, kNullAddress);
   2258 }
   2259 
   2260 LocalAllocationBuffer::LocalAllocationBuffer(
   2261     Heap* heap, LinearAllocationArea allocation_info)
   2262     : heap_(heap), allocation_info_(allocation_info) {
   2263   if (IsValid()) {
   2264     heap_->CreateFillerObjectAt(
   2265         allocation_info_.top(),
   2266         static_cast<int>(allocation_info_.limit() - allocation_info_.top()),
   2267         ClearRecordedSlots::kNo);
   2268   }
   2269 }
   2270 
   2271 
   2272 LocalAllocationBuffer::LocalAllocationBuffer(
   2273     const LocalAllocationBuffer& other) {
   2274   *this = other;
   2275 }
   2276 
   2277 
   2278 LocalAllocationBuffer& LocalAllocationBuffer::operator=(
   2279     const LocalAllocationBuffer& other) {
   2280   Close();
   2281   heap_ = other.heap_;
   2282   allocation_info_ = other.allocation_info_;
   2283 
   2284   // This is needed since we (a) cannot yet use move-semantics, and (b) want
   2285   // to make the use of the class easy by it as value and (c) implicitly call
   2286   // {Close} upon copy.
   2287   const_cast<LocalAllocationBuffer&>(other).allocation_info_.Reset(
   2288       kNullAddress, kNullAddress);
   2289   return *this;
   2290 }
   2291 
   2292 void NewSpace::UpdateLinearAllocationArea() {
   2293   // Make sure there is no unaccounted allocations.
   2294   DCHECK(!AllocationObserversActive() || top_on_previous_step_ == top());
   2295 
   2296   Address new_top = to_space_.page_low();
   2297   MemoryChunk::UpdateHighWaterMark(allocation_info_.top());
   2298   allocation_info_.Reset(new_top, to_space_.page_high());
   2299   original_top_ = top();
   2300   original_limit_ = limit();
   2301   StartNextInlineAllocationStep();
   2302   DCHECK_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_);
   2303 }
   2304 
   2305 void NewSpace::ResetLinearAllocationArea() {
   2306   // Do a step to account for memory allocated so far before resetting.
   2307   InlineAllocationStep(top(), top(), kNullAddress, 0);
   2308   to_space_.Reset();
   2309   UpdateLinearAllocationArea();
   2310   // Clear all mark-bits in the to-space.
   2311   IncrementalMarking::NonAtomicMarkingState* marking_state =
   2312       heap()->incremental_marking()->non_atomic_marking_state();
   2313   for (Page* p : to_space_) {
   2314     marking_state->ClearLiveness(p);
   2315     // Concurrent marking may have local live bytes for this page.
   2316     heap()->concurrent_marking()->ClearLiveness(p);
   2317   }
   2318 }
   2319 
   2320 void NewSpace::UpdateInlineAllocationLimit(size_t min_size) {
   2321   Address new_limit = ComputeLimit(top(), to_space_.page_high(), min_size);
   2322   allocation_info_.set_limit(new_limit);
   2323   DCHECK_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_);
   2324 }
   2325 
   2326 void PagedSpace::UpdateInlineAllocationLimit(size_t min_size) {
   2327   Address new_limit = ComputeLimit(top(), limit(), min_size);
   2328   DCHECK_LE(new_limit, limit());
   2329   DecreaseLimit(new_limit);
   2330 }
   2331 
   2332 bool NewSpace::AddFreshPage() {
   2333   Address top = allocation_info_.top();
   2334   DCHECK(!Page::IsAtObjectStart(top));
   2335 
   2336   // Do a step to account for memory allocated on previous page.
   2337   InlineAllocationStep(top, top, kNullAddress, 0);
   2338 
   2339   if (!to_space_.AdvancePage()) {
   2340     // No more pages left to advance.
   2341     return false;
   2342   }
   2343 
   2344   // Clear remainder of current page.
   2345   Address limit = Page::FromAllocationAreaAddress(top)->area_end();
   2346   int remaining_in_page = static_cast<int>(limit - top);
   2347   heap()->CreateFillerObjectAt(top, remaining_in_page, ClearRecordedSlots::kNo);
   2348   UpdateLinearAllocationArea();
   2349 
   2350   return true;
   2351 }
   2352 
   2353 
   2354 bool NewSpace::AddFreshPageSynchronized() {
   2355   base::LockGuard<base::Mutex> guard(&mutex_);
   2356   return AddFreshPage();
   2357 }
   2358 
   2359 
   2360 bool NewSpace::EnsureAllocation(int size_in_bytes,
   2361                                 AllocationAlignment alignment) {
   2362   Address old_top = allocation_info_.top();
   2363   Address high = to_space_.page_high();
   2364   int filler_size = Heap::GetFillToAlign(old_top, alignment);
   2365   int aligned_size_in_bytes = size_in_bytes + filler_size;
   2366 
   2367   if (old_top + aligned_size_in_bytes > high) {
   2368     // Not enough room in the page, try to allocate a new one.
   2369     if (!AddFreshPage()) {
   2370       return false;
   2371     }
   2372 
   2373     old_top = allocation_info_.top();
   2374     high = to_space_.page_high();
   2375     filler_size = Heap::GetFillToAlign(old_top, alignment);
   2376   }
   2377 
   2378   DCHECK(old_top + aligned_size_in_bytes <= high);
   2379 
   2380   if (allocation_info_.limit() < high) {
   2381     // Either the limit has been lowered because linear allocation was disabled
   2382     // or because incremental marking wants to get a chance to do a step,
   2383     // or because idle scavenge job wants to get a chance to post a task.
   2384     // Set the new limit accordingly.
   2385     Address new_top = old_top + aligned_size_in_bytes;
   2386     Address soon_object = old_top + filler_size;
   2387     InlineAllocationStep(new_top, new_top, soon_object, size_in_bytes);
   2388     UpdateInlineAllocationLimit(aligned_size_in_bytes);
   2389   }
   2390   return true;
   2391 }
   2392 
   2393 size_t LargeObjectSpace::Available() {
   2394   return ObjectSizeFor(heap()->memory_allocator()->Available());
   2395 }
   2396 
   2397 void SpaceWithLinearArea::StartNextInlineAllocationStep() {
   2398   if (heap()->allocation_step_in_progress()) {
   2399     // If we are mid-way through an existing step, don't start a new one.
   2400     return;
   2401   }
   2402 
   2403   if (AllocationObserversActive()) {
   2404     top_on_previous_step_ = top();
   2405     UpdateInlineAllocationLimit(0);
   2406   } else {
   2407     DCHECK_EQ(kNullAddress, top_on_previous_step_);
   2408   }
   2409 }
   2410 
   2411 void SpaceWithLinearArea::AddAllocationObserver(AllocationObserver* observer) {
   2412   InlineAllocationStep(top(), top(), kNullAddress, 0);
   2413   Space::AddAllocationObserver(observer);
   2414   DCHECK_IMPLIES(top_on_previous_step_, AllocationObserversActive());
   2415 }
   2416 
   2417 void SpaceWithLinearArea::RemoveAllocationObserver(
   2418     AllocationObserver* observer) {
   2419   Address top_for_next_step =
   2420       allocation_observers_.size() == 1 ? kNullAddress : top();
   2421   InlineAllocationStep(top(), top_for_next_step, kNullAddress, 0);
   2422   Space::RemoveAllocationObserver(observer);
   2423   DCHECK_IMPLIES(top_on_previous_step_, AllocationObserversActive());
   2424 }
   2425 
   2426 void SpaceWithLinearArea::PauseAllocationObservers() {
   2427   // Do a step to account for memory allocated so far.
   2428   InlineAllocationStep(top(), kNullAddress, kNullAddress, 0);
   2429   Space::PauseAllocationObservers();
   2430   DCHECK_EQ(kNullAddress, top_on_previous_step_);
   2431   UpdateInlineAllocationLimit(0);
   2432 }
   2433 
   2434 void SpaceWithLinearArea::ResumeAllocationObservers() {
   2435   DCHECK_EQ(kNullAddress, top_on_previous_step_);
   2436   Space::ResumeAllocationObservers();
   2437   StartNextInlineAllocationStep();
   2438 }
   2439 
   2440 void SpaceWithLinearArea::InlineAllocationStep(Address top,
   2441                                                Address top_for_next_step,
   2442                                                Address soon_object,
   2443                                                size_t size) {
   2444   if (heap()->allocation_step_in_progress()) {
   2445     // Avoid starting a new step if we are mid-way through an existing one.
   2446     return;
   2447   }
   2448 
   2449   if (top_on_previous_step_) {
   2450     if (top < top_on_previous_step_) {
   2451       // Generated code decreased the top pointer to do folded allocations.
   2452       DCHECK_NE(top, kNullAddress);
   2453       DCHECK_EQ(Page::FromAllocationAreaAddress(top),
   2454                 Page::FromAllocationAreaAddress(top_on_previous_step_));
   2455       top_on_previous_step_ = top;
   2456     }
   2457     int bytes_allocated = static_cast<int>(top - top_on_previous_step_);
   2458     AllocationStep(bytes_allocated, soon_object, static_cast<int>(size));
   2459     top_on_previous_step_ = top_for_next_step;
   2460   }
   2461 }
   2462 
   2463 std::unique_ptr<ObjectIterator> NewSpace::GetObjectIterator() {
   2464   return std::unique_ptr<ObjectIterator>(new SemiSpaceIterator(this));
   2465 }
   2466 
   2467 #ifdef VERIFY_HEAP
   2468 // We do not use the SemiSpaceIterator because verification doesn't assume
   2469 // that it works (it depends on the invariants we are checking).
   2470 void NewSpace::Verify(Isolate* isolate) {
   2471   // The allocation pointer should be in the space or at the very end.
   2472   DCHECK_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_);
   2473 
   2474   // There should be objects packed in from the low address up to the
   2475   // allocation pointer.
   2476   Address current = to_space_.first_page()->area_start();
   2477   CHECK_EQ(current, to_space_.space_start());
   2478 
   2479   size_t external_space_bytes[kNumTypes];
   2480   for (int i = 0; i < kNumTypes; i++) {
   2481     external_space_bytes[static_cast<ExternalBackingStoreType>(i)] = 0;
   2482   }
   2483 
   2484   while (current != top()) {
   2485     if (!Page::IsAlignedToPageSize(current)) {
   2486       // The allocation pointer should not be in the middle of an object.
   2487       CHECK(!Page::FromAllocationAreaAddress(current)->ContainsLimit(top()) ||
   2488             current < top());
   2489 
   2490       HeapObject* object = HeapObject::FromAddress(current);
   2491 
   2492       // The first word should be a map, and we expect all map pointers to
   2493       // be in map space or read-only space.
   2494       Map* map = object->map();
   2495       CHECK(map->IsMap());
   2496       CHECK(heap()->map_space()->Contains(map) ||
   2497             heap()->read_only_space()->Contains(map));
   2498 
   2499       // The object should not be code or a map.
   2500       CHECK(!object->IsMap());
   2501       CHECK(!object->IsAbstractCode());
   2502 
   2503       // The object itself should look OK.
   2504       object->ObjectVerify(isolate);
   2505 
   2506       // All the interior pointers should be contained in the heap.
   2507       VerifyPointersVisitor visitor(heap());
   2508       int size = object->Size();
   2509       object->IterateBody(map, size, &visitor);
   2510 
   2511       if (object->IsExternalString()) {
   2512         ExternalString* external_string = ExternalString::cast(object);
   2513         size_t size = external_string->ExternalPayloadSize();
   2514         external_space_bytes[ExternalBackingStoreType::kExternalString] += size;
   2515       } else if (object->IsJSArrayBuffer()) {
   2516         JSArrayBuffer* array_buffer = JSArrayBuffer::cast(object);
   2517         if (ArrayBufferTracker::IsTracked(array_buffer)) {
   2518           size_t size = NumberToSize(array_buffer->byte_length());
   2519           external_space_bytes[ExternalBackingStoreType::kArrayBuffer] += size;
   2520         }
   2521       }
   2522 
   2523       current += size;
   2524     } else {
   2525       // At end of page, switch to next page.
   2526       Page* page = Page::FromAllocationAreaAddress(current)->next_page();
   2527       current = page->area_start();
   2528     }
   2529   }
   2530 
   2531   for (int i = 0; i < kNumTypes; i++) {
   2532     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2533     CHECK_EQ(external_space_bytes[t], ExternalBackingStoreBytes(t));
   2534   }
   2535 
   2536   // Check semi-spaces.
   2537   CHECK_EQ(from_space_.id(), kFromSpace);
   2538   CHECK_EQ(to_space_.id(), kToSpace);
   2539   from_space_.Verify();
   2540   to_space_.Verify();
   2541 }
   2542 #endif
   2543 
   2544 // -----------------------------------------------------------------------------
   2545 // SemiSpace implementation
   2546 
   2547 void SemiSpace::SetUp(size_t initial_capacity, size_t maximum_capacity) {
   2548   DCHECK_GE(maximum_capacity, static_cast<size_t>(Page::kPageSize));
   2549   minimum_capacity_ = RoundDown(initial_capacity, Page::kPageSize);
   2550   current_capacity_ = minimum_capacity_;
   2551   maximum_capacity_ = RoundDown(maximum_capacity, Page::kPageSize);
   2552   committed_ = false;
   2553 }
   2554 
   2555 
   2556 void SemiSpace::TearDown() {
   2557   // Properly uncommit memory to keep the allocator counters in sync.
   2558   if (is_committed()) {
   2559     Uncommit();
   2560   }
   2561   current_capacity_ = maximum_capacity_ = 0;
   2562 }
   2563 
   2564 
   2565 bool SemiSpace::Commit() {
   2566   DCHECK(!is_committed());
   2567   const int num_pages = static_cast<int>(current_capacity_ / Page::kPageSize);
   2568   for (int pages_added = 0; pages_added < num_pages; pages_added++) {
   2569     Page* new_page =
   2570         heap()->memory_allocator()->AllocatePage<MemoryAllocator::kPooled>(
   2571             Page::kAllocatableMemory, this, NOT_EXECUTABLE);
   2572     if (new_page == nullptr) {
   2573       if (pages_added) RewindPages(pages_added);
   2574       return false;
   2575     }
   2576     memory_chunk_list_.PushBack(new_page);
   2577   }
   2578   Reset();
   2579   AccountCommitted(current_capacity_);
   2580   if (age_mark_ == kNullAddress) {
   2581     age_mark_ = first_page()->area_start();
   2582   }
   2583   committed_ = true;
   2584   return true;
   2585 }
   2586 
   2587 
   2588 bool SemiSpace::Uncommit() {
   2589   DCHECK(is_committed());
   2590   while (!memory_chunk_list_.Empty()) {
   2591     MemoryChunk* chunk = memory_chunk_list_.front();
   2592     memory_chunk_list_.Remove(chunk);
   2593     heap()->memory_allocator()->Free<MemoryAllocator::kPooledAndQueue>(chunk);
   2594   }
   2595   current_page_ = nullptr;
   2596   AccountUncommitted(current_capacity_);
   2597   committed_ = false;
   2598   heap()->memory_allocator()->unmapper()->FreeQueuedChunks();
   2599   return true;
   2600 }
   2601 
   2602 
   2603 size_t SemiSpace::CommittedPhysicalMemory() {
   2604   if (!is_committed()) return 0;
   2605   size_t size = 0;
   2606   for (Page* p : *this) {
   2607     size += p->CommittedPhysicalMemory();
   2608   }
   2609   return size;
   2610 }
   2611 
   2612 bool SemiSpace::GrowTo(size_t new_capacity) {
   2613   if (!is_committed()) {
   2614     if (!Commit()) return false;
   2615   }
   2616   DCHECK_EQ(new_capacity & kPageAlignmentMask, 0u);
   2617   DCHECK_LE(new_capacity, maximum_capacity_);
   2618   DCHECK_GT(new_capacity, current_capacity_);
   2619   const size_t delta = new_capacity - current_capacity_;
   2620   DCHECK(IsAligned(delta, AllocatePageSize()));
   2621   const int delta_pages = static_cast<int>(delta / Page::kPageSize);
   2622   DCHECK(last_page());
   2623   IncrementalMarking::NonAtomicMarkingState* marking_state =
   2624       heap()->incremental_marking()->non_atomic_marking_state();
   2625   for (int pages_added = 0; pages_added < delta_pages; pages_added++) {
   2626     Page* new_page =
   2627         heap()->memory_allocator()->AllocatePage<MemoryAllocator::kPooled>(
   2628             Page::kAllocatableMemory, this, NOT_EXECUTABLE);
   2629     if (new_page == nullptr) {
   2630       if (pages_added) RewindPages(pages_added);
   2631       return false;
   2632     }
   2633     memory_chunk_list_.PushBack(new_page);
   2634     marking_state->ClearLiveness(new_page);
   2635     // Duplicate the flags that was set on the old page.
   2636     new_page->SetFlags(last_page()->GetFlags(), Page::kCopyOnFlipFlagsMask);
   2637   }
   2638   AccountCommitted(delta);
   2639   current_capacity_ = new_capacity;
   2640   return true;
   2641 }
   2642 
   2643 void SemiSpace::RewindPages(int num_pages) {
   2644   DCHECK_GT(num_pages, 0);
   2645   DCHECK(last_page());
   2646   while (num_pages > 0) {
   2647     MemoryChunk* last = last_page();
   2648     memory_chunk_list_.Remove(last);
   2649     heap()->memory_allocator()->Free<MemoryAllocator::kPooledAndQueue>(last);
   2650     num_pages--;
   2651   }
   2652 }
   2653 
   2654 bool SemiSpace::ShrinkTo(size_t new_capacity) {
   2655   DCHECK_EQ(new_capacity & kPageAlignmentMask, 0u);
   2656   DCHECK_GE(new_capacity, minimum_capacity_);
   2657   DCHECK_LT(new_capacity, current_capacity_);
   2658   if (is_committed()) {
   2659     const size_t delta = current_capacity_ - new_capacity;
   2660     DCHECK(IsAligned(delta, Page::kPageSize));
   2661     int delta_pages = static_cast<int>(delta / Page::kPageSize);
   2662     RewindPages(delta_pages);
   2663     AccountUncommitted(delta);
   2664     heap()->memory_allocator()->unmapper()->FreeQueuedChunks();
   2665   }
   2666   current_capacity_ = new_capacity;
   2667   return true;
   2668 }
   2669 
   2670 void SemiSpace::FixPagesFlags(intptr_t flags, intptr_t mask) {
   2671   for (Page* page : *this) {
   2672     page->set_owner(this);
   2673     page->SetFlags(flags, mask);
   2674     if (id_ == kToSpace) {
   2675       page->ClearFlag(MemoryChunk::IN_FROM_SPACE);
   2676       page->SetFlag(MemoryChunk::IN_TO_SPACE);
   2677       page->ClearFlag(MemoryChunk::NEW_SPACE_BELOW_AGE_MARK);
   2678       heap()->incremental_marking()->non_atomic_marking_state()->SetLiveBytes(
   2679           page, 0);
   2680     } else {
   2681       page->SetFlag(MemoryChunk::IN_FROM_SPACE);
   2682       page->ClearFlag(MemoryChunk::IN_TO_SPACE);
   2683     }
   2684     DCHECK(page->IsFlagSet(MemoryChunk::IN_TO_SPACE) ||
   2685            page->IsFlagSet(MemoryChunk::IN_FROM_SPACE));
   2686   }
   2687 }
   2688 
   2689 
   2690 void SemiSpace::Reset() {
   2691   DCHECK(first_page());
   2692   DCHECK(last_page());
   2693   current_page_ = first_page();
   2694   pages_used_ = 0;
   2695 }
   2696 
   2697 void SemiSpace::RemovePage(Page* page) {
   2698   if (current_page_ == page) {
   2699     if (page->prev_page()) {
   2700       current_page_ = page->prev_page();
   2701     }
   2702   }
   2703   memory_chunk_list_.Remove(page);
   2704   for (size_t i = 0; i < ExternalBackingStoreType::kNumTypes; i++) {
   2705     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2706     DecrementExternalBackingStoreBytes(t, page->ExternalBackingStoreBytes(t));
   2707   }
   2708 }
   2709 
   2710 void SemiSpace::PrependPage(Page* page) {
   2711   page->SetFlags(current_page()->GetFlags(),
   2712                  static_cast<uintptr_t>(Page::kCopyAllFlags));
   2713   page->set_owner(this);
   2714   memory_chunk_list_.PushFront(page);
   2715   pages_used_++;
   2716   for (size_t i = 0; i < ExternalBackingStoreType::kNumTypes; i++) {
   2717     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2718     IncrementExternalBackingStoreBytes(t, page->ExternalBackingStoreBytes(t));
   2719   }
   2720 }
   2721 
   2722 void SemiSpace::Swap(SemiSpace* from, SemiSpace* to) {
   2723   // We won't be swapping semispaces without data in them.
   2724   DCHECK(from->first_page());
   2725   DCHECK(to->first_page());
   2726 
   2727   intptr_t saved_to_space_flags = to->current_page()->GetFlags();
   2728 
   2729   // We swap all properties but id_.
   2730   std::swap(from->current_capacity_, to->current_capacity_);
   2731   std::swap(from->maximum_capacity_, to->maximum_capacity_);
   2732   std::swap(from->minimum_capacity_, to->minimum_capacity_);
   2733   std::swap(from->age_mark_, to->age_mark_);
   2734   std::swap(from->committed_, to->committed_);
   2735   std::swap(from->memory_chunk_list_, to->memory_chunk_list_);
   2736   std::swap(from->current_page_, to->current_page_);
   2737   std::swap(from->external_backing_store_bytes_,
   2738             to->external_backing_store_bytes_);
   2739 
   2740   to->FixPagesFlags(saved_to_space_flags, Page::kCopyOnFlipFlagsMask);
   2741   from->FixPagesFlags(0, 0);
   2742 }
   2743 
   2744 void SemiSpace::set_age_mark(Address mark) {
   2745   DCHECK_EQ(Page::FromAllocationAreaAddress(mark)->owner(), this);
   2746   age_mark_ = mark;
   2747   // Mark all pages up to the one containing mark.
   2748   for (Page* p : PageRange(space_start(), mark)) {
   2749     p->SetFlag(MemoryChunk::NEW_SPACE_BELOW_AGE_MARK);
   2750   }
   2751 }
   2752 
   2753 std::unique_ptr<ObjectIterator> SemiSpace::GetObjectIterator() {
   2754   // Use the NewSpace::NewObjectIterator to iterate the ToSpace.
   2755   UNREACHABLE();
   2756 }
   2757 
   2758 #ifdef DEBUG
   2759 void SemiSpace::Print() {}
   2760 #endif
   2761 
   2762 #ifdef VERIFY_HEAP
   2763 void SemiSpace::Verify() {
   2764   bool is_from_space = (id_ == kFromSpace);
   2765   size_t external_backing_store_bytes[kNumTypes];
   2766 
   2767   for (int i = 0; i < kNumTypes; i++) {
   2768     external_backing_store_bytes[static_cast<ExternalBackingStoreType>(i)] = 0;
   2769   }
   2770 
   2771   for (Page* page : *this) {
   2772     CHECK_EQ(page->owner(), this);
   2773     CHECK(page->InNewSpace());
   2774     CHECK(page->IsFlagSet(is_from_space ? MemoryChunk::IN_FROM_SPACE
   2775                                         : MemoryChunk::IN_TO_SPACE));
   2776     CHECK(!page->IsFlagSet(is_from_space ? MemoryChunk::IN_TO_SPACE
   2777                                          : MemoryChunk::IN_FROM_SPACE));
   2778     CHECK(page->IsFlagSet(MemoryChunk::POINTERS_TO_HERE_ARE_INTERESTING));
   2779     if (!is_from_space) {
   2780       // The pointers-from-here-are-interesting flag isn't updated dynamically
   2781       // on from-space pages, so it might be out of sync with the marking state.
   2782       if (page->heap()->incremental_marking()->IsMarking()) {
   2783         CHECK(page->IsFlagSet(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING));
   2784       } else {
   2785         CHECK(
   2786             !page->IsFlagSet(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING));
   2787       }
   2788     }
   2789     for (int i = 0; i < kNumTypes; i++) {
   2790       ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2791       external_backing_store_bytes[t] += page->ExternalBackingStoreBytes(t);
   2792     }
   2793 
   2794     CHECK_IMPLIES(page->list_node().prev(),
   2795                   page->list_node().prev()->list_node().next() == page);
   2796   }
   2797   for (int i = 0; i < kNumTypes; i++) {
   2798     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   2799     CHECK_EQ(external_backing_store_bytes[t], ExternalBackingStoreBytes(t));
   2800   }
   2801 }
   2802 #endif
   2803 
   2804 #ifdef DEBUG
   2805 void SemiSpace::AssertValidRange(Address start, Address end) {
   2806   // Addresses belong to same semi-space
   2807   Page* page = Page::FromAllocationAreaAddress(start);
   2808   Page* end_page = Page::FromAllocationAreaAddress(end);
   2809   SemiSpace* space = reinterpret_cast<SemiSpace*>(page->owner());
   2810   DCHECK_EQ(space, end_page->owner());
   2811   // Start address is before end address, either on same page,
   2812   // or end address is on a later page in the linked list of
   2813   // semi-space pages.
   2814   if (page == end_page) {
   2815     DCHECK_LE(start, end);
   2816   } else {
   2817     while (page != end_page) {
   2818       page = page->next_page();
   2819     }
   2820     DCHECK(page);
   2821   }
   2822 }
   2823 #endif
   2824 
   2825 
   2826 // -----------------------------------------------------------------------------
   2827 // SemiSpaceIterator implementation.
   2828 
   2829 SemiSpaceIterator::SemiSpaceIterator(NewSpace* space) {
   2830   Initialize(space->first_allocatable_address(), space->top());
   2831 }
   2832 
   2833 
   2834 void SemiSpaceIterator::Initialize(Address start, Address end) {
   2835   SemiSpace::AssertValidRange(start, end);
   2836   current_ = start;
   2837   limit_ = end;
   2838 }
   2839 
   2840 size_t NewSpace::CommittedPhysicalMemory() {
   2841   if (!base::OS::HasLazyCommits()) return CommittedMemory();
   2842   MemoryChunk::UpdateHighWaterMark(allocation_info_.top());
   2843   size_t size = to_space_.CommittedPhysicalMemory();
   2844   if (from_space_.is_committed()) {
   2845     size += from_space_.CommittedPhysicalMemory();
   2846   }
   2847   return size;
   2848 }
   2849 
   2850 
   2851 // -----------------------------------------------------------------------------
   2852 // Free lists for old object spaces implementation
   2853 
   2854 
   2855 void FreeListCategory::Reset() {
   2856   set_top(nullptr);
   2857   set_prev(nullptr);
   2858   set_next(nullptr);
   2859   available_ = 0;
   2860 }
   2861 
   2862 FreeSpace* FreeListCategory::PickNodeFromList(size_t minimum_size,
   2863                                               size_t* node_size) {
   2864   DCHECK(page()->CanAllocate());
   2865   FreeSpace* node = top();
   2866   if (node == nullptr || static_cast<size_t>(node->Size()) < minimum_size) {
   2867     *node_size = 0;
   2868     return nullptr;
   2869   }
   2870   set_top(node->next());
   2871   *node_size = node->Size();
   2872   available_ -= *node_size;
   2873   return node;
   2874 }
   2875 
   2876 FreeSpace* FreeListCategory::SearchForNodeInList(size_t minimum_size,
   2877                                                  size_t* node_size) {
   2878   DCHECK(page()->CanAllocate());
   2879   FreeSpace* prev_non_evac_node = nullptr;
   2880   for (FreeSpace* cur_node = top(); cur_node != nullptr;
   2881        cur_node = cur_node->next()) {
   2882     size_t size = cur_node->size();
   2883     if (size >= minimum_size) {
   2884       DCHECK_GE(available_, size);
   2885       available_ -= size;
   2886       if (cur_node == top()) {
   2887         set_top(cur_node->next());
   2888       }
   2889       if (prev_non_evac_node != nullptr) {
   2890         MemoryChunk* chunk =
   2891             MemoryChunk::FromAddress(prev_non_evac_node->address());
   2892         if (chunk->owner()->identity() == CODE_SPACE) {
   2893           chunk->heap()->UnprotectAndRegisterMemoryChunk(chunk);
   2894         }
   2895         prev_non_evac_node->set_next(cur_node->next());
   2896       }
   2897       *node_size = size;
   2898       return cur_node;
   2899     }
   2900 
   2901     prev_non_evac_node = cur_node;
   2902   }
   2903   return nullptr;
   2904 }
   2905 
   2906 void FreeListCategory::Free(Address start, size_t size_in_bytes,
   2907                             FreeMode mode) {
   2908   DCHECK(page()->CanAllocate());
   2909   FreeSpace* free_space = FreeSpace::cast(HeapObject::FromAddress(start));
   2910   free_space->set_next(top());
   2911   set_top(free_space);
   2912   available_ += size_in_bytes;
   2913   if ((mode == kLinkCategory) && (prev() == nullptr) && (next() == nullptr)) {
   2914     owner()->AddCategory(this);
   2915   }
   2916 }
   2917 
   2918 
   2919 void FreeListCategory::RepairFreeList(Heap* heap) {
   2920   FreeSpace* n = top();
   2921   while (n != nullptr) {
   2922     Map** map_location = reinterpret_cast<Map**>(n->address());
   2923     if (*map_location == nullptr) {
   2924       *map_location = ReadOnlyRoots(heap).free_space_map();
   2925     } else {
   2926       DCHECK(*map_location == ReadOnlyRoots(heap).free_space_map());
   2927     }
   2928     n = n->next();
   2929   }
   2930 }
   2931 
   2932 void FreeListCategory::Relink() {
   2933   DCHECK(!is_linked());
   2934   owner()->AddCategory(this);
   2935 }
   2936 
   2937 FreeList::FreeList() : wasted_bytes_(0) {
   2938   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
   2939     categories_[i] = nullptr;
   2940   }
   2941   Reset();
   2942 }
   2943 
   2944 
   2945 void FreeList::Reset() {
   2946   ForAllFreeListCategories(
   2947       [](FreeListCategory* category) { category->Reset(); });
   2948   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
   2949     categories_[i] = nullptr;
   2950   }
   2951   ResetStats();
   2952 }
   2953 
   2954 size_t FreeList::Free(Address start, size_t size_in_bytes, FreeMode mode) {
   2955   Page* page = Page::FromAddress(start);
   2956   page->DecreaseAllocatedBytes(size_in_bytes);
   2957 
   2958   // Blocks have to be a minimum size to hold free list items.
   2959   if (size_in_bytes < kMinBlockSize) {
   2960     page->add_wasted_memory(size_in_bytes);
   2961     wasted_bytes_ += size_in_bytes;
   2962     return size_in_bytes;
   2963   }
   2964 
   2965   // Insert other blocks at the head of a free list of the appropriate
   2966   // magnitude.
   2967   FreeListCategoryType type = SelectFreeListCategoryType(size_in_bytes);
   2968   page->free_list_category(type)->Free(start, size_in_bytes, mode);
   2969   DCHECK_EQ(page->AvailableInFreeList(),
   2970             page->AvailableInFreeListFromAllocatedBytes());
   2971   return 0;
   2972 }
   2973 
   2974 FreeSpace* FreeList::FindNodeIn(FreeListCategoryType type, size_t minimum_size,
   2975                                 size_t* node_size) {
   2976   FreeListCategoryIterator it(this, type);
   2977   FreeSpace* node = nullptr;
   2978   while (it.HasNext()) {
   2979     FreeListCategory* current = it.Next();
   2980     node = current->PickNodeFromList(minimum_size, node_size);
   2981     if (node != nullptr) {
   2982       DCHECK(IsVeryLong() || Available() == SumFreeLists());
   2983       return node;
   2984     }
   2985     RemoveCategory(current);
   2986   }
   2987   return node;
   2988 }
   2989 
   2990 FreeSpace* FreeList::TryFindNodeIn(FreeListCategoryType type,
   2991                                    size_t minimum_size, size_t* node_size) {
   2992   if (categories_[type] == nullptr) return nullptr;
   2993   FreeSpace* node =
   2994       categories_[type]->PickNodeFromList(minimum_size, node_size);
   2995   if (node != nullptr) {
   2996     DCHECK(IsVeryLong() || Available() == SumFreeLists());
   2997   }
   2998   return node;
   2999 }
   3000 
   3001 FreeSpace* FreeList::SearchForNodeInList(FreeListCategoryType type,
   3002                                          size_t* node_size,
   3003                                          size_t minimum_size) {
   3004   FreeListCategoryIterator it(this, type);
   3005   FreeSpace* node = nullptr;
   3006   while (it.HasNext()) {
   3007     FreeListCategory* current = it.Next();
   3008     node = current->SearchForNodeInList(minimum_size, node_size);
   3009     if (node != nullptr) {
   3010       DCHECK(IsVeryLong() || Available() == SumFreeLists());
   3011       return node;
   3012     }
   3013     if (current->is_empty()) {
   3014       RemoveCategory(current);
   3015     }
   3016   }
   3017   return node;
   3018 }
   3019 
   3020 FreeSpace* FreeList::Allocate(size_t size_in_bytes, size_t* node_size) {
   3021   DCHECK_GE(kMaxBlockSize, size_in_bytes);
   3022   FreeSpace* node = nullptr;
   3023   // First try the allocation fast path: try to allocate the minimum element
   3024   // size of a free list category. This operation is constant time.
   3025   FreeListCategoryType type =
   3026       SelectFastAllocationFreeListCategoryType(size_in_bytes);
   3027   for (int i = type; i < kHuge && node == nullptr; i++) {
   3028     node = FindNodeIn(static_cast<FreeListCategoryType>(i), size_in_bytes,
   3029                       node_size);
   3030   }
   3031 
   3032   if (node == nullptr) {
   3033     // Next search the huge list for free list nodes. This takes linear time in
   3034     // the number of huge elements.
   3035     node = SearchForNodeInList(kHuge, node_size, size_in_bytes);
   3036   }
   3037 
   3038   if (node == nullptr && type != kHuge) {
   3039     // We didn't find anything in the huge list. Now search the best fitting
   3040     // free list for a node that has at least the requested size.
   3041     type = SelectFreeListCategoryType(size_in_bytes);
   3042     node = TryFindNodeIn(type, size_in_bytes, node_size);
   3043   }
   3044 
   3045   if (node != nullptr) {
   3046     Page::FromAddress(node->address())->IncreaseAllocatedBytes(*node_size);
   3047   }
   3048 
   3049   DCHECK(IsVeryLong() || Available() == SumFreeLists());
   3050   return node;
   3051 }
   3052 
   3053 size_t FreeList::EvictFreeListItems(Page* page) {
   3054   size_t sum = 0;
   3055   page->ForAllFreeListCategories([this, &sum](FreeListCategory* category) {
   3056     DCHECK_EQ(this, category->owner());
   3057     sum += category->available();
   3058     RemoveCategory(category);
   3059     category->Reset();
   3060   });
   3061   return sum;
   3062 }
   3063 
   3064 bool FreeList::ContainsPageFreeListItems(Page* page) {
   3065   bool contained = false;
   3066   page->ForAllFreeListCategories(
   3067       [this, &contained](FreeListCategory* category) {
   3068         if (category->owner() == this && category->is_linked()) {
   3069           contained = true;
   3070         }
   3071       });
   3072   return contained;
   3073 }
   3074 
   3075 void FreeList::RepairLists(Heap* heap) {
   3076   ForAllFreeListCategories(
   3077       [heap](FreeListCategory* category) { category->RepairFreeList(heap); });
   3078 }
   3079 
   3080 bool FreeList::AddCategory(FreeListCategory* category) {
   3081   FreeListCategoryType type = category->type_;
   3082   DCHECK_LT(type, kNumberOfCategories);
   3083   FreeListCategory* top = categories_[type];
   3084 
   3085   if (category->is_empty()) return false;
   3086   if (top == category) return false;
   3087 
   3088   // Common double-linked list insertion.
   3089   if (top != nullptr) {
   3090     top->set_prev(category);
   3091   }
   3092   category->set_next(top);
   3093   categories_[type] = category;
   3094   return true;
   3095 }
   3096 
   3097 void FreeList::RemoveCategory(FreeListCategory* category) {
   3098   FreeListCategoryType type = category->type_;
   3099   DCHECK_LT(type, kNumberOfCategories);
   3100   FreeListCategory* top = categories_[type];
   3101 
   3102   // Common double-linked list removal.
   3103   if (top == category) {
   3104     categories_[type] = category->next();
   3105   }
   3106   if (category->prev() != nullptr) {
   3107     category->prev()->set_next(category->next());
   3108   }
   3109   if (category->next() != nullptr) {
   3110     category->next()->set_prev(category->prev());
   3111   }
   3112   category->set_next(nullptr);
   3113   category->set_prev(nullptr);
   3114 }
   3115 
   3116 void FreeList::PrintCategories(FreeListCategoryType type) {
   3117   FreeListCategoryIterator it(this, type);
   3118   PrintF("FreeList[%p, top=%p, %d] ", static_cast<void*>(this),
   3119          static_cast<void*>(categories_[type]), type);
   3120   while (it.HasNext()) {
   3121     FreeListCategory* current = it.Next();
   3122     PrintF("%p -> ", static_cast<void*>(current));
   3123   }
   3124   PrintF("null\n");
   3125 }
   3126 
   3127 
   3128 #ifdef DEBUG
   3129 size_t FreeListCategory::SumFreeList() {
   3130   size_t sum = 0;
   3131   FreeSpace* cur = top();
   3132   while (cur != nullptr) {
   3133     DCHECK(cur->map() == page()->heap()->root(Heap::kFreeSpaceMapRootIndex));
   3134     sum += cur->relaxed_read_size();
   3135     cur = cur->next();
   3136   }
   3137   return sum;
   3138 }
   3139 
   3140 int FreeListCategory::FreeListLength() {
   3141   int length = 0;
   3142   FreeSpace* cur = top();
   3143   while (cur != nullptr) {
   3144     length++;
   3145     cur = cur->next();
   3146     if (length == kVeryLongFreeList) return length;
   3147   }
   3148   return length;
   3149 }
   3150 
   3151 bool FreeList::IsVeryLong() {
   3152   int len = 0;
   3153   for (int i = kFirstCategory; i < kNumberOfCategories; i++) {
   3154     FreeListCategoryIterator it(this, static_cast<FreeListCategoryType>(i));
   3155     while (it.HasNext()) {
   3156       len += it.Next()->FreeListLength();
   3157       if (len >= FreeListCategory::kVeryLongFreeList) return true;
   3158     }
   3159   }
   3160   return false;
   3161 }
   3162 
   3163 
   3164 // This can take a very long time because it is linear in the number of entries
   3165 // on the free list, so it should not be called if FreeListLength returns
   3166 // kVeryLongFreeList.
   3167 size_t FreeList::SumFreeLists() {
   3168   size_t sum = 0;
   3169   ForAllFreeListCategories(
   3170       [&sum](FreeListCategory* category) { sum += category->SumFreeList(); });
   3171   return sum;
   3172 }
   3173 #endif
   3174 
   3175 
   3176 // -----------------------------------------------------------------------------
   3177 // OldSpace implementation
   3178 
   3179 void PagedSpace::PrepareForMarkCompact() {
   3180   // We don't have a linear allocation area while sweeping.  It will be restored
   3181   // on the first allocation after the sweep.
   3182   FreeLinearAllocationArea();
   3183 
   3184   // Clear the free list before a full GC---it will be rebuilt afterward.
   3185   free_list_.Reset();
   3186 }
   3187 
   3188 size_t PagedSpace::SizeOfObjects() {
   3189   CHECK_GE(limit(), top());
   3190   DCHECK_GE(Size(), static_cast<size_t>(limit() - top()));
   3191   return Size() - (limit() - top());
   3192 }
   3193 
   3194 // After we have booted, we have created a map which represents free space
   3195 // on the heap.  If there was already a free list then the elements on it
   3196 // were created with the wrong FreeSpaceMap (normally nullptr), so we need to
   3197 // fix them.
   3198 void PagedSpace::RepairFreeListsAfterDeserialization() {
   3199   free_list_.RepairLists(heap());
   3200   // Each page may have a small free space that is not tracked by a free list.
   3201   // Those free spaces still contain null as their map pointer.
   3202   // Overwrite them with new fillers.
   3203   for (Page* page : *this) {
   3204     int size = static_cast<int>(page->wasted_memory());
   3205     if (size == 0) {
   3206       // If there is no wasted memory then all free space is in the free list.
   3207       continue;
   3208     }
   3209     Address start = page->HighWaterMark();
   3210     Address end = page->area_end();
   3211     if (start < end - size) {
   3212       // A region at the high watermark is already in free list.
   3213       HeapObject* filler = HeapObject::FromAddress(start);
   3214       CHECK(filler->IsFiller());
   3215       start += filler->Size();
   3216     }
   3217     CHECK_EQ(size, static_cast<int>(end - start));
   3218     heap()->CreateFillerObjectAt(start, size, ClearRecordedSlots::kNo);
   3219   }
   3220 }
   3221 
   3222 bool PagedSpace::SweepAndRetryAllocation(int size_in_bytes) {
   3223   MarkCompactCollector* collector = heap()->mark_compact_collector();
   3224   if (collector->sweeping_in_progress()) {
   3225     // Wait for the sweeper threads here and complete the sweeping phase.
   3226     collector->EnsureSweepingCompleted();
   3227 
   3228     // After waiting for the sweeper threads, there may be new free-list
   3229     // entries.
   3230     return RefillLinearAllocationAreaFromFreeList(size_in_bytes);
   3231   }
   3232   return false;
   3233 }
   3234 
   3235 bool CompactionSpace::SweepAndRetryAllocation(int size_in_bytes) {
   3236   MarkCompactCollector* collector = heap()->mark_compact_collector();
   3237   if (FLAG_concurrent_sweeping && collector->sweeping_in_progress()) {
   3238     collector->sweeper()->ParallelSweepSpace(identity(), 0);
   3239     RefillFreeList();
   3240     return RefillLinearAllocationAreaFromFreeList(size_in_bytes);
   3241   }
   3242   return false;
   3243 }
   3244 
   3245 bool PagedSpace::SlowRefillLinearAllocationArea(int size_in_bytes) {
   3246   VMState<GC> state(heap()->isolate());
   3247   RuntimeCallTimerScope runtime_timer(
   3248       heap()->isolate(), RuntimeCallCounterId::kGC_Custom_SlowAllocateRaw);
   3249   return RawSlowRefillLinearAllocationArea(size_in_bytes);
   3250 }
   3251 
   3252 bool CompactionSpace::SlowRefillLinearAllocationArea(int size_in_bytes) {
   3253   return RawSlowRefillLinearAllocationArea(size_in_bytes);
   3254 }
   3255 
   3256 bool PagedSpace::RawSlowRefillLinearAllocationArea(int size_in_bytes) {
   3257   // Allocation in this space has failed.
   3258   DCHECK_GE(size_in_bytes, 0);
   3259   const int kMaxPagesToSweep = 1;
   3260 
   3261   if (RefillLinearAllocationAreaFromFreeList(size_in_bytes)) return true;
   3262 
   3263   MarkCompactCollector* collector = heap()->mark_compact_collector();
   3264   // Sweeping is still in progress.
   3265   if (collector->sweeping_in_progress()) {
   3266     if (FLAG_concurrent_sweeping && !is_local() &&
   3267         !collector->sweeper()->AreSweeperTasksRunning()) {
   3268       collector->EnsureSweepingCompleted();
   3269     }
   3270 
   3271     // First try to refill the free-list, concurrent sweeper threads
   3272     // may have freed some objects in the meantime.
   3273     RefillFreeList();
   3274 
   3275     // Retry the free list allocation.
   3276     if (RefillLinearAllocationAreaFromFreeList(
   3277             static_cast<size_t>(size_in_bytes)))
   3278       return true;
   3279 
   3280     // If sweeping is still in progress try to sweep pages.
   3281     int max_freed = collector->sweeper()->ParallelSweepSpace(
   3282         identity(), size_in_bytes, kMaxPagesToSweep);
   3283     RefillFreeList();
   3284     if (max_freed >= size_in_bytes) {
   3285       if (RefillLinearAllocationAreaFromFreeList(
   3286               static_cast<size_t>(size_in_bytes)))
   3287         return true;
   3288     }
   3289   } else if (is_local()) {
   3290     // Sweeping not in progress and we are on a {CompactionSpace}. This can
   3291     // only happen when we are evacuating for the young generation.
   3292     PagedSpace* main_space = heap()->paged_space(identity());
   3293     Page* page = main_space->RemovePageSafe(size_in_bytes);
   3294     if (page != nullptr) {
   3295       AddPage(page);
   3296       if (RefillLinearAllocationAreaFromFreeList(
   3297               static_cast<size_t>(size_in_bytes)))
   3298         return true;
   3299     }
   3300   }
   3301 
   3302   if (heap()->ShouldExpandOldGenerationOnSlowAllocation() && Expand()) {
   3303     DCHECK((CountTotalPages() > 1) ||
   3304            (static_cast<size_t>(size_in_bytes) <= free_list_.Available()));
   3305     return RefillLinearAllocationAreaFromFreeList(
   3306         static_cast<size_t>(size_in_bytes));
   3307   }
   3308 
   3309   // If sweeper threads are active, wait for them at that point and steal
   3310   // elements form their free-lists. Allocation may still fail their which
   3311   // would indicate that there is not enough memory for the given allocation.
   3312   return SweepAndRetryAllocation(size_in_bytes);
   3313 }
   3314 
   3315 // -----------------------------------------------------------------------------
   3316 // MapSpace implementation
   3317 
   3318 #ifdef VERIFY_HEAP
   3319 void MapSpace::VerifyObject(HeapObject* object) { CHECK(object->IsMap()); }
   3320 #endif
   3321 
   3322 ReadOnlySpace::ReadOnlySpace(Heap* heap)
   3323     : PagedSpace(heap, RO_SPACE, NOT_EXECUTABLE),
   3324       is_string_padding_cleared_(heap->isolate()->initialized_from_snapshot()) {
   3325 }
   3326 
   3327 void ReadOnlyPage::MakeHeaderRelocatable() {
   3328   if (mutex_ != nullptr) {
   3329     // TODO(v8:7464): heap_ and owner_ need to be cleared as well.
   3330     delete mutex_;
   3331     mutex_ = nullptr;
   3332     local_tracker_ = nullptr;
   3333     reservation_.Reset();
   3334   }
   3335 }
   3336 
   3337 void ReadOnlySpace::SetPermissionsForPages(PageAllocator::Permission access) {
   3338   const size_t page_size = MemoryAllocator::GetCommitPageSize();
   3339   const size_t area_start_offset = RoundUp(Page::kObjectStartOffset, page_size);
   3340   for (Page* p : *this) {
   3341     ReadOnlyPage* page = static_cast<ReadOnlyPage*>(p);
   3342     if (access == PageAllocator::kRead) {
   3343       page->MakeHeaderRelocatable();
   3344     }
   3345     CHECK(SetPermissions(page->address() + area_start_offset,
   3346                          page->size() - area_start_offset, access));
   3347   }
   3348 }
   3349 
   3350 void ReadOnlySpace::ClearStringPaddingIfNeeded() {
   3351   if (is_string_padding_cleared_) return;
   3352 
   3353   WritableScope writable_scope(this);
   3354   for (Page* page : *this) {
   3355     HeapObjectIterator iterator(page);
   3356     for (HeapObject* o = iterator.Next(); o != nullptr; o = iterator.Next()) {
   3357       if (o->IsSeqOneByteString()) {
   3358         SeqOneByteString::cast(o)->clear_padding();
   3359       } else if (o->IsSeqTwoByteString()) {
   3360         SeqTwoByteString::cast(o)->clear_padding();
   3361       }
   3362     }
   3363   }
   3364   is_string_padding_cleared_ = true;
   3365 }
   3366 
   3367 void ReadOnlySpace::MarkAsReadOnly() {
   3368   DCHECK(!is_marked_read_only_);
   3369   FreeLinearAllocationArea();
   3370   is_marked_read_only_ = true;
   3371   SetPermissionsForPages(PageAllocator::kRead);
   3372 }
   3373 
   3374 void ReadOnlySpace::MarkAsReadWrite() {
   3375   DCHECK(is_marked_read_only_);
   3376   SetPermissionsForPages(PageAllocator::kReadWrite);
   3377   is_marked_read_only_ = false;
   3378 }
   3379 
   3380 Address LargePage::GetAddressToShrink(Address object_address,
   3381                                       size_t object_size) {
   3382   if (executable() == EXECUTABLE) {
   3383     return 0;
   3384   }
   3385   size_t used_size = ::RoundUp((object_address - address()) + object_size,
   3386                                MemoryAllocator::GetCommitPageSize());
   3387   if (used_size < CommittedPhysicalMemory()) {
   3388     return address() + used_size;
   3389   }
   3390   return 0;
   3391 }
   3392 
   3393 void LargePage::ClearOutOfLiveRangeSlots(Address free_start) {
   3394   RememberedSet<OLD_TO_NEW>::RemoveRange(this, free_start, area_end(),
   3395                                          SlotSet::FREE_EMPTY_BUCKETS);
   3396   RememberedSet<OLD_TO_OLD>::RemoveRange(this, free_start, area_end(),
   3397                                          SlotSet::FREE_EMPTY_BUCKETS);
   3398   RememberedSet<OLD_TO_NEW>::RemoveRangeTyped(this, free_start, area_end());
   3399   RememberedSet<OLD_TO_OLD>::RemoveRangeTyped(this, free_start, area_end());
   3400 }
   3401 
   3402 // -----------------------------------------------------------------------------
   3403 // LargeObjectIterator
   3404 
   3405 LargeObjectIterator::LargeObjectIterator(LargeObjectSpace* space) {
   3406   current_ = space->first_page();
   3407 }
   3408 
   3409 
   3410 HeapObject* LargeObjectIterator::Next() {
   3411   if (current_ == nullptr) return nullptr;
   3412 
   3413   HeapObject* object = current_->GetObject();
   3414   current_ = current_->next_page();
   3415   return object;
   3416 }
   3417 
   3418 
   3419 // -----------------------------------------------------------------------------
   3420 // LargeObjectSpace
   3421 
   3422 LargeObjectSpace::LargeObjectSpace(Heap* heap)
   3423     : LargeObjectSpace(heap, LO_SPACE) {}
   3424 
   3425 LargeObjectSpace::LargeObjectSpace(Heap* heap, AllocationSpace id)
   3426     : Space(heap, id),
   3427       size_(0),
   3428       page_count_(0),
   3429       objects_size_(0),
   3430       chunk_map_(1024) {}
   3431 
   3432 void LargeObjectSpace::TearDown() {
   3433   while (!memory_chunk_list_.Empty()) {
   3434     LargePage* page = first_page();
   3435     LOG(heap()->isolate(),
   3436         DeleteEvent("LargeObjectChunk",
   3437                     reinterpret_cast<void*>(page->address())));
   3438     memory_chunk_list_.Remove(page);
   3439     heap()->memory_allocator()->Free<MemoryAllocator::kFull>(page);
   3440   }
   3441 }
   3442 
   3443 AllocationResult LargeObjectSpace::AllocateRaw(int object_size,
   3444                                                Executability executable) {
   3445   // Check if we want to force a GC before growing the old space further.
   3446   // If so, fail the allocation.
   3447   if (!heap()->CanExpandOldGeneration(object_size) ||
   3448       !heap()->ShouldExpandOldGenerationOnSlowAllocation()) {
   3449     return AllocationResult::Retry(identity());
   3450   }
   3451 
   3452   LargePage* page = AllocateLargePage(object_size, executable);
   3453   if (page == nullptr) return AllocationResult::Retry(identity());
   3454   page->SetOldGenerationPageFlags(heap()->incremental_marking()->IsMarking());
   3455   HeapObject* object = page->GetObject();
   3456   heap()->StartIncrementalMarkingIfAllocationLimitIsReached(
   3457       heap()->GCFlagsForIncrementalMarking(),
   3458       kGCCallbackScheduleIdleGarbageCollection);
   3459   if (heap()->incremental_marking()->black_allocation()) {
   3460     heap()->incremental_marking()->marking_state()->WhiteToBlack(object);
   3461   }
   3462   DCHECK_IMPLIES(
   3463       heap()->incremental_marking()->black_allocation(),
   3464       heap()->incremental_marking()->marking_state()->IsBlack(object));
   3465   page->InitializationMemoryFence();
   3466   return object;
   3467 }
   3468 
   3469 LargePage* LargeObjectSpace::AllocateLargePage(int object_size,
   3470                                                Executability executable) {
   3471   LargePage* page = heap()->memory_allocator()->AllocateLargePage(
   3472       object_size, this, executable);
   3473   if (page == nullptr) return nullptr;
   3474   DCHECK_GE(page->area_size(), static_cast<size_t>(object_size));
   3475 
   3476   size_ += static_cast<int>(page->size());
   3477   AccountCommitted(page->size());
   3478   objects_size_ += object_size;
   3479   page_count_++;
   3480   memory_chunk_list_.PushBack(page);
   3481 
   3482   InsertChunkMapEntries(page);
   3483 
   3484   HeapObject* object = page->GetObject();
   3485 
   3486   if (Heap::ShouldZapGarbage()) {
   3487     // Make the object consistent so the heap can be verified in OldSpaceStep.
   3488     // We only need to do this in debug builds or if verify_heap is on.
   3489     reinterpret_cast<Object**>(object->address())[0] =
   3490         ReadOnlyRoots(heap()).fixed_array_map();
   3491     reinterpret_cast<Object**>(object->address())[1] = Smi::kZero;
   3492   }
   3493   heap()->CreateFillerObjectAt(object->address(), object_size,
   3494                                ClearRecordedSlots::kNo);
   3495   AllocationStep(object_size, object->address(), object_size);
   3496   return page;
   3497 }
   3498 
   3499 
   3500 size_t LargeObjectSpace::CommittedPhysicalMemory() {
   3501   // On a platform that provides lazy committing of memory, we over-account
   3502   // the actually committed memory. There is no easy way right now to support
   3503   // precise accounting of committed memory in large object space.
   3504   return CommittedMemory();
   3505 }
   3506 
   3507 
   3508 // GC support
   3509 Object* LargeObjectSpace::FindObject(Address a) {
   3510   LargePage* page = FindPage(a);
   3511   if (page != nullptr) {
   3512     return page->GetObject();
   3513   }
   3514   return Smi::kZero;  // Signaling not found.
   3515 }
   3516 
   3517 LargePage* LargeObjectSpace::FindPageThreadSafe(Address a) {
   3518   base::LockGuard<base::Mutex> guard(&chunk_map_mutex_);
   3519   return FindPage(a);
   3520 }
   3521 
   3522 LargePage* LargeObjectSpace::FindPage(Address a) {
   3523   const Address key = MemoryChunk::FromAddress(a)->address();
   3524   auto it = chunk_map_.find(key);
   3525   if (it != chunk_map_.end()) {
   3526     LargePage* page = it->second;
   3527     if (page->Contains(a)) {
   3528       return page;
   3529     }
   3530   }
   3531   return nullptr;
   3532 }
   3533 
   3534 
   3535 void LargeObjectSpace::ClearMarkingStateOfLiveObjects() {
   3536   IncrementalMarking::NonAtomicMarkingState* marking_state =
   3537       heap()->incremental_marking()->non_atomic_marking_state();
   3538   LargeObjectIterator it(this);
   3539   for (HeapObject* obj = it.Next(); obj != nullptr; obj = it.Next()) {
   3540     if (marking_state->IsBlackOrGrey(obj)) {
   3541       Marking::MarkWhite(marking_state->MarkBitFrom(obj));
   3542       MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address());
   3543       RememberedSet<OLD_TO_NEW>::FreeEmptyBuckets(chunk);
   3544       chunk->ResetProgressBar();
   3545       marking_state->SetLiveBytes(chunk, 0);
   3546     }
   3547     DCHECK(marking_state->IsWhite(obj));
   3548   }
   3549 }
   3550 
   3551 void LargeObjectSpace::InsertChunkMapEntries(LargePage* page) {
   3552   // There may be concurrent access on the chunk map. We have to take the lock
   3553   // here.
   3554   base::LockGuard<base::Mutex> guard(&chunk_map_mutex_);
   3555   for (Address current = reinterpret_cast<Address>(page);
   3556        current < reinterpret_cast<Address>(page) + page->size();
   3557        current += MemoryChunk::kPageSize) {
   3558     chunk_map_[current] = page;
   3559   }
   3560 }
   3561 
   3562 void LargeObjectSpace::RemoveChunkMapEntries(LargePage* page) {
   3563   RemoveChunkMapEntries(page, page->address());
   3564 }
   3565 
   3566 void LargeObjectSpace::RemoveChunkMapEntries(LargePage* page,
   3567                                              Address free_start) {
   3568   for (Address current = ::RoundUp(free_start, MemoryChunk::kPageSize);
   3569        current < reinterpret_cast<Address>(page) + page->size();
   3570        current += MemoryChunk::kPageSize) {
   3571     chunk_map_.erase(current);
   3572   }
   3573 }
   3574 
   3575 void LargeObjectSpace::FreeUnmarkedObjects() {
   3576   LargePage* current = first_page();
   3577   IncrementalMarking::NonAtomicMarkingState* marking_state =
   3578       heap()->incremental_marking()->non_atomic_marking_state();
   3579   objects_size_ = 0;
   3580   while (current) {
   3581     LargePage* next_current = current->next_page();
   3582     HeapObject* object = current->GetObject();
   3583     DCHECK(!marking_state->IsGrey(object));
   3584     if (marking_state->IsBlack(object)) {
   3585       Address free_start;
   3586       size_t size = static_cast<size_t>(object->Size());
   3587       objects_size_ += size;
   3588       if ((free_start = current->GetAddressToShrink(object->address(), size)) !=
   3589           0) {
   3590         DCHECK(!current->IsFlagSet(Page::IS_EXECUTABLE));
   3591         current->ClearOutOfLiveRangeSlots(free_start);
   3592         RemoveChunkMapEntries(current, free_start);
   3593         const size_t bytes_to_free =
   3594             current->size() - (free_start - current->address());
   3595         heap()->memory_allocator()->PartialFreeMemory(
   3596             current, free_start, bytes_to_free,
   3597             current->area_start() + object->Size());
   3598         size_ -= bytes_to_free;
   3599         AccountUncommitted(bytes_to_free);
   3600       }
   3601     } else {
   3602       memory_chunk_list_.Remove(current);
   3603 
   3604       // Free the chunk.
   3605       size_ -= static_cast<int>(current->size());
   3606       AccountUncommitted(current->size());
   3607       page_count_--;
   3608 
   3609       RemoveChunkMapEntries(current);
   3610       heap()->memory_allocator()->Free<MemoryAllocator::kPreFreeAndQueue>(
   3611           current);
   3612     }
   3613     current = next_current;
   3614   }
   3615 }
   3616 
   3617 
   3618 bool LargeObjectSpace::Contains(HeapObject* object) {
   3619   Address address = object->address();
   3620   MemoryChunk* chunk = MemoryChunk::FromAddress(address);
   3621 
   3622   bool owned = (chunk->owner() == this);
   3623 
   3624   SLOW_DCHECK(!owned || FindObject(address)->IsHeapObject());
   3625 
   3626   return owned;
   3627 }
   3628 
   3629 std::unique_ptr<ObjectIterator> LargeObjectSpace::GetObjectIterator() {
   3630   return std::unique_ptr<ObjectIterator>(new LargeObjectIterator(this));
   3631 }
   3632 
   3633 #ifdef VERIFY_HEAP
   3634 // We do not assume that the large object iterator works, because it depends
   3635 // on the invariants we are checking during verification.
   3636 void LargeObjectSpace::Verify(Isolate* isolate) {
   3637   size_t external_backing_store_bytes[kNumTypes];
   3638 
   3639   for (int i = 0; i < kNumTypes; i++) {
   3640     external_backing_store_bytes[static_cast<ExternalBackingStoreType>(i)] = 0;
   3641   }
   3642 
   3643   for (LargePage* chunk = first_page(); chunk != nullptr;
   3644        chunk = chunk->next_page()) {
   3645     // Each chunk contains an object that starts at the large object page's
   3646     // object area start.
   3647     HeapObject* object = chunk->GetObject();
   3648     Page* page = Page::FromAddress(object->address());
   3649     CHECK(object->address() == page->area_start());
   3650 
   3651     // The first word should be a map, and we expect all map pointers to be
   3652     // in map space or read-only space.
   3653     Map* map = object->map();
   3654     CHECK(map->IsMap());
   3655     CHECK(heap()->map_space()->Contains(map) ||
   3656           heap()->read_only_space()->Contains(map));
   3657 
   3658     // We have only the following types in the large object space:
   3659     CHECK(object->IsAbstractCode() || object->IsSeqString() ||
   3660           object->IsExternalString() || object->IsThinString() ||
   3661           object->IsFixedArray() || object->IsFixedDoubleArray() ||
   3662           object->IsWeakFixedArray() || object->IsWeakArrayList() ||
   3663           object->IsPropertyArray() || object->IsByteArray() ||
   3664           object->IsFeedbackVector() || object->IsBigInt() ||
   3665           object->IsFreeSpace() || object->IsFeedbackMetadata());
   3666 
   3667     // The object itself should look OK.
   3668     object->ObjectVerify(isolate);
   3669 
   3670     if (!FLAG_verify_heap_skip_remembered_set) {
   3671       heap()->VerifyRememberedSetFor(object);
   3672     }
   3673 
   3674     // Byte arrays and strings don't have interior pointers.
   3675     if (object->IsAbstractCode()) {
   3676       VerifyPointersVisitor code_visitor(heap());
   3677       object->IterateBody(map, object->Size(), &code_visitor);
   3678     } else if (object->IsFixedArray()) {
   3679       FixedArray* array = FixedArray::cast(object);
   3680       for (int j = 0; j < array->length(); j++) {
   3681         Object* element = array->get(j);
   3682         if (element->IsHeapObject()) {
   3683           HeapObject* element_object = HeapObject::cast(element);
   3684           CHECK(heap()->Contains(element_object));
   3685           CHECK(element_object->map()->IsMap());
   3686         }
   3687       }
   3688     } else if (object->IsPropertyArray()) {
   3689       PropertyArray* array = PropertyArray::cast(object);
   3690       for (int j = 0; j < array->length(); j++) {
   3691         Object* property = array->get(j);
   3692         if (property->IsHeapObject()) {
   3693           HeapObject* property_object = HeapObject::cast(property);
   3694           CHECK(heap()->Contains(property_object));
   3695           CHECK(property_object->map()->IsMap());
   3696         }
   3697       }
   3698     }
   3699     for (int i = 0; i < kNumTypes; i++) {
   3700       ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   3701       external_backing_store_bytes[t] += chunk->ExternalBackingStoreBytes(t);
   3702     }
   3703   }
   3704   for (int i = 0; i < kNumTypes; i++) {
   3705     ExternalBackingStoreType t = static_cast<ExternalBackingStoreType>(i);
   3706     CHECK_EQ(external_backing_store_bytes[t], ExternalBackingStoreBytes(t));
   3707   }
   3708 }
   3709 #endif
   3710 
   3711 #ifdef DEBUG
   3712 void LargeObjectSpace::Print() {
   3713   StdoutStream os;
   3714   LargeObjectIterator it(this);
   3715   for (HeapObject* obj = it.Next(); obj != nullptr; obj = it.Next()) {
   3716     obj->Print(os);
   3717   }
   3718 }
   3719 
   3720 void Page::Print() {
   3721   // Make a best-effort to print the objects in the page.
   3722   PrintF("Page@%p in %s\n", reinterpret_cast<void*>(this->address()),
   3723          this->owner()->name());
   3724   printf(" --------------------------------------\n");
   3725   HeapObjectIterator objects(this);
   3726   unsigned mark_size = 0;
   3727   for (HeapObject* object = objects.Next(); object != nullptr;
   3728        object = objects.Next()) {
   3729     bool is_marked =
   3730         heap()->incremental_marking()->marking_state()->IsBlackOrGrey(object);
   3731     PrintF(" %c ", (is_marked ? '!' : ' '));  // Indent a little.
   3732     if (is_marked) {
   3733       mark_size += object->Size();
   3734     }
   3735     object->ShortPrint();
   3736     PrintF("\n");
   3737   }
   3738   printf(" --------------------------------------\n");
   3739   printf(" Marked: %x, LiveCount: %" V8PRIdPTR "\n", mark_size,
   3740          heap()->incremental_marking()->marking_state()->live_bytes(this));
   3741 }
   3742 
   3743 #endif  // DEBUG
   3744 
   3745 NewLargeObjectSpace::NewLargeObjectSpace(Heap* heap)
   3746     : LargeObjectSpace(heap, NEW_LO_SPACE) {}
   3747 
   3748 AllocationResult NewLargeObjectSpace::AllocateRaw(int object_size) {
   3749   // TODO(hpayer): Add heap growing strategy here.
   3750   LargePage* page = AllocateLargePage(object_size, NOT_EXECUTABLE);
   3751   if (page == nullptr) return AllocationResult::Retry(identity());
   3752   page->SetYoungGenerationPageFlags(heap()->incremental_marking()->IsMarking());
   3753   page->SetFlag(MemoryChunk::IN_TO_SPACE);
   3754   page->InitializationMemoryFence();
   3755   return page->GetObject();
   3756 }
   3757 
   3758 size_t NewLargeObjectSpace::Available() {
   3759   // TODO(hpayer): Update as soon as we have a growing strategy.
   3760   return 0;
   3761 }
   3762 }  // namespace internal
   3763 }  // namespace v8
   3764