Home | History | Annotate | Download | only in heap
      1 // Copyright 2015 the V8 project authors. All rights reserved.
      2 // Use of this source code is governed by a BSD-style license that can be
      3 // found in the LICENSE file.
      4 
      5 #include "src/heap/memory-reducer.h"
      6 
      7 #include "src/flags.h"
      8 #include "src/heap/gc-tracer.h"
      9 #include "src/heap/heap-inl.h"
     10 #include "src/utils.h"
     11 #include "src/v8.h"
     12 
     13 namespace v8 {
     14 namespace internal {
     15 
     16 const int MemoryReducer::kLongDelayMs = 8000;
     17 const int MemoryReducer::kShortDelayMs = 500;
     18 const int MemoryReducer::kWatchdogDelayMs = 100000;
     19 const int MemoryReducer::kMaxNumberOfGCs = 3;
     20 
     21 MemoryReducer::TimerTask::TimerTask(MemoryReducer* memory_reducer)
     22     : CancelableTask(memory_reducer->heap()->isolate()),
     23       memory_reducer_(memory_reducer) {}
     24 
     25 
     26 void MemoryReducer::TimerTask::RunInternal() {
     27   const double kJsCallsPerMsThreshold = 0.5;
     28   Heap* heap = memory_reducer_->heap();
     29   Event event;
     30   double time_ms = heap->MonotonicallyIncreasingTimeInMs();
     31   heap->tracer()->SampleAllocation(time_ms, heap->NewSpaceAllocationCounter(),
     32                                    heap->OldGenerationAllocationCounter());
     33   double js_call_rate = memory_reducer_->SampleAndGetJsCallsPerMs(time_ms);
     34   bool low_allocation_rate = heap->HasLowAllocationRate();
     35   bool is_idle = js_call_rate < kJsCallsPerMsThreshold && low_allocation_rate;
     36   bool optimize_for_memory = heap->ShouldOptimizeForMemoryUsage();
     37   if (FLAG_trace_gc_verbose) {
     38     PrintIsolate(heap->isolate(), "Memory reducer: call rate %.3lf, %s, %s\n",
     39                  js_call_rate, low_allocation_rate ? "low alloc" : "high alloc",
     40                  optimize_for_memory ? "background" : "foreground");
     41   }
     42   event.type = kTimer;
     43   event.time_ms = time_ms;
     44   // The memory reducer will start incremental markig if
     45   // 1) mutator is likely idle: js call rate is low and allocation rate is low.
     46   // 2) mutator is in background: optimize for memory flag is set.
     47   event.should_start_incremental_gc = is_idle || optimize_for_memory;
     48   event.can_start_incremental_gc =
     49       heap->incremental_marking()->IsStopped() &&
     50       (heap->incremental_marking()->CanBeActivated() || optimize_for_memory);
     51   memory_reducer_->NotifyTimer(event);
     52 }
     53 
     54 
     55 double MemoryReducer::SampleAndGetJsCallsPerMs(double time_ms) {
     56   unsigned int counter = heap()->isolate()->js_calls_from_api_counter();
     57   unsigned int call_delta = counter - js_calls_counter_;
     58   double time_delta_ms = time_ms - js_calls_sample_time_ms_;
     59   js_calls_counter_ = counter;
     60   js_calls_sample_time_ms_ = time_ms;
     61   return time_delta_ms > 0 ? call_delta / time_delta_ms : 0;
     62 }
     63 
     64 
     65 void MemoryReducer::NotifyTimer(const Event& event) {
     66   DCHECK_EQ(kTimer, event.type);
     67   DCHECK_EQ(kWait, state_.action);
     68   state_ = Step(state_, event);
     69   if (state_.action == kRun) {
     70     DCHECK(heap()->incremental_marking()->IsStopped());
     71     DCHECK(FLAG_incremental_marking);
     72     if (FLAG_trace_gc_verbose) {
     73       PrintIsolate(heap()->isolate(), "Memory reducer: started GC #%d\n",
     74                    state_.started_gcs);
     75     }
     76     heap()->StartIdleIncrementalMarking();
     77   } else if (state_.action == kWait) {
     78     if (!heap()->incremental_marking()->IsStopped() &&
     79         heap()->ShouldOptimizeForMemoryUsage()) {
     80       // Make progress with pending incremental marking if memory usage has
     81       // higher priority than latency. This is important for background tabs
     82       // that do not send idle notifications.
     83       const int kIncrementalMarkingDelayMs = 500;
     84       double deadline = heap()->MonotonicallyIncreasingTimeInMs() +
     85                         kIncrementalMarkingDelayMs;
     86       heap()->incremental_marking()->AdvanceIncrementalMarking(
     87           deadline, i::IncrementalMarking::StepActions(
     88                         i::IncrementalMarking::NO_GC_VIA_STACK_GUARD,
     89                         i::IncrementalMarking::FORCE_MARKING,
     90                         i::IncrementalMarking::FORCE_COMPLETION));
     91       heap()->FinalizeIncrementalMarkingIfComplete(
     92           "Memory reducer: finalize incremental marking");
     93     }
     94     // Re-schedule the timer.
     95     ScheduleTimer(event.time_ms, state_.next_gc_start_ms - event.time_ms);
     96     if (FLAG_trace_gc_verbose) {
     97       PrintIsolate(heap()->isolate(), "Memory reducer: waiting for %.f ms\n",
     98                    state_.next_gc_start_ms - event.time_ms);
     99     }
    100   }
    101 }
    102 
    103 
    104 void MemoryReducer::NotifyMarkCompact(const Event& event) {
    105   DCHECK_EQ(kMarkCompact, event.type);
    106   Action old_action = state_.action;
    107   state_ = Step(state_, event);
    108   if (old_action != kWait && state_.action == kWait) {
    109     // If we are transitioning to the WAIT state, start the timer.
    110     ScheduleTimer(event.time_ms, state_.next_gc_start_ms - event.time_ms);
    111   }
    112   if (old_action == kRun) {
    113     if (FLAG_trace_gc_verbose) {
    114       PrintIsolate(heap()->isolate(), "Memory reducer: finished GC #%d (%s)\n",
    115                    state_.started_gcs,
    116                    state_.action == kWait ? "will do more" : "done");
    117     }
    118   }
    119 }
    120 
    121 void MemoryReducer::NotifyPossibleGarbage(const Event& event) {
    122   DCHECK_EQ(kPossibleGarbage, event.type);
    123   Action old_action = state_.action;
    124   state_ = Step(state_, event);
    125   if (old_action != kWait && state_.action == kWait) {
    126     // If we are transitioning to the WAIT state, start the timer.
    127     ScheduleTimer(event.time_ms, state_.next_gc_start_ms - event.time_ms);
    128   }
    129 }
    130 
    131 
    132 bool MemoryReducer::WatchdogGC(const State& state, const Event& event) {
    133   return state.last_gc_time_ms != 0 &&
    134          event.time_ms > state.last_gc_time_ms + kWatchdogDelayMs;
    135 }
    136 
    137 
    138 // For specification of this function see the comment for MemoryReducer class.
    139 MemoryReducer::State MemoryReducer::Step(const State& state,
    140                                          const Event& event) {
    141   if (!FLAG_incremental_marking || !FLAG_memory_reducer) {
    142     return State(kDone, 0, 0, state.last_gc_time_ms);
    143   }
    144   switch (state.action) {
    145     case kDone:
    146       if (event.type == kTimer) {
    147         return state;
    148       } else {
    149         DCHECK(event.type == kPossibleGarbage || event.type == kMarkCompact);
    150         return State(
    151             kWait, 0, event.time_ms + kLongDelayMs,
    152             event.type == kMarkCompact ? event.time_ms : state.last_gc_time_ms);
    153       }
    154     case kWait:
    155       switch (event.type) {
    156         case kPossibleGarbage:
    157           return state;
    158         case kTimer:
    159           if (state.started_gcs >= kMaxNumberOfGCs) {
    160             return State(kDone, kMaxNumberOfGCs, 0.0, state.last_gc_time_ms);
    161           } else if (event.can_start_incremental_gc &&
    162                      (event.should_start_incremental_gc ||
    163                       WatchdogGC(state, event))) {
    164             if (state.next_gc_start_ms <= event.time_ms) {
    165               return State(kRun, state.started_gcs + 1, 0.0,
    166                            state.last_gc_time_ms);
    167             } else {
    168               return state;
    169             }
    170           } else {
    171             return State(kWait, state.started_gcs, event.time_ms + kLongDelayMs,
    172                          state.last_gc_time_ms);
    173           }
    174         case kMarkCompact:
    175           return State(kWait, state.started_gcs, event.time_ms + kLongDelayMs,
    176                        event.time_ms);
    177       }
    178     case kRun:
    179       if (event.type != kMarkCompact) {
    180         return state;
    181       } else {
    182         if (state.started_gcs < kMaxNumberOfGCs &&
    183             (event.next_gc_likely_to_collect_more || state.started_gcs == 1)) {
    184           return State(kWait, state.started_gcs, event.time_ms + kShortDelayMs,
    185                        event.time_ms);
    186         } else {
    187           return State(kDone, kMaxNumberOfGCs, 0.0, event.time_ms);
    188         }
    189       }
    190   }
    191   UNREACHABLE();
    192   return State(kDone, 0, 0, 0.0);  // Make the compiler happy.
    193 }
    194 
    195 
    196 void MemoryReducer::ScheduleTimer(double time_ms, double delay_ms) {
    197   DCHECK(delay_ms > 0);
    198   // Record the time and the js call counter.
    199   SampleAndGetJsCallsPerMs(time_ms);
    200   // Leave some room for precision error in task scheduler.
    201   const double kSlackMs = 100;
    202   v8::Isolate* isolate = reinterpret_cast<v8::Isolate*>(heap()->isolate());
    203   auto timer_task = new MemoryReducer::TimerTask(this);
    204   V8::GetCurrentPlatform()->CallDelayedOnForegroundThread(
    205       isolate, timer_task, (delay_ms + kSlackMs) / 1000.0);
    206 }
    207 
    208 
    209 void MemoryReducer::TearDown() { state_ = State(kDone, 0, 0, 0.0); }
    210 
    211 }  // namespace internal
    212 }  // namespace v8
    213