1 // Copyright 2015 the V8 project authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style license that can be 3 // found in the LICENSE file. 4 5 #include "src/heap/memory-reducer.h" 6 7 #include "src/flags.h" 8 #include "src/heap/gc-tracer.h" 9 #include "src/heap/heap-inl.h" 10 #include "src/utils.h" 11 #include "src/v8.h" 12 13 namespace v8 { 14 namespace internal { 15 16 const int MemoryReducer::kLongDelayMs = 8000; 17 const int MemoryReducer::kShortDelayMs = 500; 18 const int MemoryReducer::kWatchdogDelayMs = 100000; 19 const int MemoryReducer::kMaxNumberOfGCs = 3; 20 const double MemoryReducer::kCommittedMemoryFactor = 1.1; 21 const size_t MemoryReducer::kCommittedMemoryDelta = 10 * MB; 22 TimerTask(MemoryReducer * memory_reducer)23 MemoryReducer::TimerTask::TimerTask(MemoryReducer* memory_reducer) 24 : CancelableTask(memory_reducer->heap()->isolate()), 25 memory_reducer_(memory_reducer) {} 26 27 RunInternal()28 void MemoryReducer::TimerTask::RunInternal() { 29 Heap* heap = memory_reducer_->heap(); 30 Event event; 31 double time_ms = heap->MonotonicallyIncreasingTimeInMs(); 32 heap->tracer()->SampleAllocation(time_ms, heap->NewSpaceAllocationCounter(), 33 heap->OldGenerationAllocationCounter()); 34 bool low_allocation_rate = heap->HasLowAllocationRate(); 35 bool optimize_for_memory = heap->ShouldOptimizeForMemoryUsage(); 36 if (FLAG_trace_gc_verbose) { 37 heap->isolate()->PrintWithTimestamp( 38 "Memory reducer: %s, %s\n", 39 low_allocation_rate ? "low alloc" : "high alloc", 40 optimize_for_memory ? "background" : "foreground"); 41 } 42 event.type = kTimer; 43 event.time_ms = time_ms; 44 // The memory reducer will start incremental markig if 45 // 1) mutator is likely idle: js call rate is low and allocation rate is low. 46 // 2) mutator is in background: optimize for memory flag is set. 47 event.should_start_incremental_gc = 48 low_allocation_rate || optimize_for_memory; 49 event.can_start_incremental_gc = 50 heap->incremental_marking()->IsStopped() && 51 (heap->incremental_marking()->CanBeActivated() || optimize_for_memory); 52 event.committed_memory = heap->CommittedOldGenerationMemory(); 53 memory_reducer_->NotifyTimer(event); 54 } 55 56 NotifyTimer(const Event & event)57 void MemoryReducer::NotifyTimer(const Event& event) { 58 DCHECK_EQ(kTimer, event.type); 59 DCHECK_EQ(kWait, state_.action); 60 state_ = Step(state_, event); 61 if (state_.action == kRun) { 62 DCHECK(heap()->incremental_marking()->IsStopped()); 63 DCHECK(FLAG_incremental_marking); 64 if (FLAG_trace_gc_verbose) { 65 heap()->isolate()->PrintWithTimestamp("Memory reducer: started GC #%d\n", 66 state_.started_gcs); 67 } 68 heap()->StartIdleIncrementalMarking( 69 GarbageCollectionReason::kMemoryReducer); 70 } else if (state_.action == kWait) { 71 if (!heap()->incremental_marking()->IsStopped() && 72 heap()->ShouldOptimizeForMemoryUsage()) { 73 // Make progress with pending incremental marking if memory usage has 74 // higher priority than latency. This is important for background tabs 75 // that do not send idle notifications. 76 const int kIncrementalMarkingDelayMs = 500; 77 double deadline = heap()->MonotonicallyIncreasingTimeInMs() + 78 kIncrementalMarkingDelayMs; 79 heap()->incremental_marking()->AdvanceIncrementalMarking( 80 deadline, IncrementalMarking::NO_GC_VIA_STACK_GUARD, 81 IncrementalMarking::FORCE_COMPLETION, StepOrigin::kTask); 82 heap()->FinalizeIncrementalMarkingIfComplete( 83 GarbageCollectionReason::kFinalizeMarkingViaTask); 84 } 85 // Re-schedule the timer. 86 ScheduleTimer(event.time_ms, state_.next_gc_start_ms - event.time_ms); 87 if (FLAG_trace_gc_verbose) { 88 heap()->isolate()->PrintWithTimestamp( 89 "Memory reducer: waiting for %.f ms\n", 90 state_.next_gc_start_ms - event.time_ms); 91 } 92 } 93 } 94 95 NotifyMarkCompact(const Event & event)96 void MemoryReducer::NotifyMarkCompact(const Event& event) { 97 DCHECK_EQ(kMarkCompact, event.type); 98 Action old_action = state_.action; 99 state_ = Step(state_, event); 100 if (old_action != kWait && state_.action == kWait) { 101 // If we are transitioning to the WAIT state, start the timer. 102 ScheduleTimer(event.time_ms, state_.next_gc_start_ms - event.time_ms); 103 } 104 if (old_action == kRun) { 105 if (FLAG_trace_gc_verbose) { 106 heap()->isolate()->PrintWithTimestamp( 107 "Memory reducer: finished GC #%d (%s)\n", state_.started_gcs, 108 state_.action == kWait ? "will do more" : "done"); 109 } 110 } 111 } 112 NotifyPossibleGarbage(const Event & event)113 void MemoryReducer::NotifyPossibleGarbage(const Event& event) { 114 DCHECK_EQ(kPossibleGarbage, event.type); 115 Action old_action = state_.action; 116 state_ = Step(state_, event); 117 if (old_action != kWait && state_.action == kWait) { 118 // If we are transitioning to the WAIT state, start the timer. 119 ScheduleTimer(event.time_ms, state_.next_gc_start_ms - event.time_ms); 120 } 121 } 122 123 WatchdogGC(const State & state,const Event & event)124 bool MemoryReducer::WatchdogGC(const State& state, const Event& event) { 125 return state.last_gc_time_ms != 0 && 126 event.time_ms > state.last_gc_time_ms + kWatchdogDelayMs; 127 } 128 129 130 // For specification of this function see the comment for MemoryReducer class. Step(const State & state,const Event & event)131 MemoryReducer::State MemoryReducer::Step(const State& state, 132 const Event& event) { 133 if (!FLAG_incremental_marking || !FLAG_memory_reducer) { 134 return State(kDone, 0, 0, state.last_gc_time_ms, 0); 135 } 136 switch (state.action) { 137 case kDone: 138 if (event.type == kTimer) { 139 return state; 140 } else if (event.type == kMarkCompact) { 141 if (event.committed_memory < 142 Max(static_cast<size_t>(state.committed_memory_at_last_run * 143 kCommittedMemoryFactor), 144 state.committed_memory_at_last_run + kCommittedMemoryDelta)) { 145 return state; 146 } else { 147 return State(kWait, 0, event.time_ms + kLongDelayMs, 148 event.type == kMarkCompact ? event.time_ms 149 : state.last_gc_time_ms, 150 0); 151 } 152 } else { 153 DCHECK_EQ(kPossibleGarbage, event.type); 154 return State( 155 kWait, 0, event.time_ms + kLongDelayMs, 156 event.type == kMarkCompact ? event.time_ms : state.last_gc_time_ms, 157 0); 158 } 159 case kWait: 160 switch (event.type) { 161 case kPossibleGarbage: 162 return state; 163 case kTimer: 164 if (state.started_gcs >= kMaxNumberOfGCs) { 165 return State(kDone, kMaxNumberOfGCs, 0.0, state.last_gc_time_ms, 166 event.committed_memory); 167 } else if (event.can_start_incremental_gc && 168 (event.should_start_incremental_gc || 169 WatchdogGC(state, event))) { 170 if (state.next_gc_start_ms <= event.time_ms) { 171 return State(kRun, state.started_gcs + 1, 0.0, 172 state.last_gc_time_ms, 0); 173 } else { 174 return state; 175 } 176 } else { 177 return State(kWait, state.started_gcs, event.time_ms + kLongDelayMs, 178 state.last_gc_time_ms, 0); 179 } 180 case kMarkCompact: 181 return State(kWait, state.started_gcs, event.time_ms + kLongDelayMs, 182 event.time_ms, 0); 183 } 184 case kRun: 185 if (event.type != kMarkCompact) { 186 return state; 187 } else { 188 if (state.started_gcs < kMaxNumberOfGCs && 189 (event.next_gc_likely_to_collect_more || state.started_gcs == 1)) { 190 return State(kWait, state.started_gcs, event.time_ms + kShortDelayMs, 191 event.time_ms, 0); 192 } else { 193 return State(kDone, kMaxNumberOfGCs, 0.0, event.time_ms, 194 event.committed_memory); 195 } 196 } 197 } 198 UNREACHABLE(); 199 return State(kDone, 0, 0, 0.0, 0); // Make the compiler happy. 200 } 201 202 ScheduleTimer(double time_ms,double delay_ms)203 void MemoryReducer::ScheduleTimer(double time_ms, double delay_ms) { 204 DCHECK(delay_ms > 0); 205 // Leave some room for precision error in task scheduler. 206 const double kSlackMs = 100; 207 v8::Isolate* isolate = reinterpret_cast<v8::Isolate*>(heap()->isolate()); 208 auto timer_task = new MemoryReducer::TimerTask(this); 209 V8::GetCurrentPlatform()->CallDelayedOnForegroundThread( 210 isolate, timer_task, (delay_ms + kSlackMs) / 1000.0); 211 } 212 TearDown()213 void MemoryReducer::TearDown() { state_ = State(kDone, 0, 0, 0.0, 0); } 214 215 } // namespace internal 216 } // namespace v8 217