Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(530)

Unified Diff: src/heap/incremental-marking.cc

Issue 2359903002: [heap] New heuristics for incremental marking step size. (Closed)
Patch Set: use space iterator Created 4 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/heap/incremental-marking.h ('k') | src/ia32/code-stubs-ia32.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/heap/incremental-marking.cc
diff --git a/src/heap/incremental-marking.cc b/src/heap/incremental-marking.cc
index 2e7128edd1eab859663eef6ae83dc83dd160f65d..9f334f74a1045d44cb8cd20fdf780f32d2b4eb3e 100644
--- a/src/heap/incremental-marking.cc
+++ b/src/heap/incremental-marking.cc
@@ -21,26 +21,20 @@ namespace internal {
IncrementalMarking::IncrementalMarking(Heap* heap)
: heap_(heap),
- observer_(*this, kAllocatedThreshold),
state_(STOPPED),
- is_compacting_(false),
- steps_count_(0),
- old_generation_space_available_at_start_of_incremental_(0),
- old_generation_space_used_at_start_of_incremental_(0),
- bytes_rescanned_(0),
- should_hurry_(false),
- marking_speed_(0),
- bytes_scanned_(0),
- allocated_(0),
- write_barriers_invoked_since_last_step_(0),
+ initial_old_generation_size_(0),
bytes_marked_ahead_of_schedule_(0),
- idle_marking_delay_counter_(0),
unscanned_bytes_of_large_object_(0),
+ idle_marking_delay_counter_(0),
+ incremental_marking_finalization_rounds_(0),
+ is_compacting_(false),
+ should_hurry_(false),
was_activated_(false),
black_allocation_(false),
finalize_marking_completed_(false),
- incremental_marking_finalization_rounds_(0),
- request_type_(NONE) {}
+ request_type_(NONE),
+ new_generation_observer_(*this, kAllocatedThreshold),
+ old_generation_observer_(*this, kAllocatedThreshold) {}
bool IncrementalMarking::BaseRecordWrite(HeapObject* obj, Object* value) {
HeapObject* value_heap_obj = HeapObject::cast(value);
@@ -71,19 +65,7 @@ void IncrementalMarking::RecordWriteSlow(HeapObject* obj, Object** slot,
void IncrementalMarking::RecordWriteFromCode(HeapObject* obj, Object** slot,
Isolate* isolate) {
DCHECK(obj->IsHeapObject());
- IncrementalMarking* marking = isolate->heap()->incremental_marking();
-
- MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address());
- int counter = chunk->write_barrier_counter();
- if (counter < (MemoryChunk::kWriteBarrierCounterGranularity / 2)) {
- marking->write_barriers_invoked_since_last_step_ +=
- MemoryChunk::kWriteBarrierCounterGranularity -
- chunk->write_barrier_counter();
- chunk->set_write_barrier_counter(
- MemoryChunk::kWriteBarrierCounterGranularity);
- }
-
- marking->RecordWrite(obj, slot, *slot);
+ isolate->heap()->incremental_marking()->RecordWrite(obj, slot, *slot);
}
// static
@@ -462,21 +444,6 @@ void IncrementalMarking::ActivateGeneratedStub(Code* stub) {
}
-void IncrementalMarking::NotifyOfHighPromotionRate() {
- if (IsMarking()) {
- if (marking_speed_ < kFastMarking) {
- if (FLAG_trace_gc) {
- heap()->isolate()->PrintWithTimestamp(
- "Increasing marking speed to %d "
- "due to high promotion rate\n",
- static_cast<int>(kFastMarking));
- }
- marking_speed_ = kFastMarking;
- }
- }
-}
-
-
static void PatchIncrementalMarkingRecordWriteStubs(
Heap* heap, RecordWriteStub::Mode mode) {
UnseededNumberDictionary* stubs = heap->code_stubs();
@@ -523,9 +490,14 @@ void IncrementalMarking::Start(GarbageCollectionReason gc_reason) {
HistogramTimerScope incremental_marking_scope(
counters->gc_incremental_marking_start());
TRACE_EVENT0("v8", "V8.GCIncrementalMarkingStart");
- ResetStepCounters();
heap_->tracer()->NotifyIncrementalMarkingStart();
+ start_time_ms_ = heap()->MonotonicallyIncreasingTimeInMs();
+ initial_old_generation_size_ = heap_->PromotedSpaceSizeOfObjects();
+ old_generation_allocation_counter_ = heap_->OldGenerationAllocationCounter();
+ bytes_allocated_ = 0;
+ bytes_marked_ahead_of_schedule_ = 0;
+ should_hurry_ = false;
was_activated_ = true;
if (!heap_->mark_compact_collector()->sweeping_in_progress()) {
@@ -538,7 +510,15 @@ void IncrementalMarking::Start(GarbageCollectionReason gc_reason) {
state_ = SWEEPING;
}
- heap_->new_space()->AddAllocationObserver(&observer_);
+ SpaceIterator it(heap_);
+ while (it.has_next()) {
+ Space* space = it.next();
+ if (space == heap_->new_space()) {
+ space->AddAllocationObserver(&new_generation_observer_);
+ } else {
+ space->AddAllocationObserver(&old_generation_observer_);
+ }
+ }
incremental_marking_job()->Start(heap_);
}
@@ -1002,9 +982,17 @@ void IncrementalMarking::Stop() {
Max(0, old_generation_size_mb - old_generation_limit_mb));
}
- heap_->new_space()->RemoveAllocationObserver(&observer_);
+ SpaceIterator it(heap_);
+ while (it.has_next()) {
+ Space* space = it.next();
+ if (space == heap_->new_space()) {
+ space->RemoveAllocationObserver(&new_generation_observer_);
+ } else {
+ space->RemoveAllocationObserver(&old_generation_observer_);
+ }
+ }
+
IncrementalMarking::set_should_hurry(false);
- ResetStepCounters();
if (IsMarking()) {
PatchIncrementalMarkingRecordWriteStubs(heap_,
RecordWriteStub::STORE_BUFFER_ONLY);
@@ -1082,82 +1070,6 @@ double IncrementalMarking::AdvanceIncrementalMarking(
}
-void IncrementalMarking::SpeedUp() {
- bool speed_up = false;
-
- if ((steps_count_ % kMarkingSpeedAccellerationInterval) == 0) {
- if (FLAG_trace_incremental_marking) {
- heap()->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Speed up marking after %d steps\n",
- static_cast<int>(kMarkingSpeedAccellerationInterval));
- }
- speed_up = true;
- }
-
- bool space_left_is_very_small =
- (old_generation_space_available_at_start_of_incremental_ < 10 * MB);
-
- bool only_1_nth_of_space_that_was_available_still_left =
- (SpaceLeftInOldSpace() * (marking_speed_ + 1) <
- old_generation_space_available_at_start_of_incremental_);
-
- if (space_left_is_very_small ||
- only_1_nth_of_space_that_was_available_still_left) {
- if (FLAG_trace_incremental_marking)
- heap()->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Speed up marking because of low space left\n");
- speed_up = true;
- }
-
- bool size_of_old_space_multiplied_by_n_during_marking =
- (heap_->PromotedTotalSize() >
- (marking_speed_ + 1) *
- old_generation_space_used_at_start_of_incremental_);
- if (size_of_old_space_multiplied_by_n_during_marking) {
- speed_up = true;
- if (FLAG_trace_incremental_marking) {
- heap()->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Speed up marking because of heap size "
- "increase\n");
- }
- }
-
- int64_t promoted_during_marking =
- heap_->PromotedTotalSize() -
- old_generation_space_used_at_start_of_incremental_;
- intptr_t delay = marking_speed_ * MB;
- intptr_t scavenge_slack = heap_->MaxSemiSpaceSize();
-
- // We try to scan at at least twice the speed that we are allocating.
- if (promoted_during_marking > bytes_scanned_ / 2 + scavenge_slack + delay) {
- if (FLAG_trace_incremental_marking) {
- heap()->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Speed up marking because marker was not "
- "keeping up\n");
- }
- speed_up = true;
- }
-
- if (speed_up) {
- if (state_ != MARKING) {
- if (FLAG_trace_incremental_marking) {
- heap()->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Postponing speeding up marking until marking "
- "starts\n");
- }
- } else {
- marking_speed_ += kMarkingSpeedAccelleration;
- marking_speed_ = static_cast<int>(
- Min(kMaxMarkingSpeed, static_cast<intptr_t>(marking_speed_ * 1.3)));
- if (FLAG_trace_incremental_marking) {
- heap()->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Marking speed increased to %d\n",
- marking_speed_);
- }
- }
- }
-}
-
void IncrementalMarking::FinalizeSweeping() {
DCHECK(state_ == SWEEPING);
if (heap_->mark_compact_collector()->sweeping_in_progress() &&
@@ -1166,59 +1078,80 @@ void IncrementalMarking::FinalizeSweeping() {
heap_->mark_compact_collector()->EnsureSweepingCompleted();
}
if (!heap_->mark_compact_collector()->sweeping_in_progress()) {
- bytes_scanned_ = 0;
StartMarking();
}
}
-void IncrementalMarking::NotifyAllocatedBytes(intptr_t allocated_bytes) {
+size_t IncrementalMarking::StepSizeToKeepUpWithAllocations() {
+ // Update bytes_allocated_ based on the allocation counter.
+ size_t current_counter = heap_->OldGenerationAllocationCounter();
+ bytes_allocated_ += current_counter - old_generation_allocation_counter_;
+ old_generation_allocation_counter_ = current_counter;
+ return bytes_allocated_;
+}
+
+size_t IncrementalMarking::StepSizeToMakeProgress() {
+ // We increase step size gradually based on the time passed in order to
+ // leave marking work to standalone tasks. The ramp up duration and the
+ // target step count are chosen based on benchmarks.
+ const int kRampUpIntervalMs = 300;
+ const size_t kTargetStepCount = 128;
+ size_t step_size = Max(initial_old_generation_size_ / kTargetStepCount,
+ IncrementalMarking::kAllocatedThreshold);
+ double time_passed_ms =
+ heap_->MonotonicallyIncreasingTimeInMs() - start_time_ms_;
+ double factor = Min(time_passed_ms / kRampUpIntervalMs, 1.0);
+ return static_cast<size_t>(factor * step_size);
+}
+
+void IncrementalMarking::AdvanceIncrementalMarkingOnAllocation() {
if (heap_->gc_state() != Heap::NOT_IN_GC || !FLAG_incremental_marking ||
(state_ != SWEEPING && state_ != MARKING)) {
return;
}
- allocated_ += allocated_bytes;
-
- if (allocated_ >= kAllocatedThreshold ||
- write_barriers_invoked_since_last_step_ >=
- kWriteBarriersInvokedThreshold) {
- // The marking speed is driven either by the allocation rate or by the rate
- // at which we are having to check the color of objects in the write
- // barrier.
- // It is possible for a tight non-allocating loop to run a lot of write
- // barriers before we get here and check them (marking can only take place
- // on
- // allocation), so to reduce the lumpiness we don't use the write barriers
- // invoked since last step directly to determine the amount of work to do.
- intptr_t bytes_to_process =
- marking_speed_ *
- Max(allocated_, write_barriers_invoked_since_last_step_);
- Step(bytes_to_process, GC_VIA_STACK_GUARD, FORCE_COMPLETION,
- StepOrigin::kV8);
+ size_t bytes_to_process =
+ StepSizeToKeepUpWithAllocations() + StepSizeToMakeProgress();
+
+ if (bytes_to_process >= IncrementalMarking::kAllocatedThreshold) {
+ // The first step after Scavenge will see many allocated bytes.
+ // Cap the step size to distribute the marking work more uniformly.
+ size_t max_step_size = GCIdleTimeHandler::EstimateMarkingStepSize(
+ kMaxStepSizeInMs,
+ heap()->tracer()->IncrementalMarkingSpeedInBytesPerMillisecond());
+ bytes_to_process = Min(bytes_to_process, max_step_size);
+
+ intptr_t bytes_processed = 0;
+ if (bytes_marked_ahead_of_schedule_ >= bytes_to_process) {
+ // Steps performed in tasks have put us ahead of schedule.
+ // We skip processing of marking dequeue here and thus
+ // shift marking time from inside V8 to standalone tasks.
+ bytes_marked_ahead_of_schedule_ -= bytes_to_process;
+ bytes_processed = bytes_to_process;
+ } else {
+ bytes_processed = Step(bytes_to_process, GC_VIA_STACK_GUARD,
+ FORCE_COMPLETION, StepOrigin::kV8);
+ }
+ bytes_allocated_ -= Min(bytes_allocated_, bytes_to_process);
}
}
-void IncrementalMarking::Step(intptr_t bytes_to_process,
- CompletionAction action,
- ForceCompletionAction completion,
- StepOrigin step_origin) {
+size_t IncrementalMarking::Step(size_t bytes_to_process,
+ CompletionAction action,
+ ForceCompletionAction completion,
+ StepOrigin step_origin) {
HistogramTimerScope incremental_marking_scope(
heap_->isolate()->counters()->gc_incremental_marking());
TRACE_EVENT0("v8", "V8.GCIncrementalMarking");
TRACE_GC(heap_->tracer(), GCTracer::Scope::MC_INCREMENTAL);
double start = heap_->MonotonicallyIncreasingTimeInMs();
- bytes_scanned_ += bytes_to_process;
-
- allocated_ = 0;
- write_barriers_invoked_since_last_step_ = 0;
-
if (state_ == SWEEPING) {
TRACE_GC(heap_->tracer(), GCTracer::Scope::MC_INCREMENTAL_SWEEPING);
FinalizeSweeping();
}
- intptr_t bytes_processed = 0;
+ size_t bytes_processed = 0;
if (state_ == MARKING) {
const bool incremental_wrapper_tracing =
FLAG_incremental_marking_wrappers && heap_->UsingEmbedderHeapTracer();
@@ -1229,17 +1162,9 @@ void IncrementalMarking::Step(intptr_t bytes_to_process,
heap_->mark_compact_collector()->marking_deque()->IsEmpty());
bool wrapper_work_left = incremental_wrapper_tracing;
if (!process_wrappers) {
- if (step_origin == StepOrigin::kV8 &&
- bytes_marked_ahead_of_schedule_ >= bytes_to_process) {
- // Steps performed in tasks have put us ahead of schedule.
- // We skip processing of marking dequeue here and thus
- // shift marking time from inside V8 to standalone tasks.
- bytes_marked_ahead_of_schedule_ -= bytes_to_process;
- } else {
- bytes_processed = ProcessMarkingDeque(bytes_to_process);
- if (step_origin == StepOrigin::kTask) {
- bytes_marked_ahead_of_schedule_ += bytes_processed;
- }
+ bytes_processed = ProcessMarkingDeque(bytes_to_process);
+ if (step_origin == StepOrigin::kTask) {
+ bytes_marked_ahead_of_schedule_ += bytes_processed;
}
} else {
const double wrapper_deadline =
@@ -1271,12 +1196,6 @@ void IncrementalMarking::Step(intptr_t bytes_to_process,
}
}
- steps_count_++;
-
- // Speed up marking if we are marking too slow or if we are almost done
- // with marking.
- SpeedUp();
-
double end = heap_->MonotonicallyIncreasingTimeInMs();
double duration = (end - start);
// Note that we report zero bytes here when sweeping was in progress or
@@ -1285,30 +1204,11 @@ void IncrementalMarking::Step(intptr_t bytes_to_process,
heap_->tracer()->AddIncrementalMarkingStep(duration, bytes_processed);
if (FLAG_trace_incremental_marking) {
heap_->isolate()->PrintWithTimestamp(
- "[IncrementalMarking] Step %s %d bytes (%d) in %.1f\n",
- step_origin == StepOrigin::kV8 ? "in v8" : "in task",
- static_cast<int>(bytes_processed), static_cast<int>(bytes_to_process),
- duration);
+ "[IncrementalMarking] Step %s %zu bytes (%zu) in %.1f\n",
+ step_origin == StepOrigin::kV8 ? "in v8" : "in task", bytes_processed,
+ bytes_to_process, duration);
}
-}
-
-
-void IncrementalMarking::ResetStepCounters() {
- steps_count_ = 0;
- old_generation_space_available_at_start_of_incremental_ =
- SpaceLeftInOldSpace();
- old_generation_space_used_at_start_of_incremental_ =
- heap_->PromotedTotalSize();
- bytes_rescanned_ = 0;
- marking_speed_ = kInitialMarkingSpeed;
- bytes_scanned_ = 0;
- write_barriers_invoked_since_last_step_ = 0;
- bytes_marked_ahead_of_schedule_ = 0;
-}
-
-
-int64_t IncrementalMarking::SpaceLeftInOldSpace() {
- return heap_->MaxOldGenerationSize() - heap_->PromotedSpaceSizeOfObjects();
+ return bytes_processed;
}
« no previous file with comments | « src/heap/incremental-marking.h ('k') | src/ia32/code-stubs-ia32.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698