Index: src/incremental-marking.cc |
diff --git a/src/incremental-marking.cc b/src/incremental-marking.cc |
index d3df705c2169ea81a715c4b212e961eb50abdf84..4150b19f94f13a1a77351e23f5e75e02ba9503aa 100644 |
--- a/src/incremental-marking.cc |
+++ b/src/incremental-marking.cc |
@@ -853,104 +853,112 @@ void IncrementalMarking::Step(intptr_t allocated_bytes, |
if (state_ == MARKING && no_marking_scope_depth_ > 0) return; |
- // The marking speed is driven either by the allocation rate or by the rate |
- // at which we are having to check the color of objects in the write barrier. |
- // It is possible for a tight non-allocating loop to run a lot of write |
- // barriers before we get here and check them (marking can only take place on |
- // allocation), so to reduce the lumpiness we don't use the write barriers |
- // invoked since last step directly to determine the amount of work to do. |
- intptr_t bytes_to_process = |
- marking_speed_ * Max(allocated_, write_barriers_invoked_since_last_step_); |
- allocated_ = 0; |
- write_barriers_invoked_since_last_step_ = 0; |
- |
- bytes_scanned_ += bytes_to_process; |
- |
- double start = base::OS::TimeCurrentMillis(); |
- |
- if (state_ == SWEEPING) { |
- if (heap_->mark_compact_collector()->sweeping_in_progress() && |
- heap_->mark_compact_collector()->IsSweepingCompleted()) { |
- heap_->mark_compact_collector()->EnsureSweepingCompleted(); |
- } |
- if (!heap_->mark_compact_collector()->sweeping_in_progress()) { |
- bytes_scanned_ = 0; |
- StartMarking(PREVENT_COMPACTION); |
+ { |
+ HistogramTimerScope incremental_marking_scope( |
+ heap_->isolate()->counters()->gc_incremental_marking()); |
+ double start = base::OS::TimeCurrentMillis(); |
+ |
+ // The marking speed is driven either by the allocation rate or by the rate |
+ // at which we are having to check the color of objects in the write |
+ // barrier. |
+ // It is possible for a tight non-allocating loop to run a lot of write |
+ // barriers before we get here and check them (marking can only take place |
+ // on |
+ // allocation), so to reduce the lumpiness we don't use the write barriers |
+ // invoked since last step directly to determine the amount of work to do. |
+ intptr_t bytes_to_process = |
+ marking_speed_ * |
+ Max(allocated_, write_barriers_invoked_since_last_step_); |
+ allocated_ = 0; |
+ write_barriers_invoked_since_last_step_ = 0; |
+ |
+ bytes_scanned_ += bytes_to_process; |
+ |
+ if (state_ == SWEEPING) { |
+ if (heap_->mark_compact_collector()->sweeping_in_progress() && |
+ heap_->mark_compact_collector()->IsSweepingCompleted()) { |
+ heap_->mark_compact_collector()->EnsureSweepingCompleted(); |
+ } |
+ if (!heap_->mark_compact_collector()->sweeping_in_progress()) { |
+ bytes_scanned_ = 0; |
+ StartMarking(PREVENT_COMPACTION); |
+ } |
+ } else if (state_ == MARKING) { |
+ ProcessMarkingDeque(bytes_to_process); |
+ if (marking_deque_.IsEmpty()) MarkingComplete(action); |
} |
- } else if (state_ == MARKING) { |
- ProcessMarkingDeque(bytes_to_process); |
- if (marking_deque_.IsEmpty()) MarkingComplete(action); |
- } |
- steps_count_++; |
+ steps_count_++; |
- bool speed_up = false; |
+ bool speed_up = false; |
- if ((steps_count_ % kMarkingSpeedAccellerationInterval) == 0) { |
- if (FLAG_trace_gc) { |
- PrintPID("Speed up marking after %d steps\n", |
- static_cast<int>(kMarkingSpeedAccellerationInterval)); |
+ if ((steps_count_ % kMarkingSpeedAccellerationInterval) == 0) { |
+ if (FLAG_trace_gc) { |
+ PrintPID("Speed up marking after %d steps\n", |
+ static_cast<int>(kMarkingSpeedAccellerationInterval)); |
+ } |
+ speed_up = true; |
} |
- speed_up = true; |
- } |
- bool space_left_is_very_small = |
- (old_generation_space_available_at_start_of_incremental_ < 10 * MB); |
+ bool space_left_is_very_small = |
+ (old_generation_space_available_at_start_of_incremental_ < 10 * MB); |
- bool only_1_nth_of_space_that_was_available_still_left = |
- (SpaceLeftInOldSpace() * (marking_speed_ + 1) < |
- old_generation_space_available_at_start_of_incremental_); |
+ bool only_1_nth_of_space_that_was_available_still_left = |
+ (SpaceLeftInOldSpace() * (marking_speed_ + 1) < |
+ old_generation_space_available_at_start_of_incremental_); |
- if (space_left_is_very_small || |
- only_1_nth_of_space_that_was_available_still_left) { |
- if (FLAG_trace_gc) PrintPID("Speed up marking because of low space left\n"); |
- speed_up = true; |
- } |
- |
- bool size_of_old_space_multiplied_by_n_during_marking = |
- (heap_->PromotedTotalSize() > |
- (marking_speed_ + 1) * |
- old_generation_space_used_at_start_of_incremental_); |
- if (size_of_old_space_multiplied_by_n_during_marking) { |
- speed_up = true; |
- if (FLAG_trace_gc) { |
- PrintPID("Speed up marking because of heap size increase\n"); |
+ if (space_left_is_very_small || |
+ only_1_nth_of_space_that_was_available_still_left) { |
+ if (FLAG_trace_gc) |
+ PrintPID("Speed up marking because of low space left\n"); |
+ speed_up = true; |
} |
- } |
- int64_t promoted_during_marking = heap_->PromotedTotalSize() |
- - old_generation_space_used_at_start_of_incremental_; |
- intptr_t delay = marking_speed_ * MB; |
- intptr_t scavenge_slack = heap_->MaxSemiSpaceSize(); |
- |
- // We try to scan at at least twice the speed that we are allocating. |
- if (promoted_during_marking > bytes_scanned_ / 2 + scavenge_slack + delay) { |
- if (FLAG_trace_gc) { |
- PrintPID("Speed up marking because marker was not keeping up\n"); |
+ bool size_of_old_space_multiplied_by_n_during_marking = |
+ (heap_->PromotedTotalSize() > |
+ (marking_speed_ + 1) * |
+ old_generation_space_used_at_start_of_incremental_); |
+ if (size_of_old_space_multiplied_by_n_during_marking) { |
+ speed_up = true; |
+ if (FLAG_trace_gc) { |
+ PrintPID("Speed up marking because of heap size increase\n"); |
+ } |
} |
- speed_up = true; |
- } |
- if (speed_up) { |
- if (state_ != MARKING) { |
+ int64_t promoted_during_marking = |
+ heap_->PromotedTotalSize() - |
+ old_generation_space_used_at_start_of_incremental_; |
+ intptr_t delay = marking_speed_ * MB; |
+ intptr_t scavenge_slack = heap_->MaxSemiSpaceSize(); |
+ |
+ // We try to scan at at least twice the speed that we are allocating. |
+ if (promoted_during_marking > bytes_scanned_ / 2 + scavenge_slack + delay) { |
if (FLAG_trace_gc) { |
- PrintPID("Postponing speeding up marking until marking starts\n"); |
+ PrintPID("Speed up marking because marker was not keeping up\n"); |
} |
- } else { |
- marking_speed_ += kMarkingSpeedAccelleration; |
- marking_speed_ = static_cast<int>( |
- Min(kMaxMarkingSpeed, |
- static_cast<intptr_t>(marking_speed_ * 1.3))); |
- if (FLAG_trace_gc) { |
- PrintPID("Marking speed increased to %d\n", marking_speed_); |
+ speed_up = true; |
+ } |
+ |
+ if (speed_up) { |
+ if (state_ != MARKING) { |
+ if (FLAG_trace_gc) { |
+ PrintPID("Postponing speeding up marking until marking starts\n"); |
+ } |
+ } else { |
+ marking_speed_ += kMarkingSpeedAccelleration; |
+ marking_speed_ = static_cast<int>( |
+ Min(kMaxMarkingSpeed, static_cast<intptr_t>(marking_speed_ * 1.3))); |
+ if (FLAG_trace_gc) { |
+ PrintPID("Marking speed increased to %d\n", marking_speed_); |
+ } |
} |
} |
- } |
- double end = base::OS::TimeCurrentMillis(); |
- double delta = (end - start); |
- heap_->tracer()->AddIncrementalMarkingStep(delta); |
- heap_->AddMarkingTime(delta); |
+ double end = base::OS::TimeCurrentMillis(); |
+ double delta = (end - start); |
+ heap_->tracer()->AddIncrementalMarkingStep(delta); |
+ heap_->AddMarkingTime(delta); |
+ } |
} |