| Index: src/heap/incremental-marking.cc
|
| diff --git a/src/heap/incremental-marking.cc b/src/heap/incremental-marking.cc
|
| index f585387dc3fa84d5601a39f02e91c868a808c018..57ed020228ab600222bdb25ca05018cec545462a 100644
|
| --- a/src/heap/incremental-marking.cc
|
| +++ b/src/heap/incremental-marking.cc
|
| @@ -30,13 +30,7 @@ IncrementalMarking::IncrementalMarking(Heap* heap)
|
| observer_(*this, kAllocatedThreshold),
|
| state_(STOPPED),
|
| is_compacting_(false),
|
| - steps_count_(0),
|
| - old_generation_space_available_at_start_of_incremental_(0),
|
| - old_generation_space_used_at_start_of_incremental_(0),
|
| - bytes_rescanned_(0),
|
| should_hurry_(false),
|
| - marking_speed_(0),
|
| - bytes_scanned_(0),
|
| allocated_(0),
|
| write_barriers_invoked_since_last_step_(0),
|
| idle_marking_delay_counter_(0),
|
| @@ -81,9 +75,11 @@ void IncrementalMarking::RecordWriteFromCode(HeapObject* obj, Object** slot,
|
| MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address());
|
| int counter = chunk->write_barrier_counter();
|
| if (counter < (MemoryChunk::kWriteBarrierCounterGranularity / 2)) {
|
| - marking->write_barriers_invoked_since_last_step_ +=
|
| - MemoryChunk::kWriteBarrierCounterGranularity -
|
| - chunk->write_barrier_counter();
|
| + marking->write_barriers_invoked_since_last_step_ =
|
| + Min(kMaxWriteBarrierCounter,
|
| + marking->write_barriers_invoked_since_last_step_ +
|
| + MemoryChunk::kWriteBarrierCounterGranularity -
|
| + chunk->write_barrier_counter());
|
| chunk->set_write_barrier_counter(
|
| MemoryChunk::kWriteBarrierCounterGranularity);
|
| }
|
| @@ -467,21 +463,6 @@ void IncrementalMarking::ActivateGeneratedStub(Code* stub) {
|
| }
|
|
|
|
|
| -void IncrementalMarking::NotifyOfHighPromotionRate() {
|
| - if (IsMarking()) {
|
| - if (marking_speed_ < kFastMarking) {
|
| - if (FLAG_trace_gc) {
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "Increasing marking speed to %d "
|
| - "due to high promotion rate\n",
|
| - static_cast<int>(kFastMarking));
|
| - }
|
| - marking_speed_ = kFastMarking;
|
| - }
|
| - }
|
| -}
|
| -
|
| -
|
| static void PatchIncrementalMarkingRecordWriteStubs(
|
| Heap* heap, RecordWriteStub::Mode mode) {
|
| UnseededNumberDictionary* stubs = heap->code_stubs();
|
| @@ -1069,87 +1050,11 @@ void IncrementalMarking::OldSpaceStep(intptr_t allocated) {
|
| heap()->StartIncrementalMarking(Heap::kNoGCFlags, kNoGCCallbackFlags,
|
| "old space step");
|
| } else {
|
| - Step(allocated * kFastMarking / kInitialMarkingSpeed, GC_VIA_STACK_GUARD);
|
| + Step(allocated, GC_VIA_STACK_GUARD);
|
| }
|
| }
|
|
|
|
|
| -void IncrementalMarking::SpeedUp() {
|
| - bool speed_up = false;
|
| -
|
| - if ((steps_count_ % kMarkingSpeedAccellerationInterval) == 0) {
|
| - if (FLAG_trace_incremental_marking) {
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "[IncrementalMarking] Speed up marking after %d steps\n",
|
| - static_cast<int>(kMarkingSpeedAccellerationInterval));
|
| - }
|
| - speed_up = true;
|
| - }
|
| -
|
| - bool space_left_is_very_small =
|
| - (old_generation_space_available_at_start_of_incremental_ < 10 * MB);
|
| -
|
| - bool only_1_nth_of_space_that_was_available_still_left =
|
| - (SpaceLeftInOldSpace() * (marking_speed_ + 1) <
|
| - old_generation_space_available_at_start_of_incremental_);
|
| -
|
| - if (space_left_is_very_small ||
|
| - only_1_nth_of_space_that_was_available_still_left) {
|
| - if (FLAG_trace_incremental_marking)
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "[IncrementalMarking] Speed up marking because of low space left\n");
|
| - speed_up = true;
|
| - }
|
| -
|
| - bool size_of_old_space_multiplied_by_n_during_marking =
|
| - (heap_->PromotedTotalSize() >
|
| - (marking_speed_ + 1) *
|
| - old_generation_space_used_at_start_of_incremental_);
|
| - if (size_of_old_space_multiplied_by_n_during_marking) {
|
| - speed_up = true;
|
| - if (FLAG_trace_incremental_marking) {
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "[IncrementalMarking] Speed up marking because of heap size "
|
| - "increase\n");
|
| - }
|
| - }
|
| -
|
| - int64_t promoted_during_marking =
|
| - heap_->PromotedTotalSize() -
|
| - old_generation_space_used_at_start_of_incremental_;
|
| - intptr_t delay = marking_speed_ * MB;
|
| - intptr_t scavenge_slack = heap_->MaxSemiSpaceSize();
|
| -
|
| - // We try to scan at at least twice the speed that we are allocating.
|
| - if (promoted_during_marking > bytes_scanned_ / 2 + scavenge_slack + delay) {
|
| - if (FLAG_trace_incremental_marking) {
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "[IncrementalMarking] Speed up marking because marker was not "
|
| - "keeping up\n");
|
| - }
|
| - speed_up = true;
|
| - }
|
| -
|
| - if (speed_up) {
|
| - if (state_ != MARKING) {
|
| - if (FLAG_trace_incremental_marking) {
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "[IncrementalMarking] Postponing speeding up marking until marking "
|
| - "starts\n");
|
| - }
|
| - } else {
|
| - marking_speed_ += kMarkingSpeedAccelleration;
|
| - marking_speed_ = static_cast<int>(
|
| - Min(kMaxMarkingSpeed, static_cast<intptr_t>(marking_speed_ * 1.3)));
|
| - if (FLAG_trace_incremental_marking) {
|
| - heap()->isolate()->PrintWithTimestamp(
|
| - "[IncrementalMarking] Marking speed increased to %d\n",
|
| - marking_speed_);
|
| - }
|
| - }
|
| - }
|
| -}
|
| -
|
| void IncrementalMarking::FinalizeSweeping() {
|
| DCHECK(state_ == SWEEPING);
|
| if (heap_->mark_compact_collector()->sweeping_in_progress() &&
|
| @@ -1158,7 +1063,6 @@ void IncrementalMarking::FinalizeSweeping() {
|
| heap_->mark_compact_collector()->EnsureSweepingCompleted();
|
| }
|
| if (!heap_->mark_compact_collector()->sweeping_in_progress()) {
|
| - bytes_scanned_ = 0;
|
| StartMarking();
|
| }
|
| }
|
| @@ -1196,22 +1100,24 @@ intptr_t IncrementalMarking::Step(intptr_t allocated_bytes,
|
| TRACE_GC(heap_->tracer(), GCTracer::Scope::MC_INCREMENTAL);
|
| double start = heap_->MonotonicallyIncreasingTimeInMs();
|
|
|
| + // Make sure that the step size is large enough to justify the overhead
|
| + // of interrupting the generated code to perform the step.
|
| + intptr_t min_bytes_to_process = GCIdleTimeHandler::EstimateMarkingStepSize(
|
| + kMinIncrementalStepDurationInMs,
|
| + heap()->tracer()->IncrementalMarkingSpeedInBytesPerMillisecond());
|
| // The marking speed is driven either by the allocation rate or by the rate
|
| // at which we are having to check the color of objects in the write
|
| // barrier.
|
| // It is possible for a tight non-allocating loop to run a lot of write
|
| // barriers before we get here and check them (marking can only take place
|
| - // on
|
| - // allocation), so to reduce the lumpiness we don't use the write barriers
|
| - // invoked since last step directly to determine the amount of work to do.
|
| - intptr_t bytes_to_process =
|
| - marking_speed_ *
|
| - Max(allocated_, write_barriers_invoked_since_last_step_);
|
| + // on allocation).
|
| + intptr_t bytes_to_process = Max(
|
| + min_bytes_to_process, kBytesToMarkPerAllocatedByte * allocated_ +
|
| + kBytesToMarkPerWriteBarrier *
|
| + write_barriers_invoked_since_last_step_);
|
| allocated_ = 0;
|
| write_barriers_invoked_since_last_step_ = 0;
|
|
|
| - bytes_scanned_ += bytes_to_process;
|
| -
|
| if (state_ == SWEEPING) {
|
| TRACE_GC(heap_->tracer(), GCTracer::Scope::MC_INCREMENTAL_SWEEPING);
|
| FinalizeSweeping();
|
| @@ -1260,12 +1166,6 @@ intptr_t IncrementalMarking::Step(intptr_t allocated_bytes,
|
| }
|
| }
|
|
|
| - steps_count_++;
|
| -
|
| - // Speed up marking if we are marking too slow or if we are almost done
|
| - // with marking.
|
| - SpeedUp();
|
| -
|
| double end = heap_->MonotonicallyIncreasingTimeInMs();
|
| double duration = (end - start);
|
| // Note that we report zero bytes here when sweeping was in progress or
|
| @@ -1278,14 +1178,7 @@ intptr_t IncrementalMarking::Step(intptr_t allocated_bytes,
|
|
|
|
|
| void IncrementalMarking::ResetStepCounters() {
|
| - steps_count_ = 0;
|
| - old_generation_space_available_at_start_of_incremental_ =
|
| - SpaceLeftInOldSpace();
|
| - old_generation_space_used_at_start_of_incremental_ =
|
| - heap_->PromotedTotalSize();
|
| - bytes_rescanned_ = 0;
|
| - marking_speed_ = kInitialMarkingSpeed;
|
| - bytes_scanned_ = 0;
|
| + allocated_ = 0;
|
| write_barriers_invoked_since_last_step_ = 0;
|
| }
|
|
|
|
|