Chromium Code Reviews| Index: src/heap/incremental-marking.cc |
| diff --git a/src/heap/incremental-marking.cc b/src/heap/incremental-marking.cc |
| index 0dcce20e7bd8cd1efa70a84ca2ab32ee9ec6888d..9b74b346d3ad591f80e2a7fe001921f21f540d1c 100644 |
| --- a/src/heap/incremental-marking.cc |
| +++ b/src/heap/incremental-marking.cc |
| @@ -21,26 +21,20 @@ namespace internal { |
| IncrementalMarking::IncrementalMarking(Heap* heap) |
| : heap_(heap), |
| - observer_(*this, kAllocatedThreshold), |
| state_(STOPPED), |
| - is_compacting_(false), |
| - steps_count_(0), |
| - old_generation_space_available_at_start_of_incremental_(0), |
| - old_generation_space_used_at_start_of_incremental_(0), |
| - bytes_rescanned_(0), |
| - should_hurry_(false), |
| - marking_speed_(0), |
| - bytes_scanned_(0), |
| - allocated_(0), |
| - write_barriers_invoked_since_last_step_(0), |
| + initial_old_generation_size_(0), |
| bytes_marked_ahead_of_schedule_(0), |
| - idle_marking_delay_counter_(0), |
| unscanned_bytes_of_large_object_(0), |
| + idle_marking_delay_counter_(0), |
| + incremental_marking_finalization_rounds_(0), |
| + is_compacting_(false), |
| + should_hurry_(false), |
| was_activated_(false), |
| black_allocation_(false), |
| finalize_marking_completed_(false), |
| - incremental_marking_finalization_rounds_(0), |
| - request_type_(NONE) {} |
| + request_type_(NONE), |
| + new_generation_observer_(*this, kAllocatedThreshold), |
| + old_generation_observer_(*this, kAllocatedThreshold) {} |
| bool IncrementalMarking::BaseRecordWrite(HeapObject* obj, Object* value) { |
| HeapObject* value_heap_obj = HeapObject::cast(value); |
| @@ -72,17 +66,6 @@ void IncrementalMarking::RecordWriteFromCode(HeapObject* obj, Object** slot, |
| Isolate* isolate) { |
| DCHECK(obj->IsHeapObject()); |
| IncrementalMarking* marking = isolate->heap()->incremental_marking(); |
|
Hannes Payer (out of office)
2016/09/26 14:32:05
Nice. Change to: isolate->heap()->incremental_mark
ulan
2016/09/27 17:02:08
Done.
|
| - |
| - MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address()); |
| - int counter = chunk->write_barrier_counter(); |
| - if (counter < (MemoryChunk::kWriteBarrierCounterGranularity / 2)) { |
| - marking->write_barriers_invoked_since_last_step_ += |
| - MemoryChunk::kWriteBarrierCounterGranularity - |
| - chunk->write_barrier_counter(); |
| - chunk->set_write_barrier_counter( |
| - MemoryChunk::kWriteBarrierCounterGranularity); |
| - } |
| - |
| marking->RecordWrite(obj, slot, *slot); |
| } |
| @@ -462,21 +445,6 @@ void IncrementalMarking::ActivateGeneratedStub(Code* stub) { |
| } |
| -void IncrementalMarking::NotifyOfHighPromotionRate() { |
| - if (IsMarking()) { |
| - if (marking_speed_ < kFastMarking) { |
| - if (FLAG_trace_gc) { |
| - heap()->isolate()->PrintWithTimestamp( |
| - "Increasing marking speed to %d " |
| - "due to high promotion rate\n", |
| - static_cast<int>(kFastMarking)); |
| - } |
| - marking_speed_ = kFastMarking; |
| - } |
| - } |
| -} |
| - |
| - |
| static void PatchIncrementalMarkingRecordWriteStubs( |
| Heap* heap, RecordWriteStub::Mode mode) { |
| UnseededNumberDictionary* stubs = heap->code_stubs(); |
| @@ -523,9 +491,14 @@ void IncrementalMarking::Start(GarbageCollectionReason gc_reason) { |
| HistogramTimerScope incremental_marking_scope( |
| counters->gc_incremental_marking_start()); |
| TRACE_EVENT0("v8", "V8.GCIncrementalMarkingStart"); |
| - ResetStepCounters(); |
| heap_->tracer()->NotifyIncrementalMarkingStart(); |
| + start_time_ms_ = heap()->MonotonicallyIncreasingTimeInMs(); |
| + initial_old_generation_size_ = heap_->PromotedSpaceSizeOfObjects(); |
| + old_generation_allocation_counter_ = heap_->OldGenerationAllocationCounter(); |
| + bytes_allocated_ = 0; |
| + bytes_marked_ahead_of_schedule_ = 0; |
| + should_hurry_ = false; |
| was_activated_ = true; |
| if (!heap_->mark_compact_collector()->sweeping_in_progress()) { |
| @@ -538,7 +511,11 @@ void IncrementalMarking::Start(GarbageCollectionReason gc_reason) { |
| state_ = SWEEPING; |
| } |
| - heap_->new_space()->AddAllocationObserver(&observer_); |
| + heap_->new_space()->AddAllocationObserver(&new_generation_observer_); |
|
Hannes Payer (out of office)
2016/09/26 14:32:05
Use the space iterator since we are iterating over
|
| + heap_->code_space()->AddAllocationObserver(&old_generation_observer_); |
| + heap_->lo_space()->AddAllocationObserver(&old_generation_observer_); |
| + heap_->map_space()->AddAllocationObserver(&old_generation_observer_); |
| + heap_->old_space()->AddAllocationObserver(&old_generation_observer_); |
| incremental_marking_job()->Start(heap_); |
| } |
| @@ -1004,9 +981,13 @@ void IncrementalMarking::Stop() { |
| Max(0, old_generation_size_mb - old_generation_limit_mb)); |
| } |
| - heap_->new_space()->RemoveAllocationObserver(&observer_); |
| + heap_->new_space()->RemoveAllocationObserver(&new_generation_observer_); |
| + heap_->old_space()->RemoveAllocationObserver(&old_generation_observer_); |
|
Hannes Payer (out of office)
2016/09/26 14:32:05
Use the space iterator since we are iterating over
ulan
2016/09/27 17:02:08
Done.
|
| + heap_->code_space()->RemoveAllocationObserver(&old_generation_observer_); |
| + heap_->map_space()->RemoveAllocationObserver(&old_generation_observer_); |
| + heap_->lo_space()->RemoveAllocationObserver(&old_generation_observer_); |
| + |
| IncrementalMarking::set_should_hurry(false); |
| - ResetStepCounters(); |
| if (IsMarking()) { |
| PatchIncrementalMarkingRecordWriteStubs(heap_, |
| RecordWriteStub::STORE_BUFFER_ONLY); |
| @@ -1084,82 +1065,6 @@ double IncrementalMarking::AdvanceIncrementalMarking( |
| } |
| -void IncrementalMarking::SpeedUp() { |
| - bool speed_up = false; |
| - |
| - if ((steps_count_ % kMarkingSpeedAccellerationInterval) == 0) { |
| - if (FLAG_trace_incremental_marking) { |
| - heap()->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Speed up marking after %d steps\n", |
| - static_cast<int>(kMarkingSpeedAccellerationInterval)); |
| - } |
| - speed_up = true; |
| - } |
| - |
| - bool space_left_is_very_small = |
| - (old_generation_space_available_at_start_of_incremental_ < 10 * MB); |
| - |
| - bool only_1_nth_of_space_that_was_available_still_left = |
| - (SpaceLeftInOldSpace() * (marking_speed_ + 1) < |
| - old_generation_space_available_at_start_of_incremental_); |
| - |
| - if (space_left_is_very_small || |
| - only_1_nth_of_space_that_was_available_still_left) { |
| - if (FLAG_trace_incremental_marking) |
| - heap()->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Speed up marking because of low space left\n"); |
| - speed_up = true; |
| - } |
| - |
| - bool size_of_old_space_multiplied_by_n_during_marking = |
| - (heap_->PromotedTotalSize() > |
| - (marking_speed_ + 1) * |
| - old_generation_space_used_at_start_of_incremental_); |
| - if (size_of_old_space_multiplied_by_n_during_marking) { |
| - speed_up = true; |
| - if (FLAG_trace_incremental_marking) { |
| - heap()->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Speed up marking because of heap size " |
| - "increase\n"); |
| - } |
| - } |
| - |
| - int64_t promoted_during_marking = |
| - heap_->PromotedTotalSize() - |
| - old_generation_space_used_at_start_of_incremental_; |
| - intptr_t delay = marking_speed_ * MB; |
| - intptr_t scavenge_slack = heap_->MaxSemiSpaceSize(); |
| - |
| - // We try to scan at at least twice the speed that we are allocating. |
| - if (promoted_during_marking > bytes_scanned_ / 2 + scavenge_slack + delay) { |
| - if (FLAG_trace_incremental_marking) { |
| - heap()->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Speed up marking because marker was not " |
| - "keeping up\n"); |
| - } |
| - speed_up = true; |
| - } |
| - |
| - if (speed_up) { |
| - if (state_ != MARKING) { |
| - if (FLAG_trace_incremental_marking) { |
| - heap()->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Postponing speeding up marking until marking " |
| - "starts\n"); |
| - } |
| - } else { |
| - marking_speed_ += kMarkingSpeedAccelleration; |
| - marking_speed_ = static_cast<int>( |
| - Min(kMaxMarkingSpeed, static_cast<intptr_t>(marking_speed_ * 1.3))); |
| - if (FLAG_trace_incremental_marking) { |
| - heap()->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Marking speed increased to %d\n", |
| - marking_speed_); |
| - } |
| - } |
| - } |
| -} |
| - |
| void IncrementalMarking::FinalizeSweeping() { |
| DCHECK(state_ == SWEEPING); |
| if (heap_->mark_compact_collector()->sweeping_in_progress() && |
| @@ -1168,59 +1073,80 @@ void IncrementalMarking::FinalizeSweeping() { |
| heap_->mark_compact_collector()->EnsureSweepingCompleted(); |
| } |
| if (!heap_->mark_compact_collector()->sweeping_in_progress()) { |
| - bytes_scanned_ = 0; |
| StartMarking(); |
| } |
| } |
| -void IncrementalMarking::NotifyAllocatedBytes(intptr_t allocated_bytes) { |
| +size_t IncrementalMarking::StepSizeToKeepUpWithAllocations() { |
| + // Update bytes_allocated_ based on the allocation counter. |
| + size_t current_counter = heap_->OldGenerationAllocationCounter(); |
|
Hannes Payer (out of office)
2016/09/26 14:32:05
Can we not use the same name for OldGenerationAllo
ulan
2016/09/27 17:02:08
Done.
|
| + bytes_allocated_ += current_counter - old_generation_allocation_counter_; |
| + old_generation_allocation_counter_ = current_counter; |
| + return bytes_allocated_; |
| +} |
| + |
| +size_t IncrementalMarking::StepSizeToMakeProgress() { |
| + // We increase step size gradually based on the time passed in order to |
| + // leave marking work to standalone tasks. The ramp up duration and the |
| + // target step count are chosen based on benchmarks. |
| + const int kRampUpIntervalMs = 300; |
| + const size_t kTargetStepCount = 128; |
| + size_t step_size = Max(initial_old_generation_size_ / kTargetStepCount, |
| + IncrementalMarking::kAllocatedThreshold); |
| + double time_passed_ms = |
| + heap_->MonotonicallyIncreasingTimeInMs() - start_time_ms_; |
| + double factor = Min(time_passed_ms / kRampUpIntervalMs, 1.0); |
| + return static_cast<size_t>(factor * step_size); |
| +} |
| + |
| +void IncrementalMarking::AdvanceIncrementalMarkingOnAllocation() { |
| if (heap_->gc_state() != Heap::NOT_IN_GC || !FLAG_incremental_marking || |
| (state_ != SWEEPING && state_ != MARKING)) { |
| return; |
| } |
| - allocated_ += allocated_bytes; |
| - |
| - if (allocated_ >= kAllocatedThreshold || |
| - write_barriers_invoked_since_last_step_ >= |
| - kWriteBarriersInvokedThreshold) { |
| - // The marking speed is driven either by the allocation rate or by the rate |
| - // at which we are having to check the color of objects in the write |
| - // barrier. |
| - // It is possible for a tight non-allocating loop to run a lot of write |
| - // barriers before we get here and check them (marking can only take place |
| - // on |
| - // allocation), so to reduce the lumpiness we don't use the write barriers |
| - // invoked since last step directly to determine the amount of work to do. |
| - intptr_t bytes_to_process = |
| - marking_speed_ * |
| - Max(allocated_, write_barriers_invoked_since_last_step_); |
| - Step(bytes_to_process, GC_VIA_STACK_GUARD, FORCE_COMPLETION, |
| - StepOrigin::kV8); |
| + size_t bytes_to_process = |
| + StepSizeToKeepUpWithAllocations() + StepSizeToMakeProgress(); |
| + |
| + if (bytes_to_process >= IncrementalMarking::kAllocatedThreshold) { |
| + // The first step after Scavenge will see many allocated bytes. |
| + // Cap the step size to distribute the marking work more uniformly. |
| + size_t max_step_size = GCIdleTimeHandler::EstimateMarkingStepSize( |
| + kMaxStepSizeInMs, |
| + heap()->tracer()->IncrementalMarkingSpeedInBytesPerMillisecond()); |
| + bytes_to_process = Min(bytes_to_process, max_step_size); |
|
Hannes Payer (out of office)
2016/09/26 14:32:05
Can we think of a way to avoid this big step after
ulan
2016/09/27 17:02:08
Acknowledged.
|
| + |
| + intptr_t bytes_processed = 0; |
| + if (bytes_marked_ahead_of_schedule_ >= bytes_to_process) { |
| + // Steps performed in tasks have put us ahead of schedule. |
| + // We skip processing of marking dequeue here and thus |
| + // shift marking time from inside V8 to standalone tasks. |
| + bytes_marked_ahead_of_schedule_ -= bytes_to_process; |
| + bytes_processed = bytes_to_process; |
| + } else { |
| + bytes_processed = Step(bytes_to_process, GC_VIA_STACK_GUARD, |
| + FORCE_COMPLETION, StepOrigin::kV8); |
| + } |
| + bytes_allocated_ -= Min(bytes_allocated_, bytes_to_process); |
| } |
| } |
| -void IncrementalMarking::Step(intptr_t bytes_to_process, |
| - CompletionAction action, |
| - ForceCompletionAction completion, |
| - StepOrigin step_origin) { |
| +size_t IncrementalMarking::Step(size_t bytes_to_process, |
| + CompletionAction action, |
| + ForceCompletionAction completion, |
| + StepOrigin step_origin) { |
| HistogramTimerScope incremental_marking_scope( |
| heap_->isolate()->counters()->gc_incremental_marking()); |
| TRACE_EVENT0("v8", "V8.GCIncrementalMarking"); |
| TRACE_GC(heap_->tracer(), GCTracer::Scope::MC_INCREMENTAL); |
| double start = heap_->MonotonicallyIncreasingTimeInMs(); |
| - bytes_scanned_ += bytes_to_process; |
| - |
| - allocated_ = 0; |
| - write_barriers_invoked_since_last_step_ = 0; |
| - |
| if (state_ == SWEEPING) { |
| TRACE_GC(heap_->tracer(), GCTracer::Scope::MC_INCREMENTAL_SWEEPING); |
| FinalizeSweeping(); |
| } |
| - intptr_t bytes_processed = 0; |
| + size_t bytes_processed = 0; |
| if (state_ == MARKING) { |
| const bool incremental_wrapper_tracing = |
| FLAG_incremental_marking_wrappers && heap_->UsingEmbedderHeapTracer(); |
| @@ -1231,17 +1157,9 @@ void IncrementalMarking::Step(intptr_t bytes_to_process, |
| heap_->mark_compact_collector()->marking_deque()->IsEmpty()); |
| bool wrapper_work_left = incremental_wrapper_tracing; |
| if (!process_wrappers) { |
| - if (step_origin == StepOrigin::kV8 && |
| - bytes_marked_ahead_of_schedule_ >= bytes_to_process) { |
| - // Steps performed in tasks have put us ahead of schedule. |
| - // We skip processing of marking dequeue here and thus |
| - // shift marking time from inside V8 to standalone tasks. |
| - bytes_marked_ahead_of_schedule_ -= bytes_to_process; |
| - } else { |
| - bytes_processed = ProcessMarkingDeque(bytes_to_process); |
| - if (step_origin == StepOrigin::kTask) { |
| - bytes_marked_ahead_of_schedule_ += bytes_processed; |
| - } |
| + bytes_processed = ProcessMarkingDeque(bytes_to_process); |
| + if (step_origin == StepOrigin::kTask) { |
| + bytes_marked_ahead_of_schedule_ += bytes_processed; |
| } |
| } else { |
| const double wrapper_deadline = |
| @@ -1273,12 +1191,6 @@ void IncrementalMarking::Step(intptr_t bytes_to_process, |
| } |
| } |
| - steps_count_++; |
| - |
| - // Speed up marking if we are marking too slow or if we are almost done |
| - // with marking. |
| - SpeedUp(); |
| - |
| double end = heap_->MonotonicallyIncreasingTimeInMs(); |
| double duration = (end - start); |
| // Note that we report zero bytes here when sweeping was in progress or |
| @@ -1287,30 +1199,11 @@ void IncrementalMarking::Step(intptr_t bytes_to_process, |
| heap_->tracer()->AddIncrementalMarkingStep(duration, bytes_processed); |
| if (FLAG_trace_incremental_marking) { |
| heap_->isolate()->PrintWithTimestamp( |
| - "[IncrementalMarking] Step %s %d bytes (%d) in %.1f\n", |
| - step_origin == StepOrigin::kV8 ? "in v8" : "in task", |
| - static_cast<int>(bytes_processed), static_cast<int>(bytes_to_process), |
| - duration); |
| + "[IncrementalMarking] Step %s %zu bytes (%zu) in %.1f\n", |
| + step_origin == StepOrigin::kV8 ? "in v8" : "in task", bytes_processed, |
| + bytes_to_process, duration); |
| } |
| -} |
| - |
| - |
| -void IncrementalMarking::ResetStepCounters() { |
| - steps_count_ = 0; |
| - old_generation_space_available_at_start_of_incremental_ = |
| - SpaceLeftInOldSpace(); |
| - old_generation_space_used_at_start_of_incremental_ = |
| - heap_->PromotedTotalSize(); |
| - bytes_rescanned_ = 0; |
| - marking_speed_ = kInitialMarkingSpeed; |
| - bytes_scanned_ = 0; |
| - write_barriers_invoked_since_last_step_ = 0; |
| - bytes_marked_ahead_of_schedule_ = 0; |
| -} |
| - |
| - |
| -int64_t IncrementalMarking::SpaceLeftInOldSpace() { |
| - return heap_->MaxOldGenerationSize() - heap_->PromotedSpaceSizeOfObjects(); |
| + return bytes_processed; |
| } |