Chromium Code Reviews| Index: src/heap/spaces.cc |
| diff --git a/src/heap/spaces.cc b/src/heap/spaces.cc |
| index 23323a4a7daa224fd8f862ec9e4d45b7aa87fa66..15f6c1908b9cbbf29608d6248036ee5744596d24 100644 |
| --- a/src/heap/spaces.cc |
| +++ b/src/heap/spaces.cc |
| @@ -2200,6 +2200,7 @@ void FreeList::Reset() { |
| medium_list_.Reset(); |
| large_list_.Reset(); |
| huge_list_.Reset(); |
| + unreported_allocation_ = 0; |
| } |
| @@ -2347,6 +2348,15 @@ FreeSpace* FreeList::FindNodeFor(int size_in_bytes, int* node_size) { |
| } |
| +void PagedSpace::SetTopAndLimit(Address top, Address limit) { |
| + DCHECK(top == limit || |
| + Page::FromAddress(top) == Page::FromAddress(limit - 1)); |
| + MemoryChunk::UpdateHighWaterMark(allocation_info_.top()); |
| + allocation_info_.set_top(top); |
| + allocation_info_.set_limit(limit); |
| +} |
| + |
| + |
| // Allocation on the old space free list. If it succeeds then a new linear |
| // allocation space has been set up with the top and limit of the space. If |
| // the allocation fails then NULL is returned, and the caller can perform a GC |
| @@ -2364,9 +2374,6 @@ HeapObject* FreeList::Allocate(int size_in_bytes) { |
| // if it is big enough. |
| owner_->Free(owner_->top(), old_linear_size); |
| - owner_->heap()->incremental_marking()->OldSpaceStep(size_in_bytes - |
| - old_linear_size); |
| - |
| int new_node_size = 0; |
| FreeSpace* new_node = FindNodeFor(size_in_bytes, &new_node_size); |
| if (new_node == NULL) { |
| @@ -2389,21 +2396,27 @@ HeapObject* FreeList::Allocate(int size_in_bytes) { |
| // candidate. |
| DCHECK(!MarkCompactCollector::IsOnEvacuationCandidate(new_node)); |
| - const int kThreshold = IncrementalMarking::kAllocatedThreshold; |
| + // An old-space step will mark more data per byte allocated, because old space |
| + // allocation is more serious. We don't want the pause to be bigger, so we |
| + // do marking after a smaller amount of allocation. |
| + const int kThreshold = IncrementalMarking::kAllocatedThreshold * |
| + IncrementalMarking::kOldSpaceAllocationMarkingFactor; |
| // Memory in the linear allocation area is counted as allocated. We may free |
| // a little of this again immediately - see below. |
| owner_->Allocate(new_node_size); |
| + unreported_allocation_ += new_node_size; |
| + |
| if (owner_->heap()->inline_allocation_disabled()) { |
| // Keep the linear allocation area empty if requested to do so, just |
| // return area back to the free list instead. |
| owner_->Free(new_node->address() + size_in_bytes, bytes_left); |
| DCHECK(owner_->top() == NULL && owner_->limit() == NULL); |
| } else if (bytes_left > kThreshold && |
| - owner_->heap()->incremental_marking()->IsMarkingIncomplete() && |
| - FLAG_incremental_marking_steps) { |
| + owner_->heap()->incremental_marking()->CanDoSteps()) { |
| int linear_size = owner_->RoundSizeDownToObjectAlignment(kThreshold); |
| + |
| // We don't want to give too large linear areas to the allocator while |
| // incremental marking is going on, because we won't check again whether |
| // we want to do another increment until the linear area is used up. |
| @@ -2411,15 +2424,30 @@ HeapObject* FreeList::Allocate(int size_in_bytes) { |
| new_node_size - size_in_bytes - linear_size); |
| owner_->SetTopAndLimit(new_node->address() + size_in_bytes, |
| new_node->address() + size_in_bytes + linear_size); |
| - } else if (bytes_left > 0) { |
| - // Normally we give the rest of the node to the allocator as its new |
| - // linear allocation area. |
| - owner_->SetTopAndLimit(new_node->address() + size_in_bytes, |
| - new_node->address() + new_node_size); |
| + // It is important that we are done updating top and limit before we call |
| + // this. |
|
Hannes Payer (out of office)
2015/04/22 14:06:21
Can you say way in the comment to make it clear?
Erik Corry Chromium.org
2015/06/29 14:45:30
Done.
|
| + owner_->heap()->incremental_marking()->OldSpaceStep(size_in_bytes + |
| + linear_size); |
| + unreported_allocation_ = 0; |
| } else { |
|
Hannes Payer (out of office)
2015/04/22 14:06:21
Please add the same comment from above also here.
Erik Corry Chromium.org
2015/06/29 14:45:30
The comment is there, just before the call to OldS
|
| - // TODO(gc) Try not freeing linear allocation region when bytes_left |
| - // are zero. |
| - owner_->SetTopAndLimit(NULL, NULL); |
| + if (bytes_left > 0) { |
| + // Normally we give the rest of the node to the allocator as its new |
| + // linear allocation area. |
| + owner_->SetTopAndLimit(new_node->address() + size_in_bytes, |
| + new_node->address() + new_node_size); |
| + } else { |
| + // TODO(gc) Try not freeing linear allocation region when bytes_left |
| + // are zero. |
| + owner_->SetTopAndLimit(NULL, NULL); |
| + } |
| + if (unreported_allocation_ > kThreshold) { |
| + // This may start the incremental marker, or do a little work if it's |
| + // already started. It is important that we are finished updating top |
| + // and limit before we call this. |
| + owner_->heap()->incremental_marking()->OldSpaceStep( |
| + Min(kThreshold, unreported_allocation_)); |
| + unreported_allocation_ = 0; |
| + } |
| } |
| return new_node; |
| @@ -2906,7 +2934,16 @@ AllocationResult LargeObjectSpace::AllocateRaw(int object_size, |
| reinterpret_cast<Object**>(object->address())[1] = Smi::FromInt(0); |
| } |
| - heap()->incremental_marking()->OldSpaceStep(object_size); |
| + // We would like to tell the incremental marker to do a lot of work, since |
| + // we just made a large allocation in old space, but that might cause a huge |
| + // pause. Underreporting here may cause the marker to speed up because it |
| + // will perceive that it is not keeping up with allocation. Although this |
| + // causes some big incremental marking steps they are not as big as this one |
| + // might have been. In testing, a very large pause was divided up into about |
| + // 12 parts. |
| + const int kThreshold = IncrementalMarking::kAllocatedThreshold * |
| + IncrementalMarking::kOldSpaceAllocationMarkingFactor; |
| + heap()->incremental_marking()->OldSpaceStep(kThreshold); |
| return object; |
| } |