Index: src/heap/spaces.cc |
diff --git a/src/heap/spaces.cc b/src/heap/spaces.cc |
index 7a411e52e56ff9cc3ced656c6e02a2c30847da9e..0806b2565da68f802c9b06041efd8fe9858312dc 100644 |
--- a/src/heap/spaces.cc |
+++ b/src/heap/spaces.cc |
@@ -2213,7 +2213,6 @@ |
medium_list_.Reset(); |
large_list_.Reset(); |
huge_list_.Reset(); |
- unreported_allocation_ = 0; |
} |
@@ -2361,22 +2360,6 @@ |
} |
-void PagedSpace::SetTopAndLimit(Address top, Address limit) { |
- DCHECK(top == limit || |
- Page::FromAddress(top) == Page::FromAddress(limit - 1)); |
- MemoryChunk::UpdateHighWaterMark(allocation_info_.top()); |
- allocation_info_.set_top(top); |
- allocation_info_.set_limit(limit); |
-} |
- |
- |
-void PagedSpace::ReturnLinearAllocationAreaToFreeList() { |
- int old_linear_size = static_cast<int>(limit() - top()); |
- Free(top(), old_linear_size); |
- SetTopAndLimit(NULL, NULL); |
-} |
- |
- |
// Allocation on the old space free list. If it succeeds then a new linear |
// allocation space has been set up with the top and limit of the space. If |
// the allocation fails then NULL is returned, and the caller can perform a GC |
@@ -2394,6 +2377,9 @@ |
// if it is big enough. |
owner_->Free(owner_->top(), old_linear_size); |
+ owner_->heap()->incremental_marking()->OldSpaceStep(size_in_bytes - |
+ old_linear_size); |
+ |
int new_node_size = 0; |
FreeSpace* new_node = FindNodeFor(size_in_bytes, &new_node_size); |
if (new_node == NULL) { |
@@ -2416,17 +2402,11 @@ |
// candidate. |
DCHECK(!MarkCompactCollector::IsOnEvacuationCandidate(new_node)); |
- // An old-space step will mark more data per byte allocated, because old space |
- // allocation is more serious. We don't want the pause to be bigger, so we |
- // do marking after a smaller amount of allocation. |
- const int kThreshold = IncrementalMarking::kAllocatedThreshold * |
- IncrementalMarking::kOldSpaceAllocationMarkingFactor; |
+ const int kThreshold = IncrementalMarking::kAllocatedThreshold; |
// Memory in the linear allocation area is counted as allocated. We may free |
// a little of this again immediately - see below. |
owner_->Allocate(new_node_size); |
- |
- unreported_allocation_ += new_node_size; |
if (owner_->heap()->inline_allocation_disabled()) { |
// Keep the linear allocation area empty if requested to do so, just |
@@ -2434,9 +2414,9 @@ |
owner_->Free(new_node->address() + size_in_bytes, bytes_left); |
DCHECK(owner_->top() == NULL && owner_->limit() == NULL); |
} else if (bytes_left > kThreshold && |
- owner_->heap()->incremental_marking()->CanDoSteps()) { |
+ owner_->heap()->incremental_marking()->IsMarkingIncomplete() && |
+ FLAG_incremental_marking_steps) { |
int linear_size = owner_->RoundSizeDownToObjectAlignment(kThreshold); |
- |
// We don't want to give too large linear areas to the allocator while |
// incremental marking is going on, because we won't check again whether |
// we want to do another increment until the linear area is used up. |
@@ -2444,32 +2424,15 @@ |
new_node_size - size_in_bytes - linear_size); |
owner_->SetTopAndLimit(new_node->address() + size_in_bytes, |
new_node->address() + size_in_bytes + linear_size); |
- // It is important that we are done updating top and limit before we call |
- // this, because it might add the free space between top and limit to the |
- // free list, and that would be very bad if top and new_node were still |
- // pointing to the same place. |
- owner_->heap()->incremental_marking()->OldSpaceStep(size_in_bytes + |
- linear_size); |
- unreported_allocation_ = 0; |
+ } else if (bytes_left > 0) { |
+ // Normally we give the rest of the node to the allocator as its new |
+ // linear allocation area. |
+ owner_->SetTopAndLimit(new_node->address() + size_in_bytes, |
+ new_node->address() + new_node_size); |
} else { |
- if (bytes_left > 0) { |
- // Normally we give the rest of the node to the allocator as its new |
- // linear allocation area. |
- owner_->SetTopAndLimit(new_node->address() + size_in_bytes, |
- new_node->address() + new_node_size); |
- } else { |
- // TODO(gc) Try not freeing linear allocation region when bytes_left |
- // are zero. |
- owner_->SetTopAndLimit(NULL, NULL); |
- } |
- if (unreported_allocation_ > kThreshold) { |
- // This may start the incremental marker, or do a little work if it's |
- // already started. It is important that we are finished updating top |
- // and limit before we call this (see above). |
- owner_->heap()->incremental_marking()->OldSpaceStep( |
- Min(kThreshold, unreported_allocation_)); |
- unreported_allocation_ = 0; |
- } |
+ // TODO(gc) Try not freeing linear allocation region when bytes_left |
+ // are zero. |
+ owner_->SetTopAndLimit(NULL, NULL); |
} |
return new_node; |
@@ -2956,16 +2919,7 @@ |
reinterpret_cast<Object**>(object->address())[1] = Smi::FromInt(0); |
} |
- // We would like to tell the incremental marker to do a lot of work, since |
- // we just made a large allocation in old space, but that might cause a huge |
- // pause. Underreporting here may cause the marker to speed up because it |
- // will perceive that it is not keeping up with allocation. Although this |
- // causes some big incremental marking steps they are not as big as this one |
- // might have been. In testing, a very large pause was divided up into about |
- // 12 parts. |
- const int kThreshold = IncrementalMarking::kAllocatedThreshold * |
- IncrementalMarking::kOldSpaceAllocationMarkingFactor; |
- heap()->incremental_marking()->OldSpaceStep(kThreshold); |
+ heap()->incremental_marking()->OldSpaceStep(object_size); |
return object; |
} |