Index: src/heap/mark-compact.cc |
diff --git a/src/heap/mark-compact.cc b/src/heap/mark-compact.cc |
index 84a2a7f31898b63cd4836210c854b1461ce5c74e..de381aab356657374f1e62f083173045d949ded0 100644 |
--- a/src/heap/mark-compact.cc |
+++ b/src/heap/mark-compact.cc |
@@ -135,7 +135,8 @@ static void VerifyMarking(PagedSpace* space) { |
static void VerifyMarking(Heap* heap) { |
- VerifyMarking(heap->old_space()); |
+ VerifyMarking(heap->old_pointer_space()); |
+ VerifyMarking(heap->old_data_space()); |
VerifyMarking(heap->code_space()); |
VerifyMarking(heap->cell_space()); |
VerifyMarking(heap->map_space()); |
@@ -199,7 +200,8 @@ static void VerifyEvacuation(NewSpace* space) { |
static void VerifyEvacuation(Heap* heap, PagedSpace* space) { |
- if (FLAG_use_allocation_folding && (space == heap->old_space())) { |
+ if (FLAG_use_allocation_folding && |
+ (space == heap->old_pointer_space() || space == heap->old_data_space())) { |
return; |
} |
PageIterator it(space); |
@@ -213,7 +215,8 @@ static void VerifyEvacuation(Heap* heap, PagedSpace* space) { |
static void VerifyEvacuation(Heap* heap) { |
- VerifyEvacuation(heap, heap->old_space()); |
+ VerifyEvacuation(heap, heap->old_pointer_space()); |
+ VerifyEvacuation(heap, heap->old_data_space()); |
VerifyEvacuation(heap, heap->code_space()); |
VerifyEvacuation(heap, heap->cell_space()); |
VerifyEvacuation(heap, heap->map_space()); |
@@ -226,7 +229,8 @@ static void VerifyEvacuation(Heap* heap) { |
void MarkCompactCollector::SetUp() { |
- free_list_old_space_.Reset(new FreeList(heap_->old_space())); |
+ free_list_old_data_space_.Reset(new FreeList(heap_->old_data_space())); |
+ free_list_old_pointer_space_.Reset(new FreeList(heap_->old_pointer_space())); |
} |
@@ -257,7 +261,8 @@ bool MarkCompactCollector::StartCompaction(CompactionMode mode) { |
if (!compacting_) { |
DCHECK(evacuation_candidates_.length() == 0); |
- CollectEvacuationCandidates(heap()->old_space()); |
+ CollectEvacuationCandidates(heap()->old_pointer_space()); |
+ CollectEvacuationCandidates(heap()->old_data_space()); |
if (FLAG_compact_code_space && (mode == NON_INCREMENTAL_COMPACTION || |
FLAG_incremental_code_compaction)) { |
@@ -271,7 +276,8 @@ bool MarkCompactCollector::StartCompaction(CompactionMode mode) { |
TraceFragmentation(heap()->cell_space()); |
} |
- heap()->old_space()->EvictEvacuationCandidatesFromFreeLists(); |
+ heap()->old_pointer_space()->EvictEvacuationCandidatesFromFreeLists(); |
+ heap()->old_data_space()->EvictEvacuationCandidatesFromFreeLists(); |
heap()->code_space()->EvictEvacuationCandidatesFromFreeLists(); |
compacting_ = evacuation_candidates_.length() > 0; |
@@ -358,7 +364,8 @@ void MarkCompactCollector::VerifyMarkbitsAreClean(NewSpace* space) { |
void MarkCompactCollector::VerifyMarkbitsAreClean() { |
- VerifyMarkbitsAreClean(heap_->old_space()); |
+ VerifyMarkbitsAreClean(heap_->old_pointer_space()); |
+ VerifyMarkbitsAreClean(heap_->old_data_space()); |
VerifyMarkbitsAreClean(heap_->code_space()); |
VerifyMarkbitsAreClean(heap_->cell_space()); |
VerifyMarkbitsAreClean(heap_->map_space()); |
@@ -416,7 +423,8 @@ static void ClearMarkbitsInNewSpace(NewSpace* space) { |
void MarkCompactCollector::ClearMarkbits() { |
ClearMarkbitsInPagedSpace(heap_->code_space()); |
ClearMarkbitsInPagedSpace(heap_->map_space()); |
- ClearMarkbitsInPagedSpace(heap_->old_space()); |
+ ClearMarkbitsInPagedSpace(heap_->old_pointer_space()); |
+ ClearMarkbitsInPagedSpace(heap_->old_data_space()); |
ClearMarkbitsInPagedSpace(heap_->cell_space()); |
ClearMarkbitsInNewSpace(heap_->new_space()); |
@@ -452,9 +460,13 @@ class MarkCompactCollector::SweeperTask : public v8::Task { |
void MarkCompactCollector::StartSweeperThreads() { |
- DCHECK(free_list_old_space_.get()->IsEmpty()); |
+ DCHECK(free_list_old_pointer_space_.get()->IsEmpty()); |
+ DCHECK(free_list_old_data_space_.get()->IsEmpty()); |
V8::GetCurrentPlatform()->CallOnBackgroundThread( |
- new SweeperTask(heap(), heap()->old_space()), |
+ new SweeperTask(heap(), heap()->old_data_space()), |
+ v8::Platform::kShortRunningTask); |
+ V8::GetCurrentPlatform()->CallOnBackgroundThread( |
+ new SweeperTask(heap(), heap()->old_pointer_space()), |
v8::Platform::kShortRunningTask); |
} |
@@ -465,16 +477,20 @@ void MarkCompactCollector::EnsureSweepingCompleted() { |
// If sweeping is not completed or not running at all, we try to complete it |
// here. |
if (!heap()->concurrent_sweeping_enabled() || !IsSweepingCompleted()) { |
- SweepInParallel(heap()->paged_space(OLD_SPACE), 0); |
+ SweepInParallel(heap()->paged_space(OLD_DATA_SPACE), 0); |
+ SweepInParallel(heap()->paged_space(OLD_POINTER_SPACE), 0); |
} |
// Wait twice for both jobs. |
if (heap()->concurrent_sweeping_enabled()) { |
pending_sweeper_jobs_semaphore_.Wait(); |
+ pending_sweeper_jobs_semaphore_.Wait(); |
} |
ParallelSweepSpacesComplete(); |
sweeping_in_progress_ = false; |
- RefillFreeList(heap()->paged_space(OLD_SPACE)); |
- heap()->paged_space(OLD_SPACE)->ResetUnsweptFreeBytes(); |
+ RefillFreeList(heap()->paged_space(OLD_DATA_SPACE)); |
+ RefillFreeList(heap()->paged_space(OLD_POINTER_SPACE)); |
+ heap()->paged_space(OLD_DATA_SPACE)->ResetUnsweptFreeBytes(); |
+ heap()->paged_space(OLD_POINTER_SPACE)->ResetUnsweptFreeBytes(); |
#ifdef VERIFY_HEAP |
if (FLAG_verify_heap && !evacuation()) { |
@@ -497,11 +513,13 @@ bool MarkCompactCollector::IsSweepingCompleted() { |
void MarkCompactCollector::RefillFreeList(PagedSpace* space) { |
FreeList* free_list; |
- if (space == heap()->old_space()) { |
- free_list = free_list_old_space_.get(); |
+ if (space == heap()->old_pointer_space()) { |
+ free_list = free_list_old_pointer_space_.get(); |
+ } else if (space == heap()->old_data_space()) { |
+ free_list = free_list_old_data_space_.get(); |
} else { |
// Any PagedSpace might invoke RefillFreeLists, so we need to make sure |
- // to only refill them for the old space. |
+ // to only refill them for old data and pointer spaces. |
return; |
} |
@@ -555,8 +573,10 @@ const char* AllocationSpaceName(AllocationSpace space) { |
switch (space) { |
case NEW_SPACE: |
return "NEW_SPACE"; |
- case OLD_SPACE: |
- return "OLD_SPACE"; |
+ case OLD_POINTER_SPACE: |
+ return "OLD_POINTER_SPACE"; |
+ case OLD_DATA_SPACE: |
+ return "OLD_DATA_SPACE"; |
case CODE_SPACE: |
return "CODE_SPACE"; |
case MAP_SPACE: |
@@ -625,7 +645,9 @@ static int FreeListFragmentation(PagedSpace* space, Page* p) { |
void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) { |
- DCHECK(space->identity() == OLD_SPACE || space->identity() == CODE_SPACE); |
+ DCHECK(space->identity() == OLD_POINTER_SPACE || |
+ space->identity() == OLD_DATA_SPACE || |
+ space->identity() == CODE_SPACE); |
static const int kMaxMaxEvacuationCandidates = 1000; |
int number_of_pages = space->CountTotalPages(); |
@@ -1791,7 +1813,7 @@ static void DiscoverGreyObjectsOnPage(MarkingDeque* marking_deque, |
int trailing_zeros = base::bits::CountTrailingZeros32(grey_objects); |
grey_objects >>= trailing_zeros; |
offset += trailing_zeros; |
- MarkBit markbit(cell, 1 << offset); |
+ MarkBit markbit(cell, 1 << offset, false); |
DCHECK(Marking::IsGrey(markbit)); |
Marking::GreyToBlack(markbit); |
Address addr = cell_base + offset * kPointerSize; |
@@ -2015,7 +2037,11 @@ void MarkCompactCollector::RefillMarkingDeque() { |
DiscoverGreyObjectsInNewSpace(heap(), &marking_deque_); |
if (marking_deque_.IsFull()) return; |
- DiscoverGreyObjectsInSpace(heap(), &marking_deque_, heap()->old_space()); |
+ DiscoverGreyObjectsInSpace(heap(), &marking_deque_, |
+ heap()->old_pointer_space()); |
+ if (marking_deque_.IsFull()) return; |
+ |
+ DiscoverGreyObjectsInSpace(heap(), &marking_deque_, heap()->old_data_space()); |
if (marking_deque_.IsFull()) return; |
DiscoverGreyObjectsInSpace(heap(), &marking_deque_, heap()->code_space()); |
@@ -2687,7 +2713,7 @@ void MarkCompactCollector::MigrateObject(HeapObject* dst, HeapObject* src, |
Address src_addr = src->address(); |
DCHECK(heap()->AllowedToBeMigrated(src, dest)); |
DCHECK(dest != LO_SPACE && size <= Page::kMaxRegularHeapObjectSize); |
- if (dest == OLD_SPACE) { |
+ if (dest == OLD_POINTER_SPACE) { |
Address src_slot = src_addr; |
Address dst_slot = dst_addr; |
DCHECK(IsAligned(size, kPointerSize)); |
@@ -2759,7 +2785,7 @@ void MarkCompactCollector::MigrateObject(HeapObject* dst, HeapObject* src, |
SlotsBuffer::IGNORE_OVERFLOW); |
Code::cast(dst)->Relocate(dst_addr - src_addr); |
} else { |
- DCHECK(dest == NEW_SPACE); |
+ DCHECK(dest == OLD_DATA_SPACE || dest == NEW_SPACE); |
heap()->MoveBlock(dst_addr, src_addr, size); |
} |
heap()->OnMoveEvent(dst, src, size); |
@@ -2881,17 +2907,19 @@ void PointersUpdatingVisitor::CheckLayoutDescriptorAndDie(Heap* heap, |
space_owner_id = 1; |
} else if (heap->new_space()->FromSpaceContains(slot_address)) { |
space_owner_id = 2; |
- } else if (heap->old_space()->ContainsSafe(slot_address)) { |
+ } else if (heap->old_pointer_space()->ContainsSafe(slot_address)) { |
space_owner_id = 3; |
- } else if (heap->code_space()->ContainsSafe(slot_address)) { |
+ } else if (heap->old_data_space()->ContainsSafe(slot_address)) { |
space_owner_id = 4; |
- } else if (heap->map_space()->ContainsSafe(slot_address)) { |
+ } else if (heap->code_space()->ContainsSafe(slot_address)) { |
space_owner_id = 5; |
- } else if (heap->cell_space()->ContainsSafe(slot_address)) { |
+ } else if (heap->map_space()->ContainsSafe(slot_address)) { |
space_owner_id = 6; |
+ } else if (heap->cell_space()->ContainsSafe(slot_address)) { |
+ space_owner_id = 7; |
} else { |
// Lo space or other. |
- space_owner_id = 7; |
+ space_owner_id = 8; |
} |
data[index++] = space_owner_id; |
data[index++] = 0x20aaaaaaaaUL; |
@@ -2982,12 +3010,14 @@ bool MarkCompactCollector::TryPromoteObject(HeapObject* object, |
int object_size) { |
DCHECK(object_size <= Page::kMaxRegularHeapObjectSize); |
- OldSpace* old_space = heap()->old_space(); |
+ OldSpace* target_space = heap()->TargetSpace(object); |
+ DCHECK(target_space == heap()->old_pointer_space() || |
+ target_space == heap()->old_data_space()); |
HeapObject* target; |
- AllocationResult allocation = old_space->AllocateRaw(object_size); |
+ AllocationResult allocation = target_space->AllocateRaw(object_size); |
if (allocation.To(&target)) { |
- MigrateObject(target, object, object_size, old_space->identity()); |
+ MigrateObject(target, object, object_size, target_space->identity()); |
heap()->IncrementPromotedObjectsSize(object_size); |
return true; |
} |
@@ -3473,8 +3503,9 @@ static bool IsOnInvalidatedCodeObject(Address addr) { |
// we can safely go to the page from the slot address. |
Page* p = Page::FromAddress(addr); |
- // First check owner's identity because old space is swept concurrently or |
- // lazily and might still have non-zero mark-bits on some pages. |
+ // First check owner's identity because old pointer and old data spaces |
+ // are swept lazily and might still have non-zero mark-bits on some |
+ // pages. |
if (p->owner()->identity() != CODE_SPACE) return false; |
// In code space only bits on evacuation candidates (but we don't record |
@@ -3652,7 +3683,12 @@ void MarkCompactCollector::EvacuateNewSpaceAndCandidates() { |
p->ClearFlag(MemoryChunk::RESCAN_ON_EVACUATION); |
switch (space->identity()) { |
- case OLD_SPACE: |
+ case OLD_DATA_SPACE: |
+ Sweep<SWEEP_AND_VISIT_LIVE_OBJECTS, SWEEP_ON_MAIN_THREAD, |
+ IGNORE_SKIP_LIST, IGNORE_FREE_SPACE>(space, NULL, p, |
+ &updating_visitor); |
+ break; |
+ case OLD_POINTER_SPACE: |
Sweep<SWEEP_AND_VISIT_LIVE_OBJECTS, SWEEP_ON_MAIN_THREAD, |
IGNORE_SKIP_LIST, IGNORE_FREE_SPACE>(space, NULL, p, |
&updating_visitor); |
@@ -4163,7 +4199,9 @@ int MarkCompactCollector::SweepInParallel(PagedSpace* space, |
int MarkCompactCollector::SweepInParallel(Page* page, PagedSpace* space) { |
int max_freed = 0; |
if (page->TryParallelSweeping()) { |
- FreeList* free_list = free_list_old_space_.get(); |
+ FreeList* free_list = space == heap()->old_pointer_space() |
+ ? free_list_old_pointer_space_.get() |
+ : free_list_old_data_space_.get(); |
FreeList private_free_list(space); |
max_freed = Sweep<SWEEP_ONLY, SWEEP_IN_PARALLEL, IGNORE_SKIP_LIST, |
IGNORE_FREE_SPACE>(space, &private_free_list, page, NULL); |
@@ -4288,7 +4326,10 @@ void MarkCompactCollector::SweepSpaces() { |
{ |
GCTracer::Scope sweep_scope(heap()->tracer(), |
GCTracer::Scope::MC_SWEEP_OLDSPACE); |
- { SweepSpace(heap()->old_space(), CONCURRENT_SWEEPING); } |
+ { |
+ SweepSpace(heap()->old_pointer_space(), CONCURRENT_SWEEPING); |
+ SweepSpace(heap()->old_data_space(), CONCURRENT_SWEEPING); |
+ } |
sweeping_in_progress_ = true; |
if (heap()->concurrent_sweeping_enabled()) { |
StartSweeperThreads(); |
@@ -4350,7 +4391,8 @@ void MarkCompactCollector::ParallelSweepSpaceComplete(PagedSpace* space) { |
void MarkCompactCollector::ParallelSweepSpacesComplete() { |
- ParallelSweepSpaceComplete(heap()->old_space()); |
+ ParallelSweepSpaceComplete(heap()->old_pointer_space()); |
+ ParallelSweepSpaceComplete(heap()->old_data_space()); |
} |