Index: src/heap/mark-compact.cc |
diff --git a/src/heap/mark-compact.cc b/src/heap/mark-compact.cc |
index b7bcf9e80035e8768c3a95eb9c13cf387f03e3f6..075a2d5a7de38208ee633e218b91d19281469d4a 100644 |
--- a/src/heap/mark-compact.cc |
+++ b/src/heap/mark-compact.cc |
@@ -683,6 +683,8 @@ void MarkCompactCollector::CollectEvacuationCandidates(PagedSpace* space) { |
PageIterator it(space); |
while (it.has_next()) { |
Page* p = it.next(); |
+ // Invariant: No page should be marked as aborted after a GC. |
+ DCHECK(!p->IsFlagSet(Page::COMPACTION_WAS_ABORTED)); |
if (p->NeverEvacuate()) continue; |
if (p->IsFlagSet(Page::POPULAR_PAGE)) { |
// This page had slots buffer overflow on previous GC, skip it. |
@@ -3278,8 +3280,13 @@ void MarkCompactCollector::EvacuatePagesInParallel() { |
// happens upon moving (which we potentially didn't do). |
// - Leave the page in the list of pages of a space since we could not |
// fully evacuate it. |
+ // - Mark them for rescanning for store buffer entries as we otherwise |
+ // might have stale store buffer entries that become "valid" again |
+ // after reusing the memory. Note that all existing store buffer |
+ // entries of such pages are filtered before rescanning. |
DCHECK(p->IsEvacuationCandidate()); |
p->SetFlag(Page::COMPACTION_WAS_ABORTED); |
+ p->set_scan_on_scavenge(true); |
abandoned_pages++; |
break; |
case MemoryChunk::kCompactingFinalize: |
@@ -3658,14 +3665,6 @@ void MarkCompactCollector::EvacuateNewSpaceAndCandidates() { |
heap_->IterateRoots(&updating_visitor, VISIT_ALL_IN_SWEEP_NEWSPACE); |
} |
- { |
- GCTracer::Scope gc_scope(heap()->tracer(), |
- GCTracer::Scope::MC_UPDATE_OLD_TO_NEW_POINTERS); |
- StoreBufferRebuildScope scope(heap_, heap_->store_buffer(), |
- &Heap::ScavengeStoreBufferCallback); |
- heap_->store_buffer()->IteratePointersToNewSpace(&UpdatePointer); |
- } |
- |
int npages = evacuation_candidates_.length(); |
{ |
GCTracer::Scope gc_scope( |
@@ -3754,6 +3753,16 @@ void MarkCompactCollector::EvacuateNewSpaceAndCandidates() { |
SweepAbortedPages(); |
} |
+ { |
+ // Note that this phase needs to happen after making aborted pages iterable |
+ // in the previous (sweeping) phase. |
+ GCTracer::Scope gc_scope(heap()->tracer(), |
+ GCTracer::Scope::MC_UPDATE_OLD_TO_NEW_POINTERS); |
+ StoreBufferRebuildScope scope(heap_, heap_->store_buffer(), |
+ &Heap::ScavengeStoreBufferCallback); |
+ heap_->store_buffer()->IteratePointersToNewSpace(&UpdatePointer); |
+ } |
+ |
heap_->isolate()->inner_pointer_to_code_cache()->Flush(); |
// The hashing of weak_object_to_code_table is no longer valid. |