| Index: src/mark-compact.cc
|
| diff --git a/src/mark-compact.cc b/src/mark-compact.cc
|
| index cbb9dfbff0cef15f00545fdd142e41dc9e18db49..fe0c685c6995ef39bd8ad6b3f74219194332b386 100644
|
| --- a/src/mark-compact.cc
|
| +++ b/src/mark-compact.cc
|
| @@ -585,13 +585,6 @@ bool MarkCompactCollector::IsConcurrentSweepingInProgress() {
|
| }
|
|
|
|
|
| -void MarkCompactCollector::FinalizeSweeping() {
|
| - ASSERT(sweeping_pending_ == false);
|
| - ReleaseEvacuationCandidates();
|
| - heap()->FreeQueuedChunks();
|
| -}
|
| -
|
| -
|
| void MarkCompactCollector::MarkInParallel() {
|
| for (int i = 0; i < FLAG_marking_threads; i++) {
|
| heap()->isolate()->marking_threads()[i]->StartMarking();
|
| @@ -911,7 +904,6 @@ void MarkCompactCollector::Prepare(GCTracer* tracer) {
|
| if (IsConcurrentSweepingInProgress()) {
|
| // Instead of waiting we could also abort the sweeper threads here.
|
| WaitUntilSweepingCompleted();
|
| - FinalizeSweeping();
|
| }
|
|
|
| // Clear marking bits if incremental marking is aborted.
|
| @@ -2849,6 +2841,7 @@ void MarkCompactCollector::EvacuatePages() {
|
| slots_buffer_allocator_.DeallocateChain(page->slots_buffer_address());
|
| page->ClearEvacuationCandidate();
|
| page->SetFlag(Page::RESCAN_ON_EVACUATION);
|
| + page->InsertAfter(static_cast<PagedSpace*>(page->owner())->anchor());
|
| }
|
| return;
|
| }
|
| @@ -3309,6 +3302,18 @@ void MarkCompactCollector::EvacuateNewSpaceAndCandidates() {
|
| }
|
|
|
|
|
| +void MarkCompactCollector::UnlinkEvacuationCandidates() {
|
| + int npages = evacuation_candidates_.length();
|
| + for (int i = 0; i < npages; i++) {
|
| + Page* p = evacuation_candidates_[i];
|
| + if (!p->IsEvacuationCandidate()) continue;
|
| + p->Unlink();
|
| + p->ClearSweptPrecisely();
|
| + p->ClearSweptConservatively();
|
| + }
|
| +}
|
| +
|
| +
|
| void MarkCompactCollector::ReleaseEvacuationCandidates() {
|
| int npages = evacuation_candidates_.length();
|
| for (int i = 0; i < npages; i++) {
|
| @@ -3319,10 +3324,11 @@ void MarkCompactCollector::ReleaseEvacuationCandidates() {
|
| p->set_scan_on_scavenge(false);
|
| slots_buffer_allocator_.DeallocateChain(p->slots_buffer_address());
|
| p->ResetLiveBytes();
|
| - space->ReleasePage(p);
|
| + space->ReleasePage(p, false);
|
| }
|
| evacuation_candidates_.Rewind(0);
|
| compacting_ = false;
|
| + heap()->FreeQueuedChunks();
|
| }
|
|
|
|
|
| @@ -3794,7 +3800,7 @@ void MarkCompactCollector::SweepSpace(PagedSpace* space, SweeperType sweeper) {
|
| // Adjust unswept free bytes because releasing a page expects said
|
| // counter to be accurate for unswept pages.
|
| space->IncreaseUnsweptFreeBytes(p);
|
| - space->ReleasePage(p);
|
| + space->ReleasePage(p, true);
|
| continue;
|
| }
|
| unused_page_present = true;
|
| @@ -3899,6 +3905,10 @@ void MarkCompactCollector::SweepSpaces() {
|
| SweepSpace(heap()->old_pointer_space(), how_to_sweep);
|
| SweepSpace(heap()->old_data_space(), how_to_sweep);
|
|
|
| + // Unlink evacuation candidates before sweeper threads access the list of
|
| + // pages to avoid race condition.
|
| + UnlinkEvacuationCandidates();
|
| +
|
| if (how_to_sweep == PARALLEL_CONSERVATIVE ||
|
| how_to_sweep == CONCURRENT_CONSERVATIVE) {
|
| // TODO(hpayer): fix race with concurrent sweeper
|
| @@ -3924,9 +3934,7 @@ void MarkCompactCollector::SweepSpaces() {
|
| // Deallocate unmarked objects and clear marked bits for marked objects.
|
| heap_->lo_space()->FreeUnmarkedObjects();
|
|
|
| - if (how_to_sweep != CONCURRENT_CONSERVATIVE) {
|
| - FinalizeSweeping();
|
| - }
|
| + ReleaseEvacuationCandidates();
|
| }
|
|
|
|
|
|
|