| Index: src/heap/mark-compact.cc
|
| diff --git a/src/heap/mark-compact.cc b/src/heap/mark-compact.cc
|
| index 487617fc3d14d59932df2af62196707bc2f7e2f4..b8c318010e028b394ac6a3c1165089399cb7a91c 100644
|
| --- a/src/heap/mark-compact.cc
|
| +++ b/src/heap/mark-compact.cc
|
| @@ -14,7 +14,7 @@
|
| #include "src/frames-inl.h"
|
| #include "src/gdb-jit.h"
|
| #include "src/global-handles.h"
|
| -#include "src/heap/array-buffer-tracker.h"
|
| +#include "src/heap/array-buffer-tracker-inl.h"
|
| #include "src/heap/gc-tracer.h"
|
| #include "src/heap/incremental-marking.h"
|
| #include "src/heap/mark-compact-inl.h"
|
| @@ -872,6 +872,10 @@ void MarkCompactCollector::Prepare() {
|
| space = spaces.next()) {
|
| space->PrepareForMarkCompact();
|
| }
|
| + if (!was_marked_incrementally_) {
|
| + ArrayBufferTracker::ResetTrackersInOldSpace(heap_);
|
| + }
|
| + heap()->account_amount_of_external_allocated_freed_memory();
|
|
|
| #ifdef VERIFY_HEAP
|
| if (!was_marked_incrementally_ && FLAG_verify_heap) {
|
| @@ -1727,20 +1731,12 @@ class MarkCompactCollector::EvacuateNewSpaceVisitor final
|
| if (heap_->ShouldBePromoted(object->address(), size) &&
|
| TryEvacuateObject(compaction_spaces_->Get(OLD_SPACE), object,
|
| &target_object)) {
|
| - // If we end up needing more special cases, we should factor this out.
|
| - if (V8_UNLIKELY(target_object->IsJSArrayBuffer())) {
|
| - heap_->array_buffer_tracker()->Promote(
|
| - JSArrayBuffer::cast(target_object));
|
| - }
|
| promoted_size_ += size;
|
| return true;
|
| }
|
| HeapObject* target = nullptr;
|
| AllocationSpace space = AllocateTargetObject(object, &target);
|
| MigrateObject(HeapObject::cast(target), object, size, space);
|
| - if (V8_UNLIKELY(target->IsJSArrayBuffer())) {
|
| - heap_->array_buffer_tracker()->MarkLive(JSArrayBuffer::cast(target));
|
| - }
|
| semispace_copied_size_ += size;
|
| return true;
|
| }
|
| @@ -1865,10 +1861,6 @@ class MarkCompactCollector::EvacuateNewSpacePageVisitor final
|
| }
|
|
|
| inline bool Visit(HeapObject* object) {
|
| - if (V8_UNLIKELY(object->IsJSArrayBuffer())) {
|
| - object->GetHeap()->array_buffer_tracker()->Promote(
|
| - JSArrayBuffer::cast(object));
|
| - }
|
| RecordMigratedSlotVisitor visitor(heap_->mark_compact_collector());
|
| object->IterateBodyFast(&visitor);
|
| promoted_size_ += object->Size();
|
| @@ -1909,6 +1901,9 @@ class MarkCompactCollector::EvacuateRecordOnlyVisitor final
|
| inline bool Visit(HeapObject* object) {
|
| RecordMigratedSlotVisitor visitor(heap_->mark_compact_collector());
|
| object->IterateBody(&visitor);
|
| + if (V8_UNLIKELY(object->IsJSArrayBuffer())) {
|
| + ArrayBufferTracker::MarkLive(heap_, JSArrayBuffer::cast(object));
|
| + }
|
| return true;
|
| }
|
|
|
| @@ -3127,11 +3122,14 @@ bool MarkCompactCollector::Evacuator::EvacuatePage(Page* page) {
|
| switch (ComputeEvacuationMode(page)) {
|
| case kObjectsNewToOld:
|
| result = EvacuateSinglePage<kClearMarkbits>(page, &new_space_visitor_);
|
| + ArrayBufferTracker::ScanAndFreeDeadArrayBuffers<
|
| + LocalArrayBufferTracker::kForwardingPointer>(page);
|
| DCHECK(result);
|
| USE(result);
|
| break;
|
| case kPageNewToOld:
|
| result = EvacuateSinglePage<kKeepMarking>(page, &new_space_page_visitor);
|
| + // ArrayBufferTracker will be updated during sweeping.
|
| DCHECK(result);
|
| USE(result);
|
| break;
|
| @@ -3145,12 +3143,16 @@ bool MarkCompactCollector::Evacuator::EvacuatePage(Page* page) {
|
| page->SetFlag(Page::COMPACTION_WAS_ABORTED);
|
| EvacuateRecordOnlyVisitor record_visitor(collector_->heap());
|
| result = EvacuateSinglePage<kKeepMarking>(page, &record_visitor);
|
| + ArrayBufferTracker::ScanAndFreeDeadArrayBuffers<
|
| + LocalArrayBufferTracker::kForwardingPointerOrMarkBit>(page);
|
| DCHECK(result);
|
| USE(result);
|
| // We need to return failure here to indicate that we want this page
|
| // added to the sweeper.
|
| return false;
|
| }
|
| + ArrayBufferTracker::ScanAndFreeDeadArrayBuffers<
|
| + LocalArrayBufferTracker::kForwardingPointer>(page);
|
| break;
|
| default:
|
| UNREACHABLE();
|
| @@ -3386,6 +3388,7 @@ int MarkCompactCollector::Sweeper::RawSweep(PagedSpace* space, Page* p,
|
| freed_bytes = space->UnaccountedFree(free_start, size);
|
| max_freed_bytes = Max(freed_bytes, max_freed_bytes);
|
| }
|
| + ArrayBufferTracker::FreeDead(p);
|
| p->concurrent_sweeping_state().SetValue(Page::kSweepingDone);
|
| return FreeList::GuaranteedAllocatable(static_cast<int>(max_freed_bytes));
|
| }
|
| @@ -3529,11 +3532,6 @@ void MarkCompactCollector::EvacuateNewSpaceAndCandidates() {
|
| }
|
| }
|
|
|
| - // EvacuateNewSpaceAndCandidates iterates over new space objects and for
|
| - // ArrayBuffers either re-registers them as live or promotes them. This is
|
| - // needed to properly free them.
|
| - heap()->array_buffer_tracker()->FreeDead(false);
|
| -
|
| // Deallocate evacuated candidate pages.
|
| ReleaseEvacuationCandidates();
|
| }
|
|
|