| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "src/heap/mark-compact.h" | 5 #include "src/heap/mark-compact.h" |
| 6 | 6 |
| 7 #include "src/base/atomicops.h" | 7 #include "src/base/atomicops.h" |
| 8 #include "src/base/bits.h" | 8 #include "src/base/bits.h" |
| 9 #include "src/base/sys-info.h" | 9 #include "src/base/sys-info.h" |
| 10 #include "src/code-stubs.h" | 10 #include "src/code-stubs.h" |
| 11 #include "src/compilation-cache.h" | 11 #include "src/compilation-cache.h" |
| 12 #include "src/deoptimizer.h" | 12 #include "src/deoptimizer.h" |
| 13 #include "src/execution.h" | 13 #include "src/execution.h" |
| 14 #include "src/frames-inl.h" | 14 #include "src/frames-inl.h" |
| 15 #include "src/gdb-jit.h" | 15 #include "src/gdb-jit.h" |
| 16 #include "src/global-handles.h" | 16 #include "src/global-handles.h" |
| 17 #include "src/heap/array-buffer-tracker-inl.h" | 17 #include "src/heap/array-buffer-tracker.h" |
| 18 #include "src/heap/gc-tracer.h" | 18 #include "src/heap/gc-tracer.h" |
| 19 #include "src/heap/incremental-marking.h" | 19 #include "src/heap/incremental-marking.h" |
| 20 #include "src/heap/mark-compact-inl.h" | 20 #include "src/heap/mark-compact-inl.h" |
| 21 #include "src/heap/object-stats.h" | 21 #include "src/heap/object-stats.h" |
| 22 #include "src/heap/objects-visiting-inl.h" | 22 #include "src/heap/objects-visiting-inl.h" |
| 23 #include "src/heap/objects-visiting.h" | 23 #include "src/heap/objects-visiting.h" |
| 24 #include "src/heap/page-parallel-job.h" | 24 #include "src/heap/page-parallel-job.h" |
| 25 #include "src/heap/spaces-inl.h" | 25 #include "src/heap/spaces-inl.h" |
| 26 #include "src/ic/ic.h" | 26 #include "src/ic/ic.h" |
| 27 #include "src/ic/stub-cache.h" | 27 #include "src/ic/stub-cache.h" |
| (...skipping 837 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 865 // marking cycle. We did not collect any slots. | 865 // marking cycle. We did not collect any slots. |
| 866 if (!FLAG_never_compact && !was_marked_incrementally_) { | 866 if (!FLAG_never_compact && !was_marked_incrementally_) { |
| 867 StartCompaction(NON_INCREMENTAL_COMPACTION); | 867 StartCompaction(NON_INCREMENTAL_COMPACTION); |
| 868 } | 868 } |
| 869 | 869 |
| 870 PagedSpaces spaces(heap()); | 870 PagedSpaces spaces(heap()); |
| 871 for (PagedSpace* space = spaces.next(); space != NULL; | 871 for (PagedSpace* space = spaces.next(); space != NULL; |
| 872 space = spaces.next()) { | 872 space = spaces.next()) { |
| 873 space->PrepareForMarkCompact(); | 873 space->PrepareForMarkCompact(); |
| 874 } | 874 } |
| 875 if (!was_marked_incrementally_) { | |
| 876 heap_->array_buffer_tracker()->ResetTrackersInOldSpace(); | |
| 877 } | |
| 878 heap()->account_amount_of_external_allocated_freed_memory(); | |
| 879 | 875 |
| 880 #ifdef VERIFY_HEAP | 876 #ifdef VERIFY_HEAP |
| 881 if (!was_marked_incrementally_ && FLAG_verify_heap) { | 877 if (!was_marked_incrementally_ && FLAG_verify_heap) { |
| 882 VerifyMarkbitsAreClean(); | 878 VerifyMarkbitsAreClean(); |
| 883 } | 879 } |
| 884 #endif | 880 #endif |
| 885 } | 881 } |
| 886 | 882 |
| 887 | 883 |
| 888 void MarkCompactCollector::Finish() { | 884 void MarkCompactCollector::Finish() { |
| (...skipping 835 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1724 local_pretenuring_feedback_(local_pretenuring_feedback) {} | 1720 local_pretenuring_feedback_(local_pretenuring_feedback) {} |
| 1725 | 1721 |
| 1726 inline bool Visit(HeapObject* object) override { | 1722 inline bool Visit(HeapObject* object) override { |
| 1727 heap_->UpdateAllocationSite<Heap::kCached>(object, | 1723 heap_->UpdateAllocationSite<Heap::kCached>(object, |
| 1728 local_pretenuring_feedback_); | 1724 local_pretenuring_feedback_); |
| 1729 int size = object->Size(); | 1725 int size = object->Size(); |
| 1730 HeapObject* target_object = nullptr; | 1726 HeapObject* target_object = nullptr; |
| 1731 if (heap_->ShouldBePromoted(object->address(), size) && | 1727 if (heap_->ShouldBePromoted(object->address(), size) && |
| 1732 TryEvacuateObject(compaction_spaces_->Get(OLD_SPACE), object, | 1728 TryEvacuateObject(compaction_spaces_->Get(OLD_SPACE), object, |
| 1733 &target_object)) { | 1729 &target_object)) { |
| 1730 // If we end up needing more special cases, we should factor this out. |
| 1731 if (V8_UNLIKELY(target_object->IsJSArrayBuffer())) { |
| 1732 heap_->array_buffer_tracker()->Promote( |
| 1733 JSArrayBuffer::cast(target_object)); |
| 1734 } |
| 1734 promoted_size_ += size; | 1735 promoted_size_ += size; |
| 1735 return true; | 1736 return true; |
| 1736 } | 1737 } |
| 1737 HeapObject* target = nullptr; | 1738 HeapObject* target = nullptr; |
| 1738 AllocationSpace space = AllocateTargetObject(object, &target); | 1739 AllocationSpace space = AllocateTargetObject(object, &target); |
| 1739 MigrateObject(HeapObject::cast(target), object, size, space); | 1740 MigrateObject(HeapObject::cast(target), object, size, space); |
| 1741 if (V8_UNLIKELY(target->IsJSArrayBuffer())) { |
| 1742 heap_->array_buffer_tracker()->MarkLive(JSArrayBuffer::cast(target)); |
| 1743 } |
| 1740 semispace_copied_size_ += size; | 1744 semispace_copied_size_ += size; |
| 1741 return true; | 1745 return true; |
| 1742 } | 1746 } |
| 1743 | 1747 |
| 1744 intptr_t promoted_size() { return promoted_size_; } | 1748 intptr_t promoted_size() { return promoted_size_; } |
| 1745 intptr_t semispace_copied_size() { return semispace_copied_size_; } | 1749 intptr_t semispace_copied_size() { return semispace_copied_size_; } |
| 1746 | 1750 |
| 1747 private: | 1751 private: |
| 1748 enum NewSpaceAllocationMode { | 1752 enum NewSpaceAllocationMode { |
| 1749 kNonstickyBailoutOldSpace, | 1753 kNonstickyBailoutOldSpace, |
| (...skipping 104 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1854 : heap_(heap), promoted_size_(0) {} | 1858 : heap_(heap), promoted_size_(0) {} |
| 1855 | 1859 |
| 1856 static void TryMoveToOldSpace(Page* page, PagedSpace* owner) { | 1860 static void TryMoveToOldSpace(Page* page, PagedSpace* owner) { |
| 1857 if (page->heap()->new_space()->ReplaceWithEmptyPage(page)) { | 1861 if (page->heap()->new_space()->ReplaceWithEmptyPage(page)) { |
| 1858 Page* new_page = Page::ConvertNewToOld(page, owner); | 1862 Page* new_page = Page::ConvertNewToOld(page, owner); |
| 1859 new_page->SetFlag(Page::PAGE_NEW_OLD_PROMOTION); | 1863 new_page->SetFlag(Page::PAGE_NEW_OLD_PROMOTION); |
| 1860 } | 1864 } |
| 1861 } | 1865 } |
| 1862 | 1866 |
| 1863 inline bool Visit(HeapObject* object) { | 1867 inline bool Visit(HeapObject* object) { |
| 1868 if (V8_UNLIKELY(object->IsJSArrayBuffer())) { |
| 1869 object->GetHeap()->array_buffer_tracker()->Promote( |
| 1870 JSArrayBuffer::cast(object)); |
| 1871 } |
| 1864 RecordMigratedSlotVisitor visitor(heap_->mark_compact_collector()); | 1872 RecordMigratedSlotVisitor visitor(heap_->mark_compact_collector()); |
| 1865 object->IterateBodyFast(&visitor); | 1873 object->IterateBodyFast(&visitor); |
| 1866 promoted_size_ += object->Size(); | 1874 promoted_size_ += object->Size(); |
| 1867 return true; | 1875 return true; |
| 1868 } | 1876 } |
| 1869 | 1877 |
| 1870 intptr_t promoted_size() { return promoted_size_; } | 1878 intptr_t promoted_size() { return promoted_size_; } |
| 1871 | 1879 |
| 1872 private: | 1880 private: |
| 1873 Heap* heap_; | 1881 Heap* heap_; |
| (...skipping 20 matching lines...) Expand all Loading... |
| 1894 }; | 1902 }; |
| 1895 | 1903 |
| 1896 class MarkCompactCollector::EvacuateRecordOnlyVisitor final | 1904 class MarkCompactCollector::EvacuateRecordOnlyVisitor final |
| 1897 : public MarkCompactCollector::HeapObjectVisitor { | 1905 : public MarkCompactCollector::HeapObjectVisitor { |
| 1898 public: | 1906 public: |
| 1899 explicit EvacuateRecordOnlyVisitor(Heap* heap) : heap_(heap) {} | 1907 explicit EvacuateRecordOnlyVisitor(Heap* heap) : heap_(heap) {} |
| 1900 | 1908 |
| 1901 inline bool Visit(HeapObject* object) { | 1909 inline bool Visit(HeapObject* object) { |
| 1902 RecordMigratedSlotVisitor visitor(heap_->mark_compact_collector()); | 1910 RecordMigratedSlotVisitor visitor(heap_->mark_compact_collector()); |
| 1903 object->IterateBody(&visitor); | 1911 object->IterateBody(&visitor); |
| 1904 if (V8_UNLIKELY(object->IsJSArrayBuffer())) { | |
| 1905 heap_->array_buffer_tracker()->MarkLive(JSArrayBuffer::cast(object)); | |
| 1906 } | |
| 1907 return true; | 1912 return true; |
| 1908 } | 1913 } |
| 1909 | 1914 |
| 1910 private: | 1915 private: |
| 1911 Heap* heap_; | 1916 Heap* heap_; |
| 1912 }; | 1917 }; |
| 1913 | 1918 |
| 1914 void MarkCompactCollector::DiscoverGreyObjectsInSpace(PagedSpace* space) { | 1919 void MarkCompactCollector::DiscoverGreyObjectsInSpace(PagedSpace* space) { |
| 1915 PageIterator it(space); | 1920 PageIterator it(space); |
| 1916 while (it.has_next()) { | 1921 while (it.has_next()) { |
| (...skipping 1194 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3111 } | 3116 } |
| 3112 if (success) { | 3117 if (success) { |
| 3113 ReportCompactionProgress(evacuation_time, saved_live_bytes); | 3118 ReportCompactionProgress(evacuation_time, saved_live_bytes); |
| 3114 } | 3119 } |
| 3115 return success; | 3120 return success; |
| 3116 } | 3121 } |
| 3117 | 3122 |
| 3118 bool MarkCompactCollector::Evacuator::EvacuatePage(Page* page) { | 3123 bool MarkCompactCollector::Evacuator::EvacuatePage(Page* page) { |
| 3119 bool result = false; | 3124 bool result = false; |
| 3120 DCHECK(page->SweepingDone()); | 3125 DCHECK(page->SweepingDone()); |
| 3121 Heap* heap = page->heap(); | |
| 3122 switch (ComputeEvacuationMode(page)) { | 3126 switch (ComputeEvacuationMode(page)) { |
| 3123 case kObjectsNewToOld: | 3127 case kObjectsNewToOld: |
| 3124 result = EvacuateSinglePage<kClearMarkbits>(page, &new_space_visitor_); | 3128 result = EvacuateSinglePage<kClearMarkbits>(page, &new_space_visitor_); |
| 3125 heap->array_buffer_tracker() | |
| 3126 ->ScanAndFreeDeadArrayBuffers< | |
| 3127 LocalArrayBufferTracker::kForwardingPointer>(page); | |
| 3128 DCHECK(result); | 3129 DCHECK(result); |
| 3129 USE(result); | 3130 USE(result); |
| 3130 break; | 3131 break; |
| 3131 case kPageNewToOld: | 3132 case kPageNewToOld: |
| 3132 result = EvacuateSinglePage<kKeepMarking>(page, &new_space_page_visitor); | 3133 result = EvacuateSinglePage<kKeepMarking>(page, &new_space_page_visitor); |
| 3133 // ArrayBufferTracker will be updated during sweeping. | |
| 3134 DCHECK(result); | 3134 DCHECK(result); |
| 3135 USE(result); | 3135 USE(result); |
| 3136 break; | 3136 break; |
| 3137 case kObjectsOldToOld: | 3137 case kObjectsOldToOld: |
| 3138 result = EvacuateSinglePage<kClearMarkbits>(page, &old_space_visitor_); | 3138 result = EvacuateSinglePage<kClearMarkbits>(page, &old_space_visitor_); |
| 3139 heap->array_buffer_tracker() | |
| 3140 ->ScanAndFreeDeadArrayBuffers< | |
| 3141 LocalArrayBufferTracker::kForwardingPointer>(page); | |
| 3142 if (!result) { | 3139 if (!result) { |
| 3143 // Aborted compaction page. We can record slots here to have them | 3140 // Aborted compaction page. We can record slots here to have them |
| 3144 // processed in parallel later on. | 3141 // processed in parallel later on. |
| 3145 EvacuateRecordOnlyVisitor record_visitor(collector_->heap()); | 3142 EvacuateRecordOnlyVisitor record_visitor(collector_->heap()); |
| 3146 result = EvacuateSinglePage<kKeepMarking>(page, &record_visitor); | 3143 result = EvacuateSinglePage<kKeepMarking>(page, &record_visitor); |
| 3147 heap->array_buffer_tracker() | |
| 3148 ->ScanAndFreeDeadArrayBuffers<LocalArrayBufferTracker::kMarkBit>( | |
| 3149 page); | |
| 3150 DCHECK(result); | 3144 DCHECK(result); |
| 3151 USE(result); | 3145 USE(result); |
| 3152 // We need to return failure here to indicate that we want this page | 3146 // We need to return failure here to indicate that we want this page |
| 3153 // added to the sweeper. | 3147 // added to the sweeper. |
| 3154 return false; | 3148 return false; |
| 3155 } | 3149 } |
| 3156 break; | 3150 break; |
| 3157 default: | 3151 default: |
| 3158 UNREACHABLE(); | 3152 UNREACHABLE(); |
| 3159 } | 3153 } |
| (...skipping 222 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3382 Bitmap::Clear(p); | 3376 Bitmap::Clear(p); |
| 3383 | 3377 |
| 3384 if (free_start != p->area_end()) { | 3378 if (free_start != p->area_end()) { |
| 3385 int size = static_cast<int>(p->area_end() - free_start); | 3379 int size = static_cast<int>(p->area_end() - free_start); |
| 3386 if (free_space_mode == ZAP_FREE_SPACE) { | 3380 if (free_space_mode == ZAP_FREE_SPACE) { |
| 3387 memset(free_start, 0xcc, size); | 3381 memset(free_start, 0xcc, size); |
| 3388 } | 3382 } |
| 3389 freed_bytes = space->UnaccountedFree(free_start, size); | 3383 freed_bytes = space->UnaccountedFree(free_start, size); |
| 3390 max_freed_bytes = Max(freed_bytes, max_freed_bytes); | 3384 max_freed_bytes = Max(freed_bytes, max_freed_bytes); |
| 3391 } | 3385 } |
| 3392 p->heap()->array_buffer_tracker()->FreeDead(p); | |
| 3393 p->concurrent_sweeping_state().SetValue(Page::kSweepingDone); | 3386 p->concurrent_sweeping_state().SetValue(Page::kSweepingDone); |
| 3394 return FreeList::GuaranteedAllocatable(static_cast<int>(max_freed_bytes)); | 3387 return FreeList::GuaranteedAllocatable(static_cast<int>(max_freed_bytes)); |
| 3395 } | 3388 } |
| 3396 | 3389 |
| 3397 void MarkCompactCollector::InvalidateCode(Code* code) { | 3390 void MarkCompactCollector::InvalidateCode(Code* code) { |
| 3398 if (heap_->incremental_marking()->IsCompacting() && | 3391 if (heap_->incremental_marking()->IsCompacting() && |
| 3399 !ShouldSkipEvacuationSlotRecording(code)) { | 3392 !ShouldSkipEvacuationSlotRecording(code)) { |
| 3400 DCHECK(compacting_); | 3393 DCHECK(compacting_); |
| 3401 | 3394 |
| 3402 // If the object is white than no slots were recorded on it yet. | 3395 // If the object is white than no slots were recorded on it yet. |
| (...skipping 123 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3526 // because root iteration traverses the stack and might have to find | 3519 // because root iteration traverses the stack and might have to find |
| 3527 // code objects from non-updated pc pointing into evacuation candidate. | 3520 // code objects from non-updated pc pointing into evacuation candidate. |
| 3528 SkipList* list = p->skip_list(); | 3521 SkipList* list = p->skip_list(); |
| 3529 if (list != NULL) list->Clear(); | 3522 if (list != NULL) list->Clear(); |
| 3530 if (p->IsFlagSet(Page::COMPACTION_WAS_ABORTED)) { | 3523 if (p->IsFlagSet(Page::COMPACTION_WAS_ABORTED)) { |
| 3531 sweeper().AddLatePage(p->owner()->identity(), p); | 3524 sweeper().AddLatePage(p->owner()->identity(), p); |
| 3532 p->ClearFlag(Page::COMPACTION_WAS_ABORTED); | 3525 p->ClearFlag(Page::COMPACTION_WAS_ABORTED); |
| 3533 } | 3526 } |
| 3534 } | 3527 } |
| 3535 | 3528 |
| 3529 // EvacuateNewSpaceAndCandidates iterates over new space objects and for |
| 3530 // ArrayBuffers either re-registers them as live or promotes them. This is |
| 3531 // needed to properly free them. |
| 3532 heap()->array_buffer_tracker()->FreeDead(false); |
| 3533 |
| 3536 // Deallocate evacuated candidate pages. | 3534 // Deallocate evacuated candidate pages. |
| 3537 ReleaseEvacuationCandidates(); | 3535 ReleaseEvacuationCandidates(); |
| 3538 } | 3536 } |
| 3539 | 3537 |
| 3540 #ifdef VERIFY_HEAP | 3538 #ifdef VERIFY_HEAP |
| 3541 if (FLAG_verify_heap && !sweeper().sweeping_in_progress()) { | 3539 if (FLAG_verify_heap && !sweeper().sweeping_in_progress()) { |
| 3542 VerifyEvacuation(heap()); | 3540 VerifyEvacuation(heap()); |
| 3543 } | 3541 } |
| 3544 #endif | 3542 #endif |
| 3545 } | 3543 } |
| (...skipping 395 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3941 MarkBit mark_bit = Marking::MarkBitFrom(host); | 3939 MarkBit mark_bit = Marking::MarkBitFrom(host); |
| 3942 if (Marking::IsBlack(mark_bit)) { | 3940 if (Marking::IsBlack(mark_bit)) { |
| 3943 RelocInfo rinfo(isolate(), pc, RelocInfo::CODE_TARGET, 0, host); | 3941 RelocInfo rinfo(isolate(), pc, RelocInfo::CODE_TARGET, 0, host); |
| 3944 RecordRelocSlot(host, &rinfo, target); | 3942 RecordRelocSlot(host, &rinfo, target); |
| 3945 } | 3943 } |
| 3946 } | 3944 } |
| 3947 } | 3945 } |
| 3948 | 3946 |
| 3949 } // namespace internal | 3947 } // namespace internal |
| 3950 } // namespace v8 | 3948 } // namespace v8 |
| OLD | NEW |