OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
45 | 45 |
46 | 46 |
47 void MarkCompactCollector::SetFlags(int flags) { | 47 void MarkCompactCollector::SetFlags(int flags) { |
48 sweep_precisely_ = ((flags & Heap::kSweepPreciselyMask) != 0); | 48 sweep_precisely_ = ((flags & Heap::kSweepPreciselyMask) != 0); |
49 reduce_memory_footprint_ = ((flags & Heap::kReduceMemoryFootprintMask) != 0); | 49 reduce_memory_footprint_ = ((flags & Heap::kReduceMemoryFootprintMask) != 0); |
50 abort_incremental_marking_ = | 50 abort_incremental_marking_ = |
51 ((flags & Heap::kAbortIncrementalMarkingMask) != 0); | 51 ((flags & Heap::kAbortIncrementalMarkingMask) != 0); |
52 } | 52 } |
53 | 53 |
54 | 54 |
55 bool MarkCompactCollector::MarkObjectAndPush(HeapObject* obj) { | |
56 if (MarkObjectWithoutPush(obj)) { | |
57 marking_deque_.PushBlack(obj); | |
58 return true; | |
59 } | |
60 return false; | |
61 } | |
62 | |
63 | |
64 void MarkCompactCollector::MarkObject(HeapObject* obj, MarkBit mark_bit) { | 55 void MarkCompactCollector::MarkObject(HeapObject* obj, MarkBit mark_bit) { |
65 ASSERT(Marking::MarkBitFrom(obj) == mark_bit); | 56 ASSERT(Marking::MarkBitFrom(obj) == mark_bit); |
66 if (!mark_bit.Get()) { | 57 if (!mark_bit.Get()) { |
67 mark_bit.Set(); | 58 mark_bit.Set(); |
68 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), obj->Size()); | 59 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), obj->Size()); |
69 ProcessNewlyMarkedObject(obj); | 60 ASSERT(IsMarked(obj)); |
| 61 ASSERT(HEAP->Contains(obj)); |
| 62 marking_deque_.PushBlack(obj); |
70 } | 63 } |
71 } | 64 } |
72 | 65 |
73 | 66 |
74 bool MarkCompactCollector::MarkObjectWithoutPush(HeapObject* obj) { | |
75 MarkBit mark_bit = Marking::MarkBitFrom(obj); | |
76 if (!mark_bit.Get()) { | |
77 SetMark(obj, mark_bit); | |
78 return true; | |
79 } | |
80 return false; | |
81 } | |
82 | |
83 | |
84 void MarkCompactCollector::SetMark(HeapObject* obj, MarkBit mark_bit) { | 67 void MarkCompactCollector::SetMark(HeapObject* obj, MarkBit mark_bit) { |
85 ASSERT(!mark_bit.Get()); | 68 ASSERT(!mark_bit.Get()); |
86 ASSERT(Marking::MarkBitFrom(obj) == mark_bit); | 69 ASSERT(Marking::MarkBitFrom(obj) == mark_bit); |
87 mark_bit.Set(); | 70 mark_bit.Set(); |
88 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), obj->Size()); | 71 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), obj->Size()); |
89 if (obj->IsMap()) { | |
90 heap_->ClearCacheOnMap(Map::cast(obj)); | |
91 } | |
92 } | 72 } |
93 | 73 |
94 | 74 |
95 bool MarkCompactCollector::IsMarked(Object* obj) { | 75 bool MarkCompactCollector::IsMarked(Object* obj) { |
96 ASSERT(obj->IsHeapObject()); | 76 ASSERT(obj->IsHeapObject()); |
97 HeapObject* heap_object = HeapObject::cast(obj); | 77 HeapObject* heap_object = HeapObject::cast(obj); |
98 return Marking::MarkBitFrom(heap_object).Get(); | 78 return Marking::MarkBitFrom(heap_object).Get(); |
99 } | 79 } |
100 | 80 |
101 | 81 |
102 void MarkCompactCollector::RecordSlot(Object** anchor_slot, | 82 void MarkCompactCollector::RecordSlot(Object** anchor_slot, |
103 Object** slot, | 83 Object** slot, |
104 Object* object) { | 84 Object* object) { |
105 Page* object_page = Page::FromAddress(reinterpret_cast<Address>(object)); | 85 Page* object_page = Page::FromAddress(reinterpret_cast<Address>(object)); |
106 if (object_page->IsEvacuationCandidate() && | 86 if (object_page->IsEvacuationCandidate() && |
107 !ShouldSkipEvacuationSlotRecording(anchor_slot)) { | 87 !ShouldSkipEvacuationSlotRecording(anchor_slot)) { |
108 if (!SlotsBuffer::AddTo(&slots_buffer_allocator_, | 88 if (!SlotsBuffer::AddTo(&slots_buffer_allocator_, |
109 object_page->slots_buffer_address(), | 89 object_page->slots_buffer_address(), |
110 slot, | 90 slot, |
111 SlotsBuffer::FAIL_ON_OVERFLOW)) { | 91 SlotsBuffer::FAIL_ON_OVERFLOW)) { |
112 EvictEvacuationCandidate(object_page); | 92 EvictEvacuationCandidate(object_page); |
113 } | 93 } |
114 } | 94 } |
115 } | 95 } |
116 | 96 |
117 | 97 |
118 } } // namespace v8::internal | 98 } } // namespace v8::internal |
119 | 99 |
120 #endif // V8_MARK_COMPACT_INL_H_ | 100 #endif // V8_MARK_COMPACT_INL_H_ |
OLD | NEW |