OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_HEAP_MARK_COMPACT_H_ | 5 #ifndef V8_HEAP_MARK_COMPACT_H_ |
6 #define V8_HEAP_MARK_COMPACT_H_ | 6 #define V8_HEAP_MARK_COMPACT_H_ |
7 | 7 |
8 #include "src/base/bits.h" | 8 #include "src/base/bits.h" |
9 #include "src/heap/spaces.h" | 9 #include "src/heap/spaces.h" |
10 | 10 |
11 namespace v8 { | 11 namespace v8 { |
12 namespace internal { | 12 namespace internal { |
13 | 13 |
14 // Callback function, returns whether an object is alive. The heap size | 14 // Callback function, returns whether an object is alive. The heap size |
15 // of the object is returned in size. It optionally updates the offset | 15 // of the object is returned in size. It optionally updates the offset |
16 // to the first live object in the page (only used for old and map objects). | 16 // to the first live object in the page (only used for old and map objects). |
17 typedef bool (*IsAliveFunction)(HeapObject* obj, int* size, int* offset); | 17 typedef bool (*IsAliveFunction)(HeapObject* obj, int* size, int* offset); |
18 | 18 |
19 // Callback function to mark an object in a given heap. | 19 // Callback function to mark an object in a given heap. |
20 typedef void (*MarkObjectFunction)(Heap* heap, HeapObject* object); | 20 typedef void (*MarkObjectFunction)(Heap* heap, HeapObject* object); |
21 | 21 |
22 // Forward declarations. | 22 // Forward declarations. |
23 class CodeFlusher; | 23 class CodeFlusher; |
24 class MarkCompactCollector; | 24 class MarkCompactCollector; |
25 class MarkingVisitor; | 25 class MarkingVisitor; |
26 class RootMarkingVisitor; | 26 class RootMarkingVisitor; |
| 27 class SlotsBuffer; |
| 28 class SlotsBufferAllocator; |
27 | 29 |
28 | 30 |
29 class Marking : public AllStatic { | 31 class Marking : public AllStatic { |
30 public: | 32 public: |
31 INLINE(static MarkBit MarkBitFrom(Address addr)) { | 33 INLINE(static MarkBit MarkBitFrom(Address addr)) { |
32 MemoryChunk* p = MemoryChunk::FromAddress(addr); | 34 MemoryChunk* p = MemoryChunk::FromAddress(addr); |
33 return p->markbits()->MarkBitFromIndex(p->AddressToMarkbitIndex(addr)); | 35 return p->markbits()->MarkBitFromIndex(p->AddressToMarkbitIndex(addr)); |
34 } | 36 } |
35 | 37 |
36 INLINE(static MarkBit MarkBitFrom(HeapObject* obj)) { | 38 INLINE(static MarkBit MarkBitFrom(HeapObject* obj)) { |
(...skipping 214 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
251 int top_; | 253 int top_; |
252 int bottom_; | 254 int bottom_; |
253 int mask_; | 255 int mask_; |
254 bool overflowed_; | 256 bool overflowed_; |
255 bool in_use_; | 257 bool in_use_; |
256 | 258 |
257 DISALLOW_COPY_AND_ASSIGN(MarkingDeque); | 259 DISALLOW_COPY_AND_ASSIGN(MarkingDeque); |
258 }; | 260 }; |
259 | 261 |
260 | 262 |
261 class SlotsBufferAllocator { | |
262 public: | |
263 SlotsBuffer* AllocateBuffer(SlotsBuffer* next_buffer); | |
264 void DeallocateBuffer(SlotsBuffer* buffer); | |
265 | |
266 void DeallocateChain(SlotsBuffer** buffer_address); | |
267 }; | |
268 | |
269 | |
270 // SlotsBuffer records a sequence of slots that has to be updated | |
271 // after live objects were relocated from evacuation candidates. | |
272 // All slots are either untyped or typed: | |
273 // - Untyped slots are expected to contain a tagged object pointer. | |
274 // They are recorded by an address. | |
275 // - Typed slots are expected to contain an encoded pointer to a heap | |
276 // object where the way of encoding depends on the type of the slot. | |
277 // They are recorded as a pair (SlotType, slot address). | |
278 // We assume that zero-page is never mapped this allows us to distinguish | |
279 // untyped slots from typed slots during iteration by a simple comparison: | |
280 // if element of slots buffer is less than NUMBER_OF_SLOT_TYPES then it | |
281 // is the first element of typed slot's pair. | |
282 class SlotsBuffer { | |
283 public: | |
284 typedef Object** ObjectSlot; | |
285 | |
286 explicit SlotsBuffer(SlotsBuffer* next_buffer) | |
287 : idx_(0), chain_length_(1), next_(next_buffer) { | |
288 if (next_ != NULL) { | |
289 chain_length_ = next_->chain_length_ + 1; | |
290 } | |
291 } | |
292 | |
293 ~SlotsBuffer() {} | |
294 | |
295 void Add(ObjectSlot slot) { | |
296 DCHECK(0 <= idx_ && idx_ < kNumberOfElements); | |
297 #ifdef DEBUG | |
298 if (slot >= reinterpret_cast<ObjectSlot>(NUMBER_OF_SLOT_TYPES)) { | |
299 DCHECK_NOT_NULL(*slot); | |
300 } | |
301 #endif | |
302 slots_[idx_++] = slot; | |
303 } | |
304 | |
305 // Should be used for testing only. | |
306 ObjectSlot Get(intptr_t i) { | |
307 DCHECK(i >= 0 && i < kNumberOfElements); | |
308 return slots_[i]; | |
309 } | |
310 | |
311 enum SlotType { | |
312 EMBEDDED_OBJECT_SLOT, | |
313 OBJECT_SLOT, | |
314 RELOCATED_CODE_OBJECT, | |
315 CELL_TARGET_SLOT, | |
316 CODE_TARGET_SLOT, | |
317 CODE_ENTRY_SLOT, | |
318 DEBUG_TARGET_SLOT, | |
319 NUMBER_OF_SLOT_TYPES | |
320 }; | |
321 | |
322 static const char* SlotTypeToString(SlotType type) { | |
323 switch (type) { | |
324 case EMBEDDED_OBJECT_SLOT: | |
325 return "EMBEDDED_OBJECT_SLOT"; | |
326 case OBJECT_SLOT: | |
327 return "OBJECT_SLOT"; | |
328 case RELOCATED_CODE_OBJECT: | |
329 return "RELOCATED_CODE_OBJECT"; | |
330 case CELL_TARGET_SLOT: | |
331 return "CELL_TARGET_SLOT"; | |
332 case CODE_TARGET_SLOT: | |
333 return "CODE_TARGET_SLOT"; | |
334 case CODE_ENTRY_SLOT: | |
335 return "CODE_ENTRY_SLOT"; | |
336 case DEBUG_TARGET_SLOT: | |
337 return "DEBUG_TARGET_SLOT"; | |
338 case NUMBER_OF_SLOT_TYPES: | |
339 return "NUMBER_OF_SLOT_TYPES"; | |
340 } | |
341 return "UNKNOWN SlotType"; | |
342 } | |
343 | |
344 void UpdateSlots(Heap* heap); | |
345 | |
346 void UpdateSlotsWithFilter(Heap* heap); | |
347 | |
348 SlotsBuffer* next() { return next_; } | |
349 | |
350 static int SizeOfChain(SlotsBuffer* buffer) { | |
351 if (buffer == NULL) return 0; | |
352 return static_cast<int>(buffer->idx_ + | |
353 (buffer->chain_length_ - 1) * kNumberOfElements); | |
354 } | |
355 | |
356 inline bool IsFull() { return idx_ == kNumberOfElements; } | |
357 | |
358 inline bool HasSpaceForTypedSlot() { return idx_ < kNumberOfElements - 1; } | |
359 | |
360 static void UpdateSlotsRecordedIn(Heap* heap, SlotsBuffer* buffer) { | |
361 while (buffer != NULL) { | |
362 buffer->UpdateSlots(heap); | |
363 buffer = buffer->next(); | |
364 } | |
365 } | |
366 | |
367 enum AdditionMode { FAIL_ON_OVERFLOW, IGNORE_OVERFLOW }; | |
368 | |
369 static bool ChainLengthThresholdReached(SlotsBuffer* buffer) { | |
370 return buffer != NULL && buffer->chain_length_ >= kChainLengthThreshold; | |
371 } | |
372 | |
373 INLINE(static bool AddToSynchronized(SlotsBufferAllocator* allocator, | |
374 SlotsBuffer** buffer_address, | |
375 base::Mutex* buffer_mutex, | |
376 ObjectSlot slot, AdditionMode mode)) { | |
377 base::LockGuard<base::Mutex> lock_guard(buffer_mutex); | |
378 return AddTo(allocator, buffer_address, slot, mode); | |
379 } | |
380 | |
381 INLINE(static bool AddTo(SlotsBufferAllocator* allocator, | |
382 SlotsBuffer** buffer_address, ObjectSlot slot, | |
383 AdditionMode mode)) { | |
384 SlotsBuffer* buffer = *buffer_address; | |
385 if (buffer == NULL || buffer->IsFull()) { | |
386 if (mode == FAIL_ON_OVERFLOW && ChainLengthThresholdReached(buffer)) { | |
387 allocator->DeallocateChain(buffer_address); | |
388 return false; | |
389 } | |
390 buffer = allocator->AllocateBuffer(buffer); | |
391 *buffer_address = buffer; | |
392 } | |
393 buffer->Add(slot); | |
394 return true; | |
395 } | |
396 | |
397 static bool IsTypedSlot(ObjectSlot slot); | |
398 | |
399 static bool AddToSynchronized(SlotsBufferAllocator* allocator, | |
400 SlotsBuffer** buffer_address, | |
401 base::Mutex* buffer_mutex, SlotType type, | |
402 Address addr, AdditionMode mode); | |
403 | |
404 static bool AddTo(SlotsBufferAllocator* allocator, | |
405 SlotsBuffer** buffer_address, SlotType type, Address addr, | |
406 AdditionMode mode); | |
407 | |
408 // Eliminates all stale entries from the slots buffer, i.e., slots that | |
409 // are not part of live objects anymore. This method must be called after | |
410 // marking, when the whole transitive closure is known and must be called | |
411 // before sweeping when mark bits are still intact. | |
412 static void RemoveInvalidSlots(Heap* heap, SlotsBuffer* buffer); | |
413 | |
414 // Eliminate all slots that are within the given address range. | |
415 static void RemoveObjectSlots(Heap* heap, SlotsBuffer* buffer, | |
416 Address start_slot, Address end_slot); | |
417 | |
418 // Ensures that there are no invalid slots in the chain of slots buffers. | |
419 static void VerifySlots(Heap* heap, SlotsBuffer* buffer); | |
420 | |
421 static const int kNumberOfElements = 1021; | |
422 | |
423 private: | |
424 static const int kChainLengthThreshold = 15; | |
425 | |
426 intptr_t idx_; | |
427 intptr_t chain_length_; | |
428 SlotsBuffer* next_; | |
429 ObjectSlot slots_[kNumberOfElements]; | |
430 }; | |
431 | |
432 | |
433 // CodeFlusher collects candidates for code flushing during marking and | 263 // CodeFlusher collects candidates for code flushing during marking and |
434 // processes those candidates after marking has completed in order to | 264 // processes those candidates after marking has completed in order to |
435 // reset those functions referencing code objects that would otherwise | 265 // reset those functions referencing code objects that would otherwise |
436 // be unreachable. Code objects can be referenced in three ways: | 266 // be unreachable. Code objects can be referenced in three ways: |
437 // - SharedFunctionInfo references unoptimized code. | 267 // - SharedFunctionInfo references unoptimized code. |
438 // - JSFunction references either unoptimized or optimized code. | 268 // - JSFunction references either unoptimized or optimized code. |
439 // - OptimizedCodeMap references optimized code. | 269 // - OptimizedCodeMap references optimized code. |
440 // We are not allowed to flush unoptimized code for functions that got | 270 // We are not allowed to flush unoptimized code for functions that got |
441 // optimized or inlined into optimized code, because we might bailout | 271 // optimized or inlined into optimized code, because we might bailout |
442 // into the unoptimized code again during deoptimization. | 272 // into the unoptimized code again during deoptimization. |
(...skipping 139 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
582 } | 412 } |
583 | 413 |
584 INLINE(static bool IsOnEvacuationCandidate(Object* obj)) { | 414 INLINE(static bool IsOnEvacuationCandidate(Object* obj)) { |
585 return Page::FromAddress(reinterpret_cast<Address>(obj)) | 415 return Page::FromAddress(reinterpret_cast<Address>(obj)) |
586 ->IsEvacuationCandidate(); | 416 ->IsEvacuationCandidate(); |
587 } | 417 } |
588 | 418 |
589 void RecordRelocSlot(RelocInfo* rinfo, Object* target); | 419 void RecordRelocSlot(RelocInfo* rinfo, Object* target); |
590 void RecordCodeEntrySlot(HeapObject* object, Address slot, Code* target); | 420 void RecordCodeEntrySlot(HeapObject* object, Address slot, Code* target); |
591 void RecordCodeTargetPatch(Address pc, Code* target); | 421 void RecordCodeTargetPatch(Address pc, Code* target); |
| 422 void RecordSlot(HeapObject* object, Object** slot, Object* target); |
| 423 void ForceRecordSlot(HeapObject* object, Object** slot, Object* target); |
592 | 424 |
593 INLINE(void RecordSlot( | 425 void UpdateSlots(SlotsBuffer* buffer); |
594 HeapObject* object, Object** slot, Object* target, | 426 void UpdateSlotsRecordedIn(SlotsBuffer* buffer); |
595 SlotsBuffer::AdditionMode mode = SlotsBuffer::FAIL_ON_OVERFLOW)); | |
596 | 427 |
597 void MigrateObject(HeapObject* dst, HeapObject* src, int size, | 428 void MigrateObject(HeapObject* dst, HeapObject* src, int size, |
598 AllocationSpace to_old_space); | 429 AllocationSpace to_old_space); |
599 | 430 |
600 void MigrateObjectTagged(HeapObject* dst, HeapObject* src, int size); | 431 void MigrateObjectTagged(HeapObject* dst, HeapObject* src, int size); |
601 void MigrateObjectMixed(HeapObject* dst, HeapObject* src, int size); | 432 void MigrateObjectMixed(HeapObject* dst, HeapObject* src, int size); |
602 void MigrateObjectRaw(HeapObject* dst, HeapObject* src, int size); | 433 void MigrateObjectRaw(HeapObject* dst, HeapObject* src, int size); |
603 | 434 |
604 bool TryPromoteObject(HeapObject* object, int object_size); | 435 bool TryPromoteObject(HeapObject* object, int object_size); |
605 | 436 |
(...skipping 114 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
720 bool parallel_compaction_in_progress_; | 551 bool parallel_compaction_in_progress_; |
721 | 552 |
722 // Synchronize sweeper threads. | 553 // Synchronize sweeper threads. |
723 base::Semaphore pending_sweeper_jobs_semaphore_; | 554 base::Semaphore pending_sweeper_jobs_semaphore_; |
724 | 555 |
725 // Synchronize compaction threads. | 556 // Synchronize compaction threads. |
726 base::Semaphore pending_compaction_jobs_semaphore_; | 557 base::Semaphore pending_compaction_jobs_semaphore_; |
727 | 558 |
728 bool evacuation_; | 559 bool evacuation_; |
729 | 560 |
730 SlotsBufferAllocator slots_buffer_allocator_; | 561 SlotsBufferAllocator* slots_buffer_allocator_; |
731 | 562 |
732 SlotsBuffer* migration_slots_buffer_; | 563 SlotsBuffer* migration_slots_buffer_; |
733 | 564 |
734 base::Mutex migration_slots_buffer_mutex_; | 565 base::Mutex migration_slots_buffer_mutex_; |
735 | 566 |
736 // Finishes GC, performs heap verification if enabled. | 567 // Finishes GC, performs heap verification if enabled. |
737 void Finish(); | 568 void Finish(); |
738 | 569 |
739 // ----------------------------------------------------------------------- | 570 // ----------------------------------------------------------------------- |
740 // Phase 1: Marking live objects. | 571 // Phase 1: Marking live objects. |
(...skipping 251 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
992 private: | 823 private: |
993 MarkCompactCollector* collector_; | 824 MarkCompactCollector* collector_; |
994 }; | 825 }; |
995 | 826 |
996 | 827 |
997 const char* AllocationSpaceName(AllocationSpace space); | 828 const char* AllocationSpaceName(AllocationSpace space); |
998 } | 829 } |
999 } // namespace v8::internal | 830 } // namespace v8::internal |
1000 | 831 |
1001 #endif // V8_HEAP_MARK_COMPACT_H_ | 832 #endif // V8_HEAP_MARK_COMPACT_H_ |
OLD | NEW |