OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 392 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
403 if (FLAG_trace_incremental_marking) { | 403 if (FLAG_trace_incremental_marking) { |
404 PrintF("[IncrementalMarking] Start\n"); | 404 PrintF("[IncrementalMarking] Start\n"); |
405 } | 405 } |
406 ASSERT(FLAG_incremental_marking); | 406 ASSERT(FLAG_incremental_marking); |
407 ASSERT(state_ == STOPPED); | 407 ASSERT(state_ == STOPPED); |
408 | 408 |
409 ResetStepCounters(); | 409 ResetStepCounters(); |
410 | 410 |
411 if (heap_->old_pointer_space()->IsSweepingComplete() && | 411 if (heap_->old_pointer_space()->IsSweepingComplete() && |
412 heap_->old_data_space()->IsSweepingComplete()) { | 412 heap_->old_data_space()->IsSweepingComplete()) { |
413 StartMarking(); | 413 StartMarking(ALLOW_COMPACTION); |
414 } else { | 414 } else { |
415 if (FLAG_trace_incremental_marking) { | 415 if (FLAG_trace_incremental_marking) { |
416 PrintF("[IncrementalMarking] Start sweeping.\n"); | 416 PrintF("[IncrementalMarking] Start sweeping.\n"); |
417 } | 417 } |
418 state_ = SWEEPING; | 418 state_ = SWEEPING; |
419 } | 419 } |
420 | 420 |
421 heap_->new_space()->LowerInlineAllocationLimit(kAllocatedThreshold); | 421 heap_->new_space()->LowerInlineAllocationLimit(kAllocatedThreshold); |
422 } | 422 } |
423 | 423 |
424 | 424 |
425 static void MarkObjectGreyDoNotEnqueue(Object* obj) { | 425 static void MarkObjectGreyDoNotEnqueue(Object* obj) { |
426 if (obj->IsHeapObject()) { | 426 if (obj->IsHeapObject()) { |
427 HeapObject* heap_obj = HeapObject::cast(obj); | 427 HeapObject* heap_obj = HeapObject::cast(obj); |
428 MarkBit mark_bit = Marking::MarkBitFrom(HeapObject::cast(obj)); | 428 MarkBit mark_bit = Marking::MarkBitFrom(HeapObject::cast(obj)); |
429 if (Marking::IsBlack(mark_bit)) { | 429 if (Marking::IsBlack(mark_bit)) { |
430 MemoryChunk::IncrementLiveBytes(heap_obj->address(), | 430 MemoryChunk::IncrementLiveBytes(heap_obj->address(), |
431 -heap_obj->Size()); | 431 -heap_obj->Size()); |
432 } | 432 } |
433 Marking::AnyToGrey(mark_bit); | 433 Marking::AnyToGrey(mark_bit); |
434 } | 434 } |
435 } | 435 } |
436 | 436 |
437 | 437 |
438 void IncrementalMarking::StartMarking() { | 438 void IncrementalMarking::StartMarking(CompactionFlag flag) { |
439 if (FLAG_trace_incremental_marking) { | 439 if (FLAG_trace_incremental_marking) { |
440 PrintF("[IncrementalMarking] Start marking\n"); | 440 PrintF("[IncrementalMarking] Start marking\n"); |
441 } | 441 } |
442 | 442 |
443 is_compacting_ = !FLAG_never_compact && | 443 is_compacting_ = !FLAG_never_compact && (flag == ALLOW_COMPACTION) && |
444 heap_->mark_compact_collector()->StartCompaction(); | 444 heap_->mark_compact_collector()->StartCompaction(); |
445 | 445 |
446 state_ = MARKING; | 446 state_ = MARKING; |
447 | 447 |
448 RecordWriteStub::Mode mode = is_compacting_ ? | 448 RecordWriteStub::Mode mode = is_compacting_ ? |
449 RecordWriteStub::INCREMENTAL_COMPACTION : RecordWriteStub::INCREMENTAL; | 449 RecordWriteStub::INCREMENTAL_COMPACTION : RecordWriteStub::INCREMENTAL; |
450 | 450 |
451 PatchIncrementalMarkingRecordWriteStubs(heap_, mode); | 451 PatchIncrementalMarkingRecordWriteStubs(heap_, mode); |
452 | 452 |
453 EnsureMarkingDequeIsCommitted(); | 453 EnsureMarkingDequeIsCommitted(); |
(...skipping 56 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
510 HeapObject* obj = array[current]; | 510 HeapObject* obj = array[current]; |
511 ASSERT(obj->IsHeapObject()); | 511 ASSERT(obj->IsHeapObject()); |
512 current = ((current + 1) & mask); | 512 current = ((current + 1) & mask); |
513 if (heap_->InNewSpace(obj)) { | 513 if (heap_->InNewSpace(obj)) { |
514 MapWord map_word = obj->map_word(); | 514 MapWord map_word = obj->map_word(); |
515 if (map_word.IsForwardingAddress()) { | 515 if (map_word.IsForwardingAddress()) { |
516 HeapObject* dest = map_word.ToForwardingAddress(); | 516 HeapObject* dest = map_word.ToForwardingAddress(); |
517 array[new_top] = dest; | 517 array[new_top] = dest; |
518 new_top = ((new_top + 1) & mask); | 518 new_top = ((new_top + 1) & mask); |
519 ASSERT(new_top != marking_deque_.bottom()); | 519 ASSERT(new_top != marking_deque_.bottom()); |
520 ASSERT(Marking::IsGrey(Marking::MarkBitFrom(obj))); | 520 #ifdef DEBUG |
| 521 MarkBit mark_bit = Marking::MarkBitFrom(obj); |
| 522 ASSERT(Marking::IsGrey(mark_bit) || |
| 523 (obj->IsFiller() && Marking::IsWhite(mark_bit))); |
| 524 #endif |
521 } | 525 } |
522 } else if (obj->map() != filler_map) { | 526 } else if (obj->map() != filler_map) { |
523 // Skip one word filler objects that appear on the | 527 // Skip one word filler objects that appear on the |
524 // stack when we perform in place array shift. | 528 // stack when we perform in place array shift. |
525 array[new_top] = obj; | 529 array[new_top] = obj; |
526 new_top = ((new_top + 1) & mask); | 530 new_top = ((new_top + 1) & mask); |
527 ASSERT(new_top != marking_deque_.bottom()); | 531 ASSERT(new_top != marking_deque_.bottom()); |
528 ASSERT(Marking::IsGrey(Marking::MarkBitFrom(obj))); | 532 #ifdef DEBUG |
| 533 MarkBit mark_bit = Marking::MarkBitFrom(obj); |
| 534 ASSERT(Marking::IsGrey(mark_bit) || |
| 535 (obj->IsFiller() && Marking::IsWhite(mark_bit))); |
| 536 #endif |
529 } | 537 } |
530 } | 538 } |
531 marking_deque_.set_top(new_top); | 539 marking_deque_.set_top(new_top); |
532 | 540 |
533 steps_took_since_last_gc_ = 0; | 541 steps_took_since_last_gc_ = 0; |
534 steps_count_since_last_gc_ = 0; | 542 steps_count_since_last_gc_ = 0; |
535 longest_step_ = 0.0; | 543 longest_step_ = 0.0; |
536 } | 544 } |
537 | 545 |
538 | 546 |
(...skipping 150 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
689 | 697 |
690 double start = 0; | 698 double start = 0; |
691 | 699 |
692 if (FLAG_trace_incremental_marking || FLAG_trace_gc) { | 700 if (FLAG_trace_incremental_marking || FLAG_trace_gc) { |
693 start = OS::TimeCurrentMillis(); | 701 start = OS::TimeCurrentMillis(); |
694 } | 702 } |
695 | 703 |
696 if (state_ == SWEEPING) { | 704 if (state_ == SWEEPING) { |
697 if (heap_->old_pointer_space()->AdvanceSweeper(bytes_to_process) && | 705 if (heap_->old_pointer_space()->AdvanceSweeper(bytes_to_process) && |
698 heap_->old_data_space()->AdvanceSweeper(bytes_to_process)) { | 706 heap_->old_data_space()->AdvanceSweeper(bytes_to_process)) { |
699 StartMarking(); | 707 StartMarking(PREVENT_COMPACTION); |
700 } | 708 } |
701 } else if (state_ == MARKING) { | 709 } else if (state_ == MARKING) { |
702 Map* filler_map = heap_->one_pointer_filler_map(); | 710 Map* filler_map = heap_->one_pointer_filler_map(); |
703 Map* global_context_map = heap_->global_context_map(); | 711 Map* global_context_map = heap_->global_context_map(); |
704 IncrementalMarkingMarkingVisitor marking_visitor(heap_, this); | 712 IncrementalMarkingMarkingVisitor marking_visitor(heap_, this); |
705 while (!marking_deque_.IsEmpty() && bytes_to_process > 0) { | 713 while (!marking_deque_.IsEmpty() && bytes_to_process > 0) { |
706 HeapObject* obj = marking_deque_.Pop(); | 714 HeapObject* obj = marking_deque_.Pop(); |
707 | 715 |
708 // Explicitly skip one word fillers. Incremental markbit patterns are | 716 // Explicitly skip one word fillers. Incremental markbit patterns are |
709 // correct only for objects that occupy at least two words. | 717 // correct only for objects that occupy at least two words. |
710 Map* map = obj->map(); | 718 Map* map = obj->map(); |
711 if (map == filler_map) continue; | 719 if (map == filler_map) continue; |
712 | 720 |
713 ASSERT(Marking::IsGrey(Marking::MarkBitFrom(obj))); | |
714 int size = obj->SizeFromMap(map); | 721 int size = obj->SizeFromMap(map); |
715 bytes_to_process -= size; | 722 bytes_to_process -= size; |
716 MarkBit map_mark_bit = Marking::MarkBitFrom(map); | 723 MarkBit map_mark_bit = Marking::MarkBitFrom(map); |
717 if (Marking::IsWhite(map_mark_bit)) { | 724 if (Marking::IsWhite(map_mark_bit)) { |
718 WhiteToGreyAndPush(map, map_mark_bit); | 725 WhiteToGreyAndPush(map, map_mark_bit); |
719 } | 726 } |
720 | 727 |
721 // TODO(gc) switch to static visitor instead of normal visitor. | 728 // TODO(gc) switch to static visitor instead of normal visitor. |
722 if (map == global_context_map) { | 729 if (map == global_context_map) { |
723 // Global contexts have weak fields. | 730 // Global contexts have weak fields. |
724 Context* ctx = Context::cast(obj); | 731 Context* ctx = Context::cast(obj); |
725 | 732 |
726 // We will mark cache black with a separate pass | 733 // We will mark cache black with a separate pass |
727 // when we finish marking. | 734 // when we finish marking. |
728 MarkObjectGreyDoNotEnqueue(ctx->normalized_map_cache()); | 735 MarkObjectGreyDoNotEnqueue(ctx->normalized_map_cache()); |
729 | 736 |
730 VisitGlobalContext(ctx, &marking_visitor); | 737 VisitGlobalContext(ctx, &marking_visitor); |
731 } else { | 738 } else { |
732 obj->IterateBody(map->instance_type(), size, &marking_visitor); | 739 obj->IterateBody(map->instance_type(), size, &marking_visitor); |
733 } | 740 } |
734 | 741 |
735 MarkBit obj_mark_bit = Marking::MarkBitFrom(obj); | 742 MarkBit obj_mark_bit = Marking::MarkBitFrom(obj); |
736 ASSERT(!Marking::IsBlack(obj_mark_bit)); | 743 ASSERT(Marking::IsGrey(obj_mark_bit) || |
| 744 (obj->IsFiller() && Marking::IsWhite(obj_mark_bit))); |
737 Marking::MarkBlack(obj_mark_bit); | 745 Marking::MarkBlack(obj_mark_bit); |
738 MemoryChunk::IncrementLiveBytes(obj->address(), size); | 746 MemoryChunk::IncrementLiveBytes(obj->address(), size); |
739 } | 747 } |
740 if (marking_deque_.IsEmpty()) MarkingComplete(); | 748 if (marking_deque_.IsEmpty()) MarkingComplete(); |
741 } | 749 } |
742 | 750 |
743 allocated_ = 0; | 751 allocated_ = 0; |
744 | 752 |
745 steps_count_++; | 753 steps_count_++; |
746 steps_count_since_last_gc_++; | 754 steps_count_since_last_gc_++; |
(...skipping 55 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
802 bytes_rescanned_ = 0; | 810 bytes_rescanned_ = 0; |
803 allocation_marking_factor_ = kInitialAllocationMarkingFactor; | 811 allocation_marking_factor_ = kInitialAllocationMarkingFactor; |
804 } | 812 } |
805 | 813 |
806 | 814 |
807 int64_t IncrementalMarking::SpaceLeftInOldSpace() { | 815 int64_t IncrementalMarking::SpaceLeftInOldSpace() { |
808 return heap_->MaxOldGenerationSize() - heap_->PromotedSpaceSize(); | 816 return heap_->MaxOldGenerationSize() - heap_->PromotedSpaceSize(); |
809 } | 817 } |
810 | 818 |
811 } } // namespace v8::internal | 819 } } // namespace v8::internal |
OLD | NEW |