OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include <algorithm> | 5 #include <algorithm> |
6 | 6 |
7 #include "src/v8.h" | 7 #include "src/v8.h" |
8 | 8 |
9 #include "src/counters.h" | 9 #include "src/counters.h" |
10 #include "src/heap/store-buffer-inl.h" | 10 #include "src/heap/store-buffer-inl.h" |
(...skipping 450 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
461 // We skip free space objects. | 461 // We skip free space objects. |
462 if (!heap_object->IsFiller()) { | 462 if (!heap_object->IsFiller()) { |
463 DCHECK(heap_object->IsMap()); | 463 DCHECK(heap_object->IsMap()); |
464 FindPointersToNewSpaceInRegion( | 464 FindPointersToNewSpaceInRegion( |
465 heap_object->address() + Map::kPointerFieldsBeginOffset, | 465 heap_object->address() + Map::kPointerFieldsBeginOffset, |
466 heap_object->address() + Map::kPointerFieldsEndOffset, | 466 heap_object->address() + Map::kPointerFieldsEndOffset, |
467 slot_callback); | 467 slot_callback); |
468 } | 468 } |
469 } | 469 } |
470 } else { | 470 } else { |
471 if (!page->SweepingCompleted()) { | |
472 heap_->mark_compact_collector()->SweepInParallel(page, owner); | |
473 if (!page->SweepingCompleted()) { | |
474 // We were not able to sweep that page, i.e., a concurrent | |
475 // sweeper thread currently owns this page. | |
476 // TODO(hpayer): This may introduce a huge pause here. We | |
477 // just care about finish sweeping of the scan on scavenge page. | |
478 heap_->mark_compact_collector()->EnsureSweepingCompleted(); | |
479 } | |
480 } | |
481 CHECK(page->owner() == heap_->old_space()); | 471 CHECK(page->owner() == heap_->old_space()); |
| 472 heap_->mark_compact_collector()->EnsureSweepingCompleted(page, |
| 473 owner); |
482 HeapObjectIterator iterator(page, NULL); | 474 HeapObjectIterator iterator(page, NULL); |
483 for (HeapObject* heap_object = iterator.Next(); heap_object != NULL; | 475 for (HeapObject* heap_object = iterator.Next(); heap_object != NULL; |
484 heap_object = iterator.Next()) { | 476 heap_object = iterator.Next()) { |
485 // We iterate over objects that contain new space pointers only. | 477 // We iterate over objects that contain new space pointers only. |
486 Address obj_address = heap_object->address(); | 478 Address obj_address = heap_object->address(); |
487 const int start_offset = HeapObject::kHeaderSize; | 479 const int start_offset = HeapObject::kHeaderSize; |
488 const int end_offset = heap_object->Size(); | 480 const int end_offset = heap_object->Size(); |
489 | 481 |
490 switch (heap_object->ContentType()) { | 482 switch (heap_object->ContentType()) { |
491 case HeapObjectContents::kTaggedValues: { | 483 case HeapObjectContents::kTaggedValues: { |
(...skipping 93 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
585 } | 577 } |
586 old_buffer_is_sorted_ = false; | 578 old_buffer_is_sorted_ = false; |
587 old_buffer_is_filtered_ = false; | 579 old_buffer_is_filtered_ = false; |
588 *old_top_++ = reinterpret_cast<Address>(int_addr << kPointerSizeLog2); | 580 *old_top_++ = reinterpret_cast<Address>(int_addr << kPointerSizeLog2); |
589 DCHECK(old_top_ <= old_limit_); | 581 DCHECK(old_top_ <= old_limit_); |
590 } | 582 } |
591 heap_->isolate()->counters()->store_buffer_compactions()->Increment(); | 583 heap_->isolate()->counters()->store_buffer_compactions()->Increment(); |
592 } | 584 } |
593 } // namespace internal | 585 } // namespace internal |
594 } // namespace v8 | 586 } // namespace v8 |
OLD | NEW |