Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(325)

Side by Side Diff: runtime/vm/heap.cc

Issue 2951333002: Moves the top_ and end_ words of the Scavenger into mutator thread. (Closed)
Patch Set: Address comments from CL Created 3 years, 5 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
1 // Copyright (c) 2012, the Dart project authors. Please see the AUTHORS file 1 // Copyright (c) 2012, the Dart project authors. Please see the AUTHORS file
2 // for details. All rights reserved. Use of this source code is governed by a 2 // for details. All rights reserved. Use of this source code is governed by a
3 // BSD-style license that can be found in the LICENSE file. 3 // BSD-style license that can be found in the LICENSE file.
4 4
5 #include "vm/heap.h" 5 #include "vm/heap.h"
6 6
7 #include "platform/assert.h" 7 #include "platform/assert.h"
8 #include "platform/utils.h" 8 #include "platform/utils.h"
9 #include "vm/flags.h" 9 #include "vm/flags.h"
10 #include "vm/isolate.h" 10 #include "vm/isolate.h"
(...skipping 45 matching lines...) Expand 10 before | Expand all | Expand 10 after
56 delete new_weak_tables_[sel]; 56 delete new_weak_tables_[sel];
57 delete old_weak_tables_[sel]; 57 delete old_weak_tables_[sel];
58 } 58 }
59 } 59 }
60 60
61 61
62 uword Heap::AllocateNew(intptr_t size) { 62 uword Heap::AllocateNew(intptr_t size) {
63 ASSERT(Thread::Current()->no_safepoint_scope_depth() == 0); 63 ASSERT(Thread::Current()->no_safepoint_scope_depth() == 0);
64 // Currently, only the Dart thread may allocate in new space. 64 // Currently, only the Dart thread may allocate in new space.
65 isolate()->AssertCurrentThreadIsMutator(); 65 isolate()->AssertCurrentThreadIsMutator();
66 uword addr = new_space_.TryAllocate(size); 66 uword addr = new_space_.TryAllocateInTLAB(Thread::Current(), size);
67 if (addr == 0) { 67 if (addr == 0) {
68 // This call to CollectGarbage might end up "reusing" a collection spawned 68 // This call to CollectGarbage might end up "reusing" a collection spawned
69 // from a different thread and will be racing to allocate the requested 69 // from a different thread and will be racing to allocate the requested
70 // memory with other threads being released after the collection. 70 // memory with other threads being released after the collection.
71 CollectGarbage(kNew); 71 CollectGarbage(kNew);
72 addr = new_space_.TryAllocate(size); 72 addr = new_space_.TryAllocateInTLAB(Thread::Current(), size);
73 if (addr == 0) { 73 if (addr == 0) {
74 return AllocateOld(size, HeapPage::kData); 74 return AllocateOld(size, HeapPage::kData);
75 } 75 }
76 } 76 }
77 return addr; 77 return addr;
78 } 78 }
79 79
80 80
81 uword Heap::AllocateOld(intptr_t size, HeapPage::PageType type) { 81 uword Heap::AllocateOld(intptr_t size, HeapPage::PageType type) {
82 ASSERT(Thread::Current()->no_safepoint_scope_depth() == 0); 82 ASSERT(Thread::Current()->no_safepoint_scope_depth() == 0);
(...skipping 158 matching lines...) Expand 10 before | Expand all | Expand 10 after
241 #endif 241 #endif
242 ASSERT(old_space_->tasks() == 1); 242 ASSERT(old_space_->tasks() == 1);
243 old_space_->set_tasks(0); 243 old_space_->set_tasks(0);
244 ml.NotifyAll(); 244 ml.NotifyAll();
245 } 245 }
246 246
247 247
248 void Heap::IterateObjects(ObjectVisitor* visitor) const { 248 void Heap::IterateObjects(ObjectVisitor* visitor) const {
249 // The visitor must not allocate from the heap. 249 // The visitor must not allocate from the heap.
250 NoSafepointScope no_safepoint_scope_; 250 NoSafepointScope no_safepoint_scope_;
251 IterateNewObjects(visitor);
252 IterateOldObjects(visitor);
253 }
254
255
256 void Heap::IterateNewObjects(ObjectVisitor* visitor) const {
251 new_space_.VisitObjects(visitor); 257 new_space_.VisitObjects(visitor);
252 IterateOldObjects(visitor);
253 } 258 }
254 259
255 260
256 void Heap::IterateOldObjects(ObjectVisitor* visitor) const { 261 void Heap::IterateOldObjects(ObjectVisitor* visitor) const {
257 HeapIterationScope heap_iteration_scope; 262 HeapIterationScope heap_iteration_scope;
258 old_space_.VisitObjects(visitor); 263 old_space_.VisitObjects(visitor);
259 } 264 }
260 265
261 266
262 void Heap::IterateOldObjectsNoImagePages(ObjectVisitor* visitor) const { 267 void Heap::IterateOldObjectsNoImagePages(ObjectVisitor* visitor) const {
(...skipping 102 matching lines...) Expand 10 before | Expand all | Expand 10 after
365 #endif 370 #endif
366 371
367 372
368 void Heap::EvacuateNewSpace(Thread* thread, GCReason reason) { 373 void Heap::EvacuateNewSpace(Thread* thread, GCReason reason) {
369 ASSERT(reason == kFull); 374 ASSERT(reason == kFull);
370 if (BeginNewSpaceGC(thread)) { 375 if (BeginNewSpaceGC(thread)) {
371 RecordBeforeGC(kNew, kFull); 376 RecordBeforeGC(kNew, kFull);
372 VMTagScope tagScope(thread, VMTag::kGCNewSpaceTagId); 377 VMTagScope tagScope(thread, VMTag::kGCNewSpaceTagId);
373 TIMELINE_FUNCTION_GC_DURATION(thread, "EvacuateNewGeneration"); 378 TIMELINE_FUNCTION_GC_DURATION(thread, "EvacuateNewGeneration");
374 NOT_IN_PRODUCT(UpdateClassHeapStatsBeforeGC(kNew)); 379 NOT_IN_PRODUCT(UpdateClassHeapStatsBeforeGC(kNew));
375 new_space_.Evacuate(); 380 new_space_.Evacuate(isolate()->mutator_thread());
376 NOT_IN_PRODUCT(isolate()->class_table()->UpdatePromoted()); 381 NOT_IN_PRODUCT(isolate()->class_table()->UpdatePromoted());
377 RecordAfterGC(kNew); 382 RecordAfterGC(kNew);
378 PrintStats(); 383 PrintStats();
379 NOT_IN_PRODUCT(PrintStatsToTimeline(&tds)); 384 NOT_IN_PRODUCT(PrintStatsToTimeline(&tds));
380 EndNewSpaceGC(); 385 EndNewSpaceGC();
381 } 386 }
382 } 387 }
383 388
384 389
385 void Heap::CollectNewSpaceGarbage(Thread* thread, 390 void Heap::CollectNewSpaceGarbage(Thread* thread,
(...skipping 116 matching lines...) Expand 10 before | Expand all | Expand 10 after
502 } 507 }
503 508
504 509
505 void Heap::WriteProtect(bool read_only) { 510 void Heap::WriteProtect(bool read_only) {
506 read_only_ = read_only; 511 read_only_ = read_only;
507 new_space_.WriteProtect(read_only); 512 new_space_.WriteProtect(read_only);
508 old_space_.WriteProtect(read_only); 513 old_space_.WriteProtect(read_only);
509 } 514 }
510 515
511 516
512 intptr_t Heap::TopOffset(Heap::Space space) {
513 if (space == kNew) {
514 return OFFSET_OF(Heap, new_space_) + Scavenger::top_offset();
515 } else {
516 ASSERT(space == kOld);
517 return OFFSET_OF(Heap, old_space_) + PageSpace::top_offset();
518 }
519 }
520
521
522 intptr_t Heap::EndOffset(Heap::Space space) {
523 if (space == kNew) {
524 return OFFSET_OF(Heap, new_space_) + Scavenger::end_offset();
525 } else {
526 ASSERT(space == kOld);
527 return OFFSET_OF(Heap, old_space_) + PageSpace::end_offset();
528 }
529 }
530
531
532 void Heap::Init(Isolate* isolate, 517 void Heap::Init(Isolate* isolate,
533 intptr_t max_new_gen_words, 518 intptr_t max_new_gen_words,
534 intptr_t max_old_gen_words, 519 intptr_t max_old_gen_words,
535 intptr_t max_external_words) { 520 intptr_t max_external_words) {
536 ASSERT(isolate->heap() == NULL); 521 ASSERT(isolate->heap() == NULL);
537 Heap* heap = new Heap(isolate, max_new_gen_words, max_old_gen_words, 522 Heap* heap = new Heap(isolate, max_new_gen_words, max_old_gen_words,
538 max_external_words); 523 max_external_words);
539 isolate->set_heap(heap); 524 isolate->set_heap(heap);
540 } 525 }
541 526
(...skipping 58 matching lines...) Expand 10 before | Expand all | Expand 10 after
600 585
601 586
602 bool Heap::Verify(MarkExpectation mark_expectation) const { 587 bool Heap::Verify(MarkExpectation mark_expectation) const {
603 HeapIterationScope heap_iteration_scope; 588 HeapIterationScope heap_iteration_scope;
604 return VerifyGC(mark_expectation); 589 return VerifyGC(mark_expectation);
605 } 590 }
606 591
607 592
608 bool Heap::VerifyGC(MarkExpectation mark_expectation) const { 593 bool Heap::VerifyGC(MarkExpectation mark_expectation) const {
609 StackZone stack_zone(Thread::Current()); 594 StackZone stack_zone(Thread::Current());
595
596 // Synchronize the top_ of the heap's new space with the thread's top_
597 new_space_.FlushTLS();
598
610 ObjectSet* allocated_set = 599 ObjectSet* allocated_set =
611 CreateAllocatedObjectSet(stack_zone.GetZone(), mark_expectation); 600 CreateAllocatedObjectSet(stack_zone.GetZone(), mark_expectation);
612 VerifyPointersVisitor visitor(isolate(), allocated_set); 601 VerifyPointersVisitor visitor(isolate(), allocated_set);
613 VisitObjectPointers(&visitor); 602 VisitObjectPointers(&visitor);
614 603
615 // Only returning a value so that Heap::Validate can be called from an ASSERT. 604 // Only returning a value so that Heap::Validate can be called from an ASSERT.
616 return true; 605 return true;
617 } 606 }
618 607
619 608
(...skipping 278 matching lines...) Expand 10 before | Expand all | Expand 10 after
898 Dart::vm_isolate()->heap()->WriteProtect(false); 887 Dart::vm_isolate()->heap()->WriteProtect(false);
899 } 888 }
900 889
901 890
902 WritableVMIsolateScope::~WritableVMIsolateScope() { 891 WritableVMIsolateScope::~WritableVMIsolateScope() {
903 ASSERT(Dart::vm_isolate()->heap()->UsedInWords(Heap::kNew) == 0); 892 ASSERT(Dart::vm_isolate()->heap()->UsedInWords(Heap::kNew) == 0);
904 Dart::vm_isolate()->heap()->WriteProtect(true); 893 Dart::vm_isolate()->heap()->WriteProtect(true);
905 } 894 }
906 895
907 } // namespace dart 896 } // namespace dart
OLDNEW
« no previous file with comments | « runtime/vm/heap.h ('k') | runtime/vm/intrinsifier_arm.cc » ('j') | runtime/vm/scavenger.h » ('J')

Powered by Google App Engine
This is Rietveld 408576698