Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(4)

Side by Side Diff: src/heap/spaces.h

Issue 2005723005: Reland of "[heap] Fine-grained JSArrayBuffer tracking" (Closed) Base URL: https://chromium.googlesource.com/v8/v8.git@master
Patch Set: Addressed comment Created 4 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « src/heap/scavenger.cc ('k') | src/heap/spaces.cc » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
1 // Copyright 2011 the V8 project authors. All rights reserved. 1 // Copyright 2011 the V8 project authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be 2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file. 3 // found in the LICENSE file.
4 4
5 #ifndef V8_HEAP_SPACES_H_ 5 #ifndef V8_HEAP_SPACES_H_
6 #define V8_HEAP_SPACES_H_ 6 #define V8_HEAP_SPACES_H_
7 7
8 #include <list> 8 #include <list>
9 9
10 #include "src/allocation.h" 10 #include "src/allocation.h"
11 #include "src/base/atomic-utils.h" 11 #include "src/base/atomic-utils.h"
12 #include "src/base/atomicops.h" 12 #include "src/base/atomicops.h"
13 #include "src/base/bits.h" 13 #include "src/base/bits.h"
14 #include "src/base/platform/mutex.h" 14 #include "src/base/platform/mutex.h"
15 #include "src/flags.h" 15 #include "src/flags.h"
16 #include "src/hashmap.h" 16 #include "src/hashmap.h"
17 #include "src/heap/array-buffer-tracker.h"
17 #include "src/list.h" 18 #include "src/list.h"
18 #include "src/objects.h" 19 #include "src/objects.h"
19 #include "src/utils.h" 20 #include "src/utils.h"
20 21
21 namespace v8 { 22 namespace v8 {
22 namespace internal { 23 namespace internal {
23 24
24 class AllocationInfo; 25 class AllocationInfo;
25 class AllocationObserver; 26 class AllocationObserver;
26 class CompactionSpace; 27 class CompactionSpace;
(...skipping 434 matching lines...) Expand 10 before | Expand all | Expand 10 after
461 // not be performed on that page. Sweeper threads that are done with their 462 // not be performed on that page. Sweeper threads that are done with their
462 // work will set this value and not touch the page anymore. 463 // work will set this value and not touch the page anymore.
463 // |kSweepingPending|: This page is ready for parallel sweeping. 464 // |kSweepingPending|: This page is ready for parallel sweeping.
464 // |kSweepingInProgress|: This page is currently swept by a sweeper thread. 465 // |kSweepingInProgress|: This page is currently swept by a sweeper thread.
465 enum ConcurrentSweepingState { 466 enum ConcurrentSweepingState {
466 kSweepingDone, 467 kSweepingDone,
467 kSweepingPending, 468 kSweepingPending,
468 kSweepingInProgress, 469 kSweepingInProgress,
469 }; 470 };
470 471
472 enum ArrayBufferTrackerAccessMode { kDontCreate, kCreateIfNotPresent };
473
471 // Every n write barrier invocations we go to runtime even though 474 // Every n write barrier invocations we go to runtime even though
472 // we could have handled it in generated code. This lets us check 475 // we could have handled it in generated code. This lets us check
473 // whether we have hit the limit and should do some more marking. 476 // whether we have hit the limit and should do some more marking.
474 static const int kWriteBarrierCounterGranularity = 500; 477 static const int kWriteBarrierCounterGranularity = 500;
475 478
476 static const int kPointersToHereAreInterestingMask = 479 static const int kPointersToHereAreInterestingMask =
477 1 << POINTERS_TO_HERE_ARE_INTERESTING; 480 1 << POINTERS_TO_HERE_ARE_INTERESTING;
478 481
479 static const int kPointersFromHereAreInterestingMask = 482 static const int kPointersFromHereAreInterestingMask =
480 1 << POINTERS_FROM_HERE_ARE_INTERESTING; 483 1 << POINTERS_FROM_HERE_ARE_INTERESTING;
(...skipping 35 matching lines...) Expand 10 before | Expand all | Expand 10 after
516 static const size_t kMinHeaderSize = 519 static const size_t kMinHeaderSize =
517 kWriteBarrierCounterOffset + 520 kWriteBarrierCounterOffset +
518 kIntptrSize // intptr_t write_barrier_counter_ 521 kIntptrSize // intptr_t write_barrier_counter_
519 + kPointerSize // AtomicValue high_water_mark_ 522 + kPointerSize // AtomicValue high_water_mark_
520 + kPointerSize // base::Mutex* mutex_ 523 + kPointerSize // base::Mutex* mutex_
521 + kPointerSize // base::AtomicWord concurrent_sweeping_ 524 + kPointerSize // base::AtomicWord concurrent_sweeping_
522 + 2 * kPointerSize // AtomicNumber free-list statistics 525 + 2 * kPointerSize // AtomicNumber free-list statistics
523 + kPointerSize // AtomicValue next_chunk_ 526 + kPointerSize // AtomicValue next_chunk_
524 + kPointerSize // AtomicValue prev_chunk_ 527 + kPointerSize // AtomicValue prev_chunk_
525 // FreeListCategory categories_[kNumberOfCategories] 528 // FreeListCategory categories_[kNumberOfCategories]
526 + FreeListCategory::kSize * kNumberOfCategories; 529 + FreeListCategory::kSize * kNumberOfCategories +
530 kPointerSize; // LocalArrayBufferTracker tracker_
527 531
528 // We add some more space to the computed header size to amount for missing 532 // We add some more space to the computed header size to amount for missing
529 // alignment requirements in our computation. 533 // alignment requirements in our computation.
530 // Try to get kHeaderSize properly aligned on 32-bit and 64-bit machines. 534 // Try to get kHeaderSize properly aligned on 32-bit and 64-bit machines.
531 static const size_t kHeaderSize = kMinHeaderSize; 535 static const size_t kHeaderSize = kMinHeaderSize;
532 536
533 static const int kBodyOffset = 537 static const int kBodyOffset =
534 CODE_POINTER_ALIGN(kHeaderSize + Bitmap::kSize); 538 CODE_POINTER_ALIGN(kHeaderSize + Bitmap::kSize);
535 539
536 // The start offset of the object area in a page. Aligned to both maps and 540 // The start offset of the object area in a page. Aligned to both maps and
(...skipping 98 matching lines...) Expand 10 before | Expand all | Expand 10 after
635 639
636 void AllocateOldToNewSlots(); 640 void AllocateOldToNewSlots();
637 void ReleaseOldToNewSlots(); 641 void ReleaseOldToNewSlots();
638 void AllocateOldToOldSlots(); 642 void AllocateOldToOldSlots();
639 void ReleaseOldToOldSlots(); 643 void ReleaseOldToOldSlots();
640 void AllocateTypedOldToNewSlots(); 644 void AllocateTypedOldToNewSlots();
641 void ReleaseTypedOldToNewSlots(); 645 void ReleaseTypedOldToNewSlots();
642 void AllocateTypedOldToOldSlots(); 646 void AllocateTypedOldToOldSlots();
643 void ReleaseTypedOldToOldSlots(); 647 void ReleaseTypedOldToOldSlots();
644 648
649 template <ArrayBufferTrackerAccessMode tracker_access>
650 inline LocalArrayBufferTracker* local_tracker() {
651 if (local_tracker_ == nullptr && tracker_access == kCreateIfNotPresent) {
652 local_tracker_ = new LocalArrayBufferTracker(heap_);
653 }
654 return local_tracker_;
655 }
656
657 void ReleaseLocalTracker();
658
645 Address area_start() { return area_start_; } 659 Address area_start() { return area_start_; }
646 Address area_end() { return area_end_; } 660 Address area_end() { return area_end_; }
647 int area_size() { return static_cast<int>(area_end() - area_start()); } 661 int area_size() { return static_cast<int>(area_end() - area_start()); }
648 662
649 bool CommitArea(size_t requested); 663 bool CommitArea(size_t requested);
650 664
651 // Approximate amount of physical memory committed for this chunk. 665 // Approximate amount of physical memory committed for this chunk.
652 size_t CommittedPhysicalMemory() { return high_water_mark_.Value(); } 666 size_t CommittedPhysicalMemory() { return high_water_mark_.Value(); }
653 667
654 int progress_bar() { 668 int progress_bar() {
(...skipping 162 matching lines...) Expand 10 before | Expand all | Expand 10 after
817 base::AtomicNumber<intptr_t> available_in_free_list_; 831 base::AtomicNumber<intptr_t> available_in_free_list_;
818 base::AtomicNumber<intptr_t> wasted_memory_; 832 base::AtomicNumber<intptr_t> wasted_memory_;
819 833
820 // next_chunk_ holds a pointer of type MemoryChunk 834 // next_chunk_ holds a pointer of type MemoryChunk
821 base::AtomicValue<MemoryChunk*> next_chunk_; 835 base::AtomicValue<MemoryChunk*> next_chunk_;
822 // prev_chunk_ holds a pointer of type MemoryChunk 836 // prev_chunk_ holds a pointer of type MemoryChunk
823 base::AtomicValue<MemoryChunk*> prev_chunk_; 837 base::AtomicValue<MemoryChunk*> prev_chunk_;
824 838
825 FreeListCategory categories_[kNumberOfCategories]; 839 FreeListCategory categories_[kNumberOfCategories];
826 840
841 LocalArrayBufferTracker* local_tracker_;
842
827 private: 843 private:
828 void InitializeReservedMemory() { reservation_.Reset(); } 844 void InitializeReservedMemory() { reservation_.Reset(); }
829 845
830 friend class MemoryAllocator; 846 friend class MemoryAllocator;
831 friend class MemoryChunkValidator; 847 friend class MemoryChunkValidator;
832 }; 848 };
833 849
834 // ----------------------------------------------------------------------------- 850 // -----------------------------------------------------------------------------
835 // A page is a memory chunk of a size 1MB. Large object pages may be larger. 851 // A page is a memory chunk of a size 1MB. Large object pages may be larger.
836 // 852 //
(...skipping 1451 matching lines...) Expand 10 before | Expand all | Expand 10 after
2288 // sweeper. 2304 // sweeper.
2289 virtual void RefillFreeList(); 2305 virtual void RefillFreeList();
2290 2306
2291 FreeList* free_list() { return &free_list_; } 2307 FreeList* free_list() { return &free_list_; }
2292 2308
2293 base::Mutex* mutex() { return &space_mutex_; } 2309 base::Mutex* mutex() { return &space_mutex_; }
2294 2310
2295 inline void UnlinkFreeListCategories(Page* page); 2311 inline void UnlinkFreeListCategories(Page* page);
2296 inline intptr_t RelinkFreeListCategories(Page* page); 2312 inline intptr_t RelinkFreeListCategories(Page* page);
2297 2313
2314 // Callback signature:
2315 // void Callback(Page*);
2316 template <typename Callback>
2317 void ForAllPages(Callback callback) {
2318 PageIterator it(this);
2319 while (it.has_next()) {
2320 callback(it.next());
2321 }
2322 }
2323
2298 protected: 2324 protected:
2299 // PagedSpaces that should be included in snapshots have different, i.e., 2325 // PagedSpaces that should be included in snapshots have different, i.e.,
2300 // smaller, initial pages. 2326 // smaller, initial pages.
2301 virtual bool snapshotable() { return true; } 2327 virtual bool snapshotable() { return true; }
2302 2328
2303 bool HasPages() { return anchor_.next_page() != &anchor_; } 2329 bool HasPages() { return anchor_.next_page() != &anchor_; }
2304 2330
2305 // Cleans up the space, frees all pages in this space except those belonging 2331 // Cleans up the space, frees all pages in this space except those belonging
2306 // to the initial chunk, uncommits addresses in the initial chunk. 2332 // to the initial chunk, uncommits addresses in the initial chunk.
2307 void TearDown(); 2333 void TearDown();
(...skipping 841 matching lines...) Expand 10 before | Expand all | Expand 10 after
3149 count = 0; 3175 count = 0;
3150 } 3176 }
3151 // Must be small, since an iteration is used for lookup. 3177 // Must be small, since an iteration is used for lookup.
3152 static const int kMaxComments = 64; 3178 static const int kMaxComments = 64;
3153 }; 3179 };
3154 #endif 3180 #endif
3155 } // namespace internal 3181 } // namespace internal
3156 } // namespace v8 3182 } // namespace v8
3157 3183
3158 #endif // V8_HEAP_SPACES_H_ 3184 #endif // V8_HEAP_SPACES_H_
OLDNEW
« no previous file with comments | « src/heap/scavenger.cc ('k') | src/heap/spaces.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698