Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 /* | 1 /* |
| 2 * Copyright (C) 2013 Google Inc. All rights reserved. | 2 * Copyright (C) 2013 Google Inc. All rights reserved. |
| 3 * | 3 * |
| 4 * Redistribution and use in source and binary forms, with or without | 4 * Redistribution and use in source and binary forms, with or without |
| 5 * modification, are permitted provided that the following conditions are | 5 * modification, are permitted provided that the following conditions are |
| 6 * met: | 6 * met: |
| 7 * | 7 * |
| 8 * * Redistributions of source code must retain the above copyright | 8 * * Redistributions of source code must retain the above copyright |
| 9 * notice, this list of conditions and the following disclaimer. | 9 * notice, this list of conditions and the following disclaimer. |
| 10 * * Redistributions in binary form must reproduce the above | 10 * * Redistributions in binary form must reproduce the above |
| (...skipping 29 matching lines...) Expand all Loading... | |
| 40 #include "wtf/Assertions.h" | 40 #include "wtf/Assertions.h" |
| 41 #include "wtf/Atomics.h" | 41 #include "wtf/Atomics.h" |
| 42 #include "wtf/Forward.h" | 42 #include "wtf/Forward.h" |
| 43 | 43 |
| 44 namespace blink { | 44 namespace blink { |
| 45 | 45 |
| 46 template<typename T> class Member; | 46 template<typename T> class Member; |
| 47 template<typename T> class WeakMember; | 47 template<typename T> class WeakMember; |
| 48 template<typename T> class UntracedMember; | 48 template<typename T> class UntracedMember; |
| 49 | 49 |
| 50 // TODO(peria): Refactor following two sets of templates. | |
| 51 | |
| 50 template<typename T, bool = NeedsAdjustAndMark<T>::value> class ObjectAliveTrait ; | 52 template<typename T, bool = NeedsAdjustAndMark<T>::value> class ObjectAliveTrait ; |
| 51 | 53 |
| 52 template<typename T> | 54 template<typename T> |
| 53 class ObjectAliveTrait<T, false> { | 55 class ObjectAliveTrait<T, false> { |
| 54 public: | 56 public: |
| 55 static bool isHeapObjectAlive(T* object) | 57 static bool isHeapObjectAlive(T* object) |
| 56 { | 58 { |
| 57 static_assert(sizeof(T), "T must be fully defined"); | 59 static_assert(sizeof(T), "T must be fully defined"); |
| 58 return HeapObjectHeader::fromPayload(object)->isMarked(); | 60 return HeapObjectHeader::fromPayload(object)->isMarked(); |
| 59 } | 61 } |
| 60 }; | 62 }; |
| 61 | 63 |
| 62 template<typename T> | 64 template<typename T> |
| 63 class ObjectAliveTrait<T, true> { | 65 class ObjectAliveTrait<T, true> { |
| 64 public: | 66 public: |
| 65 static bool isHeapObjectAlive(T* object) | 67 static bool isHeapObjectAlive(T* object) |
| 66 { | 68 { |
| 67 static_assert(sizeof(T), "T must be fully defined"); | |
| 68 return object->isHeapObjectAlive(); | 69 return object->isHeapObjectAlive(); |
| 69 } | 70 } |
| 70 }; | 71 }; |
| 71 | 72 |
| 73 template<typename T, bool = IsGarbageCollectedMixin<T>::value> class HeapObjectH eaderTrait; | |
| 74 | |
| 75 template<typename T> | |
| 76 class HeapObjectHeaderTrait<T, true> { | |
| 77 public: | |
| 78 static HeapObjectHeader* heapObjectHeader(T* obj) | |
| 79 { | |
| 80 static_assert(IsFullyDefined<T>::value, "T must be fully defined."); | |
| 81 // TODO(peria): This ASSERT() is too restrictive. The ASSERT forbids | |
| 82 // to call heapObjectHeader() for an object while another | |
| 83 // (totally independent) mixin object is under construction. | |
| 84 ASSERT(!ThreadState::current()->isConstructingGCMixin()); | |
| 85 return obj->heapObjectHeader(); | |
| 86 } | |
| 87 }; | |
| 88 | |
| 89 template<typename T> | |
| 90 class HeapObjectHeaderTrait<T, false> { | |
| 91 public: | |
| 92 static HeapObjectHeader* heapObjectHeader(T* obj) | |
| 93 { | |
| 94 static_assert(IsFullyDefined<T>::value, "T must be fully defined."); | |
| 95 ASSERT(!ThreadState::current()->isConstructingGCMixin()); | |
| 96 return HeapObjectHeader::fromPayload(obj); | |
| 97 } | |
| 98 }; | |
| 99 | |
| 72 class PLATFORM_EXPORT Heap { | 100 class PLATFORM_EXPORT Heap { |
| 73 public: | 101 public: |
| 74 static void init(); | 102 static void init(); |
| 75 static void shutdown(); | 103 static void shutdown(); |
| 76 static void doShutdown(); | 104 static void doShutdown(); |
| 77 | 105 |
| 78 #if ENABLE(ASSERT) | 106 #if ENABLE(ASSERT) |
| 79 static BasePage* findPageFromAddress(Address); | 107 static BasePage* findPageFromAddress(Address); |
| 80 static BasePage* findPageFromAddress(const void* pointer) { return findPageF romAddress(reinterpret_cast<Address>(const_cast<void*>(pointer))); } | 108 static BasePage* findPageFromAddress(const void* pointer) { return findPageF romAddress(reinterpret_cast<Address>(const_cast<void*>(pointer))); } |
| 81 #endif | 109 #endif |
| 82 | 110 |
| 83 template<typename T> | 111 template<typename T> |
| 84 static inline bool isHeapObjectAlive(T* object) | 112 static inline bool isHeapObjectAlive(T* object) |
| 85 { | 113 { |
| 86 static_assert(sizeof(T), "T must be fully defined"); | 114 static_assert(sizeof(T), "T must be fully defined"); |
| 87 // The strongification of collections relies on the fact that once a | 115 // The strongification of collections relies on the fact that once a |
| 88 // collection has been strongified, there is no way that it can contain | 116 // collection has been strongified, there is no way that it can contain |
| 89 // non-live entries, so no entries will be removed. Since you can't set | 117 // non-live entries, so no entries will be removed. Since you can't set |
| 90 // the mark bit on a null pointer, that means that null pointers are | 118 // the mark bit on a null pointer, that means that null pointers are |
| 91 // always 'alive'. | 119 // always 'alive'. |
| 92 if (!object) | 120 if (!object) |
| 93 return true; | 121 return true; |
| 94 return ObjectAliveTrait<T>::isHeapObjectAlive(object); | 122 return ObjectAliveTrait<T>::isHeapObjectAlive(object); |
| 95 } | 123 } |
| 96 template<typename T> | 124 template<typename T> |
| 97 static inline bool isHeapObjectAlive(const Member<T>& member) | 125 static inline bool isHeapObjectAlive(const Member<T>& member) |
| 98 { | 126 { |
| 99 return isHeapObjectAlive(member.get()); | 127 return isHeapObjectAlive(member.unsafeGet()); |
| 100 } | 128 } |
| 101 template<typename T> | 129 template<typename T> |
| 102 static inline bool isHeapObjectAlive(const WeakMember<T>& member) | 130 static inline bool isHeapObjectAlive(const WeakMember<T>& member) |
| 103 { | 131 { |
| 104 return isHeapObjectAlive(member.get()); | 132 return isHeapObjectAlive(member.unsafeGet()); |
| 105 } | 133 } |
| 106 template<typename T> | 134 template<typename T> |
| 107 static inline bool isHeapObjectAlive(const UntracedMember<T>& member) | 135 static inline bool isHeapObjectAlive(const UntracedMember<T>& member) |
| 108 { | 136 { |
| 109 return isHeapObjectAlive(member.get()); | 137 return isHeapObjectAlive(member.unsafeGet()); |
| 110 } | 138 } |
| 111 template<typename T> | 139 template<typename T> |
| 112 static inline bool isHeapObjectAlive(const RawPtr<T>& ptr) | 140 static inline bool isHeapObjectAlive(const RawPtr<T>& ptr) |
| 113 { | 141 { |
| 114 return isHeapObjectAlive(ptr.get()); | 142 return isHeapObjectAlive(ptr.get()); |
| 115 } | 143 } |
| 116 | 144 |
| 117 // Is the finalizable GC object still alive, but slated for lazy sweeping? | 145 // Is the finalizable GC object still alive, but slated for lazy sweeping? |
| 118 // If a lazy sweep is in progress, returns true if the object was found | 146 // If a lazy sweep is in progress, returns true if the object was found |
| 119 // to be not reachable during the marking phase, but it has yet to be swept | 147 // to be not reachable during the marking phase, but it has yet to be swept |
| (...skipping 132 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 252 static size_t wrapperCount() { return acquireLoad(&s_wrapperCount); } | 280 static size_t wrapperCount() { return acquireLoad(&s_wrapperCount); } |
| 253 static size_t wrapperCountAtLastGC() { return acquireLoad(&s_wrapperCountAtL astGC); } | 281 static size_t wrapperCountAtLastGC() { return acquireLoad(&s_wrapperCountAtL astGC); } |
| 254 static void increaseCollectedWrapperCount(size_t delta) { atomicAdd(&s_colle ctedWrapperCount, static_cast<long>(delta)); } | 282 static void increaseCollectedWrapperCount(size_t delta) { atomicAdd(&s_colle ctedWrapperCount, static_cast<long>(delta)); } |
| 255 static size_t collectedWrapperCount() { return acquireLoad(&s_collectedWrapp erCount); } | 283 static size_t collectedWrapperCount() { return acquireLoad(&s_collectedWrapp erCount); } |
| 256 static size_t partitionAllocSizeAtLastGC() { return acquireLoad(&s_partition AllocSizeAtLastGC); } | 284 static size_t partitionAllocSizeAtLastGC() { return acquireLoad(&s_partition AllocSizeAtLastGC); } |
| 257 | 285 |
| 258 static double estimatedMarkingTime(); | 286 static double estimatedMarkingTime(); |
| 259 static void reportMemoryUsageHistogram(); | 287 static void reportMemoryUsageHistogram(); |
| 260 static void reportMemoryUsageForTracing(); | 288 static void reportMemoryUsageForTracing(); |
| 261 | 289 |
| 262 #if ENABLE(ASSERT) | 290 static uint32_t gcGeneration() { return s_gcGeneration; } |
|
haraken
2015/11/20 01:59:06
Can we add:
ASSERT(s_gcGeneration != gcGenerati
peria
2015/11/20 02:27:00
Done.
| |
| 263 static uint16_t gcGeneration() { return s_gcGeneration; } | |
| 264 #endif | |
| 265 | 291 |
| 266 private: | 292 private: |
| 267 // A RegionTree is a simple binary search tree of PageMemoryRegions sorted | 293 // A RegionTree is a simple binary search tree of PageMemoryRegions sorted |
| 268 // by base addresses. | 294 // by base addresses. |
| 269 class RegionTree { | 295 class RegionTree { |
| 270 public: | 296 public: |
| 271 explicit RegionTree(PageMemoryRegion* region) : m_region(region), m_left (nullptr), m_right(nullptr) { } | 297 explicit RegionTree(PageMemoryRegion* region) : m_region(region), m_left (nullptr), m_right(nullptr) { } |
| 272 ~RegionTree() | 298 ~RegionTree() |
| 273 { | 299 { |
| 274 delete m_left; | 300 delete m_left; |
| (...skipping 26 matching lines...) Expand all Loading... | |
| 301 static size_t s_allocatedSpace; | 327 static size_t s_allocatedSpace; |
| 302 static size_t s_allocatedObjectSize; | 328 static size_t s_allocatedObjectSize; |
| 303 static size_t s_objectSizeAtLastGC; | 329 static size_t s_objectSizeAtLastGC; |
| 304 static size_t s_markedObjectSize; | 330 static size_t s_markedObjectSize; |
| 305 static size_t s_markedObjectSizeAtLastCompleteSweep; | 331 static size_t s_markedObjectSizeAtLastCompleteSweep; |
| 306 static size_t s_wrapperCount; | 332 static size_t s_wrapperCount; |
| 307 static size_t s_wrapperCountAtLastGC; | 333 static size_t s_wrapperCountAtLastGC; |
| 308 static size_t s_collectedWrapperCount; | 334 static size_t s_collectedWrapperCount; |
| 309 static size_t s_partitionAllocSizeAtLastGC; | 335 static size_t s_partitionAllocSizeAtLastGC; |
| 310 static double s_estimatedMarkingTimePerByte; | 336 static double s_estimatedMarkingTimePerByte; |
| 311 #if ENABLE(ASSERT) | 337 static uint32_t s_gcGeneration; |
| 312 static uint16_t s_gcGeneration; | |
| 313 #endif | |
| 314 | 338 |
| 315 friend class ThreadState; | 339 friend class ThreadState; |
| 316 }; | 340 }; |
| 317 | 341 |
| 318 template<typename T> | 342 template<typename T> |
| 319 struct IsEagerlyFinalizedType { | 343 struct IsEagerlyFinalizedType { |
| 320 private: | 344 private: |
| 321 typedef char YesType; | 345 typedef char YesType; |
| 322 struct NoType { | 346 struct NoType { |
| 323 char padding[8]; | 347 char padding[8]; |
| (...skipping 124 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 448 #define EAGERLY_FINALIZE_WILL_BE_REMOVED() EAGERLY_FINALIZE() | 472 #define EAGERLY_FINALIZE_WILL_BE_REMOVED() EAGERLY_FINALIZE() |
| 449 #else | 473 #else |
| 450 #define EAGERLY_FINALIZE_WILL_BE_REMOVED() | 474 #define EAGERLY_FINALIZE_WILL_BE_REMOVED() |
| 451 #endif | 475 #endif |
| 452 | 476 |
| 453 inline Address Heap::allocateOnHeapIndex(ThreadState* state, size_t size, int he apIndex, size_t gcInfoIndex) | 477 inline Address Heap::allocateOnHeapIndex(ThreadState* state, size_t size, int he apIndex, size_t gcInfoIndex) |
| 454 { | 478 { |
| 455 ASSERT(state->isAllocationAllowed()); | 479 ASSERT(state->isAllocationAllowed()); |
| 456 ASSERT(heapIndex != BlinkGC::LargeObjectHeapIndex); | 480 ASSERT(heapIndex != BlinkGC::LargeObjectHeapIndex); |
| 457 NormalPageHeap* heap = static_cast<NormalPageHeap*>(state->heap(heapIndex)); | 481 NormalPageHeap* heap = static_cast<NormalPageHeap*>(state->heap(heapIndex)); |
| 458 return heap->allocateObject(allocationSizeFromSize(size), gcInfoIndex); | 482 return heap->allocateObject(allocationSizeFromSize(size), gcInfoIndex, gcGen eration()); |
| 459 } | 483 } |
| 460 | 484 |
| 461 template<typename T> | 485 template<typename T> |
| 462 Address Heap::allocate(size_t size, bool eagerlySweep) | 486 Address Heap::allocate(size_t size, bool eagerlySweep) |
| 463 { | 487 { |
| 464 ThreadState* state = ThreadStateFor<ThreadingTrait<T>::Affinity>::state(); | 488 ThreadState* state = ThreadStateFor<ThreadingTrait<T>::Affinity>::state(); |
| 465 return Heap::allocateOnHeapIndex(state, size, eagerlySweep ? BlinkGC::EagerS weepHeapIndex : Heap::heapIndexForObjectSize(size), GCInfoTrait<T>::index()); | 489 return Heap::allocateOnHeapIndex(state, size, eagerlySweep ? BlinkGC::EagerS weepHeapIndex : Heap::heapIndexForObjectSize(size), GCInfoTrait<T>::index()); |
| 466 } | 490 } |
| 467 | 491 |
| 468 template<typename T> | 492 template<typename T> |
| (...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 503 void VisitorHelper<Derived>::handleWeakCell(Visitor* self, void* object) | 527 void VisitorHelper<Derived>::handleWeakCell(Visitor* self, void* object) |
| 504 { | 528 { |
| 505 T** cell = reinterpret_cast<T**>(object); | 529 T** cell = reinterpret_cast<T**>(object); |
| 506 if (*cell && !ObjectAliveTrait<T>::isHeapObjectAlive(*cell)) | 530 if (*cell && !ObjectAliveTrait<T>::isHeapObjectAlive(*cell)) |
| 507 *cell = nullptr; | 531 *cell = nullptr; |
| 508 } | 532 } |
| 509 | 533 |
| 510 } // namespace blink | 534 } // namespace blink |
| 511 | 535 |
| 512 #endif // Heap_h | 536 #endif // Heap_h |
| OLD | NEW |