Index: src/heap/spaces.h |
diff --git a/src/heap/spaces.h b/src/heap/spaces.h |
index 0a1fcc871ef755bf2e81f44c14bedd79c19a6c77..791ee6c32ac410b63ff4de43ab2d27eec113a4dd 100644 |
--- a/src/heap/spaces.h |
+++ b/src/heap/spaces.h |
@@ -4,8 +4,6 @@ |
#ifndef V8_HEAP_SPACES_H_ |
#define V8_HEAP_SPACES_H_ |
- |
-#include <list> |
#include "src/allocation.h" |
#include "src/atomic-utils.h" |
@@ -443,10 +441,6 @@ |
// still has to be performed. |
PRE_FREED, |
- // |POOLED|: When actually freeing this chunk, only uncommit and do not |
- // give up the reservation as we still reuse the chunk at some point. |
- POOLED, |
- |
// |COMPACTION_WAS_ABORTED|: Indicates that the compaction in this page |
// has been aborted and needs special handling by the sweeper. |
COMPACTION_WAS_ABORTED, |
@@ -1262,91 +1256,14 @@ |
// A space acquires chunks of memory from the operating system. The memory |
// allocator allocated and deallocates pages for the paged heap spaces and large |
// pages for large object space. |
+// |
+// Each space has to manage it's own pages. |
+// |
class MemoryAllocator { |
public: |
- // Unmapper takes care of concurrently unmapping and uncommitting memory |
- // chunks. |
- class Unmapper { |
- public: |
- class UnmapFreeMemoryTask; |
- |
- explicit Unmapper(MemoryAllocator* allocator) |
- : allocator_(allocator), |
- pending_unmapping_tasks_semaphore_(0), |
- concurrent_unmapping_tasks_active_(0) {} |
- |
- void AddMemoryChunkSafe(MemoryChunk* chunk) { |
- if ((chunk->size() == Page::kPageSize) && |
- (chunk->executable() == EXECUTABLE)) { |
- AddMemoryChunkSafe<kRegular>(chunk); |
- } else { |
- AddMemoryChunkSafe<kNonRegular>(chunk); |
- } |
- } |
- |
- MemoryChunk* TryGetPooledMemoryChunkSafe() { |
- // Procedure: |
- // (1) Try to get a chunk that was declared as pooled and already has |
- // been uncommitted. |
- // (2) Try to steal any memory chunk of kPageSize that would've been |
- // unmapped. |
- MemoryChunk* chunk = GetMemoryChunkSafe<kPooled>(); |
- if (chunk == nullptr) { |
- chunk = GetMemoryChunkSafe<kRegular>(); |
- if (chunk != nullptr) { |
- // For stolen chunks we need to manually free any allocated memory. |
- chunk->ReleaseAllocatedMemory(); |
- } |
- } |
- return chunk; |
- } |
- |
- void FreeQueuedChunks(); |
- bool WaitUntilCompleted(); |
- |
- private: |
- enum ChunkQueueType { |
- kRegular, // Pages of kPageSize that do not live in a CodeRange and |
- // can thus be used for stealing. |
- kNonRegular, // Large chunks and executable chunks. |
- kPooled, // Pooled chunks, already uncommited and ready for reuse. |
- kNumberOfChunkQueues, |
- }; |
- |
- template <ChunkQueueType type> |
- void AddMemoryChunkSafe(MemoryChunk* chunk) { |
- base::LockGuard<base::Mutex> guard(&mutex_); |
- chunks_[type].push_back(chunk); |
- } |
- |
- template <ChunkQueueType type> |
- MemoryChunk* GetMemoryChunkSafe() { |
- base::LockGuard<base::Mutex> guard(&mutex_); |
- if (chunks_[type].empty()) return nullptr; |
- MemoryChunk* chunk = chunks_[type].front(); |
- chunks_[type].pop_front(); |
- return chunk; |
- } |
- |
- void PerformFreeMemoryOnQueuedChunks(); |
- |
- base::Mutex mutex_; |
- MemoryAllocator* allocator_; |
- std::list<MemoryChunk*> chunks_[kNumberOfChunkQueues]; |
- base::Semaphore pending_unmapping_tasks_semaphore_; |
- intptr_t concurrent_unmapping_tasks_active_; |
- |
- friend class MemoryAllocator; |
- }; |
- |
enum AllocationMode { |
kRegular, |
kPooled, |
- }; |
- enum FreeMode { |
- kFull, |
- kPreFreeAndQueue, |
- kPooledAndQueue, |
}; |
explicit MemoryAllocator(Isolate* isolate); |
@@ -1366,7 +1283,16 @@ |
PageType* AllocatePage(intptr_t size, SpaceType* owner, |
Executability executable); |
- template <MemoryAllocator::FreeMode mode = kFull> |
+ // PreFree logically frees the object, i.e., it takes care of the size |
+ // bookkeeping and calls the allocation callback. |
+ void PreFreeMemory(MemoryChunk* chunk); |
+ |
+ // FreeMemory can be called concurrently when PreFree was executed before. |
+ void PerformFreeMemory(MemoryChunk* chunk); |
+ |
+ // Free is a wrapper method. For kRegular AllocationMode it calls PreFree and |
+ // PerformFreeMemory together. For kPooled it will dispatch to pooled free. |
+ template <MemoryAllocator::AllocationMode mode = kRegular> |
void Free(MemoryChunk* chunk); |
// Returns allocated spaces in bytes. |
@@ -1472,20 +1398,15 @@ |
size_t reserved_size); |
CodeRange* code_range() { return code_range_; } |
- Unmapper* unmapper() { return &unmapper_; } |
private: |
- // PreFree logically frees the object, i.e., it takes care of the size |
- // bookkeeping and calls the allocation callback. |
- void PreFreeMemory(MemoryChunk* chunk); |
- |
- // FreeMemory can be called concurrently when PreFree was executed before. |
- void PerformFreeMemory(MemoryChunk* chunk); |
- |
// See AllocatePage for public interface. Note that currently we only support |
// pools for NOT_EXECUTABLE pages of size MemoryChunk::kPageSize. |
template <typename SpaceType> |
MemoryChunk* AllocatePagePooled(SpaceType* owner); |
+ |
+ // Free that chunk into the pool. |
+ void FreePooled(MemoryChunk* chunk); |
Isolate* isolate_; |
@@ -1542,8 +1463,9 @@ |
} while ((high > ptr) && !highest_ever_allocated_.TrySetValue(ptr, high)); |
} |
+ List<MemoryChunk*> chunk_pool_; |
+ |
base::VirtualMemory last_chunk_; |
- Unmapper unmapper_; |
friend class TestCodeRangeScope; |