Index: src/heap/store-buffer.cc |
diff --git a/src/heap/store-buffer.cc b/src/heap/store-buffer.cc |
index ffd5d96c5c663ce9f98fdc723352da07d209679c..a982eb3c40280f96d883c870161363f1a2656ae9 100644 |
--- a/src/heap/store-buffer.cc |
+++ b/src/heap/store-buffer.cc |
@@ -16,135 +16,62 @@ |
namespace internal { |
StoreBuffer::StoreBuffer(Heap* heap) |
- : heap_(heap), top_(nullptr), current_(0), virtual_memory_(nullptr) { |
- for (int i = 0; i < kStoreBuffers; i++) { |
- start_[i] = nullptr; |
- limit_[i] = nullptr; |
- lazy_top_[i] = nullptr; |
- } |
- task_running_ = false; |
-} |
+ : heap_(heap), |
+ top_(nullptr), |
+ start_(nullptr), |
+ limit_(nullptr), |
+ virtual_memory_(nullptr) {} |
void StoreBuffer::SetUp() { |
// Allocate 3x the buffer size, so that we can start the new store buffer |
// aligned to 2x the size. This lets us use a bit test to detect the end of |
// the area. |
- virtual_memory_ = new base::VirtualMemory(kStoreBufferSize * 3); |
+ virtual_memory_ = new base::VirtualMemory(kStoreBufferSize * 2); |
uintptr_t start_as_int = |
reinterpret_cast<uintptr_t>(virtual_memory_->address()); |
- start_[0] = |
- reinterpret_cast<Address*>(RoundUp(start_as_int, kStoreBufferSize)); |
- limit_[0] = start_[0] + (kStoreBufferSize / kPointerSize); |
- start_[1] = limit_[0]; |
- limit_[1] = start_[1] + (kStoreBufferSize / kPointerSize); |
+ start_ = reinterpret_cast<Address*>(RoundUp(start_as_int, kStoreBufferSize)); |
+ limit_ = start_ + (kStoreBufferSize / kPointerSize); |
+ DCHECK(reinterpret_cast<Address>(start_) >= virtual_memory_->address()); |
+ DCHECK(reinterpret_cast<Address>(limit_) >= virtual_memory_->address()); |
Address* vm_limit = reinterpret_cast<Address*>( |
reinterpret_cast<char*>(virtual_memory_->address()) + |
virtual_memory_->size()); |
+ DCHECK(start_ <= vm_limit); |
+ DCHECK(limit_ <= vm_limit); |
+ USE(vm_limit); |
+ DCHECK((reinterpret_cast<uintptr_t>(limit_) & kStoreBufferMask) == 0); |
- USE(vm_limit); |
- for (int i = 0; i < kStoreBuffers; i++) { |
- DCHECK(reinterpret_cast<Address>(start_[i]) >= virtual_memory_->address()); |
- DCHECK(reinterpret_cast<Address>(limit_[i]) >= virtual_memory_->address()); |
- DCHECK(start_[i] <= vm_limit); |
- DCHECK(limit_[i] <= vm_limit); |
- DCHECK((reinterpret_cast<uintptr_t>(limit_[i]) & kStoreBufferMask) == 0); |
- } |
- |
- if (!virtual_memory_->Commit(reinterpret_cast<Address>(start_[0]), |
- kStoreBufferSize * kStoreBuffers, |
+ if (!virtual_memory_->Commit(reinterpret_cast<Address>(start_), |
+ kStoreBufferSize, |
false)) { // Not executable. |
V8::FatalProcessOutOfMemory("StoreBuffer::SetUp"); |
} |
- current_ = 0; |
- top_ = start_[current_]; |
+ top_ = start_; |
} |
void StoreBuffer::TearDown() { |
delete virtual_memory_; |
- top_ = nullptr; |
- for (int i = 0; i < kStoreBuffers; i++) { |
- start_[i] = nullptr; |
- limit_[i] = nullptr; |
- lazy_top_[i] = nullptr; |
- } |
+ top_ = start_ = limit_ = nullptr; |
} |
void StoreBuffer::StoreBufferOverflow(Isolate* isolate) { |
- isolate->heap()->store_buffer()->FlipStoreBuffers(); |
+ isolate->heap()->store_buffer()->MoveEntriesToRememberedSet(); |
isolate->counters()->store_buffer_overflows()->Increment(); |
} |
-void StoreBuffer::FlipStoreBuffers() { |
- base::LockGuard<base::Mutex> guard(&mutex_); |
- int other = (current_ + 1) % kStoreBuffers; |
- MoveEntriesToRememberedSet(other); |
- lazy_top_[current_] = top_; |
- current_ = other; |
- top_ = start_[current_]; |
- |
- if (!task_running_) { |
- task_running_ = true; |
- Task* task = new Task(heap_->isolate(), this); |
- V8::GetCurrentPlatform()->CallOnBackgroundThread( |
- task, v8::Platform::kShortRunningTask); |
- } |
-} |
- |
-void StoreBuffer::MoveEntriesToRememberedSet(int index) { |
- if (!lazy_top_[index]) return; |
- DCHECK_GE(index, 0); |
- DCHECK_LT(index, kStoreBuffers); |
- for (Address* current = start_[index]; current < lazy_top_[index]; |
- current++) { |
+void StoreBuffer::MoveEntriesToRememberedSet() { |
+ if (top_ == start_) return; |
+ DCHECK(top_ <= limit_); |
+ for (Address* current = start_; current < top_; current++) { |
DCHECK(!heap_->code_space()->Contains(*current)); |
Address addr = *current; |
Page* page = Page::FromAnyPointerAddress(heap_, addr); |
- if (IsDeletionAddress(addr)) { |
- current++; |
- Address end = *current; |
- DCHECK(!IsDeletionAddress(end)); |
- addr = UnmarkDeletionAddress(addr); |
- if (end) { |
- RememberedSet<OLD_TO_NEW>::RemoveRange(page, addr, end, |
- SlotSet::PREFREE_EMPTY_BUCKETS); |
- } else { |
- RememberedSet<OLD_TO_NEW>::Remove(page, addr); |
- } |
- } else { |
- DCHECK(!IsDeletionAddress(addr)); |
- RememberedSet<OLD_TO_NEW>::Insert(page, addr); |
- } |
+ RememberedSet<OLD_TO_NEW>::Insert(page, addr); |
} |
- lazy_top_[index] = nullptr; |
-} |
- |
-void StoreBuffer::MoveAllEntriesToRememberedSet() { |
- base::LockGuard<base::Mutex> guard(&mutex_); |
- int other = (current_ + 1) % kStoreBuffers; |
- MoveEntriesToRememberedSet(other); |
- lazy_top_[current_] = top_; |
- MoveEntriesToRememberedSet(current_); |
- top_ = start_[current_]; |
-} |
- |
-void StoreBuffer::ConcurrentlyProcessStoreBuffer() { |
- base::LockGuard<base::Mutex> guard(&mutex_); |
- int other = (current_ + 1) % kStoreBuffers; |
- MoveEntriesToRememberedSet(other); |
- task_running_ = false; |
-} |
- |
-void StoreBuffer::DeleteEntry(Address start, Address end) { |
- if (top_ + sizeof(Address) * 2 > limit_[current_]) { |
- StoreBufferOverflow(heap_->isolate()); |
- } |
- *top_ = MarkDeletionAddress(start); |
- top_++; |
- *top_ = end; |
- top_++; |
+ top_ = start_; |
} |
} // namespace internal |