Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(376)

Unified Diff: src/heap/spaces.cc

Issue 1406533002: Revert of [heap] Divide available memory upon compaction tasks (Closed) Base URL: https://chromium.googlesource.com/v8/v8.git@counters
Patch Set: Created 5 years, 2 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/heap/spaces.h ('k') | test/cctest/heap-tester.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/heap/spaces.cc
diff --git a/src/heap/spaces.cc b/src/heap/spaces.cc
index 5c1784f2cc80d74512805ba529dcd2d0a32e7dae..50a213e8ae4f582a3a684b4639c73894616fb851 100644
--- a/src/heap/spaces.cc
+++ b/src/heap/spaces.cc
@@ -982,101 +982,6 @@
}
-void PagedSpace::AddMemory(Address start, intptr_t size) {
- accounting_stats_.ExpandSpace(static_cast<int>(size));
- Free(start, static_cast<int>(size));
-}
-
-
-FreeSpace* PagedSpace::TryRemoveMemory(intptr_t size_in_bytes) {
- FreeSpace* free_space = free_list()->TryRemoveMemory(size_in_bytes);
- if (free_space != nullptr) {
- accounting_stats_.DecreaseCapacity(free_space->size());
- }
- return free_space;
-}
-
-
-void PagedSpace::DivideUponCompactionSpaces(CompactionSpaceCollection** other,
- int num, intptr_t limit) {
- DCHECK_GT(num, 0);
- DCHECK(other != nullptr);
-
- if (limit == 0) limit = std::numeric_limits<intptr_t>::max();
-
- EmptyAllocationInfo();
-
- bool memory_available = true;
- bool spaces_need_memory = true;
- FreeSpace* node = nullptr;
- CompactionSpace* current_space = nullptr;
- // Iterate over spaces and memory as long as we have memory and there are
- // spaces in need of some.
- while (memory_available && spaces_need_memory) {
- spaces_need_memory = false;
- // Round-robin over all spaces.
- for (int i = 0; i < num; i++) {
- current_space = other[i]->Get(identity());
- if (current_space->free_list()->available() < limit) {
- // Space has not reached its limit. Try to get some memory.
- spaces_need_memory = true;
- node = TryRemoveMemory(limit - current_space->free_list()->available());
- if (node != nullptr) {
- CHECK(current_space->identity() == identity());
- current_space->AddMemory(node->address(), node->size());
- } else {
- memory_available = false;
- break;
- }
- }
- }
- }
-}
-
-
-void PagedSpace::RefillFreeList() {
- MarkCompactCollector* collector = heap()->mark_compact_collector();
- FreeList* free_list = nullptr;
- if (this == heap()->old_space()) {
- free_list = collector->free_list_old_space().get();
- } else if (this == heap()->code_space()) {
- free_list = collector->free_list_code_space().get();
- } else if (this == heap()->map_space()) {
- free_list = collector->free_list_map_space().get();
- } else {
- // Any PagedSpace might invoke RefillFreeList. We filter all but our old
- // generation spaces out.
- return;
- }
- DCHECK(free_list != nullptr);
- intptr_t added = free_list_.Concatenate(free_list);
- accounting_stats_.IncreaseCapacity(added);
-}
-
-
-void CompactionSpace::RefillFreeList() {
- MarkCompactCollector* collector = heap()->mark_compact_collector();
- FreeList* free_list = nullptr;
- if (identity() == OLD_SPACE) {
- free_list = collector->free_list_old_space().get();
- } else if (identity() == CODE_SPACE) {
- free_list = collector->free_list_code_space().get();
- } else {
- // Compaction spaces only represent old or code space.
- UNREACHABLE();
- }
- DCHECK(free_list != nullptr);
- intptr_t refilled = 0;
- while (refilled < kCompactionMemoryWanted) {
- FreeSpace* node =
- free_list->TryRemoveMemory(kCompactionMemoryWanted - refilled);
- if (node == nullptr) return;
- refilled += node->size();
- AddMemory(node->address(), node->size());
- }
-}
-
-
void PagedSpace::MoveOverFreeMemory(PagedSpace* other) {
DCHECK(identity() == other->identity());
// Destroy the linear allocation space of {other}. This is needed to
@@ -1089,7 +994,8 @@
intptr_t added = free_list_.Concatenate(other->free_list());
// Moved memory is not recorded as allocated memory, but rather increases and
- // decreases capacity of the corresponding spaces.
+ // decreases capacity of the corresponding spaces. Used size and waste size
+ // are maintained by the receiving space upon allocating and freeing blocks.
other->accounting_stats_.DecreaseCapacity(added);
accounting_stats_.IncreaseCapacity(added);
}
@@ -1098,19 +1004,16 @@
void PagedSpace::MergeCompactionSpace(CompactionSpace* other) {
// Unmerged fields:
// area_size_
+ // allocation_info_
+ // end_of_unswept_pages_
+ // unswept_free_bytes_
// anchor_
MoveOverFreeMemory(other);
// Update and clear accounting statistics.
accounting_stats_.Merge(other->accounting_stats_);
- other->accounting_stats_.Clear();
-
- // The linear allocation area of {other} should be destroyed now.
- DCHECK(other->top() == nullptr);
- DCHECK(other->limit() == nullptr);
-
- DCHECK(other->end_of_unswept_pages_ == nullptr);
+ other->accounting_stats_.Reset();
AccountCommitted(other->CommittedMemory());
@@ -2489,28 +2392,6 @@
}
-FreeSpace* FreeList::TryRemoveMemory(intptr_t hint_size_in_bytes) {
- base::LockGuard<base::Mutex> guard(&mutex_);
- FreeSpace* node = nullptr;
- int node_size = 0;
- // Try to find a node that fits exactly.
- node = FindNodeFor(static_cast<int>(hint_size_in_bytes), &node_size);
- // If no node could be found get as much memory as possible.
- if (node == nullptr) node = FindNodeIn(kHuge, &node_size);
- if (node == nullptr) node = FindNodeIn(kLarge, &node_size);
- if (node != nullptr) {
- // Give back left overs that were not required by {size_in_bytes}.
- intptr_t aligned_size = RoundUp(hint_size_in_bytes, kPointerSize);
- intptr_t left_over = node_size - aligned_size;
- if (left_over > 0) {
- Free(node->address() + aligned_size, static_cast<int>(left_over));
- node->set_size(static_cast<int>(aligned_size));
- }
- }
- return node;
-}
-
-
// Allocation on the old space free list. If it succeeds then a new linear
// allocation space has been set up with the top and limit of the space. If
// the allocation fails then NULL is returned, and the caller can perform a GC
@@ -2746,7 +2627,7 @@
if (collector->sweeping_in_progress()) {
// First try to refill the free-list, concurrent sweeper threads
// may have freed some objects in the meantime.
- RefillFreeList();
+ collector->RefillFreeList(this);
// Retry the free list allocation.
HeapObject* object = free_list_.Allocate(size_in_bytes);
@@ -2754,7 +2635,7 @@
// If sweeping is still in progress try to sweep pages on the main thread.
int free_chunk = collector->SweepInParallel(this, size_in_bytes);
- RefillFreeList();
+ collector->RefillFreeList(this);
if (free_chunk >= size_in_bytes) {
HeapObject* object = free_list_.Allocate(size_in_bytes);
// We should be able to allocate an object here since we just freed that
« no previous file with comments | « src/heap/spaces.h ('k') | test/cctest/heap-tester.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698