| Index: src/spaces.cc
|
| diff --git a/src/spaces.cc b/src/spaces.cc
|
| index 73abff3db9b3142b247a799f1ea4051fe9e6bb04..ac3c068d1509756ea810886fabd05a67048555fb 100644
|
| --- a/src/spaces.cc
|
| +++ b/src/spaces.cc
|
| @@ -757,7 +757,7 @@ int PagedSpace::CountTotalPages() {
|
|
|
|
|
| void PagedSpace::Shrink() {
|
| - // TODO(gc) release half of pages?
|
| + // TODO(1614) Not implemented.
|
| }
|
|
|
|
|
| @@ -2309,8 +2309,8 @@ void LargeObjectSpace::TearDown() {
|
| }
|
|
|
|
|
| -MaybeObject* LargeObjectSpace::AllocateRawInternal(int object_size,
|
| - Executability executable) {
|
| +MaybeObject* LargeObjectSpace::AllocateRaw(int object_size,
|
| + Executability executable) {
|
| // Check if we want to force a GC before growing the old space further.
|
| // If so, fail the allocation.
|
| if (!heap()->always_allocate() &&
|
| @@ -2318,7 +2318,6 @@ MaybeObject* LargeObjectSpace::AllocateRawInternal(int object_size,
|
| return Failure::RetryAfterGC(identity());
|
| }
|
|
|
| - // TODO(gc) isolates merge
|
| LargePage* page = heap()->isolate()->memory_allocator()->
|
| AllocateLargePage(object_size, executable, this);
|
| if (page == NULL) return Failure::RetryAfterGC(identity());
|
| @@ -2336,24 +2335,6 @@ MaybeObject* LargeObjectSpace::AllocateRawInternal(int object_size,
|
| }
|
|
|
|
|
| -MaybeObject* LargeObjectSpace::AllocateRawCode(int size_in_bytes) {
|
| - ASSERT(0 < size_in_bytes);
|
| - return AllocateRawInternal(size_in_bytes, EXECUTABLE);
|
| -}
|
| -
|
| -
|
| -MaybeObject* LargeObjectSpace::AllocateRawFixedArray(int size_in_bytes) {
|
| - ASSERT(0 < size_in_bytes);
|
| - return AllocateRawInternal(size_in_bytes, NOT_EXECUTABLE);
|
| -}
|
| -
|
| -
|
| -MaybeObject* LargeObjectSpace::AllocateRawData(int size_in_bytes) {
|
| - ASSERT(0 < size_in_bytes);
|
| - return AllocateRawInternal(size_in_bytes, NOT_EXECUTABLE);
|
| -}
|
| -
|
| -
|
| // GC support
|
| MaybeObject* LargeObjectSpace::FindObject(Address a) {
|
| for (LargePage* page = first_page_;
|
|
|