Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(860)

Unified Diff: src/platform/virtual-memory.cc

Issue 23970004: Revert r16648, r16641, r16638 and r16637. (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: Created 7 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/platform/virtual-memory.h ('k') | src/spaces.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/platform/virtual-memory.cc
diff --git a/src/platform/virtual-memory.cc b/src/platform/virtual-memory.cc
deleted file mode 100644
index f72bc9054c392daa7f5837daee76563eb81678fd..0000000000000000000000000000000000000000
--- a/src/platform/virtual-memory.cc
+++ /dev/null
@@ -1,513 +0,0 @@
-// Copyright 2013 the V8 project authors. All rights reserved.
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-// * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-// * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following
-// disclaimer in the documentation and/or other materials provided
-// with the distribution.
-// * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived
-// from this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-#include "platform/virtual-memory.h"
-
-#if V8_OS_POSIX
-#include <sys/types.h>
-#include <sys/mman.h>
-#include <sys/time.h>
-#include <sys/resource.h>
-
-#include <unistd.h>
-#endif
-
-#if V8_OS_MACOSX
-#include <mach/vm_statistics.h>
-#endif
-
-#include <cerrno>
-
-#include "platform/mutex.h"
-#include "utils.h"
-#include "utils/random-number-generator.h"
-#if V8_OS_CYGIN || V8_OS_WIN
-#include "win32-headers.h"
-#endif
-
-namespace v8 {
-namespace internal {
-
-class RandomAddressGenerator V8_FINAL {
- public:
- V8_INLINE uintptr_t NextAddress() {
- LockGuard<Mutex> lock_guard(&mutex_);
- uintptr_t address = rng_.NextInt();
-#if V8_HOST_ARCH_64_BIT
- address = (address << 32) + static_cast<uintptr_t>(rng_.NextInt());
-#endif
- return address;
- }
-
- private:
- Mutex mutex_;
- RandomNumberGenerator rng_;
-};
-
-typedef LazyInstance<RandomAddressGenerator,
- DefaultConstructTrait<RandomAddressGenerator>,
- ThreadSafeInitOnceTrait>::type LazyRandomAddressGenerator;
-
-#define LAZY_RANDOM_ADDRESS_GENERATOR_INITIALIZER LAZY_INSTANCE_INITIALIZER
-
-
-static V8_INLINE void* GenerateRandomAddress() {
-#if V8_OS_NACL
- // TODO(bradchen): Restore randomization once Native Client gets smarter
- // about using mmap address hints.
- // See http://code.google.com/p/nativeclient/issues/3341
- return NULL;
-#else // V8_OS_NACL
- LazyRandomAddressGenerator random_address_generator =
- LAZY_RANDOM_ADDRESS_GENERATOR_INITIALIZER;
- uintptr_t address = random_address_generator.Pointer()->NextAddress();
-
-# if V8_TARGET_ARCH_X64
-# if V8_OS_CYGWIN || V8_OS_WIN
- // Try not to map pages into the default range that windows loads DLLs.
- // Use a multiple of 64KiB to prevent committing unused memory.
- address += V8_UINT64_C(0x00080000000);
- address &= V8_UINT64_C(0x3ffffff0000);
-# else // V8_OS_CYGWIN || V8_OS_WIN
- // Currently available CPUs have 48 bits of virtual addressing. Truncate
- // the hint address to 46 bits to give the kernel a fighting chance of
- // fulfilling our placement request.
- address &= V8_UINT64_C(0x3ffffffff000);
-# endif // V8_OS_CYGWIN || V8_OS_WIN
-# else // V8_TARGET_ARCH_X64
-# if V8_OS_CYGWIN || V8_OS_WIN
- // Try not to map pages into the default range that windows loads DLLs.
- // Use a multiple of 64KiB to prevent committing unused memory.
- address += 0x04000000;
- address &= 0x3fff0000;
-# elif V8_OS_SOLARIS
- // For our Solaris/illumos mmap hint, we pick a random address in the bottom
- // half of the top half of the address space (that is, the third quarter).
- // Because we do not MAP_FIXED, this will be treated only as a hint -- the
- // system will not fail to mmap() because something else happens to already
- // be mapped at our random address. We deliberately set the hint high enough
- // to get well above the system's break (that is, the heap); Solaris and
- // illumos will try the hint and if that fails allocate as if there were
- // no hint at all. The high hint prevents the break from getting hemmed in
- // at low values, ceding half of the address space to the system heap.
- address &= 0x3ffff000;
- address += 0x80000000;
-# else // V8_OS_CYGWIN || V8_OS_WIN
- // The range 0x20000000 - 0x60000000 is relatively unpopulated across a
- // variety of ASLR modes (PAE kernel, NX compat mode, etc) and on Mac OS X
- // 10.6 and 10.7.
- address &= 0x3ffff000;
- address += 0x20000000;
-# endif // V8_OS_CYGIN || V8_OS_WIN
-# endif // V8_TARGET_ARCH_X64
- return reinterpret_cast<void*>(address);
-#endif // V8_OS_NACL
-}
-
-
-// static
-void* VirtualMemory::AllocateRegion(size_t size,
- size_t* size_return,
- Executability executability) {
- ASSERT_LT(0, size);
- ASSERT_NE(NULL, size_return);
- void* address = ReserveRegion(size, &size);
- if (address == NULL) return NULL;
- if (!CommitRegion(address, size, executability)) {
- bool result = ReleaseRegion(address, size);
- ASSERT(result);
- USE(result);
- return NULL;
- }
- *size_return = size;
- return address;
-}
-
-#if V8_OS_CYGWIN || V8_OS_WIN
-
-// static
-void* VirtualMemory::ReserveRegion(size_t size, size_t* size_return) {
- ASSERT_LT(0, size);
- ASSERT_NE(NULL, size_return);
- // The minimum size that can be reserved is 64KiB, see
- // http://msdn.microsoft.com/en-us/library/ms810627.aspx
- if (size < 64 * KB) {
- size = 64 * KB;
- }
- size = RoundUp(size, GetPageSize());
- LPVOID address = NULL;
- // Try and randomize the allocation address (up to three attempts).
- for (unsigned attempts = 0; address == NULL && attempts < 3; ++attempts) {
- address = VirtualAlloc(GenerateRandomAddress(),
- size,
- MEM_RESERVE,
- PAGE_NOACCESS);
- }
- if (address == NULL) {
- // After three attempts give up and let the kernel find an address.
- address = VirtualAlloc(NULL, size, MEM_RESERVE, PAGE_NOACCESS);
- }
- if (address == NULL) {
- return NULL;
- }
- ASSERT(IsAligned(reinterpret_cast<uintptr_t>(address),
- GetAllocationGranularity()));
- *size_return = size;
- return address;
-}
-
-
-// static
-void* VirtualMemory::ReserveRegion(size_t size,
- size_t* size_return,
- size_t alignment) {
- ASSERT_LT(0, size);
- ASSERT_NE(NULL, size_return);
- ASSERT(IsAligned(alignment, GetAllocationGranularity()));
-
- size_t reserved_size = RoundUp(size + alignment, GetAllocationGranularity());
- Address reserved_base = static_cast<Address>(
- ReserveRegion(reserved_size, &reserved_size));
- if (reserved_base == NULL) {
- return NULL;
- }
- ASSERT_LE(size, reserved_size);
- ASSERT_LE(size + alignment, reserved_size);
- ASSERT(IsAligned(reserved_size, GetPageSize()));
-
- // Try reducing the size by freeing and then reallocating a specific area.
- bool result = ReleaseRegion(reserved_base, reserved_size);
- USE(result);
- ASSERT(result);
- size_t aligned_size = RoundUp(size, GetPageSize());
- Address aligned_base = static_cast<Address>(
- VirtualAlloc(RoundUp(reserved_base, alignment),
- aligned_size,
- MEM_RESERVE,
- PAGE_NOACCESS));
- if (aligned_base != NULL) {
- ASSERT(aligned_base == RoundUp(reserved_base, alignment));
- ASSERT(IsAligned(reinterpret_cast<uintptr_t>(aligned_base),
- GetAllocationGranularity()));
- ASSERT(IsAligned(aligned_size, GetPageSize()));
- *size_return = aligned_size;
- return aligned_base;
- }
-
- // Resizing failed, just go with a bigger area.
- ASSERT(IsAligned(reserved_size, GetAllocationGranularity()));
- return ReserveRegion(reserved_size, size_return);
-}
-
-
-// static
-bool VirtualMemory::CommitRegion(void* address,
- size_t size,
- Executability executability) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- DWORD protect = 0;
- switch (executability) {
- case NOT_EXECUTABLE:
- protect = PAGE_READWRITE;
- break;
-
- case EXECUTABLE:
- protect = PAGE_EXECUTE_READWRITE;
- break;
- }
- LPVOID result = VirtualAlloc(address, size, MEM_COMMIT, protect);
- if (result == NULL) {
- ASSERT(GetLastError() != ERROR_INVALID_ADDRESS);
- return false;
- }
- ASSERT_EQ(address, result);
- return true;
-}
-
-
-// static
-bool VirtualMemory::UncommitRegion(void* address, size_t size) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- int result = VirtualFree(address, size, MEM_DECOMMIT);
- if (result == 0) {
- return false;
- }
- return true;
-}
-
-
-// static
-bool VirtualMemory::WriteProtectRegion(void* address, size_t size) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- DWORD old_protect;
- return VirtualProtect(address, size, PAGE_EXECUTE_READ, &old_protect);
-}
-
-
-// static
-bool VirtualMemory::ReleaseRegion(void* address, size_t size) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- USE(size);
- int result = VirtualFree(address, 0, MEM_RELEASE);
- if (result == 0) {
- return false;
- }
- return true;
-}
-
-
-// static
-size_t VirtualMemory::GetAllocationGranularity() {
- static size_t allocation_granularity = 0;
- if (allocation_granularity == 0) {
- SYSTEM_INFO system_info;
- GetSystemInfo(&system_info);
- allocation_granularity = system_info.dwAllocationGranularity;
- MemoryBarrier();
- }
- ASSERT_GE(allocation_granularity, GetPageSize());
- return allocation_granularity;
-}
-
-
-// static
-size_t VirtualMemory::GetLimit() {
- return 0;
-}
-
-
-// static
-size_t VirtualMemory::GetPageSize() {
- static size_t page_size = 0;
- if (page_size == 0) {
- SYSTEM_INFO system_info;
- GetSystemInfo(&system_info);
- page_size = system_info.dwPageSize;
- MemoryBarrier();
- }
- return page_size;
-}
-
-
-#else // V8_OS_CYGIN || V8_OS_WIN
-
-
-// Constants used for mmap.
-#if V8_OS_MACOSX
-// kMmapFd is used to pass vm_alloc flags to tag the region with the user
-// defined tag 255 This helps identify V8-allocated regions in memory analysis
-// tools like vmmap(1).
-static const int kMmapFd = VM_MAKE_TAG(255);
-#else
-static const int kMmapFd = -1;
-#endif // V8_OS_MACOSX
-static const off_t kMmapFdOffset = 0;
-
-
-// static
-void* VirtualMemory::ReserveRegion(size_t size, size_t* size_return) {
- ASSERT_LT(0, size);
- ASSERT_NE(NULL, size_return);
-
- size = RoundUp(size, GetPageSize());
- void* address = mmap(GenerateRandomAddress(),
- size,
- PROT_NONE,
- MAP_ANON | MAP_NORESERVE | MAP_PRIVATE,
- kMmapFd,
- kMmapFdOffset);
- if (address == MAP_FAILED) {
- ASSERT_NE(EINVAL, errno);
- return NULL;
- }
- *size_return = size;
- return address;
-}
-
-
-// static
-void* VirtualMemory::ReserveRegion(size_t size,
- size_t* size_return,
- size_t alignment) {
- ASSERT_LT(0, size);
- ASSERT_NE(NULL, size_return);
- ASSERT(IsAligned(alignment, GetPageSize()));
-
- size_t reserved_size;
- Address reserved_base = static_cast<Address>(
- ReserveRegion(size + alignment, &reserved_size));
- if (reserved_base == NULL) {
- return NULL;
- }
-
- Address aligned_base = RoundUp(reserved_base, alignment);
- ASSERT_LE(reserved_base, aligned_base);
-
- // Unmap extra memory reserved before the aligned region.
- if (aligned_base != reserved_base) {
- size_t prefix_size = static_cast<size_t>(aligned_base - reserved_base);
- bool result = ReleaseRegion(reserved_base, prefix_size);
- ASSERT(result);
- USE(result);
- reserved_size -= prefix_size;
- }
-
- size_t aligned_size = RoundUp(size, GetPageSize());
- ASSERT_LE(aligned_size, reserved_size);
-
- // Unmap extra memory reserved after the aligned region.
- if (aligned_size != reserved_size) {
- size_t suffix_size = reserved_size - aligned_size;
- bool result = ReleaseRegion(aligned_base + aligned_size, suffix_size);
- ASSERT(result);
- USE(result);
- reserved_size -= suffix_size;
- }
-
- ASSERT(aligned_size == reserved_size);
- ASSERT_NE(NULL, aligned_base);
-
- *size_return = aligned_size;
- return aligned_base;
-}
-
-
-// static
-bool VirtualMemory::CommitRegion(void* address,
- size_t size,
- Executability executability) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- int prot = 0;
- // The Native Client port of V8 uses an interpreter,
- // so code pages don't need PROT_EXEC.
-#if V8_OS_NACL
- executability = NOT_EXECUTABLE;
-#endif
- switch (executability) {
- case NOT_EXECUTABLE:
- prot = PROT_READ | PROT_WRITE;
- break;
-
- case EXECUTABLE:
- prot = PROT_EXEC | PROT_READ | PROT_WRITE;
- break;
- }
- void* result = mmap(address,
- size,
- prot,
- MAP_ANON | MAP_FIXED | MAP_PRIVATE,
- kMmapFd,
- kMmapFdOffset);
- if (result == MAP_FAILED) {
- ASSERT_NE(EINVAL, errno);
- return false;
- }
- return true;
-}
-
-
-// static
-bool VirtualMemory::UncommitRegion(void* address, size_t size) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- void* result = mmap(address,
- size,
- PROT_NONE,
- MAP_ANON | MAP_FIXED | MAP_NORESERVE | MAP_PRIVATE,
- kMmapFd,
- kMmapFdOffset);
- if (result == MAP_FAILED) {
- ASSERT_NE(EINVAL, errno);
- return false;
- }
- return true;
-}
-
-
-// static
-bool VirtualMemory::WriteProtectRegion(void* address, size_t size) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
-#if V8_OS_NACL
- // The Native Client port of V8 uses an interpreter,
- // so code pages don't need PROT_EXEC.
- int prot = PROT_READ;
-#else
- int prot = PROT_EXEC | PROT_READ;
-#endif
- int result = mprotect(address, size, prot);
- if (result < 0) {
- ASSERT_NE(EINVAL, errno);
- return false;
- }
- return true;
-}
-
-
-// static
-bool VirtualMemory::ReleaseRegion(void* address, size_t size) {
- ASSERT_NE(NULL, address);
- ASSERT_LT(0, size);
- int result = munmap(address, size);
- if (result < 0) {
- ASSERT_NE(EINVAL, errno);
- return false;
- }
- return true;
-}
-
-
-// static
-size_t VirtualMemory::GetAllocationGranularity() {
- return GetPageSize();
-}
-
-
-// static
-size_t VirtualMemory::GetLimit() {
- struct rlimit rlim;
- int result = getrlimit(RLIMIT_DATA, &rlim);
- ASSERT_EQ(0, result);
- USE(result);
- return rlim.rlim_cur;
-}
-
-
-// static
-size_t VirtualMemory::GetPageSize() {
- static const size_t kPageSize = getpagesize();
- return kPageSize;
-}
-
-#endif // V8_OS_CYGWIN || V8_OS_WIN
-
-} } // namespace v8::internal
« no previous file with comments | « src/platform/virtual-memory.h ('k') | src/spaces.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698