| Index: src/platform/virtual-memory.cc
|
| diff --git a/src/platform/virtual-memory.cc b/src/platform/virtual-memory.cc
|
| deleted file mode 100644
|
| index f72bc9054c392daa7f5837daee76563eb81678fd..0000000000000000000000000000000000000000
|
| --- a/src/platform/virtual-memory.cc
|
| +++ /dev/null
|
| @@ -1,513 +0,0 @@
|
| -// Copyright 2013 the V8 project authors. All rights reserved.
|
| -// Redistribution and use in source and binary forms, with or without
|
| -// modification, are permitted provided that the following conditions are
|
| -// met:
|
| -//
|
| -// * Redistributions of source code must retain the above copyright
|
| -// notice, this list of conditions and the following disclaimer.
|
| -// * Redistributions in binary form must reproduce the above
|
| -// copyright notice, this list of conditions and the following
|
| -// disclaimer in the documentation and/or other materials provided
|
| -// with the distribution.
|
| -// * Neither the name of Google Inc. nor the names of its
|
| -// contributors may be used to endorse or promote products derived
|
| -// from this software without specific prior written permission.
|
| -//
|
| -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
| -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
| -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
| -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
| -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
| -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
| -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
| -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
| -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
| -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
| -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
| -
|
| -#include "platform/virtual-memory.h"
|
| -
|
| -#if V8_OS_POSIX
|
| -#include <sys/types.h>
|
| -#include <sys/mman.h>
|
| -#include <sys/time.h>
|
| -#include <sys/resource.h>
|
| -
|
| -#include <unistd.h>
|
| -#endif
|
| -
|
| -#if V8_OS_MACOSX
|
| -#include <mach/vm_statistics.h>
|
| -#endif
|
| -
|
| -#include <cerrno>
|
| -
|
| -#include "platform/mutex.h"
|
| -#include "utils.h"
|
| -#include "utils/random-number-generator.h"
|
| -#if V8_OS_CYGIN || V8_OS_WIN
|
| -#include "win32-headers.h"
|
| -#endif
|
| -
|
| -namespace v8 {
|
| -namespace internal {
|
| -
|
| -class RandomAddressGenerator V8_FINAL {
|
| - public:
|
| - V8_INLINE uintptr_t NextAddress() {
|
| - LockGuard<Mutex> lock_guard(&mutex_);
|
| - uintptr_t address = rng_.NextInt();
|
| -#if V8_HOST_ARCH_64_BIT
|
| - address = (address << 32) + static_cast<uintptr_t>(rng_.NextInt());
|
| -#endif
|
| - return address;
|
| - }
|
| -
|
| - private:
|
| - Mutex mutex_;
|
| - RandomNumberGenerator rng_;
|
| -};
|
| -
|
| -typedef LazyInstance<RandomAddressGenerator,
|
| - DefaultConstructTrait<RandomAddressGenerator>,
|
| - ThreadSafeInitOnceTrait>::type LazyRandomAddressGenerator;
|
| -
|
| -#define LAZY_RANDOM_ADDRESS_GENERATOR_INITIALIZER LAZY_INSTANCE_INITIALIZER
|
| -
|
| -
|
| -static V8_INLINE void* GenerateRandomAddress() {
|
| -#if V8_OS_NACL
|
| - // TODO(bradchen): Restore randomization once Native Client gets smarter
|
| - // about using mmap address hints.
|
| - // See http://code.google.com/p/nativeclient/issues/3341
|
| - return NULL;
|
| -#else // V8_OS_NACL
|
| - LazyRandomAddressGenerator random_address_generator =
|
| - LAZY_RANDOM_ADDRESS_GENERATOR_INITIALIZER;
|
| - uintptr_t address = random_address_generator.Pointer()->NextAddress();
|
| -
|
| -# if V8_TARGET_ARCH_X64
|
| -# if V8_OS_CYGWIN || V8_OS_WIN
|
| - // Try not to map pages into the default range that windows loads DLLs.
|
| - // Use a multiple of 64KiB to prevent committing unused memory.
|
| - address += V8_UINT64_C(0x00080000000);
|
| - address &= V8_UINT64_C(0x3ffffff0000);
|
| -# else // V8_OS_CYGWIN || V8_OS_WIN
|
| - // Currently available CPUs have 48 bits of virtual addressing. Truncate
|
| - // the hint address to 46 bits to give the kernel a fighting chance of
|
| - // fulfilling our placement request.
|
| - address &= V8_UINT64_C(0x3ffffffff000);
|
| -# endif // V8_OS_CYGWIN || V8_OS_WIN
|
| -# else // V8_TARGET_ARCH_X64
|
| -# if V8_OS_CYGWIN || V8_OS_WIN
|
| - // Try not to map pages into the default range that windows loads DLLs.
|
| - // Use a multiple of 64KiB to prevent committing unused memory.
|
| - address += 0x04000000;
|
| - address &= 0x3fff0000;
|
| -# elif V8_OS_SOLARIS
|
| - // For our Solaris/illumos mmap hint, we pick a random address in the bottom
|
| - // half of the top half of the address space (that is, the third quarter).
|
| - // Because we do not MAP_FIXED, this will be treated only as a hint -- the
|
| - // system will not fail to mmap() because something else happens to already
|
| - // be mapped at our random address. We deliberately set the hint high enough
|
| - // to get well above the system's break (that is, the heap); Solaris and
|
| - // illumos will try the hint and if that fails allocate as if there were
|
| - // no hint at all. The high hint prevents the break from getting hemmed in
|
| - // at low values, ceding half of the address space to the system heap.
|
| - address &= 0x3ffff000;
|
| - address += 0x80000000;
|
| -# else // V8_OS_CYGWIN || V8_OS_WIN
|
| - // The range 0x20000000 - 0x60000000 is relatively unpopulated across a
|
| - // variety of ASLR modes (PAE kernel, NX compat mode, etc) and on Mac OS X
|
| - // 10.6 and 10.7.
|
| - address &= 0x3ffff000;
|
| - address += 0x20000000;
|
| -# endif // V8_OS_CYGIN || V8_OS_WIN
|
| -# endif // V8_TARGET_ARCH_X64
|
| - return reinterpret_cast<void*>(address);
|
| -#endif // V8_OS_NACL
|
| -}
|
| -
|
| -
|
| -// static
|
| -void* VirtualMemory::AllocateRegion(size_t size,
|
| - size_t* size_return,
|
| - Executability executability) {
|
| - ASSERT_LT(0, size);
|
| - ASSERT_NE(NULL, size_return);
|
| - void* address = ReserveRegion(size, &size);
|
| - if (address == NULL) return NULL;
|
| - if (!CommitRegion(address, size, executability)) {
|
| - bool result = ReleaseRegion(address, size);
|
| - ASSERT(result);
|
| - USE(result);
|
| - return NULL;
|
| - }
|
| - *size_return = size;
|
| - return address;
|
| -}
|
| -
|
| -#if V8_OS_CYGWIN || V8_OS_WIN
|
| -
|
| -// static
|
| -void* VirtualMemory::ReserveRegion(size_t size, size_t* size_return) {
|
| - ASSERT_LT(0, size);
|
| - ASSERT_NE(NULL, size_return);
|
| - // The minimum size that can be reserved is 64KiB, see
|
| - // http://msdn.microsoft.com/en-us/library/ms810627.aspx
|
| - if (size < 64 * KB) {
|
| - size = 64 * KB;
|
| - }
|
| - size = RoundUp(size, GetPageSize());
|
| - LPVOID address = NULL;
|
| - // Try and randomize the allocation address (up to three attempts).
|
| - for (unsigned attempts = 0; address == NULL && attempts < 3; ++attempts) {
|
| - address = VirtualAlloc(GenerateRandomAddress(),
|
| - size,
|
| - MEM_RESERVE,
|
| - PAGE_NOACCESS);
|
| - }
|
| - if (address == NULL) {
|
| - // After three attempts give up and let the kernel find an address.
|
| - address = VirtualAlloc(NULL, size, MEM_RESERVE, PAGE_NOACCESS);
|
| - }
|
| - if (address == NULL) {
|
| - return NULL;
|
| - }
|
| - ASSERT(IsAligned(reinterpret_cast<uintptr_t>(address),
|
| - GetAllocationGranularity()));
|
| - *size_return = size;
|
| - return address;
|
| -}
|
| -
|
| -
|
| -// static
|
| -void* VirtualMemory::ReserveRegion(size_t size,
|
| - size_t* size_return,
|
| - size_t alignment) {
|
| - ASSERT_LT(0, size);
|
| - ASSERT_NE(NULL, size_return);
|
| - ASSERT(IsAligned(alignment, GetAllocationGranularity()));
|
| -
|
| - size_t reserved_size = RoundUp(size + alignment, GetAllocationGranularity());
|
| - Address reserved_base = static_cast<Address>(
|
| - ReserveRegion(reserved_size, &reserved_size));
|
| - if (reserved_base == NULL) {
|
| - return NULL;
|
| - }
|
| - ASSERT_LE(size, reserved_size);
|
| - ASSERT_LE(size + alignment, reserved_size);
|
| - ASSERT(IsAligned(reserved_size, GetPageSize()));
|
| -
|
| - // Try reducing the size by freeing and then reallocating a specific area.
|
| - bool result = ReleaseRegion(reserved_base, reserved_size);
|
| - USE(result);
|
| - ASSERT(result);
|
| - size_t aligned_size = RoundUp(size, GetPageSize());
|
| - Address aligned_base = static_cast<Address>(
|
| - VirtualAlloc(RoundUp(reserved_base, alignment),
|
| - aligned_size,
|
| - MEM_RESERVE,
|
| - PAGE_NOACCESS));
|
| - if (aligned_base != NULL) {
|
| - ASSERT(aligned_base == RoundUp(reserved_base, alignment));
|
| - ASSERT(IsAligned(reinterpret_cast<uintptr_t>(aligned_base),
|
| - GetAllocationGranularity()));
|
| - ASSERT(IsAligned(aligned_size, GetPageSize()));
|
| - *size_return = aligned_size;
|
| - return aligned_base;
|
| - }
|
| -
|
| - // Resizing failed, just go with a bigger area.
|
| - ASSERT(IsAligned(reserved_size, GetAllocationGranularity()));
|
| - return ReserveRegion(reserved_size, size_return);
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::CommitRegion(void* address,
|
| - size_t size,
|
| - Executability executability) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - DWORD protect = 0;
|
| - switch (executability) {
|
| - case NOT_EXECUTABLE:
|
| - protect = PAGE_READWRITE;
|
| - break;
|
| -
|
| - case EXECUTABLE:
|
| - protect = PAGE_EXECUTE_READWRITE;
|
| - break;
|
| - }
|
| - LPVOID result = VirtualAlloc(address, size, MEM_COMMIT, protect);
|
| - if (result == NULL) {
|
| - ASSERT(GetLastError() != ERROR_INVALID_ADDRESS);
|
| - return false;
|
| - }
|
| - ASSERT_EQ(address, result);
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::UncommitRegion(void* address, size_t size) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - int result = VirtualFree(address, size, MEM_DECOMMIT);
|
| - if (result == 0) {
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::WriteProtectRegion(void* address, size_t size) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - DWORD old_protect;
|
| - return VirtualProtect(address, size, PAGE_EXECUTE_READ, &old_protect);
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::ReleaseRegion(void* address, size_t size) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - USE(size);
|
| - int result = VirtualFree(address, 0, MEM_RELEASE);
|
| - if (result == 0) {
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -size_t VirtualMemory::GetAllocationGranularity() {
|
| - static size_t allocation_granularity = 0;
|
| - if (allocation_granularity == 0) {
|
| - SYSTEM_INFO system_info;
|
| - GetSystemInfo(&system_info);
|
| - allocation_granularity = system_info.dwAllocationGranularity;
|
| - MemoryBarrier();
|
| - }
|
| - ASSERT_GE(allocation_granularity, GetPageSize());
|
| - return allocation_granularity;
|
| -}
|
| -
|
| -
|
| -// static
|
| -size_t VirtualMemory::GetLimit() {
|
| - return 0;
|
| -}
|
| -
|
| -
|
| -// static
|
| -size_t VirtualMemory::GetPageSize() {
|
| - static size_t page_size = 0;
|
| - if (page_size == 0) {
|
| - SYSTEM_INFO system_info;
|
| - GetSystemInfo(&system_info);
|
| - page_size = system_info.dwPageSize;
|
| - MemoryBarrier();
|
| - }
|
| - return page_size;
|
| -}
|
| -
|
| -
|
| -#else // V8_OS_CYGIN || V8_OS_WIN
|
| -
|
| -
|
| -// Constants used for mmap.
|
| -#if V8_OS_MACOSX
|
| -// kMmapFd is used to pass vm_alloc flags to tag the region with the user
|
| -// defined tag 255 This helps identify V8-allocated regions in memory analysis
|
| -// tools like vmmap(1).
|
| -static const int kMmapFd = VM_MAKE_TAG(255);
|
| -#else
|
| -static const int kMmapFd = -1;
|
| -#endif // V8_OS_MACOSX
|
| -static const off_t kMmapFdOffset = 0;
|
| -
|
| -
|
| -// static
|
| -void* VirtualMemory::ReserveRegion(size_t size, size_t* size_return) {
|
| - ASSERT_LT(0, size);
|
| - ASSERT_NE(NULL, size_return);
|
| -
|
| - size = RoundUp(size, GetPageSize());
|
| - void* address = mmap(GenerateRandomAddress(),
|
| - size,
|
| - PROT_NONE,
|
| - MAP_ANON | MAP_NORESERVE | MAP_PRIVATE,
|
| - kMmapFd,
|
| - kMmapFdOffset);
|
| - if (address == MAP_FAILED) {
|
| - ASSERT_NE(EINVAL, errno);
|
| - return NULL;
|
| - }
|
| - *size_return = size;
|
| - return address;
|
| -}
|
| -
|
| -
|
| -// static
|
| -void* VirtualMemory::ReserveRegion(size_t size,
|
| - size_t* size_return,
|
| - size_t alignment) {
|
| - ASSERT_LT(0, size);
|
| - ASSERT_NE(NULL, size_return);
|
| - ASSERT(IsAligned(alignment, GetPageSize()));
|
| -
|
| - size_t reserved_size;
|
| - Address reserved_base = static_cast<Address>(
|
| - ReserveRegion(size + alignment, &reserved_size));
|
| - if (reserved_base == NULL) {
|
| - return NULL;
|
| - }
|
| -
|
| - Address aligned_base = RoundUp(reserved_base, alignment);
|
| - ASSERT_LE(reserved_base, aligned_base);
|
| -
|
| - // Unmap extra memory reserved before the aligned region.
|
| - if (aligned_base != reserved_base) {
|
| - size_t prefix_size = static_cast<size_t>(aligned_base - reserved_base);
|
| - bool result = ReleaseRegion(reserved_base, prefix_size);
|
| - ASSERT(result);
|
| - USE(result);
|
| - reserved_size -= prefix_size;
|
| - }
|
| -
|
| - size_t aligned_size = RoundUp(size, GetPageSize());
|
| - ASSERT_LE(aligned_size, reserved_size);
|
| -
|
| - // Unmap extra memory reserved after the aligned region.
|
| - if (aligned_size != reserved_size) {
|
| - size_t suffix_size = reserved_size - aligned_size;
|
| - bool result = ReleaseRegion(aligned_base + aligned_size, suffix_size);
|
| - ASSERT(result);
|
| - USE(result);
|
| - reserved_size -= suffix_size;
|
| - }
|
| -
|
| - ASSERT(aligned_size == reserved_size);
|
| - ASSERT_NE(NULL, aligned_base);
|
| -
|
| - *size_return = aligned_size;
|
| - return aligned_base;
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::CommitRegion(void* address,
|
| - size_t size,
|
| - Executability executability) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - int prot = 0;
|
| - // The Native Client port of V8 uses an interpreter,
|
| - // so code pages don't need PROT_EXEC.
|
| -#if V8_OS_NACL
|
| - executability = NOT_EXECUTABLE;
|
| -#endif
|
| - switch (executability) {
|
| - case NOT_EXECUTABLE:
|
| - prot = PROT_READ | PROT_WRITE;
|
| - break;
|
| -
|
| - case EXECUTABLE:
|
| - prot = PROT_EXEC | PROT_READ | PROT_WRITE;
|
| - break;
|
| - }
|
| - void* result = mmap(address,
|
| - size,
|
| - prot,
|
| - MAP_ANON | MAP_FIXED | MAP_PRIVATE,
|
| - kMmapFd,
|
| - kMmapFdOffset);
|
| - if (result == MAP_FAILED) {
|
| - ASSERT_NE(EINVAL, errno);
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::UncommitRegion(void* address, size_t size) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - void* result = mmap(address,
|
| - size,
|
| - PROT_NONE,
|
| - MAP_ANON | MAP_FIXED | MAP_NORESERVE | MAP_PRIVATE,
|
| - kMmapFd,
|
| - kMmapFdOffset);
|
| - if (result == MAP_FAILED) {
|
| - ASSERT_NE(EINVAL, errno);
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::WriteProtectRegion(void* address, size_t size) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| -#if V8_OS_NACL
|
| - // The Native Client port of V8 uses an interpreter,
|
| - // so code pages don't need PROT_EXEC.
|
| - int prot = PROT_READ;
|
| -#else
|
| - int prot = PROT_EXEC | PROT_READ;
|
| -#endif
|
| - int result = mprotect(address, size, prot);
|
| - if (result < 0) {
|
| - ASSERT_NE(EINVAL, errno);
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -bool VirtualMemory::ReleaseRegion(void* address, size_t size) {
|
| - ASSERT_NE(NULL, address);
|
| - ASSERT_LT(0, size);
|
| - int result = munmap(address, size);
|
| - if (result < 0) {
|
| - ASSERT_NE(EINVAL, errno);
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -
|
| -// static
|
| -size_t VirtualMemory::GetAllocationGranularity() {
|
| - return GetPageSize();
|
| -}
|
| -
|
| -
|
| -// static
|
| -size_t VirtualMemory::GetLimit() {
|
| - struct rlimit rlim;
|
| - int result = getrlimit(RLIMIT_DATA, &rlim);
|
| - ASSERT_EQ(0, result);
|
| - USE(result);
|
| - return rlim.rlim_cur;
|
| -}
|
| -
|
| -
|
| -// static
|
| -size_t VirtualMemory::GetPageSize() {
|
| - static const size_t kPageSize = getpagesize();
|
| - return kPageSize;
|
| -}
|
| -
|
| -#endif // V8_OS_CYGWIN || V8_OS_WIN
|
| -
|
| -} } // namespace v8::internal
|
|
|