OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 // Platform-specific code for Solaris 10 goes here. For the POSIX-compatible | 5 // Platform-specific code for Solaris 10 goes here. For the POSIX-compatible |
6 // parts, the implementation is in platform-posix.cc. | 6 // parts, the implementation is in platform-posix.cc. |
7 | 7 |
8 #ifdef __sparc | 8 #ifdef __sparc |
9 # error "V8 does not support the SPARC CPU architecture." | 9 # error "V8 does not support the SPARC CPU architecture." |
10 #endif | 10 #endif |
(...skipping 136 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
147 | 147 |
148 VirtualMemory::VirtualMemory() : address_(NULL), size_(0) { } | 148 VirtualMemory::VirtualMemory() : address_(NULL), size_(0) { } |
149 | 149 |
150 | 150 |
151 VirtualMemory::VirtualMemory(size_t size) | 151 VirtualMemory::VirtualMemory(size_t size) |
152 : address_(ReserveRegion(size)), size_(size) { } | 152 : address_(ReserveRegion(size)), size_(size) { } |
153 | 153 |
154 | 154 |
155 VirtualMemory::VirtualMemory(size_t size, size_t alignment) | 155 VirtualMemory::VirtualMemory(size_t size, size_t alignment) |
156 : address_(NULL), size_(0) { | 156 : address_(NULL), size_(0) { |
157 ASSERT(IsAligned(alignment, static_cast<intptr_t>(OS::AllocateAlignment()))); | 157 DCHECK(IsAligned(alignment, static_cast<intptr_t>(OS::AllocateAlignment()))); |
158 size_t request_size = RoundUp(size + alignment, | 158 size_t request_size = RoundUp(size + alignment, |
159 static_cast<intptr_t>(OS::AllocateAlignment())); | 159 static_cast<intptr_t>(OS::AllocateAlignment())); |
160 void* reservation = mmap(OS::GetRandomMmapAddr(), | 160 void* reservation = mmap(OS::GetRandomMmapAddr(), |
161 request_size, | 161 request_size, |
162 PROT_NONE, | 162 PROT_NONE, |
163 MAP_PRIVATE | MAP_ANONYMOUS | MAP_NORESERVE, | 163 MAP_PRIVATE | MAP_ANONYMOUS | MAP_NORESERVE, |
164 kMmapFd, | 164 kMmapFd, |
165 kMmapFdOffset); | 165 kMmapFdOffset); |
166 if (reservation == MAP_FAILED) return; | 166 if (reservation == MAP_FAILED) return; |
167 | 167 |
168 uint8_t* base = static_cast<uint8_t*>(reservation); | 168 uint8_t* base = static_cast<uint8_t*>(reservation); |
169 uint8_t* aligned_base = RoundUp(base, alignment); | 169 uint8_t* aligned_base = RoundUp(base, alignment); |
170 ASSERT_LE(base, aligned_base); | 170 DCHECK_LE(base, aligned_base); |
171 | 171 |
172 // Unmap extra memory reserved before and after the desired block. | 172 // Unmap extra memory reserved before and after the desired block. |
173 if (aligned_base != base) { | 173 if (aligned_base != base) { |
174 size_t prefix_size = static_cast<size_t>(aligned_base - base); | 174 size_t prefix_size = static_cast<size_t>(aligned_base - base); |
175 OS::Free(base, prefix_size); | 175 OS::Free(base, prefix_size); |
176 request_size -= prefix_size; | 176 request_size -= prefix_size; |
177 } | 177 } |
178 | 178 |
179 size_t aligned_size = RoundUp(size, OS::AllocateAlignment()); | 179 size_t aligned_size = RoundUp(size, OS::AllocateAlignment()); |
180 ASSERT_LE(aligned_size, request_size); | 180 DCHECK_LE(aligned_size, request_size); |
181 | 181 |
182 if (aligned_size != request_size) { | 182 if (aligned_size != request_size) { |
183 size_t suffix_size = request_size - aligned_size; | 183 size_t suffix_size = request_size - aligned_size; |
184 OS::Free(aligned_base + aligned_size, suffix_size); | 184 OS::Free(aligned_base + aligned_size, suffix_size); |
185 request_size -= suffix_size; | 185 request_size -= suffix_size; |
186 } | 186 } |
187 | 187 |
188 ASSERT(aligned_size == request_size); | 188 DCHECK(aligned_size == request_size); |
189 | 189 |
190 address_ = static_cast<void*>(aligned_base); | 190 address_ = static_cast<void*>(aligned_base); |
191 size_ = aligned_size; | 191 size_ = aligned_size; |
192 } | 192 } |
193 | 193 |
194 | 194 |
195 VirtualMemory::~VirtualMemory() { | 195 VirtualMemory::~VirtualMemory() { |
196 if (IsReserved()) { | 196 if (IsReserved()) { |
197 bool result = ReleaseRegion(address(), size()); | 197 bool result = ReleaseRegion(address(), size()); |
198 ASSERT(result); | 198 DCHECK(result); |
199 USE(result); | 199 USE(result); |
200 } | 200 } |
201 } | 201 } |
202 | 202 |
203 | 203 |
204 bool VirtualMemory::IsReserved() { | 204 bool VirtualMemory::IsReserved() { |
205 return address_ != NULL; | 205 return address_ != NULL; |
206 } | 206 } |
207 | 207 |
208 | 208 |
(...skipping 61 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
270 return munmap(base, size) == 0; | 270 return munmap(base, size) == 0; |
271 } | 271 } |
272 | 272 |
273 | 273 |
274 bool VirtualMemory::HasLazyCommits() { | 274 bool VirtualMemory::HasLazyCommits() { |
275 // TODO(alph): implement for the platform. | 275 // TODO(alph): implement for the platform. |
276 return false; | 276 return false; |
277 } | 277 } |
278 | 278 |
279 } } // namespace v8::base | 279 } } // namespace v8::base |
OLD | NEW |