| OLD | NEW |
| 1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 189 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 200 | 200 |
| 201 memory_allocator->Free(memory_chunk); | 201 memory_allocator->Free(memory_chunk); |
| 202 memory_allocator->TearDown(); | 202 memory_allocator->TearDown(); |
| 203 delete memory_allocator; | 203 delete memory_allocator; |
| 204 } | 204 } |
| 205 | 205 |
| 206 | 206 |
| 207 TEST(Regress3540) { | 207 TEST(Regress3540) { |
| 208 Isolate* isolate = CcTest::i_isolate(); | 208 Isolate* isolate = CcTest::i_isolate(); |
| 209 Heap* heap = isolate->heap(); | 209 Heap* heap = isolate->heap(); |
| 210 const int pageSize = Page::kPageSize; |
| 210 MemoryAllocator* memory_allocator = new MemoryAllocator(isolate); | 211 MemoryAllocator* memory_allocator = new MemoryAllocator(isolate); |
| 211 CHECK( | 212 CHECK( |
| 212 memory_allocator->SetUp(heap->MaxReserved(), heap->MaxExecutableSize())); | 213 memory_allocator->SetUp(heap->MaxReserved(), heap->MaxExecutableSize())); |
| 213 TestMemoryAllocatorScope test_allocator_scope(isolate, memory_allocator); | 214 TestMemoryAllocatorScope test_allocator_scope(isolate, memory_allocator); |
| 214 CodeRange* code_range = new CodeRange(isolate); | 215 CodeRange* code_range = new CodeRange(isolate); |
| 215 const size_t code_range_size = 4 * MB; | 216 const size_t code_range_size = 4 * pageSize; |
| 216 if (!code_range->SetUp( | 217 if (!code_range->SetUp( |
| 217 code_range_size + | 218 code_range_size + |
| 218 RoundUp(v8::base::OS::CommitPageSize() * kReservedCodeRangePages, | 219 RoundUp(v8::base::OS::CommitPageSize() * kReservedCodeRangePages, |
| 219 MemoryChunk::kAlignment) + | 220 MemoryChunk::kAlignment) + |
| 220 v8::internal::MemoryAllocator::CodePageAreaSize())) { | 221 v8::internal::MemoryAllocator::CodePageAreaSize())) { |
| 221 return; | 222 return; |
| 222 } | 223 } |
| 223 Address address; | 224 Address address; |
| 224 size_t size; | 225 size_t size; |
| 225 address = code_range->AllocateRawMemory(code_range_size - 2 * MB, | 226 address = code_range->AllocateRawMemory( |
| 226 code_range_size - 2 * MB, &size); | 227 code_range_size - 2 * pageSize, code_range_size - 2 * pageSize, &size); |
| 227 CHECK(address != NULL); | 228 CHECK(address != NULL); |
| 228 Address null_address; | 229 Address null_address; |
| 229 size_t null_size; | 230 size_t null_size; |
| 230 null_address = code_range->AllocateRawMemory( | 231 null_address = code_range->AllocateRawMemory( |
| 231 code_range_size - MB, code_range_size - MB, &null_size); | 232 code_range_size - pageSize, code_range_size - pageSize, &null_size); |
| 232 CHECK(null_address == NULL); | 233 CHECK(null_address == NULL); |
| 233 code_range->FreeRawMemory(address, size); | 234 code_range->FreeRawMemory(address, size); |
| 234 delete code_range; | 235 delete code_range; |
| 235 memory_allocator->TearDown(); | 236 memory_allocator->TearDown(); |
| 236 delete memory_allocator; | 237 delete memory_allocator; |
| 237 } | 238 } |
| 238 | 239 |
| 239 | 240 |
| 240 static unsigned int Pseudorandom() { | 241 static unsigned int Pseudorandom() { |
| 241 static uint32_t lo = 2345; | 242 static uint32_t lo = 2345; |
| (...skipping 173 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 415 | 416 |
| 416 CHECK(lo->FindObject(ho->address()) == obj); | 417 CHECK(lo->FindObject(ho->address()) == obj); |
| 417 | 418 |
| 418 CHECK(lo->Contains(ho)); | 419 CHECK(lo->Contains(ho)); |
| 419 | 420 |
| 420 while (true) { | 421 while (true) { |
| 421 intptr_t available = lo->Available(); | 422 intptr_t available = lo->Available(); |
| 422 { AllocationResult allocation = lo->AllocateRaw(lo_size, NOT_EXECUTABLE); | 423 { AllocationResult allocation = lo->AllocateRaw(lo_size, NOT_EXECUTABLE); |
| 423 if (allocation.IsRetry()) break; | 424 if (allocation.IsRetry()) break; |
| 424 } | 425 } |
| 425 CHECK(lo->Available() < available); | 426 // The available value is conservative such that it may report |
| 427 // zero prior to heap exhaustion. |
| 428 CHECK(lo->Available() < available || available == 0); |
| 426 } | 429 } |
| 427 | 430 |
| 428 CHECK(!lo->IsEmpty()); | 431 CHECK(!lo->IsEmpty()); |
| 429 | 432 |
| 430 CHECK(lo->AllocateRaw(lo_size, NOT_EXECUTABLE).IsRetry()); | 433 CHECK(lo->AllocateRaw(lo_size, NOT_EXECUTABLE).IsRetry()); |
| 431 } | 434 } |
| 432 | 435 |
| 433 | 436 |
| 434 TEST(SizeOfFirstPageIsLargeEnough) { | 437 TEST(SizeOfFirstPageIsLargeEnough) { |
| 435 if (i::FLAG_always_opt) return; | 438 if (i::FLAG_always_opt) return; |
| (...skipping 17 matching lines...) Expand all Loading... |
| 453 if (i == CODE_SPACE && i::FLAG_debug_code) continue; | 456 if (i == CODE_SPACE && i::FLAG_debug_code) continue; |
| 454 CHECK_EQ(1, isolate->heap()->paged_space(i)->CountTotalPages()); | 457 CHECK_EQ(1, isolate->heap()->paged_space(i)->CountTotalPages()); |
| 455 } | 458 } |
| 456 | 459 |
| 457 // No large objects required to perform the above steps. | 460 // No large objects required to perform the above steps. |
| 458 CHECK(isolate->heap()->lo_space()->IsEmpty()); | 461 CHECK(isolate->heap()->lo_space()->IsEmpty()); |
| 459 } | 462 } |
| 460 | 463 |
| 461 | 464 |
| 462 UNINITIALIZED_TEST(NewSpaceGrowsToTargetCapacity) { | 465 UNINITIALIZED_TEST(NewSpaceGrowsToTargetCapacity) { |
| 463 FLAG_target_semi_space_size = 2; | 466 FLAG_target_semi_space_size = 2 * (Page::kPageSize / MB); |
| 464 if (FLAG_optimize_for_size) return; | 467 if (FLAG_optimize_for_size) return; |
| 465 | 468 |
| 466 v8::Isolate* isolate = v8::Isolate::New(); | 469 v8::Isolate* isolate = v8::Isolate::New(); |
| 467 { | 470 { |
| 468 v8::Isolate::Scope isolate_scope(isolate); | 471 v8::Isolate::Scope isolate_scope(isolate); |
| 469 v8::HandleScope handle_scope(isolate); | 472 v8::HandleScope handle_scope(isolate); |
| 470 v8::Context::New(isolate)->Enter(); | 473 v8::Context::New(isolate)->Enter(); |
| 471 | 474 |
| 472 Isolate* i_isolate = reinterpret_cast<Isolate*>(isolate); | 475 Isolate* i_isolate = reinterpret_cast<Isolate*>(isolate); |
| 473 | 476 |
| (...skipping 16 matching lines...) Expand all Loading... |
| 490 | 493 |
| 491 // Turn the allocation into a proper object so isolate teardown won't | 494 // Turn the allocation into a proper object so isolate teardown won't |
| 492 // crash. | 495 // crash. |
| 493 HeapObject* free_space = NULL; | 496 HeapObject* free_space = NULL; |
| 494 CHECK(allocation.To(&free_space)); | 497 CHECK(allocation.To(&free_space)); |
| 495 new_space->heap()->CreateFillerObjectAt(free_space->address(), 80); | 498 new_space->heap()->CreateFillerObjectAt(free_space->address(), 80); |
| 496 } | 499 } |
| 497 } | 500 } |
| 498 isolate->Dispose(); | 501 isolate->Dispose(); |
| 499 } | 502 } |
| OLD | NEW |