| OLD | NEW |
| 1 // Copyright 2007-2010 the V8 project authors. All rights reserved. | 1 // Copyright 2007-2010 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 463 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 474 const int kSmallStringSize = | 474 const int kSmallStringSize = |
| 475 SeqAsciiString::kHeaderSize + kSmallStringLength; | 475 SeqAsciiString::kHeaderSize + kSmallStringLength; |
| 476 const int kMapSize = Map::kSize; | 476 const int kMapSize = Map::kSize; |
| 477 | 477 |
| 478 Object* new_last = NULL; | 478 Object* new_last = NULL; |
| 479 for (int i = 0; | 479 for (int i = 0; |
| 480 i + kSmallFixedArraySize <= new_space_size; | 480 i + kSmallFixedArraySize <= new_space_size; |
| 481 i += kSmallFixedArraySize) { | 481 i += kSmallFixedArraySize) { |
| 482 Object* obj = Heap::AllocateFixedArray(kSmallFixedArrayLength); | 482 Object* obj = Heap::AllocateFixedArray(kSmallFixedArrayLength); |
| 483 if (new_last != NULL) { | 483 if (new_last != NULL) { |
| 484 CHECK_EQ(reinterpret_cast<char*>(obj), | 484 CHECK(reinterpret_cast<char*>(obj) == |
| 485 reinterpret_cast<char*>(new_last) + kSmallFixedArraySize); | 485 reinterpret_cast<char*>(new_last) + kSmallFixedArraySize); |
| 486 } | 486 } |
| 487 new_last = obj; | 487 new_last = obj; |
| 488 } | 488 } |
| 489 | 489 |
| 490 Object* pointer_last = NULL; | 490 Object* pointer_last = NULL; |
| 491 for (int i = 0; | 491 for (int i = 0; |
| 492 i + kSmallFixedArraySize <= size; | 492 i + kSmallFixedArraySize <= size; |
| 493 i += kSmallFixedArraySize) { | 493 i += kSmallFixedArraySize) { |
| 494 Object* obj = Heap::AllocateFixedArray(kSmallFixedArrayLength, TENURED); | 494 Object* obj = Heap::AllocateFixedArray(kSmallFixedArrayLength, TENURED); |
| 495 int old_page_fullness = i % Page::kPageSize; | 495 int old_page_fullness = i % Page::kPageSize; |
| 496 int page_fullness = (i + kSmallFixedArraySize) % Page::kPageSize; | 496 int page_fullness = (i + kSmallFixedArraySize) % Page::kPageSize; |
| 497 if (page_fullness < old_page_fullness || | 497 if (page_fullness < old_page_fullness || |
| 498 page_fullness > Page::kObjectAreaSize) { | 498 page_fullness > Page::kObjectAreaSize) { |
| 499 i = RoundUp(i, Page::kPageSize); | 499 i = RoundUp(i, Page::kPageSize); |
| 500 pointer_last = NULL; | 500 pointer_last = NULL; |
| 501 } | 501 } |
| 502 if (pointer_last != NULL) { | 502 if (pointer_last != NULL) { |
| 503 CHECK_EQ(reinterpret_cast<char*>(obj), | 503 CHECK(reinterpret_cast<char*>(obj) == |
| 504 reinterpret_cast<char*>(pointer_last) + kSmallFixedArraySize); | 504 reinterpret_cast<char*>(pointer_last) + kSmallFixedArraySize); |
| 505 } | 505 } |
| 506 pointer_last = obj; | 506 pointer_last = obj; |
| 507 } | 507 } |
| 508 | 508 |
| 509 Object* data_last = NULL; | 509 Object* data_last = NULL; |
| 510 for (int i = 0; i + kSmallStringSize <= size; i += kSmallStringSize) { | 510 for (int i = 0; i + kSmallStringSize <= size; i += kSmallStringSize) { |
| 511 Object* obj = Heap::AllocateRawAsciiString(kSmallStringLength, TENURED); | 511 Object* obj = Heap::AllocateRawAsciiString(kSmallStringLength, TENURED); |
| 512 int old_page_fullness = i % Page::kPageSize; | 512 int old_page_fullness = i % Page::kPageSize; |
| 513 int page_fullness = (i + kSmallStringSize) % Page::kPageSize; | 513 int page_fullness = (i + kSmallStringSize) % Page::kPageSize; |
| 514 if (page_fullness < old_page_fullness || | 514 if (page_fullness < old_page_fullness || |
| 515 page_fullness > Page::kObjectAreaSize) { | 515 page_fullness > Page::kObjectAreaSize) { |
| 516 i = RoundUp(i, Page::kPageSize); | 516 i = RoundUp(i, Page::kPageSize); |
| 517 data_last = NULL; | 517 data_last = NULL; |
| 518 } | 518 } |
| 519 if (data_last != NULL) { | 519 if (data_last != NULL) { |
| 520 CHECK_EQ(reinterpret_cast<char*>(obj), | 520 CHECK(reinterpret_cast<char*>(obj) == |
| 521 reinterpret_cast<char*>(data_last) + kSmallStringSize); | 521 reinterpret_cast<char*>(data_last) + kSmallStringSize); |
| 522 } | 522 } |
| 523 data_last = obj; | 523 data_last = obj; |
| 524 } | 524 } |
| 525 | 525 |
| 526 Object* map_last = NULL; | 526 Object* map_last = NULL; |
| 527 for (int i = 0; i + kMapSize <= size; i += kMapSize) { | 527 for (int i = 0; i + kMapSize <= size; i += kMapSize) { |
| 528 Object* obj = Heap::AllocateMap(JS_OBJECT_TYPE, 42 * kPointerSize); | 528 Object* obj = Heap::AllocateMap(JS_OBJECT_TYPE, 42 * kPointerSize); |
| 529 int old_page_fullness = i % Page::kPageSize; | 529 int old_page_fullness = i % Page::kPageSize; |
| 530 int page_fullness = (i + kMapSize) % Page::kPageSize; | 530 int page_fullness = (i + kMapSize) % Page::kPageSize; |
| 531 if (page_fullness < old_page_fullness || | 531 if (page_fullness < old_page_fullness || |
| 532 page_fullness > Page::kObjectAreaSize) { | 532 page_fullness > Page::kObjectAreaSize) { |
| 533 i = RoundUp(i, Page::kPageSize); | 533 i = RoundUp(i, Page::kPageSize); |
| 534 map_last = NULL; | 534 map_last = NULL; |
| 535 } | 535 } |
| 536 if (map_last != NULL) { | 536 if (map_last != NULL) { |
| 537 CHECK_EQ(reinterpret_cast<char*>(obj), | 537 CHECK(reinterpret_cast<char*>(obj) == |
| 538 reinterpret_cast<char*>(map_last) + kMapSize); | 538 reinterpret_cast<char*>(map_last) + kMapSize); |
| 539 } | 539 } |
| 540 map_last = obj; | 540 map_last = obj; |
| 541 } | 541 } |
| 542 | 542 |
| 543 if (size > Page::kObjectAreaSize) { | 543 if (size > Page::kObjectAreaSize) { |
| 544 // Support for reserving space in large object space is not there yet, | 544 // Support for reserving space in large object space is not there yet, |
| 545 // but using an always-allocate scope is fine for now. | 545 // but using an always-allocate scope is fine for now. |
| 546 AlwaysAllocateScope always; | 546 AlwaysAllocateScope always; |
| 547 int large_object_array_length = | 547 int large_object_array_length = |
| 548 (size - FixedArray::kHeaderSize) / kPointerSize; | 548 (size - FixedArray::kHeaderSize) / kPointerSize; |
| (...skipping 12 matching lines...) Expand all Loading... |
| 561 TEST(TestThatAlwaysFails) { | 561 TEST(TestThatAlwaysFails) { |
| 562 bool ArtificialFailure = false; | 562 bool ArtificialFailure = false; |
| 563 CHECK(ArtificialFailure); | 563 CHECK(ArtificialFailure); |
| 564 } | 564 } |
| 565 | 565 |
| 566 | 566 |
| 567 DEPENDENT_TEST(DependentTestThatAlwaysFails, TestThatAlwaysSucceeds) { | 567 DEPENDENT_TEST(DependentTestThatAlwaysFails, TestThatAlwaysSucceeds) { |
| 568 bool ArtificialFailure2 = false; | 568 bool ArtificialFailure2 = false; |
| 569 CHECK(ArtificialFailure2); | 569 CHECK(ArtificialFailure2); |
| 570 } | 570 } |
| OLD | NEW |