OLD | NEW |
---|---|
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 2823 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
2834 | 2834 |
2835 ASSERT(!InNewSpace(empty_fixed_array())); | 2835 ASSERT(!InNewSpace(empty_fixed_array())); |
2836 return true; | 2836 return true; |
2837 } | 2837 } |
2838 | 2838 |
2839 | 2839 |
2840 MaybeObject* Heap::AllocateHeapNumber(double value, PretenureFlag pretenure) { | 2840 MaybeObject* Heap::AllocateHeapNumber(double value, PretenureFlag pretenure) { |
2841 // Statically ensure that it is safe to allocate heap numbers in paged | 2841 // Statically ensure that it is safe to allocate heap numbers in paged |
2842 // spaces. | 2842 // spaces. |
2843 STATIC_ASSERT(HeapNumber::kSize <= Page::kNonCodeObjectAreaSize); | 2843 STATIC_ASSERT(HeapNumber::kSize <= Page::kNonCodeObjectAreaSize); |
2844 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; | 2844 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; |
Hannes Payer (out of office)
2013/08/20 09:49:19
You can also use the select function here.
Michael Starzinger
2013/09/23 11:35:05
Done.
| |
2845 | 2845 |
2846 Object* result; | 2846 Object* result; |
2847 { MaybeObject* maybe_result = | 2847 { MaybeObject* maybe_result = |
2848 AllocateRaw(HeapNumber::kSize, space, OLD_DATA_SPACE); | 2848 AllocateRaw(HeapNumber::kSize, space, OLD_DATA_SPACE); |
2849 if (!maybe_result->ToObject(&result)) return maybe_result; | 2849 if (!maybe_result->ToObject(&result)) return maybe_result; |
2850 } | 2850 } |
2851 | 2851 |
2852 HeapObject::cast(result)->set_map_no_write_barrier(heap_number_map()); | 2852 HeapObject::cast(result)->set_map_no_write_barrier(heap_number_map()); |
2853 HeapNumber::cast(result)->set_value(value); | 2853 HeapNumber::cast(result)->set_value(value); |
2854 return result; | 2854 return result; |
(...skipping 1147 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4002 String* answer = String::cast(result); | 4002 String* answer = String::cast(result); |
4003 answer->Set(0, code); | 4003 answer->Set(0, code); |
4004 return answer; | 4004 return answer; |
4005 } | 4005 } |
4006 | 4006 |
4007 | 4007 |
4008 MaybeObject* Heap::AllocateByteArray(int length, PretenureFlag pretenure) { | 4008 MaybeObject* Heap::AllocateByteArray(int length, PretenureFlag pretenure) { |
4009 if (length < 0 || length > ByteArray::kMaxLength) { | 4009 if (length < 0 || length > ByteArray::kMaxLength) { |
4010 return Failure::OutOfMemoryException(0x7); | 4010 return Failure::OutOfMemoryException(0x7); |
4011 } | 4011 } |
4012 if (pretenure == NOT_TENURED) { | |
4013 return AllocateByteArray(length); | |
4014 } | |
4015 int size = ByteArray::SizeFor(length); | 4012 int size = ByteArray::SizeFor(length); |
4016 AllocationSpace space = | 4013 AllocationSpace space = SelectSpace(size, OLD_DATA_SPACE, pretenure); |
4017 (size > Page::kMaxNonCodeHeapObjectSize) ? LO_SPACE : OLD_DATA_SPACE; | |
4018 Object* result; | |
4019 { MaybeObject* maybe_result = AllocateRaw(size, space, space); | |
4020 if (!maybe_result->ToObject(&result)) return maybe_result; | |
4021 } | |
4022 | |
4023 reinterpret_cast<ByteArray*>(result)->set_map_no_write_barrier( | |
4024 byte_array_map()); | |
4025 reinterpret_cast<ByteArray*>(result)->set_length(length); | |
4026 return result; | |
4027 } | |
4028 | |
4029 | |
4030 MaybeObject* Heap::AllocateByteArray(int length) { | |
4031 if (length < 0 || length > ByteArray::kMaxLength) { | |
4032 return Failure::OutOfMemoryException(0x8); | |
4033 } | |
4034 int size = ByteArray::SizeFor(length); | |
4035 AllocationSpace space = | |
4036 (size > Page::kMaxNonCodeHeapObjectSize) ? LO_SPACE : NEW_SPACE; | |
4037 Object* result; | 4014 Object* result; |
4038 { MaybeObject* maybe_result = AllocateRaw(size, space, OLD_DATA_SPACE); | 4015 { MaybeObject* maybe_result = AllocateRaw(size, space, OLD_DATA_SPACE); |
4039 if (!maybe_result->ToObject(&result)) return maybe_result; | 4016 if (!maybe_result->ToObject(&result)) return maybe_result; |
4040 } | 4017 } |
4041 | 4018 |
4042 reinterpret_cast<ByteArray*>(result)->set_map_no_write_barrier( | 4019 reinterpret_cast<ByteArray*>(result)->set_map_no_write_barrier( |
4043 byte_array_map()); | 4020 byte_array_map()); |
4044 reinterpret_cast<ByteArray*>(result)->set_length(length); | 4021 reinterpret_cast<ByteArray*>(result)->set_length(length); |
4045 return result; | 4022 return result; |
4046 } | 4023 } |
(...skipping 10 matching lines...) Expand all Loading... | |
4057 filler->set_map_no_write_barrier(free_space_map()); | 4034 filler->set_map_no_write_barrier(free_space_map()); |
4058 FreeSpace::cast(filler)->set_size(size); | 4035 FreeSpace::cast(filler)->set_size(size); |
4059 } | 4036 } |
4060 } | 4037 } |
4061 | 4038 |
4062 | 4039 |
4063 MaybeObject* Heap::AllocateExternalArray(int length, | 4040 MaybeObject* Heap::AllocateExternalArray(int length, |
4064 ExternalArrayType array_type, | 4041 ExternalArrayType array_type, |
4065 void* external_pointer, | 4042 void* external_pointer, |
4066 PretenureFlag pretenure) { | 4043 PretenureFlag pretenure) { |
4067 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; | 4044 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; |
Hannes Payer (out of office)
2013/08/20 09:49:19
You can also use the select function here.
Michael Starzinger
2013/09/23 11:35:05
Done.
| |
4068 Object* result; | 4045 Object* result; |
4069 { MaybeObject* maybe_result = AllocateRaw(ExternalArray::kAlignedSize, | 4046 { MaybeObject* maybe_result = AllocateRaw(ExternalArray::kAlignedSize, |
4070 space, | 4047 space, |
4071 OLD_DATA_SPACE); | 4048 OLD_DATA_SPACE); |
4072 if (!maybe_result->ToObject(&result)) return maybe_result; | 4049 if (!maybe_result->ToObject(&result)) return maybe_result; |
4073 } | 4050 } |
4074 | 4051 |
4075 reinterpret_cast<ExternalArray*>(result)->set_map_no_write_barrier( | 4052 reinterpret_cast<ExternalArray*>(result)->set_map_no_write_barrier( |
4076 MapForExternalArrayType(array_type)); | 4053 MapForExternalArrayType(array_type)); |
4077 reinterpret_cast<ExternalArray*>(result)->set_length(length); | 4054 reinterpret_cast<ExternalArray*>(result)->set_length(length); |
(...skipping 79 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4157 code->Verify(); | 4134 code->Verify(); |
4158 } | 4135 } |
4159 #endif | 4136 #endif |
4160 return code; | 4137 return code; |
4161 } | 4138 } |
4162 | 4139 |
4163 | 4140 |
4164 MaybeObject* Heap::CopyCode(Code* code) { | 4141 MaybeObject* Heap::CopyCode(Code* code) { |
4165 // Allocate an object the same size as the code object. | 4142 // Allocate an object the same size as the code object. |
4166 int obj_size = code->Size(); | 4143 int obj_size = code->Size(); |
4167 MaybeObject* maybe_result; | 4144 MaybeObject* maybe_result; |
Hannes Payer (out of office)
2013/08/20 09:49:19
We could also clean up the code allocation with a
Michael Starzinger
2013/09/23 11:35:05
As discussed offline: Let's do this in a separate
| |
4168 if (obj_size > code_space()->AreaSize()) { | 4145 if (obj_size > code_space()->AreaSize()) { |
4169 maybe_result = lo_space_->AllocateRaw(obj_size, EXECUTABLE); | 4146 maybe_result = lo_space_->AllocateRaw(obj_size, EXECUTABLE); |
4170 } else { | 4147 } else { |
4171 maybe_result = code_space_->AllocateRaw(obj_size); | 4148 maybe_result = code_space_->AllocateRaw(obj_size); |
4172 } | 4149 } |
4173 | 4150 |
4174 Object* result; | 4151 Object* result; |
4175 if (!maybe_result->ToObject(&result)) return maybe_result; | 4152 if (!maybe_result->ToObject(&result)) return maybe_result; |
4176 | 4153 |
4177 // Copy code object. | 4154 // Copy code object. |
(...skipping 192 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4370 // duplicate the check here. | 4347 // duplicate the check here. |
4371 ASSERT(AllowHeapAllocation::IsAllowed() && gc_state_ == NOT_IN_GC); | 4348 ASSERT(AllowHeapAllocation::IsAllowed() && gc_state_ == NOT_IN_GC); |
4372 | 4349 |
4373 // Check that the size of the boilerplate matches our | 4350 // Check that the size of the boilerplate matches our |
4374 // expectations. The ArgumentsAccessStub::GenerateNewObject relies | 4351 // expectations. The ArgumentsAccessStub::GenerateNewObject relies |
4375 // on the size being a known constant. | 4352 // on the size being a known constant. |
4376 ASSERT(arguments_object_size == boilerplate->map()->instance_size()); | 4353 ASSERT(arguments_object_size == boilerplate->map()->instance_size()); |
4377 | 4354 |
4378 // Do the allocation. | 4355 // Do the allocation. |
4379 Object* result; | 4356 Object* result; |
4380 { MaybeObject* maybe_result = | 4357 { MaybeObject* maybe_result = |
Hannes Payer (out of office)
2013/08/20 09:49:19
We could do a proper space selection here as well.
Michael Starzinger
2013/09/23 11:35:05
As discussed offline: Let's do this in a separate
| |
4381 AllocateRaw(arguments_object_size, NEW_SPACE, OLD_POINTER_SPACE); | 4358 AllocateRaw(arguments_object_size, NEW_SPACE, OLD_POINTER_SPACE); |
4382 if (!maybe_result->ToObject(&result)) return maybe_result; | 4359 if (!maybe_result->ToObject(&result)) return maybe_result; |
4383 } | 4360 } |
4384 | 4361 |
4385 // Copy the content. The arguments boilerplate doesn't have any | 4362 // Copy the content. The arguments boilerplate doesn't have any |
4386 // fields that point to new space so it's safe to skip the write | 4363 // fields that point to new space so it's safe to skip the write |
4387 // barrier here. | 4364 // barrier here. |
4388 CopyBlock(HeapObject::cast(result)->address(), | 4365 CopyBlock(HeapObject::cast(result)->address(), |
4389 boilerplate->address(), | 4366 boilerplate->address(), |
4390 JSObject::kHeaderSize); | 4367 JSObject::kHeaderSize); |
(...skipping 107 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4498 int prop_size = map->InitialPropertiesLength(); | 4475 int prop_size = map->InitialPropertiesLength(); |
4499 ASSERT(prop_size >= 0); | 4476 ASSERT(prop_size >= 0); |
4500 { MaybeObject* maybe_properties = AllocateFixedArray(prop_size, pretenure); | 4477 { MaybeObject* maybe_properties = AllocateFixedArray(prop_size, pretenure); |
4501 if (!maybe_properties->To(&properties)) return maybe_properties; | 4478 if (!maybe_properties->To(&properties)) return maybe_properties; |
4502 } | 4479 } |
4503 } else { | 4480 } else { |
4504 properties = empty_fixed_array(); | 4481 properties = empty_fixed_array(); |
4505 } | 4482 } |
4506 | 4483 |
4507 // Allocate the JSObject. | 4484 // Allocate the JSObject. |
4508 AllocationSpace space = | 4485 int size = map->instance_size(); |
4509 (pretenure == TENURED) ? OLD_POINTER_SPACE : NEW_SPACE; | 4486 AllocationSpace space = SelectSpace(size, OLD_POINTER_SPACE, pretenure); |
4510 if (map->instance_size() > Page::kMaxNonCodeHeapObjectSize) space = LO_SPACE; | |
4511 Object* obj; | 4487 Object* obj; |
4512 MaybeObject* maybe_obj = Allocate(map, space); | 4488 MaybeObject* maybe_obj = Allocate(map, space); |
4513 if (!maybe_obj->To(&obj)) return maybe_obj; | 4489 if (!maybe_obj->To(&obj)) return maybe_obj; |
4514 | 4490 |
4515 // Initialize the JSObject. | 4491 // Initialize the JSObject. |
4516 InitializeJSObjectFromMap(JSObject::cast(obj), properties, map); | 4492 InitializeJSObjectFromMap(JSObject::cast(obj), properties, map); |
4517 ASSERT(JSObject::cast(obj)->HasFastElements() || | 4493 ASSERT(JSObject::cast(obj)->HasFastElements() || |
4518 JSObject::cast(obj)->HasExternalArrayElements()); | 4494 JSObject::cast(obj)->HasExternalArrayElements()); |
4519 return obj; | 4495 return obj; |
4520 } | 4496 } |
(...skipping 12 matching lines...) Expand all Loading... | |
4533 | 4509 |
4534 // Allocate the backing storage for the properties. | 4510 // Allocate the backing storage for the properties. |
4535 int prop_size = map->InitialPropertiesLength(); | 4511 int prop_size = map->InitialPropertiesLength(); |
4536 ASSERT(prop_size >= 0); | 4512 ASSERT(prop_size >= 0); |
4537 FixedArray* properties; | 4513 FixedArray* properties; |
4538 { MaybeObject* maybe_properties = AllocateFixedArray(prop_size); | 4514 { MaybeObject* maybe_properties = AllocateFixedArray(prop_size); |
4539 if (!maybe_properties->To(&properties)) return maybe_properties; | 4515 if (!maybe_properties->To(&properties)) return maybe_properties; |
4540 } | 4516 } |
4541 | 4517 |
4542 // Allocate the JSObject. | 4518 // Allocate the JSObject. |
4543 AllocationSpace space = NEW_SPACE; | 4519 int size = map->instance_size(); |
4544 if (map->instance_size() > Page::kMaxNonCodeHeapObjectSize) space = LO_SPACE; | 4520 AllocationSpace space = SelectSpace(size, OLD_POINTER_SPACE, NOT_TENURED); |
4545 Object* obj; | 4521 Object* obj; |
4546 MaybeObject* maybe_obj = | 4522 MaybeObject* maybe_obj = |
4547 AllocateWithAllocationSite(map, space, allocation_site); | 4523 AllocateWithAllocationSite(map, space, allocation_site); |
4548 if (!maybe_obj->To(&obj)) return maybe_obj; | 4524 if (!maybe_obj->To(&obj)) return maybe_obj; |
4549 | 4525 |
4550 // Initialize the JSObject. | 4526 // Initialize the JSObject. |
4551 InitializeJSObjectFromMap(JSObject::cast(obj), properties, map); | 4527 InitializeJSObjectFromMap(JSObject::cast(obj), properties, map); |
4552 ASSERT(JSObject::cast(obj)->HasFastElements()); | 4528 ASSERT(JSObject::cast(obj)->HasFastElements()); |
4553 return obj; | 4529 return obj; |
4554 } | 4530 } |
(...skipping 332 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4887 | 4863 |
4888 // Make the clone. | 4864 // Make the clone. |
4889 Map* map = source->map(); | 4865 Map* map = source->map(); |
4890 int object_size = map->instance_size(); | 4866 int object_size = map->instance_size(); |
4891 Object* clone; | 4867 Object* clone; |
4892 | 4868 |
4893 WriteBarrierMode wb_mode = UPDATE_WRITE_BARRIER; | 4869 WriteBarrierMode wb_mode = UPDATE_WRITE_BARRIER; |
4894 | 4870 |
4895 // If we're forced to always allocate, we use the general allocation | 4871 // If we're forced to always allocate, we use the general allocation |
4896 // functions which may leave us with an object in old space. | 4872 // functions which may leave us with an object in old space. |
4897 if (always_allocate()) { | 4873 if (always_allocate()) { |
Hannes Payer (out of office)
2013/08/20 09:49:19
You could also use the select function here... but
| |
4898 { MaybeObject* maybe_clone = | 4874 { MaybeObject* maybe_clone = |
4899 AllocateRaw(object_size, NEW_SPACE, OLD_POINTER_SPACE); | 4875 AllocateRaw(object_size, NEW_SPACE, OLD_POINTER_SPACE); |
4900 if (!maybe_clone->ToObject(&clone)) return maybe_clone; | 4876 if (!maybe_clone->ToObject(&clone)) return maybe_clone; |
4901 } | 4877 } |
4902 Address clone_address = HeapObject::cast(clone)->address(); | 4878 Address clone_address = HeapObject::cast(clone)->address(); |
4903 CopyBlock(clone_address, | 4879 CopyBlock(clone_address, |
4904 source->address(), | 4880 source->address(), |
4905 object_size); | 4881 object_size); |
4906 // Update write barrier for all fields that lie beyond the header. | 4882 // Update write barrier for all fields that lie beyond the header. |
4907 RecordWrites(clone_address, | 4883 RecordWrites(clone_address, |
(...skipping 57 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4965 int object_size = map->instance_size(); | 4941 int object_size = map->instance_size(); |
4966 Object* clone; | 4942 Object* clone; |
4967 | 4943 |
4968 ASSERT(map->CanTrackAllocationSite()); | 4944 ASSERT(map->CanTrackAllocationSite()); |
4969 ASSERT(map->instance_type() == JS_ARRAY_TYPE); | 4945 ASSERT(map->instance_type() == JS_ARRAY_TYPE); |
4970 WriteBarrierMode wb_mode = UPDATE_WRITE_BARRIER; | 4946 WriteBarrierMode wb_mode = UPDATE_WRITE_BARRIER; |
4971 | 4947 |
4972 // If we're forced to always allocate, we use the general allocation | 4948 // If we're forced to always allocate, we use the general allocation |
4973 // functions which may leave us with an object in old space. | 4949 // functions which may leave us with an object in old space. |
4974 int adjusted_object_size = object_size; | 4950 int adjusted_object_size = object_size; |
4975 if (always_allocate()) { | 4951 if (always_allocate()) { |
Hannes Payer (out of office)
2013/08/20 09:49:19
Likewise.
| |
4976 // We'll only track origin if we are certain to allocate in new space | 4952 // We'll only track origin if we are certain to allocate in new space |
4977 const int kMinFreeNewSpaceAfterGC = InitialSemiSpaceSize() * 3/4; | 4953 const int kMinFreeNewSpaceAfterGC = InitialSemiSpaceSize() * 3/4; |
4978 if ((object_size + AllocationMemento::kSize) < kMinFreeNewSpaceAfterGC) { | 4954 if ((object_size + AllocationMemento::kSize) < kMinFreeNewSpaceAfterGC) { |
4979 adjusted_object_size += AllocationMemento::kSize; | 4955 adjusted_object_size += AllocationMemento::kSize; |
4980 } | 4956 } |
4981 | 4957 |
4982 { MaybeObject* maybe_clone = | 4958 { MaybeObject* maybe_clone = |
4983 AllocateRaw(adjusted_object_size, NEW_SPACE, OLD_POINTER_SPACE); | 4959 AllocateRaw(adjusted_object_size, NEW_SPACE, OLD_POINTER_SPACE); |
4984 if (!maybe_clone->ToObject(&clone)) return maybe_clone; | 4960 if (!maybe_clone->ToObject(&clone)) return maybe_clone; |
4985 } | 4961 } |
(...skipping 325 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
5311 } | 5287 } |
5312 map = ascii_internalized_string_map(); | 5288 map = ascii_internalized_string_map(); |
5313 size = SeqOneByteString::SizeFor(chars); | 5289 size = SeqOneByteString::SizeFor(chars); |
5314 } else { | 5290 } else { |
5315 if (chars > SeqTwoByteString::kMaxLength) { | 5291 if (chars > SeqTwoByteString::kMaxLength) { |
5316 return Failure::OutOfMemoryException(0xa); | 5292 return Failure::OutOfMemoryException(0xa); |
5317 } | 5293 } |
5318 map = internalized_string_map(); | 5294 map = internalized_string_map(); |
5319 size = SeqTwoByteString::SizeFor(chars); | 5295 size = SeqTwoByteString::SizeFor(chars); |
5320 } | 5296 } |
5297 AllocationSpace space = SelectSpace(size, OLD_DATA_SPACE, TENURED); | |
5321 | 5298 |
5322 // Allocate string. | 5299 // Allocate string. |
5323 Object* result; | 5300 Object* result; |
5324 { MaybeObject* maybe_result = (size > Page::kMaxNonCodeHeapObjectSize) | 5301 { MaybeObject* maybe_result = AllocateRaw(size, space, OLD_DATA_SPACE); |
5325 ? lo_space_->AllocateRaw(size, NOT_EXECUTABLE) | |
5326 : old_data_space_->AllocateRaw(size); | |
5327 if (!maybe_result->ToObject(&result)) return maybe_result; | 5302 if (!maybe_result->ToObject(&result)) return maybe_result; |
5328 } | 5303 } |
5329 | 5304 |
5330 reinterpret_cast<HeapObject*>(result)->set_map_no_write_barrier(map); | 5305 reinterpret_cast<HeapObject*>(result)->set_map_no_write_barrier(map); |
5331 // Set length and hash fields of the allocated string. | 5306 // Set length and hash fields of the allocated string. |
5332 String* answer = String::cast(result); | 5307 String* answer = String::cast(result); |
5333 answer->set_length(chars); | 5308 answer->set_length(chars); |
5334 answer->set_hash_field(hash_field); | 5309 answer->set_hash_field(hash_field); |
5335 | 5310 |
5336 ASSERT_EQ(size, answer->Size()); | 5311 ASSERT_EQ(size, answer->Size()); |
(...skipping 18 matching lines...) Expand all Loading... | |
5355 Vector<const char>, int, uint32_t); | 5330 Vector<const char>, int, uint32_t); |
5356 | 5331 |
5357 | 5332 |
5358 MaybeObject* Heap::AllocateRawOneByteString(int length, | 5333 MaybeObject* Heap::AllocateRawOneByteString(int length, |
5359 PretenureFlag pretenure) { | 5334 PretenureFlag pretenure) { |
5360 if (length < 0 || length > SeqOneByteString::kMaxLength) { | 5335 if (length < 0 || length > SeqOneByteString::kMaxLength) { |
5361 return Failure::OutOfMemoryException(0xb); | 5336 return Failure::OutOfMemoryException(0xb); |
5362 } | 5337 } |
5363 int size = SeqOneByteString::SizeFor(length); | 5338 int size = SeqOneByteString::SizeFor(length); |
5364 ASSERT(size <= SeqOneByteString::kMaxSize); | 5339 ASSERT(size <= SeqOneByteString::kMaxSize); |
5365 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; | 5340 AllocationSpace space = SelectSpace(size, OLD_DATA_SPACE, pretenure); |
5366 AllocationSpace retry_space = OLD_DATA_SPACE; | |
5367 | |
5368 if (size > Page::kMaxNonCodeHeapObjectSize) { | |
5369 // Allocate in large object space, retry space will be ignored. | |
5370 space = LO_SPACE; | |
5371 } | |
5372 | 5341 |
5373 Object* result; | 5342 Object* result; |
5374 { MaybeObject* maybe_result = AllocateRaw(size, space, retry_space); | 5343 { MaybeObject* maybe_result = AllocateRaw(size, space, OLD_DATA_SPACE); |
5375 if (!maybe_result->ToObject(&result)) return maybe_result; | 5344 if (!maybe_result->ToObject(&result)) return maybe_result; |
5376 } | 5345 } |
5377 | 5346 |
5378 // Partially initialize the object. | 5347 // Partially initialize the object. |
5379 HeapObject::cast(result)->set_map_no_write_barrier(ascii_string_map()); | 5348 HeapObject::cast(result)->set_map_no_write_barrier(ascii_string_map()); |
5380 String::cast(result)->set_length(length); | 5349 String::cast(result)->set_length(length); |
5381 String::cast(result)->set_hash_field(String::kEmptyHashField); | 5350 String::cast(result)->set_hash_field(String::kEmptyHashField); |
5382 ASSERT_EQ(size, HeapObject::cast(result)->Size()); | 5351 ASSERT_EQ(size, HeapObject::cast(result)->Size()); |
5383 | 5352 |
5384 return result; | 5353 return result; |
5385 } | 5354 } |
5386 | 5355 |
5387 | 5356 |
5388 MaybeObject* Heap::AllocateRawTwoByteString(int length, | 5357 MaybeObject* Heap::AllocateRawTwoByteString(int length, |
5389 PretenureFlag pretenure) { | 5358 PretenureFlag pretenure) { |
5390 if (length < 0 || length > SeqTwoByteString::kMaxLength) { | 5359 if (length < 0 || length > SeqTwoByteString::kMaxLength) { |
5391 return Failure::OutOfMemoryException(0xc); | 5360 return Failure::OutOfMemoryException(0xc); |
5392 } | 5361 } |
5393 int size = SeqTwoByteString::SizeFor(length); | 5362 int size = SeqTwoByteString::SizeFor(length); |
5394 ASSERT(size <= SeqTwoByteString::kMaxSize); | 5363 ASSERT(size <= SeqTwoByteString::kMaxSize); |
5395 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; | 5364 AllocationSpace space = SelectSpace(size, OLD_DATA_SPACE, pretenure); |
5396 AllocationSpace retry_space = OLD_DATA_SPACE; | |
5397 | |
5398 if (size > Page::kMaxNonCodeHeapObjectSize) { | |
5399 // Allocate in large object space, retry space will be ignored. | |
5400 space = LO_SPACE; | |
5401 } | |
5402 | 5365 |
5403 Object* result; | 5366 Object* result; |
5404 { MaybeObject* maybe_result = AllocateRaw(size, space, retry_space); | 5367 { MaybeObject* maybe_result = AllocateRaw(size, space, OLD_DATA_SPACE); |
5405 if (!maybe_result->ToObject(&result)) return maybe_result; | 5368 if (!maybe_result->ToObject(&result)) return maybe_result; |
5406 } | 5369 } |
5407 | 5370 |
5408 // Partially initialize the object. | 5371 // Partially initialize the object. |
5409 HeapObject::cast(result)->set_map_no_write_barrier(string_map()); | 5372 HeapObject::cast(result)->set_map_no_write_barrier(string_map()); |
5410 String::cast(result)->set_length(length); | 5373 String::cast(result)->set_length(length); |
5411 String::cast(result)->set_hash_field(String::kEmptyHashField); | 5374 String::cast(result)->set_hash_field(String::kEmptyHashField); |
5412 ASSERT_EQ(size, HeapObject::cast(result)->Size()); | 5375 ASSERT_EQ(size, HeapObject::cast(result)->Size()); |
5413 return result; | 5376 return result; |
5414 } | 5377 } |
(...skipping 26 matching lines...) Expand all Loading... | |
5441 } | 5404 } |
5442 } | 5405 } |
5443 return AllocateJSObjectFromMapWithAllocationSite(map, allocation_site); | 5406 return AllocateJSObjectFromMapWithAllocationSite(map, allocation_site); |
5444 } | 5407 } |
5445 | 5408 |
5446 | 5409 |
5447 MaybeObject* Heap::AllocateEmptyFixedArray() { | 5410 MaybeObject* Heap::AllocateEmptyFixedArray() { |
5448 int size = FixedArray::SizeFor(0); | 5411 int size = FixedArray::SizeFor(0); |
5449 Object* result; | 5412 Object* result; |
5450 { MaybeObject* maybe_result = | 5413 { MaybeObject* maybe_result = |
5451 AllocateRaw(size, OLD_DATA_SPACE, OLD_DATA_SPACE); | 5414 AllocateRaw(size, OLD_DATA_SPACE, OLD_DATA_SPACE); |
Hannes Payer (out of office)
2013/08/20 09:49:19
For cases like that we could add an AllocateRaw(si
Michael Starzinger
2013/09/23 11:35:05
As discussed offline: Let's do this in a separate
| |
5452 if (!maybe_result->ToObject(&result)) return maybe_result; | 5415 if (!maybe_result->ToObject(&result)) return maybe_result; |
5453 } | 5416 } |
5454 // Initialize the object. | 5417 // Initialize the object. |
5455 reinterpret_cast<FixedArray*>(result)->set_map_no_write_barrier( | 5418 reinterpret_cast<FixedArray*>(result)->set_map_no_write_barrier( |
5456 fixed_array_map()); | 5419 fixed_array_map()); |
5457 reinterpret_cast<FixedArray*>(result)->set_length(0); | 5420 reinterpret_cast<FixedArray*>(result)->set_length(0); |
5458 return result; | 5421 return result; |
5459 } | 5422 } |
5460 | 5423 |
5461 | 5424 |
5462 MaybeObject* Heap::AllocateEmptyExternalArray(ExternalArrayType array_type) { | 5425 MaybeObject* Heap::AllocateEmptyExternalArray(ExternalArrayType array_type) { |
5463 return AllocateExternalArray(0, array_type, NULL, TENURED); | 5426 return AllocateExternalArray(0, array_type, NULL, TENURED); |
5464 } | 5427 } |
5465 | 5428 |
5466 | 5429 |
5467 MaybeObject* Heap::AllocateRawFixedArray(int length) { | 5430 MaybeObject* Heap::AllocateRawFixedArray(int length) { |
5468 if (length < 0 || length > FixedArray::kMaxLength) { | 5431 if (length < 0 || length > FixedArray::kMaxLength) { |
5469 return Failure::OutOfMemoryException(0xd); | 5432 return Failure::OutOfMemoryException(0xd); |
5470 } | 5433 } |
5471 ASSERT(length > 0); | 5434 ASSERT(length > 0); |
5472 // Use the general function if we're forced to always allocate. | 5435 // Use the general function if we're forced to always allocate. |
5473 if (always_allocate()) return AllocateFixedArray(length, TENURED); | 5436 if (always_allocate()) return AllocateFixedArray(length, TENURED); |
5474 // Allocate the raw data for a fixed array. | 5437 // Allocate the raw data for a fixed array. |
5475 int size = FixedArray::SizeFor(length); | 5438 int size = FixedArray::SizeFor(length); |
5476 return size <= Page::kMaxNonCodeHeapObjectSize | 5439 return size <= Page::kMaxNonCodeHeapObjectSize |
Hannes Payer (out of office)
2013/08/20 09:49:19
Can we merge this function with AllocateRawFixedAr
Michael Starzinger
2013/09/23 11:35:05
As discussed offline: Let's do this in a separate
| |
5477 ? new_space_.AllocateRaw(size) | 5440 ? new_space_.AllocateRaw(size) |
5478 : lo_space_->AllocateRaw(size, NOT_EXECUTABLE); | 5441 : lo_space_->AllocateRaw(size, NOT_EXECUTABLE); |
5479 } | 5442 } |
5480 | 5443 |
5481 | 5444 |
5482 MaybeObject* Heap::CopyFixedArrayWithMap(FixedArray* src, Map* map) { | 5445 MaybeObject* Heap::CopyFixedArrayWithMap(FixedArray* src, Map* map) { |
5483 int len = src->length(); | 5446 int len = src->length(); |
5484 Object* obj; | 5447 Object* obj; |
5485 { MaybeObject* maybe_obj = AllocateRawFixedArray(len); | 5448 { MaybeObject* maybe_obj = AllocateRawFixedArray(len); |
5486 if (!maybe_obj->ToObject(&obj)) return maybe_obj; | 5449 if (!maybe_obj->ToObject(&obj)) return maybe_obj; |
(...skipping 51 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
5538 MemsetPointer(array->data_start(), undefined_value(), length); | 5501 MemsetPointer(array->data_start(), undefined_value(), length); |
5539 return result; | 5502 return result; |
5540 } | 5503 } |
5541 | 5504 |
5542 | 5505 |
5543 MaybeObject* Heap::AllocateRawFixedArray(int length, PretenureFlag pretenure) { | 5506 MaybeObject* Heap::AllocateRawFixedArray(int length, PretenureFlag pretenure) { |
5544 if (length < 0 || length > FixedArray::kMaxLength) { | 5507 if (length < 0 || length > FixedArray::kMaxLength) { |
5545 return Failure::OutOfMemoryException(0xe); | 5508 return Failure::OutOfMemoryException(0xe); |
5546 } | 5509 } |
5547 int size = FixedArray::SizeFor(length); | 5510 int size = FixedArray::SizeFor(length); |
5548 AllocationSpace space = | 5511 AllocationSpace space = SelectSpace(size, OLD_POINTER_SPACE, pretenure); |
5549 (pretenure == TENURED) ? OLD_POINTER_SPACE : NEW_SPACE; | |
5550 AllocationSpace retry_space = OLD_POINTER_SPACE; | |
5551 | 5512 |
5552 if (size > Page::kMaxNonCodeHeapObjectSize) { | 5513 return AllocateRaw(size, space, OLD_POINTER_SPACE); |
5553 // Allocate in large object space, retry space will be ignored. | |
5554 space = LO_SPACE; | |
5555 } | |
5556 | |
5557 return AllocateRaw(size, space, retry_space); | |
5558 } | 5514 } |
5559 | 5515 |
5560 | 5516 |
5561 MUST_USE_RESULT static MaybeObject* AllocateFixedArrayWithFiller( | 5517 MUST_USE_RESULT static MaybeObject* AllocateFixedArrayWithFiller( |
5562 Heap* heap, | 5518 Heap* heap, |
5563 int length, | 5519 int length, |
5564 PretenureFlag pretenure, | 5520 PretenureFlag pretenure, |
5565 Object* filler) { | 5521 Object* filler) { |
5566 ASSERT(length >= 0); | 5522 ASSERT(length >= 0); |
5567 ASSERT(heap->empty_fixed_array()->IsFixedArray()); | 5523 ASSERT(heap->empty_fixed_array()->IsFixedArray()); |
(...skipping 42 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
5610 fixed_array_map()); | 5566 fixed_array_map()); |
5611 FixedArray::cast(obj)->set_length(length); | 5567 FixedArray::cast(obj)->set_length(length); |
5612 return obj; | 5568 return obj; |
5613 } | 5569 } |
5614 | 5570 |
5615 | 5571 |
5616 MaybeObject* Heap::AllocateEmptyFixedDoubleArray() { | 5572 MaybeObject* Heap::AllocateEmptyFixedDoubleArray() { |
5617 int size = FixedDoubleArray::SizeFor(0); | 5573 int size = FixedDoubleArray::SizeFor(0); |
5618 Object* result; | 5574 Object* result; |
5619 { MaybeObject* maybe_result = | 5575 { MaybeObject* maybe_result = |
5620 AllocateRaw(size, OLD_DATA_SPACE, OLD_DATA_SPACE); | 5576 AllocateRaw(size, OLD_DATA_SPACE, OLD_DATA_SPACE); |
Hannes Payer (out of office)
2013/08/20 09:49:19
here we could use the AllocateRaw(size, space) wra
Michael Starzinger
2013/09/23 11:35:05
As discussed offline: Let's do this in a separate
| |
5621 if (!maybe_result->ToObject(&result)) return maybe_result; | 5577 if (!maybe_result->ToObject(&result)) return maybe_result; |
5622 } | 5578 } |
5623 // Initialize the object. | 5579 // Initialize the object. |
5624 reinterpret_cast<FixedDoubleArray*>(result)->set_map_no_write_barrier( | 5580 reinterpret_cast<FixedDoubleArray*>(result)->set_map_no_write_barrier( |
5625 fixed_double_array_map()); | 5581 fixed_double_array_map()); |
5626 reinterpret_cast<FixedDoubleArray*>(result)->set_length(0); | 5582 reinterpret_cast<FixedDoubleArray*>(result)->set_length(0); |
5627 return result; | 5583 return result; |
5628 } | 5584 } |
5629 | 5585 |
5630 | 5586 |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
5665 return elements; | 5621 return elements; |
5666 } | 5622 } |
5667 | 5623 |
5668 | 5624 |
5669 MaybeObject* Heap::AllocateRawFixedDoubleArray(int length, | 5625 MaybeObject* Heap::AllocateRawFixedDoubleArray(int length, |
5670 PretenureFlag pretenure) { | 5626 PretenureFlag pretenure) { |
5671 if (length < 0 || length > FixedDoubleArray::kMaxLength) { | 5627 if (length < 0 || length > FixedDoubleArray::kMaxLength) { |
5672 return Failure::OutOfMemoryException(0xf); | 5628 return Failure::OutOfMemoryException(0xf); |
5673 } | 5629 } |
5674 int size = FixedDoubleArray::SizeFor(length); | 5630 int size = FixedDoubleArray::SizeFor(length); |
5675 AllocationSpace space = (pretenure == TENURED) ? OLD_DATA_SPACE : NEW_SPACE; | |
5676 AllocationSpace retry_space = OLD_DATA_SPACE; | |
5677 | |
5678 #ifndef V8_HOST_ARCH_64_BIT | 5631 #ifndef V8_HOST_ARCH_64_BIT |
5679 size += kPointerSize; | 5632 size += kPointerSize; |
5680 #endif | 5633 #endif |
5681 | 5634 AllocationSpace space = SelectSpace(size, OLD_DATA_SPACE, pretenure); |
5682 if (size > Page::kMaxNonCodeHeapObjectSize) { | |
5683 // Allocate in large object space, retry space will be ignored. | |
5684 space = LO_SPACE; | |
5685 } | |
5686 | 5635 |
5687 HeapObject* object; | 5636 HeapObject* object; |
5688 { MaybeObject* maybe_object = AllocateRaw(size, space, retry_space); | 5637 { MaybeObject* maybe_object = AllocateRaw(size, space, OLD_DATA_SPACE); |
5689 if (!maybe_object->To<HeapObject>(&object)) return maybe_object; | 5638 if (!maybe_object->To<HeapObject>(&object)) return maybe_object; |
5690 } | 5639 } |
5691 | 5640 |
5692 return EnsureDoubleAligned(this, object, size); | 5641 return EnsureDoubleAligned(this, object, size); |
5693 } | 5642 } |
5694 | 5643 |
5695 | 5644 |
5696 MaybeObject* Heap::AllocateHashTable(int length, PretenureFlag pretenure) { | 5645 MaybeObject* Heap::AllocateHashTable(int length, PretenureFlag pretenure) { |
5697 Object* result; | 5646 Object* result; |
5698 { MaybeObject* maybe_result = AllocateFixedArray(length, pretenure); | 5647 { MaybeObject* maybe_result = AllocateFixedArray(length, pretenure); |
(...skipping 183 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
5882 switch (type) { | 5831 switch (type) { |
5883 #define MAKE_CASE(NAME, Name, name) \ | 5832 #define MAKE_CASE(NAME, Name, name) \ |
5884 case NAME##_TYPE: map = name##_map(); break; | 5833 case NAME##_TYPE: map = name##_map(); break; |
5885 STRUCT_LIST(MAKE_CASE) | 5834 STRUCT_LIST(MAKE_CASE) |
5886 #undef MAKE_CASE | 5835 #undef MAKE_CASE |
5887 default: | 5836 default: |
5888 UNREACHABLE(); | 5837 UNREACHABLE(); |
5889 return Failure::InternalError(); | 5838 return Failure::InternalError(); |
5890 } | 5839 } |
5891 int size = map->instance_size(); | 5840 int size = map->instance_size(); |
5892 AllocationSpace space = | 5841 AllocationSpace space = SelectSpace(size, OLD_POINTER_SPACE, TENURED); |
5893 (size > Page::kMaxNonCodeHeapObjectSize) ? LO_SPACE : OLD_POINTER_SPACE; | |
5894 Object* result; | 5842 Object* result; |
5895 { MaybeObject* maybe_result = Allocate(map, space); | 5843 { MaybeObject* maybe_result = Allocate(map, space); |
5896 if (!maybe_result->ToObject(&result)) return maybe_result; | 5844 if (!maybe_result->ToObject(&result)) return maybe_result; |
5897 } | 5845 } |
5898 Struct::cast(result)->InitializeBody(size); | 5846 Struct::cast(result)->InitializeBody(size); |
5899 return result; | 5847 return result; |
5900 } | 5848 } |
5901 | 5849 |
5902 | 5850 |
5903 bool Heap::IsHeapIterable() { | 5851 bool Heap::IsHeapIterable() { |
(...skipping 2141 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
8045 if (FLAG_parallel_recompilation) { | 7993 if (FLAG_parallel_recompilation) { |
8046 heap_->relocation_mutex_->Lock(); | 7994 heap_->relocation_mutex_->Lock(); |
8047 #ifdef DEBUG | 7995 #ifdef DEBUG |
8048 heap_->relocation_mutex_locked_by_optimizer_thread_ = | 7996 heap_->relocation_mutex_locked_by_optimizer_thread_ = |
8049 heap_->isolate()->optimizing_compiler_thread()->IsOptimizerThread(); | 7997 heap_->isolate()->optimizing_compiler_thread()->IsOptimizerThread(); |
8050 #endif // DEBUG | 7998 #endif // DEBUG |
8051 } | 7999 } |
8052 } | 8000 } |
8053 | 8001 |
8054 } } // namespace v8::internal | 8002 } } // namespace v8::internal |
OLD | NEW |