| Index: src/x64/lithium-codegen-x64.cc
|
| diff --git a/src/x64/lithium-codegen-x64.cc b/src/x64/lithium-codegen-x64.cc
|
| index 0bb676b9dbe12ff0690e7d06e68e0f3660ba83f6..81c67c748fadc7acd09697a5e4c73cb699a63590 100644
|
| --- a/src/x64/lithium-codegen-x64.cc
|
| +++ b/src/x64/lithium-codegen-x64.cc
|
| @@ -5206,163 +5206,6 @@ void LCodeGen::DoArrayLiteral(LArrayLiteral* instr) {
|
| }
|
|
|
|
|
| -void LCodeGen::EmitDeepCopy(Handle<JSObject> object,
|
| - Register result,
|
| - Register source,
|
| - int* offset,
|
| - AllocationSiteMode mode) {
|
| - ASSERT(!source.is(rcx));
|
| - ASSERT(!result.is(rcx));
|
| -
|
| - bool create_allocation_site_info = mode == TRACK_ALLOCATION_SITE &&
|
| - object->map()->CanTrackAllocationSite();
|
| -
|
| - // Only elements backing stores for non-COW arrays need to be copied.
|
| - Handle<FixedArrayBase> elements(object->elements());
|
| - bool has_elements = elements->length() > 0 &&
|
| - elements->map() != isolate()->heap()->fixed_cow_array_map();
|
| -
|
| - // Increase the offset so that subsequent objects end up right after
|
| - // this object and its backing store.
|
| - int object_offset = *offset;
|
| - int object_size = object->map()->instance_size();
|
| - int elements_size = has_elements ? elements->Size() : 0;
|
| - int elements_offset = *offset + object_size;
|
| - if (create_allocation_site_info) {
|
| - elements_offset += AllocationSiteInfo::kSize;
|
| - *offset += AllocationSiteInfo::kSize;
|
| - }
|
| -
|
| - *offset += object_size + elements_size;
|
| -
|
| - // Copy object header.
|
| - ASSERT(object->properties()->length() == 0);
|
| - int inobject_properties = object->map()->inobject_properties();
|
| - int header_size = object_size - inobject_properties * kPointerSize;
|
| - for (int i = 0; i < header_size; i += kPointerSize) {
|
| - if (has_elements && i == JSObject::kElementsOffset) {
|
| - __ lea(rcx, Operand(result, elements_offset));
|
| - } else {
|
| - __ movq(rcx, FieldOperand(source, i));
|
| - }
|
| - __ movq(FieldOperand(result, object_offset + i), rcx);
|
| - }
|
| -
|
| - // Copy in-object properties.
|
| - for (int i = 0; i < inobject_properties; i++) {
|
| - int total_offset = object_offset + object->GetInObjectPropertyOffset(i);
|
| - Handle<Object> value = Handle<Object>(object->InObjectPropertyAt(i),
|
| - isolate());
|
| - if (value->IsJSObject()) {
|
| - Handle<JSObject> value_object = Handle<JSObject>::cast(value);
|
| - __ lea(rcx, Operand(result, *offset));
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - __ LoadHeapObject(source, value_object);
|
| - EmitDeepCopy(value_object, result, source, offset,
|
| - DONT_TRACK_ALLOCATION_SITE);
|
| - } else if (value->IsHeapObject()) {
|
| - __ LoadHeapObject(rcx, Handle<HeapObject>::cast(value));
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - } else {
|
| - __ movq(rcx, value, RelocInfo::NONE64);
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - }
|
| - }
|
| -
|
| - // Build Allocation Site Info if desired
|
| - if (create_allocation_site_info) {
|
| - __ LoadRoot(kScratchRegister, Heap::kAllocationSiteInfoMapRootIndex);
|
| - __ movq(FieldOperand(result, object_size), kScratchRegister);
|
| - __ movq(FieldOperand(result, object_size + kPointerSize), source);
|
| - }
|
| -
|
| - if (has_elements) {
|
| - // Copy elements backing store header.
|
| - __ LoadHeapObject(source, elements);
|
| - for (int i = 0; i < FixedArray::kHeaderSize; i += kPointerSize) {
|
| - __ movq(rcx, FieldOperand(source, i));
|
| - __ movq(FieldOperand(result, elements_offset + i), rcx);
|
| - }
|
| -
|
| - // Copy elements backing store content.
|
| - int elements_length = elements->length();
|
| - if (elements->IsFixedDoubleArray()) {
|
| - Handle<FixedDoubleArray> double_array =
|
| - Handle<FixedDoubleArray>::cast(elements);
|
| - for (int i = 0; i < elements_length; i++) {
|
| - int64_t value = double_array->get_representation(i);
|
| - int total_offset =
|
| - elements_offset + FixedDoubleArray::OffsetOfElementAt(i);
|
| - __ movq(rcx, value, RelocInfo::NONE64);
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - }
|
| - } else if (elements->IsFixedArray()) {
|
| - Handle<FixedArray> fast_elements = Handle<FixedArray>::cast(elements);
|
| - for (int i = 0; i < elements_length; i++) {
|
| - int total_offset = elements_offset + FixedArray::OffsetOfElementAt(i);
|
| - Handle<Object> value(fast_elements->get(i), isolate());
|
| - if (value->IsJSObject()) {
|
| - Handle<JSObject> value_object = Handle<JSObject>::cast(value);
|
| - __ lea(rcx, Operand(result, *offset));
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - __ LoadHeapObject(source, value_object);
|
| - EmitDeepCopy(value_object, result, source, offset,
|
| - DONT_TRACK_ALLOCATION_SITE);
|
| - } else if (value->IsHeapObject()) {
|
| - __ LoadHeapObject(rcx, Handle<HeapObject>::cast(value));
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - } else {
|
| - __ movq(rcx, value, RelocInfo::NONE64);
|
| - __ movq(FieldOperand(result, total_offset), rcx);
|
| - }
|
| - }
|
| - } else {
|
| - UNREACHABLE();
|
| - }
|
| - }
|
| -}
|
| -
|
| -
|
| -void LCodeGen::DoFastLiteral(LFastLiteral* instr) {
|
| - int size = instr->hydrogen()->total_size();
|
| - ElementsKind boilerplate_elements_kind =
|
| - instr->hydrogen()->boilerplate()->GetElementsKind();
|
| -
|
| - // Deopt if the array literal boilerplate ElementsKind is of a type different
|
| - // than the expected one. The check isn't necessary if the boilerplate has
|
| - // already been converted to TERMINAL_FAST_ELEMENTS_KIND.
|
| - if (CanTransitionToMoreGeneralFastElementsKind(
|
| - boilerplate_elements_kind, true)) {
|
| - __ LoadHeapObject(rbx, instr->hydrogen()->boilerplate());
|
| - __ movq(rcx, FieldOperand(rbx, HeapObject::kMapOffset));
|
| - // Load the map's "bit field 2".
|
| - __ movb(rcx, FieldOperand(rcx, Map::kBitField2Offset));
|
| - // Retrieve elements_kind from bit field 2.
|
| - __ and_(rcx, Immediate(Map::kElementsKindMask));
|
| - __ cmpb(rcx, Immediate(boilerplate_elements_kind <<
|
| - Map::kElementsKindShift));
|
| - DeoptimizeIf(not_equal, instr->environment());
|
| - }
|
| -
|
| - // Allocate all objects that are part of the literal in one big
|
| - // allocation. This avoids multiple limit checks.
|
| - Label allocated, runtime_allocate;
|
| - __ Allocate(size, rax, rcx, rdx, &runtime_allocate, TAG_OBJECT);
|
| - __ jmp(&allocated);
|
| -
|
| - __ bind(&runtime_allocate);
|
| - __ Push(Smi::FromInt(size));
|
| - CallRuntime(Runtime::kAllocateInNewSpace, 1, instr);
|
| -
|
| - __ bind(&allocated);
|
| - int offset = 0;
|
| - __ LoadHeapObject(rbx, instr->hydrogen()->boilerplate());
|
| - EmitDeepCopy(instr->hydrogen()->boilerplate(), rax, rbx, &offset,
|
| - instr->hydrogen()->allocation_site_mode());
|
| - ASSERT_EQ(size, offset);
|
| -}
|
| -
|
| -
|
| void LCodeGen::DoObjectLiteral(LObjectLiteral* instr) {
|
| Handle<FixedArray> literals(instr->environment()->closure()->literals());
|
| Handle<FixedArray> constant_properties =
|
|
|