| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 3990 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4001 // Compare flat ASCII strings natively. Remove arguments from stack first. | 4001 // Compare flat ASCII strings natively. Remove arguments from stack first. |
| 4002 __ IncrementCounter(counters->string_compare_native(), 1, a2, a3); | 4002 __ IncrementCounter(counters->string_compare_native(), 1, a2, a3); |
| 4003 __ Addu(sp, sp, Operand(2 * kPointerSize)); | 4003 __ Addu(sp, sp, Operand(2 * kPointerSize)); |
| 4004 GenerateCompareFlatAsciiStrings(masm, a1, a0, a2, a3, t0, t1); | 4004 GenerateCompareFlatAsciiStrings(masm, a1, a0, a2, a3, t0, t1); |
| 4005 | 4005 |
| 4006 __ bind(&runtime); | 4006 __ bind(&runtime); |
| 4007 __ TailCallRuntime(Runtime::kHiddenStringCompare, 2, 1); | 4007 __ TailCallRuntime(Runtime::kHiddenStringCompare, 2, 1); |
| 4008 } | 4008 } |
| 4009 | 4009 |
| 4010 | 4010 |
| 4011 void ArrayPushStub::Generate(MacroAssembler* masm) { | |
| 4012 Register receiver = a0; | |
| 4013 Register scratch = a1; | |
| 4014 | |
| 4015 int argc = arguments_count(); | |
| 4016 | |
| 4017 if (argc == 0) { | |
| 4018 // Nothing to do, just return the length. | |
| 4019 __ lw(v0, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4020 __ DropAndRet(argc + 1); | |
| 4021 return; | |
| 4022 } | |
| 4023 | |
| 4024 Isolate* isolate = masm->isolate(); | |
| 4025 | |
| 4026 if (argc != 1) { | |
| 4027 __ TailCallExternalReference( | |
| 4028 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4029 return; | |
| 4030 } | |
| 4031 | |
| 4032 Label call_builtin, attempt_to_grow_elements, with_write_barrier; | |
| 4033 | |
| 4034 Register elements = t2; | |
| 4035 Register end_elements = t1; | |
| 4036 // Get the elements array of the object. | |
| 4037 __ lw(elements, FieldMemOperand(receiver, JSArray::kElementsOffset)); | |
| 4038 | |
| 4039 if (IsFastSmiOrObjectElementsKind(elements_kind())) { | |
| 4040 // Check that the elements are in fast mode and writable. | |
| 4041 __ CheckMap(elements, | |
| 4042 scratch, | |
| 4043 Heap::kFixedArrayMapRootIndex, | |
| 4044 &call_builtin, | |
| 4045 DONT_DO_SMI_CHECK); | |
| 4046 } | |
| 4047 | |
| 4048 // Get the array's length into scratch and calculate new length. | |
| 4049 __ lw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4050 __ Addu(scratch, scratch, Operand(Smi::FromInt(argc))); | |
| 4051 | |
| 4052 // Get the elements' length. | |
| 4053 __ lw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset)); | |
| 4054 | |
| 4055 const int kEndElementsOffset = | |
| 4056 FixedArray::kHeaderSize - kHeapObjectTag - argc * kPointerSize; | |
| 4057 | |
| 4058 if (IsFastSmiOrObjectElementsKind(elements_kind())) { | |
| 4059 // Check if we could survive without allocation. | |
| 4060 __ Branch(&attempt_to_grow_elements, gt, scratch, Operand(t0)); | |
| 4061 | |
| 4062 // Check if value is a smi. | |
| 4063 __ lw(t0, MemOperand(sp, (argc - 1) * kPointerSize)); | |
| 4064 __ JumpIfNotSmi(t0, &with_write_barrier); | |
| 4065 | |
| 4066 // Store the value. | |
| 4067 // We may need a register containing the address end_elements below, | |
| 4068 // so write back the value in end_elements. | |
| 4069 __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize); | |
| 4070 __ Addu(end_elements, elements, end_elements); | |
| 4071 __ Addu(end_elements, end_elements, kEndElementsOffset); | |
| 4072 __ sw(t0, MemOperand(end_elements)); | |
| 4073 } else { | |
| 4074 // Check if we could survive without allocation. | |
| 4075 __ Branch(&call_builtin, gt, scratch, Operand(t0)); | |
| 4076 | |
| 4077 __ lw(t0, MemOperand(sp, (argc - 1) * kPointerSize)); | |
| 4078 __ StoreNumberToDoubleElements(t0, scratch, elements, a3, t1, a2, | |
| 4079 &call_builtin, argc * kDoubleSize); | |
| 4080 } | |
| 4081 | |
| 4082 // Save new length. | |
| 4083 __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4084 __ mov(v0, scratch); | |
| 4085 __ DropAndRet(argc + 1); | |
| 4086 | |
| 4087 if (IsFastDoubleElementsKind(elements_kind())) { | |
| 4088 __ bind(&call_builtin); | |
| 4089 __ TailCallExternalReference( | |
| 4090 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4091 return; | |
| 4092 } | |
| 4093 | |
| 4094 __ bind(&with_write_barrier); | |
| 4095 | |
| 4096 if (IsFastSmiElementsKind(elements_kind())) { | |
| 4097 if (FLAG_trace_elements_transitions) __ jmp(&call_builtin); | |
| 4098 | |
| 4099 __ lw(t3, FieldMemOperand(t0, HeapObject::kMapOffset)); | |
| 4100 __ LoadRoot(at, Heap::kHeapNumberMapRootIndex); | |
| 4101 __ Branch(&call_builtin, eq, t3, Operand(at)); | |
| 4102 | |
| 4103 ElementsKind target_kind = IsHoleyElementsKind(elements_kind()) | |
| 4104 ? FAST_HOLEY_ELEMENTS : FAST_ELEMENTS; | |
| 4105 __ lw(a3, ContextOperand(cp, Context::GLOBAL_OBJECT_INDEX)); | |
| 4106 __ lw(a3, FieldMemOperand(a3, GlobalObject::kNativeContextOffset)); | |
| 4107 __ lw(a3, ContextOperand(a3, Context::JS_ARRAY_MAPS_INDEX)); | |
| 4108 const int header_size = FixedArrayBase::kHeaderSize; | |
| 4109 // Verify that the object can be transitioned in place. | |
| 4110 const int origin_offset = header_size + elements_kind() * kPointerSize; | |
| 4111 __ lw(a2, FieldMemOperand(receiver, origin_offset)); | |
| 4112 __ lw(at, FieldMemOperand(a3, HeapObject::kMapOffset)); | |
| 4113 __ Branch(&call_builtin, ne, a2, Operand(at)); | |
| 4114 | |
| 4115 | |
| 4116 const int target_offset = header_size + target_kind * kPointerSize; | |
| 4117 __ lw(a3, FieldMemOperand(a3, target_offset)); | |
| 4118 __ mov(a2, receiver); | |
| 4119 ElementsTransitionGenerator::GenerateMapChangeElementsTransition( | |
| 4120 masm, DONT_TRACK_ALLOCATION_SITE, NULL); | |
| 4121 } | |
| 4122 | |
| 4123 // Save new length. | |
| 4124 __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4125 | |
| 4126 // Store the value. | |
| 4127 // We may need a register containing the address end_elements below, so write | |
| 4128 // back the value in end_elements. | |
| 4129 __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize); | |
| 4130 __ Addu(end_elements, elements, end_elements); | |
| 4131 __ Addu(end_elements, end_elements, kEndElementsOffset); | |
| 4132 __ sw(t0, MemOperand(end_elements)); | |
| 4133 | |
| 4134 __ RecordWrite(elements, | |
| 4135 end_elements, | |
| 4136 t0, | |
| 4137 kRAHasNotBeenSaved, | |
| 4138 kDontSaveFPRegs, | |
| 4139 EMIT_REMEMBERED_SET, | |
| 4140 OMIT_SMI_CHECK); | |
| 4141 __ mov(v0, scratch); | |
| 4142 __ DropAndRet(argc + 1); | |
| 4143 | |
| 4144 __ bind(&attempt_to_grow_elements); | |
| 4145 // scratch: array's length + 1. | |
| 4146 | |
| 4147 if (!FLAG_inline_new) { | |
| 4148 __ bind(&call_builtin); | |
| 4149 __ TailCallExternalReference( | |
| 4150 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4151 return; | |
| 4152 } | |
| 4153 | |
| 4154 __ lw(a2, MemOperand(sp, (argc - 1) * kPointerSize)); | |
| 4155 // Growing elements that are SMI-only requires special handling in case the | |
| 4156 // new element is non-Smi. For now, delegate to the builtin. | |
| 4157 if (IsFastSmiElementsKind(elements_kind())) { | |
| 4158 __ JumpIfNotSmi(a2, &call_builtin); | |
| 4159 } | |
| 4160 | |
| 4161 // We could be lucky and the elements array could be at the top of new-space. | |
| 4162 // In this case we can just grow it in place by moving the allocation pointer | |
| 4163 // up. | |
| 4164 ExternalReference new_space_allocation_top = | |
| 4165 ExternalReference::new_space_allocation_top_address(isolate); | |
| 4166 ExternalReference new_space_allocation_limit = | |
| 4167 ExternalReference::new_space_allocation_limit_address(isolate); | |
| 4168 | |
| 4169 const int kAllocationDelta = 4; | |
| 4170 ASSERT(kAllocationDelta >= argc); | |
| 4171 // Load top and check if it is the end of elements. | |
| 4172 __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize); | |
| 4173 __ Addu(end_elements, elements, end_elements); | |
| 4174 __ Addu(end_elements, end_elements, Operand(kEndElementsOffset)); | |
| 4175 __ li(t0, Operand(new_space_allocation_top)); | |
| 4176 __ lw(a3, MemOperand(t0)); | |
| 4177 __ Branch(&call_builtin, ne, a3, Operand(end_elements)); | |
| 4178 | |
| 4179 __ li(t3, Operand(new_space_allocation_limit)); | |
| 4180 __ lw(t3, MemOperand(t3)); | |
| 4181 __ Addu(a3, a3, Operand(kAllocationDelta * kPointerSize)); | |
| 4182 __ Branch(&call_builtin, hi, a3, Operand(t3)); | |
| 4183 | |
| 4184 // We fit and could grow elements. | |
| 4185 // Update new_space_allocation_top. | |
| 4186 __ sw(a3, MemOperand(t0)); | |
| 4187 // Push the argument. | |
| 4188 __ sw(a2, MemOperand(end_elements)); | |
| 4189 // Fill the rest with holes. | |
| 4190 __ LoadRoot(a3, Heap::kTheHoleValueRootIndex); | |
| 4191 for (int i = 1; i < kAllocationDelta; i++) { | |
| 4192 __ sw(a3, MemOperand(end_elements, i * kPointerSize)); | |
| 4193 } | |
| 4194 | |
| 4195 // Update elements' and array's sizes. | |
| 4196 __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4197 __ lw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset)); | |
| 4198 __ Addu(t0, t0, Operand(Smi::FromInt(kAllocationDelta))); | |
| 4199 __ sw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset)); | |
| 4200 | |
| 4201 // Elements are in new space, so write barrier is not required. | |
| 4202 __ mov(v0, scratch); | |
| 4203 __ DropAndRet(argc + 1); | |
| 4204 | |
| 4205 __ bind(&call_builtin); | |
| 4206 __ TailCallExternalReference( | |
| 4207 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4208 } | |
| 4209 | |
| 4210 | |
| 4211 void BinaryOpICWithAllocationSiteStub::Generate(MacroAssembler* masm) { | 4011 void BinaryOpICWithAllocationSiteStub::Generate(MacroAssembler* masm) { |
| 4212 // ----------- S t a t e ------------- | 4012 // ----------- S t a t e ------------- |
| 4213 // -- a1 : left | 4013 // -- a1 : left |
| 4214 // -- a0 : right | 4014 // -- a0 : right |
| 4215 // -- ra : return address | 4015 // -- ra : return address |
| 4216 // ----------------------------------- | 4016 // ----------------------------------- |
| 4217 Isolate* isolate = masm->isolate(); | 4017 Isolate* isolate = masm->isolate(); |
| 4218 | 4018 |
| 4219 // Load a2 with the allocation site. We stick an undefined dummy value here | 4019 // Load a2 with the allocation site. We stick an undefined dummy value here |
| 4220 // and replace it with the real allocation site later when we instantiate this | 4020 // and replace it with the real allocation site later when we instantiate this |
| (...skipping 1432 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5653 MemOperand(fp, 6 * kPointerSize), | 5453 MemOperand(fp, 6 * kPointerSize), |
| 5654 NULL); | 5454 NULL); |
| 5655 } | 5455 } |
| 5656 | 5456 |
| 5657 | 5457 |
| 5658 #undef __ | 5458 #undef __ |
| 5659 | 5459 |
| 5660 } } // namespace v8::internal | 5460 } } // namespace v8::internal |
| 5661 | 5461 |
| 5662 #endif // V8_TARGET_ARCH_MIPS | 5462 #endif // V8_TARGET_ARCH_MIPS |
| OLD | NEW |