| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 3981 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3992 // Compare flat ASCII strings natively. Remove arguments from stack first. | 3992 // Compare flat ASCII strings natively. Remove arguments from stack first. |
| 3993 __ IncrementCounter(counters->string_compare_native(), 1, a2, a3); | 3993 __ IncrementCounter(counters->string_compare_native(), 1, a2, a3); |
| 3994 __ Addu(sp, sp, Operand(2 * kPointerSize)); | 3994 __ Addu(sp, sp, Operand(2 * kPointerSize)); |
| 3995 GenerateCompareFlatAsciiStrings(masm, a1, a0, a2, a3, t0, t1); | 3995 GenerateCompareFlatAsciiStrings(masm, a1, a0, a2, a3, t0, t1); |
| 3996 | 3996 |
| 3997 __ bind(&runtime); | 3997 __ bind(&runtime); |
| 3998 __ TailCallRuntime(Runtime::kHiddenStringCompare, 2, 1); | 3998 __ TailCallRuntime(Runtime::kHiddenStringCompare, 2, 1); |
| 3999 } | 3999 } |
| 4000 | 4000 |
| 4001 | 4001 |
| 4002 void ArrayPushStub::Generate(MacroAssembler* masm) { | |
| 4003 Register receiver = a0; | |
| 4004 Register scratch = a1; | |
| 4005 | |
| 4006 int argc = arguments_count(); | |
| 4007 | |
| 4008 if (argc == 0) { | |
| 4009 // Nothing to do, just return the length. | |
| 4010 __ lw(v0, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4011 __ DropAndRet(argc + 1); | |
| 4012 return; | |
| 4013 } | |
| 4014 | |
| 4015 Isolate* isolate = masm->isolate(); | |
| 4016 | |
| 4017 if (argc != 1) { | |
| 4018 __ TailCallExternalReference( | |
| 4019 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4020 return; | |
| 4021 } | |
| 4022 | |
| 4023 Label call_builtin, attempt_to_grow_elements, with_write_barrier; | |
| 4024 | |
| 4025 Register elements = t2; | |
| 4026 Register end_elements = t1; | |
| 4027 // Get the elements array of the object. | |
| 4028 __ lw(elements, FieldMemOperand(receiver, JSArray::kElementsOffset)); | |
| 4029 | |
| 4030 if (IsFastSmiOrObjectElementsKind(elements_kind())) { | |
| 4031 // Check that the elements are in fast mode and writable. | |
| 4032 __ CheckMap(elements, | |
| 4033 scratch, | |
| 4034 Heap::kFixedArrayMapRootIndex, | |
| 4035 &call_builtin, | |
| 4036 DONT_DO_SMI_CHECK); | |
| 4037 } | |
| 4038 | |
| 4039 // Get the array's length into scratch and calculate new length. | |
| 4040 __ lw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4041 __ Addu(scratch, scratch, Operand(Smi::FromInt(argc))); | |
| 4042 | |
| 4043 // Get the elements' length. | |
| 4044 __ lw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset)); | |
| 4045 | |
| 4046 const int kEndElementsOffset = | |
| 4047 FixedArray::kHeaderSize - kHeapObjectTag - argc * kPointerSize; | |
| 4048 | |
| 4049 if (IsFastSmiOrObjectElementsKind(elements_kind())) { | |
| 4050 // Check if we could survive without allocation. | |
| 4051 __ Branch(&attempt_to_grow_elements, gt, scratch, Operand(t0)); | |
| 4052 | |
| 4053 // Check if value is a smi. | |
| 4054 __ lw(t0, MemOperand(sp, (argc - 1) * kPointerSize)); | |
| 4055 __ JumpIfNotSmi(t0, &with_write_barrier); | |
| 4056 | |
| 4057 // Store the value. | |
| 4058 // We may need a register containing the address end_elements below, | |
| 4059 // so write back the value in end_elements. | |
| 4060 __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize); | |
| 4061 __ Addu(end_elements, elements, end_elements); | |
| 4062 __ Addu(end_elements, end_elements, kEndElementsOffset); | |
| 4063 __ sw(t0, MemOperand(end_elements)); | |
| 4064 } else { | |
| 4065 // Check if we could survive without allocation. | |
| 4066 __ Branch(&call_builtin, gt, scratch, Operand(t0)); | |
| 4067 | |
| 4068 __ lw(t0, MemOperand(sp, (argc - 1) * kPointerSize)); | |
| 4069 __ StoreNumberToDoubleElements(t0, scratch, elements, a3, t1, a2, | |
| 4070 &call_builtin, argc * kDoubleSize); | |
| 4071 } | |
| 4072 | |
| 4073 // Save new length. | |
| 4074 __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4075 __ mov(v0, scratch); | |
| 4076 __ DropAndRet(argc + 1); | |
| 4077 | |
| 4078 if (IsFastDoubleElementsKind(elements_kind())) { | |
| 4079 __ bind(&call_builtin); | |
| 4080 __ TailCallExternalReference( | |
| 4081 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4082 return; | |
| 4083 } | |
| 4084 | |
| 4085 __ bind(&with_write_barrier); | |
| 4086 | |
| 4087 if (IsFastSmiElementsKind(elements_kind())) { | |
| 4088 if (FLAG_trace_elements_transitions) __ jmp(&call_builtin); | |
| 4089 | |
| 4090 __ lw(t3, FieldMemOperand(t0, HeapObject::kMapOffset)); | |
| 4091 __ LoadRoot(at, Heap::kHeapNumberMapRootIndex); | |
| 4092 __ Branch(&call_builtin, eq, t3, Operand(at)); | |
| 4093 | |
| 4094 ElementsKind target_kind = IsHoleyElementsKind(elements_kind()) | |
| 4095 ? FAST_HOLEY_ELEMENTS : FAST_ELEMENTS; | |
| 4096 __ lw(a3, ContextOperand(cp, Context::GLOBAL_OBJECT_INDEX)); | |
| 4097 __ lw(a3, FieldMemOperand(a3, GlobalObject::kNativeContextOffset)); | |
| 4098 __ lw(a3, ContextOperand(a3, Context::JS_ARRAY_MAPS_INDEX)); | |
| 4099 const int header_size = FixedArrayBase::kHeaderSize; | |
| 4100 // Verify that the object can be transitioned in place. | |
| 4101 const int origin_offset = header_size + elements_kind() * kPointerSize; | |
| 4102 __ lw(a2, FieldMemOperand(receiver, origin_offset)); | |
| 4103 __ lw(at, FieldMemOperand(a3, HeapObject::kMapOffset)); | |
| 4104 __ Branch(&call_builtin, ne, a2, Operand(at)); | |
| 4105 | |
| 4106 | |
| 4107 const int target_offset = header_size + target_kind * kPointerSize; | |
| 4108 __ lw(a3, FieldMemOperand(a3, target_offset)); | |
| 4109 __ mov(a2, receiver); | |
| 4110 ElementsTransitionGenerator::GenerateMapChangeElementsTransition( | |
| 4111 masm, DONT_TRACK_ALLOCATION_SITE, NULL); | |
| 4112 } | |
| 4113 | |
| 4114 // Save new length. | |
| 4115 __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4116 | |
| 4117 // Store the value. | |
| 4118 // We may need a register containing the address end_elements below, so write | |
| 4119 // back the value in end_elements. | |
| 4120 __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize); | |
| 4121 __ Addu(end_elements, elements, end_elements); | |
| 4122 __ Addu(end_elements, end_elements, kEndElementsOffset); | |
| 4123 __ sw(t0, MemOperand(end_elements)); | |
| 4124 | |
| 4125 __ RecordWrite(elements, | |
| 4126 end_elements, | |
| 4127 t0, | |
| 4128 kRAHasNotBeenSaved, | |
| 4129 kDontSaveFPRegs, | |
| 4130 EMIT_REMEMBERED_SET, | |
| 4131 OMIT_SMI_CHECK); | |
| 4132 __ mov(v0, scratch); | |
| 4133 __ DropAndRet(argc + 1); | |
| 4134 | |
| 4135 __ bind(&attempt_to_grow_elements); | |
| 4136 // scratch: array's length + 1. | |
| 4137 | |
| 4138 if (!FLAG_inline_new) { | |
| 4139 __ bind(&call_builtin); | |
| 4140 __ TailCallExternalReference( | |
| 4141 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4142 return; | |
| 4143 } | |
| 4144 | |
| 4145 __ lw(a2, MemOperand(sp, (argc - 1) * kPointerSize)); | |
| 4146 // Growing elements that are SMI-only requires special handling in case the | |
| 4147 // new element is non-Smi. For now, delegate to the builtin. | |
| 4148 if (IsFastSmiElementsKind(elements_kind())) { | |
| 4149 __ JumpIfNotSmi(a2, &call_builtin); | |
| 4150 } | |
| 4151 | |
| 4152 // We could be lucky and the elements array could be at the top of new-space. | |
| 4153 // In this case we can just grow it in place by moving the allocation pointer | |
| 4154 // up. | |
| 4155 ExternalReference new_space_allocation_top = | |
| 4156 ExternalReference::new_space_allocation_top_address(isolate); | |
| 4157 ExternalReference new_space_allocation_limit = | |
| 4158 ExternalReference::new_space_allocation_limit_address(isolate); | |
| 4159 | |
| 4160 const int kAllocationDelta = 4; | |
| 4161 ASSERT(kAllocationDelta >= argc); | |
| 4162 // Load top and check if it is the end of elements. | |
| 4163 __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize); | |
| 4164 __ Addu(end_elements, elements, end_elements); | |
| 4165 __ Addu(end_elements, end_elements, Operand(kEndElementsOffset)); | |
| 4166 __ li(t0, Operand(new_space_allocation_top)); | |
| 4167 __ lw(a3, MemOperand(t0)); | |
| 4168 __ Branch(&call_builtin, ne, a3, Operand(end_elements)); | |
| 4169 | |
| 4170 __ li(t3, Operand(new_space_allocation_limit)); | |
| 4171 __ lw(t3, MemOperand(t3)); | |
| 4172 __ Addu(a3, a3, Operand(kAllocationDelta * kPointerSize)); | |
| 4173 __ Branch(&call_builtin, hi, a3, Operand(t3)); | |
| 4174 | |
| 4175 // We fit and could grow elements. | |
| 4176 // Update new_space_allocation_top. | |
| 4177 __ sw(a3, MemOperand(t0)); | |
| 4178 // Push the argument. | |
| 4179 __ sw(a2, MemOperand(end_elements)); | |
| 4180 // Fill the rest with holes. | |
| 4181 __ LoadRoot(a3, Heap::kTheHoleValueRootIndex); | |
| 4182 for (int i = 1; i < kAllocationDelta; i++) { | |
| 4183 __ sw(a3, MemOperand(end_elements, i * kPointerSize)); | |
| 4184 } | |
| 4185 | |
| 4186 // Update elements' and array's sizes. | |
| 4187 __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset)); | |
| 4188 __ lw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset)); | |
| 4189 __ Addu(t0, t0, Operand(Smi::FromInt(kAllocationDelta))); | |
| 4190 __ sw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset)); | |
| 4191 | |
| 4192 // Elements are in new space, so write barrier is not required. | |
| 4193 __ mov(v0, scratch); | |
| 4194 __ DropAndRet(argc + 1); | |
| 4195 | |
| 4196 __ bind(&call_builtin); | |
| 4197 __ TailCallExternalReference( | |
| 4198 ExternalReference(Builtins::c_ArrayPush, isolate), argc + 1, 1); | |
| 4199 } | |
| 4200 | |
| 4201 | |
| 4202 void BinaryOpICWithAllocationSiteStub::Generate(MacroAssembler* masm) { | 4002 void BinaryOpICWithAllocationSiteStub::Generate(MacroAssembler* masm) { |
| 4203 // ----------- S t a t e ------------- | 4003 // ----------- S t a t e ------------- |
| 4204 // -- a1 : left | 4004 // -- a1 : left |
| 4205 // -- a0 : right | 4005 // -- a0 : right |
| 4206 // -- ra : return address | 4006 // -- ra : return address |
| 4207 // ----------------------------------- | 4007 // ----------------------------------- |
| 4208 Isolate* isolate = masm->isolate(); | 4008 Isolate* isolate = masm->isolate(); |
| 4209 | 4009 |
| 4210 // Load a2 with the allocation site. We stick an undefined dummy value here | 4010 // Load a2 with the allocation site. We stick an undefined dummy value here |
| 4211 // and replace it with the real allocation site later when we instantiate this | 4011 // and replace it with the real allocation site later when we instantiate this |
| (...skipping 1432 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5644 MemOperand(fp, 6 * kPointerSize), | 5444 MemOperand(fp, 6 * kPointerSize), |
| 5645 NULL); | 5445 NULL); |
| 5646 } | 5446 } |
| 5647 | 5447 |
| 5648 | 5448 |
| 5649 #undef __ | 5449 #undef __ |
| 5650 | 5450 |
| 5651 } } // namespace v8::internal | 5451 } } // namespace v8::internal |
| 5652 | 5452 |
| 5653 #endif // V8_TARGET_ARCH_MIPS | 5453 #endif // V8_TARGET_ARCH_MIPS |
| OLD | NEW |