Index: src/mips64/code-stubs-mips64.cc |
diff --git a/src/mips64/code-stubs-mips64.cc b/src/mips64/code-stubs-mips64.cc |
index 50d1b03b31f8d4dd2383331b0fda1a1e424d5737..52a124cd8e1297d0cdae09d8a23ffc1a911e117f 100644 |
--- a/src/mips64/code-stubs-mips64.cc |
+++ b/src/mips64/code-stubs-mips64.cc |
@@ -1729,7 +1729,7 @@ void ArgumentsAccessStub::GenerateNewSloppyFast(MacroAssembler* masm) { |
__ Daddu(t1, t1, Operand(FixedArray::kHeaderSize)); |
// 3. Arguments object. |
- __ Daddu(t1, t1, Operand(Heap::kSloppyArgumentsObjectSize)); |
+ __ Daddu(t1, t1, Operand(JSSloppyArgumentsObject::kSize)); |
// Do the allocation of all three objects in one go. |
__ Allocate(t1, v0, t1, a4, &runtime, TAG_OBJECT); |
@@ -1762,23 +1762,17 @@ void ArgumentsAccessStub::GenerateNewSloppyFast(MacroAssembler* masm) { |
__ sd(t1, FieldMemOperand(v0, JSObject::kElementsOffset)); |
// Set up the callee in-object property. |
- STATIC_ASSERT(Heap::kArgumentsCalleeIndex == 1); |
__ AssertNotSmi(a1); |
- const int kCalleeOffset = JSObject::kHeaderSize + |
- Heap::kArgumentsCalleeIndex * kPointerSize; |
- __ sd(a1, FieldMemOperand(v0, kCalleeOffset)); |
+ __ sd(a1, FieldMemOperand(v0, JSSloppyArgumentsObject::kCalleeOffset)); |
// Use the length (smi tagged) and set that as an in-object property too. |
__ AssertSmi(a5); |
- STATIC_ASSERT(Heap::kArgumentsLengthIndex == 0); |
- const int kLengthOffset = JSObject::kHeaderSize + |
- Heap::kArgumentsLengthIndex * kPointerSize; |
- __ sd(a5, FieldMemOperand(v0, kLengthOffset)); |
+ __ sd(a5, FieldMemOperand(v0, JSSloppyArgumentsObject::kLengthOffset)); |
// Set up the elements pointer in the allocated arguments object. |
// If we allocated a parameter map, a4 will point there, otherwise |
// it will point to the backing store. |
- __ Daddu(a4, v0, Operand(Heap::kSloppyArgumentsObjectSize)); |
+ __ Daddu(a4, v0, Operand(JSSloppyArgumentsObject::kSize)); |
__ sd(a4, FieldMemOperand(v0, JSObject::kElementsOffset)); |
// v0 = address of new object (tagged) |
@@ -1844,7 +1838,7 @@ void ArgumentsAccessStub::GenerateNewSloppyFast(MacroAssembler* masm) { |
__ Branch(¶meters_loop, ne, a5, Operand(Smi::FromInt(0))); |
// Restore t1 = argument count (tagged). |
- __ ld(a5, FieldMemOperand(v0, kLengthOffset)); |
+ __ ld(a5, FieldMemOperand(v0, JSSloppyArgumentsObject::kLengthOffset)); |
__ bind(&skip_parameter_map); |
// v0 = address of new object (tagged) |
@@ -1907,96 +1901,6 @@ void LoadIndexedInterceptorStub::Generate(MacroAssembler* masm) { |
} |
-void ArgumentsAccessStub::GenerateNewStrict(MacroAssembler* masm) { |
- // a1 : function |
- // a2 : number of parameters (tagged) |
- // a3 : parameters pointer |
- |
- DCHECK(a1.is(ArgumentsAccessNewDescriptor::function())); |
- DCHECK(a2.is(ArgumentsAccessNewDescriptor::parameter_count())); |
- DCHECK(a3.is(ArgumentsAccessNewDescriptor::parameter_pointer())); |
- |
- // Check if the calling frame is an arguments adaptor frame. |
- Label try_allocate, runtime; |
- __ ld(a4, MemOperand(fp, StandardFrameConstants::kCallerFPOffset)); |
- __ ld(a0, MemOperand(a4, StandardFrameConstants::kContextOffset)); |
- __ Branch(&try_allocate, ne, a0, |
- Operand(Smi::FromInt(StackFrame::ARGUMENTS_ADAPTOR))); |
- |
- // Patch the arguments.length and the parameters pointer. |
- __ ld(a2, MemOperand(a4, ArgumentsAdaptorFrameConstants::kLengthOffset)); |
- __ SmiScale(at, a2, kPointerSizeLog2); |
- __ Daddu(a4, a4, Operand(at)); |
- __ Daddu(a3, a4, Operand(StandardFrameConstants::kCallerSPOffset)); |
- |
- // Try the new space allocation. Start out with computing the size |
- // of the arguments object and the elements array in words. |
- Label add_arguments_object; |
- __ bind(&try_allocate); |
- __ SmiUntag(t1, a2); |
- __ Branch(&add_arguments_object, eq, a2, Operand(zero_reg)); |
- |
- __ Daddu(t1, t1, Operand(FixedArray::kHeaderSize / kPointerSize)); |
- __ bind(&add_arguments_object); |
- __ Daddu(t1, t1, Operand(Heap::kStrictArgumentsObjectSize / kPointerSize)); |
- |
- // Do the allocation of both objects in one go. |
- __ Allocate(t1, v0, a4, a5, &runtime, |
- static_cast<AllocationFlags>(TAG_OBJECT | SIZE_IN_WORDS)); |
- |
- // Get the arguments boilerplate from the current native context. |
- __ LoadNativeContextSlot(Context::STRICT_ARGUMENTS_MAP_INDEX, a4); |
- |
- __ sd(a4, FieldMemOperand(v0, JSObject::kMapOffset)); |
- __ LoadRoot(a5, Heap::kEmptyFixedArrayRootIndex); |
- __ sd(a5, FieldMemOperand(v0, JSObject::kPropertiesOffset)); |
- __ sd(a5, FieldMemOperand(v0, JSObject::kElementsOffset)); |
- |
- // Get the length (smi tagged) and set that as an in-object property too. |
- STATIC_ASSERT(Heap::kArgumentsLengthIndex == 0); |
- __ AssertSmi(a2); |
- __ sd(a2, |
- FieldMemOperand(v0, JSObject::kHeaderSize + |
- Heap::kArgumentsLengthIndex * kPointerSize)); |
- |
- Label done; |
- __ Branch(&done, eq, a2, Operand(zero_reg)); |
- |
- // Set up the elements pointer in the allocated arguments object and |
- // initialize the header in the elements fixed array. |
- __ Daddu(a4, v0, Operand(Heap::kStrictArgumentsObjectSize)); |
- __ sd(a4, FieldMemOperand(v0, JSObject::kElementsOffset)); |
- __ LoadRoot(a5, Heap::kFixedArrayMapRootIndex); |
- __ sd(a5, FieldMemOperand(a4, FixedArray::kMapOffset)); |
- __ sd(a2, FieldMemOperand(a4, FixedArray::kLengthOffset)); |
- __ SmiUntag(a2); |
- |
- // Copy the fixed array slots. |
- Label loop; |
- // Set up a4 to point to the first array slot. |
- __ Daddu(a4, a4, Operand(FixedArray::kHeaderSize - kHeapObjectTag)); |
- __ bind(&loop); |
- // Pre-decrement a3 with kPointerSize on each iteration. |
- // Pre-decrement in order to skip receiver. |
- __ Daddu(a3, a3, Operand(-kPointerSize)); |
- __ ld(a5, MemOperand(a3)); |
- // Post-increment a4 with kPointerSize on each iteration. |
- __ sd(a5, MemOperand(a4)); |
- __ Daddu(a4, a4, Operand(kPointerSize)); |
- __ Dsubu(a2, a2, Operand(1)); |
- __ Branch(&loop, ne, a2, Operand(zero_reg)); |
- |
- // Return. |
- __ bind(&done); |
- __ Ret(); |
- |
- // Do the runtime call to allocate the arguments object. |
- __ bind(&runtime); |
- __ Push(a1, a3, a2); |
- __ TailCallRuntime(Runtime::kNewStrictArguments); |
-} |
- |
- |
void RegExpExecStub::Generate(MacroAssembler* masm) { |
// Just jump directly to runtime if native RegExp is not selected at compile |
// time or if regexp entry in generated code is turned off runtime switch or |
@@ -5309,6 +5213,116 @@ void FastNewRestParameterStub::Generate(MacroAssembler* masm) { |
} |
+void FastNewStrictArgumentsStub::Generate(MacroAssembler* masm) { |
+ // ----------- S t a t e ------------- |
+ // -- a1 : function |
+ // -- cp : context |
+ // -- fp : frame pointer |
+ // -- ra : return address |
+ // ----------------------------------- |
+ __ AssertFunction(a1); |
+ |
+ // For Ignition we need to skip all possible handler/stub frames until |
+ // we reach the JavaScript frame for the function (similar to what the |
+ // runtime fallback implementation does). So make a2 point to that |
+ // JavaScript frame. |
+ { |
+ Label loop, loop_entry; |
+ __ Branch(USE_DELAY_SLOT, &loop_entry); |
+ __ mov(a2, fp); // In delay slot. |
+ __ bind(&loop); |
+ __ ld(a2, MemOperand(a2, StandardFrameConstants::kCallerFPOffset)); |
+ __ bind(&loop_entry); |
+ __ ld(a3, MemOperand(a2, StandardFrameConstants::kMarkerOffset)); |
+ __ Branch(&loop, ne, a1, Operand(a3)); |
+ } |
+ |
+ // Check if we have an arguments adaptor frame below the function frame. |
+ Label arguments_adaptor, arguments_done; |
+ __ ld(a3, MemOperand(a2, StandardFrameConstants::kCallerFPOffset)); |
+ __ ld(a0, MemOperand(a3, StandardFrameConstants::kContextOffset)); |
+ __ Branch(&arguments_adaptor, eq, a0, |
+ Operand(Smi::FromInt(StackFrame::ARGUMENTS_ADAPTOR))); |
+ { |
+ __ ld(a1, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); |
+ __ ld(a0, |
+ FieldMemOperand(a1, SharedFunctionInfo::kFormalParameterCountOffset)); |
+ __ Dlsa(a2, a2, a0, kPointerSizeLog2); |
+ __ Daddu(a2, a2, Operand(StandardFrameConstants::kCallerSPOffset - |
+ 1 * kPointerSize)); |
+ } |
+ __ Branch(&arguments_done); |
+ __ bind(&arguments_adaptor); |
+ { |
+ __ SmiLoadUntag( |
+ a0, MemOperand(a2, ArgumentsAdaptorFrameConstants::kLengthOffset)); |
+ __ Dlsa(a2, a3, a0, kPointerSizeLog2); |
+ __ Daddu(a2, a2, Operand(StandardFrameConstants::kCallerSPOffset - |
+ 1 * kPointerSize)); |
+ } |
+ __ bind(&arguments_done); |
+ |
+ // ----------- S t a t e ------------- |
+ // -- cp : context |
+ // -- a0 : number of rest parameters |
+ // -- a2 : pointer to first rest parameters |
+ // -- ra : return address |
+ // ----------------------------------- |
+ |
+ // Allocate space for the rest parameter array plus the backing store. |
+ Label allocate, done_allocate; |
+ __ li(a1, Operand(JSStrictArgumentsObject::kSize + FixedArray::kHeaderSize)); |
+ __ Dlsa(a1, a1, a0, kPointerSizeLog2); |
+ __ Allocate(a1, v0, a3, a4, &allocate, TAG_OBJECT); |
+ __ bind(&done_allocate); |
+ |
+ // Compute arguments.length in a4. |
+ __ SmiTag(a4, a0); |
+ |
+ // Setup the elements array in v0. |
+ __ LoadRoot(at, Heap::kFixedArrayMapRootIndex); |
+ __ sd(at, FieldMemOperand(v0, FixedArray::kMapOffset)); |
+ __ sd(a4, FieldMemOperand(v0, FixedArray::kLengthOffset)); |
+ __ Daddu(a3, v0, Operand(FixedArray::kHeaderSize)); |
+ { |
+ Label loop, done_loop; |
+ __ Dlsa(a1, a3, a0, kPointerSizeLog2); |
+ __ bind(&loop); |
+ __ Branch(&done_loop, eq, a1, Operand(a3)); |
+ __ ld(at, MemOperand(a2, 0 * kPointerSize)); |
+ __ sd(at, FieldMemOperand(a3, 0 * kPointerSize)); |
+ __ Dsubu(a2, a2, Operand(1 * kPointerSize)); |
+ __ Daddu(a3, a3, Operand(1 * kPointerSize)); |
+ __ b(&loop); |
+ __ bind(&done_loop); |
+ } |
+ |
+ // Setup the rest parameter array in a3. |
+ __ LoadNativeContextSlot(Context::JS_ARRAY_FAST_ELEMENTS_MAP_INDEX, at); |
+ __ sd(at, FieldMemOperand(a3, JSStrictArgumentsObject::kMapOffset)); |
+ __ LoadRoot(at, Heap::kEmptyFixedArrayRootIndex); |
+ __ sd(at, FieldMemOperand(a3, JSStrictArgumentsObject::kPropertiesOffset)); |
+ __ sd(v0, FieldMemOperand(a3, JSStrictArgumentsObject::kElementsOffset)); |
+ __ sd(a4, FieldMemOperand(a3, JSStrictArgumentsObject::kLengthOffset)); |
+ STATIC_ASSERT(JSStrictArgumentsObject::kSize == 4 * kPointerSize); |
+ __ Ret(USE_DELAY_SLOT); |
+ __ mov(v0, a3); // In delay slot |
+ |
+ // Fall back to %AllocateInNewSpace. |
+ __ bind(&allocate); |
+ { |
+ FrameScope scope(masm, StackFrame::INTERNAL); |
+ __ SmiTag(a0); |
+ __ SmiTag(a1); |
+ __ Push(a0, a2, a1); |
+ __ CallRuntime(Runtime::kAllocateInNewSpace); |
+ __ Pop(a0, a2); |
+ __ SmiUntag(a0); |
+ } |
+ __ jmp(&done_allocate); |
+} |
+ |
+ |
void LoadGlobalViaContextStub::Generate(MacroAssembler* masm) { |
Register context_reg = cp; |
Register slot_reg = a2; |