Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(61)

Unified Diff: src/builtins/arm64/builtins-arm64.cc

Issue 2649143002: [Turbofan] Implement call with spread bytecode in assembly code. (Closed)
Patch Set: Rename PushArgsMode to InterpreterPushArgsMode Created 3 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/builtins/arm/builtins-arm.cc ('k') | src/builtins/builtins.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/builtins/arm64/builtins-arm64.cc
diff --git a/src/builtins/arm64/builtins-arm64.cc b/src/builtins/arm64/builtins-arm64.cc
index 2cd3aae2d1fdf8e48c819e2d285c66fcba49894b..cc4ba869a357b1e9b1d4eb8fd08852a66cc0a887 100644
--- a/src/builtins/arm64/builtins-arm64.cc
+++ b/src/builtins/arm64/builtins-arm64.cc
@@ -1162,7 +1162,7 @@ static void Generate_InterpreterPushArgs(MacroAssembler* masm,
// static
void Builtins::Generate_InterpreterPushArgsAndCallImpl(
MacroAssembler* masm, TailCallMode tail_call_mode,
- CallableType function_type) {
+ InterpreterPushArgsMode mode) {
// ----------- S t a t e -------------
// -- x0 : the number of arguments (not including the receiver)
// -- x2 : the address of the first argument to be pushed. Subsequent
@@ -1179,12 +1179,14 @@ void Builtins::Generate_InterpreterPushArgsAndCallImpl(
Generate_InterpreterPushArgs(masm, x3, x2, x4, x5, x6, &stack_overflow);
// Call the target.
- if (function_type == CallableType::kJSFunction) {
+ if (mode == InterpreterPushArgsMode::kJSFunction) {
__ Jump(masm->isolate()->builtins()->CallFunction(ConvertReceiverMode::kAny,
tail_call_mode),
RelocInfo::CODE_TARGET);
+ } else if (mode == InterpreterPushArgsMode::kWithFinalSpread) {
+ __ Jump(masm->isolate()->builtins()->CallWithSpread(),
+ RelocInfo::CODE_TARGET);
} else {
- DCHECK_EQ(function_type, CallableType::kAny);
__ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny,
tail_call_mode),
RelocInfo::CODE_TARGET);
@@ -1199,7 +1201,7 @@ void Builtins::Generate_InterpreterPushArgsAndCallImpl(
// static
void Builtins::Generate_InterpreterPushArgsAndConstructImpl(
- MacroAssembler* masm, PushArgsConstructMode mode) {
+ MacroAssembler* masm, InterpreterPushArgsMode mode) {
// ----------- S t a t e -------------
// -- x0 : argument count (not including receiver)
// -- x3 : new target
@@ -1216,7 +1218,7 @@ void Builtins::Generate_InterpreterPushArgsAndConstructImpl(
Generate_InterpreterPushArgs(masm, x0, x4, x5, x6, x7, &stack_overflow);
__ AssertUndefinedOrAllocationSite(x2, x6);
- if (mode == PushArgsConstructMode::kJSFunction) {
+ if (mode == InterpreterPushArgsMode::kJSFunction) {
__ AssertFunction(x1);
// Tail call to the function-specific construct stub (still in the caller
@@ -1225,12 +1227,12 @@ void Builtins::Generate_InterpreterPushArgsAndConstructImpl(
__ Ldr(x4, FieldMemOperand(x4, SharedFunctionInfo::kConstructStubOffset));
__ Add(x4, x4, Code::kHeaderSize - kHeapObjectTag);
__ Br(x4);
- } else if (mode == PushArgsConstructMode::kWithFinalSpread) {
+ } else if (mode == InterpreterPushArgsMode::kWithFinalSpread) {
// Call the constructor with x0, x1, and x3 unmodified.
__ Jump(masm->isolate()->builtins()->ConstructWithSpread(),
RelocInfo::CODE_TARGET);
} else {
- DCHECK_EQ(PushArgsConstructMode::kOther, mode);
+ DCHECK_EQ(InterpreterPushArgsMode::kOther, mode);
// Call the constructor with x0, x1, and x3 unmodified.
__ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET);
}
@@ -2696,6 +2698,151 @@ void Builtins::Generate_Call(MacroAssembler* masm, ConvertReceiverMode mode,
}
}
+static void CheckSpreadAndPushToStack(MacroAssembler* masm) {
+ Register argc = x0;
+ Register constructor = x1;
+ Register new_target = x3;
+
+ Register scratch = x2;
+ Register scratch2 = x6;
+
+ Register spread = x4;
+ Register spread_map = x5;
+
+ Register spread_len = x5;
+
+ __ Peek(spread, 0);
+ __ Ldr(spread_map, FieldMemOperand(spread, HeapObject::kMapOffset));
+
+ Label runtime_call, push_args;
+ // Check that the spread is an array.
+ __ CompareInstanceType(spread_map, scratch, JS_ARRAY_TYPE);
+ __ B(ne, &runtime_call);
+
+ // Check that we have the original ArrayPrototype.
+ __ Ldr(scratch, FieldMemOperand(spread_map, Map::kPrototypeOffset));
+ __ Ldr(scratch2, NativeContextMemOperand());
+ __ Ldr(scratch2,
+ ContextMemOperand(scratch2, Context::INITIAL_ARRAY_PROTOTYPE_INDEX));
+ __ Cmp(scratch, scratch2);
+ __ B(ne, &runtime_call);
+
+ // Check that the ArrayPrototype hasn't been modified in a way that would
+ // affect iteration.
+ __ LoadRoot(scratch, Heap::kArrayIteratorProtectorRootIndex);
+ __ Ldr(scratch, FieldMemOperand(scratch, Cell::kValueOffset));
+ __ Cmp(scratch, Smi::FromInt(Isolate::kProtectorValid));
+ __ B(ne, &runtime_call);
+
+ // Check that the map of the initial array iterator hasn't changed.
+ __ Ldr(scratch2, NativeContextMemOperand());
+ __ Ldr(scratch,
+ ContextMemOperand(scratch2,
+ Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_INDEX));
+ __ Ldr(scratch, FieldMemOperand(scratch, HeapObject::kMapOffset));
+ __ Ldr(scratch2,
+ ContextMemOperand(
+ scratch2, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_MAP_INDEX));
+ __ Cmp(scratch, scratch2);
+ __ B(ne, &runtime_call);
+
+ // For FastPacked kinds, iteration will have the same effect as simply
+ // accessing each property in order.
+ Label no_protector_check;
+ __ Ldr(scratch, FieldMemOperand(spread_map, Map::kBitField2Offset));
+ __ DecodeField<Map::ElementsKindBits>(scratch);
+ __ Cmp(scratch, FAST_HOLEY_ELEMENTS);
+ __ B(hi, &runtime_call);
+ // For non-FastHoley kinds, we can skip the protector check.
+ __ Cmp(scratch, FAST_SMI_ELEMENTS);
+ __ B(eq, &no_protector_check);
+ __ Cmp(scratch, FAST_ELEMENTS);
+ __ B(eq, &no_protector_check);
+ // Check the ArrayProtector cell.
+ __ LoadRoot(scratch, Heap::kArrayProtectorRootIndex);
+ __ Ldr(scratch, FieldMemOperand(scratch, PropertyCell::kValueOffset));
+ __ Cmp(scratch, Smi::FromInt(Isolate::kProtectorValid));
+ __ B(ne, &runtime_call);
+
+ __ Bind(&no_protector_check);
+ // Load the FixedArray backing store, but use the length from the array.
+ __ Ldrsw(spread_len, UntagSmiFieldMemOperand(spread, JSArray::kLengthOffset));
+ __ Ldr(spread, FieldMemOperand(spread, JSArray::kElementsOffset));
+ __ B(&push_args);
+
+ __ Bind(&runtime_call);
+ {
+ // Call the builtin for the result of the spread.
+ FrameScope scope(masm, StackFrame::INTERNAL);
+ __ SmiTag(argc);
+ __ Push(constructor, new_target, argc, spread);
+ __ CallRuntime(Runtime::kSpreadIterableFixed);
+ __ Mov(spread, x0);
+ __ Pop(argc, new_target, constructor);
+ __ SmiUntag(argc);
+ }
+
+ {
+ // Calculate the new nargs including the result of the spread.
+ __ Ldrsw(spread_len,
+ UntagSmiFieldMemOperand(spread, FixedArray::kLengthOffset));
+
+ __ Bind(&push_args);
+ // argc += spread_len - 1. Subtract 1 for the spread itself.
+ __ Add(argc, argc, spread_len);
+ __ Sub(argc, argc, 1);
+
+ // Pop the spread argument off the stack.
+ __ Pop(scratch);
+ }
+
+ // Check for stack overflow.
+ {
+ // Check the stack for overflow. We are not trying to catch interruptions
+ // (i.e. debug break and preemption) here, so check the "real stack limit".
+ Label done;
+ __ LoadRoot(scratch, Heap::kRealStackLimitRootIndex);
+ // Make scratch the space we have left. The stack might already be
+ // overflowed here which will cause scratch to become negative.
+ __ Sub(scratch, masm->StackPointer(), scratch);
+ // Check if the arguments will overflow the stack.
+ __ Cmp(scratch, Operand(spread_len, LSL, kPointerSizeLog2));
+ __ B(gt, &done); // Signed comparison.
+ __ TailCallRuntime(Runtime::kThrowStackOverflow);
+ __ Bind(&done);
+ }
+
+ // Put the evaluated spread onto the stack as additional arguments.
+ {
+ __ Mov(scratch, 0);
+ Label done, loop;
+ __ Bind(&loop);
+ __ Cmp(scratch, spread_len);
+ __ B(eq, &done);
+ __ Add(scratch2, spread, Operand(scratch, LSL, kPointerSizeLog2));
+ __ Ldr(scratch2, FieldMemOperand(scratch2, FixedArray::kHeaderSize));
+ __ Push(scratch2);
+ __ Add(scratch, scratch, Operand(1));
+ __ B(&loop);
+ __ Bind(&done);
+ }
+}
+
+// static
+void Builtins::Generate_CallWithSpread(MacroAssembler* masm) {
+ // ----------- S t a t e -------------
+ // -- x0 : the number of arguments (not including the receiver)
+ // -- x1 : the constructor to call (can be any Object)
+ // -----------------------------------
+
+ // CheckSpreadAndPushToStack will push r3 to save it.
+ __ LoadRoot(x3, Heap::kUndefinedValueRootIndex);
+ CheckSpreadAndPushToStack(masm);
+ __ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny,
+ TailCallMode::kDisallow),
+ RelocInfo::CODE_TARGET);
+}
+
// static
void Builtins::Generate_ConstructFunction(MacroAssembler* masm) {
// ----------- S t a t e -------------
@@ -2825,130 +2972,7 @@ void Builtins::Generate_ConstructWithSpread(MacroAssembler* masm) {
// the JSFunction on which new was invoked initially)
// -----------------------------------
- Register argc = x0;
- Register constructor = x1;
- Register new_target = x3;
-
- Register scratch = x2;
- Register scratch2 = x6;
-
- Register spread = x4;
- Register spread_map = x5;
- __ Peek(spread, 0);
- __ Ldr(spread_map, FieldMemOperand(spread, HeapObject::kMapOffset));
-
- Label runtime_call, push_args;
- // Check that the spread is an array.
- __ CompareInstanceType(spread_map, scratch, JS_ARRAY_TYPE);
- __ B(ne, &runtime_call);
-
- // Check that we have the original ArrayPrototype.
- __ Ldr(scratch, FieldMemOperand(spread_map, Map::kPrototypeOffset));
- __ Ldr(scratch2, NativeContextMemOperand());
- __ Ldr(scratch2,
- ContextMemOperand(scratch2, Context::INITIAL_ARRAY_PROTOTYPE_INDEX));
- __ Cmp(scratch, scratch2);
- __ B(ne, &runtime_call);
-
- // Check that the ArrayPrototype hasn't been modified in a way that would
- // affect iteration.
- __ LoadRoot(scratch, Heap::kArrayIteratorProtectorRootIndex);
- __ Ldr(scratch, FieldMemOperand(scratch, Cell::kValueOffset));
- __ Cmp(scratch, Smi::FromInt(Isolate::kProtectorValid));
- __ B(ne, &runtime_call);
-
- // Check that the map of the initial array iterator hasn't changed.
- __ Ldr(scratch2, NativeContextMemOperand());
- __ Ldr(scratch,
- ContextMemOperand(scratch2,
- Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_INDEX));
- __ Ldr(scratch, FieldMemOperand(scratch, HeapObject::kMapOffset));
- __ Ldr(scratch2,
- ContextMemOperand(
- scratch2, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_MAP_INDEX));
- __ Cmp(scratch, scratch2);
- __ B(ne, &runtime_call);
-
- // For FastPacked kinds, iteration will have the same effect as simply
- // accessing each property in order.
- Label no_protector_check;
- __ Ldr(scratch, FieldMemOperand(spread_map, Map::kBitField2Offset));
- __ DecodeField<Map::ElementsKindBits>(scratch);
- __ Cmp(scratch, FAST_HOLEY_ELEMENTS);
- __ B(hi, &runtime_call);
- // For non-FastHoley kinds, we can skip the protector check.
- __ Cmp(scratch, FAST_SMI_ELEMENTS);
- __ B(eq, &no_protector_check);
- __ Cmp(scratch, FAST_ELEMENTS);
- __ B(eq, &no_protector_check);
- // Check the ArrayProtector cell.
- __ LoadRoot(scratch, Heap::kArrayProtectorRootIndex);
- __ Ldr(scratch, FieldMemOperand(scratch, PropertyCell::kValueOffset));
- __ Cmp(scratch, Smi::FromInt(Isolate::kProtectorValid));
- __ B(ne, &runtime_call);
-
- __ Bind(&no_protector_check);
- // Load the FixedArray backing store.
- __ Ldr(spread, FieldMemOperand(spread, JSArray::kElementsOffset));
- __ B(&push_args);
-
- __ Bind(&runtime_call);
- {
- // Call the builtin for the result of the spread.
- FrameScope scope(masm, StackFrame::INTERNAL);
- __ SmiTag(argc);
- __ Push(constructor, new_target, argc, spread);
- __ CallRuntime(Runtime::kSpreadIterableFixed);
- __ Mov(spread, x0);
- __ Pop(argc, new_target, constructor);
- __ SmiUntag(argc);
- }
-
- Register spread_len = x5;
- __ Bind(&push_args);
- {
- // Pop the spread argument off the stack.
- __ Pop(scratch);
- // Calculate the new nargs including the result of the spread.
- __ Ldrsw(spread_len,
- UntagSmiFieldMemOperand(spread, FixedArray::kLengthOffset));
- // argc += spread_len - 1. Subtract 1 for the spread itself.
- __ Add(argc, argc, spread_len);
- __ Sub(argc, argc, 1);
- }
-
- // Check for stack overflow.
- {
- // Check the stack for overflow. We are not trying to catch interruptions
- // (i.e. debug break and preemption) here, so check the "real stack limit".
- Label done;
- __ LoadRoot(scratch, Heap::kRealStackLimitRootIndex);
- // Make scratch the space we have left. The stack might already be
- // overflowed here which will cause scratch to become negative.
- __ Sub(scratch, masm->StackPointer(), scratch);
- // Check if the arguments will overflow the stack.
- __ Cmp(scratch, Operand(spread_len, LSL, kPointerSizeLog2));
- __ B(gt, &done); // Signed comparison.
- __ TailCallRuntime(Runtime::kThrowStackOverflow);
- __ Bind(&done);
- }
-
- // Put the evaluated spread onto the stack as additional arguments.
- {
- __ Mov(scratch, 0);
- Label done, loop;
- __ Bind(&loop);
- __ Cmp(scratch, spread_len);
- __ B(eq, &done);
- __ Add(scratch2, spread, Operand(scratch, LSL, kPointerSizeLog2));
- __ Ldr(scratch2, FieldMemOperand(scratch2, FixedArray::kHeaderSize));
- __ Push(scratch2);
- __ Add(scratch, scratch, Operand(1));
- __ B(&loop);
- __ Bind(&done);
- }
-
- // Dispatch.
+ CheckSpreadAndPushToStack(masm);
__ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET);
}
« no previous file with comments | « src/builtins/arm/builtins-arm.cc ('k') | src/builtins/builtins.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698