Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(71)

Unified Diff: src/builtins/x64/builtins-x64.cc

Issue 2649143002: [Turbofan] Implement call with spread bytecode in assembly code. (Closed)
Patch Set: Rename PushArgsMode to InterpreterPushArgsMode Created 3 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/builtins/s390/builtins-s390.cc ('k') | src/builtins/x87/builtins-x87.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/builtins/x64/builtins-x64.cc
diff --git a/src/builtins/x64/builtins-x64.cc b/src/builtins/x64/builtins-x64.cc
index b232077e6584f71ce99ee47d18d00b74b8bed61e..58d38cf138a6236728f3a6fc9e42f971d98537c3 100644
--- a/src/builtins/x64/builtins-x64.cc
+++ b/src/builtins/x64/builtins-x64.cc
@@ -765,7 +765,7 @@ static void Generate_InterpreterPushArgs(MacroAssembler* masm,
// static
void Builtins::Generate_InterpreterPushArgsAndCallImpl(
MacroAssembler* masm, TailCallMode tail_call_mode,
- CallableType function_type) {
+ InterpreterPushArgsMode mode) {
// ----------- S t a t e -------------
// -- rax : the number of arguments (not including the receiver)
// -- rbx : the address of the first argument to be pushed. Subsequent
@@ -791,12 +791,14 @@ void Builtins::Generate_InterpreterPushArgsAndCallImpl(
// Call the target.
__ PushReturnAddressFrom(kScratchRegister); // Re-push return address.
- if (function_type == CallableType::kJSFunction) {
+ if (mode == InterpreterPushArgsMode::kJSFunction) {
__ Jump(masm->isolate()->builtins()->CallFunction(ConvertReceiverMode::kAny,
tail_call_mode),
RelocInfo::CODE_TARGET);
+ } else if (mode == InterpreterPushArgsMode::kWithFinalSpread) {
+ __ Jump(masm->isolate()->builtins()->CallWithSpread(),
+ RelocInfo::CODE_TARGET);
} else {
- DCHECK_EQ(function_type, CallableType::kAny);
__ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny,
tail_call_mode),
RelocInfo::CODE_TARGET);
@@ -813,7 +815,7 @@ void Builtins::Generate_InterpreterPushArgsAndCallImpl(
// static
void Builtins::Generate_InterpreterPushArgsAndConstructImpl(
- MacroAssembler* masm, PushArgsConstructMode mode) {
+ MacroAssembler* masm, InterpreterPushArgsMode mode) {
// ----------- S t a t e -------------
// -- rax : the number of arguments (not including the receiver)
// -- rdx : the new target (either the same as the constructor or
@@ -842,7 +844,7 @@ void Builtins::Generate_InterpreterPushArgsAndConstructImpl(
__ PushReturnAddressFrom(kScratchRegister);
__ AssertUndefinedOrAllocationSite(rbx);
- if (mode == PushArgsConstructMode::kJSFunction) {
+ if (mode == InterpreterPushArgsMode::kJSFunction) {
// Tail call to the function-specific construct stub (still in the caller
// context at this point).
__ AssertFunction(rdi);
@@ -852,12 +854,12 @@ void Builtins::Generate_InterpreterPushArgsAndConstructImpl(
__ leap(rcx, FieldOperand(rcx, Code::kHeaderSize));
// Jump to the constructor function (rax, rbx, rdx passed on).
__ jmp(rcx);
- } else if (mode == PushArgsConstructMode::kWithFinalSpread) {
+ } else if (mode == InterpreterPushArgsMode::kWithFinalSpread) {
// Call the constructor (rax, rdx, rdi passed on).
__ Jump(masm->isolate()->builtins()->ConstructWithSpread(),
RelocInfo::CODE_TARGET);
} else {
- DCHECK_EQ(PushArgsConstructMode::kOther, mode);
+ DCHECK_EQ(InterpreterPushArgsMode::kOther, mode);
// Call the constructor (rax, rdx, rdi passed on).
__ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET);
}
@@ -2761,6 +2763,143 @@ void Builtins::Generate_Call(MacroAssembler* masm, ConvertReceiverMode mode,
}
}
+static void CheckSpreadAndPushToStack(MacroAssembler* masm) {
+ // Load the spread argument into rbx.
+ __ movp(rbx, Operand(rsp, kPointerSize));
+ // Load the map of the spread into r15.
+ __ movp(r15, FieldOperand(rbx, HeapObject::kMapOffset));
+ // Load native context into r14.
+ __ movp(r14, NativeContextOperand());
+
+ Label runtime_call, push_args;
+ // Check that the spread is an array.
+ __ CmpInstanceType(r15, JS_ARRAY_TYPE);
+ __ j(not_equal, &runtime_call);
+
+ // Check that we have the original ArrayPrototype.
+ __ movp(rcx, FieldOperand(r15, Map::kPrototypeOffset));
+ __ cmpp(rcx, ContextOperand(r14, Context::INITIAL_ARRAY_PROTOTYPE_INDEX));
+ __ j(not_equal, &runtime_call);
+
+ // Check that the ArrayPrototype hasn't been modified in a way that would
+ // affect iteration.
+ __ LoadRoot(rcx, Heap::kArrayIteratorProtectorRootIndex);
+ __ Cmp(FieldOperand(rcx, Cell::kValueOffset),
+ Smi::FromInt(Isolate::kProtectorValid));
+ __ j(not_equal, &runtime_call);
+
+ // Check that the map of the initial array iterator hasn't changed.
+ __ movp(rcx,
+ ContextOperand(r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_INDEX));
+ __ movp(rcx, FieldOperand(rcx, HeapObject::kMapOffset));
+ __ cmpp(rcx, ContextOperand(
+ r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_MAP_INDEX));
+ __ j(not_equal, &runtime_call);
+
+ // For FastPacked kinds, iteration will have the same effect as simply
+ // accessing each property in order.
+ Label no_protector_check;
+ __ movzxbp(rcx, FieldOperand(r15, Map::kBitField2Offset));
+ __ DecodeField<Map::ElementsKindBits>(rcx);
+ __ cmpp(rcx, Immediate(FAST_HOLEY_ELEMENTS));
+ __ j(above, &runtime_call);
+ // For non-FastHoley kinds, we can skip the protector check.
+ __ cmpp(rcx, Immediate(FAST_SMI_ELEMENTS));
+ __ j(equal, &no_protector_check);
+ __ cmpp(rcx, Immediate(FAST_ELEMENTS));
+ __ j(equal, &no_protector_check);
+ // Check the ArrayProtector cell.
+ __ LoadRoot(rcx, Heap::kArrayProtectorRootIndex);
+ __ Cmp(FieldOperand(rcx, PropertyCell::kValueOffset),
+ Smi::FromInt(Isolate::kProtectorValid));
+ __ j(not_equal, &runtime_call);
+
+ __ bind(&no_protector_check);
+ // Load the FixedArray backing store, but use the length from the array.
+ __ SmiToInteger32(r9, FieldOperand(rbx, JSArray::kLengthOffset));
+ __ movp(rbx, FieldOperand(rbx, JSArray::kElementsOffset));
+ __ jmp(&push_args);
+
+ __ bind(&runtime_call);
+ {
+ // Call the builtin for the result of the spread.
+ FrameScope scope(masm, StackFrame::INTERNAL);
+ __ Push(rdi); // target
+ __ Push(rdx); // new target
+ __ Integer32ToSmi(rax, rax);
+ __ Push(rax); // nargs
+ __ Push(rbx);
+ __ CallRuntime(Runtime::kSpreadIterableFixed);
+ __ movp(rbx, rax);
+ __ Pop(rax); // nargs
+ __ SmiToInteger32(rax, rax);
+ __ Pop(rdx); // new target
+ __ Pop(rdi); // target
+ }
+
+ {
+ // Calculate the new nargs including the result of the spread.
+ __ SmiToInteger32(r9, FieldOperand(rbx, FixedArray::kLengthOffset));
+
+ __ bind(&push_args);
+ // rax += r9 - 1. Subtract 1 for the spread itself.
+ __ leap(rax, Operand(rax, r9, times_1, -1));
+
+ // Pop the return address and spread argument.
+ __ PopReturnAddressTo(r8);
+ __ Pop(rcx);
+ }
+
+ // Check for stack overflow.
+ {
+ // Check the stack for overflow. We are not trying to catch interruptions
+ // (i.e. debug break and preemption) here, so check the "real stack limit".
+ Label done;
+ __ LoadRoot(kScratchRegister, Heap::kRealStackLimitRootIndex);
+ __ movp(rcx, rsp);
+ // Make rcx the space we have left. The stack might already be overflowed
+ // here which will cause rcx to become negative.
+ __ subp(rcx, kScratchRegister);
+ __ sarp(rcx, Immediate(kPointerSizeLog2));
+ // Check if the arguments will overflow the stack.
+ __ cmpp(rcx, r9);
+ __ j(greater, &done, Label::kNear); // Signed comparison.
+ __ TailCallRuntime(Runtime::kThrowStackOverflow);
+ __ bind(&done);
+ }
+
+ // Put the evaluated spread onto the stack as additional arguments.
+ {
+ __ Set(rcx, 0);
+ Label done, loop;
+ __ bind(&loop);
+ __ cmpl(rcx, r9);
+ __ j(equal, &done, Label::kNear);
+ __ movp(kScratchRegister, FieldOperand(rbx, rcx, times_pointer_size,
+ FixedArray::kHeaderSize));
+ __ Push(kScratchRegister);
+ __ incl(rcx);
+ __ jmp(&loop);
+ __ bind(&done);
+ __ PushReturnAddressFrom(r8);
+ }
+}
+
+// static
+void Builtins::Generate_CallWithSpread(MacroAssembler* masm) {
+ // ----------- S t a t e -------------
+ // -- rax : the number of arguments (not including the receiver)
+ // -- rdi : the target to call (can be any Object)
+ // -----------------------------------
+
+ // CheckSpreadAndPushToStack will push rdx to save it.
+ __ LoadRoot(rdx, Heap::kUndefinedValueRootIndex);
+ CheckSpreadAndPushToStack(masm);
+ __ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny,
+ TailCallMode::kDisallow),
+ RelocInfo::CODE_TARGET);
+}
+
// static
void Builtins::Generate_ConstructFunction(MacroAssembler* masm) {
// ----------- S t a t e -------------
@@ -2892,124 +3031,7 @@ void Builtins::Generate_ConstructWithSpread(MacroAssembler* masm) {
// -- rdi : the constructor to call (can be any Object)
// -----------------------------------
- // Load the spread argument into rbx.
- __ movp(rbx, Operand(rsp, kPointerSize));
- // Load the map of the spread into r15.
- __ movp(r15, FieldOperand(rbx, HeapObject::kMapOffset));
- // Load native context into r14.
- __ movp(r14, NativeContextOperand());
-
- Label runtime_call, push_args;
- // Check that the spread is an array.
- __ CmpInstanceType(r15, JS_ARRAY_TYPE);
- __ j(not_equal, &runtime_call);
-
- // Check that we have the original ArrayPrototype.
- __ movp(rcx, FieldOperand(r15, Map::kPrototypeOffset));
- __ cmpp(rcx, ContextOperand(r14, Context::INITIAL_ARRAY_PROTOTYPE_INDEX));
- __ j(not_equal, &runtime_call);
-
- // Check that the ArrayPrototype hasn't been modified in a way that would
- // affect iteration.
- __ LoadRoot(rcx, Heap::kArrayIteratorProtectorRootIndex);
- __ Cmp(FieldOperand(rcx, Cell::kValueOffset),
- Smi::FromInt(Isolate::kProtectorValid));
- __ j(not_equal, &runtime_call);
-
- // Check that the map of the initial array iterator hasn't changed.
- __ movp(rcx,
- ContextOperand(r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_INDEX));
- __ movp(rcx, FieldOperand(rcx, HeapObject::kMapOffset));
- __ cmpp(rcx, ContextOperand(
- r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_MAP_INDEX));
- __ j(not_equal, &runtime_call);
-
- // For FastPacked kinds, iteration will have the same effect as simply
- // accessing each property in order.
- Label no_protector_check;
- __ movzxbp(rcx, FieldOperand(r15, Map::kBitField2Offset));
- __ DecodeField<Map::ElementsKindBits>(rcx);
- __ cmpp(rcx, Immediate(FAST_HOLEY_ELEMENTS));
- __ j(above, &runtime_call);
- // For non-FastHoley kinds, we can skip the protector check.
- __ cmpp(rcx, Immediate(FAST_SMI_ELEMENTS));
- __ j(equal, &no_protector_check);
- __ cmpp(rcx, Immediate(FAST_ELEMENTS));
- __ j(equal, &no_protector_check);
- // Check the ArrayProtector cell.
- __ LoadRoot(rcx, Heap::kArrayProtectorRootIndex);
- __ Cmp(FieldOperand(rcx, PropertyCell::kValueOffset),
- Smi::FromInt(Isolate::kProtectorValid));
- __ j(not_equal, &runtime_call);
-
- __ bind(&no_protector_check);
- // Load the FixedArray backing store.
- __ movp(rbx, FieldOperand(rbx, JSArray::kElementsOffset));
- __ jmp(&push_args);
-
- __ bind(&runtime_call);
- {
- // Call the builtin for the result of the spread.
- FrameScope scope(masm, StackFrame::INTERNAL);
- __ Push(rdi); // target
- __ Push(rdx); // new target
- __ Integer32ToSmi(rax, rax);
- __ Push(rax); // nargs
- __ Push(rbx);
- __ CallRuntime(Runtime::kSpreadIterableFixed);
- __ movp(rbx, rax);
- __ Pop(rax); // nargs
- __ SmiToInteger32(rax, rax);
- __ Pop(rdx); // new target
- __ Pop(rdi); // target
- }
-
- __ bind(&push_args);
- {
- // Pop the return address and spread argument.
- __ PopReturnAddressTo(r8);
- __ Pop(rcx);
-
- // Calculate the new nargs including the result of the spread.
- __ SmiToInteger32(r9, FieldOperand(rbx, FixedArray::kLengthOffset));
- // rax += r9 - 1. Subtract 1 for the spread itself.
- __ leap(rax, Operand(rax, r9, times_1, -1));
- }
-
- // Check for stack overflow.
- {
- // Check the stack for overflow. We are not trying to catch interruptions
- // (i.e. debug break and preemption) here, so check the "real stack limit".
- Label done;
- __ LoadRoot(kScratchRegister, Heap::kRealStackLimitRootIndex);
- __ movp(rcx, rsp);
- // Make rcx the space we have left. The stack might already be overflowed
- // here which will cause rcx to become negative.
- __ subp(rcx, kScratchRegister);
- __ sarp(rcx, Immediate(kPointerSizeLog2));
- // Check if the arguments will overflow the stack.
- __ cmpp(rcx, r9);
- __ j(greater, &done, Label::kNear); // Signed comparison.
- __ TailCallRuntime(Runtime::kThrowStackOverflow);
- __ bind(&done);
- }
-
- // Put the evaluated spread onto the stack as additional arguments.
- {
- __ Set(rcx, 0);
- Label done, loop;
- __ bind(&loop);
- __ cmpl(rcx, r9);
- __ j(equal, &done, Label::kNear);
- __ movp(kScratchRegister, FieldOperand(rbx, rcx, times_pointer_size,
- FixedArray::kHeaderSize));
- __ Push(kScratchRegister);
- __ incl(rcx);
- __ jmp(&loop);
- __ bind(&done);
- __ PushReturnAddressFrom(r8);
- }
- // Dispatch.
+ CheckSpreadAndPushToStack(masm);
__ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET);
}
« no previous file with comments | « src/builtins/s390/builtins-s390.cc ('k') | src/builtins/x87/builtins-x87.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698