| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #if V8_TARGET_ARCH_X64 | 5 #if V8_TARGET_ARCH_X64 |
| 6 | 6 |
| 7 #include "src/code-factory.h" | 7 #include "src/code-factory.h" |
| 8 #include "src/codegen.h" | 8 #include "src/codegen.h" |
| 9 #include "src/deoptimizer.h" | 9 #include "src/deoptimizer.h" |
| 10 #include "src/full-codegen/full-codegen.h" | 10 #include "src/full-codegen/full-codegen.h" |
| (...skipping 747 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 758 __ Push(Operand(start_address, 0)); | 758 __ Push(Operand(start_address, 0)); |
| 759 __ subp(start_address, Immediate(kPointerSize)); | 759 __ subp(start_address, Immediate(kPointerSize)); |
| 760 __ bind(&loop_check); | 760 __ bind(&loop_check); |
| 761 __ cmpp(start_address, scratch); | 761 __ cmpp(start_address, scratch); |
| 762 __ j(greater, &loop_header, Label::kNear); | 762 __ j(greater, &loop_header, Label::kNear); |
| 763 } | 763 } |
| 764 | 764 |
| 765 // static | 765 // static |
| 766 void Builtins::Generate_InterpreterPushArgsAndCallImpl( | 766 void Builtins::Generate_InterpreterPushArgsAndCallImpl( |
| 767 MacroAssembler* masm, TailCallMode tail_call_mode, | 767 MacroAssembler* masm, TailCallMode tail_call_mode, |
| 768 CallableType function_type) { | 768 InterpreterPushArgsMode mode) { |
| 769 // ----------- S t a t e ------------- | 769 // ----------- S t a t e ------------- |
| 770 // -- rax : the number of arguments (not including the receiver) | 770 // -- rax : the number of arguments (not including the receiver) |
| 771 // -- rbx : the address of the first argument to be pushed. Subsequent | 771 // -- rbx : the address of the first argument to be pushed. Subsequent |
| 772 // arguments should be consecutive above this, in the same order as | 772 // arguments should be consecutive above this, in the same order as |
| 773 // they are to be pushed onto the stack. | 773 // they are to be pushed onto the stack. |
| 774 // -- rdi : the target to call (can be any Object). | 774 // -- rdi : the target to call (can be any Object). |
| 775 // ----------------------------------- | 775 // ----------------------------------- |
| 776 Label stack_overflow; | 776 Label stack_overflow; |
| 777 | 777 |
| 778 // Number of values to be pushed. | 778 // Number of values to be pushed. |
| 779 __ Move(rcx, rax); | 779 __ Move(rcx, rax); |
| 780 __ addp(rcx, Immediate(1)); // Add one for receiver. | 780 __ addp(rcx, Immediate(1)); // Add one for receiver. |
| 781 | 781 |
| 782 // Add a stack check before pushing arguments. | 782 // Add a stack check before pushing arguments. |
| 783 Generate_StackOverflowCheck(masm, rcx, rdx, r8, &stack_overflow); | 783 Generate_StackOverflowCheck(masm, rcx, rdx, r8, &stack_overflow); |
| 784 | 784 |
| 785 // Pop return address to allow tail-call after pushing arguments. | 785 // Pop return address to allow tail-call after pushing arguments. |
| 786 __ PopReturnAddressTo(kScratchRegister); | 786 __ PopReturnAddressTo(kScratchRegister); |
| 787 | 787 |
| 788 // rbx and rdx will be modified. | 788 // rbx and rdx will be modified. |
| 789 Generate_InterpreterPushArgs(masm, rcx, rbx, rdx); | 789 Generate_InterpreterPushArgs(masm, rcx, rbx, rdx); |
| 790 | 790 |
| 791 // Call the target. | 791 // Call the target. |
| 792 __ PushReturnAddressFrom(kScratchRegister); // Re-push return address. | 792 __ PushReturnAddressFrom(kScratchRegister); // Re-push return address. |
| 793 | 793 |
| 794 if (function_type == CallableType::kJSFunction) { | 794 if (mode == InterpreterPushArgsMode::kJSFunction) { |
| 795 __ Jump(masm->isolate()->builtins()->CallFunction(ConvertReceiverMode::kAny, | 795 __ Jump(masm->isolate()->builtins()->CallFunction(ConvertReceiverMode::kAny, |
| 796 tail_call_mode), | 796 tail_call_mode), |
| 797 RelocInfo::CODE_TARGET); | 797 RelocInfo::CODE_TARGET); |
| 798 } else if (mode == InterpreterPushArgsMode::kWithFinalSpread) { |
| 799 __ Jump(masm->isolate()->builtins()->CallWithSpread(), |
| 800 RelocInfo::CODE_TARGET); |
| 798 } else { | 801 } else { |
| 799 DCHECK_EQ(function_type, CallableType::kAny); | |
| 800 __ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny, | 802 __ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny, |
| 801 tail_call_mode), | 803 tail_call_mode), |
| 802 RelocInfo::CODE_TARGET); | 804 RelocInfo::CODE_TARGET); |
| 803 } | 805 } |
| 804 | 806 |
| 805 // Throw stack overflow exception. | 807 // Throw stack overflow exception. |
| 806 __ bind(&stack_overflow); | 808 __ bind(&stack_overflow); |
| 807 { | 809 { |
| 808 __ TailCallRuntime(Runtime::kThrowStackOverflow); | 810 __ TailCallRuntime(Runtime::kThrowStackOverflow); |
| 809 // This should be unreachable. | 811 // This should be unreachable. |
| 810 __ int3(); | 812 __ int3(); |
| 811 } | 813 } |
| 812 } | 814 } |
| 813 | 815 |
| 814 // static | 816 // static |
| 815 void Builtins::Generate_InterpreterPushArgsAndConstructImpl( | 817 void Builtins::Generate_InterpreterPushArgsAndConstructImpl( |
| 816 MacroAssembler* masm, PushArgsConstructMode mode) { | 818 MacroAssembler* masm, InterpreterPushArgsMode mode) { |
| 817 // ----------- S t a t e ------------- | 819 // ----------- S t a t e ------------- |
| 818 // -- rax : the number of arguments (not including the receiver) | 820 // -- rax : the number of arguments (not including the receiver) |
| 819 // -- rdx : the new target (either the same as the constructor or | 821 // -- rdx : the new target (either the same as the constructor or |
| 820 // the JSFunction on which new was invoked initially) | 822 // the JSFunction on which new was invoked initially) |
| 821 // -- rdi : the constructor to call (can be any Object) | 823 // -- rdi : the constructor to call (can be any Object) |
| 822 // -- rbx : the allocation site feedback if available, undefined otherwise | 824 // -- rbx : the allocation site feedback if available, undefined otherwise |
| 823 // -- rcx : the address of the first argument to be pushed. Subsequent | 825 // -- rcx : the address of the first argument to be pushed. Subsequent |
| 824 // arguments should be consecutive above this, in the same order as | 826 // arguments should be consecutive above this, in the same order as |
| 825 // they are to be pushed onto the stack. | 827 // they are to be pushed onto the stack. |
| 826 // ----------------------------------- | 828 // ----------------------------------- |
| 827 Label stack_overflow; | 829 Label stack_overflow; |
| 828 | 830 |
| 829 // Add a stack check before pushing arguments. | 831 // Add a stack check before pushing arguments. |
| 830 Generate_StackOverflowCheck(masm, rax, r8, r9, &stack_overflow); | 832 Generate_StackOverflowCheck(masm, rax, r8, r9, &stack_overflow); |
| 831 | 833 |
| 832 // Pop return address to allow tail-call after pushing arguments. | 834 // Pop return address to allow tail-call after pushing arguments. |
| 833 __ PopReturnAddressTo(kScratchRegister); | 835 __ PopReturnAddressTo(kScratchRegister); |
| 834 | 836 |
| 835 // Push slot for the receiver to be constructed. | 837 // Push slot for the receiver to be constructed. |
| 836 __ Push(Immediate(0)); | 838 __ Push(Immediate(0)); |
| 837 | 839 |
| 838 // rcx and r8 will be modified. | 840 // rcx and r8 will be modified. |
| 839 Generate_InterpreterPushArgs(masm, rax, rcx, r8); | 841 Generate_InterpreterPushArgs(masm, rax, rcx, r8); |
| 840 | 842 |
| 841 // Push return address in preparation for the tail-call. | 843 // Push return address in preparation for the tail-call. |
| 842 __ PushReturnAddressFrom(kScratchRegister); | 844 __ PushReturnAddressFrom(kScratchRegister); |
| 843 | 845 |
| 844 __ AssertUndefinedOrAllocationSite(rbx); | 846 __ AssertUndefinedOrAllocationSite(rbx); |
| 845 if (mode == PushArgsConstructMode::kJSFunction) { | 847 if (mode == InterpreterPushArgsMode::kJSFunction) { |
| 846 // Tail call to the function-specific construct stub (still in the caller | 848 // Tail call to the function-specific construct stub (still in the caller |
| 847 // context at this point). | 849 // context at this point). |
| 848 __ AssertFunction(rdi); | 850 __ AssertFunction(rdi); |
| 849 | 851 |
| 850 __ movp(rcx, FieldOperand(rdi, JSFunction::kSharedFunctionInfoOffset)); | 852 __ movp(rcx, FieldOperand(rdi, JSFunction::kSharedFunctionInfoOffset)); |
| 851 __ movp(rcx, FieldOperand(rcx, SharedFunctionInfo::kConstructStubOffset)); | 853 __ movp(rcx, FieldOperand(rcx, SharedFunctionInfo::kConstructStubOffset)); |
| 852 __ leap(rcx, FieldOperand(rcx, Code::kHeaderSize)); | 854 __ leap(rcx, FieldOperand(rcx, Code::kHeaderSize)); |
| 853 // Jump to the constructor function (rax, rbx, rdx passed on). | 855 // Jump to the constructor function (rax, rbx, rdx passed on). |
| 854 __ jmp(rcx); | 856 __ jmp(rcx); |
| 855 } else if (mode == PushArgsConstructMode::kWithFinalSpread) { | 857 } else if (mode == InterpreterPushArgsMode::kWithFinalSpread) { |
| 856 // Call the constructor (rax, rdx, rdi passed on). | 858 // Call the constructor (rax, rdx, rdi passed on). |
| 857 __ Jump(masm->isolate()->builtins()->ConstructWithSpread(), | 859 __ Jump(masm->isolate()->builtins()->ConstructWithSpread(), |
| 858 RelocInfo::CODE_TARGET); | 860 RelocInfo::CODE_TARGET); |
| 859 } else { | 861 } else { |
| 860 DCHECK_EQ(PushArgsConstructMode::kOther, mode); | 862 DCHECK_EQ(InterpreterPushArgsMode::kOther, mode); |
| 861 // Call the constructor (rax, rdx, rdi passed on). | 863 // Call the constructor (rax, rdx, rdi passed on). |
| 862 __ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET); | 864 __ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET); |
| 863 } | 865 } |
| 864 | 866 |
| 865 // Throw stack overflow exception. | 867 // Throw stack overflow exception. |
| 866 __ bind(&stack_overflow); | 868 __ bind(&stack_overflow); |
| 867 { | 869 { |
| 868 __ TailCallRuntime(Runtime::kThrowStackOverflow); | 870 __ TailCallRuntime(Runtime::kThrowStackOverflow); |
| 869 // This should be unreachable. | 871 // This should be unreachable. |
| 870 __ int3(); | 872 __ int3(); |
| (...skipping 1883 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2754 | 2756 |
| 2755 // 3. Call to something that is not callable. | 2757 // 3. Call to something that is not callable. |
| 2756 __ bind(&non_callable); | 2758 __ bind(&non_callable); |
| 2757 { | 2759 { |
| 2758 FrameScope scope(masm, StackFrame::INTERNAL); | 2760 FrameScope scope(masm, StackFrame::INTERNAL); |
| 2759 __ Push(rdi); | 2761 __ Push(rdi); |
| 2760 __ CallRuntime(Runtime::kThrowCalledNonCallable); | 2762 __ CallRuntime(Runtime::kThrowCalledNonCallable); |
| 2761 } | 2763 } |
| 2762 } | 2764 } |
| 2763 | 2765 |
| 2766 static void CheckSpreadAndPushToStack(MacroAssembler* masm) { |
| 2767 // Load the spread argument into rbx. |
| 2768 __ movp(rbx, Operand(rsp, kPointerSize)); |
| 2769 // Load the map of the spread into r15. |
| 2770 __ movp(r15, FieldOperand(rbx, HeapObject::kMapOffset)); |
| 2771 // Load native context into r14. |
| 2772 __ movp(r14, NativeContextOperand()); |
| 2773 |
| 2774 Label runtime_call, push_args; |
| 2775 // Check that the spread is an array. |
| 2776 __ CmpInstanceType(r15, JS_ARRAY_TYPE); |
| 2777 __ j(not_equal, &runtime_call); |
| 2778 |
| 2779 // Check that we have the original ArrayPrototype. |
| 2780 __ movp(rcx, FieldOperand(r15, Map::kPrototypeOffset)); |
| 2781 __ cmpp(rcx, ContextOperand(r14, Context::INITIAL_ARRAY_PROTOTYPE_INDEX)); |
| 2782 __ j(not_equal, &runtime_call); |
| 2783 |
| 2784 // Check that the ArrayPrototype hasn't been modified in a way that would |
| 2785 // affect iteration. |
| 2786 __ LoadRoot(rcx, Heap::kArrayIteratorProtectorRootIndex); |
| 2787 __ Cmp(FieldOperand(rcx, Cell::kValueOffset), |
| 2788 Smi::FromInt(Isolate::kProtectorValid)); |
| 2789 __ j(not_equal, &runtime_call); |
| 2790 |
| 2791 // Check that the map of the initial array iterator hasn't changed. |
| 2792 __ movp(rcx, |
| 2793 ContextOperand(r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_INDEX)); |
| 2794 __ movp(rcx, FieldOperand(rcx, HeapObject::kMapOffset)); |
| 2795 __ cmpp(rcx, ContextOperand( |
| 2796 r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_MAP_INDEX)); |
| 2797 __ j(not_equal, &runtime_call); |
| 2798 |
| 2799 // For FastPacked kinds, iteration will have the same effect as simply |
| 2800 // accessing each property in order. |
| 2801 Label no_protector_check; |
| 2802 __ movzxbp(rcx, FieldOperand(r15, Map::kBitField2Offset)); |
| 2803 __ DecodeField<Map::ElementsKindBits>(rcx); |
| 2804 __ cmpp(rcx, Immediate(FAST_HOLEY_ELEMENTS)); |
| 2805 __ j(above, &runtime_call); |
| 2806 // For non-FastHoley kinds, we can skip the protector check. |
| 2807 __ cmpp(rcx, Immediate(FAST_SMI_ELEMENTS)); |
| 2808 __ j(equal, &no_protector_check); |
| 2809 __ cmpp(rcx, Immediate(FAST_ELEMENTS)); |
| 2810 __ j(equal, &no_protector_check); |
| 2811 // Check the ArrayProtector cell. |
| 2812 __ LoadRoot(rcx, Heap::kArrayProtectorRootIndex); |
| 2813 __ Cmp(FieldOperand(rcx, PropertyCell::kValueOffset), |
| 2814 Smi::FromInt(Isolate::kProtectorValid)); |
| 2815 __ j(not_equal, &runtime_call); |
| 2816 |
| 2817 __ bind(&no_protector_check); |
| 2818 // Load the FixedArray backing store, but use the length from the array. |
| 2819 __ SmiToInteger32(r9, FieldOperand(rbx, JSArray::kLengthOffset)); |
| 2820 __ movp(rbx, FieldOperand(rbx, JSArray::kElementsOffset)); |
| 2821 __ jmp(&push_args); |
| 2822 |
| 2823 __ bind(&runtime_call); |
| 2824 { |
| 2825 // Call the builtin for the result of the spread. |
| 2826 FrameScope scope(masm, StackFrame::INTERNAL); |
| 2827 __ Push(rdi); // target |
| 2828 __ Push(rdx); // new target |
| 2829 __ Integer32ToSmi(rax, rax); |
| 2830 __ Push(rax); // nargs |
| 2831 __ Push(rbx); |
| 2832 __ CallRuntime(Runtime::kSpreadIterableFixed); |
| 2833 __ movp(rbx, rax); |
| 2834 __ Pop(rax); // nargs |
| 2835 __ SmiToInteger32(rax, rax); |
| 2836 __ Pop(rdx); // new target |
| 2837 __ Pop(rdi); // target |
| 2838 } |
| 2839 |
| 2840 { |
| 2841 // Calculate the new nargs including the result of the spread. |
| 2842 __ SmiToInteger32(r9, FieldOperand(rbx, FixedArray::kLengthOffset)); |
| 2843 |
| 2844 __ bind(&push_args); |
| 2845 // rax += r9 - 1. Subtract 1 for the spread itself. |
| 2846 __ leap(rax, Operand(rax, r9, times_1, -1)); |
| 2847 |
| 2848 // Pop the return address and spread argument. |
| 2849 __ PopReturnAddressTo(r8); |
| 2850 __ Pop(rcx); |
| 2851 } |
| 2852 |
| 2853 // Check for stack overflow. |
| 2854 { |
| 2855 // Check the stack for overflow. We are not trying to catch interruptions |
| 2856 // (i.e. debug break and preemption) here, so check the "real stack limit". |
| 2857 Label done; |
| 2858 __ LoadRoot(kScratchRegister, Heap::kRealStackLimitRootIndex); |
| 2859 __ movp(rcx, rsp); |
| 2860 // Make rcx the space we have left. The stack might already be overflowed |
| 2861 // here which will cause rcx to become negative. |
| 2862 __ subp(rcx, kScratchRegister); |
| 2863 __ sarp(rcx, Immediate(kPointerSizeLog2)); |
| 2864 // Check if the arguments will overflow the stack. |
| 2865 __ cmpp(rcx, r9); |
| 2866 __ j(greater, &done, Label::kNear); // Signed comparison. |
| 2867 __ TailCallRuntime(Runtime::kThrowStackOverflow); |
| 2868 __ bind(&done); |
| 2869 } |
| 2870 |
| 2871 // Put the evaluated spread onto the stack as additional arguments. |
| 2872 { |
| 2873 __ Set(rcx, 0); |
| 2874 Label done, loop; |
| 2875 __ bind(&loop); |
| 2876 __ cmpl(rcx, r9); |
| 2877 __ j(equal, &done, Label::kNear); |
| 2878 __ movp(kScratchRegister, FieldOperand(rbx, rcx, times_pointer_size, |
| 2879 FixedArray::kHeaderSize)); |
| 2880 __ Push(kScratchRegister); |
| 2881 __ incl(rcx); |
| 2882 __ jmp(&loop); |
| 2883 __ bind(&done); |
| 2884 __ PushReturnAddressFrom(r8); |
| 2885 } |
| 2886 } |
| 2887 |
| 2888 // static |
| 2889 void Builtins::Generate_CallWithSpread(MacroAssembler* masm) { |
| 2890 // ----------- S t a t e ------------- |
| 2891 // -- rax : the number of arguments (not including the receiver) |
| 2892 // -- rdi : the target to call (can be any Object) |
| 2893 // ----------------------------------- |
| 2894 |
| 2895 // CheckSpreadAndPushToStack will push rdx to save it. |
| 2896 __ LoadRoot(rdx, Heap::kUndefinedValueRootIndex); |
| 2897 CheckSpreadAndPushToStack(masm); |
| 2898 __ Jump(masm->isolate()->builtins()->Call(ConvertReceiverMode::kAny, |
| 2899 TailCallMode::kDisallow), |
| 2900 RelocInfo::CODE_TARGET); |
| 2901 } |
| 2902 |
| 2764 // static | 2903 // static |
| 2765 void Builtins::Generate_ConstructFunction(MacroAssembler* masm) { | 2904 void Builtins::Generate_ConstructFunction(MacroAssembler* masm) { |
| 2766 // ----------- S t a t e ------------- | 2905 // ----------- S t a t e ------------- |
| 2767 // -- rax : the number of arguments (not including the receiver) | 2906 // -- rax : the number of arguments (not including the receiver) |
| 2768 // -- rdx : the new target (checked to be a constructor) | 2907 // -- rdx : the new target (checked to be a constructor) |
| 2769 // -- rdi : the constructor to call (checked to be a JSFunction) | 2908 // -- rdi : the constructor to call (checked to be a JSFunction) |
| 2770 // ----------------------------------- | 2909 // ----------------------------------- |
| 2771 __ AssertFunction(rdi); | 2910 __ AssertFunction(rdi); |
| 2772 | 2911 |
| 2773 // Calling convention for function specific ConstructStubs require | 2912 // Calling convention for function specific ConstructStubs require |
| (...skipping 111 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2885 | 3024 |
| 2886 // static | 3025 // static |
| 2887 void Builtins::Generate_ConstructWithSpread(MacroAssembler* masm) { | 3026 void Builtins::Generate_ConstructWithSpread(MacroAssembler* masm) { |
| 2888 // ----------- S t a t e ------------- | 3027 // ----------- S t a t e ------------- |
| 2889 // -- rax : the number of arguments (not including the receiver) | 3028 // -- rax : the number of arguments (not including the receiver) |
| 2890 // -- rdx : the new target (either the same as the constructor or | 3029 // -- rdx : the new target (either the same as the constructor or |
| 2891 // the JSFunction on which new was invoked initially) | 3030 // the JSFunction on which new was invoked initially) |
| 2892 // -- rdi : the constructor to call (can be any Object) | 3031 // -- rdi : the constructor to call (can be any Object) |
| 2893 // ----------------------------------- | 3032 // ----------------------------------- |
| 2894 | 3033 |
| 2895 // Load the spread argument into rbx. | 3034 CheckSpreadAndPushToStack(masm); |
| 2896 __ movp(rbx, Operand(rsp, kPointerSize)); | |
| 2897 // Load the map of the spread into r15. | |
| 2898 __ movp(r15, FieldOperand(rbx, HeapObject::kMapOffset)); | |
| 2899 // Load native context into r14. | |
| 2900 __ movp(r14, NativeContextOperand()); | |
| 2901 | |
| 2902 Label runtime_call, push_args; | |
| 2903 // Check that the spread is an array. | |
| 2904 __ CmpInstanceType(r15, JS_ARRAY_TYPE); | |
| 2905 __ j(not_equal, &runtime_call); | |
| 2906 | |
| 2907 // Check that we have the original ArrayPrototype. | |
| 2908 __ movp(rcx, FieldOperand(r15, Map::kPrototypeOffset)); | |
| 2909 __ cmpp(rcx, ContextOperand(r14, Context::INITIAL_ARRAY_PROTOTYPE_INDEX)); | |
| 2910 __ j(not_equal, &runtime_call); | |
| 2911 | |
| 2912 // Check that the ArrayPrototype hasn't been modified in a way that would | |
| 2913 // affect iteration. | |
| 2914 __ LoadRoot(rcx, Heap::kArrayIteratorProtectorRootIndex); | |
| 2915 __ Cmp(FieldOperand(rcx, Cell::kValueOffset), | |
| 2916 Smi::FromInt(Isolate::kProtectorValid)); | |
| 2917 __ j(not_equal, &runtime_call); | |
| 2918 | |
| 2919 // Check that the map of the initial array iterator hasn't changed. | |
| 2920 __ movp(rcx, | |
| 2921 ContextOperand(r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_INDEX)); | |
| 2922 __ movp(rcx, FieldOperand(rcx, HeapObject::kMapOffset)); | |
| 2923 __ cmpp(rcx, ContextOperand( | |
| 2924 r14, Context::INITIAL_ARRAY_ITERATOR_PROTOTYPE_MAP_INDEX)); | |
| 2925 __ j(not_equal, &runtime_call); | |
| 2926 | |
| 2927 // For FastPacked kinds, iteration will have the same effect as simply | |
| 2928 // accessing each property in order. | |
| 2929 Label no_protector_check; | |
| 2930 __ movzxbp(rcx, FieldOperand(r15, Map::kBitField2Offset)); | |
| 2931 __ DecodeField<Map::ElementsKindBits>(rcx); | |
| 2932 __ cmpp(rcx, Immediate(FAST_HOLEY_ELEMENTS)); | |
| 2933 __ j(above, &runtime_call); | |
| 2934 // For non-FastHoley kinds, we can skip the protector check. | |
| 2935 __ cmpp(rcx, Immediate(FAST_SMI_ELEMENTS)); | |
| 2936 __ j(equal, &no_protector_check); | |
| 2937 __ cmpp(rcx, Immediate(FAST_ELEMENTS)); | |
| 2938 __ j(equal, &no_protector_check); | |
| 2939 // Check the ArrayProtector cell. | |
| 2940 __ LoadRoot(rcx, Heap::kArrayProtectorRootIndex); | |
| 2941 __ Cmp(FieldOperand(rcx, PropertyCell::kValueOffset), | |
| 2942 Smi::FromInt(Isolate::kProtectorValid)); | |
| 2943 __ j(not_equal, &runtime_call); | |
| 2944 | |
| 2945 __ bind(&no_protector_check); | |
| 2946 // Load the FixedArray backing store. | |
| 2947 __ movp(rbx, FieldOperand(rbx, JSArray::kElementsOffset)); | |
| 2948 __ jmp(&push_args); | |
| 2949 | |
| 2950 __ bind(&runtime_call); | |
| 2951 { | |
| 2952 // Call the builtin for the result of the spread. | |
| 2953 FrameScope scope(masm, StackFrame::INTERNAL); | |
| 2954 __ Push(rdi); // target | |
| 2955 __ Push(rdx); // new target | |
| 2956 __ Integer32ToSmi(rax, rax); | |
| 2957 __ Push(rax); // nargs | |
| 2958 __ Push(rbx); | |
| 2959 __ CallRuntime(Runtime::kSpreadIterableFixed); | |
| 2960 __ movp(rbx, rax); | |
| 2961 __ Pop(rax); // nargs | |
| 2962 __ SmiToInteger32(rax, rax); | |
| 2963 __ Pop(rdx); // new target | |
| 2964 __ Pop(rdi); // target | |
| 2965 } | |
| 2966 | |
| 2967 __ bind(&push_args); | |
| 2968 { | |
| 2969 // Pop the return address and spread argument. | |
| 2970 __ PopReturnAddressTo(r8); | |
| 2971 __ Pop(rcx); | |
| 2972 | |
| 2973 // Calculate the new nargs including the result of the spread. | |
| 2974 __ SmiToInteger32(r9, FieldOperand(rbx, FixedArray::kLengthOffset)); | |
| 2975 // rax += r9 - 1. Subtract 1 for the spread itself. | |
| 2976 __ leap(rax, Operand(rax, r9, times_1, -1)); | |
| 2977 } | |
| 2978 | |
| 2979 // Check for stack overflow. | |
| 2980 { | |
| 2981 // Check the stack for overflow. We are not trying to catch interruptions | |
| 2982 // (i.e. debug break and preemption) here, so check the "real stack limit". | |
| 2983 Label done; | |
| 2984 __ LoadRoot(kScratchRegister, Heap::kRealStackLimitRootIndex); | |
| 2985 __ movp(rcx, rsp); | |
| 2986 // Make rcx the space we have left. The stack might already be overflowed | |
| 2987 // here which will cause rcx to become negative. | |
| 2988 __ subp(rcx, kScratchRegister); | |
| 2989 __ sarp(rcx, Immediate(kPointerSizeLog2)); | |
| 2990 // Check if the arguments will overflow the stack. | |
| 2991 __ cmpp(rcx, r9); | |
| 2992 __ j(greater, &done, Label::kNear); // Signed comparison. | |
| 2993 __ TailCallRuntime(Runtime::kThrowStackOverflow); | |
| 2994 __ bind(&done); | |
| 2995 } | |
| 2996 | |
| 2997 // Put the evaluated spread onto the stack as additional arguments. | |
| 2998 { | |
| 2999 __ Set(rcx, 0); | |
| 3000 Label done, loop; | |
| 3001 __ bind(&loop); | |
| 3002 __ cmpl(rcx, r9); | |
| 3003 __ j(equal, &done, Label::kNear); | |
| 3004 __ movp(kScratchRegister, FieldOperand(rbx, rcx, times_pointer_size, | |
| 3005 FixedArray::kHeaderSize)); | |
| 3006 __ Push(kScratchRegister); | |
| 3007 __ incl(rcx); | |
| 3008 __ jmp(&loop); | |
| 3009 __ bind(&done); | |
| 3010 __ PushReturnAddressFrom(r8); | |
| 3011 } | |
| 3012 // Dispatch. | |
| 3013 __ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET); | 3035 __ Jump(masm->isolate()->builtins()->Construct(), RelocInfo::CODE_TARGET); |
| 3014 } | 3036 } |
| 3015 | 3037 |
| 3016 static void CompatibleReceiverCheck(MacroAssembler* masm, Register receiver, | 3038 static void CompatibleReceiverCheck(MacroAssembler* masm, Register receiver, |
| 3017 Register function_template_info, | 3039 Register function_template_info, |
| 3018 Register scratch0, Register scratch1, | 3040 Register scratch0, Register scratch1, |
| 3019 Register scratch2, | 3041 Register scratch2, |
| 3020 Label* receiver_check_failed) { | 3042 Label* receiver_check_failed) { |
| 3021 Register signature = scratch0; | 3043 Register signature = scratch0; |
| 3022 Register map = scratch1; | 3044 Register map = scratch1; |
| (...skipping 155 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3178 void Builtins::Generate_InterpreterOnStackReplacement(MacroAssembler* masm) { | 3200 void Builtins::Generate_InterpreterOnStackReplacement(MacroAssembler* masm) { |
| 3179 Generate_OnStackReplacementHelper(masm, true); | 3201 Generate_OnStackReplacementHelper(masm, true); |
| 3180 } | 3202 } |
| 3181 | 3203 |
| 3182 #undef __ | 3204 #undef __ |
| 3183 | 3205 |
| 3184 } // namespace internal | 3206 } // namespace internal |
| 3185 } // namespace v8 | 3207 } // namespace v8 |
| 3186 | 3208 |
| 3187 #endif // V8_TARGET_ARCH_X64 | 3209 #endif // V8_TARGET_ARCH_X64 |
| OLD | NEW |