| Index: src/x64/builtins-x64.cc
|
| diff --git a/src/x64/builtins-x64.cc b/src/x64/builtins-x64.cc
|
| index a8160302196c2e835ee8563e239e89cc224082b0..c9be9b7c9c3cedf9e4d9dbffc791ac880c837eab 100644
|
| --- a/src/x64/builtins-x64.cc
|
| +++ b/src/x64/builtins-x64.cc
|
| @@ -69,7 +69,7 @@ void Builtins::Generate_Adaptor(MacroAssembler* masm,
|
|
|
| // JumpToExternalReference expects rax to contain the number of arguments
|
| // including the receiver and the extra arguments.
|
| - __ addq(rax, Immediate(num_extra_args + 1));
|
| + __ addp(rax, Immediate(num_extra_args + 1));
|
| __ JumpToExternalReference(ExternalReference(id, masm->isolate()), 1);
|
| }
|
|
|
| @@ -289,10 +289,10 @@ static void Generate_JSConstructStubHelper(MacroAssembler* masm,
|
| __ movzxbq(rdx, FieldOperand(rax, Map::kUnusedPropertyFieldsOffset));
|
| __ movzxbq(rcx,
|
| FieldOperand(rax, Map::kPreAllocatedPropertyFieldsOffset));
|
| - __ addq(rdx, rcx);
|
| + __ addp(rdx, rcx);
|
| // Calculate unused properties past the end of the in-object properties.
|
| __ movzxbq(rcx, FieldOperand(rax, Map::kInObjectPropertiesOffset));
|
| - __ subq(rdx, rcx);
|
| + __ subp(rdx, rcx);
|
| // Done if no extra properties are to be allocated.
|
| __ j(zero, &allocated);
|
| __ Assert(positive, kPropertyAllocationCountFailed);
|
| @@ -332,7 +332,7 @@ static void Generate_JSConstructStubHelper(MacroAssembler* masm,
|
| __ jmp(&entry);
|
| __ bind(&loop);
|
| __ movp(Operand(rcx, 0), rdx);
|
| - __ addq(rcx, Immediate(kPointerSize));
|
| + __ addp(rcx, Immediate(kPointerSize));
|
| __ bind(&entry);
|
| __ cmpq(rcx, rax);
|
| __ j(below, &loop);
|
| @@ -590,7 +590,7 @@ static void Generate_JSEntryTrampolineHelper(MacroAssembler* masm,
|
| __ bind(&loop);
|
| __ movp(kScratchRegister, Operand(rbx, rcx, times_pointer_size, 0));
|
| __ Push(Operand(kScratchRegister, 0)); // dereference handle
|
| - __ addq(rcx, Immediate(1));
|
| + __ addp(rcx, Immediate(1));
|
| __ bind(&entry);
|
| __ cmpq(rcx, rax);
|
| __ j(not_equal, &loop);
|
| @@ -670,7 +670,7 @@ static void GenerateMakeCodeYoungAgainCommon(MacroAssembler* masm) {
|
|
|
| // Re-execute the code that was patched back to the young age when
|
| // the stub returns.
|
| - __ subq(Operand(rsp, 0), Immediate(5));
|
| + __ subp(Operand(rsp, 0), Immediate(5));
|
| __ Pushad();
|
| __ Move(arg_reg_2, ExternalReference::isolate_address(masm->isolate()));
|
| __ movp(arg_reg_1, Operand(rsp, kNumSafepointRegisters * kPointerSize));
|
| @@ -706,7 +706,7 @@ void Builtins::Generate_MarkCodeAsExecutedOnce(MacroAssembler* masm) {
|
| __ Pushad();
|
| __ Move(arg_reg_2, ExternalReference::isolate_address(masm->isolate()));
|
| __ movp(arg_reg_1, Operand(rsp, kNumSafepointRegisters * kPointerSize));
|
| - __ subq(arg_reg_1, Immediate(Assembler::kShortCallInstructionLength));
|
| + __ subp(arg_reg_1, Immediate(Assembler::kShortCallInstructionLength));
|
| { // NOLINT
|
| FrameScope scope(masm, StackFrame::MANUAL);
|
| __ PrepareCallCFunction(2);
|
| @@ -1007,7 +1007,7 @@ void Builtins::Generate_FunctionApply(MacroAssembler* masm) {
|
| __ movp(rcx, rsp);
|
| // Make rcx the space we have left. The stack might already be overflowed
|
| // here which will cause rcx to become negative.
|
| - __ subq(rcx, kScratchRegister);
|
| + __ subp(rcx, kScratchRegister);
|
| // Make rdx the space we need for the array when it is unrolled onto the
|
| // stack.
|
| __ PositiveSmiTimesPowerOfTwoToInteger64(rdx, rax, kPointerSizeLog2);
|
| @@ -1388,7 +1388,7 @@ void Builtins::Generate_ArgumentsAdaptorTrampoline(MacroAssembler* masm) {
|
| __ bind(©);
|
| __ incq(r8);
|
| __ Push(Operand(rax, 0));
|
| - __ subq(rax, Immediate(kPointerSize));
|
| + __ subp(rax, Immediate(kPointerSize));
|
| __ cmpq(r8, rbx);
|
| __ j(less, ©);
|
| __ jmp(&invoke);
|
| @@ -1407,7 +1407,7 @@ void Builtins::Generate_ArgumentsAdaptorTrampoline(MacroAssembler* masm) {
|
| __ bind(©);
|
| __ incq(r8);
|
| __ Push(Operand(rdi, 0));
|
| - __ subq(rdi, Immediate(kPointerSize));
|
| + __ subp(rdi, Immediate(kPointerSize));
|
| __ cmpq(r8, rax);
|
| __ j(less, ©);
|
|
|
|
|