| Index: src/x64/stub-cache-x64.cc
|
| diff --git a/src/x64/stub-cache-x64.cc b/src/x64/stub-cache-x64.cc
|
| index 5c31133db4a53862990e5a9a5c9182ebd18867ff..3b1f3a52ea8aa018671ea4da27e0c05bd434c1a4 100644
|
| --- a/src/x64/stub-cache-x64.cc
|
| +++ b/src/x64/stub-cache-x64.cc
|
| @@ -2322,7 +2322,7 @@ Handle<Code> CallStubCompiler::CompileMathFloorCall(
|
| Label already_round;
|
| __ bind(&conversion_failure);
|
| int64_t kTwoMantissaBits= V8_INT64_C(0x4330000000000000);
|
| - __ movq(rbx, kTwoMantissaBits, RelocInfo::NONE64);
|
| + __ movq(rbx, kTwoMantissaBits);
|
| __ movq(xmm1, rbx);
|
| __ ucomisd(xmm0, xmm1);
|
| __ j(above_equal, &already_round);
|
| @@ -2343,7 +2343,7 @@ Handle<Code> CallStubCompiler::CompileMathFloorCall(
|
|
|
| // Subtract 1 if the argument was less than the tentative result.
|
| int64_t kOne = V8_INT64_C(0x3ff0000000000000);
|
| - __ movq(rbx, kOne, RelocInfo::NONE64);
|
| + __ movq(rbx, kOne);
|
| __ movq(xmm1, rbx);
|
| __ andpd(xmm1, xmm2);
|
| __ subsd(xmm0, xmm1);
|
| @@ -2447,8 +2447,7 @@ Handle<Code> CallStubCompiler::CompileMathAbsCall(
|
| Label negative_sign;
|
| const int sign_mask_shift =
|
| (HeapNumber::kExponentOffset - HeapNumber::kValueOffset) * kBitsPerByte;
|
| - __ movq(rdi, static_cast<int64_t>(HeapNumber::kSignMask) << sign_mask_shift,
|
| - RelocInfo::NONE64);
|
| + __ movq(rdi, static_cast<int64_t>(HeapNumber::kSignMask) << sign_mask_shift);
|
| __ testq(rbx, rdi);
|
| __ j(not_zero, &negative_sign);
|
| __ ret(2 * kPointerSize);
|
|
|