OLD | NEW |
1 // Copyright 2013 the V8 project authors. All rights reserved. | 1 // Copyright 2013 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #if V8_TARGET_ARCH_X64 | 5 #if V8_TARGET_ARCH_X64 |
6 | 6 |
7 #include "src/bootstrapper.h" | 7 #include "src/bootstrapper.h" |
8 #include "src/code-stubs.h" | 8 #include "src/code-stubs.h" |
9 #include "src/codegen.h" | 9 #include "src/codegen.h" |
10 #include "src/ic/handler-compiler.h" | 10 #include "src/ic/handler-compiler.h" |
(...skipping 162 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
173 Register result_reg = final_result_reg.is(rcx) ? rax : final_result_reg; | 173 Register result_reg = final_result_reg.is(rcx) ? rax : final_result_reg; |
174 // Save ecx if it isn't the return register and therefore volatile, or if it | 174 // Save ecx if it isn't the return register and therefore volatile, or if it |
175 // is the return register, then save the temp register we use in its stead | 175 // is the return register, then save the temp register we use in its stead |
176 // for the result. | 176 // for the result. |
177 Register save_reg = final_result_reg.is(rcx) ? rax : rcx; | 177 Register save_reg = final_result_reg.is(rcx) ? rax : rcx; |
178 __ pushq(scratch1); | 178 __ pushq(scratch1); |
179 __ pushq(save_reg); | 179 __ pushq(save_reg); |
180 | 180 |
181 bool stash_exponent_copy = !input_reg.is(rsp); | 181 bool stash_exponent_copy = !input_reg.is(rsp); |
182 __ movl(scratch1, mantissa_operand); | 182 __ movl(scratch1, mantissa_operand); |
183 __ movsd(xmm0, mantissa_operand); | 183 __ Movsd(xmm0, mantissa_operand); |
184 __ movl(rcx, exponent_operand); | 184 __ movl(rcx, exponent_operand); |
185 if (stash_exponent_copy) __ pushq(rcx); | 185 if (stash_exponent_copy) __ pushq(rcx); |
186 | 186 |
187 __ andl(rcx, Immediate(HeapNumber::kExponentMask)); | 187 __ andl(rcx, Immediate(HeapNumber::kExponentMask)); |
188 __ shrl(rcx, Immediate(HeapNumber::kExponentShift)); | 188 __ shrl(rcx, Immediate(HeapNumber::kExponentShift)); |
189 __ leal(result_reg, MemOperand(rcx, -HeapNumber::kExponentBias)); | 189 __ leal(result_reg, MemOperand(rcx, -HeapNumber::kExponentBias)); |
190 __ cmpl(result_reg, Immediate(HeapNumber::kMantissaBits)); | 190 __ cmpl(result_reg, Immediate(HeapNumber::kMantissaBits)); |
191 __ j(below, &process_64_bits); | 191 __ j(below, &process_64_bits); |
192 | 192 |
193 // Result is entirely in lower 32-bits of mantissa | 193 // Result is entirely in lower 32-bits of mantissa |
(...skipping 36 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
230 | 230 |
231 | 231 |
232 void FloatingPointHelper::LoadSSE2UnknownOperands(MacroAssembler* masm, | 232 void FloatingPointHelper::LoadSSE2UnknownOperands(MacroAssembler* masm, |
233 Label* not_numbers) { | 233 Label* not_numbers) { |
234 Label load_smi_rdx, load_nonsmi_rax, load_smi_rax, load_float_rax, done; | 234 Label load_smi_rdx, load_nonsmi_rax, load_smi_rax, load_float_rax, done; |
235 // Load operand in rdx into xmm0, or branch to not_numbers. | 235 // Load operand in rdx into xmm0, or branch to not_numbers. |
236 __ LoadRoot(rcx, Heap::kHeapNumberMapRootIndex); | 236 __ LoadRoot(rcx, Heap::kHeapNumberMapRootIndex); |
237 __ JumpIfSmi(rdx, &load_smi_rdx); | 237 __ JumpIfSmi(rdx, &load_smi_rdx); |
238 __ cmpp(FieldOperand(rdx, HeapObject::kMapOffset), rcx); | 238 __ cmpp(FieldOperand(rdx, HeapObject::kMapOffset), rcx); |
239 __ j(not_equal, not_numbers); // Argument in rdx is not a number. | 239 __ j(not_equal, not_numbers); // Argument in rdx is not a number. |
240 __ movsd(xmm0, FieldOperand(rdx, HeapNumber::kValueOffset)); | 240 __ Movsd(xmm0, FieldOperand(rdx, HeapNumber::kValueOffset)); |
241 // Load operand in rax into xmm1, or branch to not_numbers. | 241 // Load operand in rax into xmm1, or branch to not_numbers. |
242 __ JumpIfSmi(rax, &load_smi_rax); | 242 __ JumpIfSmi(rax, &load_smi_rax); |
243 | 243 |
244 __ bind(&load_nonsmi_rax); | 244 __ bind(&load_nonsmi_rax); |
245 __ cmpp(FieldOperand(rax, HeapObject::kMapOffset), rcx); | 245 __ cmpp(FieldOperand(rax, HeapObject::kMapOffset), rcx); |
246 __ j(not_equal, not_numbers); | 246 __ j(not_equal, not_numbers); |
247 __ movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); | 247 __ Movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); |
248 __ jmp(&done); | 248 __ jmp(&done); |
249 | 249 |
250 __ bind(&load_smi_rdx); | 250 __ bind(&load_smi_rdx); |
251 __ SmiToInteger32(kScratchRegister, rdx); | 251 __ SmiToInteger32(kScratchRegister, rdx); |
252 __ Cvtlsi2sd(xmm0, kScratchRegister); | 252 __ Cvtlsi2sd(xmm0, kScratchRegister); |
253 __ JumpIfNotSmi(rax, &load_nonsmi_rax); | 253 __ JumpIfNotSmi(rax, &load_nonsmi_rax); |
254 | 254 |
255 __ bind(&load_smi_rax); | 255 __ bind(&load_smi_rax); |
256 __ SmiToInteger32(kScratchRegister, rax); | 256 __ SmiToInteger32(kScratchRegister, rax); |
257 __ Cvtlsi2sd(xmm1, kScratchRegister); | 257 __ Cvtlsi2sd(xmm1, kScratchRegister); |
(...skipping 23 matching lines...) Expand all Loading... |
281 // This can only happen if the stub is called from non-optimized code. | 281 // This can only happen if the stub is called from non-optimized code. |
282 // Load input parameters from stack. | 282 // Load input parameters from stack. |
283 StackArgumentsAccessor args(rsp, 2, ARGUMENTS_DONT_CONTAIN_RECEIVER); | 283 StackArgumentsAccessor args(rsp, 2, ARGUMENTS_DONT_CONTAIN_RECEIVER); |
284 __ movp(base, args.GetArgumentOperand(0)); | 284 __ movp(base, args.GetArgumentOperand(0)); |
285 __ movp(exponent, args.GetArgumentOperand(1)); | 285 __ movp(exponent, args.GetArgumentOperand(1)); |
286 __ JumpIfSmi(base, &base_is_smi, Label::kNear); | 286 __ JumpIfSmi(base, &base_is_smi, Label::kNear); |
287 __ CompareRoot(FieldOperand(base, HeapObject::kMapOffset), | 287 __ CompareRoot(FieldOperand(base, HeapObject::kMapOffset), |
288 Heap::kHeapNumberMapRootIndex); | 288 Heap::kHeapNumberMapRootIndex); |
289 __ j(not_equal, &call_runtime); | 289 __ j(not_equal, &call_runtime); |
290 | 290 |
291 __ movsd(double_base, FieldOperand(base, HeapNumber::kValueOffset)); | 291 __ Movsd(double_base, FieldOperand(base, HeapNumber::kValueOffset)); |
292 __ jmp(&unpack_exponent, Label::kNear); | 292 __ jmp(&unpack_exponent, Label::kNear); |
293 | 293 |
294 __ bind(&base_is_smi); | 294 __ bind(&base_is_smi); |
295 __ SmiToInteger32(base, base); | 295 __ SmiToInteger32(base, base); |
296 __ Cvtlsi2sd(double_base, base); | 296 __ Cvtlsi2sd(double_base, base); |
297 __ bind(&unpack_exponent); | 297 __ bind(&unpack_exponent); |
298 | 298 |
299 __ JumpIfNotSmi(exponent, &exponent_not_smi, Label::kNear); | 299 __ JumpIfNotSmi(exponent, &exponent_not_smi, Label::kNear); |
300 __ SmiToInteger32(exponent, exponent); | 300 __ SmiToInteger32(exponent, exponent); |
301 __ jmp(&int_exponent); | 301 __ jmp(&int_exponent); |
302 | 302 |
303 __ bind(&exponent_not_smi); | 303 __ bind(&exponent_not_smi); |
304 __ CompareRoot(FieldOperand(exponent, HeapObject::kMapOffset), | 304 __ CompareRoot(FieldOperand(exponent, HeapObject::kMapOffset), |
305 Heap::kHeapNumberMapRootIndex); | 305 Heap::kHeapNumberMapRootIndex); |
306 __ j(not_equal, &call_runtime); | 306 __ j(not_equal, &call_runtime); |
307 __ movsd(double_exponent, FieldOperand(exponent, HeapNumber::kValueOffset)); | 307 __ Movsd(double_exponent, FieldOperand(exponent, HeapNumber::kValueOffset)); |
308 } else if (exponent_type() == TAGGED) { | 308 } else if (exponent_type() == TAGGED) { |
309 __ JumpIfNotSmi(exponent, &exponent_not_smi, Label::kNear); | 309 __ JumpIfNotSmi(exponent, &exponent_not_smi, Label::kNear); |
310 __ SmiToInteger32(exponent, exponent); | 310 __ SmiToInteger32(exponent, exponent); |
311 __ jmp(&int_exponent); | 311 __ jmp(&int_exponent); |
312 | 312 |
313 __ bind(&exponent_not_smi); | 313 __ bind(&exponent_not_smi); |
314 __ movsd(double_exponent, FieldOperand(exponent, HeapNumber::kValueOffset)); | 314 __ Movsd(double_exponent, FieldOperand(exponent, HeapNumber::kValueOffset)); |
315 } | 315 } |
316 | 316 |
317 if (exponent_type() != INTEGER) { | 317 if (exponent_type() != INTEGER) { |
318 Label fast_power, try_arithmetic_simplification; | 318 Label fast_power, try_arithmetic_simplification; |
319 // Detect integer exponents stored as double. | 319 // Detect integer exponents stored as double. |
320 __ DoubleToI(exponent, double_exponent, double_scratch, | 320 __ DoubleToI(exponent, double_exponent, double_scratch, |
321 TREAT_MINUS_ZERO_AS_ZERO, &try_arithmetic_simplification, | 321 TREAT_MINUS_ZERO_AS_ZERO, &try_arithmetic_simplification, |
322 &try_arithmetic_simplification, | 322 &try_arithmetic_simplification, |
323 &try_arithmetic_simplification); | 323 &try_arithmetic_simplification); |
324 __ jmp(&int_exponent); | 324 __ jmp(&int_exponent); |
(...skipping 73 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
398 __ divsd(double_result, double_exponent); | 398 __ divsd(double_result, double_exponent); |
399 __ jmp(&done); | 399 __ jmp(&done); |
400 } | 400 } |
401 | 401 |
402 // Using FPU instructions to calculate power. | 402 // Using FPU instructions to calculate power. |
403 Label fast_power_failed; | 403 Label fast_power_failed; |
404 __ bind(&fast_power); | 404 __ bind(&fast_power); |
405 __ fnclex(); // Clear flags to catch exceptions later. | 405 __ fnclex(); // Clear flags to catch exceptions later. |
406 // Transfer (B)ase and (E)xponent onto the FPU register stack. | 406 // Transfer (B)ase and (E)xponent onto the FPU register stack. |
407 __ subp(rsp, Immediate(kDoubleSize)); | 407 __ subp(rsp, Immediate(kDoubleSize)); |
408 __ movsd(Operand(rsp, 0), double_exponent); | 408 __ Movsd(Operand(rsp, 0), double_exponent); |
409 __ fld_d(Operand(rsp, 0)); // E | 409 __ fld_d(Operand(rsp, 0)); // E |
410 __ movsd(Operand(rsp, 0), double_base); | 410 __ Movsd(Operand(rsp, 0), double_base); |
411 __ fld_d(Operand(rsp, 0)); // B, E | 411 __ fld_d(Operand(rsp, 0)); // B, E |
412 | 412 |
413 // Exponent is in st(1) and base is in st(0) | 413 // Exponent is in st(1) and base is in st(0) |
414 // B ^ E = (2^(E * log2(B)) - 1) + 1 = (2^X - 1) + 1 for X = E * log2(B) | 414 // B ^ E = (2^(E * log2(B)) - 1) + 1 = (2^X - 1) + 1 for X = E * log2(B) |
415 // FYL2X calculates st(1) * log2(st(0)) | 415 // FYL2X calculates st(1) * log2(st(0)) |
416 __ fyl2x(); // X | 416 __ fyl2x(); // X |
417 __ fld(0); // X, X | 417 __ fld(0); // X, X |
418 __ frndint(); // rnd(X), X | 418 __ frndint(); // rnd(X), X |
419 __ fsub(1); // rnd(X), X-rnd(X) | 419 __ fsub(1); // rnd(X), X-rnd(X) |
420 __ fxch(1); // X - rnd(X), rnd(X) | 420 __ fxch(1); // X - rnd(X), rnd(X) |
421 // F2XM1 calculates 2^st(0) - 1 for -1 < st(0) < 1 | 421 // F2XM1 calculates 2^st(0) - 1 for -1 < st(0) < 1 |
422 __ f2xm1(); // 2^(X-rnd(X)) - 1, rnd(X) | 422 __ f2xm1(); // 2^(X-rnd(X)) - 1, rnd(X) |
423 __ fld1(); // 1, 2^(X-rnd(X)) - 1, rnd(X) | 423 __ fld1(); // 1, 2^(X-rnd(X)) - 1, rnd(X) |
424 __ faddp(1); // 2^(X-rnd(X)), rnd(X) | 424 __ faddp(1); // 2^(X-rnd(X)), rnd(X) |
425 // FSCALE calculates st(0) * 2^st(1) | 425 // FSCALE calculates st(0) * 2^st(1) |
426 __ fscale(); // 2^X, rnd(X) | 426 __ fscale(); // 2^X, rnd(X) |
427 __ fstp(1); | 427 __ fstp(1); |
428 // Bail out to runtime in case of exceptions in the status word. | 428 // Bail out to runtime in case of exceptions in the status word. |
429 __ fnstsw_ax(); | 429 __ fnstsw_ax(); |
430 __ testb(rax, Immediate(0x5F)); // Check for all but precision exception. | 430 __ testb(rax, Immediate(0x5F)); // Check for all but precision exception. |
431 __ j(not_zero, &fast_power_failed, Label::kNear); | 431 __ j(not_zero, &fast_power_failed, Label::kNear); |
432 __ fstp_d(Operand(rsp, 0)); | 432 __ fstp_d(Operand(rsp, 0)); |
433 __ movsd(double_result, Operand(rsp, 0)); | 433 __ Movsd(double_result, Operand(rsp, 0)); |
434 __ addp(rsp, Immediate(kDoubleSize)); | 434 __ addp(rsp, Immediate(kDoubleSize)); |
435 __ jmp(&done); | 435 __ jmp(&done); |
436 | 436 |
437 __ bind(&fast_power_failed); | 437 __ bind(&fast_power_failed); |
438 __ fninit(); | 438 __ fninit(); |
439 __ addp(rsp, Immediate(kDoubleSize)); | 439 __ addp(rsp, Immediate(kDoubleSize)); |
440 __ jmp(&call_runtime); | 440 __ jmp(&call_runtime); |
441 } | 441 } |
442 | 442 |
443 // Calculate power with integer exponent. | 443 // Calculate power with integer exponent. |
444 __ bind(&int_exponent); | 444 __ bind(&int_exponent); |
445 const XMMRegister double_scratch2 = double_exponent; | 445 const XMMRegister double_scratch2 = double_exponent; |
446 // Back up exponent as we need to check if exponent is negative later. | 446 // Back up exponent as we need to check if exponent is negative later. |
447 __ movp(scratch, exponent); // Back up exponent. | 447 __ movp(scratch, exponent); // Back up exponent. |
448 __ movsd(double_scratch, double_base); // Back up base. | 448 __ Movsd(double_scratch, double_base); // Back up base. |
449 __ movsd(double_scratch2, double_result); // Load double_exponent with 1. | 449 __ Movsd(double_scratch2, double_result); // Load double_exponent with 1. |
450 | 450 |
451 // Get absolute value of exponent. | 451 // Get absolute value of exponent. |
452 Label no_neg, while_true, while_false; | 452 Label no_neg, while_true, while_false; |
453 __ testl(scratch, scratch); | 453 __ testl(scratch, scratch); |
454 __ j(positive, &no_neg, Label::kNear); | 454 __ j(positive, &no_neg, Label::kNear); |
455 __ negl(scratch); | 455 __ negl(scratch); |
456 __ bind(&no_neg); | 456 __ bind(&no_neg); |
457 | 457 |
458 __ j(zero, &while_false, Label::kNear); | 458 __ j(zero, &while_false, Label::kNear); |
459 __ shrl(scratch, Immediate(1)); | 459 __ shrl(scratch, Immediate(1)); |
460 // Above condition means CF==0 && ZF==0. This means that the | 460 // Above condition means CF==0 && ZF==0. This means that the |
461 // bit that has been shifted out is 0 and the result is not 0. | 461 // bit that has been shifted out is 0 and the result is not 0. |
462 __ j(above, &while_true, Label::kNear); | 462 __ j(above, &while_true, Label::kNear); |
463 __ movsd(double_result, double_scratch); | 463 __ Movsd(double_result, double_scratch); |
464 __ j(zero, &while_false, Label::kNear); | 464 __ j(zero, &while_false, Label::kNear); |
465 | 465 |
466 __ bind(&while_true); | 466 __ bind(&while_true); |
467 __ shrl(scratch, Immediate(1)); | 467 __ shrl(scratch, Immediate(1)); |
468 __ mulsd(double_scratch, double_scratch); | 468 __ mulsd(double_scratch, double_scratch); |
469 __ j(above, &while_true, Label::kNear); | 469 __ j(above, &while_true, Label::kNear); |
470 __ mulsd(double_result, double_scratch); | 470 __ mulsd(double_result, double_scratch); |
471 __ j(not_zero, &while_true); | 471 __ j(not_zero, &while_true); |
472 | 472 |
473 __ bind(&while_false); | 473 __ bind(&while_false); |
474 // If the exponent is negative, return 1/result. | 474 // If the exponent is negative, return 1/result. |
475 __ testl(exponent, exponent); | 475 __ testl(exponent, exponent); |
476 __ j(greater, &done); | 476 __ j(greater, &done); |
477 __ divsd(double_scratch2, double_result); | 477 __ divsd(double_scratch2, double_result); |
478 __ movsd(double_result, double_scratch2); | 478 __ Movsd(double_result, double_scratch2); |
479 // Test whether result is zero. Bail out to check for subnormal result. | 479 // Test whether result is zero. Bail out to check for subnormal result. |
480 // Due to subnormals, x^-y == (1/x)^y does not hold in all cases. | 480 // Due to subnormals, x^-y == (1/x)^y does not hold in all cases. |
481 __ xorps(double_scratch2, double_scratch2); | 481 __ xorps(double_scratch2, double_scratch2); |
482 __ ucomisd(double_scratch2, double_result); | 482 __ ucomisd(double_scratch2, double_result); |
483 // double_exponent aliased as double_scratch2 has already been overwritten | 483 // double_exponent aliased as double_scratch2 has already been overwritten |
484 // and may not have contained the exponent value in the first place when the | 484 // and may not have contained the exponent value in the first place when the |
485 // input was a smi. We reset it with exponent value before bailing out. | 485 // input was a smi. We reset it with exponent value before bailing out. |
486 __ j(not_equal, &done); | 486 __ j(not_equal, &done); |
487 __ Cvtlsi2sd(double_exponent, exponent); | 487 __ Cvtlsi2sd(double_exponent, exponent); |
488 | 488 |
489 // Returning or bailing out. | 489 // Returning or bailing out. |
490 Counters* counters = isolate()->counters(); | 490 Counters* counters = isolate()->counters(); |
491 if (exponent_type() == ON_STACK) { | 491 if (exponent_type() == ON_STACK) { |
492 // The arguments are still on the stack. | 492 // The arguments are still on the stack. |
493 __ bind(&call_runtime); | 493 __ bind(&call_runtime); |
494 __ TailCallRuntime(Runtime::kMathPowRT, 2, 1); | 494 __ TailCallRuntime(Runtime::kMathPowRT, 2, 1); |
495 | 495 |
496 // The stub is called from non-optimized code, which expects the result | 496 // The stub is called from non-optimized code, which expects the result |
497 // as heap number in rax. | 497 // as heap number in rax. |
498 __ bind(&done); | 498 __ bind(&done); |
499 __ AllocateHeapNumber(rax, rcx, &call_runtime); | 499 __ AllocateHeapNumber(rax, rcx, &call_runtime); |
500 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), double_result); | 500 __ Movsd(FieldOperand(rax, HeapNumber::kValueOffset), double_result); |
501 __ IncrementCounter(counters->math_pow(), 1); | 501 __ IncrementCounter(counters->math_pow(), 1); |
502 __ ret(2 * kPointerSize); | 502 __ ret(2 * kPointerSize); |
503 } else { | 503 } else { |
504 __ bind(&call_runtime); | 504 __ bind(&call_runtime); |
505 // Move base to the correct argument register. Exponent is already in xmm1. | 505 // Move base to the correct argument register. Exponent is already in xmm1. |
506 __ movsd(xmm0, double_base); | 506 __ Movsd(xmm0, double_base); |
507 DCHECK(double_exponent.is(xmm1)); | 507 DCHECK(double_exponent.is(xmm1)); |
508 { | 508 { |
509 AllowExternalCallThatCantCauseGC scope(masm); | 509 AllowExternalCallThatCantCauseGC scope(masm); |
510 __ PrepareCallCFunction(2); | 510 __ PrepareCallCFunction(2); |
511 __ CallCFunction( | 511 __ CallCFunction( |
512 ExternalReference::power_double_double_function(isolate()), 2); | 512 ExternalReference::power_double_double_function(isolate()), 2); |
513 } | 513 } |
514 // Return value is in xmm0. | 514 // Return value is in xmm0. |
515 __ movsd(double_result, xmm0); | 515 __ Movsd(double_result, xmm0); |
516 | 516 |
517 __ bind(&done); | 517 __ bind(&done); |
518 __ IncrementCounter(counters->math_pow(), 1); | 518 __ IncrementCounter(counters->math_pow(), 1); |
519 __ ret(0); | 519 __ ret(0); |
520 } | 520 } |
521 } | 521 } |
522 | 522 |
523 | 523 |
524 void FunctionPrototypeStub::Generate(MacroAssembler* masm) { | 524 void FunctionPrototypeStub::Generate(MacroAssembler* masm) { |
525 Label miss; | 525 Label miss; |
(...skipping 1032 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1558 } | 1558 } |
1559 __ Set(rax, EQUAL); | 1559 __ Set(rax, EQUAL); |
1560 __ ret(0); | 1560 __ ret(0); |
1561 | 1561 |
1562 __ bind(&heap_number); | 1562 __ bind(&heap_number); |
1563 // It is a heap number, so return equal if it's not NaN. | 1563 // It is a heap number, so return equal if it's not NaN. |
1564 // For NaN, return 1 for every condition except greater and | 1564 // For NaN, return 1 for every condition except greater and |
1565 // greater-equal. Return -1 for them, so the comparison yields | 1565 // greater-equal. Return -1 for them, so the comparison yields |
1566 // false for all conditions except not-equal. | 1566 // false for all conditions except not-equal. |
1567 __ Set(rax, EQUAL); | 1567 __ Set(rax, EQUAL); |
1568 __ movsd(xmm0, FieldOperand(rdx, HeapNumber::kValueOffset)); | 1568 __ Movsd(xmm0, FieldOperand(rdx, HeapNumber::kValueOffset)); |
1569 __ ucomisd(xmm0, xmm0); | 1569 __ ucomisd(xmm0, xmm0); |
1570 __ setcc(parity_even, rax); | 1570 __ setcc(parity_even, rax); |
1571 // rax is 0 for equal non-NaN heapnumbers, 1 for NaNs. | 1571 // rax is 0 for equal non-NaN heapnumbers, 1 for NaNs. |
1572 if (cc == greater_equal || cc == greater) { | 1572 if (cc == greater_equal || cc == greater) { |
1573 __ negp(rax); | 1573 __ negp(rax); |
1574 } | 1574 } |
1575 __ ret(0); | 1575 __ ret(0); |
1576 | 1576 |
1577 __ bind(¬_identical); | 1577 __ bind(¬_identical); |
1578 } | 1578 } |
(...skipping 1848 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3427 } | 3427 } |
3428 if (right() == CompareICState::SMI) { | 3428 if (right() == CompareICState::SMI) { |
3429 __ JumpIfNotSmi(rax, &miss); | 3429 __ JumpIfNotSmi(rax, &miss); |
3430 } | 3430 } |
3431 | 3431 |
3432 // Load left and right operand. | 3432 // Load left and right operand. |
3433 Label done, left, left_smi, right_smi; | 3433 Label done, left, left_smi, right_smi; |
3434 __ JumpIfSmi(rax, &right_smi, Label::kNear); | 3434 __ JumpIfSmi(rax, &right_smi, Label::kNear); |
3435 __ CompareMap(rax, isolate()->factory()->heap_number_map()); | 3435 __ CompareMap(rax, isolate()->factory()->heap_number_map()); |
3436 __ j(not_equal, &maybe_undefined1, Label::kNear); | 3436 __ j(not_equal, &maybe_undefined1, Label::kNear); |
3437 __ movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); | 3437 __ Movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); |
3438 __ jmp(&left, Label::kNear); | 3438 __ jmp(&left, Label::kNear); |
3439 __ bind(&right_smi); | 3439 __ bind(&right_smi); |
3440 __ SmiToInteger32(rcx, rax); // Can't clobber rax yet. | 3440 __ SmiToInteger32(rcx, rax); // Can't clobber rax yet. |
3441 __ Cvtlsi2sd(xmm1, rcx); | 3441 __ Cvtlsi2sd(xmm1, rcx); |
3442 | 3442 |
3443 __ bind(&left); | 3443 __ bind(&left); |
3444 __ JumpIfSmi(rdx, &left_smi, Label::kNear); | 3444 __ JumpIfSmi(rdx, &left_smi, Label::kNear); |
3445 __ CompareMap(rdx, isolate()->factory()->heap_number_map()); | 3445 __ CompareMap(rdx, isolate()->factory()->heap_number_map()); |
3446 __ j(not_equal, &maybe_undefined2, Label::kNear); | 3446 __ j(not_equal, &maybe_undefined2, Label::kNear); |
3447 __ movsd(xmm0, FieldOperand(rdx, HeapNumber::kValueOffset)); | 3447 __ Movsd(xmm0, FieldOperand(rdx, HeapNumber::kValueOffset)); |
3448 __ jmp(&done); | 3448 __ jmp(&done); |
3449 __ bind(&left_smi); | 3449 __ bind(&left_smi); |
3450 __ SmiToInteger32(rcx, rdx); // Can't clobber rdx yet. | 3450 __ SmiToInteger32(rcx, rdx); // Can't clobber rdx yet. |
3451 __ Cvtlsi2sd(xmm0, rcx); | 3451 __ Cvtlsi2sd(xmm0, rcx); |
3452 | 3452 |
3453 __ bind(&done); | 3453 __ bind(&done); |
3454 // Compare operands | 3454 // Compare operands |
3455 __ ucomisd(xmm0, xmm1); | 3455 __ ucomisd(xmm0, xmm1); |
3456 | 3456 |
3457 // Don't base result on EFLAGS when a NaN is involved. | 3457 // Don't base result on EFLAGS when a NaN is involved. |
(...skipping 2118 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5576 kStackSpace, nullptr, return_value_operand, NULL); | 5576 kStackSpace, nullptr, return_value_operand, NULL); |
5577 } | 5577 } |
5578 | 5578 |
5579 | 5579 |
5580 #undef __ | 5580 #undef __ |
5581 | 5581 |
5582 } // namespace internal | 5582 } // namespace internal |
5583 } // namespace v8 | 5583 } // namespace v8 |
5584 | 5584 |
5585 #endif // V8_TARGET_ARCH_X64 | 5585 #endif // V8_TARGET_ARCH_X64 |
OLD | NEW |