OLD | NEW |
---|---|
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 1488 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
1499 | 1499 |
1500 void TypeRecordingBinaryOpStub::GenerateRegisterArgsPush(MacroAssembler* masm) { | 1500 void TypeRecordingBinaryOpStub::GenerateRegisterArgsPush(MacroAssembler* masm) { |
1501 __ pop(rcx); | 1501 __ pop(rcx); |
1502 __ push(rdx); | 1502 __ push(rdx); |
1503 __ push(rax); | 1503 __ push(rax); |
1504 __ push(rcx); | 1504 __ push(rcx); |
1505 } | 1505 } |
1506 | 1506 |
1507 | 1507 |
1508 void TranscendentalCacheStub::Generate(MacroAssembler* masm) { | 1508 void TranscendentalCacheStub::Generate(MacroAssembler* masm) { |
1509 // Input on stack: | 1509 // TAGGED case: |
1510 // rsp[8]: argument (should be number). | 1510 // Input: |
1511 // rsp[0]: return address. | 1511 // rsp[8]: argument (should be number). |
1512 // rsp[0]: return address. | |
1513 // Output: | |
1514 // rax: tagged double result. | |
1515 // UNTAGGED case: | |
1516 // Input:: | |
1517 // rsp[0]: return address. | |
1518 // xmm1: untagged double input argument | |
1519 // Output: | |
1520 // xmm1: untagged double result. | |
1521 | |
1512 Label runtime_call; | 1522 Label runtime_call; |
1513 Label runtime_call_clear_stack; | 1523 Label runtime_call_clear_stack; |
1514 Label input_not_smi; | 1524 Label skip_cache; |
1515 NearLabel loaded; | 1525 const bool tagged = (argument_type_ == TAGGED); |
1516 // Test that rax is a number. | 1526 if (tagged) { |
1517 __ movq(rax, Operand(rsp, kPointerSize)); | 1527 NearLabel input_not_smi; |
1518 __ JumpIfNotSmi(rax, &input_not_smi); | 1528 NearLabel loaded; |
1519 // Input is a smi. Untag and load it onto the FPU stack. | 1529 // Test that rax is a number. |
1520 // Then load the bits of the double into rbx. | 1530 __ movq(rax, Operand(rsp, kPointerSize)); |
1521 __ SmiToInteger32(rax, rax); | 1531 __ JumpIfNotSmi(rax, &input_not_smi); |
1522 __ subq(rsp, Immediate(kPointerSize)); | 1532 // Input is a smi. Untag and load it onto the FPU stack. |
1523 __ cvtlsi2sd(xmm1, rax); | 1533 // Then load the bits of the double into rbx. |
1524 __ movsd(Operand(rsp, 0), xmm1); | 1534 __ SmiToInteger32(rax, rax); |
1525 __ movq(rbx, xmm1); | 1535 __ subq(rsp, Immediate(kPointerSize)); |
Lasse Reichstein
2011/02/25 06:49:26
Use kDoubleSize instead (id it exists, otherwise i
William Hesse
2011/02/25 12:12:38
Done.
| |
1526 __ movq(rdx, xmm1); | 1536 __ cvtlsi2sd(xmm1, rax); |
1527 __ fld_d(Operand(rsp, 0)); | 1537 __ movsd(Operand(rsp, 0), xmm1); |
1528 __ addq(rsp, Immediate(kPointerSize)); | 1538 __ movq(rbx, xmm1); |
1529 __ jmp(&loaded); | 1539 __ movq(rdx, xmm1); |
1540 __ fld_d(Operand(rsp, 0)); | |
1541 __ addq(rsp, Immediate(kPointerSize)); | |
1542 __ jmp(&loaded); | |
1530 | 1543 |
1531 __ bind(&input_not_smi); | 1544 __ bind(&input_not_smi); |
1532 // Check if input is a HeapNumber. | 1545 // Check if input is a HeapNumber. |
1533 __ Move(rbx, Factory::heap_number_map()); | 1546 __ Move(rbx, Factory::heap_number_map()); |
Lasse Reichstein
2011/02/25 06:49:26
Use LoadRoot.
William Hesse
2011/02/25 12:12:38
Done.
| |
1534 __ cmpq(rbx, FieldOperand(rax, HeapObject::kMapOffset)); | 1547 __ cmpq(rbx, FieldOperand(rax, HeapObject::kMapOffset)); |
1535 __ j(not_equal, &runtime_call); | 1548 __ j(not_equal, &runtime_call); |
1536 // Input is a HeapNumber. Push it on the FPU stack and load its | 1549 // Input is a HeapNumber. Push it on the FPU stack and load its |
1537 // bits into rbx. | 1550 // bits into rbx. |
1538 __ fld_d(FieldOperand(rax, HeapNumber::kValueOffset)); | 1551 __ fld_d(FieldOperand(rax, HeapNumber::kValueOffset)); |
1539 __ movq(rbx, FieldOperand(rax, HeapNumber::kValueOffset)); | 1552 __ movq(rbx, FieldOperand(rax, HeapNumber::kValueOffset)); |
1540 __ movq(rdx, rbx); | 1553 __ movq(rdx, rbx); |
1541 __ bind(&loaded); | 1554 |
1555 __ bind(&loaded); | |
1556 } else { // UNTAGGED. | |
1557 __ movq(rbx, xmm1); | |
1558 __ movq(rdx, rbx); | |
Lasse Reichstein
2011/02/25 06:49:26
Use movq(rdx,xmm1) to avoid dependencies.
William Hesse
2011/02/25 12:12:38
Done.
| |
1559 } | |
1560 | |
1542 // ST[0] == double value | 1561 // ST[0] == double value |
Lasse Reichstein
2011/02/25 06:49:26
The untagged case haven't pushed the value on the
William Hesse
2011/02/25 12:12:38
Added comment to this effect. All code paths corr
| |
1543 // rbx = bits of double value. | 1562 // rbx = bits of double value. |
1544 // rdx = also bits of double value. | 1563 // rdx = also bits of double value. |
1545 // Compute hash (h is 32 bits, bits are 64 and the shifts are arithmetic): | 1564 // Compute hash (h is 32 bits, bits are 64 and the shifts are arithmetic): |
1546 // h = h0 = bits ^ (bits >> 32); | 1565 // h = h0 = bits ^ (bits >> 32); |
1547 // h ^= h >> 16; | 1566 // h ^= h >> 16; |
1548 // h ^= h >> 8; | 1567 // h ^= h >> 8; |
1549 // h = h & (cacheSize - 1); | 1568 // h = h & (cacheSize - 1); |
1550 // or h = (h0 ^ (h0 >> 8) ^ (h0 >> 16) ^ (h0 >> 24)) & (cacheSize - 1) | 1569 // or h = (h0 ^ (h0 >> 8) ^ (h0 >> 16) ^ (h0 >> 24)) & (cacheSize - 1) |
1551 __ sar(rdx, Immediate(32)); | 1570 __ sar(rdx, Immediate(32)); |
1552 __ xorl(rdx, rbx); | 1571 __ xorl(rdx, rbx); |
(...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
1590 #endif | 1609 #endif |
1591 // Find the address of the rcx'th entry in the cache, i.e., &rax[rcx*16]. | 1610 // Find the address of the rcx'th entry in the cache, i.e., &rax[rcx*16]. |
1592 __ addl(rcx, rcx); | 1611 __ addl(rcx, rcx); |
1593 __ lea(rcx, Operand(rax, rcx, times_8, 0)); | 1612 __ lea(rcx, Operand(rax, rcx, times_8, 0)); |
1594 // Check if cache matches: Double value is stored in uint32_t[2] array. | 1613 // Check if cache matches: Double value is stored in uint32_t[2] array. |
1595 NearLabel cache_miss; | 1614 NearLabel cache_miss; |
1596 __ cmpq(rbx, Operand(rcx, 0)); | 1615 __ cmpq(rbx, Operand(rcx, 0)); |
1597 __ j(not_equal, &cache_miss); | 1616 __ j(not_equal, &cache_miss); |
1598 // Cache hit! | 1617 // Cache hit! |
1599 __ movq(rax, Operand(rcx, 2 * kIntSize)); | 1618 __ movq(rax, Operand(rcx, 2 * kIntSize)); |
1600 __ fstp(0); // Clear FPU stack. | 1619 if (tagged) { |
1601 __ ret(kPointerSize); | 1620 __ fstp(0); // Clear FPU stack. |
1621 __ ret(kPointerSize); | |
1622 } else { // UNTAGGED. | |
1623 __ movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); | |
1624 __ Ret(); | |
1625 } | |
1602 | 1626 |
1603 __ bind(&cache_miss); | 1627 __ bind(&cache_miss); |
1604 // Update cache with new value. | 1628 // Update cache with new value. |
1605 Label nan_result; | 1629 if (tagged) { |
1606 GenerateOperation(masm, &nan_result); | |
1607 __ AllocateHeapNumber(rax, rdi, &runtime_call_clear_stack); | 1630 __ AllocateHeapNumber(rax, rdi, &runtime_call_clear_stack); |
1631 } else { // UNTAGGED. | |
1632 __ AllocateHeapNumber(rax, rdi, &skip_cache); | |
1633 __ subq(rsp, Immediate(kDoubleSize)); | |
1634 __ movsd(Operand(rsp, 0), xmm1); | |
1635 __ fld_d(Operand(rsp, 0)); | |
1636 __ addq(rsp, Immediate(kDoubleSize)); | |
1637 } | |
1638 GenerateOperation(masm); | |
1608 __ movq(Operand(rcx, 0), rbx); | 1639 __ movq(Operand(rcx, 0), rbx); |
1609 __ movq(Operand(rcx, 2 * kIntSize), rax); | 1640 __ movq(Operand(rcx, 2 * kIntSize), rax); |
1610 __ fstp_d(FieldOperand(rax, HeapNumber::kValueOffset)); | 1641 __ fstp_d(FieldOperand(rax, HeapNumber::kValueOffset)); |
1611 __ ret(kPointerSize); | 1642 if (tagged) { |
1643 __ ret(kPointerSize); | |
1644 } else { // UNTAGGED. | |
1645 __ movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); | |
1646 __ Ret(); | |
1612 | 1647 |
1613 __ bind(&runtime_call_clear_stack); | 1648 // Skip cache and return answer directly, only in untagged case. |
1614 __ fstp(0); | 1649 __ bind(&skip_cache); |
1615 __ bind(&runtime_call); | 1650 __ subq(rsp, Immediate(kDoubleSize)); |
1616 __ TailCallExternalReference(ExternalReference(RuntimeFunction()), 1, 1); | 1651 __ movsd(Operand(rsp, 0), xmm1); |
1652 __ fld_d(Operand(rsp, 0)); | |
1653 GenerateOperation(masm); | |
1654 __ fstp_d(Operand(rsp, 0)); | |
1655 __ movsd(xmm1, Operand(rsp, 0)); | |
1656 __ addq(rsp, Immediate(kDoubleSize)); | |
1657 // We return the value in xmm1 without adding it to the cache, but | |
1658 // we cause a scavenging GC so that future allocations will succeed. | |
Lasse Reichstein
2011/02/25 06:49:26
Why don't we just allocate the size of a HeapNumbe
William Hesse
2011/02/25 12:12:38
Because this was added to the code after it was do
| |
1659 __ EnterInternalFrame(); | |
1660 // Allocate an unused object bigger than a HeapNumber. | |
1661 __ Push(Smi::FromInt(2 * kDoubleSize)); | |
1662 __ CallRuntimeSaveDoubles(Runtime::kAllocateInNewSpace); | |
1663 __ LeaveInternalFrame(); | |
1664 __ Ret(); | |
1665 } | |
1617 | 1666 |
1618 __ bind(&nan_result); | 1667 // Call runtime, doing whatever allocation and cleanup is necessary. |
1619 __ fstp(0); // Remove argument from FPU stack. | 1668 if (tagged) { |
1620 __ LoadRoot(rax, Heap::kNanValueRootIndex); | 1669 __ bind(&runtime_call_clear_stack); |
1621 __ movq(Operand(rcx, 0), rbx); | 1670 __ fstp(0); |
1622 __ movq(Operand(rcx, 2 * kIntSize), rax); | 1671 __ bind(&runtime_call); |
1623 __ ret(kPointerSize); | 1672 __ TailCallExternalReference(ExternalReference(RuntimeFunction()), 1, 1); |
1673 } else { // UNTAGGED. | |
1674 __ bind(&runtime_call_clear_stack); | |
1675 __ bind(&runtime_call); | |
1676 __ AllocateHeapNumber(rax, rdi, &skip_cache); | |
1677 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm1); | |
1678 __ EnterInternalFrame(); | |
1679 __ push(rax); | |
1680 __ CallRuntime(RuntimeFunction(), 1); | |
1681 __ LeaveInternalFrame(); | |
1682 __ movsd(xmm1, FieldOperand(rax, HeapNumber::kValueOffset)); | |
1683 __ Ret(); | |
1684 } | |
1624 } | 1685 } |
1625 | 1686 |
1626 | 1687 |
1627 Runtime::FunctionId TranscendentalCacheStub::RuntimeFunction() { | 1688 Runtime::FunctionId TranscendentalCacheStub::RuntimeFunction() { |
1628 switch (type_) { | 1689 switch (type_) { |
1629 // Add more cases when necessary. | 1690 // Add more cases when necessary. |
1630 case TranscendentalCache::SIN: return Runtime::kMath_sin; | 1691 case TranscendentalCache::SIN: return Runtime::kMath_sin; |
1631 case TranscendentalCache::COS: return Runtime::kMath_cos; | 1692 case TranscendentalCache::COS: return Runtime::kMath_cos; |
1632 case TranscendentalCache::LOG: return Runtime::kMath_log; | 1693 case TranscendentalCache::LOG: return Runtime::kMath_log; |
1633 default: | 1694 default: |
1634 UNIMPLEMENTED(); | 1695 UNIMPLEMENTED(); |
1635 return Runtime::kAbort; | 1696 return Runtime::kAbort; |
1636 } | 1697 } |
1637 } | 1698 } |
1638 | 1699 |
1639 | 1700 |
1640 void TranscendentalCacheStub::GenerateOperation(MacroAssembler* masm, | 1701 void TranscendentalCacheStub::GenerateOperation(MacroAssembler* masm) { |
1641 Label* on_nan_result) { | |
1642 // Registers: | 1702 // Registers: |
1643 // rbx: Bits of input double. Must be preserved. | 1703 // rbx: Bits of input double. Must be preserved. |
1644 // rcx: Pointer to cache entry. Must be preserved. | 1704 // rcx: Pointer to cache entry. Must be preserved. |
1645 // st(0): Input double | 1705 // st(0): Input double |
1646 Label done; | 1706 Label done; |
1647 if (type_ == TranscendentalCache::SIN || type_ == TranscendentalCache::COS) { | 1707 if (type_ == TranscendentalCache::SIN || type_ == TranscendentalCache::COS) { |
1648 // Both fsin and fcos require arguments in the range +/-2^63 and | 1708 // Both fsin and fcos require arguments in the range +/-2^63 and |
1649 // return NaN for infinities and NaN. They can share all code except | 1709 // return NaN for infinities and NaN. They can share all code except |
1650 // the actual fsin/fcos operation. | 1710 // the actual fsin/fcos operation. |
1651 Label in_range; | 1711 Label in_range; |
1652 // If argument is outside the range -2^63..2^63, fsin/cos doesn't | 1712 // If argument is outside the range -2^63..2^63, fsin/cos doesn't |
1653 // work. We must reduce it to the appropriate range. | 1713 // work. We must reduce it to the appropriate range. |
1654 __ movq(rdi, rbx); | 1714 __ movq(rdi, rbx); |
1655 // Move exponent and sign bits to low bits. | 1715 // Move exponent and sign bits to low bits. |
1656 __ shr(rdi, Immediate(HeapNumber::kMantissaBits)); | 1716 __ shr(rdi, Immediate(HeapNumber::kMantissaBits)); |
1657 // Remove sign bit. | 1717 // Remove sign bit. |
1658 __ andl(rdi, Immediate((1 << HeapNumber::kExponentBits) - 1)); | 1718 __ andl(rdi, Immediate((1 << HeapNumber::kExponentBits) - 1)); |
1659 int supported_exponent_limit = (63 + HeapNumber::kExponentBias); | 1719 int supported_exponent_limit = (63 + HeapNumber::kExponentBias); |
1660 __ cmpl(rdi, Immediate(supported_exponent_limit)); | 1720 __ cmpl(rdi, Immediate(supported_exponent_limit)); |
1661 __ j(below, &in_range); | 1721 __ j(below, &in_range); |
1662 // Check for infinity and NaN. Both return NaN for sin. | 1722 // Check for infinity and NaN. Both return NaN for sin. |
1663 __ cmpl(rdi, Immediate(0x7ff)); | 1723 __ cmpl(rdi, Immediate(0x7ff)); |
1664 __ j(equal, on_nan_result); | 1724 NearLabel non_nan_result; |
1725 __ j(not_equal, &non_nan_result); | |
1726 // Input is +/-Infinity or NaN. Result is NaN. | |
1727 __ fstp(0); | |
1728 __ LoadRoot(kScratchRegister, Heap::kNanValueRootIndex); | |
1729 __ push(FieldOperand(kScratchRegister, HeapNumber::kValueOffset)); | |
1730 __ fld_d(Operand(rsp, 0)); | |
Lasse Reichstein
2011/02/25 06:49:26
Can't you just fld_d(FieldOperand(kScratchRegister
William Hesse
2011/02/25 12:12:38
Done.
| |
1731 __ addq(rsp, Immediate(2 * kPointerSize)); | |
1732 __ jmp(&done); | |
1733 | |
1734 __ bind(&non_nan_result); | |
1665 | 1735 |
1666 // Use fpmod to restrict argument to the range +/-2*PI. | 1736 // Use fpmod to restrict argument to the range +/-2*PI. |
1667 __ fldpi(); | 1737 __ fldpi(); |
1668 __ fadd(0); | 1738 __ fadd(0); |
1669 __ fld(1); | 1739 __ fld(1); |
1670 // FPU Stack: input, 2*pi, input. | 1740 // FPU Stack: input, 2*pi, input. |
1671 { | 1741 { |
1672 Label no_exceptions; | 1742 Label no_exceptions; |
1673 __ fwait(); | 1743 __ fwait(); |
1674 __ fnstsw_ax(); | 1744 __ fnstsw_ax(); |
(...skipping 3191 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
4866 FieldOperand(elements, PixelArray::kExternalPointerOffset)); | 4936 FieldOperand(elements, PixelArray::kExternalPointerOffset)); |
4867 __ movb(Operand(external_pointer, untagged_key, times_1, 0), untagged_value); | 4937 __ movb(Operand(external_pointer, untagged_key, times_1, 0), untagged_value); |
4868 __ ret(0); // Return value in eax. | 4938 __ ret(0); // Return value in eax. |
4869 } | 4939 } |
4870 | 4940 |
4871 #undef __ | 4941 #undef __ |
4872 | 4942 |
4873 } } // namespace v8::internal | 4943 } } // namespace v8::internal |
4874 | 4944 |
4875 #endif // V8_TARGET_ARCH_X64 | 4945 #endif // V8_TARGET_ARCH_X64 |
OLD | NEW |