OLD | NEW |
1 // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file | 1 // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file |
2 // for details. All rights reserved. Use of this source code is governed by a | 2 // for details. All rights reserved. Use of this source code is governed by a |
3 // BSD-style license that can be found in the LICENSE file. | 3 // BSD-style license that can be found in the LICENSE file. |
4 | 4 |
5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_IA32. | 5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_IA32. |
6 #if defined(TARGET_ARCH_IA32) | 6 #if defined(TARGET_ARCH_IA32) |
7 | 7 |
8 #include "vm/flow_graph_compiler.h" | 8 #include "vm/flow_graph_compiler.h" |
9 | 9 |
10 #include "lib/error.h" | 10 #include "lib/error.h" |
(...skipping 1578 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1589 } | 1589 } |
1590 } else if (source.IsFpuRegister()) { | 1590 } else if (source.IsFpuRegister()) { |
1591 if (destination.IsFpuRegister()) { | 1591 if (destination.IsFpuRegister()) { |
1592 // Optimization manual recommends using MOVAPS for register | 1592 // Optimization manual recommends using MOVAPS for register |
1593 // to register moves. | 1593 // to register moves. |
1594 __ movaps(destination.fpu_reg(), source.fpu_reg()); | 1594 __ movaps(destination.fpu_reg(), source.fpu_reg()); |
1595 } else { | 1595 } else { |
1596 if (destination.IsDoubleStackSlot()) { | 1596 if (destination.IsDoubleStackSlot()) { |
1597 __ movsd(destination.ToStackSlotAddress(), source.fpu_reg()); | 1597 __ movsd(destination.ToStackSlotAddress(), source.fpu_reg()); |
1598 } else { | 1598 } else { |
1599 ASSERT(destination.IsFloat32x4StackSlot() || | 1599 ASSERT(destination.IsQuadStackSlot()); |
1600 destination.IsUint32x4StackSlot()); | |
1601 __ movups(destination.ToStackSlotAddress(), source.fpu_reg()); | 1600 __ movups(destination.ToStackSlotAddress(), source.fpu_reg()); |
1602 } | 1601 } |
1603 } | 1602 } |
1604 } else if (source.IsDoubleStackSlot()) { | 1603 } else if (source.IsDoubleStackSlot()) { |
1605 if (destination.IsFpuRegister()) { | 1604 if (destination.IsFpuRegister()) { |
1606 __ movsd(destination.fpu_reg(), source.ToStackSlotAddress()); | 1605 __ movsd(destination.fpu_reg(), source.ToStackSlotAddress()); |
1607 } else { | 1606 } else { |
1608 ASSERT(destination.IsDoubleStackSlot()); | 1607 ASSERT(destination.IsDoubleStackSlot()); |
1609 __ movsd(XMM0, source.ToStackSlotAddress()); | 1608 __ movsd(XMM0, source.ToStackSlotAddress()); |
1610 __ movsd(destination.ToStackSlotAddress(), XMM0); | 1609 __ movsd(destination.ToStackSlotAddress(), XMM0); |
1611 } | 1610 } |
1612 } else if (source.IsFloat32x4StackSlot() || source.IsUint32x4StackSlot()) { | 1611 } else if (source.IsQuadStackSlot()) { |
1613 if (destination.IsFpuRegister()) { | 1612 if (destination.IsFpuRegister()) { |
1614 __ movups(destination.fpu_reg(), source.ToStackSlotAddress()); | 1613 __ movups(destination.fpu_reg(), source.ToStackSlotAddress()); |
1615 } else { | 1614 } else { |
1616 ASSERT(destination.IsFloat32x4StackSlot() || | 1615 ASSERT(destination.IsQuadStackSlot()); |
1617 destination.IsUint32x4StackSlot()); | |
1618 __ movups(XMM0, source.ToStackSlotAddress()); | 1616 __ movups(XMM0, source.ToStackSlotAddress()); |
1619 __ movups(destination.ToStackSlotAddress(), XMM0); | 1617 __ movups(destination.ToStackSlotAddress(), XMM0); |
1620 } | 1618 } |
1621 } else { | 1619 } else { |
1622 ASSERT(source.IsConstant()); | 1620 ASSERT(source.IsConstant()); |
1623 if (destination.IsRegister()) { | 1621 if (destination.IsRegister()) { |
1624 const Object& constant = source.constant(); | 1622 const Object& constant = source.constant(); |
1625 if (constant.IsSmi() && (Smi::Cast(constant).Value() == 0)) { | 1623 if (constant.IsSmi() && (Smi::Cast(constant).Value() == 0)) { |
1626 __ xorl(destination.reg(), destination.reg()); | 1624 __ xorl(destination.reg(), destination.reg()); |
1627 } else { | 1625 } else { |
(...skipping 21 matching lines...) Expand all Loading... |
1649 } else if (source.IsStackSlot() && destination.IsRegister()) { | 1647 } else if (source.IsStackSlot() && destination.IsRegister()) { |
1650 Exchange(destination.reg(), source.ToStackSlotAddress()); | 1648 Exchange(destination.reg(), source.ToStackSlotAddress()); |
1651 } else if (source.IsStackSlot() && destination.IsStackSlot()) { | 1649 } else if (source.IsStackSlot() && destination.IsStackSlot()) { |
1652 Exchange(destination.ToStackSlotAddress(), source.ToStackSlotAddress()); | 1650 Exchange(destination.ToStackSlotAddress(), source.ToStackSlotAddress()); |
1653 } else if (source.IsFpuRegister() && destination.IsFpuRegister()) { | 1651 } else if (source.IsFpuRegister() && destination.IsFpuRegister()) { |
1654 __ movaps(XMM0, source.fpu_reg()); | 1652 __ movaps(XMM0, source.fpu_reg()); |
1655 __ movaps(source.fpu_reg(), destination.fpu_reg()); | 1653 __ movaps(source.fpu_reg(), destination.fpu_reg()); |
1656 __ movaps(destination.fpu_reg(), XMM0); | 1654 __ movaps(destination.fpu_reg(), XMM0); |
1657 } else if (source.IsFpuRegister() || destination.IsFpuRegister()) { | 1655 } else if (source.IsFpuRegister() || destination.IsFpuRegister()) { |
1658 ASSERT(destination.IsDoubleStackSlot() || | 1656 ASSERT(destination.IsDoubleStackSlot() || |
1659 destination.IsFloat32x4StackSlot() || | 1657 destination.IsQuadStackSlot() || |
1660 destination.IsUint32x4StackSlot() || | |
1661 source.IsDoubleStackSlot() || | 1658 source.IsDoubleStackSlot() || |
1662 source.IsFloat32x4StackSlot() || | 1659 source.IsQuadStackSlot()); |
1663 source.IsUint32x4StackSlot()); | |
1664 bool double_width = destination.IsDoubleStackSlot() || | 1660 bool double_width = destination.IsDoubleStackSlot() || |
1665 source.IsDoubleStackSlot(); | 1661 source.IsDoubleStackSlot(); |
1666 XmmRegister reg = source.IsFpuRegister() ? source.fpu_reg() | 1662 XmmRegister reg = source.IsFpuRegister() ? source.fpu_reg() |
1667 : destination.fpu_reg(); | 1663 : destination.fpu_reg(); |
1668 const Address& slot_address = source.IsFpuRegister() | 1664 const Address& slot_address = source.IsFpuRegister() |
1669 ? destination.ToStackSlotAddress() | 1665 ? destination.ToStackSlotAddress() |
1670 : source.ToStackSlotAddress(); | 1666 : source.ToStackSlotAddress(); |
1671 | 1667 |
1672 if (double_width) { | 1668 if (double_width) { |
1673 __ movsd(XMM0, slot_address); | 1669 __ movsd(XMM0, slot_address); |
1674 __ movsd(slot_address, reg); | 1670 __ movsd(slot_address, reg); |
1675 } else { | 1671 } else { |
1676 __ movups(XMM0, slot_address); | 1672 __ movups(XMM0, slot_address); |
1677 __ movups(slot_address, reg); | 1673 __ movups(slot_address, reg); |
1678 } | 1674 } |
1679 __ movaps(reg, XMM0); | 1675 __ movaps(reg, XMM0); |
| 1676 } else if (source.IsDoubleStackSlot() && destination.IsDoubleStackSlot()) { |
| 1677 const Address& source_slot_address = source.ToStackSlotAddress(); |
| 1678 const Address& destination_slot_address = destination.ToStackSlotAddress(); |
| 1679 |
| 1680 ScratchFpuRegisterScope ensure_scratch(this, XMM0); |
| 1681 __ movsd(XMM0, source_slot_address); |
| 1682 __ movsd(ensure_scratch.reg(), destination_slot_address); |
| 1683 __ movsd(destination_slot_address, XMM0); |
| 1684 __ movsd(source_slot_address, ensure_scratch.reg()); |
| 1685 } else if (source.IsQuadStackSlot() && destination.IsQuadStackSlot()) { |
| 1686 const Address& source_slot_address = source.ToStackSlotAddress(); |
| 1687 const Address& destination_slot_address = destination.ToStackSlotAddress(); |
| 1688 |
| 1689 ScratchFpuRegisterScope ensure_scratch(this, XMM0); |
| 1690 __ movups(XMM0, source_slot_address); |
| 1691 __ movups(ensure_scratch.reg(), destination_slot_address); |
| 1692 __ movups(destination_slot_address, XMM0); |
| 1693 __ movups(source_slot_address, ensure_scratch.reg()); |
1680 } else { | 1694 } else { |
1681 UNREACHABLE(); | 1695 UNREACHABLE(); |
1682 } | 1696 } |
1683 | 1697 |
1684 // The swap of source and destination has executed a move from source to | 1698 // The swap of source and destination has executed a move from source to |
1685 // destination. | 1699 // destination. |
1686 move->Eliminate(); | 1700 move->Eliminate(); |
1687 | 1701 |
1688 // Any unperformed (including pending) move with a source of either | 1702 // Any unperformed (including pending) move with a source of either |
1689 // this move's source or destination needs to have their source | 1703 // this move's source or destination needs to have their source |
1690 // changed to reflect the state of affairs after the swap. | 1704 // changed to reflect the state of affairs after the swap. |
1691 for (int i = 0; i < moves_.length(); ++i) { | 1705 for (int i = 0; i < moves_.length(); ++i) { |
1692 const MoveOperands& other_move = *moves_[i]; | 1706 const MoveOperands& other_move = *moves_[i]; |
1693 if (other_move.Blocks(source)) { | 1707 if (other_move.Blocks(source)) { |
1694 moves_[i]->set_src(destination); | 1708 moves_[i]->set_src(destination); |
1695 } else if (other_move.Blocks(destination)) { | 1709 } else if (other_move.Blocks(destination)) { |
1696 moves_[i]->set_src(source); | 1710 moves_[i]->set_src(source); |
1697 } | 1711 } |
1698 } | 1712 } |
1699 } | 1713 } |
1700 | 1714 |
1701 | 1715 |
1702 void ParallelMoveResolver::MoveMemoryToMemory(const Address& dst, | 1716 void ParallelMoveResolver::MoveMemoryToMemory(const Address& dst, |
1703 const Address& src) { | 1717 const Address& src) { |
1704 // TODO(vegorov): allocate temporary register for such moves. | 1718 ScratchRegisterScope ensure_scratch(this, kNoRegister); |
1705 __ pushl(EAX); | 1719 __ movl(ensure_scratch.reg(), src); |
1706 __ movl(EAX, src); | 1720 __ movl(dst, ensure_scratch.reg()); |
1707 __ movl(dst, EAX); | |
1708 __ popl(EAX); | |
1709 } | 1721 } |
1710 | 1722 |
1711 | 1723 |
1712 void ParallelMoveResolver::StoreObject(const Address& dst, const Object& obj) { | 1724 void ParallelMoveResolver::StoreObject(const Address& dst, const Object& obj) { |
1713 // TODO(vegorov): allocate temporary register for such moves. | |
1714 if (obj.IsSmi() || obj.IsNull()) { | 1725 if (obj.IsSmi() || obj.IsNull()) { |
1715 __ movl(dst, Immediate(reinterpret_cast<int32_t>(obj.raw()))); | 1726 __ movl(dst, Immediate(reinterpret_cast<int32_t>(obj.raw()))); |
1716 } else { | 1727 } else { |
1717 __ pushl(EAX); | 1728 ScratchRegisterScope ensure_scratch(this, kNoRegister); |
1718 __ LoadObject(EAX, obj); | 1729 __ LoadObject(ensure_scratch.reg(), obj); |
1719 __ movl(dst, EAX); | 1730 __ movl(dst, ensure_scratch.reg()); |
1720 __ popl(EAX); | |
1721 } | 1731 } |
1722 } | 1732 } |
1723 | 1733 |
1724 | 1734 |
1725 void ParallelMoveResolver::Exchange(Register reg, const Address& mem) { | 1735 void ParallelMoveResolver::Exchange(Register reg, const Address& mem) { |
1726 // TODO(vegorov): allocate temporary register for such moves. | 1736 ScratchRegisterScope ensure_scratch(this, reg); |
1727 Register scratch = (reg == EAX) ? ECX : EAX; | 1737 __ movl(ensure_scratch.reg(), mem); |
1728 __ pushl(scratch); | 1738 __ movl(mem, reg); |
1729 __ movl(scratch, mem); | 1739 __ movl(reg, ensure_scratch.reg()); |
1730 __ xchgl(scratch, reg); | |
1731 __ movl(mem, scratch); | |
1732 __ popl(scratch); | |
1733 } | 1740 } |
1734 | 1741 |
1735 | 1742 |
1736 void ParallelMoveResolver::Exchange(const Address& mem1, const Address& mem2) { | 1743 void ParallelMoveResolver::Exchange(const Address& mem1, const Address& mem2) { |
1737 // TODO(vegorov): allocate temporary registers for such moves. | 1744 ScratchRegisterScope ensure_scratch1(this, kNoRegister); |
1738 __ pushl(EAX); | 1745 ScratchRegisterScope ensure_scratch2(this, ensure_scratch1.reg()); |
1739 __ pushl(ECX); | 1746 __ movl(ensure_scratch1.reg(), mem1); |
1740 __ movl(EAX, mem1); | 1747 __ movl(ensure_scratch2.reg(), mem2); |
1741 __ movl(ECX, mem2); | 1748 __ movl(mem2, ensure_scratch1.reg()); |
1742 __ movl(mem1, ECX); | 1749 __ movl(mem1, ensure_scratch2.reg()); |
1743 __ movl(mem2, EAX); | |
1744 __ popl(ECX); | |
1745 __ popl(EAX); | |
1746 } | 1750 } |
1747 | 1751 |
1748 | 1752 |
| 1753 void ParallelMoveResolver::SpillScratch(Register reg) { |
| 1754 __ pushl(reg); |
| 1755 } |
| 1756 |
| 1757 |
| 1758 void ParallelMoveResolver::RestoreScratch(Register reg) { |
| 1759 __ popl(reg); |
| 1760 } |
| 1761 |
| 1762 |
| 1763 void ParallelMoveResolver::SpillFpuScratch(FpuRegister reg) { |
| 1764 __ subl(ESP, Immediate(kFpuRegisterSize)); |
| 1765 __ movups(Address(ESP, 0), reg); |
| 1766 } |
| 1767 |
| 1768 |
| 1769 void ParallelMoveResolver::RestoreFpuScratch(FpuRegister reg) { |
| 1770 __ movups(reg, Address(ESP, 0)); |
| 1771 __ addl(ESP, Immediate(kFpuRegisterSize)); |
| 1772 } |
| 1773 |
| 1774 |
1749 #undef __ | 1775 #undef __ |
1750 | 1776 |
1751 } // namespace dart | 1777 } // namespace dart |
1752 | 1778 |
1753 #endif // defined TARGET_ARCH_IA32 | 1779 #endif // defined TARGET_ARCH_IA32 |
OLD | NEW |