Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(718)

Unified Diff: unittest/AssemblerX8664/TestUtil.h

Issue 1224173006: Adds the x86-64 assembler. (Closed) Base URL: https://chromium.googlesource.com/native_client/pnacl-subzero.git@master
Patch Set: Addresses comments; make format Created 5 years, 5 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « unittest/AssemblerX8664/Other.cpp ('k') | unittest/AssemblerX8664/XmmArith.cpp » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: unittest/AssemblerX8664/TestUtil.h
diff --git a/unittest/AssemblerX8664/TestUtil.h b/unittest/AssemblerX8664/TestUtil.h
new file mode 100644
index 0000000000000000000000000000000000000000..4615459b59a130e26212b827c6e10c30424f9c55
--- /dev/null
+++ b/unittest/AssemblerX8664/TestUtil.h
@@ -0,0 +1,1097 @@
+
+//===- subzero/unittest/unittest/AssemblerX8664/TestUtil.h ------*- C++ -*-===//
+//
+// The Subzero Code Generator
+//
+// This file is distributed under the University of Illinois Open Source
+// License. See LICENSE.TXT for details.
+//
+//===----------------------------------------------------------------------===//
+//
+// Utility classes for testing the X8664 Assembler.
+//
+//===----------------------------------------------------------------------===//
+
+#ifndef ASSEMBLERX8664_TESTUTIL_H_
+#define ASSEMBLERX8664_TESTUTIL_H_
+
+#include "IceAssemblerX8664.h"
+
+#include "gtest/gtest.h"
+
+#include <cassert>
+#include <sys/mman.h>
+
+namespace Ice {
+namespace X8664 {
+namespace Test {
+
+class AssemblerX8664TestBase : public ::testing::Test {
+protected:
+ using Address = AssemblerX8664::Traits::Address;
+ using ByteRegister = AssemblerX8664::Traits::ByteRegister;
+ using Cond = AssemblerX8664::Traits::Cond;
+ using GPRRegister = AssemblerX8664::Traits::GPRRegister;
+ using Traits = AssemblerX8664::Traits;
+ using XmmRegister = AssemblerX8664::Traits::XmmRegister;
+
+// The following are "nicknames" for all possible GPRs in x86-64. With those, we
+// can use, e.g.,
+//
+// Encoded_GPR_al()
+//
+// instead of GPRRegister::Encoded_Reg_eax for 8 bit operands. They also
+// introduce "regular" nicknames for legacy x86-32 register (e.g., eax becomes
+// r1; esp, r0).
+#define LegacyRegAliases(NewName, Name64, Name32, Name16, Name8) \
+ static constexpr GPRRegister Encoded_GPR_##NewName() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##NewName##q() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##NewName##d() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##NewName##w() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##NewName##l() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name64() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name32() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name16() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name8() { \
+ return GPRRegister::Encoded_Reg_##Name32; \
+ }
+#define NewRegAliases(Name) \
+ static constexpr GPRRegister Encoded_GPR_##Name() { \
+ return GPRRegister::Encoded_Reg_##Name##d; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name##q() { \
+ return GPRRegister::Encoded_Reg_##Name##d; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name##d() { \
+ return GPRRegister::Encoded_Reg_##Name##d; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name##w() { \
+ return GPRRegister::Encoded_Reg_##Name##d; \
+ } \
+ static constexpr GPRRegister Encoded_GPR_##Name##l() { \
+ return GPRRegister::Encoded_Reg_##Name##d; \
+ }
+#define XmmRegAliases(Name) \
+ static constexpr XmmRegister Encoded_Xmm_##Name() { \
+ return XmmRegister::Encoded_Reg_##Name; \
+ }
+ LegacyRegAliases(r0, rsp, esp, sp, spl);
+ LegacyRegAliases(r1, rax, eax, ax, al);
+ LegacyRegAliases(r2, rbx, ebx, bx, bl);
+ LegacyRegAliases(r3, rcx, ecx, cx, cl);
+ LegacyRegAliases(r4, rdx, edx, dx, dl);
+ LegacyRegAliases(r5, rbp, ebp, bp, bpl);
+ LegacyRegAliases(r6, rsi, esi, si, sil);
+ LegacyRegAliases(r7, rdi, edi, di, dil);
+ NewRegAliases(r8);
+ NewRegAliases(r9);
+ NewRegAliases(r10);
+ NewRegAliases(r11);
+ NewRegAliases(r12);
+ NewRegAliases(r13);
+ NewRegAliases(r14);
+ NewRegAliases(r15);
+ XmmRegAliases(xmm0);
+ XmmRegAliases(xmm1);
+ XmmRegAliases(xmm2);
+ XmmRegAliases(xmm3);
+ XmmRegAliases(xmm4);
+ XmmRegAliases(xmm5);
+ XmmRegAliases(xmm6);
+ XmmRegAliases(xmm7);
+ XmmRegAliases(xmm8);
+ XmmRegAliases(xmm9);
+ XmmRegAliases(xmm10);
+ XmmRegAliases(xmm11);
+ XmmRegAliases(xmm12);
+ XmmRegAliases(xmm13);
+ XmmRegAliases(xmm14);
+ XmmRegAliases(xmm15);
+#undef XmmRegAliases
+#undef NewRegAliases
+#undef LegacyRegAliases
+
+ AssemblerX8664TestBase() { reset(); }
+
+ void reset() { Assembler.reset(new AssemblerX8664()); }
+
+ AssemblerX8664 *assembler() const { return Assembler.get(); }
+
+ size_t codeBytesSize() const { return Assembler->getBufferView().size(); }
+
+ const uint8_t *codeBytes() const {
+ return static_cast<const uint8_t *>(
+ static_cast<const void *>(Assembler->getBufferView().data()));
+ }
+
+private:
+ std::unique_ptr<AssemblerX8664> Assembler;
+};
+
+// __ is a helper macro. It allows test cases to emit X8664 assembly
+// instructions with
+//
+// __ mov(GPRRegister::Reg_Eax, 1);
+// __ ret();
+//
+// and so on. The idea of having this was "stolen" from dart's unit tests.
+#define __ (this->assembler())->
+
+// AssemblerX8664LowLevelTest verify that the "basic" instructions the tests
+// rely on are encoded correctly. Therefore, instead of executing the assembled
+// code, these tests will verify that the assembled bytes are sane.
+class AssemblerX8664LowLevelTest : public AssemblerX8664TestBase {
+protected:
+ // verifyBytes is a template helper that takes a Buffer, and a variable number
+ // of bytes. As the name indicates, it is used to verify the bytes for an
+ // instruction encoding.
+ template <int N, int I> static bool verifyBytes(const uint8_t *) {
+ static_assert(I == N, "Invalid template instantiation.");
+ return true;
+ }
+
+ template <int N, int I = 0, typename... Args>
+ static bool verifyBytes(const uint8_t *Buffer, uint8_t Byte,
+ Args... OtherBytes) {
+ static_assert(I < N, "Invalid template instantiation.");
+ EXPECT_EQ(Byte, Buffer[I]) << "Byte " << (I + 1) << " of " << N;
+ return verifyBytes<N, I + 1>(Buffer, OtherBytes...) && Buffer[I] == Byte;
+ }
+};
+
+// After these tests we should have a sane environment; we know the following
+// work:
+//
+// (*) zeroing eax, ebx, ecx, edx, edi, and esi;
+// (*) call $4 instruction (used for ip materialization);
+// (*) register push and pop;
+// (*) cmp reg, reg; and
+// (*) returning from functions.
+//
+// We can now dive into testing each emitting method in AssemblerX8664. Each
+// test will emit some instructions for performing the test. The assembled
+// instructions will operate in a "safe" environment. All x86-64 registers are
+// spilled to the program stack, and the registers are then zeroed out, with the
+// exception of %esp and %r9.
+//
+// The jitted code and the unittest code will share the same stack. Therefore,
+// test harnesses need to ensure it does not leave anything it pushed on the
+// stack.
+//
+// %r9 is initialized with a pointer for rIP-based addressing. This pointer is
+// used for position-independent access to a scratchpad area for use in tests.
+// In theory we could use rip-based addressing, but in practice that would
+// require creating fixups, which would, in turn, require creating a global
+// context. We therefore rely on the same technique used for pic code in x86-32
+// (i.e., IP materialization). Upon a test start up, a call(NextInstruction) is
+// executed. We then pop the return address from the stack, and use it for pic
+// addressing.
+//
+// The jitted code will look like the following:
+//
+// test:
+// push %r9
+// call test$materialize_ip
+// test$materialize_ip: <<------- %r9 will point here
+// pop %r9
+// push %rax
+// push %rbx
+// push %rcx
+// push %rdx
+// push %rbp
+// push %rdi
+// push %rsi
+// push %r8
+// push %r10
+// push %r11
+// push %r12
+// push %r13
+// push %r14
+// push %r15
+// mov $0, %rax
+// mov $0, %rbx
+// mov $0, %rcx
+// mov $0, %rdx
+// mov $0, %rbp
+// mov $0, %rdi
+// mov $0, %rsi
+// mov $0, %r8
+// mov $0, %r10
+// mov $0, %r11
+// mov $0, %r12
+// mov $0, %r13
+// mov $0, %r14
+// mov $0, %r15
+//
+// << test code goes here >>
+//
+// mov %rax, { 0 + $ScratchpadOffset}(%rbp)
+// mov %rbx, { 8 + $ScratchpadOffset}(%rbp)
+// mov %rcx, { 16 + $ScratchpadOffset}(%rbp)
+// mov %rdx, { 24 + $ScratchpadOffset}(%rbp)
+// mov %rdi, { 32 + $ScratchpadOffset}(%rbp)
+// mov %rsi, { 40 + $ScratchpadOffset}(%rbp)
+// mov %rbp, { 48 + $ScratchpadOffset}(%rbp)
+// mov %rsp, { 56 + $ScratchpadOffset}(%rbp)
+// mov %r8, { 64 + $ScratchpadOffset}(%rbp)
+// mov %r9, { 72 + $ScratchpadOffset}(%rbp)
+// mov %r10, { 80 + $ScratchpadOffset}(%rbp)
+// mov %r11, { 88 + $ScratchpadOffset}(%rbp)
+// mov %r12, { 96 + $ScratchpadOffset}(%rbp)
+// mov %r13, {104 + $ScratchpadOffset}(%rbp)
+// mov %r14, {112 + $ScratchpadOffset}(%rbp)
+// mov %r15, {120 + $ScratchpadOffset}(%rbp)
+// movups %xmm0, {128 + $ScratchpadOffset}(%rbp)
+// movups %xmm1, {136 + $ScratchpadOffset}(%rbp)
+// movups %xmm2, {144 + $ScratchpadOffset}(%rbp)
+// movups %xmm3, {152 + $ScratchpadOffset}(%rbp)
+// movups %xmm4, {160 + $ScratchpadOffset}(%rbp)
+// movups %xmm5, {168 + $ScratchpadOffset}(%rbp)
+// movups %xmm6, {176 + $ScratchpadOffset}(%rbp)
+// movups %xmm7, {184 + $ScratchpadOffset}(%rbp)
+// movups %xmm8, {192 + $ScratchpadOffset}(%rbp)
+// movups %xmm9, {200 + $ScratchpadOffset}(%rbp)
+// movups %xmm10, {208 + $ScratchpadOffset}(%rbp)
+// movups %xmm11, {216 + $ScratchpadOffset}(%rbp)
+// movups %xmm12, {224 + $ScratchpadOffset}(%rbp)
+// movups %xmm13, {232 + $ScratchpadOffset}(%rbp)
+// movups %xmm14, {240 + $ScratchpadOffset}(%rbp)
+// movups %xmm15, {248 + $ScratchpadOffset}(%rbp)
+//
+// pop %r15
+// pop %r14
+// pop %r13
+// pop %r12
+// pop %r11
+// pop %r10
+// pop %r8
+// pop %rsi
+// pop %rdi
+// pop %rbp
+// pop %rdx
+// pop %rcx
+// pop %rbx
+// pop %rax
+// pop %r9
+// ret
+//
+// << ... >>
+//
+// scratchpad: <<------- accessed via $Offset(%ebp)
+//
+// << test scratch area >>
+//
+// TODO(jpp): test the
+//
+// mov %reg, $Offset(%ebp)
+// movups %xmm, $Offset(%ebp)
+//
+// encodings using the low level assembler test ensuring that the register
+// values can be written to the scratchpad area.
+//
+// r9 was deliberately choosen so that every instruction accessing memory would
+// fail if the rex prefix was not emitted for it.
+class AssemblerX8664Test : public AssemblerX8664TestBase {
+protected:
+ // Dqword is used to represent 128-bit data types. The Dqword's contents are
+ // the same as the contents read from memory. Tests can then use the union
+ // members to verify the tests' outputs.
+ //
+ // NOTE: We want sizeof(Dqword) == sizeof(uint64_t) * 2. In other words, we
+ // want Dqword's contents to be **exactly** what the memory contents were so
+ // that we can do, e.g.,
+ //
+ // ...
+ // float Ret[4];
+ // // populate Ret
+ // return *reinterpret_cast<Dqword *>(&Ret);
+ //
+ // While being an ugly hack, this kind of return statements are used
+ // extensively in the PackedArith (see below) class.
+ union Dqword {
+ template <typename T0, typename T1, typename T2, typename T3,
+ typename = typename std::enable_if<
+ std::is_floating_point<T0>::value>::type>
+ Dqword(T0 F0, T1 F1, T2 F2, T3 F3) {
+ F32[0] = F0;
+ F32[1] = F1;
+ F32[2] = F2;
+ F32[3] = F3;
+ }
+
+ template <typename T>
+ Dqword(typename std::enable_if<std::is_same<T, int32_t>::value, T>::type I0,
+ T I1, T I2, T I3) {
+ I32[0] = I0;
+ I32[1] = I1;
+ I32[2] = I2;
+ I32[3] = I3;
+ }
+
+ template <typename T>
+ Dqword(typename std::enable_if<std::is_same<T, uint64_t>::value, T>::type
+ U64_0,
+ T U64_1) {
+ U64[0] = U64_0;
+ U64[1] = U64_1;
+ }
+
+ template <typename T>
+ Dqword(typename std::enable_if<std::is_same<T, double>::value, T>::type D0,
+ T D1) {
+ F64[0] = D0;
+ F64[1] = D1;
+ }
+
+ bool operator==(const Dqword &Rhs) const {
+ return std::memcmp(this, &Rhs, sizeof(*this)) == 0;
+ }
+
+ double F64[2];
+ uint64_t U64[2];
+ int64_t I64[2];
+
+ float F32[4];
+ uint32_t U32[4];
+ int32_t I32[4];
+
+ uint16_t U16[8];
+ int16_t I16[8];
+
+ uint8_t U8[16];
+ int8_t I8[16];
+
+ private:
+ Dqword() = delete;
+ };
+
+ // As stated, we want this condition to hold, so we assert.
+ static_assert(sizeof(Dqword) == 2 * sizeof(uint64_t),
+ "Dqword has the wrong size.");
+
+ // PackedArith is an interface provider for Dqwords. PackedArith's C argument
+ // is the undelying Dqword's type, which is then used so that we can define
+ // operators in terms of C++ operators on the underlying elements' type.
+ template <typename C> class PackedArith {
+ public:
+ static constexpr uint32_t N = sizeof(Dqword) / sizeof(C);
+ static_assert(N * sizeof(C) == sizeof(Dqword),
+ "Invalid template paramenter.");
+ static_assert((N & 1) == 0, "N should be divisible by 2");
+
+#define DefinePackedComparisonOperator(Op) \
+ template <typename Container = C, int Size = N> \
+ typename std::enable_if<std::is_floating_point<Container>::value, \
+ Dqword>::type \
+ operator Op(const Dqword &Rhs) const { \
+ using ElemType = \
+ typename std::conditional<std::is_same<float, Container>::value, \
+ int32_t, int64_t>::type; \
+ static_assert(sizeof(ElemType) == sizeof(Container), \
+ "Check ElemType definition."); \
+ const ElemType *const RhsPtr = \
+ reinterpret_cast<const ElemType *const>(&Rhs); \
+ const ElemType *const LhsPtr = \
+ reinterpret_cast<const ElemType *const>(&Lhs); \
+ ElemType Ret[N]; \
+ for (uint32_t i = 0; i < N; ++i) { \
+ Ret[i] = (LhsPtr[i] Op RhsPtr[i]) ? -1 : 0; \
+ } \
+ return *reinterpret_cast<Dqword *>(&Ret); \
+ }
+
+ DefinePackedComparisonOperator(< );
+ DefinePackedComparisonOperator(<= );
+ DefinePackedComparisonOperator(> );
+ DefinePackedComparisonOperator(>= );
+ DefinePackedComparisonOperator(== );
+ DefinePackedComparisonOperator(!= );
+
+#undef DefinePackedComparisonOperator
+
+#define DefinePackedOrdUnordComparisonOperator(Op, Ordered) \
+ template <typename Container = C, int Size = N> \
+ typename std::enable_if<std::is_floating_point<Container>::value, \
+ Dqword>::type \
+ Op(const Dqword &Rhs) const { \
+ using ElemType = \
+ typename std::conditional<std::is_same<float, Container>::value, \
+ int32_t, int64_t>::type; \
+ static_assert(sizeof(ElemType) == sizeof(Container), \
+ "Check ElemType definition."); \
+ const Container *const RhsPtr = \
+ reinterpret_cast<const Container *const>(&Rhs); \
+ const Container *const LhsPtr = \
+ reinterpret_cast<const Container *const>(&Lhs); \
+ ElemType Ret[N]; \
+ for (uint32_t i = 0; i < N; ++i) { \
+ Ret[i] = (!(LhsPtr[i] == LhsPtr[i]) || !(RhsPtr[i] == RhsPtr[i])) != \
+ (Ordered) \
+ ? -1 \
+ : 0; \
+ } \
+ return *reinterpret_cast<Dqword *>(&Ret); \
+ }
+
+ DefinePackedOrdUnordComparisonOperator(ord, true);
+ DefinePackedOrdUnordComparisonOperator(unord, false);
+#undef DefinePackedOrdUnordComparisonOperator
+
+#define DefinePackedArithOperator(Op, RhsIndexChanges, NeedsInt) \
+ template <typename Container = C, int Size = N> \
+ Dqword operator Op(const Dqword &Rhs) const { \
+ using ElemTypeForFp = typename std::conditional< \
+ !(NeedsInt), Container, \
+ typename std::conditional< \
+ std::is_same<Container, float>::value, uint32_t, \
+ typename std::conditional<std::is_same<Container, double>::value, \
+ uint64_t, void>::type>::type>::type; \
+ using ElemType = \
+ typename std::conditional<std::is_integral<Container>::value, \
+ Container, ElemTypeForFp>::type; \
+ static_assert(!std::is_same<void, ElemType>::value, \
+ "Check ElemType definition."); \
+ const ElemType *const RhsPtr = \
+ reinterpret_cast<const ElemType *const>(&Rhs); \
+ const ElemType *const LhsPtr = \
+ reinterpret_cast<const ElemType *const>(&Lhs); \
+ ElemType Ret[N]; \
+ for (uint32_t i = 0; i < N; ++i) { \
+ Ret[i] = LhsPtr[i] Op RhsPtr[(RhsIndexChanges) ? i : 0]; \
+ } \
+ return *reinterpret_cast<Dqword *>(&Ret); \
+ }
+
+ DefinePackedArithOperator(>>, false, true);
+ DefinePackedArithOperator(<<, false, true);
+ DefinePackedArithOperator(+, true, false);
+ DefinePackedArithOperator(-, true, false);
+ DefinePackedArithOperator(/, true, false);
+ DefinePackedArithOperator(&, true, true);
+ DefinePackedArithOperator(|, true, true);
+ DefinePackedArithOperator (^, true, true);
+
+#undef DefinePackedArithOperator
+
+#define DefinePackedArithShiftImm(Op) \
+ template <typename Container = C, int Size = N> \
+ Dqword operator Op(uint8_t imm) const { \
+ const Container *const LhsPtr = \
+ reinterpret_cast<const Container *const>(&Lhs); \
+ Container Ret[N]; \
+ for (uint32_t i = 0; i < N; ++i) { \
+ Ret[i] = LhsPtr[i] Op imm; \
+ } \
+ return *reinterpret_cast<Dqword *>(&Ret); \
+ }
+
+ DefinePackedArithShiftImm(>> );
+ DefinePackedArithShiftImm(<< );
+
+#undef DefinePackedArithShiftImm
+
+ template <typename Container = C, int Size = N>
+ typename std::enable_if<std::is_signed<Container>::value ||
+ std::is_floating_point<Container>::value,
+ Dqword>::type
+ operator*(const Dqword &Rhs) const {
+ static_assert((std::is_integral<Container>::value &&
+ sizeof(Container) < sizeof(uint64_t)) ||
+ std::is_floating_point<Container>::value,
+ "* is only defined for i(8|16|32), and fp types.");
+
+ const Container *const RhsPtr =
+ reinterpret_cast<const Container *const>(&Rhs);
+ const Container *const LhsPtr =
+ reinterpret_cast<const Container *const>(&Lhs);
+ Container Ret[Size];
+ for (uint32_t i = 0; i < Size; ++i) {
+ Ret[i] = LhsPtr[i] * RhsPtr[i];
+ }
+ return *reinterpret_cast<Dqword *>(&Ret);
+ }
+
+ template <typename Container = C, int Size = N,
+ typename = typename std::enable_if<
+ !std::is_signed<Container>::value>::type>
+ Dqword operator*(const Dqword &Rhs) const {
+ static_assert(std::is_integral<Container>::value &&
+ sizeof(Container) < sizeof(uint64_t),
+ "* is only defined for ui(8|16|32)");
+ using NextType = typename std::conditional<
+ sizeof(Container) == 1, uint16_t,
+ typename std::conditional<sizeof(Container) == 2, uint32_t,
+ uint64_t>::type>::type;
+ static_assert(sizeof(Container) * 2 == sizeof(NextType),
+ "Unexpected size");
+
+ const Container *const RhsPtr =
+ reinterpret_cast<const Container *const>(&Rhs);
+ const Container *const LhsPtr =
+ reinterpret_cast<const Container *const>(&Lhs);
+ NextType Ret[Size / 2];
+ for (uint32_t i = 0; i < Size; i += 2) {
+ Ret[i / 2] =
+ static_cast<NextType>(LhsPtr[i]) * static_cast<NextType>(RhsPtr[i]);
+ }
+ return *reinterpret_cast<Dqword *>(&Ret);
+ }
+
+ template <typename Container = C, int Size = N>
+ PackedArith<Container> operator~() const {
+ const Container *const LhsPtr =
+ reinterpret_cast<const Container *const>(&Lhs);
+ Container Ret[Size];
+ for (uint32_t i = 0; i < Size; ++i) {
+ Ret[i] = ~LhsPtr[i];
+ }
+ return PackedArith<Container>(*reinterpret_cast<Dqword *>(&Ret));
+ }
+
+#define MinMaxOperations(Name, Suffix) \
+ template <typename Container = C, int Size = N> \
+ Dqword Name##Suffix(const Dqword &Rhs) const { \
+ static_assert(std::is_floating_point<Container>::value, \
+ #Name #Suffix "ps is only available for fp."); \
+ const Container *const RhsPtr = \
+ reinterpret_cast<const Container *const>(&Rhs); \
+ const Container *const LhsPtr = \
+ reinterpret_cast<const Container *const>(&Lhs); \
+ Container Ret[Size]; \
+ for (uint32_t i = 0; i < Size; ++i) { \
+ Ret[i] = std::Name(LhsPtr[i], RhsPtr[i]); \
+ } \
+ return *reinterpret_cast<Dqword *>(&Ret); \
+ }
+
+ MinMaxOperations(max, ps);
+ MinMaxOperations(max, pd);
+ MinMaxOperations(min, ps);
+ MinMaxOperations(min, pd);
+#undef MinMaxOperations
+
+ template <typename Container = C, int Size = N>
+ Dqword blendWith(const Dqword &Rhs, const Dqword &Mask) const {
+ using MaskType = typename std::conditional<
+ sizeof(Container) == 1, int8_t,
+ typename std::conditional<sizeof(Container) == 2, int16_t,
+ int32_t>::type>::type;
+ static_assert(sizeof(MaskType) == sizeof(Container),
+ "MaskType has the wrong size.");
+ const Container *const RhsPtr =
+ reinterpret_cast<const Container *const>(&Rhs);
+ const Container *const LhsPtr =
+ reinterpret_cast<const Container *const>(&Lhs);
+ const MaskType *const MaskPtr =
+ reinterpret_cast<const MaskType *const>(&Mask);
+ Container Ret[Size];
+ for (int i = 0; i < Size; ++i) {
+ Ret[i] = ((MaskPtr[i] < 0) ? RhsPtr : LhsPtr)[i];
+ }
+ return *reinterpret_cast<Dqword *>(&Ret);
+ }
+
+ private:
+ // The AssemblerX8664Test class needs to be a friend so that it can create
+ // PackedArith objects (see below.)
+ friend class AssemblerX8664Test;
+
+ explicit PackedArith(const Dqword &MyLhs) : Lhs(MyLhs) {}
+
+ // Lhs can't be a & because operator~ returns a temporary object that needs
+ // access to its own Dqword.
+ const Dqword Lhs;
+ };
+
+ // Named constructor for PackedArith objects.
+ template <typename C> static PackedArith<C> packedAs(const Dqword &D) {
+ return PackedArith<C>(D);
+ }
+
+ AssemblerX8664Test() { reset(); }
+
+ void reset() {
+ AssemblerX8664TestBase::reset();
+
+ NeedsEpilogue = true;
+ // These dwords are allocated for saving the GPR state after the jitted code
+ // runs.
+ NumAllocatedDwords = AssembledTest::ScratchpadSlots;
+ addPrologue();
+ }
+
+ // AssembledTest is a wrapper around a PROT_EXEC mmap'ed buffer. This buffer
+ // contains both the test code as well as prologue/epilogue, and the
+ // scratchpad area that tests may use -- all tests use this scratchpad area
+ // for storing the processor's registers after the tests executed. This class
+ // also exposes helper methods for reading the register state after test
+ // execution, as well as for reading the scratchpad area.
+ class AssembledTest {
+ AssembledTest() = delete;
+ AssembledTest(const AssembledTest &) = delete;
+ AssembledTest &operator=(const AssembledTest &) = delete;
+
+ public:
+ static constexpr uint32_t MaximumCodeSize = 1 << 20;
+ static constexpr uint32_t raxSlot() { return 0; }
+ static constexpr uint32_t rbxSlot() { return 2; }
+ static constexpr uint32_t rcxSlot() { return 4; }
+ static constexpr uint32_t rdxSlot() { return 6; }
+ static constexpr uint32_t rdiSlot() { return 8; }
+ static constexpr uint32_t rsiSlot() { return 10; }
+ static constexpr uint32_t rbpSlot() { return 12; }
+ static constexpr uint32_t rspSlot() { return 14; }
+ static constexpr uint32_t r8Slot() { return 16; }
+ static constexpr uint32_t r9Slot() { return 18; }
+ static constexpr uint32_t r10Slot() { return 20; }
+ static constexpr uint32_t r11Slot() { return 22; }
+ static constexpr uint32_t r12Slot() { return 24; }
+ static constexpr uint32_t r13Slot() { return 26; }
+ static constexpr uint32_t r14Slot() { return 28; }
+ static constexpr uint32_t r15Slot() { return 30; }
+
+ // save 4 dwords for each xmm registers.
+ static constexpr uint32_t xmm0Slot() { return 32; }
+ static constexpr uint32_t xmm1Slot() { return 36; }
+ static constexpr uint32_t xmm2Slot() { return 40; }
+ static constexpr uint32_t xmm3Slot() { return 44; }
+ static constexpr uint32_t xmm4Slot() { return 48; }
+ static constexpr uint32_t xmm5Slot() { return 52; }
+ static constexpr uint32_t xmm6Slot() { return 56; }
+ static constexpr uint32_t xmm7Slot() { return 60; }
+ static constexpr uint32_t xmm8Slot() { return 64; }
+ static constexpr uint32_t xmm9Slot() { return 68; }
+ static constexpr uint32_t xmm10Slot() { return 72; }
+ static constexpr uint32_t xmm11Slot() { return 76; }
+ static constexpr uint32_t xmm12Slot() { return 80; }
+ static constexpr uint32_t xmm13Slot() { return 84; }
+ static constexpr uint32_t xmm14Slot() { return 88; }
+ static constexpr uint32_t xmm15Slot() { return 92; }
+
+ static constexpr uint32_t ScratchpadSlots = 96;
+
+ AssembledTest(const uint8_t *Data, const size_t MySize,
+ const size_t ExtraStorageDwords)
+ : Size(MaximumCodeSize + 4 * ExtraStorageDwords) {
+ // MaxCodeSize is needed because EXPECT_LT needs a symbol with a name --
+ // probably a compiler bug?
+ uint32_t MaxCodeSize = MaximumCodeSize;
+ EXPECT_LT(MySize, MaxCodeSize);
+ assert(MySize < MaximumCodeSize);
+ ExecutableData = mmap(nullptr, Size, PROT_WRITE | PROT_READ | PROT_EXEC,
+ MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
+ EXPECT_NE(MAP_FAILED, ExecutableData) << strerror(errno);
+ assert(MAP_FAILED != ExecutableData);
+ std::memcpy(ExecutableData, Data, MySize);
+ }
+
+ // We allow AssembledTest to be moved so that we can return objects of
+ // this type.
+ AssembledTest(AssembledTest &&Buffer)
+ : ExecutableData(Buffer.ExecutableData), Size(Buffer.Size) {
+ Buffer.ExecutableData = nullptr;
+ Buffer.Size = 0;
+ }
+
+ AssembledTest &operator=(AssembledTest &&Buffer) {
+ ExecutableData = Buffer.ExecutableData;
+ Buffer.ExecutableData = nullptr;
+ Size = Buffer.Size;
+ Buffer.Size = 0;
+ return *this;
+ }
+
+ ~AssembledTest() {
+ if (ExecutableData != nullptr) {
+ munmap(ExecutableData, Size);
+ ExecutableData = nullptr;
+ }
+ }
+
+ void run() const { reinterpret_cast<void (*)()>(ExecutableData)(); }
+
+#define LegacyRegAccessors(NewName, Name64, Name32, Name16, Name8) \
+ static_assert(Encoded_GPR_##NewName() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint64_t NewName() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##NewName##q() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint64_t NewName##q() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##NewName##d() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint32_t NewName##d() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##NewName##w() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint16_t NewName##w() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##NewName##l() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint8_t NewName##l() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##Name64() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint64_t Name64() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##Name32() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint32_t Name32() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##Name16() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint16_t Name16() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ } \
+ static_assert(Encoded_GPR_##Name8() == Encoded_GPR_##Name64(), \
+ "Invalid aliasing."); \
+ uint8_t Name8() const { \
+ return contentsOfQword(AssembledTest::Name64##Slot()); \
+ }
+#define NewRegAccessors(NewName) \
+ uint64_t NewName() const { \
+ return contentsOfQword(AssembledTest::NewName##Slot()); \
+ } \
+ uint64_t NewName##q() const { \
+ return contentsOfQword(AssembledTest::NewName##Slot()); \
+ } \
+ uint32_t NewName##d() const { \
+ return contentsOfQword(AssembledTest::NewName##Slot()); \
+ } \
+ uint16_t NewName##w() const { \
+ return contentsOfQword(AssembledTest::NewName##Slot()); \
+ } \
+ uint8_t NewName##l() const { \
+ return contentsOfQword(AssembledTest::NewName##Slot()); \
+ }
+#define XmmRegAccessor(Name) \
+ template <typename T> T Name() const { \
+ return xmm<T>(AssembledTest::Name##Slot()); \
+ }
+ LegacyRegAccessors(r0, rsp, esp, sp, spl);
+ LegacyRegAccessors(r1, rax, eax, ax, al);
+ LegacyRegAccessors(r2, rbx, ebx, bx, bl);
+ LegacyRegAccessors(r3, rcx, ecx, cx, cl);
+ LegacyRegAccessors(r4, rdx, edx, dx, dl);
+ LegacyRegAccessors(r5, rbp, ebp, bp, bpl);
+ LegacyRegAccessors(r6, rsi, esi, si, sil);
+ LegacyRegAccessors(r7, rdi, edi, di, dil);
+ NewRegAccessors(r8);
+ NewRegAccessors(r9);
+ NewRegAccessors(r10);
+ NewRegAccessors(r11);
+ NewRegAccessors(r12);
+ NewRegAccessors(r13);
+ NewRegAccessors(r14);
+ NewRegAccessors(r15);
+ XmmRegAccessor(xmm0);
+ XmmRegAccessor(xmm1);
+ XmmRegAccessor(xmm2);
+ XmmRegAccessor(xmm3);
+ XmmRegAccessor(xmm4);
+ XmmRegAccessor(xmm5);
+ XmmRegAccessor(xmm6);
+ XmmRegAccessor(xmm7);
+ XmmRegAccessor(xmm8);
+ XmmRegAccessor(xmm9);
+ XmmRegAccessor(xmm10);
+ XmmRegAccessor(xmm11);
+ XmmRegAccessor(xmm12);
+ XmmRegAccessor(xmm13);
+ XmmRegAccessor(xmm14);
+ XmmRegAccessor(xmm15);
+#undef XmmRegAccessor
+#undef NewRegAccessors
+#undef LegacyRegAccessors
+
+ // contentsOfDword is used for reading the values in the scratchpad area.
+ // Valid arguments are the dword ids returned by
+ // AssemblerX8664Test::allocateDword() -- other inputs are considered
+ // invalid, and are not guaranteed to work if the implementation changes.
+ template <typename T = uint32_t, typename = typename std::enable_if<
+ sizeof(T) == sizeof(uint32_t)>::type>
+ T contentsOfDword(uint32_t Dword) const {
+ return *reinterpret_cast<T *>(static_cast<uint8_t *>(ExecutableData) +
+ dwordOffset(Dword));
+ }
+
+ template <typename T = uint64_t, typename = typename std::enable_if<
+ sizeof(T) == sizeof(uint64_t)>::type>
+ T contentsOfQword(uint32_t InitialDword) const {
+ return *reinterpret_cast<T *>(static_cast<uint8_t *>(ExecutableData) +
+ dwordOffset(InitialDword));
+ }
+
+ Dqword contentsOfDqword(uint32_t InitialDword) const {
+ return *reinterpret_cast<Dqword *>(
+ static_cast<uint8_t *>(ExecutableData) +
+ dwordOffset(InitialDword));
+ }
+
+ template <typename T = uint32_t, typename = typename std::enable_if<
+ sizeof(T) == sizeof(uint32_t)>::type>
+ void setDwordTo(uint32_t Dword, T value) {
+ *reinterpret_cast<uint32_t *>(static_cast<uint8_t *>(ExecutableData) +
+ dwordOffset(Dword)) =
+ *reinterpret_cast<uint32_t *>(&value);
+ }
+
+ template <typename T = uint64_t, typename = typename std::enable_if<
+ sizeof(T) == sizeof(uint64_t)>::type>
+ void setQwordTo(uint32_t InitialDword, T value) {
+ *reinterpret_cast<uint64_t *>(static_cast<uint8_t *>(ExecutableData) +
+ dwordOffset(InitialDword)) =
+ *reinterpret_cast<uint64_t *>(&value);
+ }
+
+ void setDqwordTo(uint32_t InitialDword, const Dqword &qdword) {
+ setQwordTo(InitialDword, qdword.U64[0]);
+ setQwordTo(InitialDword + 2, qdword.U64[1]);
+ }
+
+ private:
+ template <typename T>
+ typename std::enable_if<std::is_same<T, Dqword>::value, Dqword>::type
+ xmm(uint8_t Slot) const {
+ return contentsOfDqword(Slot);
+ }
+
+ template <typename T>
+ typename std::enable_if<!std::is_same<T, Dqword>::value, T>::type
+ xmm(uint8_t Slot) const {
+ constexpr bool TIs64Bit = sizeof(T) == sizeof(uint64_t);
+ using _64BitType = typename std::conditional<TIs64Bit, T, uint64_t>::type;
+ using _32BitType = typename std::conditional<TIs64Bit, uint32_t, T>::type;
+ if (TIs64Bit) {
+ return contentsOfQword<_64BitType>(Slot);
+ }
+ return contentsOfDword<_32BitType>(Slot);
+ }
+
+ static uint32_t dwordOffset(uint32_t Index) {
+ return MaximumCodeSize + (Index * 4);
+ }
+
+ void *ExecutableData = nullptr;
+ size_t Size;
+ };
+
+ // assemble created an AssembledTest with the jitted code. The first time
+ // assemble is executed it will add the epilogue to the jitted code (which is
+ // the reason why this method is not const qualified.
+ AssembledTest assemble() {
+ if (NeedsEpilogue) {
+ addEpilogue();
+ }
+
+ NeedsEpilogue = false;
+ return AssembledTest(codeBytes(), codeBytesSize(), NumAllocatedDwords);
+ }
+
+ // Allocates a new dword slot in the test's scratchpad area.
+ uint32_t allocateDword() { return NumAllocatedDwords++; }
+
+ // Allocates a new qword slot in the test's scratchpad area.
+ uint32_t allocateQword() {
+ uint32_t InitialDword = allocateDword();
+ allocateDword();
+ return InitialDword;
+ }
+
+ // Allocates a new dqword slot in the test's scratchpad area.
+ uint32_t allocateDqword() {
+ uint32_t InitialDword = allocateQword();
+ allocateQword();
+ return InitialDword;
+ }
+
+ Address dwordAddress(uint32_t Dword) {
+ return Address(Encoded_GPR_r9(), dwordDisp(Dword));
+ }
+
+private:
+ // e??SlotAddress returns an AssemblerX8664::Traits::Address that can be used
+ // by the test cases to encode an address operand for accessing the slot for
+ // the specified register. These are all private for, when jitting the test
+ // code, tests should not tamper with these values. Besides, during the test
+ // execution these slots' contents are undefined and should not be accessed.
+ Address raxSlotAddress() { return dwordAddress(AssembledTest::raxSlot()); }
+ Address rbxSlotAddress() { return dwordAddress(AssembledTest::rbxSlot()); }
+ Address rcxSlotAddress() { return dwordAddress(AssembledTest::rcxSlot()); }
+ Address rdxSlotAddress() { return dwordAddress(AssembledTest::rdxSlot()); }
+ Address rdiSlotAddress() { return dwordAddress(AssembledTest::rdiSlot()); }
+ Address rsiSlotAddress() { return dwordAddress(AssembledTest::rsiSlot()); }
+ Address rbpSlotAddress() { return dwordAddress(AssembledTest::rbpSlot()); }
+ Address rspSlotAddress() { return dwordAddress(AssembledTest::rspSlot()); }
+ Address r8SlotAddress() { return dwordAddress(AssembledTest::r8Slot()); }
+ Address r9SlotAddress() { return dwordAddress(AssembledTest::r9Slot()); }
+ Address r10SlotAddress() { return dwordAddress(AssembledTest::r10Slot()); }
+ Address r11SlotAddress() { return dwordAddress(AssembledTest::r11Slot()); }
+ Address r12SlotAddress() { return dwordAddress(AssembledTest::r12Slot()); }
+ Address r13SlotAddress() { return dwordAddress(AssembledTest::r13Slot()); }
+ Address r14SlotAddress() { return dwordAddress(AssembledTest::r14Slot()); }
+ Address r15SlotAddress() { return dwordAddress(AssembledTest::r15Slot()); }
+ Address xmm0SlotAddress() { return dwordAddress(AssembledTest::xmm0Slot()); }
+ Address xmm1SlotAddress() { return dwordAddress(AssembledTest::xmm1Slot()); }
+ Address xmm2SlotAddress() { return dwordAddress(AssembledTest::xmm2Slot()); }
+ Address xmm3SlotAddress() { return dwordAddress(AssembledTest::xmm3Slot()); }
+ Address xmm4SlotAddress() { return dwordAddress(AssembledTest::xmm4Slot()); }
+ Address xmm5SlotAddress() { return dwordAddress(AssembledTest::xmm5Slot()); }
+ Address xmm6SlotAddress() { return dwordAddress(AssembledTest::xmm6Slot()); }
+ Address xmm7SlotAddress() { return dwordAddress(AssembledTest::xmm7Slot()); }
+ Address xmm8SlotAddress() { return dwordAddress(AssembledTest::xmm8Slot()); }
+ Address xmm9SlotAddress() { return dwordAddress(AssembledTest::xmm9Slot()); }
+ Address xmm10SlotAddress() {
+ return dwordAddress(AssembledTest::xmm10Slot());
+ }
+ Address xmm11SlotAddress() {
+ return dwordAddress(AssembledTest::xmm11Slot());
+ }
+ Address xmm12SlotAddress() {
+ return dwordAddress(AssembledTest::xmm12Slot());
+ }
+ Address xmm13SlotAddress() {
+ return dwordAddress(AssembledTest::xmm13Slot());
+ }
+ Address xmm14SlotAddress() {
+ return dwordAddress(AssembledTest::xmm14Slot());
+ }
+ Address xmm15SlotAddress() {
+ return dwordAddress(AssembledTest::xmm15Slot());
+ }
+
+ // Returns the displacement that should be used when accessing the specified
+ // Dword in the scratchpad area. It needs to adjust for the initial
+ // instructions that are emitted before the call that materializes the IP
+ // register.
+ uint32_t dwordDisp(uint32_t Dword) const {
+ EXPECT_LT(Dword, NumAllocatedDwords);
+ assert(Dword < NumAllocatedDwords);
+ static constexpr uint8_t PushR9Bytes = 2;
+ static constexpr uint8_t CallImmBytes = 5;
+ return AssembledTest::MaximumCodeSize + (Dword * 4) -
+ (PushR9Bytes + CallImmBytes);
+ }
+
+ void addPrologue() {
+ __ pushl(Encoded_GPR_r9());
+ __ call(Immediate(4));
+ __ popl(Encoded_GPR_r9());
+
+ __ pushl(Encoded_GPR_rax());
+ __ pushl(Encoded_GPR_rbx());
+ __ pushl(Encoded_GPR_rcx());
+ __ pushl(Encoded_GPR_rdx());
+ __ pushl(Encoded_GPR_rbp());
+ __ pushl(Encoded_GPR_rdi());
+ __ pushl(Encoded_GPR_rsi());
+ __ pushl(Encoded_GPR_r8());
+ __ pushl(Encoded_GPR_r10());
+ __ pushl(Encoded_GPR_r11());
+ __ pushl(Encoded_GPR_r12());
+ __ pushl(Encoded_GPR_r13());
+ __ pushl(Encoded_GPR_r14());
+ __ pushl(Encoded_GPR_r15());
+
+ __ mov(IceType_i32, Encoded_GPR_rax(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_rbx(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_rcx(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_rdx(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_rbp(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_rdi(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_rsi(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r8(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r10(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r11(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r12(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r13(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r14(), Immediate(0x00));
+ __ mov(IceType_i32, Encoded_GPR_r15(), Immediate(0x00));
+ }
+
+ void addEpilogue() {
+ __ mov(IceType_i64, raxSlotAddress(), Encoded_GPR_rax());
+ __ mov(IceType_i64, rbxSlotAddress(), Encoded_GPR_rbx());
+ __ mov(IceType_i64, rcxSlotAddress(), Encoded_GPR_rcx());
+ __ mov(IceType_i64, rdxSlotAddress(), Encoded_GPR_rdx());
+ __ mov(IceType_i64, rdiSlotAddress(), Encoded_GPR_rdi());
+ __ mov(IceType_i64, rsiSlotAddress(), Encoded_GPR_rsi());
+ __ mov(IceType_i64, rbpSlotAddress(), Encoded_GPR_rbp());
+ __ mov(IceType_i64, rspSlotAddress(), Encoded_GPR_rsp());
+ __ mov(IceType_i64, r8SlotAddress(), Encoded_GPR_r8());
+ __ mov(IceType_i64, r9SlotAddress(), Encoded_GPR_r9());
+ __ mov(IceType_i64, r10SlotAddress(), Encoded_GPR_r10());
+ __ mov(IceType_i64, r11SlotAddress(), Encoded_GPR_r11());
+ __ mov(IceType_i64, r12SlotAddress(), Encoded_GPR_r12());
+ __ mov(IceType_i64, r13SlotAddress(), Encoded_GPR_r13());
+ __ mov(IceType_i64, r14SlotAddress(), Encoded_GPR_r14());
+ __ mov(IceType_i64, r15SlotAddress(), Encoded_GPR_r15());
+ __ movups(xmm0SlotAddress(), Encoded_Xmm_xmm0());
+ __ movups(xmm1SlotAddress(), Encoded_Xmm_xmm1());
+ __ movups(xmm2SlotAddress(), Encoded_Xmm_xmm2());
+ __ movups(xmm3SlotAddress(), Encoded_Xmm_xmm3());
+ __ movups(xmm4SlotAddress(), Encoded_Xmm_xmm4());
+ __ movups(xmm5SlotAddress(), Encoded_Xmm_xmm5());
+ __ movups(xmm6SlotAddress(), Encoded_Xmm_xmm6());
+ __ movups(xmm7SlotAddress(), Encoded_Xmm_xmm7());
+ __ movups(xmm8SlotAddress(), Encoded_Xmm_xmm8());
+ __ movups(xmm9SlotAddress(), Encoded_Xmm_xmm9());
+ __ movups(xmm10SlotAddress(), Encoded_Xmm_xmm10());
+ __ movups(xmm11SlotAddress(), Encoded_Xmm_xmm11());
+ __ movups(xmm12SlotAddress(), Encoded_Xmm_xmm12());
+ __ movups(xmm13SlotAddress(), Encoded_Xmm_xmm13());
+ __ movups(xmm14SlotAddress(), Encoded_Xmm_xmm14());
+ __ movups(xmm15SlotAddress(), Encoded_Xmm_xmm15());
+
+ __ popl(Encoded_GPR_r15());
+ __ popl(Encoded_GPR_r14());
+ __ popl(Encoded_GPR_r13());
+ __ popl(Encoded_GPR_r12());
+ __ popl(Encoded_GPR_r11());
+ __ popl(Encoded_GPR_r10());
+ __ popl(Encoded_GPR_r8());
+ __ popl(Encoded_GPR_rsi());
+ __ popl(Encoded_GPR_rdi());
+ __ popl(Encoded_GPR_rbp());
+ __ popl(Encoded_GPR_rdx());
+ __ popl(Encoded_GPR_rcx());
+ __ popl(Encoded_GPR_rbx());
+ __ popl(Encoded_GPR_rax());
+ __ popl(Encoded_GPR_r9());
+
+ __ ret();
+ }
+
+ bool NeedsEpilogue;
+ uint32_t NumAllocatedDwords;
+};
+
+} // end of namespace Test
+} // end of namespace X8664
+} // end of namespace Ice
+
+#endif // ASSEMBLERX8664_TESTUTIL_H_
« no previous file with comments | « unittest/AssemblerX8664/Other.cpp ('k') | unittest/AssemblerX8664/XmmArith.cpp » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698