Index: test/cctest/test-macro-assembler-mips64.cc |
diff --git a/test/cctest/test-macro-assembler-mips64.cc b/test/cctest/test-macro-assembler-mips64.cc |
index e5f16e264ff3a8e8a0a8af4aaf520379046fb211..f0180c18e336acfd865c38339d9ca5f626152ed1 100644 |
--- a/test/cctest/test-macro-assembler-mips64.cc |
+++ b/test/cctest/test-macro-assembler-mips64.cc |
@@ -567,12 +567,21 @@ static const std::vector<int64_t> cvt_trunc_int64_test_values() { |
for (std::vector<ctype>::iterator var = var##_vec.begin(); \ |
var != var##_vec.end(); ++var) |
+#define FOR_INPUTS2(ctype, itype, var, var2, test_vector) \ |
+ std::vector<ctype> var##_vec = test_vector(); \ |
+ std::vector<ctype>::iterator var; \ |
+ std::vector<ctype>::reverse_iterator var2; \ |
+ for (var = var##_vec.begin(), var2 = var##_vec.rbegin(); \ |
+ var != var##_vec.end(); ++var, ++var2) |
+ |
#define FOR_ENUM_INPUTS(var, type, test_vector) \ |
FOR_INPUTS(enum type, type, var, test_vector) |
#define FOR_STRUCT_INPUTS(var, type, test_vector) \ |
FOR_INPUTS(struct type, type, var, test_vector) |
#define FOR_INT32_INPUTS(var, test_vector) \ |
FOR_INPUTS(int32_t, int32, var, test_vector) |
+#define FOR_INT32_INPUTS2(var, var2, test_vector) \ |
+ FOR_INPUTS2(int32_t, int32, var, var2, test_vector) |
#define FOR_INT64_INPUTS(var, test_vector) \ |
FOR_INPUTS(int64_t, int64, var, test_vector) |
#define FOR_UINT32_INPUTS(var, test_vector) \ |
@@ -1396,4 +1405,344 @@ TEST(min_max_nan) { |
} |
} |
+template <typename IN_TYPE, typename Func> |
+bool run_Unaligned(char* memory_buffer, int32_t in_offset, int32_t out_offset, |
+ IN_TYPE value, Func GenerateUnalignedInstructionFunc) { |
+ typedef int32_t (*F_CVT)(char* x0, int x1, int x2, int x3, int x4); |
+ |
+ Isolate* isolate = CcTest::i_isolate(); |
+ HandleScope scope(isolate); |
+ MacroAssembler assm(isolate, nullptr, 0, |
+ v8::internal::CodeObjectRequired::kYes); |
+ MacroAssembler* masm = &assm; |
+ IN_TYPE res; |
+ |
+ GenerateUnalignedInstructionFunc(masm, in_offset, out_offset); |
+ __ jr(ra); |
+ __ nop(); |
+ |
+ CodeDesc desc; |
+ assm.GetCode(&desc); |
+ Handle<Code> code = isolate->factory()->NewCode( |
+ desc, Code::ComputeFlags(Code::STUB), Handle<Code>()); |
+ |
+ F_CVT f = FUNCTION_CAST<F_CVT>(code->entry()); |
+ |
+ MemCopy(memory_buffer + in_offset, &value, sizeof(IN_TYPE)); |
+ CALL_GENERATED_CODE(isolate, f, memory_buffer, 0, 0, 0, 0); |
+ MemCopy(&res, memory_buffer + out_offset, sizeof(IN_TYPE)); |
+ |
+ return res == value; |
+} |
+ |
+static const std::vector<uint64_t> unsigned_test_values() { |
+ static const uint64_t kValues[] = { |
+ 0x2180f18a06384414, 0x000a714532102277, 0xbc1acccf180649f0, |
+ 0x8000000080008000, 0x0000000000000001, 0xffffffffffffffff, |
+ }; |
+ return std::vector<uint64_t>(&kValues[0], &kValues[arraysize(kValues)]); |
+} |
+ |
+static const std::vector<int32_t> unsigned_test_offset() { |
+ static const int32_t kValues[] = {// value, offset |
+ -132 * KB, -21 * KB, 0, 19 * KB, 135 * KB}; |
+ return std::vector<int32_t>(&kValues[0], &kValues[arraysize(kValues)]); |
+} |
+ |
+static const std::vector<int32_t> unsigned_test_offset_increment() { |
+ static const int32_t kValues[] = {-5, -4, -3, -2, -1, 0, 1, 2, 3, 4, 5}; |
+ return std::vector<int32_t>(&kValues[0], &kValues[arraysize(kValues)]); |
+} |
+ |
+TEST(Ulh) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ uint16_t value = static_cast<uint64_t>(*i & 0xFFFF); |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<uint16_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Ulh(v0, MemOperand(a0, in_offset)); |
+ __ Ush(v0, MemOperand(a0, out_offset), v0); |
+ })); |
+ CHECK_EQ(true, run_Unaligned<uint16_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ mov(t0, a0); |
+ __ Ulh(a0, MemOperand(a0, in_offset)); |
+ __ Ush(a0, MemOperand(t0, out_offset), v0); |
+ })); |
+ CHECK_EQ(true, run_Unaligned<uint16_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ mov(t0, a0); |
+ __ Ulhu(a0, MemOperand(a0, in_offset)); |
+ __ Ush(a0, MemOperand(t0, out_offset), t1); |
+ })); |
+ CHECK_EQ(true, run_Unaligned<uint16_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Ulhu(v0, MemOperand(a0, in_offset)); |
+ __ Ush(v0, MemOperand(a0, out_offset), t1); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
+TEST(Ulh_bitextension) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ uint16_t value = static_cast<uint64_t>(*i & 0xFFFF); |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<uint16_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ Label success, fail, end, different; |
+ __ Ulh(t0, MemOperand(a0, in_offset)); |
+ __ Ulhu(t1, MemOperand(a0, in_offset)); |
+ __ Branch(&different, ne, t0, Operand(t1)); |
+ |
+ // If signed and unsigned values are same, check |
+ // the upper bits to see if they are zero |
+ __ sra(t0, t0, 15); |
+ __ Branch(&success, eq, t0, Operand(zero_reg)); |
+ __ Branch(&fail); |
+ |
+ // If signed and unsigned values are different, |
+ // check that the upper bits are complementary |
+ __ bind(&different); |
+ __ sra(t1, t1, 15); |
+ __ Branch(&fail, ne, t1, Operand(1)); |
+ __ sra(t0, t0, 15); |
+ __ addiu(t0, t0, 1); |
+ __ Branch(&fail, ne, t0, Operand(zero_reg)); |
+ // Fall through to success |
+ |
+ __ bind(&success); |
+ __ Ulh(t0, MemOperand(a0, in_offset)); |
+ __ Ush(t0, MemOperand(a0, out_offset), v0); |
+ __ Branch(&end); |
+ __ bind(&fail); |
+ __ Ush(zero_reg, MemOperand(a0, out_offset), v0); |
+ __ bind(&end); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
+TEST(Ulw) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ uint32_t value = static_cast<uint32_t>(*i & 0xFFFFFFFF); |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<uint32_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Ulw(v0, MemOperand(a0, in_offset)); |
+ __ Usw(v0, MemOperand(a0, out_offset)); |
+ })); |
+ CHECK_EQ(true, |
+ run_Unaligned<uint32_t>( |
+ buffer_middle, in_offset, out_offset, (uint32_t)value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ mov(t0, a0); |
+ __ Ulw(a0, MemOperand(a0, in_offset)); |
+ __ Usw(a0, MemOperand(t0, out_offset)); |
+ })); |
+ CHECK_EQ(true, run_Unaligned<uint32_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Ulwu(v0, MemOperand(a0, in_offset)); |
+ __ Usw(v0, MemOperand(a0, out_offset)); |
+ })); |
+ CHECK_EQ(true, |
+ run_Unaligned<uint32_t>( |
+ buffer_middle, in_offset, out_offset, (uint32_t)value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ mov(t0, a0); |
+ __ Ulwu(a0, MemOperand(a0, in_offset)); |
+ __ Usw(a0, MemOperand(t0, out_offset)); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
+TEST(Ulw_extension) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ uint32_t value = static_cast<uint32_t>(*i & 0xFFFFFFFF); |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<uint32_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ Label success, fail, end, different; |
+ __ Ulw(t0, MemOperand(a0, in_offset)); |
+ __ Ulwu(t1, MemOperand(a0, in_offset)); |
+ __ Branch(&different, ne, t0, Operand(t1)); |
+ |
+ // If signed and unsigned values are same, check |
+ // the upper bits to see if they are zero |
+ __ dsra(t0, t0, 31); |
+ __ Branch(&success, eq, t0, Operand(zero_reg)); |
+ __ Branch(&fail); |
+ |
+ // If signed and unsigned values are different, |
+ // check that the upper bits are complementary |
+ __ bind(&different); |
+ __ dsra(t1, t1, 31); |
+ __ Branch(&fail, ne, t1, Operand(1)); |
+ __ dsra(t0, t0, 31); |
+ __ daddiu(t0, t0, 1); |
+ __ Branch(&fail, ne, t0, Operand(zero_reg)); |
+ // Fall through to success |
+ |
+ __ bind(&success); |
+ __ Ulw(t0, MemOperand(a0, in_offset)); |
+ __ Usw(t0, MemOperand(a0, out_offset)); |
+ __ Branch(&end); |
+ __ bind(&fail); |
+ __ Usw(zero_reg, MemOperand(a0, out_offset)); |
+ __ bind(&end); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
+TEST(Uld) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ uint64_t value = *i; |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<uint64_t>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Uld(v0, MemOperand(a0, in_offset)); |
+ __ Usd(v0, MemOperand(a0, out_offset)); |
+ })); |
+ CHECK_EQ(true, |
+ run_Unaligned<uint64_t>( |
+ buffer_middle, in_offset, out_offset, (uint32_t)value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ mov(t0, a0); |
+ __ Uld(a0, MemOperand(a0, in_offset)); |
+ __ Usd(a0, MemOperand(t0, out_offset)); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
+TEST(Ulwc1) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ float value = static_cast<float>(*i & 0xFFFFFFFF); |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<float>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Ulwc1(f0, MemOperand(a0, in_offset), t0); |
+ __ Uswc1(f0, MemOperand(a0, out_offset), t0); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
+TEST(Uldc1) { |
+ CcTest::InitializeVM(); |
+ |
+ static const int kBufferSize = 300 * KB; |
+ char memory_buffer[kBufferSize]; |
+ char* buffer_middle = memory_buffer + (kBufferSize / 2); |
+ |
+ FOR_UINT64_INPUTS(i, unsigned_test_values) { |
+ FOR_INT32_INPUTS2(j1, j2, unsigned_test_offset) { |
+ FOR_INT32_INPUTS2(k1, k2, unsigned_test_offset_increment) { |
+ double value = static_cast<double>(*i); |
+ int32_t in_offset = *j1 + *k1; |
+ int32_t out_offset = *j2 + *k2; |
+ |
+ CHECK_EQ(true, run_Unaligned<double>( |
+ buffer_middle, in_offset, out_offset, value, |
+ [](MacroAssembler* masm, int32_t in_offset, |
+ int32_t out_offset) { |
+ __ Uldc1(f0, MemOperand(a0, in_offset), t0); |
+ __ Usdc1(f0, MemOperand(a0, out_offset), t0); |
+ })); |
+ } |
+ } |
+ } |
+} |
+ |
#undef __ |