| Index: test/cctest/test-simulator-neon-arm64.cc
|
| diff --git a/test/cctest/test-simulator-neon-arm64.cc b/test/cctest/test-simulator-neon-arm64.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..7468e3d854d70fe20c98243c59e71ab81654a4b4
|
| --- /dev/null
|
| +++ b/test/cctest/test-simulator-neon-arm64.cc
|
| @@ -0,0 +1,2562 @@
|
| +// Copyright 2016 the V8 project authors. All rights reserved.
|
| +// Use of this source code is governed by a BSD-style license that can be
|
| +// found in the LICENSE file.
|
| +
|
| +#include <stdio.h>
|
| +#include <stdlib.h>
|
| +#include <string.h>
|
| +#include <cmath>
|
| +#include <limits>
|
| +
|
| +#include "src/arm64/decoder-arm64-inl.h"
|
| +#include "src/arm64/disasm-arm64.h"
|
| +#include "src/arm64/simulator-arm64.h"
|
| +#include "src/arm64/utils-arm64.h"
|
| +#include "src/base/platform/platform.h"
|
| +#include "src/base/utils/random-number-generator.h"
|
| +#include "src/macro-assembler-inl.h"
|
| +#include "test/cctest/cctest.h"
|
| +#include "test/cctest/test-simulator-neon-inputs-arm64.h"
|
| +#include "test/cctest/test-simulator-neon-traces-arm64.h"
|
| +#include "test/cctest/test-utils-arm64.h"
|
| +
|
| +using namespace v8::internal;
|
| +
|
| +// Test infrastructure.
|
| +//
|
| +// Tests are functions which accept no parameters and have no return values.
|
| +// The testing code should not perform an explicit return once completed. For
|
| +// example to test the mov immediate instruction a very simple test would be:
|
| +//
|
| +// SIMTEST(mov_x0_one) {
|
| +// SETUP();
|
| +//
|
| +// START();
|
| +// __ mov(x0, Operand(1));
|
| +// END();
|
| +//
|
| +// RUN();
|
| +//
|
| +// CHECK_EQUAL_64(1, x0);
|
| +//
|
| +// TEARDOWN();
|
| +// }
|
| +//
|
| +// Within a START ... END block all registers but sp can be modified. sp has to
|
| +// be explicitly saved/restored. The END() macro replaces the function return
|
| +// so it may appear multiple times in a test if the test has multiple exit
|
| +// points.
|
| +//
|
| +// Once the test has been run all integer and floating point registers as well
|
| +// as flags are accessible through a RegisterDump instance, see
|
| +// test-utils-arm64.h for more info on RegisterDump.
|
| +//
|
| +// We provide some helper assert to handle common cases:
|
| +//
|
| +// CHECK_EQUAL_32(int32_t, int32_t)
|
| +// CHECK_EQUAL_FP32(float, float)
|
| +// CHECK_EQUAL_32(int32_t, W register)
|
| +// CHECK_EQUAL_FP32(float, S register)
|
| +// CHECK_EQUAL_64(int64_t, int64_t)
|
| +// CHECK_EQUAL_FP64(double, double)
|
| +// CHECK_EQUAL_64(int64_t, X register)
|
| +// CHECK_EQUAL_64(X register, X register)
|
| +// CHECK_EQUAL_FP64(double, D register)
|
| +//
|
| +// e.g. CHECK_EQUAL_64(0.5, d30);
|
| +//
|
| +// If more advance computation is required before the assert then access the
|
| +// RegisterDump named core directly:
|
| +//
|
| +// CHECK_EQUAL_64(0x1234, core.xreg(0) & 0xffff);
|
| +
|
| +#if 0 // TODO(all): enable.
|
| +static v8::Persistent<v8::Context> env;
|
| +
|
| +static void InitializeVM() {
|
| + if (env.IsEmpty()) {
|
| + env = v8::Context::New();
|
| + }
|
| +}
|
| +#endif
|
| +
|
| +#define __ masm.
|
| +#define SIMTEST(name) TEST(SIM_##name)
|
| +
|
| +#define BUF_SIZE 8192
|
| +#define SETUP() SETUP_SIZE(BUF_SIZE)
|
| +
|
| +#define INIT_V8() CcTest::InitializeVM();
|
| +
|
| +#ifdef USE_SIMULATOR
|
| +
|
| +// Run tests with the simulator.
|
| +#define SETUP_SIZE(buf_size) \
|
| + Isolate* isolate = CcTest::i_isolate(); \
|
| + HandleScope scope(isolate); \
|
| + CHECK(isolate != NULL); \
|
| + byte* buf = new byte[buf_size]; \
|
| + MacroAssembler masm(isolate, buf, buf_size, \
|
| + v8::internal::CodeObjectRequired::kYes); \
|
| + Decoder<DispatchingDecoderVisitor>* decoder = \
|
| + new Decoder<DispatchingDecoderVisitor>(); \
|
| + Simulator simulator(decoder); \
|
| + RegisterDump core;
|
| +
|
| +// Reset the assembler and simulator, so that instructions can be generated,
|
| +// but don't actually emit any code. This can be used by tests that need to
|
| +// emit instructions at the start of the buffer. Note that START_AFTER_RESET
|
| +// must be called before any callee-saved register is modified, and before an
|
| +// END is encountered.
|
| +//
|
| +// Most tests should call START, rather than call RESET directly.
|
| +#define RESET() \
|
| + __ Reset(); \
|
| + simulator.ResetState();
|
| +
|
| +#define START_AFTER_RESET() \
|
| + __ SetStackPointer(csp); \
|
| + __ PushCalleeSavedRegisters(); \
|
| + __ Debug("Start test.", __LINE__, TRACE_ENABLE | LOG_ALL);
|
| +
|
| +#define START() \
|
| + RESET(); \
|
| + START_AFTER_RESET();
|
| +
|
| +#define RUN() simulator.RunFrom(reinterpret_cast<Instruction*>(buf))
|
| +
|
| +#define END() \
|
| + __ Debug("End test.", __LINE__, TRACE_DISABLE | LOG_ALL); \
|
| + core.Dump(&masm); \
|
| + __ PopCalleeSavedRegisters(); \
|
| + __ Ret(); \
|
| + __ GetCode(NULL);
|
| +
|
| +#define TEARDOWN() delete[] buf;
|
| +
|
| +#else // ifdef USE_SIMULATOR.
|
| +// Run the test on real hardware or models.
|
| +#define SETUP_SIZE(buf_size) \
|
| + Isolate* isolate = CcTest::i_isolate(); \
|
| + HandleScope scope(isolate); \
|
| + CHECK(isolate != NULL); \
|
| + size_t actual_size; \
|
| + byte* buf = static_cast<byte*>( \
|
| + v8::base::OS::Allocate(buf_size, &actual_size, true)); \
|
| + MacroAssembler masm(isolate, buf, actual_size, \
|
| + v8::internal::CodeObjectRequired::kYes); \
|
| + RegisterDump core;
|
| +
|
| +#define RESET() \
|
| + __ Reset(); \
|
| + /* Reset the machine state (like simulator.ResetState()). */ \
|
| + __ Msr(NZCV, xzr); \
|
| + __ Msr(FPCR, xzr);
|
| +
|
| +#define START_AFTER_RESET() \
|
| + __ SetStackPointer(csp); \
|
| + __ PushCalleeSavedRegisters();
|
| +
|
| +#define START() \
|
| + RESET(); \
|
| + START_AFTER_RESET();
|
| +
|
| +#define RUN() \
|
| + Assembler::FlushICache(isolate, buf, masm.SizeOfGeneratedCode()); \
|
| + { \
|
| + void (*test_function)(void); \
|
| + memcpy(&test_function, &buf, sizeof(buf)); \
|
| + test_function(); \
|
| + }
|
| +
|
| +#define END() \
|
| + core.Dump(&masm); \
|
| + __ PopCalleeSavedRegisters(); \
|
| + __ Ret(); \
|
| + __ GetCode(NULL);
|
| +
|
| +#define TEARDOWN() v8::base::OS::Free(buf, actual_size);
|
| +
|
| +#endif // ifdef USE_SIMULATOR.
|
| +
|
| +#define CHECK_EQUAL_NZCV(expected) CHECK(EqualNzcv(expected, core.flags_nzcv()))
|
| +
|
| +#define CHECK_EQUAL_REGISTERS(expected) CHECK(EqualRegisters(&expected, &core))
|
| +
|
| +#define CHECK_EQUAL_32(expected, result) \
|
| + CHECK(Equal32(static_cast<uint32_t>(expected), &core, result))
|
| +
|
| +#define CHECK_EQUAL_FP32(expected, result) \
|
| + CHECK(EqualFP32(expected, &core, result))
|
| +
|
| +#define CHECK_EQUAL_64(expected, result) CHECK(Equal64(expected, &core, result))
|
| +
|
| +#define CHECK_EQUAL_FP64(expected, result) \
|
| + CHECK(EqualFP64(expected, &core, result))
|
| +
|
| +#ifdef DEBUG
|
| +#define CHECK_LITERAL_POOL_SIZE(expected) \
|
| + CHECK((expected) == (__ LiteralPoolSize()))
|
| +#else
|
| +#define CHECK_LITERAL_POOL_SIZE(expected) ((void)0)
|
| +#endif
|
| +
|
| +// The maximum number of errors to report in detail for each test.
|
| +static const unsigned kErrorReportLimit = 8;
|
| +
|
| +typedef void (MacroAssembler::*Test1OpNEONHelper_t)(const VRegister& vd,
|
| + const VRegister& vn);
|
| +typedef void (MacroAssembler::*Test2OpNEONHelper_t)(const VRegister& vd,
|
| + const VRegister& vn,
|
| + const VRegister& vm);
|
| +typedef void (MacroAssembler::*TestByElementNEONHelper_t)(const VRegister& vd,
|
| + const VRegister& vn,
|
| + const VRegister& vm,
|
| + int vm_index);
|
| +typedef void (MacroAssembler::*TestOpImmOpImmVdUpdateNEONHelper_t)(
|
| + const VRegister& vd, int imm1, const VRegister& vn, int imm2);
|
| +
|
| +// This helps using the same typename for both the function pointer
|
| +// and the array of immediates passed to helper routines.
|
| +template <typename T>
|
| +class Test2OpImmediateNEONHelper_t {
|
| + public:
|
| + typedef void (MacroAssembler::*mnemonic)(const VRegister& vd,
|
| + const VRegister& vn, T imm);
|
| +};
|
| +
|
| +namespace {
|
| +
|
| +// Maximum number of hex characters required to represent values of either
|
| +// templated type.
|
| +template <typename Ta, typename Tb>
|
| +unsigned MaxHexCharCount() {
|
| + unsigned count = static_cast<unsigned>(std::max(sizeof(Ta), sizeof(Tb)));
|
| + return (count * 8) / 4;
|
| +}
|
| +
|
| +// ==== Tests for instructions of the form <INST> VReg, VReg. ====
|
| +
|
| +void Test1OpNEON_Helper(Test1OpNEONHelper_t helper, uintptr_t inputs_n,
|
| + unsigned inputs_n_length, uintptr_t results,
|
| + VectorFormat vd_form, VectorFormat vn_form) {
|
| + DCHECK_NE(vd_form, kFormatUndefined);
|
| + DCHECK_NE(vn_form, kFormatUndefined);
|
| +
|
| + SETUP();
|
| + START();
|
| +
|
| + // Roll up the loop to keep the code size down.
|
| + Label loop_n;
|
| +
|
| + Register out = x0;
|
| + Register inputs_n_base = x1;
|
| + Register inputs_n_last_16bytes = x3;
|
| + Register index_n = x5;
|
| +
|
| + const unsigned vd_bits = RegisterSizeInBitsFromFormat(vd_form);
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned vn_bits = RegisterSizeInBitsFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vn_form);
|
| + const unsigned vn_lane_bits = LaneSizeInBitsFromFormat(vn_form);
|
| +
|
| + // These will be either a D- or a Q-register form, with a single lane
|
| + // (for use in scalar load and store operations).
|
| + VRegister vd = VRegister::Create(0, vd_bits);
|
| + VRegister vn = v1.V16B();
|
| + VRegister vntmp = v3.V16B();
|
| +
|
| + // These will have the correct format for use when calling 'helper'.
|
| + VRegister vd_helper = VRegister::Create(0, vd_bits, vd_lane_count);
|
| + VRegister vn_helper = VRegister::Create(1, vn_bits, vn_lane_count);
|
| +
|
| + // 'v*tmp_single' will be either 'Vt.B', 'Vt.H', 'Vt.S' or 'Vt.D'.
|
| + VRegister vntmp_single = VRegister::Create(3, vn_lane_bits);
|
| +
|
| + __ Mov(out, results);
|
| +
|
| + __ Mov(inputs_n_base, inputs_n);
|
| + __ Mov(inputs_n_last_16bytes,
|
| + inputs_n + (vn_lane_bytes * inputs_n_length) - 16);
|
| +
|
| + __ Ldr(vn, MemOperand(inputs_n_last_16bytes));
|
| +
|
| + __ Mov(index_n, 0);
|
| + __ Bind(&loop_n);
|
| +
|
| + __ Ldr(vntmp_single,
|
| + MemOperand(inputs_n_base, index_n, LSL, vn_lane_bytes_log2));
|
| + __ Ext(vn, vn, vntmp, vn_lane_bytes);
|
| +
|
| + // Set the destination to zero.
|
| +
|
| + // TODO(all): Setting the destination to values other than zero might be a
|
| + // better test for instructions such as sqxtn2 which may leave parts of V
|
| + // registers unchanged.
|
| + __ Movi(vd.V16B(), 0);
|
| +
|
| + (masm.*helper)(vd_helper, vn_helper);
|
| +
|
| + __ Str(vd, MemOperand(out, vd.SizeInBytes(), PostIndex));
|
| +
|
| + __ Add(index_n, index_n, 1);
|
| + __ Cmp(index_n, inputs_n_length);
|
| + __ B(lo, &loop_n);
|
| +
|
| + END();
|
| + RUN();
|
| + TEARDOWN();
|
| +}
|
| +
|
| +// Test NEON instructions. The inputs_*[] and expected[] arrays should be
|
| +// arrays of rawbit representation of input values. This ensures that
|
| +// exact bit comparisons can be performed.
|
| +template <typename Td, typename Tn>
|
| +void Test1OpNEON(const char* name, Test1OpNEONHelper_t helper,
|
| + const Tn inputs_n[], unsigned inputs_n_length,
|
| + const Td expected[], unsigned expected_length,
|
| + VectorFormat vd_form, VectorFormat vn_form) {
|
| + DCHECK_GT(inputs_n_length, 0U);
|
| +
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| +
|
| + const unsigned results_length = inputs_n_length;
|
| + std::vector<Td> results(results_length * vd_lane_count, 0);
|
| + const unsigned lane_len_in_hex = MaxHexCharCount<Td, Tn>();
|
| +
|
| + Test1OpNEON_Helper(
|
| + helper, reinterpret_cast<uintptr_t>(inputs_n), inputs_n_length,
|
| + reinterpret_cast<uintptr_t>(results.data()), vd_form, vn_form);
|
| +
|
| + // Check the results.
|
| + CHECK(expected_length == results_length);
|
| + unsigned error_count = 0;
|
| + unsigned d = 0;
|
| + const char* padding = " ";
|
| + DCHECK_GE(strlen(padding), (lane_len_in_hex + 1));
|
| + for (unsigned n = 0; n < inputs_n_length; n++, d++) {
|
| + bool error_in_vector = false;
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index = (n * vd_lane_count) + lane;
|
| +
|
| + if (results[output_index] != expected[output_index]) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| + }
|
| +
|
| + if (error_in_vector && (++error_count <= kErrorReportLimit)) {
|
| + printf("%s\n", name);
|
| + printf(" Vn%.*s| Vd%.*s| Expected\n", lane_len_in_hex + 1, padding,
|
| + lane_len_in_hex + 1, padding);
|
| +
|
| + const unsigned first_index_n =
|
| + inputs_n_length - (16 / vn_lane_bytes) + n + 1;
|
| +
|
| + for (unsigned lane = 0; lane < std::max(vd_lane_count, vn_lane_count);
|
| + lane++) {
|
| + unsigned output_index = (n * vd_lane_count) + lane;
|
| + unsigned input_index_n = (first_index_n + lane) % inputs_n_length;
|
| +
|
| + printf("%c0x%0*" PRIx64 " | 0x%0*" PRIx64
|
| + " "
|
| + "| 0x%0*" PRIx64 "\n",
|
| + results[output_index] != expected[output_index] ? '*' : ' ',
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_n[input_index_n]),
|
| + lane_len_in_hex, static_cast<uint64_t>(results[output_index]),
|
| + lane_len_in_hex, static_cast<uint64_t>(expected[output_index]));
|
| + }
|
| + }
|
| + }
|
| + DCHECK_EQ(d, expected_length);
|
| + if (error_count > kErrorReportLimit) {
|
| + printf("%u other errors follow.\n", error_count - kErrorReportLimit);
|
| + }
|
| + DCHECK_EQ(error_count, 0U);
|
| +}
|
| +
|
| +// ==== Tests for instructions of the form <mnemonic> <V><d>, <Vn>.<T> ====
|
| +// where <V> is one of B, H, S or D registers.
|
| +// e.g. saddlv H1, v0.8B
|
| +
|
| +// TODO(all): Change tests to store all lanes of the resulting V register.
|
| +// Some tests store all 128 bits of the resulting V register to
|
| +// check the simulator's behaviour on the rest of the register.
|
| +// This is better than storing the affected lanes only.
|
| +// Change any tests such as the 'Across' template to do the same.
|
| +
|
| +void Test1OpAcrossNEON_Helper(Test1OpNEONHelper_t helper, uintptr_t inputs_n,
|
| + unsigned inputs_n_length, uintptr_t results,
|
| + VectorFormat vd_form, VectorFormat vn_form) {
|
| + DCHECK_NE(vd_form, kFormatUndefined);
|
| + DCHECK_NE(vn_form, kFormatUndefined);
|
| +
|
| + SETUP();
|
| + START();
|
| +
|
| + // Roll up the loop to keep the code size down.
|
| + Label loop_n;
|
| +
|
| + Register out = x0;
|
| + Register inputs_n_base = x1;
|
| + Register inputs_n_last_vector = x3;
|
| + Register index_n = x5;
|
| +
|
| + const unsigned vd_bits = RegisterSizeInBitsFromFormat(vd_form);
|
| + const unsigned vn_bits = RegisterSizeInBitsFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vn_form);
|
| + const unsigned vn_lane_bits = LaneSizeInBitsFromFormat(vn_form);
|
| +
|
| + // Test destructive operations by (arbitrarily) using the same register for
|
| + // B and S lane sizes.
|
| + bool destructive = (vd_bits == kBRegSize) || (vd_bits == kSRegSize);
|
| +
|
| + // These will be either a D- or a Q-register form, with a single lane
|
| + // (for use in scalar load and store operations).
|
| + // Create two aliases for v8; the first is the destination for the tested
|
| + // instruction, the second, the whole Q register to check the results.
|
| + VRegister vd = VRegister::Create(0, vd_bits);
|
| + VRegister vdstr = VRegister::Create(0, kQRegSizeInBits);
|
| +
|
| + VRegister vn = VRegister::Create(1, vn_bits);
|
| + VRegister vntmp = VRegister::Create(3, vn_bits);
|
| +
|
| + // These will have the correct format for use when calling 'helper'.
|
| + VRegister vd_helper = VRegister::Create(0, vn_bits, vn_lane_count);
|
| + VRegister vn_helper = VRegister::Create(1, vn_bits, vn_lane_count);
|
| +
|
| + // 'v*tmp_single' will be either 'Vt.B', 'Vt.H', 'Vt.S' or 'Vt.D'.
|
| + VRegister vntmp_single = VRegister::Create(3, vn_lane_bits);
|
| +
|
| + // Same registers for use in the 'ext' instructions.
|
| + VRegister vn_ext = (kDRegSizeInBits == vn_bits) ? vn.V8B() : vn.V16B();
|
| + VRegister vntmp_ext =
|
| + (kDRegSizeInBits == vn_bits) ? vntmp.V8B() : vntmp.V16B();
|
| +
|
| + __ Mov(out, results);
|
| +
|
| + __ Mov(inputs_n_base, inputs_n);
|
| + __ Mov(inputs_n_last_vector,
|
| + inputs_n + vn_lane_bytes * (inputs_n_length - vn_lane_count));
|
| +
|
| + __ Ldr(vn, MemOperand(inputs_n_last_vector));
|
| +
|
| + __ Mov(index_n, 0);
|
| + __ Bind(&loop_n);
|
| +
|
| + __ Ldr(vntmp_single,
|
| + MemOperand(inputs_n_base, index_n, LSL, vn_lane_bytes_log2));
|
| + __ Ext(vn_ext, vn_ext, vntmp_ext, vn_lane_bytes);
|
| +
|
| + if (destructive) {
|
| + __ Mov(vd_helper, vn_helper);
|
| + (masm.*helper)(vd, vd_helper);
|
| + } else {
|
| + (masm.*helper)(vd, vn_helper);
|
| + }
|
| +
|
| + __ Str(vdstr, MemOperand(out, kQRegSize, PostIndex));
|
| +
|
| + __ Add(index_n, index_n, 1);
|
| + __ Cmp(index_n, inputs_n_length);
|
| + __ B(lo, &loop_n);
|
| +
|
| + END();
|
| + RUN();
|
| + TEARDOWN();
|
| +}
|
| +
|
| +// Test NEON instructions. The inputs_*[] and expected[] arrays should be
|
| +// arrays of rawbit representation of input values. This ensures that
|
| +// exact bit comparisons can be performed.
|
| +template <typename Td, typename Tn>
|
| +void Test1OpAcrossNEON(const char* name, Test1OpNEONHelper_t helper,
|
| + const Tn inputs_n[], unsigned inputs_n_length,
|
| + const Td expected[], unsigned expected_length,
|
| + VectorFormat vd_form, VectorFormat vn_form) {
|
| + DCHECK_GT(inputs_n_length, 0U);
|
| +
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| + const unsigned vd_lanes_per_q = MaxLaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned results_length = inputs_n_length;
|
| + std::vector<Td> results(results_length * vd_lanes_per_q, 0);
|
| + const unsigned lane_len_in_hex = MaxHexCharCount<Td, Tn>();
|
| +
|
| + Test1OpAcrossNEON_Helper(
|
| + helper, reinterpret_cast<uintptr_t>(inputs_n), inputs_n_length,
|
| + reinterpret_cast<uintptr_t>(results.data()), vd_form, vn_form);
|
| +
|
| + // Check the results.
|
| + DCHECK_EQ(expected_length, results_length);
|
| + unsigned error_count = 0;
|
| + unsigned d = 0;
|
| + const char* padding = " ";
|
| + DCHECK_GE(strlen(padding), (lane_len_in_hex + 1));
|
| + for (unsigned n = 0; n < inputs_n_length; n++, d++) {
|
| + bool error_in_vector = false;
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned expected_index = (n * vd_lane_count) + lane;
|
| + unsigned results_index = (n * vd_lanes_per_q) + lane;
|
| +
|
| + if (results[results_index] != expected[expected_index]) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| +
|
| + // For across operations, the remaining lanes should be zero.
|
| + for (unsigned lane = vd_lane_count; lane < vd_lanes_per_q; lane++) {
|
| + unsigned results_index = (n * vd_lanes_per_q) + lane;
|
| + if (results[results_index] != 0) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| + }
|
| + }
|
| +
|
| + if (error_in_vector && (++error_count <= kErrorReportLimit)) {
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| +
|
| + printf("%s\n", name);
|
| + printf(" Vn%.*s| Vd%.*s| Expected\n", lane_len_in_hex + 1, padding,
|
| + lane_len_in_hex + 1, padding);
|
| +
|
| + for (unsigned lane = 0; lane < vn_lane_count; lane++) {
|
| + unsigned results_index =
|
| + (n * vd_lanes_per_q) + ((vn_lane_count - 1) - lane);
|
| + unsigned input_index_n =
|
| + (inputs_n_length - vn_lane_count + n + 1 + lane) % inputs_n_length;
|
| +
|
| + Td expect = 0;
|
| + if ((vn_lane_count - 1) == lane) {
|
| + // This is the last lane to be printed, ie. the least-significant
|
| + // lane, so use the expected value; any other lane should be zero.
|
| + unsigned expected_index = n * vd_lane_count;
|
| + expect = expected[expected_index];
|
| + }
|
| + printf("%c0x%0*" PRIx64 " | 0x%0*" PRIx64 " | 0x%0*" PRIx64 "\n",
|
| + results[results_index] != expect ? '*' : ' ', lane_len_in_hex,
|
| + static_cast<uint64_t>(inputs_n[input_index_n]), lane_len_in_hex,
|
| + static_cast<uint64_t>(results[results_index]), lane_len_in_hex,
|
| + static_cast<uint64_t>(expect));
|
| + }
|
| + }
|
| + }
|
| + DCHECK_EQ(d, expected_length);
|
| + if (error_count > kErrorReportLimit) {
|
| + printf("%u other errors follow.\n", error_count - kErrorReportLimit);
|
| + }
|
| + DCHECK_EQ(error_count, 0U);
|
| +}
|
| +
|
| +// ==== Tests for instructions of the form <INST> VReg, VReg, VReg. ====
|
| +
|
| +void Test2OpNEON_Helper(Test2OpNEONHelper_t helper, uintptr_t inputs_d,
|
| + uintptr_t inputs_n, unsigned inputs_n_length,
|
| + uintptr_t inputs_m, unsigned inputs_m_length,
|
| + uintptr_t results, VectorFormat vd_form,
|
| + VectorFormat vn_form, VectorFormat vm_form) {
|
| + DCHECK_NE(vd_form, kFormatUndefined);
|
| + DCHECK_NE(vn_form, kFormatUndefined);
|
| + DCHECK_NE(vm_form, kFormatUndefined);
|
| +
|
| + SETUP();
|
| + START();
|
| +
|
| + // Roll up the loop to keep the code size down.
|
| + Label loop_n, loop_m;
|
| +
|
| + Register out = x0;
|
| + Register inputs_n_base = x1;
|
| + Register inputs_m_base = x2;
|
| + Register inputs_d_base = x3;
|
| + Register inputs_n_last_16bytes = x4;
|
| + Register inputs_m_last_16bytes = x5;
|
| + Register index_n = x6;
|
| + Register index_m = x7;
|
| +
|
| + const unsigned vd_bits = RegisterSizeInBitsFromFormat(vd_form);
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned vn_bits = RegisterSizeInBitsFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vn_form);
|
| + const unsigned vn_lane_bits = LaneSizeInBitsFromFormat(vn_form);
|
| +
|
| + const unsigned vm_bits = RegisterSizeInBitsFromFormat(vm_form);
|
| + const unsigned vm_lane_count = LaneCountFromFormat(vm_form);
|
| + const unsigned vm_lane_bytes = LaneSizeInBytesFromFormat(vm_form);
|
| + const unsigned vm_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vm_form);
|
| + const unsigned vm_lane_bits = LaneSizeInBitsFromFormat(vm_form);
|
| +
|
| + // Always load and store 128 bits regardless of the format.
|
| + VRegister vd = v0.V16B();
|
| + VRegister vn = v1.V16B();
|
| + VRegister vm = v2.V16B();
|
| + VRegister vntmp = v3.V16B();
|
| + VRegister vmtmp = v4.V16B();
|
| + VRegister vres = v5.V16B();
|
| +
|
| + // These will have the correct format for calling the 'helper'.
|
| + VRegister vn_helper = VRegister::Create(1, vn_bits, vn_lane_count);
|
| + VRegister vm_helper = VRegister::Create(2, vm_bits, vm_lane_count);
|
| + VRegister vres_helper = VRegister::Create(5, vd_bits, vd_lane_count);
|
| +
|
| + // 'v*tmp_single' will be either 'Vt.B', 'Vt.H', 'Vt.S' or 'Vt.D'.
|
| + VRegister vntmp_single = VRegister::Create(3, vn_lane_bits);
|
| + VRegister vmtmp_single = VRegister::Create(4, vm_lane_bits);
|
| +
|
| + __ Mov(out, results);
|
| +
|
| + __ Mov(inputs_d_base, inputs_d);
|
| +
|
| + __ Mov(inputs_n_base, inputs_n);
|
| + __ Mov(inputs_n_last_16bytes, inputs_n + (inputs_n_length - 16));
|
| + __ Mov(inputs_m_base, inputs_m);
|
| + __ Mov(inputs_m_last_16bytes, inputs_m + (inputs_m_length - 16));
|
| +
|
| + __ Ldr(vd, MemOperand(inputs_d_base));
|
| + __ Ldr(vn, MemOperand(inputs_n_last_16bytes));
|
| + __ Ldr(vm, MemOperand(inputs_m_last_16bytes));
|
| +
|
| + __ Mov(index_n, 0);
|
| + __ Bind(&loop_n);
|
| +
|
| + __ Ldr(vntmp_single,
|
| + MemOperand(inputs_n_base, index_n, LSL, vn_lane_bytes_log2));
|
| + __ Ext(vn, vn, vntmp, vn_lane_bytes);
|
| +
|
| + __ Mov(index_m, 0);
|
| + __ Bind(&loop_m);
|
| +
|
| + __ Ldr(vmtmp_single,
|
| + MemOperand(inputs_m_base, index_m, LSL, vm_lane_bytes_log2));
|
| + __ Ext(vm, vm, vmtmp, vm_lane_bytes);
|
| +
|
| + __ Mov(vres, vd);
|
| +
|
| + (masm.*helper)(vres_helper, vn_helper, vm_helper);
|
| +
|
| + __ Str(vres, MemOperand(out, vd.SizeInBytes(), PostIndex));
|
| +
|
| + __ Add(index_m, index_m, 1);
|
| + __ Cmp(index_m, inputs_m_length);
|
| + __ B(lo, &loop_m);
|
| +
|
| + __ Add(index_n, index_n, 1);
|
| + __ Cmp(index_n, inputs_n_length);
|
| + __ B(lo, &loop_n);
|
| +
|
| + END();
|
| + RUN();
|
| + TEARDOWN();
|
| +}
|
| +
|
| +// Test NEON instructions. The inputs_*[] and expected[] arrays should be
|
| +// arrays of rawbit representation of input values. This ensures that
|
| +// exact bit comparisons can be performed.
|
| +template <typename Td, typename Tn, typename Tm>
|
| +void Test2OpNEON(const char* name, Test2OpNEONHelper_t helper,
|
| + const Td inputs_d[], const Tn inputs_n[],
|
| + unsigned inputs_n_length, const Tm inputs_m[],
|
| + unsigned inputs_m_length, const Td expected[],
|
| + unsigned expected_length, VectorFormat vd_form,
|
| + VectorFormat vn_form, VectorFormat vm_form) {
|
| + DCHECK(inputs_n_length > 0 && inputs_m_length > 0);
|
| +
|
| + const unsigned vd_lane_count = MaxLaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned results_length = inputs_n_length * inputs_m_length;
|
| + std::vector<Td> results(results_length * vd_lane_count);
|
| + const unsigned lane_len_in_hex =
|
| + static_cast<unsigned>(std::max(sizeof(Td), sizeof(Tm)) * 8) / 4;
|
| +
|
| + Test2OpNEON_Helper(helper, reinterpret_cast<uintptr_t>(inputs_d),
|
| + reinterpret_cast<uintptr_t>(inputs_n), inputs_n_length,
|
| + reinterpret_cast<uintptr_t>(inputs_m), inputs_m_length,
|
| + reinterpret_cast<uintptr_t>(results.data()), vd_form,
|
| + vn_form, vm_form);
|
| +
|
| + // Check the results.
|
| + CHECK(expected_length == results_length);
|
| + unsigned error_count = 0;
|
| + unsigned d = 0;
|
| + const char* padding = " ";
|
| + DCHECK_GE(strlen(padding), lane_len_in_hex + 1);
|
| + for (unsigned n = 0; n < inputs_n_length; n++) {
|
| + for (unsigned m = 0; m < inputs_m_length; m++, d++) {
|
| + bool error_in_vector = false;
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index =
|
| + (n * inputs_m_length * vd_lane_count) + (m * vd_lane_count) + lane;
|
| +
|
| + if (results[output_index] != expected[output_index]) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| + }
|
| +
|
| + if (error_in_vector && (++error_count <= kErrorReportLimit)) {
|
| + printf("%s\n", name);
|
| + printf(" Vd%.*s| Vn%.*s| Vm%.*s| Vd%.*s| Expected\n",
|
| + lane_len_in_hex + 1, padding, lane_len_in_hex + 1, padding,
|
| + lane_len_in_hex + 1, padding, lane_len_in_hex + 1, padding);
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index = (n * inputs_m_length * vd_lane_count) +
|
| + (m * vd_lane_count) + lane;
|
| + unsigned input_index_n =
|
| + (inputs_n_length - vd_lane_count + n + 1 + lane) %
|
| + inputs_n_length;
|
| + unsigned input_index_m =
|
| + (inputs_m_length - vd_lane_count + m + 1 + lane) %
|
| + inputs_m_length;
|
| +
|
| + printf(
|
| + "%c0x%0*" PRIx64 " | 0x%0*" PRIx64 " | 0x%0*" PRIx64
|
| + " "
|
| + "| 0x%0*" PRIx64 " | 0x%0*" PRIx64 "\n",
|
| + results[output_index] != expected[output_index] ? '*' : ' ',
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_d[lane]),
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_n[input_index_n]),
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_m[input_index_m]),
|
| + lane_len_in_hex, static_cast<uint64_t>(results[output_index]),
|
| + lane_len_in_hex, static_cast<uint64_t>(expected[output_index]));
|
| + }
|
| + }
|
| + }
|
| + }
|
| + DCHECK_EQ(d, expected_length);
|
| + if (error_count > kErrorReportLimit) {
|
| + printf("%u other errors follow.\n", error_count - kErrorReportLimit);
|
| + }
|
| + DCHECK_EQ(error_count, 0U);
|
| +}
|
| +
|
| +// ==== Tests for instructions of the form <INST> Vd, Vn, Vm[<#index>]. ====
|
| +
|
| +void TestByElementNEON_Helper(TestByElementNEONHelper_t helper,
|
| + uintptr_t inputs_d, uintptr_t inputs_n,
|
| + unsigned inputs_n_length, uintptr_t inputs_m,
|
| + unsigned inputs_m_length, const int indices[],
|
| + unsigned indices_length, uintptr_t results,
|
| + VectorFormat vd_form, VectorFormat vn_form,
|
| + VectorFormat vm_form) {
|
| + DCHECK_NE(vd_form, kFormatUndefined);
|
| + DCHECK_NE(vn_form, kFormatUndefined);
|
| + DCHECK_NE(vm_form, kFormatUndefined);
|
| +
|
| + SETUP();
|
| + START();
|
| +
|
| + // Roll up the loop to keep the code size down.
|
| + Label loop_n, loop_m;
|
| +
|
| + Register out = x0;
|
| + Register inputs_n_base = x1;
|
| + Register inputs_m_base = x2;
|
| + Register inputs_d_base = x3;
|
| + Register inputs_n_last_16bytes = x4;
|
| + Register inputs_m_last_16bytes = x5;
|
| + Register index_n = x6;
|
| + Register index_m = x7;
|
| +
|
| + const unsigned vd_bits = RegisterSizeInBitsFromFormat(vd_form);
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned vn_bits = RegisterSizeInBitsFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vn_form);
|
| + const unsigned vn_lane_bits = LaneSizeInBitsFromFormat(vn_form);
|
| +
|
| + const unsigned vm_bits = RegisterSizeInBitsFromFormat(vm_form);
|
| + const unsigned vm_lane_count = LaneCountFromFormat(vm_form);
|
| + const unsigned vm_lane_bytes = LaneSizeInBytesFromFormat(vm_form);
|
| + const unsigned vm_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vm_form);
|
| + const unsigned vm_lane_bits = LaneSizeInBitsFromFormat(vm_form);
|
| +
|
| + // Always load and store 128 bits regardless of the format.
|
| + VRegister vd = v0.V16B();
|
| + VRegister vn = v1.V16B();
|
| + VRegister vm = v2.V16B();
|
| + VRegister vntmp = v3.V16B();
|
| + VRegister vmtmp = v4.V16B();
|
| + VRegister vres = v5.V16B();
|
| +
|
| + // These will have the correct format for calling the 'helper'.
|
| + VRegister vn_helper = VRegister::Create(1, vn_bits, vn_lane_count);
|
| + VRegister vm_helper = VRegister::Create(2, vm_bits, vm_lane_count);
|
| + VRegister vres_helper = VRegister::Create(5, vd_bits, vd_lane_count);
|
| +
|
| + // 'v*tmp_single' will be either 'Vt.B', 'Vt.H', 'Vt.S' or 'Vt.D'.
|
| + VRegister vntmp_single = VRegister::Create(3, vn_lane_bits);
|
| + VRegister vmtmp_single = VRegister::Create(4, vm_lane_bits);
|
| +
|
| + __ Mov(out, results);
|
| +
|
| + __ Mov(inputs_d_base, inputs_d);
|
| +
|
| + __ Mov(inputs_n_base, inputs_n);
|
| + __ Mov(inputs_n_last_16bytes, inputs_n + (inputs_n_length - 16));
|
| + __ Mov(inputs_m_base, inputs_m);
|
| + __ Mov(inputs_m_last_16bytes, inputs_m + (inputs_m_length - 16));
|
| +
|
| + __ Ldr(vd, MemOperand(inputs_d_base));
|
| + __ Ldr(vn, MemOperand(inputs_n_last_16bytes));
|
| + __ Ldr(vm, MemOperand(inputs_m_last_16bytes));
|
| +
|
| + __ Mov(index_n, 0);
|
| + __ Bind(&loop_n);
|
| +
|
| + __ Ldr(vntmp_single,
|
| + MemOperand(inputs_n_base, index_n, LSL, vn_lane_bytes_log2));
|
| + __ Ext(vn, vn, vntmp, vn_lane_bytes);
|
| +
|
| + __ Mov(index_m, 0);
|
| + __ Bind(&loop_m);
|
| +
|
| + __ Ldr(vmtmp_single,
|
| + MemOperand(inputs_m_base, index_m, LSL, vm_lane_bytes_log2));
|
| + __ Ext(vm, vm, vmtmp, vm_lane_bytes);
|
| +
|
| + __ Mov(vres, vd);
|
| + {
|
| + for (unsigned i = 0; i < indices_length; i++) {
|
| + (masm.*helper)(vres_helper, vn_helper, vm_helper, indices[i]);
|
| + __ Str(vres, MemOperand(out, vd.SizeInBytes(), PostIndex));
|
| + }
|
| + }
|
| +
|
| + __ Add(index_m, index_m, 1);
|
| + __ Cmp(index_m, inputs_m_length);
|
| + __ B(lo, &loop_m);
|
| +
|
| + __ Add(index_n, index_n, 1);
|
| + __ Cmp(index_n, inputs_n_length);
|
| + __ B(lo, &loop_n);
|
| +
|
| + END();
|
| + RUN();
|
| + TEARDOWN();
|
| +}
|
| +
|
| +// Test NEON instructions. The inputs_*[] and expected[] arrays should be
|
| +// arrays of rawbit representation of input values. This ensures that
|
| +// exact bit comparisons can be performed.
|
| +template <typename Td, typename Tn, typename Tm>
|
| +void TestByElementNEON(const char* name, TestByElementNEONHelper_t helper,
|
| + const Td inputs_d[], const Tn inputs_n[],
|
| + unsigned inputs_n_length, const Tm inputs_m[],
|
| + unsigned inputs_m_length, const int indices[],
|
| + unsigned indices_length, const Td expected[],
|
| + unsigned expected_length, VectorFormat vd_form,
|
| + VectorFormat vn_form, VectorFormat vm_form) {
|
| + DCHECK_GT(inputs_n_length, 0U);
|
| + DCHECK_GT(inputs_m_length, 0U);
|
| + DCHECK_GT(indices_length, 0U);
|
| +
|
| + const unsigned vd_lane_count = MaxLaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned results_length =
|
| + inputs_n_length * inputs_m_length * indices_length;
|
| + std::vector<Td> results(results_length * vd_lane_count, 0);
|
| + const unsigned lane_len_in_hex = MaxHexCharCount<Td, Tm>();
|
| +
|
| + TestByElementNEON_Helper(
|
| + helper, reinterpret_cast<uintptr_t>(inputs_d),
|
| + reinterpret_cast<uintptr_t>(inputs_n), inputs_n_length,
|
| + reinterpret_cast<uintptr_t>(inputs_m), inputs_m_length, indices,
|
| + indices_length, reinterpret_cast<uintptr_t>(results.data()), vd_form,
|
| + vn_form, vm_form);
|
| +
|
| + // Check the results.
|
| + CHECK(expected_length == results_length);
|
| + unsigned error_count = 0;
|
| + unsigned d = 0;
|
| + const char* padding = " ";
|
| + DCHECK_GE(strlen(padding), lane_len_in_hex + 1);
|
| + for (unsigned n = 0; n < inputs_n_length; n++) {
|
| + for (unsigned m = 0; m < inputs_m_length; m++) {
|
| + for (unsigned index = 0; index < indices_length; index++, d++) {
|
| + bool error_in_vector = false;
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index =
|
| + (n * inputs_m_length * indices_length * vd_lane_count) +
|
| + (m * indices_length * vd_lane_count) + (index * vd_lane_count) +
|
| + lane;
|
| +
|
| + if (results[output_index] != expected[output_index]) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| + }
|
| +
|
| + if (error_in_vector && (++error_count <= kErrorReportLimit)) {
|
| + printf("%s\n", name);
|
| + printf(" Vd%.*s| Vn%.*s| Vm%.*s| Index | Vd%.*s| Expected\n",
|
| + lane_len_in_hex + 1, padding, lane_len_in_hex + 1, padding,
|
| + lane_len_in_hex + 1, padding, lane_len_in_hex + 1, padding);
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index =
|
| + (n * inputs_m_length * indices_length * vd_lane_count) +
|
| + (m * indices_length * vd_lane_count) + (index * vd_lane_count) +
|
| + lane;
|
| + unsigned input_index_n =
|
| + (inputs_n_length - vd_lane_count + n + 1 + lane) %
|
| + inputs_n_length;
|
| + unsigned input_index_m =
|
| + (inputs_m_length - vd_lane_count + m + 1 + lane) %
|
| + inputs_m_length;
|
| +
|
| + printf(
|
| + "%c0x%0*" PRIx64 " | 0x%0*" PRIx64 " | 0x%0*" PRIx64
|
| + " "
|
| + "| [%3d] | 0x%0*" PRIx64 " | 0x%0*" PRIx64 "\n",
|
| + results[output_index] != expected[output_index] ? '*' : ' ',
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_d[lane]),
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_n[input_index_n]),
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_m[input_index_m]),
|
| + indices[index], lane_len_in_hex,
|
| + static_cast<uint64_t>(results[output_index]), lane_len_in_hex,
|
| + static_cast<uint64_t>(expected[output_index]));
|
| + }
|
| + }
|
| + }
|
| + }
|
| + }
|
| + DCHECK_EQ(d, expected_length);
|
| + if (error_count > kErrorReportLimit) {
|
| + printf("%u other errors follow.\n", error_count - kErrorReportLimit);
|
| + }
|
| + CHECK(error_count == 0);
|
| +}
|
| +
|
| +// ==== Tests for instructions of the form <INST> VReg, VReg, #Immediate. ====
|
| +
|
| +template <typename Tm>
|
| +void Test2OpImmNEON_Helper(
|
| + typename Test2OpImmediateNEONHelper_t<Tm>::mnemonic helper,
|
| + uintptr_t inputs_n, unsigned inputs_n_length, const Tm inputs_m[],
|
| + unsigned inputs_m_length, uintptr_t results, VectorFormat vd_form,
|
| + VectorFormat vn_form) {
|
| + DCHECK(vd_form != kFormatUndefined && vn_form != kFormatUndefined);
|
| +
|
| + SETUP();
|
| + START();
|
| +
|
| + // Roll up the loop to keep the code size down.
|
| + Label loop_n;
|
| +
|
| + Register out = x0;
|
| + Register inputs_n_base = x1;
|
| + Register inputs_n_last_16bytes = x3;
|
| + Register index_n = x5;
|
| +
|
| + const unsigned vd_bits = RegisterSizeInBitsFromFormat(vd_form);
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned vn_bits = RegisterSizeInBitsFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vn_form);
|
| + const unsigned vn_lane_bits = LaneSizeInBitsFromFormat(vn_form);
|
| +
|
| + // These will be either a D- or a Q-register form, with a single lane
|
| + // (for use in scalar load and store operations).
|
| + VRegister vd = VRegister::Create(0, vd_bits);
|
| + VRegister vn = v1.V16B();
|
| + VRegister vntmp = v3.V16B();
|
| +
|
| + // These will have the correct format for use when calling 'helper'.
|
| + VRegister vd_helper = VRegister::Create(0, vd_bits, vd_lane_count);
|
| + VRegister vn_helper = VRegister::Create(1, vn_bits, vn_lane_count);
|
| +
|
| + // 'v*tmp_single' will be either 'Vt.B', 'Vt.H', 'Vt.S' or 'Vt.D'.
|
| + VRegister vntmp_single = VRegister::Create(3, vn_lane_bits);
|
| +
|
| + __ Mov(out, results);
|
| +
|
| + __ Mov(inputs_n_base, inputs_n);
|
| + __ Mov(inputs_n_last_16bytes,
|
| + inputs_n + (vn_lane_bytes * inputs_n_length) - 16);
|
| +
|
| + __ Ldr(vn, MemOperand(inputs_n_last_16bytes));
|
| +
|
| + __ Mov(index_n, 0);
|
| + __ Bind(&loop_n);
|
| +
|
| + __ Ldr(vntmp_single,
|
| + MemOperand(inputs_n_base, index_n, LSL, vn_lane_bytes_log2));
|
| + __ Ext(vn, vn, vntmp, vn_lane_bytes);
|
| +
|
| + // Set the destination to zero for tests such as '[r]shrn2'.
|
| + // TODO(all): Setting the destination to values other than zero might be a
|
| + // better test for shift and accumulate instructions (srsra/ssra/usra/ursra).
|
| + __ Movi(vd.V16B(), 0);
|
| +
|
| + {
|
| + for (unsigned i = 0; i < inputs_m_length; i++) {
|
| + (masm.*helper)(vd_helper, vn_helper, inputs_m[i]);
|
| + __ Str(vd, MemOperand(out, vd.SizeInBytes(), PostIndex));
|
| + }
|
| + }
|
| +
|
| + __ Add(index_n, index_n, 1);
|
| + __ Cmp(index_n, inputs_n_length);
|
| + __ B(lo, &loop_n);
|
| +
|
| + END();
|
| + RUN();
|
| + TEARDOWN();
|
| +}
|
| +
|
| +// Test NEON instructions. The inputs_*[] and expected[] arrays should be
|
| +// arrays of rawbit representation of input values. This ensures that
|
| +// exact bit comparisons can be performed.
|
| +template <typename Td, typename Tn, typename Tm>
|
| +void Test2OpImmNEON(const char* name,
|
| + typename Test2OpImmediateNEONHelper_t<Tm>::mnemonic helper,
|
| + const Tn inputs_n[], unsigned inputs_n_length,
|
| + const Tm inputs_m[], unsigned inputs_m_length,
|
| + const Td expected[], unsigned expected_length,
|
| + VectorFormat vd_form, VectorFormat vn_form) {
|
| + DCHECK(inputs_n_length > 0 && inputs_m_length > 0);
|
| +
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| +
|
| + const unsigned results_length = inputs_n_length * inputs_m_length;
|
| + std::vector<Td> results(results_length * vd_lane_count, 0);
|
| + const unsigned lane_len_in_hex = MaxHexCharCount<Td, Tn>();
|
| +
|
| + Test2OpImmNEON_Helper(helper, reinterpret_cast<uintptr_t>(inputs_n),
|
| + inputs_n_length, inputs_m, inputs_m_length,
|
| + reinterpret_cast<uintptr_t>(results.data()), vd_form,
|
| + vn_form);
|
| +
|
| + // Check the results.
|
| + CHECK(expected_length == results_length);
|
| + unsigned error_count = 0;
|
| + unsigned d = 0;
|
| + const char* padding = " ";
|
| + DCHECK_GE(strlen(padding), lane_len_in_hex + 1);
|
| + for (unsigned n = 0; n < inputs_n_length; n++) {
|
| + for (unsigned m = 0; m < inputs_m_length; m++, d++) {
|
| + bool error_in_vector = false;
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index =
|
| + (n * inputs_m_length * vd_lane_count) + (m * vd_lane_count) + lane;
|
| +
|
| + if (results[output_index] != expected[output_index]) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| + }
|
| +
|
| + if (error_in_vector && (++error_count <= kErrorReportLimit)) {
|
| + printf("%s\n", name);
|
| + printf(" Vn%.*s| Imm%.*s| Vd%.*s| Expected\n", lane_len_in_hex + 1,
|
| + padding, lane_len_in_hex, padding, lane_len_in_hex + 1, padding);
|
| +
|
| + const unsigned first_index_n =
|
| + inputs_n_length - (16 / vn_lane_bytes) + n + 1;
|
| +
|
| + for (unsigned lane = 0; lane < std::max(vd_lane_count, vn_lane_count);
|
| + lane++) {
|
| + unsigned output_index = (n * inputs_m_length * vd_lane_count) +
|
| + (m * vd_lane_count) + lane;
|
| + unsigned input_index_n = (first_index_n + lane) % inputs_n_length;
|
| + unsigned input_index_m = m;
|
| +
|
| + printf(
|
| + "%c0x%0*" PRIx64 " | 0x%0*" PRIx64
|
| + " "
|
| + "| 0x%0*" PRIx64 " | 0x%0*" PRIx64 "\n",
|
| + results[output_index] != expected[output_index] ? '*' : ' ',
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_n[input_index_n]),
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_m[input_index_m]),
|
| + lane_len_in_hex, static_cast<uint64_t>(results[output_index]),
|
| + lane_len_in_hex, static_cast<uint64_t>(expected[output_index]));
|
| + }
|
| + }
|
| + }
|
| + }
|
| + DCHECK_EQ(d, expected_length);
|
| + if (error_count > kErrorReportLimit) {
|
| + printf("%u other errors follow.\n", error_count - kErrorReportLimit);
|
| + }
|
| + CHECK(error_count == 0);
|
| +}
|
| +
|
| +// ==== Tests for instructions of the form <INST> VReg, #Imm, VReg, #Imm. ====
|
| +
|
| +void TestOpImmOpImmNEON_Helper(TestOpImmOpImmVdUpdateNEONHelper_t helper,
|
| + uintptr_t inputs_d, const int inputs_imm1[],
|
| + unsigned inputs_imm1_length, uintptr_t inputs_n,
|
| + unsigned inputs_n_length,
|
| + const int inputs_imm2[],
|
| + unsigned inputs_imm2_length, uintptr_t results,
|
| + VectorFormat vd_form, VectorFormat vn_form) {
|
| + DCHECK_NE(vd_form, kFormatUndefined);
|
| + DCHECK_NE(vn_form, kFormatUndefined);
|
| +
|
| + SETUP();
|
| + START();
|
| +
|
| + // Roll up the loop to keep the code size down.
|
| + Label loop_n;
|
| +
|
| + Register out = x0;
|
| + Register inputs_d_base = x1;
|
| + Register inputs_n_base = x2;
|
| + Register inputs_n_last_vector = x4;
|
| + Register index_n = x6;
|
| +
|
| + const unsigned vd_bits = RegisterSizeInBitsFromFormat(vd_form);
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned vn_bits = RegisterSizeInBitsFromFormat(vn_form);
|
| + const unsigned vn_lane_count = LaneCountFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes = LaneSizeInBytesFromFormat(vn_form);
|
| + const unsigned vn_lane_bytes_log2 = LaneSizeInBytesLog2FromFormat(vn_form);
|
| + const unsigned vn_lane_bits = LaneSizeInBitsFromFormat(vn_form);
|
| +
|
| + // These will be either a D- or a Q-register form, with a single lane
|
| + // (for use in scalar load and store operations).
|
| + VRegister vd = VRegister::Create(0, vd_bits);
|
| + VRegister vn = VRegister::Create(1, vn_bits);
|
| + VRegister vntmp = VRegister::Create(4, vn_bits);
|
| + VRegister vres = VRegister::Create(5, vn_bits);
|
| +
|
| + VRegister vn_helper = VRegister::Create(1, vn_bits, vn_lane_count);
|
| + VRegister vres_helper = VRegister::Create(5, vd_bits, vd_lane_count);
|
| +
|
| + // 'v*tmp_single' will be either 'Vt.B', 'Vt.H', 'Vt.S' or 'Vt.D'.
|
| + VRegister vntmp_single = VRegister::Create(4, vn_lane_bits);
|
| +
|
| + // Same registers for use in the 'ext' instructions.
|
| + VRegister vn_ext = (kDRegSize == vn_bits) ? vn.V8B() : vn.V16B();
|
| + VRegister vntmp_ext = (kDRegSize == vn_bits) ? vntmp.V8B() : vntmp.V16B();
|
| +
|
| + __ Mov(out, results);
|
| +
|
| + __ Mov(inputs_d_base, inputs_d);
|
| +
|
| + __ Mov(inputs_n_base, inputs_n);
|
| + __ Mov(inputs_n_last_vector,
|
| + inputs_n + vn_lane_bytes * (inputs_n_length - vn_lane_count));
|
| +
|
| + __ Ldr(vd, MemOperand(inputs_d_base));
|
| +
|
| + __ Ldr(vn, MemOperand(inputs_n_last_vector));
|
| +
|
| + __ Mov(index_n, 0);
|
| + __ Bind(&loop_n);
|
| +
|
| + __ Ldr(vntmp_single,
|
| + MemOperand(inputs_n_base, index_n, LSL, vn_lane_bytes_log2));
|
| + __ Ext(vn_ext, vn_ext, vntmp_ext, vn_lane_bytes);
|
| +
|
| + for (unsigned i = 0; i < inputs_imm1_length; i++) {
|
| + for (unsigned j = 0; j < inputs_imm2_length; j++) {
|
| + __ Mov(vres, vd);
|
| + (masm.*helper)(vres_helper, inputs_imm1[i], vn_helper, inputs_imm2[j]);
|
| + __ Str(vres, MemOperand(out, vd.SizeInBytes(), PostIndex));
|
| + }
|
| + }
|
| +
|
| + __ Add(index_n, index_n, 1);
|
| + __ Cmp(index_n, inputs_n_length);
|
| + __ B(lo, &loop_n);
|
| +
|
| + END();
|
| + RUN();
|
| + TEARDOWN();
|
| +}
|
| +
|
| +// Test NEON instructions. The inputs_*[] and expected[] arrays should be
|
| +// arrays of rawbit representation of input values. This ensures that
|
| +// exact bit comparisons can be performed.
|
| +template <typename Td, typename Tn>
|
| +void TestOpImmOpImmNEON(const char* name,
|
| + TestOpImmOpImmVdUpdateNEONHelper_t helper,
|
| + const Td inputs_d[], const int inputs_imm1[],
|
| + unsigned inputs_imm1_length, const Tn inputs_n[],
|
| + unsigned inputs_n_length, const int inputs_imm2[],
|
| + unsigned inputs_imm2_length, const Td expected[],
|
| + unsigned expected_length, VectorFormat vd_form,
|
| + VectorFormat vn_form) {
|
| + DCHECK_GT(inputs_n_length, 0U);
|
| + DCHECK_GT(inputs_imm1_length, 0U);
|
| + DCHECK_GT(inputs_imm2_length, 0U);
|
| +
|
| + const unsigned vd_lane_count = LaneCountFromFormat(vd_form);
|
| +
|
| + const unsigned results_length =
|
| + inputs_n_length * inputs_imm1_length * inputs_imm2_length;
|
| +
|
| + std::vector<Td> results(results_length * vd_lane_count, 0);
|
| + const unsigned lane_len_in_hex = MaxHexCharCount<Td, Tn>();
|
| +
|
| + TestOpImmOpImmNEON_Helper(
|
| + helper, reinterpret_cast<uintptr_t>(inputs_d), inputs_imm1,
|
| + inputs_imm1_length, reinterpret_cast<uintptr_t>(inputs_n),
|
| + inputs_n_length, inputs_imm2, inputs_imm2_length,
|
| + reinterpret_cast<uintptr_t>(results.data()), vd_form, vn_form);
|
| +
|
| + // Check the results.
|
| + CHECK(expected_length == results_length);
|
| + unsigned error_count = 0;
|
| + unsigned counted_length = 0;
|
| + const char* padding = " ";
|
| + DCHECK(strlen(padding) >= (lane_len_in_hex + 1));
|
| + for (unsigned n = 0; n < inputs_n_length; n++) {
|
| + for (unsigned imm1 = 0; imm1 < inputs_imm1_length; imm1++) {
|
| + for (unsigned imm2 = 0; imm2 < inputs_imm2_length; imm2++) {
|
| + bool error_in_vector = false;
|
| +
|
| + counted_length++;
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index =
|
| + (n * inputs_imm1_length * inputs_imm2_length * vd_lane_count) +
|
| + (imm1 * inputs_imm2_length * vd_lane_count) +
|
| + (imm2 * vd_lane_count) + lane;
|
| +
|
| + if (results[output_index] != expected[output_index]) {
|
| + error_in_vector = true;
|
| + break;
|
| + }
|
| + }
|
| +
|
| + if (error_in_vector && (++error_count <= kErrorReportLimit)) {
|
| + printf("%s\n", name);
|
| + printf(" Vd%.*s| Imm%.*s| Vn%.*s| Imm%.*s| Vd%.*s| Expected\n",
|
| + lane_len_in_hex + 1, padding, lane_len_in_hex, padding,
|
| + lane_len_in_hex + 1, padding, lane_len_in_hex, padding,
|
| + lane_len_in_hex + 1, padding);
|
| +
|
| + for (unsigned lane = 0; lane < vd_lane_count; lane++) {
|
| + unsigned output_index =
|
| + (n * inputs_imm1_length * inputs_imm2_length * vd_lane_count) +
|
| + (imm1 * inputs_imm2_length * vd_lane_count) +
|
| + (imm2 * vd_lane_count) + lane;
|
| + unsigned input_index_n =
|
| + (inputs_n_length - vd_lane_count + n + 1 + lane) %
|
| + inputs_n_length;
|
| + unsigned input_index_imm1 = imm1;
|
| + unsigned input_index_imm2 = imm2;
|
| +
|
| + printf(
|
| + "%c0x%0*" PRIx64 " | 0x%0*" PRIx64 " | 0x%0*" PRIx64
|
| + " "
|
| + "| 0x%0*" PRIx64 " | 0x%0*" PRIx64 " | 0x%0*" PRIx64 "\n",
|
| + results[output_index] != expected[output_index] ? '*' : ' ',
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_d[lane]),
|
| + lane_len_in_hex,
|
| + static_cast<uint64_t>(inputs_imm1[input_index_imm1]),
|
| + lane_len_in_hex, static_cast<uint64_t>(inputs_n[input_index_n]),
|
| + lane_len_in_hex,
|
| + static_cast<uint64_t>(inputs_imm2[input_index_imm2]),
|
| + lane_len_in_hex, static_cast<uint64_t>(results[output_index]),
|
| + lane_len_in_hex, static_cast<uint64_t>(expected[output_index]));
|
| + }
|
| + }
|
| + }
|
| + }
|
| + }
|
| + DCHECK_EQ(counted_length, expected_length);
|
| + if (error_count > kErrorReportLimit) {
|
| + printf("%u other errors follow.\n", error_count - kErrorReportLimit);
|
| + }
|
| + CHECK(error_count == 0);
|
| +}
|
| +
|
| +} // anonymous namespace
|
| +
|
| +// ==== NEON Tests. ====
|
| +
|
| +// clang-format off
|
| +
|
| +#define CALL_TEST_NEON_HELPER_1Op(mnemonic, vdform, vnform, input_n) \
|
| + Test1OpNEON(STRINGIFY(mnemonic) "_" STRINGIFY(vdform), \
|
| + &MacroAssembler::mnemonic, input_n, \
|
| + (sizeof(input_n) / sizeof(input_n[0])), \
|
| + kExpected_NEON_##mnemonic##_##vdform, \
|
| + kExpectedCount_NEON_##mnemonic##_##vdform, kFormat##vdform, \
|
| + kFormat##vnform)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_1OpAcross(mnemonic, vdform, vnform, input_n) \
|
| + Test1OpAcrossNEON( \
|
| + STRINGIFY(mnemonic) "_" STRINGIFY(vdform) "_" STRINGIFY(vnform), \
|
| + &MacroAssembler::mnemonic, input_n, \
|
| + (sizeof(input_n) / sizeof(input_n[0])), \
|
| + kExpected_NEON_##mnemonic##_##vdform##_##vnform, \
|
| + kExpectedCount_NEON_##mnemonic##_##vdform##_##vnform, kFormat##vdform, \
|
| + kFormat##vnform)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_2Op(mnemonic, vdform, vnform, vmform, input_d, \
|
| + input_n, input_m) \
|
| + Test2OpNEON(STRINGIFY(mnemonic) "_" STRINGIFY(vdform), \
|
| + &MacroAssembler::mnemonic, input_d, input_n, \
|
| + (sizeof(input_n) / sizeof(input_n[0])), input_m, \
|
| + (sizeof(input_m) / sizeof(input_m[0])), \
|
| + kExpected_NEON_##mnemonic##_##vdform, \
|
| + kExpectedCount_NEON_##mnemonic##_##vdform, kFormat##vdform, \
|
| + kFormat##vnform, kFormat##vmform)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_2OpImm(mnemonic, vdform, vnform, input_n, \
|
| + input_m) \
|
| + Test2OpImmNEON(STRINGIFY(mnemonic) "_" STRINGIFY(vdform) "_2OPIMM", \
|
| + &MacroAssembler::mnemonic, input_n, \
|
| + (sizeof(input_n) / sizeof(input_n[0])), input_m, \
|
| + (sizeof(input_m) / sizeof(input_m[0])), \
|
| + kExpected_NEON_##mnemonic##_##vdform##_2OPIMM, \
|
| + kExpectedCount_NEON_##mnemonic##_##vdform##_2OPIMM, \
|
| + kFormat##vdform, kFormat##vnform)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_ByElement(mnemonic, vdform, vnform, vmform, \
|
| + input_d, input_n, input_m, indices) \
|
| + TestByElementNEON( \
|
| + STRINGIFY(mnemonic) "_" STRINGIFY(vdform) "_" STRINGIFY( \
|
| + vnform) "_" STRINGIFY(vmform), \
|
| + &MacroAssembler::mnemonic, input_d, input_n, \
|
| + (sizeof(input_n) / sizeof(input_n[0])), input_m, \
|
| + (sizeof(input_m) / sizeof(input_m[0])), indices, \
|
| + (sizeof(indices) / sizeof(indices[0])), \
|
| + kExpected_NEON_##mnemonic##_##vdform##_##vnform##_##vmform, \
|
| + kExpectedCount_NEON_##mnemonic##_##vdform##_##vnform##_##vmform, \
|
| + kFormat##vdform, kFormat##vnform, kFormat##vmform)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_OpImmOpImm(helper, mnemonic, vdform, vnform, \
|
| + input_d, input_imm1, input_n, \
|
| + input_imm2) \
|
| + TestOpImmOpImmNEON(STRINGIFY(mnemonic) "_" STRINGIFY(vdform), helper, \
|
| + input_d, input_imm1, \
|
| + (sizeof(input_imm1) / sizeof(input_imm1[0])), input_n, \
|
| + (sizeof(input_n) / sizeof(input_n[0])), input_imm2, \
|
| + (sizeof(input_imm2) / sizeof(input_imm2[0])), \
|
| + kExpected_NEON_##mnemonic##_##vdform, \
|
| + kExpectedCount_NEON_##mnemonic##_##vdform, \
|
| + kFormat##vdform, kFormat##vnform)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_2SAME(mnemonic, variant, input) \
|
| + CALL_TEST_NEON_HELPER_1Op(mnemonic, variant, variant, input)
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_8B_16B(mnemonic, input) \
|
| + SIMTEST(mnemonic##_8B) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 8B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_16B) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 16B, kInput8bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_4H_8H(mnemonic, input) \
|
| + SIMTEST(mnemonic##_4H) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 4H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_8H) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 8H, kInput16bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_2S_4S(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 2S, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 4S, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_BH(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_8B_16B(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_4H_8H(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_NO2D(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_BH(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_2S_4S(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_NO2D(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 2D, kInput64bits##input); \
|
| + }
|
| +#define DEFINE_TEST_NEON_2SAME_SD(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_2S_4S(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 2D, kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_FP(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 2S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 4S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, 2D, kInputDouble##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_FP_SCALAR(mnemonic, input) \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, D, kInputDouble##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_SCALAR_B(mnemonic, input) \
|
| + SIMTEST(mnemonic##_B) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, B, kInput8bits##input); \
|
| + }
|
| +#define DEFINE_TEST_NEON_2SAME_SCALAR_H(mnemonic, input) \
|
| + SIMTEST(mnemonic##_H) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, H, kInput16bits##input); \
|
| + }
|
| +#define DEFINE_TEST_NEON_2SAME_SCALAR_S(mnemonic, input) \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, S, kInput32bits##input); \
|
| + }
|
| +#define DEFINE_TEST_NEON_2SAME_SCALAR_D(mnemonic, input) \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_2SAME(mnemonic, D, kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_SCALAR(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_SCALAR_B(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_SCALAR_H(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_SCALAR_S(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_SCALAR_D(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_2SAME_SCALAR_SD(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_SCALAR_S(mnemonic, input) \
|
| + DEFINE_TEST_NEON_2SAME_SCALAR_D(mnemonic, input)
|
| +
|
| +#define CALL_TEST_NEON_HELPER_ACROSS(mnemonic, vd_form, vn_form, input_n) \
|
| + CALL_TEST_NEON_HELPER_1OpAcross(mnemonic, vd_form, vn_form, input_n)
|
| +
|
| +#define DEFINE_TEST_NEON_ACROSS(mnemonic, input) \
|
| + SIMTEST(mnemonic##_B_8B) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, B, 8B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_B_16B) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, B, 16B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_H_4H) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, H, 4H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_H_8H) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, H, 8H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_S_4S) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, S, 4S, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_ACROSS_LONG(mnemonic, input) \
|
| + SIMTEST(mnemonic##_H_8B) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, H, 8B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_H_16B) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, H, 16B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_S_4H) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, S, 4H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_S_8H) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, S, 8H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_D_4S) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, D, 4S, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_ACROSS_FP(mnemonic, input) \
|
| + SIMTEST(mnemonic##_S_4S) { \
|
| + CALL_TEST_NEON_HELPER_ACROSS(mnemonic, S, 4S, kInputFloat##input); \
|
| + }
|
| +
|
| +#define CALL_TEST_NEON_HELPER_2DIFF(mnemonic, vdform, vnform, input_n) \
|
| + CALL_TEST_NEON_HELPER_1Op(mnemonic, vdform, vnform, input_n)
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_LONG(mnemonic, input) \
|
| + SIMTEST(mnemonic##_4H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 4H, 8B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_8H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 8H, 16B, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 2S, 4H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 4S, 8H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_1D) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 1D, 2S, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 2D, 4S, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_NARROW(mnemonic, input) \
|
| + SIMTEST(mnemonic##_8B) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 8B, 8H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 4H, 4S, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 2S, 2D, kInput64bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_16B) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 16B, 8H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 8H, 4S, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 4S, 2D, kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_FP_LONG(mnemonic, input) \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 4S, 4H, kInputFloat16##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 2D, 2S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 4S, 8H, kInputFloat16##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_2D) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 2D, 4S, kInputFloat##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_FP_NARROW(mnemonic, input) \
|
| + SIMTEST(mnemonic##_4H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 4H, 4S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 2S, 2D, kInputDouble##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 8H, 4S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 4S, 2D, kInputDouble##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_FP_NARROW_2S(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, 2S, 2D, kInputDouble##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic##2, 4S, 2D, kInputDouble##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_SCALAR_NARROW(mnemonic, input) \
|
| + SIMTEST(mnemonic##_B) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, B, H, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_H) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, H, S, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, S, D, kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2DIFF_FP_SCALAR_SD(mnemonic, input) \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, S, 2S, kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_2DIFF(mnemonic, D, 2D, kInputDouble##input); \
|
| + }
|
| +
|
| +#define CALL_TEST_NEON_HELPER_3SAME(mnemonic, variant, input_d, input_nm) \
|
| + { \
|
| + CALL_TEST_NEON_HELPER_2Op(mnemonic, variant, variant, variant, input_d, \
|
| + input_nm, input_nm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_8B_16B(mnemonic, input) \
|
| + SIMTEST(mnemonic##_8B) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 8B, kInput8bitsAccDestination, \
|
| + kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_16B) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 16B, kInput8bitsAccDestination, \
|
| + kInput8bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_HS(mnemonic, input) \
|
| + SIMTEST(mnemonic##_4H) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 4H, kInput16bitsAccDestination, \
|
| + kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_8H) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 8H, kInput16bitsAccDestination, \
|
| + kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 2S, kInput32bitsAccDestination, \
|
| + kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 4S, kInput32bitsAccDestination, \
|
| + kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_NO2D(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3SAME_8B_16B(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3SAME_HS(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3SAME_NO2D(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 2D, kInput64bitsAccDestination, \
|
| + kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_FP(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 2S, kInputFloatAccDestination, \
|
| + kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 4S, kInputFloatAccDestination, \
|
| + kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, 2D, kInputDoubleAccDestination, \
|
| + kInputDouble##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_SCALAR_D(mnemonic, input) \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, D, kInput64bitsAccDestination, \
|
| + kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_SCALAR_HS(mnemonic, input) \
|
| + SIMTEST(mnemonic##_H) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, H, kInput16bitsAccDestination, \
|
| + kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, S, kInput32bitsAccDestination, \
|
| + kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_SCALAR(mnemonic, input) \
|
| + SIMTEST(mnemonic##_B) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, B, kInput8bitsAccDestination, \
|
| + kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_H) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, H, kInput16bitsAccDestination, \
|
| + kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, S, kInput32bitsAccDestination, \
|
| + kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, D, kInput64bitsAccDestination, \
|
| + kInput64bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3SAME_FP_SCALAR(mnemonic, input) \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, S, kInputFloatAccDestination, \
|
| + kInputFloat##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_3SAME(mnemonic, D, kInputDoubleAccDestination, \
|
| + kInputDouble##input); \
|
| + }
|
| +
|
| +#define CALL_TEST_NEON_HELPER_3DIFF(mnemonic, vdform, vnform, vmform, input_d, \
|
| + input_n, input_m) \
|
| + { \
|
| + CALL_TEST_NEON_HELPER_2Op(mnemonic, vdform, vnform, vmform, input_d, \
|
| + input_n, input_m); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_LONG_8H(mnemonic, input) \
|
| + SIMTEST(mnemonic##_8H) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 8H, 8B, 8B, \
|
| + kInput16bitsAccDestination, \
|
| + kInput8bits##input, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 8H, 16B, 16B, \
|
| + kInput16bitsAccDestination, \
|
| + kInput8bits##input, kInput8bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_LONG_4S(mnemonic, input) \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 4S, 4H, 4H, \
|
| + kInput32bitsAccDestination, \
|
| + kInput16bits##input, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 4S, 8H, 8H, \
|
| + kInput32bitsAccDestination, \
|
| + kInput16bits##input, kInput16bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_LONG_2D(mnemonic, input) \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 2D, 2S, 2S, \
|
| + kInput64bitsAccDestination, \
|
| + kInput32bits##input, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_2D) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 2D, 4S, 4S, \
|
| + kInput64bitsAccDestination, \
|
| + kInput32bits##input, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_LONG_SD(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_LONG_4S(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_LONG_2D(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_LONG(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_LONG_8H(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_LONG_4S(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_LONG_2D(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_S(mnemonic, input) \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, S, H, H, kInput32bitsAccDestination, \
|
| + kInput16bits##input, kInput16bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_D(mnemonic, input) \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, D, S, S, kInput64bitsAccDestination, \
|
| + kInput32bits##input, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_SD(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_S(mnemonic, input) \
|
| + DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_D(mnemonic, input)
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_WIDE(mnemonic, input) \
|
| + SIMTEST(mnemonic##_8H) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 8H, 8H, 8B, \
|
| + kInput16bitsAccDestination, \
|
| + kInput16bits##input, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 4S, 4S, 4H, \
|
| + kInput32bitsAccDestination, \
|
| + kInput32bits##input, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 2D, 2D, 2S, \
|
| + kInput64bitsAccDestination, \
|
| + kInput64bits##input, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 8H, 8H, 16B, \
|
| + kInput16bitsAccDestination, \
|
| + kInput16bits##input, kInput8bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 4S, 4S, 8H, \
|
| + kInput32bitsAccDestination, \
|
| + kInput32bits##input, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_2D) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 2D, 2D, 4S, \
|
| + kInput64bitsAccDestination, \
|
| + kInput64bits##input, kInput32bits##input); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_3DIFF_NARROW(mnemonic, input) \
|
| + SIMTEST(mnemonic##_8B) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 8B, 8H, 8H, \
|
| + kInput8bitsAccDestination, \
|
| + kInput16bits##input, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_4H) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 4H, 4S, 4S, \
|
| + kInput16bitsAccDestination, \
|
| + kInput32bits##input, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic, 2S, 2D, 2D, \
|
| + kInput32bitsAccDestination, \
|
| + kInput64bits##input, kInput64bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_16B) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 16B, 8H, 8H, \
|
| + kInput8bitsAccDestination, \
|
| + kInput16bits##input, kInput16bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 8H, 4S, 4S, \
|
| + kInput16bitsAccDestination, \
|
| + kInput32bits##input, kInput32bits##input); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S) { \
|
| + CALL_TEST_NEON_HELPER_3DIFF(mnemonic##2, 4S, 2D, 2D, \
|
| + kInput32bitsAccDestination, \
|
| + kInput64bits##input, kInput64bits##input); \
|
| + }
|
| +
|
| +#define CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, vdform, vnform, input_n, \
|
| + input_imm) \
|
| + { \
|
| + CALL_TEST_NEON_HELPER_2OpImm(mnemonic, vdform, vnform, input_n, \
|
| + input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_8B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 8B, 8B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_16B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 16B, 16B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4H, 4H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_8H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 8H, 8H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2S, 2S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4S, 4S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2D, 2D, kInput64bits##input, \
|
| + kInput64bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_COPY(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_8B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 8B, B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_16B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 16B, B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4H, H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_8H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 8H, H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2S, S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4S, S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2D, D, kInput64bits##input, \
|
| + kInput64bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_NARROW(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_8B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 8B, 8H, kInput16bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4H, 4S, kInput32bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2S, 2D, kInput64bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##2_16B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic##2, 16B, 8H, kInput16bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic##2, 8H, 4S, kInput32bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic##2, 4S, 2D, kInput64bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, B, H, kInput16bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, H, S, kInput32bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, S, D, kInput64bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_FCMP_ZERO(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_2S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2S, 2S, kInputFloat##Basic, \
|
| + kInputDoubleImm##input_imm) \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4S, 4S, kInputFloat##input, \
|
| + kInputDoubleImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2D, 2D, kInputDouble##input, \
|
| + kInputDoubleImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_FP(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_2S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2S, 2S, kInputFloat##Basic, \
|
| + kInput32bitsImm##input_imm) \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4S, 4S, kInputFloat##input, \
|
| + kInput32bitsImm##input_imm) \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2D, 2D, kInputDouble##input, \
|
| + kInput64bitsImm##input_imm) \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_FP_SCALAR(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, S, S, kInputFloat##Basic, \
|
| + kInput32bitsImm##input_imm) \
|
| + } \
|
| + SIMTEST(mnemonic##_D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, D, D, kInputDouble##input, \
|
| + kInput64bitsImm##input_imm) \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_SD(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_2S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2S, 2S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4S, 4S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2D, 2D, kInput64bits##input, \
|
| + kInput64bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_SCALAR_D(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, D, D, kInput64bits##input, \
|
| + kInput64bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_SCALAR_SD(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, S, S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + DEFINE_TEST_NEON_2OPIMM_SCALAR_D(mnemonic, input, input_imm)
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_D(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, D, D, kInputDouble##input, \
|
| + kInputDoubleImm##input_imm); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_SD(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, S, S, kInputFloat##input, \
|
| + kInputDoubleImm##input_imm); \
|
| + } \
|
| + DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_D(mnemonic, input, input_imm)
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_SCALAR(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_B_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, B, B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, H, H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + DEFINE_TEST_NEON_2OPIMM_SCALAR_SD(mnemonic, input, input_imm)
|
| +
|
| +#define DEFINE_TEST_NEON_2OPIMM_LONG(mnemonic, input, input_imm) \
|
| + SIMTEST(mnemonic##_8H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 8H, 8B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 4S, 4H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic, 2D, 2S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##2_8H_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic##2, 8H, 16B, kInput8bits##input, \
|
| + kInput8bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic##2, 4S, 8H, kInput16bits##input, \
|
| + kInput16bitsImm##input_imm); \
|
| + } \
|
| + SIMTEST(mnemonic##2_2D_2OPIMM) { \
|
| + CALL_TEST_NEON_HELPER_2OPIMM(mnemonic##2, 2D, 4S, kInput32bits##input, \
|
| + kInput32bitsImm##input_imm); \
|
| + }
|
| +
|
| +#define CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, vdform, vnform, vmform, \
|
| + input_d, input_n, input_m, indices) \
|
| + { \
|
| + CALL_TEST_NEON_HELPER_ByElement(mnemonic, vdform, vnform, vmform, input_d, \
|
| + input_n, input_m, indices); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_BYELEMENT(mnemonic, input_d, input_n, input_m) \
|
| + SIMTEST(mnemonic##_4H_4H_H) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 4H, 4H, H, kInput16bits##input_d, kInput16bits##input_n, \
|
| + kInput16bits##input_m, kInputHIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_8H_8H_H) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 8H, 8H, H, kInput16bits##input_d, kInput16bits##input_n, \
|
| + kInput16bits##input_m, kInputHIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_2S_2S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 2S, 2S, S, kInput32bits##input_d, kInput32bits##input_n, \
|
| + kInput32bits##input_m, kInputSIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_4S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 4S, 4S, S, kInput32bits##input_d, kInput32bits##input_n, \
|
| + kInput32bits##input_m, kInputSIndices); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_BYELEMENT_SCALAR(mnemonic, input_d, input_n, input_m) \
|
| + SIMTEST(mnemonic##_H_H_H) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, H, H, H, kInput16bits##input_d, \
|
| + kInput16bits##input_n, \
|
| + kInput16bits##input_m, kInputHIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_S_S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, S, S, S, kInput32bits##input_d, \
|
| + kInput32bits##input_n, \
|
| + kInput32bits##input_m, kInputSIndices); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_FP_BYELEMENT(mnemonic, input_d, input_n, input_m) \
|
| + SIMTEST(mnemonic##_2S_2S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, 2S, 2S, S, kInputFloat##input_d, \
|
| + kInputFloat##input_n, \
|
| + kInputFloat##input_m, kInputSIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_4S_4S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, 4S, 4S, S, kInputFloat##input_d, \
|
| + kInputFloat##input_n, \
|
| + kInputFloat##input_m, kInputSIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2D_D) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 2D, 2D, D, kInputDouble##input_d, kInputDouble##input_n, \
|
| + kInputDouble##input_m, kInputDIndices); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_FP_BYELEMENT_SCALAR(mnemonic, inp_d, inp_n, inp_m) \
|
| + SIMTEST(mnemonic##_S_S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, S, S, S, kInputFloat##inp_d, \
|
| + kInputFloat##inp_n, kInputFloat##inp_m, \
|
| + kInputSIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_D_D_D) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, D, D, D, kInputDouble##inp_d, \
|
| + kInputDouble##inp_n, kInputDouble##inp_m, \
|
| + kInputDIndices); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_BYELEMENT_DIFF(mnemonic, input_d, input_n, input_m) \
|
| + SIMTEST(mnemonic##_4S_4H_H) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 4S, 4H, H, kInput32bits##input_d, kInput16bits##input_n, \
|
| + kInput16bits##input_m, kInputHIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##2_4S_8H_H) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic##2, 4S, 8H, H, kInput32bits##input_d, kInput16bits##input_n, \
|
| + kInput16bits##input_m, kInputHIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_2D_2S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic, 2D, 2S, S, kInput64bits##input_d, kInput32bits##input_n, \
|
| + kInput32bits##input_m, kInputSIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##2_2D_4S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT( \
|
| + mnemonic##2, 2D, 4S, S, kInput64bits##input_d, kInput32bits##input_n, \
|
| + kInput32bits##input_m, kInputSIndices); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_BYELEMENT_DIFF_SCALAR(mnemonic, input_d, input_n, \
|
| + input_m) \
|
| + SIMTEST(mnemonic##_S_H_H) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, S, H, H, kInput32bits##input_d, \
|
| + kInput16bits##input_n, \
|
| + kInput16bits##input_m, kInputHIndices); \
|
| + } \
|
| + SIMTEST(mnemonic##_D_S_S) { \
|
| + CALL_TEST_NEON_HELPER_BYELEMENT(mnemonic, D, S, S, kInput64bits##input_d, \
|
| + kInput32bits##input_n, \
|
| + kInput32bits##input_m, kInputSIndices); \
|
| + }
|
| +
|
| +#define CALL_TEST_NEON_HELPER_2OP2IMM(mnemonic, variant, input_d, input_imm1, \
|
| + input_n, input_imm2) \
|
| + { \
|
| + CALL_TEST_NEON_HELPER_OpImmOpImm(&MacroAssembler::mnemonic, mnemonic, \
|
| + variant, variant, input_d, input_imm1, \
|
| + input_n, input_imm2); \
|
| + }
|
| +
|
| +#define DEFINE_TEST_NEON_2OP2IMM(mnemonic, input_d, input_imm1, input_n, \
|
| + input_imm2) \
|
| + SIMTEST(mnemonic##_B) { \
|
| + CALL_TEST_NEON_HELPER_2OP2IMM( \
|
| + mnemonic, 16B, kInput8bits##input_d, kInput8bitsImm##input_imm1, \
|
| + kInput8bits##input_n, kInput8bitsImm##input_imm2); \
|
| + } \
|
| + SIMTEST(mnemonic##_H) { \
|
| + CALL_TEST_NEON_HELPER_2OP2IMM( \
|
| + mnemonic, 8H, kInput16bits##input_d, kInput16bitsImm##input_imm1, \
|
| + kInput16bits##input_n, kInput16bitsImm##input_imm2); \
|
| + } \
|
| + SIMTEST(mnemonic##_S) { \
|
| + CALL_TEST_NEON_HELPER_2OP2IMM( \
|
| + mnemonic, 4S, kInput32bits##input_d, kInput32bitsImm##input_imm1, \
|
| + kInput32bits##input_n, kInput32bitsImm##input_imm2); \
|
| + } \
|
| + SIMTEST(mnemonic##_D) { \
|
| + CALL_TEST_NEON_HELPER_2OP2IMM( \
|
| + mnemonic, 2D, kInput64bits##input_d, kInput64bitsImm##input_imm1, \
|
| + kInput64bits##input_n, kInput64bitsImm##input_imm2); \
|
| + }
|
| +
|
| +// clang-format on
|
| +
|
| +// Advanced SIMD copy.
|
| +DEFINE_TEST_NEON_2OP2IMM(ins, Basic, LaneCountFromZero, Basic,
|
| + LaneCountFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_COPY(dup, Basic, LaneCountFromZero)
|
| +
|
| +// Advanced SIMD scalar copy.
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR(dup, Basic, LaneCountFromZero)
|
| +
|
| +// Advanced SIMD three same.
|
| +DEFINE_TEST_NEON_3SAME_NO2D(shadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME(sqadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(srhadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(shsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME(sqsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME(cmgt, Basic)
|
| +DEFINE_TEST_NEON_3SAME(cmge, Basic)
|
| +DEFINE_TEST_NEON_3SAME(sshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME(sqshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME(srshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME(sqrshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(smax, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(smin, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(sabd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(saba, Basic)
|
| +DEFINE_TEST_NEON_3SAME(add, Basic)
|
| +DEFINE_TEST_NEON_3SAME(cmtst, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(mla, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(mul, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(smaxp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(sminp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_HS(sqdmulh, Basic)
|
| +DEFINE_TEST_NEON_3SAME(addp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmaxnm, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmla, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmulx, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fcmeq, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmax, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(frecps, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(and_, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(bic, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fminnm, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmls, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmin, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(frsqrts, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(orr, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(orn, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(uhadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME(uqadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(urhadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(uhsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME(uqsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME(cmhi, Basic)
|
| +DEFINE_TEST_NEON_3SAME(cmhs, Basic)
|
| +DEFINE_TEST_NEON_3SAME(ushl, Basic)
|
| +DEFINE_TEST_NEON_3SAME(uqshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME(urshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME(uqrshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(umax, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(umin, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(uabd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(uaba, Basic)
|
| +DEFINE_TEST_NEON_3SAME(sub, Basic)
|
| +DEFINE_TEST_NEON_3SAME(cmeq, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(mls, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(pmul, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(uminp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_NO2D(umaxp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_HS(sqrdmulh, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmaxnmp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(faddp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmul, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fcmge, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(facge, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fmaxp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fdiv, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(eor, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(bsl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fminnmp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fabd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fcmgt, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(facgt, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP(fminp, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(bit, Basic)
|
| +DEFINE_TEST_NEON_3SAME_8B_16B(bif, Basic)
|
| +
|
| +// Advanced SIMD scalar three same.
|
| +DEFINE_TEST_NEON_3SAME_SCALAR(sqadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR(sqsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(cmgt, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(cmge, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(sshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR(sqshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(srshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR(sqrshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(add, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(cmtst, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_HS(sqdmulh, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(fmulx, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(fcmeq, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(frecps, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(frsqrts, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(uqadd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(uqsub, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(cmhi, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(cmhs, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(ushl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR(uqshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(urshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR(uqrshl, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(sub, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_D(cmeq, Basic)
|
| +DEFINE_TEST_NEON_3SAME_SCALAR_HS(sqrdmulh, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(fcmge, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(facge, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(fabd, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(fcmgt, Basic)
|
| +DEFINE_TEST_NEON_3SAME_FP_SCALAR(facgt, Basic)
|
| +
|
| +// Advanced SIMD three different.
|
| +DEFINE_TEST_NEON_3DIFF_LONG(saddl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_WIDE(saddw, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(ssubl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_WIDE(ssubw, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_NARROW(addhn, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(sabal, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_NARROW(subhn, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(sabdl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(smlal, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG_SD(sqdmlal, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(smlsl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG_SD(sqdmlsl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(smull, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG_SD(sqdmull, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG_8H(pmull, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(uaddl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_WIDE(uaddw, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(usubl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_WIDE(usubw, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_NARROW(raddhn, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(uabal, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_NARROW(rsubhn, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(uabdl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(umlal, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(umlsl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_LONG(umull, Basic)
|
| +
|
| +// Advanced SIMD scalar three different.
|
| +DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_SD(sqdmlal, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_SD(sqdmlsl, Basic)
|
| +DEFINE_TEST_NEON_3DIFF_SCALAR_LONG_SD(sqdmull, Basic)
|
| +
|
| +// Advanced SIMD scalar pairwise.
|
| +SIMTEST(addp_SCALAR) {
|
| + CALL_TEST_NEON_HELPER_2DIFF(addp, D, 2D, kInput64bitsBasic);
|
| +}
|
| +DEFINE_TEST_NEON_2DIFF_FP_SCALAR_SD(fmaxnmp, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_FP_SCALAR_SD(faddp, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_FP_SCALAR_SD(fmaxp, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_FP_SCALAR_SD(fminnmp, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_FP_SCALAR_SD(fminp, Basic)
|
| +
|
| +// Advanced SIMD shift by immediate.
|
| +DEFINE_TEST_NEON_2OPIMM(sshr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(ssra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(srshr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(srsra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(shl, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM(sqshl, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(shrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(rshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(sqshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(sqrshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_LONG(sshll, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SD(scvtf, FixedPointConversions,
|
| + TypeWidthFromZeroToWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_FP(fcvtzs, Conversions, TypeWidthFromZeroToWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(ushr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(usra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(urshr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(ursra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(sri, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM(sli, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM(sqshlu, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM(uqshl, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(sqshrun, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(sqrshrun, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(uqshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_NARROW(uqrshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_LONG(ushll, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SD(ucvtf, FixedPointConversions,
|
| + TypeWidthFromZeroToWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_FP(fcvtzu, Conversions, TypeWidthFromZeroToWidth)
|
| +
|
| +// Advanced SIMD scalar shift by immediate..
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(sshr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(ssra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(srshr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(srsra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(shl, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR(sqshl, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(sqshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(sqrshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_SD(scvtf, FixedPointConversions,
|
| + TypeWidthFromZeroToWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR(fcvtzs, Conversions, TypeWidthFromZeroToWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(ushr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(usra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(urshr, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(ursra, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(sri, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(sli, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR(sqshlu, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR(uqshl, Basic, TypeWidthFromZero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(sqshrun, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(sqrshrun, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(uqshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_NARROW(uqrshrn, Basic, TypeWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_SD(ucvtf, FixedPointConversions,
|
| + TypeWidthFromZeroToWidth)
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR(fcvtzu, Conversions, TypeWidthFromZeroToWidth)
|
| +
|
| +// Advanced SIMD two-register miscellaneous.
|
| +DEFINE_TEST_NEON_2SAME_NO2D(rev64, Basic)
|
| +DEFINE_TEST_NEON_2SAME_8B_16B(rev16, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_LONG(saddlp, Basic)
|
| +DEFINE_TEST_NEON_2SAME(suqadd, Basic)
|
| +DEFINE_TEST_NEON_2SAME_NO2D(cls, Basic)
|
| +DEFINE_TEST_NEON_2SAME_8B_16B(cnt, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_LONG(sadalp, Basic)
|
| +DEFINE_TEST_NEON_2SAME(sqabs, Basic)
|
| +DEFINE_TEST_NEON_2OPIMM(cmgt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM(cmeq, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM(cmlt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME(abs, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_NARROW(xtn, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_NARROW(sqxtn, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_FP_NARROW(fcvtn, Conversions)
|
| +DEFINE_TEST_NEON_2DIFF_FP_LONG(fcvtl, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(frintn, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(frintm, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtns, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtms, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtas, Conversions)
|
| +// SCVTF (vector, integer) covered by SCVTF(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2OPIMM_FCMP_ZERO(fcmgt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_FCMP_ZERO(fcmeq, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_FCMP_ZERO(fcmlt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME_FP(fabs, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP(frintp, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(frintz, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtps, Conversions)
|
| +// FCVTZS(vector, integer) covered by FCVTZS(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2SAME_2S_4S(urecpe, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP(frecpe, Basic)
|
| +DEFINE_TEST_NEON_2SAME_BH(rev32, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_LONG(uaddlp, Basic)
|
| +DEFINE_TEST_NEON_2SAME(usqadd, Basic)
|
| +DEFINE_TEST_NEON_2SAME_NO2D(clz, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_LONG(uadalp, Basic)
|
| +DEFINE_TEST_NEON_2SAME(sqneg, Basic)
|
| +DEFINE_TEST_NEON_2OPIMM(cmge, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM(cmle, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME(neg, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_NARROW(sqxtun, Basic)
|
| +DEFINE_TEST_NEON_2OPIMM_LONG(shll, Basic, SHLL)
|
| +DEFINE_TEST_NEON_2DIFF_NARROW(uqxtn, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_FP_NARROW_2S(fcvtxn, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(frinta, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(frintx, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtnu, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtmu, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtau, Conversions)
|
| +// UCVTF (vector, integer) covered by UCVTF(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2SAME_8B_16B(not_, Basic)
|
| +DEFINE_TEST_NEON_2SAME_8B_16B(rbit, Basic)
|
| +DEFINE_TEST_NEON_2OPIMM_FCMP_ZERO(fcmge, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_FCMP_ZERO(fcmle, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME_FP(fneg, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP(frinti, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP(fcvtpu, Conversions)
|
| +// FCVTZU(vector, integer) covered by FCVTZU(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2SAME_2S_4S(ursqrte, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP(frsqrte, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP(fsqrt, Basic)
|
| +
|
| +// Advanced SIMD scalar two-register miscellaneous.
|
| +DEFINE_TEST_NEON_2SAME_SCALAR(suqadd, Basic)
|
| +DEFINE_TEST_NEON_2SAME_SCALAR(sqabs, Basic)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(cmgt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(cmeq, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(cmlt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME_SCALAR_D(abs, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_SCALAR_NARROW(sqxtn, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtns, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtms, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtas, Conversions)
|
| +// SCVTF (vector, integer) covered by SCVTF(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_SD(fcmgt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_SD(fcmeq, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_SD(fcmlt, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtps, Conversions)
|
| +// FCVTZS(vector, integer) covered by FCVTZS(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(frecpe, Basic)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(frecpx, Basic)
|
| +DEFINE_TEST_NEON_2SAME_SCALAR(usqadd, Basic)
|
| +DEFINE_TEST_NEON_2SAME_SCALAR(sqneg, Basic)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(cmge, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_SCALAR_D(cmle, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME_SCALAR_D(neg, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_SCALAR_NARROW(sqxtun, Basic)
|
| +DEFINE_TEST_NEON_2DIFF_SCALAR_NARROW(uqxtn, Basic)
|
| +SIMTEST(fcvtxn_SCALAR) {
|
| + CALL_TEST_NEON_HELPER_2DIFF(fcvtxn, S, D, kInputDoubleConversions);
|
| +}
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtnu, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtmu, Conversions)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtau, Conversions)
|
| +// UCVTF (vector, integer) covered by UCVTF(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_SD(fcmge, Basic, Zero)
|
| +DEFINE_TEST_NEON_2OPIMM_FP_SCALAR_SD(fcmle, Basic, Zero)
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(fcvtpu, Conversions)
|
| +// FCVTZU(vector, integer) covered by FCVTZU(vector, fixed point) with fbits 0.
|
| +DEFINE_TEST_NEON_2SAME_FP_SCALAR(frsqrte, Basic)
|
| +
|
| +// Advanced SIMD across lanes.
|
| +DEFINE_TEST_NEON_ACROSS_LONG(saddlv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS(smaxv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS(sminv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS(addv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS_LONG(uaddlv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS(umaxv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS(uminv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS_FP(fmaxnmv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS_FP(fmaxv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS_FP(fminnmv, Basic)
|
| +DEFINE_TEST_NEON_ACROSS_FP(fminv, Basic)
|
| +
|
| +// Advanced SIMD permute.
|
| +DEFINE_TEST_NEON_3SAME(uzp1, Basic)
|
| +DEFINE_TEST_NEON_3SAME(trn1, Basic)
|
| +DEFINE_TEST_NEON_3SAME(zip1, Basic)
|
| +DEFINE_TEST_NEON_3SAME(uzp2, Basic)
|
| +DEFINE_TEST_NEON_3SAME(trn2, Basic)
|
| +DEFINE_TEST_NEON_3SAME(zip2, Basic)
|
| +
|
| +// Advanced SIMD vector x indexed element.
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(smlal, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(sqdmlal, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(smlsl, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(sqdmlsl, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT(mul, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(smull, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(sqdmull, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT(sqdmulh, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT(sqrdmulh, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT(fmla, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT(fmls, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT(fmul, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT(mla, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(umlal, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT(mls, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(umlsl, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF(umull, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT(fmulx, Basic, Basic, Basic)
|
| +
|
| +// Advanced SIMD scalar x indexed element.
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF_SCALAR(sqdmlal, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF_SCALAR(sqdmlsl, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_DIFF_SCALAR(sqdmull, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_SCALAR(sqdmulh, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_BYELEMENT_SCALAR(sqrdmulh, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT_SCALAR(fmla, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT_SCALAR(fmls, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT_SCALAR(fmul, Basic, Basic, Basic)
|
| +DEFINE_TEST_NEON_FP_BYELEMENT_SCALAR(fmulx, Basic, Basic, Basic)
|
|
|