Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(383)

Unified Diff: test/cctest/test-assembler-arm64.cc

Issue 430503007: Rename ASSERT* to DCHECK*. (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: REBASE and fixes Created 6 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « test/cctest/test-assembler-arm.cc ('k') | test/cctest/test-disasm-arm64.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: test/cctest/test-assembler-arm64.cc
diff --git a/test/cctest/test-assembler-arm64.cc b/test/cctest/test-assembler-arm64.cc
index 1e1b99e37494d5135c0e69f23972c61088260693..818392082e7adc73e2ee7ef9e89d0835874b11f4 100644
--- a/test/cctest/test-assembler-arm64.cc
+++ b/test/cctest/test-assembler-arm64.cc
@@ -58,7 +58,7 @@ using namespace v8::internal;
//
// RUN();
//
-// ASSERT_EQUAL_64(1, x0);
+// CHECK_EQUAL_64(1, x0);
//
// TEARDOWN();
// }
@@ -74,22 +74,22 @@ using namespace v8::internal;
//
// We provide some helper assert to handle common cases:
//
-// ASSERT_EQUAL_32(int32_t, int_32t)
-// ASSERT_EQUAL_FP32(float, float)
-// ASSERT_EQUAL_32(int32_t, W register)
-// ASSERT_EQUAL_FP32(float, S register)
-// ASSERT_EQUAL_64(int64_t, int_64t)
-// ASSERT_EQUAL_FP64(double, double)
-// ASSERT_EQUAL_64(int64_t, X register)
-// ASSERT_EQUAL_64(X register, X register)
-// ASSERT_EQUAL_FP64(double, D register)
+// CHECK_EQUAL_32(int32_t, int_32t)
+// CHECK_EQUAL_FP32(float, float)
+// CHECK_EQUAL_32(int32_t, W register)
+// CHECK_EQUAL_FP32(float, S register)
+// CHECK_EQUAL_64(int64_t, int_64t)
+// CHECK_EQUAL_FP64(double, double)
+// CHECK_EQUAL_64(int64_t, X register)
+// CHECK_EQUAL_64(X register, X register)
+// CHECK_EQUAL_FP64(double, D register)
//
-// e.g. ASSERT_EQUAL_64(0.5, d30);
+// e.g. CHECK_EQUAL_64(0.5, d30);
//
// If more advance computation is required before the assert then access the
// RegisterDump named core directly:
//
-// ASSERT_EQUAL_64(0x1234, core.xreg(0) & 0xffff);
+// CHECK_EQUAL_64(0x1234, core.xreg(0) & 0xffff);
#if 0 // TODO(all): enable.
@@ -116,7 +116,7 @@ static void InitializeVM() {
#define SETUP_SIZE(buf_size) \
Isolate* isolate = Isolate::Current(); \
HandleScope scope(isolate); \
- ASSERT(isolate != NULL); \
+ DCHECK(isolate != NULL); \
byte* buf = new byte[buf_size]; \
MacroAssembler masm(isolate, buf, buf_size); \
Decoder<DispatchingDecoderVisitor>* decoder = \
@@ -170,7 +170,7 @@ static void InitializeVM() {
#define SETUP_SIZE(buf_size) \
Isolate* isolate = Isolate::Current(); \
HandleScope scope(isolate); \
- ASSERT(isolate != NULL); \
+ DCHECK(isolate != NULL); \
byte* buf = new byte[buf_size]; \
MacroAssembler masm(isolate, buf, buf_size); \
RegisterDump core;
@@ -209,29 +209,29 @@ static void InitializeVM() {
#endif // ifdef USE_SIMULATOR.
-#define ASSERT_EQUAL_NZCV(expected) \
+#define CHECK_EQUAL_NZCV(expected) \
CHECK(EqualNzcv(expected, core.flags_nzcv()))
-#define ASSERT_EQUAL_REGISTERS(expected) \
+#define CHECK_EQUAL_REGISTERS(expected) \
CHECK(EqualRegisters(&expected, &core))
-#define ASSERT_EQUAL_32(expected, result) \
+#define CHECK_EQUAL_32(expected, result) \
CHECK(Equal32(static_cast<uint32_t>(expected), &core, result))
-#define ASSERT_EQUAL_FP32(expected, result) \
+#define CHECK_EQUAL_FP32(expected, result) \
CHECK(EqualFP32(expected, &core, result))
-#define ASSERT_EQUAL_64(expected, result) \
+#define CHECK_EQUAL_64(expected, result) \
CHECK(Equal64(expected, &core, result))
-#define ASSERT_EQUAL_FP64(expected, result) \
+#define CHECK_EQUAL_FP64(expected, result) \
CHECK(EqualFP64(expected, &core, result))
#ifdef DEBUG
-#define ASSERT_LITERAL_POOL_SIZE(expected) \
+#define DCHECK_LITERAL_POOL_SIZE(expected) \
CHECK((expected) == (__ LiteralPoolSize()))
#else
-#define ASSERT_LITERAL_POOL_SIZE(expected) \
+#define DCHECK_LITERAL_POOL_SIZE(expected) \
((void) 0)
#endif
@@ -276,12 +276,12 @@ TEST(stack_ops) {
RUN();
- ASSERT_EQUAL_64(0x1000, x0);
- ASSERT_EQUAL_64(0x1050, x1);
- ASSERT_EQUAL_64(0x104f, x2);
- ASSERT_EQUAL_64(0x1fff, x3);
- ASSERT_EQUAL_64(0xfffffff8, x4);
- ASSERT_EQUAL_64(0xfffffff8, x5);
+ CHECK_EQUAL_64(0x1000, x0);
+ CHECK_EQUAL_64(0x1050, x1);
+ CHECK_EQUAL_64(0x104f, x2);
+ CHECK_EQUAL_64(0x1fff, x3);
+ CHECK_EQUAL_64(0xfffffff8, x4);
+ CHECK_EQUAL_64(0xfffffff8, x5);
TEARDOWN();
}
@@ -312,22 +312,22 @@ TEST(mvn) {
RUN();
- ASSERT_EQUAL_64(0xfffff000, x0);
- ASSERT_EQUAL_64(0xfffffffffffff000UL, x1);
- ASSERT_EQUAL_64(0x00001fff, x2);
- ASSERT_EQUAL_64(0x0000000000003fffUL, x3);
- ASSERT_EQUAL_64(0xe00001ff, x4);
- ASSERT_EQUAL_64(0xf0000000000000ffUL, x5);
- ASSERT_EQUAL_64(0x00000001, x6);
- ASSERT_EQUAL_64(0x0, x7);
- ASSERT_EQUAL_64(0x7ff80000, x8);
- ASSERT_EQUAL_64(0x3ffc000000000000UL, x9);
- ASSERT_EQUAL_64(0xffffff00, x10);
- ASSERT_EQUAL_64(0x0000000000000001UL, x11);
- ASSERT_EQUAL_64(0xffff8003, x12);
- ASSERT_EQUAL_64(0xffffffffffff0007UL, x13);
- ASSERT_EQUAL_64(0xfffffffffffe000fUL, x14);
- ASSERT_EQUAL_64(0xfffffffffffe000fUL, x15);
+ CHECK_EQUAL_64(0xfffff000, x0);
+ CHECK_EQUAL_64(0xfffffffffffff000UL, x1);
+ CHECK_EQUAL_64(0x00001fff, x2);
+ CHECK_EQUAL_64(0x0000000000003fffUL, x3);
+ CHECK_EQUAL_64(0xe00001ff, x4);
+ CHECK_EQUAL_64(0xf0000000000000ffUL, x5);
+ CHECK_EQUAL_64(0x00000001, x6);
+ CHECK_EQUAL_64(0x0, x7);
+ CHECK_EQUAL_64(0x7ff80000, x8);
+ CHECK_EQUAL_64(0x3ffc000000000000UL, x9);
+ CHECK_EQUAL_64(0xffffff00, x10);
+ CHECK_EQUAL_64(0x0000000000000001UL, x11);
+ CHECK_EQUAL_64(0xffff8003, x12);
+ CHECK_EQUAL_64(0xffffffffffff0007UL, x13);
+ CHECK_EQUAL_64(0xfffffffffffe000fUL, x14);
+ CHECK_EQUAL_64(0xfffffffffffe000fUL, x15);
TEARDOWN();
}
@@ -384,31 +384,31 @@ TEST(mov) {
RUN();
- ASSERT_EQUAL_64(0x0123456789abcdefL, x0);
- ASSERT_EQUAL_64(0x00000000abcd0000L, x1);
- ASSERT_EQUAL_64(0xffffabcdffffffffL, x2);
- ASSERT_EQUAL_64(0x5432ffffffffffffL, x3);
- ASSERT_EQUAL_64(x4, x5);
- ASSERT_EQUAL_32(-1, w6);
- ASSERT_EQUAL_64(0x0123456789abcdefL, x7);
- ASSERT_EQUAL_32(0x89abcdefL, w8);
- ASSERT_EQUAL_64(0x0123456789abcdefL, x9);
- ASSERT_EQUAL_32(0x89abcdefL, w10);
- ASSERT_EQUAL_64(0x00000fff, x11);
- ASSERT_EQUAL_64(0x0000000000000fffUL, x12);
- ASSERT_EQUAL_64(0x00001ffe, x13);
- ASSERT_EQUAL_64(0x0000000000003ffcUL, x14);
- ASSERT_EQUAL_64(0x000001ff, x15);
- ASSERT_EQUAL_64(0x00000000000000ffUL, x18);
- ASSERT_EQUAL_64(0x00000001, x19);
- ASSERT_EQUAL_64(0x0, x20);
- ASSERT_EQUAL_64(0x7ff80000, x21);
- ASSERT_EQUAL_64(0x3ffc000000000000UL, x22);
- ASSERT_EQUAL_64(0x000000fe, x23);
- ASSERT_EQUAL_64(0xfffffffffffffffcUL, x24);
- ASSERT_EQUAL_64(0x00007ff8, x25);
- ASSERT_EQUAL_64(0x000000000000fff0UL, x26);
- ASSERT_EQUAL_64(0x000000000001ffe0UL, x27);
+ CHECK_EQUAL_64(0x0123456789abcdefL, x0);
+ CHECK_EQUAL_64(0x00000000abcd0000L, x1);
+ CHECK_EQUAL_64(0xffffabcdffffffffL, x2);
+ CHECK_EQUAL_64(0x5432ffffffffffffL, x3);
+ CHECK_EQUAL_64(x4, x5);
+ CHECK_EQUAL_32(-1, w6);
+ CHECK_EQUAL_64(0x0123456789abcdefL, x7);
+ CHECK_EQUAL_32(0x89abcdefL, w8);
+ CHECK_EQUAL_64(0x0123456789abcdefL, x9);
+ CHECK_EQUAL_32(0x89abcdefL, w10);
+ CHECK_EQUAL_64(0x00000fff, x11);
+ CHECK_EQUAL_64(0x0000000000000fffUL, x12);
+ CHECK_EQUAL_64(0x00001ffe, x13);
+ CHECK_EQUAL_64(0x0000000000003ffcUL, x14);
+ CHECK_EQUAL_64(0x000001ff, x15);
+ CHECK_EQUAL_64(0x00000000000000ffUL, x18);
+ CHECK_EQUAL_64(0x00000001, x19);
+ CHECK_EQUAL_64(0x0, x20);
+ CHECK_EQUAL_64(0x7ff80000, x21);
+ CHECK_EQUAL_64(0x3ffc000000000000UL, x22);
+ CHECK_EQUAL_64(0x000000fe, x23);
+ CHECK_EQUAL_64(0xfffffffffffffffcUL, x24);
+ CHECK_EQUAL_64(0x00007ff8, x25);
+ CHECK_EQUAL_64(0x000000000000fff0UL, x26);
+ CHECK_EQUAL_64(0x000000000001ffe0UL, x27);
TEARDOWN();
}
@@ -433,16 +433,16 @@ TEST(mov_imm_w) {
RUN();
- ASSERT_EQUAL_64(0xffffffffL, x0);
- ASSERT_EQUAL_64(0xffff1234L, x1);
- ASSERT_EQUAL_64(0x1234ffffL, x2);
- ASSERT_EQUAL_64(0x00000000L, x3);
- ASSERT_EQUAL_64(0x00001234L, x4);
- ASSERT_EQUAL_64(0x12340000L, x5);
- ASSERT_EQUAL_64(0x12345678L, x6);
- ASSERT_EQUAL_64(0x80000000L, x7);
- ASSERT_EQUAL_64(0xffff0000L, x8);
- ASSERT_EQUAL_32(kWMinInt, w9);
+ CHECK_EQUAL_64(0xffffffffL, x0);
+ CHECK_EQUAL_64(0xffff1234L, x1);
+ CHECK_EQUAL_64(0x1234ffffL, x2);
+ CHECK_EQUAL_64(0x00000000L, x3);
+ CHECK_EQUAL_64(0x00001234L, x4);
+ CHECK_EQUAL_64(0x12340000L, x5);
+ CHECK_EQUAL_64(0x12345678L, x6);
+ CHECK_EQUAL_64(0x80000000L, x7);
+ CHECK_EQUAL_64(0xffff0000L, x8);
+ CHECK_EQUAL_32(kWMinInt, w9);
TEARDOWN();
}
@@ -484,32 +484,32 @@ TEST(mov_imm_x) {
RUN();
- ASSERT_EQUAL_64(0xffffffffffff1234L, x1);
- ASSERT_EQUAL_64(0xffffffff12345678L, x2);
- ASSERT_EQUAL_64(0xffff1234ffff5678L, x3);
- ASSERT_EQUAL_64(0x1234ffffffff5678L, x4);
- ASSERT_EQUAL_64(0x1234ffff5678ffffL, x5);
- ASSERT_EQUAL_64(0x12345678ffffffffL, x6);
- ASSERT_EQUAL_64(0x1234ffffffffffffL, x7);
- ASSERT_EQUAL_64(0x123456789abcffffL, x8);
- ASSERT_EQUAL_64(0x12345678ffff9abcL, x9);
- ASSERT_EQUAL_64(0x1234ffff56789abcL, x10);
- ASSERT_EQUAL_64(0xffff123456789abcL, x11);
- ASSERT_EQUAL_64(0x0000000000000000L, x12);
- ASSERT_EQUAL_64(0x0000000000001234L, x13);
- ASSERT_EQUAL_64(0x0000000012345678L, x14);
- ASSERT_EQUAL_64(0x0000123400005678L, x15);
- ASSERT_EQUAL_64(0x1234000000005678L, x18);
- ASSERT_EQUAL_64(0x1234000056780000L, x19);
- ASSERT_EQUAL_64(0x1234567800000000L, x20);
- ASSERT_EQUAL_64(0x1234000000000000L, x21);
- ASSERT_EQUAL_64(0x123456789abc0000L, x22);
- ASSERT_EQUAL_64(0x1234567800009abcL, x23);
- ASSERT_EQUAL_64(0x1234000056789abcL, x24);
- ASSERT_EQUAL_64(0x0000123456789abcL, x25);
- ASSERT_EQUAL_64(0x123456789abcdef0L, x26);
- ASSERT_EQUAL_64(0xffff000000000001L, x27);
- ASSERT_EQUAL_64(0x8000ffff00000000L, x28);
+ CHECK_EQUAL_64(0xffffffffffff1234L, x1);
+ CHECK_EQUAL_64(0xffffffff12345678L, x2);
+ CHECK_EQUAL_64(0xffff1234ffff5678L, x3);
+ CHECK_EQUAL_64(0x1234ffffffff5678L, x4);
+ CHECK_EQUAL_64(0x1234ffff5678ffffL, x5);
+ CHECK_EQUAL_64(0x12345678ffffffffL, x6);
+ CHECK_EQUAL_64(0x1234ffffffffffffL, x7);
+ CHECK_EQUAL_64(0x123456789abcffffL, x8);
+ CHECK_EQUAL_64(0x12345678ffff9abcL, x9);
+ CHECK_EQUAL_64(0x1234ffff56789abcL, x10);
+ CHECK_EQUAL_64(0xffff123456789abcL, x11);
+ CHECK_EQUAL_64(0x0000000000000000L, x12);
+ CHECK_EQUAL_64(0x0000000000001234L, x13);
+ CHECK_EQUAL_64(0x0000000012345678L, x14);
+ CHECK_EQUAL_64(0x0000123400005678L, x15);
+ CHECK_EQUAL_64(0x1234000000005678L, x18);
+ CHECK_EQUAL_64(0x1234000056780000L, x19);
+ CHECK_EQUAL_64(0x1234567800000000L, x20);
+ CHECK_EQUAL_64(0x1234000000000000L, x21);
+ CHECK_EQUAL_64(0x123456789abc0000L, x22);
+ CHECK_EQUAL_64(0x1234567800009abcL, x23);
+ CHECK_EQUAL_64(0x1234000056789abcL, x24);
+ CHECK_EQUAL_64(0x0000123456789abcL, x25);
+ CHECK_EQUAL_64(0x123456789abcdef0L, x26);
+ CHECK_EQUAL_64(0xffff000000000001L, x27);
+ CHECK_EQUAL_64(0x8000ffff00000000L, x28);
TEARDOWN();
}
@@ -537,16 +537,16 @@ TEST(orr) {
RUN();
- ASSERT_EQUAL_64(0xf000f0ff, x2);
- ASSERT_EQUAL_64(0xf000f0f0, x3);
- ASSERT_EQUAL_64(0xf00000ff0000f0f0L, x4);
- ASSERT_EQUAL_64(0x0f00f0ff, x5);
- ASSERT_EQUAL_64(0xff00f0ff, x6);
- ASSERT_EQUAL_64(0x0f00f0ff, x7);
- ASSERT_EQUAL_64(0x0ffff0f0, x8);
- ASSERT_EQUAL_64(0x0ff00000000ff0f0L, x9);
- ASSERT_EQUAL_64(0xf0ff, x10);
- ASSERT_EQUAL_64(0xf0000000f000f0f0L, x11);
+ CHECK_EQUAL_64(0xf000f0ff, x2);
+ CHECK_EQUAL_64(0xf000f0f0, x3);
+ CHECK_EQUAL_64(0xf00000ff0000f0f0L, x4);
+ CHECK_EQUAL_64(0x0f00f0ff, x5);
+ CHECK_EQUAL_64(0xff00f0ff, x6);
+ CHECK_EQUAL_64(0x0f00f0ff, x7);
+ CHECK_EQUAL_64(0x0ffff0f0, x8);
+ CHECK_EQUAL_64(0x0ff00000000ff0f0L, x9);
+ CHECK_EQUAL_64(0xf0ff, x10);
+ CHECK_EQUAL_64(0xf0000000f000f0f0L, x11);
TEARDOWN();
}
@@ -571,14 +571,14 @@ TEST(orr_extend) {
RUN();
- ASSERT_EQUAL_64(0x00000081, x6);
- ASSERT_EQUAL_64(0x00010101, x7);
- ASSERT_EQUAL_64(0x00020201, x8);
- ASSERT_EQUAL_64(0x0000000400040401UL, x9);
- ASSERT_EQUAL_64(0x00000000ffffff81UL, x10);
- ASSERT_EQUAL_64(0xffffffffffff0101UL, x11);
- ASSERT_EQUAL_64(0xfffffffe00020201UL, x12);
- ASSERT_EQUAL_64(0x0000000400040401UL, x13);
+ CHECK_EQUAL_64(0x00000081, x6);
+ CHECK_EQUAL_64(0x00010101, x7);
+ CHECK_EQUAL_64(0x00020201, x8);
+ CHECK_EQUAL_64(0x0000000400040401UL, x9);
+ CHECK_EQUAL_64(0x00000000ffffff81UL, x10);
+ CHECK_EQUAL_64(0xffffffffffff0101UL, x11);
+ CHECK_EQUAL_64(0xfffffffe00020201UL, x12);
+ CHECK_EQUAL_64(0x0000000400040401UL, x13);
TEARDOWN();
}
@@ -601,12 +601,12 @@ TEST(bitwise_wide_imm) {
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(0xf0f0f0f0f0f0f0f0UL, x1);
- ASSERT_EQUAL_64(0x1234567890abcdefUL, x10);
- ASSERT_EQUAL_64(0xf0fbfdffUL, x11);
- ASSERT_EQUAL_32(kWMinInt, w12);
- ASSERT_EQUAL_32(kWMinInt, w13);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(0xf0f0f0f0f0f0f0f0UL, x1);
+ CHECK_EQUAL_64(0x1234567890abcdefUL, x10);
+ CHECK_EQUAL_64(0xf0fbfdffUL, x11);
+ CHECK_EQUAL_32(kWMinInt, w12);
+ CHECK_EQUAL_32(kWMinInt, w13);
TEARDOWN();
}
@@ -634,16 +634,16 @@ TEST(orn) {
RUN();
- ASSERT_EQUAL_64(0xffffffff0ffffff0L, x2);
- ASSERT_EQUAL_64(0xfffff0ff, x3);
- ASSERT_EQUAL_64(0xfffffff0fffff0ffL, x4);
- ASSERT_EQUAL_64(0xffffffff87fffff0L, x5);
- ASSERT_EQUAL_64(0x07fffff0, x6);
- ASSERT_EQUAL_64(0xffffffff87fffff0L, x7);
- ASSERT_EQUAL_64(0xff00ffff, x8);
- ASSERT_EQUAL_64(0xff00ffffffffffffL, x9);
- ASSERT_EQUAL_64(0xfffff0f0, x10);
- ASSERT_EQUAL_64(0xffff0000fffff0f0L, x11);
+ CHECK_EQUAL_64(0xffffffff0ffffff0L, x2);
+ CHECK_EQUAL_64(0xfffff0ff, x3);
+ CHECK_EQUAL_64(0xfffffff0fffff0ffL, x4);
+ CHECK_EQUAL_64(0xffffffff87fffff0L, x5);
+ CHECK_EQUAL_64(0x07fffff0, x6);
+ CHECK_EQUAL_64(0xffffffff87fffff0L, x7);
+ CHECK_EQUAL_64(0xff00ffff, x8);
+ CHECK_EQUAL_64(0xff00ffffffffffffL, x9);
+ CHECK_EQUAL_64(0xfffff0f0, x10);
+ CHECK_EQUAL_64(0xffff0000fffff0f0L, x11);
TEARDOWN();
}
@@ -668,14 +668,14 @@ TEST(orn_extend) {
RUN();
- ASSERT_EQUAL_64(0xffffff7f, x6);
- ASSERT_EQUAL_64(0xfffffffffffefefdUL, x7);
- ASSERT_EQUAL_64(0xfffdfdfb, x8);
- ASSERT_EQUAL_64(0xfffffffbfffbfbf7UL, x9);
- ASSERT_EQUAL_64(0x0000007f, x10);
- ASSERT_EQUAL_64(0x0000fefd, x11);
- ASSERT_EQUAL_64(0x00000001fffdfdfbUL, x12);
- ASSERT_EQUAL_64(0xfffffffbfffbfbf7UL, x13);
+ CHECK_EQUAL_64(0xffffff7f, x6);
+ CHECK_EQUAL_64(0xfffffffffffefefdUL, x7);
+ CHECK_EQUAL_64(0xfffdfdfb, x8);
+ CHECK_EQUAL_64(0xfffffffbfffbfbf7UL, x9);
+ CHECK_EQUAL_64(0x0000007f, x10);
+ CHECK_EQUAL_64(0x0000fefd, x11);
+ CHECK_EQUAL_64(0x00000001fffdfdfbUL, x12);
+ CHECK_EQUAL_64(0xfffffffbfffbfbf7UL, x13);
TEARDOWN();
}
@@ -703,16 +703,16 @@ TEST(and_) {
RUN();
- ASSERT_EQUAL_64(0x000000f0, x2);
- ASSERT_EQUAL_64(0x00000ff0, x3);
- ASSERT_EQUAL_64(0x00000ff0, x4);
- ASSERT_EQUAL_64(0x00000070, x5);
- ASSERT_EQUAL_64(0x0000ff00, x6);
- ASSERT_EQUAL_64(0x00000f00, x7);
- ASSERT_EQUAL_64(0x00000ff0, x8);
- ASSERT_EQUAL_64(0x00000000, x9);
- ASSERT_EQUAL_64(0x0000ff00, x10);
- ASSERT_EQUAL_64(0x000000f0, x11);
+ CHECK_EQUAL_64(0x000000f0, x2);
+ CHECK_EQUAL_64(0x00000ff0, x3);
+ CHECK_EQUAL_64(0x00000ff0, x4);
+ CHECK_EQUAL_64(0x00000070, x5);
+ CHECK_EQUAL_64(0x0000ff00, x6);
+ CHECK_EQUAL_64(0x00000f00, x7);
+ CHECK_EQUAL_64(0x00000ff0, x8);
+ CHECK_EQUAL_64(0x00000000, x9);
+ CHECK_EQUAL_64(0x0000ff00, x10);
+ CHECK_EQUAL_64(0x000000f0, x11);
TEARDOWN();
}
@@ -737,14 +737,14 @@ TEST(and_extend) {
RUN();
- ASSERT_EQUAL_64(0x00000081, x6);
- ASSERT_EQUAL_64(0x00010102, x7);
- ASSERT_EQUAL_64(0x00020204, x8);
- ASSERT_EQUAL_64(0x0000000400040408UL, x9);
- ASSERT_EQUAL_64(0xffffff81, x10);
- ASSERT_EQUAL_64(0xffffffffffff0102UL, x11);
- ASSERT_EQUAL_64(0xfffffffe00020204UL, x12);
- ASSERT_EQUAL_64(0x0000000400040408UL, x13);
+ CHECK_EQUAL_64(0x00000081, x6);
+ CHECK_EQUAL_64(0x00010102, x7);
+ CHECK_EQUAL_64(0x00020204, x8);
+ CHECK_EQUAL_64(0x0000000400040408UL, x9);
+ CHECK_EQUAL_64(0xffffff81, x10);
+ CHECK_EQUAL_64(0xffffffffffff0102UL, x11);
+ CHECK_EQUAL_64(0xfffffffe00020204UL, x12);
+ CHECK_EQUAL_64(0x0000000400040408UL, x13);
TEARDOWN();
}
@@ -761,8 +761,8 @@ TEST(ands) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0xf00000ff, x0);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0xf00000ff, x0);
START();
__ Mov(x0, 0xfff0);
@@ -772,8 +772,8 @@ TEST(ands) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0x00000000, x0);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0x00000000, x0);
START();
__ Mov(x0, 0x8000000000000000L);
@@ -783,8 +783,8 @@ TEST(ands) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x8000000000000000L, x0);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x8000000000000000L, x0);
START();
__ Mov(x0, 0xfff0);
@@ -793,8 +793,8 @@ TEST(ands) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0x00000000, x0);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0x00000000, x0);
START();
__ Mov(x0, 0xff000000);
@@ -803,8 +803,8 @@ TEST(ands) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x80000000, x0);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x80000000, x0);
TEARDOWN();
}
@@ -842,18 +842,18 @@ TEST(bic) {
RUN();
- ASSERT_EQUAL_64(0x0000ff00, x2);
- ASSERT_EQUAL_64(0x0000f000, x3);
- ASSERT_EQUAL_64(0x0000f000, x4);
- ASSERT_EQUAL_64(0x0000ff80, x5);
- ASSERT_EQUAL_64(0x000000f0, x6);
- ASSERT_EQUAL_64(0x0000f0f0, x7);
- ASSERT_EQUAL_64(0x0000f000, x8);
- ASSERT_EQUAL_64(0x0000ff00, x9);
- ASSERT_EQUAL_64(0x0000ffe0, x10);
- ASSERT_EQUAL_64(0x0000fef0, x11);
+ CHECK_EQUAL_64(0x0000ff00, x2);
+ CHECK_EQUAL_64(0x0000f000, x3);
+ CHECK_EQUAL_64(0x0000f000, x4);
+ CHECK_EQUAL_64(0x0000ff80, x5);
+ CHECK_EQUAL_64(0x000000f0, x6);
+ CHECK_EQUAL_64(0x0000f0f0, x7);
+ CHECK_EQUAL_64(0x0000f000, x8);
+ CHECK_EQUAL_64(0x0000ff00, x9);
+ CHECK_EQUAL_64(0x0000ffe0, x10);
+ CHECK_EQUAL_64(0x0000fef0, x11);
- ASSERT_EQUAL_64(0x543210, x21);
+ CHECK_EQUAL_64(0x543210, x21);
TEARDOWN();
}
@@ -878,14 +878,14 @@ TEST(bic_extend) {
RUN();
- ASSERT_EQUAL_64(0xffffff7e, x6);
- ASSERT_EQUAL_64(0xfffffffffffefefdUL, x7);
- ASSERT_EQUAL_64(0xfffdfdfb, x8);
- ASSERT_EQUAL_64(0xfffffffbfffbfbf7UL, x9);
- ASSERT_EQUAL_64(0x0000007e, x10);
- ASSERT_EQUAL_64(0x0000fefd, x11);
- ASSERT_EQUAL_64(0x00000001fffdfdfbUL, x12);
- ASSERT_EQUAL_64(0xfffffffbfffbfbf7UL, x13);
+ CHECK_EQUAL_64(0xffffff7e, x6);
+ CHECK_EQUAL_64(0xfffffffffffefefdUL, x7);
+ CHECK_EQUAL_64(0xfffdfdfb, x8);
+ CHECK_EQUAL_64(0xfffffffbfffbfbf7UL, x9);
+ CHECK_EQUAL_64(0x0000007e, x10);
+ CHECK_EQUAL_64(0x0000fefd, x11);
+ CHECK_EQUAL_64(0x00000001fffdfdfbUL, x12);
+ CHECK_EQUAL_64(0xfffffffbfffbfbf7UL, x13);
TEARDOWN();
}
@@ -902,8 +902,8 @@ TEST(bics) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0x00000000, x0);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0x00000000, x0);
START();
__ Mov(x0, 0xffffffff);
@@ -912,8 +912,8 @@ TEST(bics) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x80000000, x0);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x80000000, x0);
START();
__ Mov(x0, 0x8000000000000000L);
@@ -923,8 +923,8 @@ TEST(bics) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0x00000000, x0);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0x00000000, x0);
START();
__ Mov(x0, 0xffffffffffffffffL);
@@ -933,8 +933,8 @@ TEST(bics) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x8000000000000000L, x0);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x8000000000000000L, x0);
START();
__ Mov(w0, 0xffff0000);
@@ -943,8 +943,8 @@ TEST(bics) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0x00000000, x0);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0x00000000, x0);
TEARDOWN();
}
@@ -972,16 +972,16 @@ TEST(eor) {
RUN();
- ASSERT_EQUAL_64(0xf000ff0f, x2);
- ASSERT_EQUAL_64(0x0000f000, x3);
- ASSERT_EQUAL_64(0x0000000f0000f000L, x4);
- ASSERT_EQUAL_64(0x7800ff8f, x5);
- ASSERT_EQUAL_64(0xffff00f0, x6);
- ASSERT_EQUAL_64(0x0000f0f0, x7);
- ASSERT_EQUAL_64(0x0000f00f, x8);
- ASSERT_EQUAL_64(0x00000ff00000ffffL, x9);
- ASSERT_EQUAL_64(0xff0000f0, x10);
- ASSERT_EQUAL_64(0xff00ff00ff0000f0L, x11);
+ CHECK_EQUAL_64(0xf000ff0f, x2);
+ CHECK_EQUAL_64(0x0000f000, x3);
+ CHECK_EQUAL_64(0x0000000f0000f000L, x4);
+ CHECK_EQUAL_64(0x7800ff8f, x5);
+ CHECK_EQUAL_64(0xffff00f0, x6);
+ CHECK_EQUAL_64(0x0000f0f0, x7);
+ CHECK_EQUAL_64(0x0000f00f, x8);
+ CHECK_EQUAL_64(0x00000ff00000ffffL, x9);
+ CHECK_EQUAL_64(0xff0000f0, x10);
+ CHECK_EQUAL_64(0xff00ff00ff0000f0L, x11);
TEARDOWN();
}
@@ -1006,14 +1006,14 @@ TEST(eor_extend) {
RUN();
- ASSERT_EQUAL_64(0x11111190, x6);
- ASSERT_EQUAL_64(0x1111111111101013UL, x7);
- ASSERT_EQUAL_64(0x11131315, x8);
- ASSERT_EQUAL_64(0x1111111511151519UL, x9);
- ASSERT_EQUAL_64(0xeeeeee90, x10);
- ASSERT_EQUAL_64(0xeeeeeeeeeeee1013UL, x11);
- ASSERT_EQUAL_64(0xeeeeeeef11131315UL, x12);
- ASSERT_EQUAL_64(0x1111111511151519UL, x13);
+ CHECK_EQUAL_64(0x11111190, x6);
+ CHECK_EQUAL_64(0x1111111111101013UL, x7);
+ CHECK_EQUAL_64(0x11131315, x8);
+ CHECK_EQUAL_64(0x1111111511151519UL, x9);
+ CHECK_EQUAL_64(0xeeeeee90, x10);
+ CHECK_EQUAL_64(0xeeeeeeeeeeee1013UL, x11);
+ CHECK_EQUAL_64(0xeeeeeeef11131315UL, x12);
+ CHECK_EQUAL_64(0x1111111511151519UL, x13);
TEARDOWN();
}
@@ -1041,16 +1041,16 @@ TEST(eon) {
RUN();
- ASSERT_EQUAL_64(0xffffffff0fff00f0L, x2);
- ASSERT_EQUAL_64(0xffff0fff, x3);
- ASSERT_EQUAL_64(0xfffffff0ffff0fffL, x4);
- ASSERT_EQUAL_64(0xffffffff87ff0070L, x5);
- ASSERT_EQUAL_64(0x0000ff0f, x6);
- ASSERT_EQUAL_64(0xffffffffffff0f0fL, x7);
- ASSERT_EQUAL_64(0xffff0ff0, x8);
- ASSERT_EQUAL_64(0xfffff00fffff0000L, x9);
- ASSERT_EQUAL_64(0xfc3f03cf, x10);
- ASSERT_EQUAL_64(0xffffefffffff100fL, x11);
+ CHECK_EQUAL_64(0xffffffff0fff00f0L, x2);
+ CHECK_EQUAL_64(0xffff0fff, x3);
+ CHECK_EQUAL_64(0xfffffff0ffff0fffL, x4);
+ CHECK_EQUAL_64(0xffffffff87ff0070L, x5);
+ CHECK_EQUAL_64(0x0000ff0f, x6);
+ CHECK_EQUAL_64(0xffffffffffff0f0fL, x7);
+ CHECK_EQUAL_64(0xffff0ff0, x8);
+ CHECK_EQUAL_64(0xfffff00fffff0000L, x9);
+ CHECK_EQUAL_64(0xfc3f03cf, x10);
+ CHECK_EQUAL_64(0xffffefffffff100fL, x11);
TEARDOWN();
}
@@ -1075,14 +1075,14 @@ TEST(eon_extend) {
RUN();
- ASSERT_EQUAL_64(0xeeeeee6f, x6);
- ASSERT_EQUAL_64(0xeeeeeeeeeeefefecUL, x7);
- ASSERT_EQUAL_64(0xeeececea, x8);
- ASSERT_EQUAL_64(0xeeeeeeeaeeeaeae6UL, x9);
- ASSERT_EQUAL_64(0x1111116f, x10);
- ASSERT_EQUAL_64(0x111111111111efecUL, x11);
- ASSERT_EQUAL_64(0x11111110eeececeaUL, x12);
- ASSERT_EQUAL_64(0xeeeeeeeaeeeaeae6UL, x13);
+ CHECK_EQUAL_64(0xeeeeee6f, x6);
+ CHECK_EQUAL_64(0xeeeeeeeeeeefefecUL, x7);
+ CHECK_EQUAL_64(0xeeececea, x8);
+ CHECK_EQUAL_64(0xeeeeeeeaeeeaeae6UL, x9);
+ CHECK_EQUAL_64(0x1111116f, x10);
+ CHECK_EQUAL_64(0x111111111111efecUL, x11);
+ CHECK_EQUAL_64(0x11111110eeececeaUL, x12);
+ CHECK_EQUAL_64(0xeeeeeeeaeeeaeae6UL, x13);
TEARDOWN();
}
@@ -1121,25 +1121,25 @@ TEST(mul) {
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(0, x1);
- ASSERT_EQUAL_64(0xffffffff, x2);
- ASSERT_EQUAL_64(1, x3);
- ASSERT_EQUAL_64(0, x4);
- ASSERT_EQUAL_64(0xffffffff, x5);
- ASSERT_EQUAL_64(0xffffffff00000001UL, x6);
- ASSERT_EQUAL_64(1, x7);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(1, x10);
- ASSERT_EQUAL_64(0, x11);
- ASSERT_EQUAL_64(0, x12);
- ASSERT_EQUAL_64(1, x13);
- ASSERT_EQUAL_64(0xffffffff, x14);
- ASSERT_EQUAL_64(0, x20);
- ASSERT_EQUAL_64(0xffffffff00000001UL, x21);
- ASSERT_EQUAL_64(0xffffffff, x22);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x23);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(0, x1);
+ CHECK_EQUAL_64(0xffffffff, x2);
+ CHECK_EQUAL_64(1, x3);
+ CHECK_EQUAL_64(0, x4);
+ CHECK_EQUAL_64(0xffffffff, x5);
+ CHECK_EQUAL_64(0xffffffff00000001UL, x6);
+ CHECK_EQUAL_64(1, x7);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(1, x10);
+ CHECK_EQUAL_64(0, x11);
+ CHECK_EQUAL_64(0, x12);
+ CHECK_EQUAL_64(1, x13);
+ CHECK_EQUAL_64(0xffffffff, x14);
+ CHECK_EQUAL_64(0, x20);
+ CHECK_EQUAL_64(0xffffffff00000001UL, x21);
+ CHECK_EQUAL_64(0xffffffff, x22);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x23);
TEARDOWN();
}
@@ -1153,7 +1153,7 @@ static void SmullHelper(int64_t expected, int64_t a, int64_t b) {
__ Smull(x2, w0, w1);
END();
RUN();
- ASSERT_EQUAL_64(expected, x2);
+ CHECK_EQUAL_64(expected, x2);
TEARDOWN();
}
@@ -1209,31 +1209,31 @@ TEST(madd) {
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(0xffffffff, x2);
- ASSERT_EQUAL_64(0xffffffff, x3);
- ASSERT_EQUAL_64(1, x4);
- ASSERT_EQUAL_64(0, x5);
- ASSERT_EQUAL_64(0, x6);
- ASSERT_EQUAL_64(0xffffffff, x7);
- ASSERT_EQUAL_64(0xfffffffe, x8);
- ASSERT_EQUAL_64(2, x9);
- ASSERT_EQUAL_64(0, x10);
- ASSERT_EQUAL_64(0, x11);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(0xffffffff, x2);
+ CHECK_EQUAL_64(0xffffffff, x3);
+ CHECK_EQUAL_64(1, x4);
+ CHECK_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(0, x6);
+ CHECK_EQUAL_64(0xffffffff, x7);
+ CHECK_EQUAL_64(0xfffffffe, x8);
+ CHECK_EQUAL_64(2, x9);
+ CHECK_EQUAL_64(0, x10);
+ CHECK_EQUAL_64(0, x11);
- ASSERT_EQUAL_64(0, x12);
- ASSERT_EQUAL_64(1, x13);
- ASSERT_EQUAL_64(0xffffffff, x14);
- ASSERT_EQUAL_64(0xffffffffffffffff, x15);
- ASSERT_EQUAL_64(1, x20);
- ASSERT_EQUAL_64(0x100000000UL, x21);
- ASSERT_EQUAL_64(0, x22);
- ASSERT_EQUAL_64(0xffffffff, x23);
- ASSERT_EQUAL_64(0x1fffffffe, x24);
- ASSERT_EQUAL_64(0xfffffffe00000002UL, x25);
- ASSERT_EQUAL_64(0, x26);
- ASSERT_EQUAL_64(0, x27);
+ CHECK_EQUAL_64(0, x12);
+ CHECK_EQUAL_64(1, x13);
+ CHECK_EQUAL_64(0xffffffff, x14);
+ CHECK_EQUAL_64(0xffffffffffffffff, x15);
+ CHECK_EQUAL_64(1, x20);
+ CHECK_EQUAL_64(0x100000000UL, x21);
+ CHECK_EQUAL_64(0, x22);
+ CHECK_EQUAL_64(0xffffffff, x23);
+ CHECK_EQUAL_64(0x1fffffffe, x24);
+ CHECK_EQUAL_64(0xfffffffe00000002UL, x25);
+ CHECK_EQUAL_64(0, x26);
+ CHECK_EQUAL_64(0, x27);
TEARDOWN();
}
@@ -1279,31 +1279,31 @@ TEST(msub) {
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(0xffffffff, x2);
- ASSERT_EQUAL_64(0xffffffff, x3);
- ASSERT_EQUAL_64(1, x4);
- ASSERT_EQUAL_64(0xfffffffe, x5);
- ASSERT_EQUAL_64(0xfffffffe, x6);
- ASSERT_EQUAL_64(1, x7);
- ASSERT_EQUAL_64(0, x8);
- ASSERT_EQUAL_64(0, x9);
- ASSERT_EQUAL_64(0xfffffffe, x10);
- ASSERT_EQUAL_64(0xfffffffe, x11);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(0xffffffff, x2);
+ CHECK_EQUAL_64(0xffffffff, x3);
+ CHECK_EQUAL_64(1, x4);
+ CHECK_EQUAL_64(0xfffffffe, x5);
+ CHECK_EQUAL_64(0xfffffffe, x6);
+ CHECK_EQUAL_64(1, x7);
+ CHECK_EQUAL_64(0, x8);
+ CHECK_EQUAL_64(0, x9);
+ CHECK_EQUAL_64(0xfffffffe, x10);
+ CHECK_EQUAL_64(0xfffffffe, x11);
- ASSERT_EQUAL_64(0, x12);
- ASSERT_EQUAL_64(1, x13);
- ASSERT_EQUAL_64(0xffffffff, x14);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x15);
- ASSERT_EQUAL_64(1, x20);
- ASSERT_EQUAL_64(0xfffffffeUL, x21);
- ASSERT_EQUAL_64(0xfffffffffffffffeUL, x22);
- ASSERT_EQUAL_64(0xffffffff00000001UL, x23);
- ASSERT_EQUAL_64(0, x24);
- ASSERT_EQUAL_64(0x200000000UL, x25);
- ASSERT_EQUAL_64(0x1fffffffeUL, x26);
- ASSERT_EQUAL_64(0xfffffffffffffffeUL, x27);
+ CHECK_EQUAL_64(0, x12);
+ CHECK_EQUAL_64(1, x13);
+ CHECK_EQUAL_64(0xffffffff, x14);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x15);
+ CHECK_EQUAL_64(1, x20);
+ CHECK_EQUAL_64(0xfffffffeUL, x21);
+ CHECK_EQUAL_64(0xfffffffffffffffeUL, x22);
+ CHECK_EQUAL_64(0xffffffff00000001UL, x23);
+ CHECK_EQUAL_64(0, x24);
+ CHECK_EQUAL_64(0x200000000UL, x25);
+ CHECK_EQUAL_64(0x1fffffffeUL, x26);
+ CHECK_EQUAL_64(0xfffffffffffffffeUL, x27);
TEARDOWN();
}
@@ -1341,18 +1341,18 @@ TEST(smulh) {
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(0, x1);
- ASSERT_EQUAL_64(0, x2);
- ASSERT_EQUAL_64(0x01234567, x3);
- ASSERT_EQUAL_64(0x02468acf, x4);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x5);
- ASSERT_EQUAL_64(0x4000000000000000UL, x6);
- ASSERT_EQUAL_64(0, x7);
- ASSERT_EQUAL_64(0, x8);
- ASSERT_EQUAL_64(0x1c71c71c71c71c71UL, x9);
- ASSERT_EQUAL_64(0xe38e38e38e38e38eUL, x10);
- ASSERT_EQUAL_64(0x1c71c71c71c71c72UL, x11);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(0, x1);
+ CHECK_EQUAL_64(0, x2);
+ CHECK_EQUAL_64(0x01234567, x3);
+ CHECK_EQUAL_64(0x02468acf, x4);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x5);
+ CHECK_EQUAL_64(0x4000000000000000UL, x6);
+ CHECK_EQUAL_64(0, x7);
+ CHECK_EQUAL_64(0, x8);
+ CHECK_EQUAL_64(0x1c71c71c71c71c71UL, x9);
+ CHECK_EQUAL_64(0xe38e38e38e38e38eUL, x10);
+ CHECK_EQUAL_64(0x1c71c71c71c71c72UL, x11);
TEARDOWN();
}
@@ -1381,14 +1381,14 @@ TEST(smaddl_umaddl) {
RUN();
- ASSERT_EQUAL_64(3, x9);
- ASSERT_EQUAL_64(5, x10);
- ASSERT_EQUAL_64(5, x11);
- ASSERT_EQUAL_64(0x200000001UL, x12);
- ASSERT_EQUAL_64(0x100000003UL, x13);
- ASSERT_EQUAL_64(0xfffffffe00000005UL, x14);
- ASSERT_EQUAL_64(0xfffffffe00000005UL, x15);
- ASSERT_EQUAL_64(0x1, x22);
+ CHECK_EQUAL_64(3, x9);
+ CHECK_EQUAL_64(5, x10);
+ CHECK_EQUAL_64(5, x11);
+ CHECK_EQUAL_64(0x200000001UL, x12);
+ CHECK_EQUAL_64(0x100000003UL, x13);
+ CHECK_EQUAL_64(0xfffffffe00000005UL, x14);
+ CHECK_EQUAL_64(0xfffffffe00000005UL, x15);
+ CHECK_EQUAL_64(0x1, x22);
TEARDOWN();
}
@@ -1417,14 +1417,14 @@ TEST(smsubl_umsubl) {
RUN();
- ASSERT_EQUAL_64(5, x9);
- ASSERT_EQUAL_64(3, x10);
- ASSERT_EQUAL_64(3, x11);
- ASSERT_EQUAL_64(0x1ffffffffUL, x12);
- ASSERT_EQUAL_64(0xffffffff00000005UL, x13);
- ASSERT_EQUAL_64(0x200000003UL, x14);
- ASSERT_EQUAL_64(0x200000003UL, x15);
- ASSERT_EQUAL_64(0x3ffffffffUL, x22);
+ CHECK_EQUAL_64(5, x9);
+ CHECK_EQUAL_64(3, x10);
+ CHECK_EQUAL_64(3, x11);
+ CHECK_EQUAL_64(0x1ffffffffUL, x12);
+ CHECK_EQUAL_64(0xffffffff00000005UL, x13);
+ CHECK_EQUAL_64(0x200000003UL, x14);
+ CHECK_EQUAL_64(0x200000003UL, x15);
+ CHECK_EQUAL_64(0x3ffffffffUL, x22);
TEARDOWN();
}
@@ -1480,34 +1480,34 @@ TEST(div) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(0xffffffff, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0xffffffff, x3);
- ASSERT_EQUAL_64(1, x4);
- ASSERT_EQUAL_64(1, x5);
- ASSERT_EQUAL_64(0, x6);
- ASSERT_EQUAL_64(1, x7);
- ASSERT_EQUAL_64(0, x8);
- ASSERT_EQUAL_64(0xffffffff00000001UL, x9);
- ASSERT_EQUAL_64(0x40000000, x10);
- ASSERT_EQUAL_64(0xC0000000, x11);
- ASSERT_EQUAL_64(0x40000000, x12);
- ASSERT_EQUAL_64(0x40000000, x13);
- ASSERT_EQUAL_64(0x4000000000000000UL, x14);
- ASSERT_EQUAL_64(0xC000000000000000UL, x15);
- ASSERT_EQUAL_64(0, x22);
- ASSERT_EQUAL_64(0x80000000, x23);
- ASSERT_EQUAL_64(0, x24);
- ASSERT_EQUAL_64(0x8000000000000000UL, x25);
- ASSERT_EQUAL_64(0, x26);
- ASSERT_EQUAL_64(0, x27);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x28);
- ASSERT_EQUAL_64(0, x29);
- ASSERT_EQUAL_64(0, x18);
- ASSERT_EQUAL_64(0, x19);
- ASSERT_EQUAL_64(0, x20);
- ASSERT_EQUAL_64(0, x21);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(0xffffffff, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0xffffffff, x3);
+ CHECK_EQUAL_64(1, x4);
+ CHECK_EQUAL_64(1, x5);
+ CHECK_EQUAL_64(0, x6);
+ CHECK_EQUAL_64(1, x7);
+ CHECK_EQUAL_64(0, x8);
+ CHECK_EQUAL_64(0xffffffff00000001UL, x9);
+ CHECK_EQUAL_64(0x40000000, x10);
+ CHECK_EQUAL_64(0xC0000000, x11);
+ CHECK_EQUAL_64(0x40000000, x12);
+ CHECK_EQUAL_64(0x40000000, x13);
+ CHECK_EQUAL_64(0x4000000000000000UL, x14);
+ CHECK_EQUAL_64(0xC000000000000000UL, x15);
+ CHECK_EQUAL_64(0, x22);
+ CHECK_EQUAL_64(0x80000000, x23);
+ CHECK_EQUAL_64(0, x24);
+ CHECK_EQUAL_64(0x8000000000000000UL, x25);
+ CHECK_EQUAL_64(0, x26);
+ CHECK_EQUAL_64(0, x27);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x28);
+ CHECK_EQUAL_64(0, x29);
+ CHECK_EQUAL_64(0, x18);
+ CHECK_EQUAL_64(0, x19);
+ CHECK_EQUAL_64(0, x20);
+ CHECK_EQUAL_64(0, x21);
TEARDOWN();
}
@@ -1530,13 +1530,13 @@ TEST(rbit_rev) {
RUN();
- ASSERT_EQUAL_64(0x084c2a6e, x0);
- ASSERT_EQUAL_64(0x084c2a6e195d3b7fUL, x1);
- ASSERT_EQUAL_64(0x54761032, x2);
- ASSERT_EQUAL_64(0xdcfe98ba54761032UL, x3);
- ASSERT_EQUAL_64(0x10325476, x4);
- ASSERT_EQUAL_64(0x98badcfe10325476UL, x5);
- ASSERT_EQUAL_64(0x1032547698badcfeUL, x6);
+ CHECK_EQUAL_64(0x084c2a6e, x0);
+ CHECK_EQUAL_64(0x084c2a6e195d3b7fUL, x1);
+ CHECK_EQUAL_64(0x54761032, x2);
+ CHECK_EQUAL_64(0xdcfe98ba54761032UL, x3);
+ CHECK_EQUAL_64(0x10325476, x4);
+ CHECK_EQUAL_64(0x98badcfe10325476UL, x5);
+ CHECK_EQUAL_64(0x1032547698badcfeUL, x6);
TEARDOWN();
}
@@ -1566,18 +1566,18 @@ TEST(clz_cls) {
RUN();
- ASSERT_EQUAL_64(8, x0);
- ASSERT_EQUAL_64(12, x1);
- ASSERT_EQUAL_64(0, x2);
- ASSERT_EQUAL_64(0, x3);
- ASSERT_EQUAL_64(32, x4);
- ASSERT_EQUAL_64(64, x5);
- ASSERT_EQUAL_64(7, x6);
- ASSERT_EQUAL_64(11, x7);
- ASSERT_EQUAL_64(12, x8);
- ASSERT_EQUAL_64(8, x9);
- ASSERT_EQUAL_64(31, x10);
- ASSERT_EQUAL_64(63, x11);
+ CHECK_EQUAL_64(8, x0);
+ CHECK_EQUAL_64(12, x1);
+ CHECK_EQUAL_64(0, x2);
+ CHECK_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(32, x4);
+ CHECK_EQUAL_64(64, x5);
+ CHECK_EQUAL_64(7, x6);
+ CHECK_EQUAL_64(11, x7);
+ CHECK_EQUAL_64(12, x8);
+ CHECK_EQUAL_64(8, x9);
+ CHECK_EQUAL_64(31, x10);
+ CHECK_EQUAL_64(63, x11);
TEARDOWN();
}
@@ -1615,8 +1615,8 @@ TEST(label) {
RUN();
- ASSERT_EQUAL_64(0x1, x0);
- ASSERT_EQUAL_64(0x1, x1);
+ CHECK_EQUAL_64(0x1, x0);
+ CHECK_EQUAL_64(0x1, x1);
TEARDOWN();
}
@@ -1649,7 +1649,7 @@ TEST(branch_at_start) {
RUN();
- ASSERT_EQUAL_64(0x1, x0);
+ CHECK_EQUAL_64(0x1, x0);
TEARDOWN();
}
@@ -1693,8 +1693,8 @@ TEST(adr) {
RUN();
- ASSERT_EQUAL_64(0x0, x0);
- ASSERT_EQUAL_64(0x0, x1);
+ CHECK_EQUAL_64(0x0, x0);
+ CHECK_EQUAL_64(0x0, x1);
TEARDOWN();
}
@@ -1759,7 +1759,7 @@ TEST(adr_far) {
RUN();
- ASSERT_EQUAL_64(0xf, x0);
+ CHECK_EQUAL_64(0xf, x0);
TEARDOWN();
}
@@ -1849,7 +1849,7 @@ TEST(branch_cond) {
RUN();
- ASSERT_EQUAL_64(0x1, x0);
+ CHECK_EQUAL_64(0x1, x0);
TEARDOWN();
}
@@ -1896,9 +1896,9 @@ TEST(branch_to_reg) {
RUN();
- ASSERT_EQUAL_64(core.xreg(3) + kInstructionSize, x0);
- ASSERT_EQUAL_64(42, x1);
- ASSERT_EQUAL_64(84, x2);
+ CHECK_EQUAL_64(core.xreg(3) + kInstructionSize, x0);
+ CHECK_EQUAL_64(42, x1);
+ CHECK_EQUAL_64(84, x2);
TEARDOWN();
}
@@ -1966,12 +1966,12 @@ TEST(compare_branch) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(0, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0, x3);
- ASSERT_EQUAL_64(1, x4);
- ASSERT_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(0, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(1, x4);
+ CHECK_EQUAL_64(0, x5);
TEARDOWN();
}
@@ -2019,10 +2019,10 @@ TEST(test_branch) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(0, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(0, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0, x3);
TEARDOWN();
}
@@ -2095,8 +2095,8 @@ TEST(far_branch_backward) {
RUN();
- ASSERT_EQUAL_64(0x7, x0);
- ASSERT_EQUAL_64(0x1, x1);
+ CHECK_EQUAL_64(0x7, x0);
+ CHECK_EQUAL_64(0x1, x1);
TEARDOWN();
}
@@ -2165,8 +2165,8 @@ TEST(far_branch_simple_veneer) {
RUN();
- ASSERT_EQUAL_64(0x7, x0);
- ASSERT_EQUAL_64(0x1, x1);
+ CHECK_EQUAL_64(0x7, x0);
+ CHECK_EQUAL_64(0x1, x1);
TEARDOWN();
}
@@ -2260,8 +2260,8 @@ TEST(far_branch_veneer_link_chain) {
RUN();
- ASSERT_EQUAL_64(0x7, x0);
- ASSERT_EQUAL_64(0x1, x1);
+ CHECK_EQUAL_64(0x7, x0);
+ CHECK_EQUAL_64(0x1, x1);
TEARDOWN();
}
@@ -2350,8 +2350,8 @@ TEST(far_branch_veneer_broken_link_chain) {
RUN();
- ASSERT_EQUAL_64(0x3, x0);
- ASSERT_EQUAL_64(0x1, x1);
+ CHECK_EQUAL_64(0x3, x0);
+ CHECK_EQUAL_64(0x1, x1);
TEARDOWN();
}
@@ -2408,7 +2408,7 @@ TEST(branch_type) {
RUN();
- ASSERT_EQUAL_64(0x0, x0);
+ CHECK_EQUAL_64(0x0, x0);
TEARDOWN();
}
@@ -2440,18 +2440,18 @@ TEST(ldr_str_offset) {
RUN();
- ASSERT_EQUAL_64(0x76543210, x0);
- ASSERT_EQUAL_64(0x76543210, dst[0]);
- ASSERT_EQUAL_64(0xfedcba98, x1);
- ASSERT_EQUAL_64(0xfedcba9800000000UL, dst[1]);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, x2);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, dst[2]);
- ASSERT_EQUAL_64(0x32, x3);
- ASSERT_EQUAL_64(0x3200, dst[3]);
- ASSERT_EQUAL_64(0x7654, x4);
- ASSERT_EQUAL_64(0x765400, dst[4]);
- ASSERT_EQUAL_64(src_base, x17);
- ASSERT_EQUAL_64(dst_base, x18);
+ CHECK_EQUAL_64(0x76543210, x0);
+ CHECK_EQUAL_64(0x76543210, dst[0]);
+ CHECK_EQUAL_64(0xfedcba98, x1);
+ CHECK_EQUAL_64(0xfedcba9800000000UL, dst[1]);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, x2);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, dst[2]);
+ CHECK_EQUAL_64(0x32, x3);
+ CHECK_EQUAL_64(0x3200, dst[3]);
+ CHECK_EQUAL_64(0x7654, x4);
+ CHECK_EQUAL_64(0x765400, dst[4]);
+ CHECK_EQUAL_64(src_base, x17);
+ CHECK_EQUAL_64(dst_base, x18);
TEARDOWN();
}
@@ -2489,18 +2489,18 @@ TEST(ldr_str_wide) {
RUN();
- ASSERT_EQUAL_32(8191, w0);
- ASSERT_EQUAL_32(8191, dst[8191]);
- ASSERT_EQUAL_64(src_base, x22);
- ASSERT_EQUAL_64(dst_base, x23);
- ASSERT_EQUAL_32(0, w1);
- ASSERT_EQUAL_32(0, dst[0]);
- ASSERT_EQUAL_64(src_base + 4096 * sizeof(src[0]), x24);
- ASSERT_EQUAL_64(dst_base + 4096 * sizeof(dst[0]), x25);
- ASSERT_EQUAL_32(6144, w2);
- ASSERT_EQUAL_32(6144, dst[6144]);
- ASSERT_EQUAL_64(src_base + 6144 * sizeof(src[0]), x26);
- ASSERT_EQUAL_64(dst_base + 6144 * sizeof(dst[0]), x27);
+ CHECK_EQUAL_32(8191, w0);
+ CHECK_EQUAL_32(8191, dst[8191]);
+ CHECK_EQUAL_64(src_base, x22);
+ CHECK_EQUAL_64(dst_base, x23);
+ CHECK_EQUAL_32(0, w1);
+ CHECK_EQUAL_32(0, dst[0]);
+ CHECK_EQUAL_64(src_base + 4096 * sizeof(src[0]), x24);
+ CHECK_EQUAL_64(dst_base + 4096 * sizeof(dst[0]), x25);
+ CHECK_EQUAL_32(6144, w2);
+ CHECK_EQUAL_32(6144, dst[6144]);
+ CHECK_EQUAL_64(src_base + 6144 * sizeof(src[0]), x26);
+ CHECK_EQUAL_64(dst_base + 6144 * sizeof(dst[0]), x27);
TEARDOWN();
}
@@ -2540,26 +2540,26 @@ TEST(ldr_str_preindex) {
RUN();
- ASSERT_EQUAL_64(0xfedcba98, x0);
- ASSERT_EQUAL_64(0xfedcba9800000000UL, dst[1]);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, x1);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, dst[2]);
- ASSERT_EQUAL_64(0x01234567, x2);
- ASSERT_EQUAL_64(0x0123456700000000UL, dst[4]);
- ASSERT_EQUAL_64(0x32, x3);
- ASSERT_EQUAL_64(0x3200, dst[3]);
- ASSERT_EQUAL_64(0x9876, x4);
- ASSERT_EQUAL_64(0x987600, dst[5]);
- ASSERT_EQUAL_64(src_base + 4, x17);
- ASSERT_EQUAL_64(dst_base + 12, x18);
- ASSERT_EQUAL_64(src_base + 8, x19);
- ASSERT_EQUAL_64(dst_base + 16, x20);
- ASSERT_EQUAL_64(src_base + 12, x21);
- ASSERT_EQUAL_64(dst_base + 36, x22);
- ASSERT_EQUAL_64(src_base + 1, x23);
- ASSERT_EQUAL_64(dst_base + 25, x24);
- ASSERT_EQUAL_64(src_base + 3, x25);
- ASSERT_EQUAL_64(dst_base + 41, x26);
+ CHECK_EQUAL_64(0xfedcba98, x0);
+ CHECK_EQUAL_64(0xfedcba9800000000UL, dst[1]);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, x1);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, dst[2]);
+ CHECK_EQUAL_64(0x01234567, x2);
+ CHECK_EQUAL_64(0x0123456700000000UL, dst[4]);
+ CHECK_EQUAL_64(0x32, x3);
+ CHECK_EQUAL_64(0x3200, dst[3]);
+ CHECK_EQUAL_64(0x9876, x4);
+ CHECK_EQUAL_64(0x987600, dst[5]);
+ CHECK_EQUAL_64(src_base + 4, x17);
+ CHECK_EQUAL_64(dst_base + 12, x18);
+ CHECK_EQUAL_64(src_base + 8, x19);
+ CHECK_EQUAL_64(dst_base + 16, x20);
+ CHECK_EQUAL_64(src_base + 12, x21);
+ CHECK_EQUAL_64(dst_base + 36, x22);
+ CHECK_EQUAL_64(src_base + 1, x23);
+ CHECK_EQUAL_64(dst_base + 25, x24);
+ CHECK_EQUAL_64(src_base + 3, x25);
+ CHECK_EQUAL_64(dst_base + 41, x26);
TEARDOWN();
}
@@ -2599,26 +2599,26 @@ TEST(ldr_str_postindex) {
RUN();
- ASSERT_EQUAL_64(0xfedcba98, x0);
- ASSERT_EQUAL_64(0xfedcba9800000000UL, dst[1]);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, x1);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, dst[2]);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, x2);
- ASSERT_EQUAL_64(0x0123456789abcdefUL, dst[4]);
- ASSERT_EQUAL_64(0x32, x3);
- ASSERT_EQUAL_64(0x3200, dst[3]);
- ASSERT_EQUAL_64(0x9876, x4);
- ASSERT_EQUAL_64(0x987600, dst[5]);
- ASSERT_EQUAL_64(src_base + 8, x17);
- ASSERT_EQUAL_64(dst_base + 24, x18);
- ASSERT_EQUAL_64(src_base + 16, x19);
- ASSERT_EQUAL_64(dst_base + 32, x20);
- ASSERT_EQUAL_64(src_base, x21);
- ASSERT_EQUAL_64(dst_base, x22);
- ASSERT_EQUAL_64(src_base + 2, x23);
- ASSERT_EQUAL_64(dst_base + 30, x24);
- ASSERT_EQUAL_64(src_base, x25);
- ASSERT_EQUAL_64(dst_base, x26);
+ CHECK_EQUAL_64(0xfedcba98, x0);
+ CHECK_EQUAL_64(0xfedcba9800000000UL, dst[1]);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, x1);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, dst[2]);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, x2);
+ CHECK_EQUAL_64(0x0123456789abcdefUL, dst[4]);
+ CHECK_EQUAL_64(0x32, x3);
+ CHECK_EQUAL_64(0x3200, dst[3]);
+ CHECK_EQUAL_64(0x9876, x4);
+ CHECK_EQUAL_64(0x987600, dst[5]);
+ CHECK_EQUAL_64(src_base + 8, x17);
+ CHECK_EQUAL_64(dst_base + 24, x18);
+ CHECK_EQUAL_64(src_base + 16, x19);
+ CHECK_EQUAL_64(dst_base + 32, x20);
+ CHECK_EQUAL_64(src_base, x21);
+ CHECK_EQUAL_64(dst_base, x22);
+ CHECK_EQUAL_64(src_base + 2, x23);
+ CHECK_EQUAL_64(dst_base + 30, x24);
+ CHECK_EQUAL_64(src_base, x25);
+ CHECK_EQUAL_64(dst_base, x26);
TEARDOWN();
}
@@ -2647,16 +2647,16 @@ TEST(load_signed) {
RUN();
- ASSERT_EQUAL_64(0xffffff80, x0);
- ASSERT_EQUAL_64(0x0000007f, x1);
- ASSERT_EQUAL_64(0xffff8080, x2);
- ASSERT_EQUAL_64(0x00007f7f, x3);
- ASSERT_EQUAL_64(0xffffffffffffff80UL, x4);
- ASSERT_EQUAL_64(0x000000000000007fUL, x5);
- ASSERT_EQUAL_64(0xffffffffffff8080UL, x6);
- ASSERT_EQUAL_64(0x0000000000007f7fUL, x7);
- ASSERT_EQUAL_64(0xffffffff80008080UL, x8);
- ASSERT_EQUAL_64(0x000000007fff7f7fUL, x9);
+ CHECK_EQUAL_64(0xffffff80, x0);
+ CHECK_EQUAL_64(0x0000007f, x1);
+ CHECK_EQUAL_64(0xffff8080, x2);
+ CHECK_EQUAL_64(0x00007f7f, x3);
+ CHECK_EQUAL_64(0xffffffffffffff80UL, x4);
+ CHECK_EQUAL_64(0x000000000000007fUL, x5);
+ CHECK_EQUAL_64(0xffffffffffff8080UL, x6);
+ CHECK_EQUAL_64(0x0000000000007f7fUL, x7);
+ CHECK_EQUAL_64(0xffffffff80008080UL, x8);
+ CHECK_EQUAL_64(0x000000007fff7f7fUL, x9);
TEARDOWN();
}
@@ -2696,15 +2696,15 @@ TEST(load_store_regoffset) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(0x0000000300000002UL, x1);
- ASSERT_EQUAL_64(3, x2);
- ASSERT_EQUAL_64(3, x3);
- ASSERT_EQUAL_64(2, x4);
- ASSERT_EQUAL_32(1, dst[0]);
- ASSERT_EQUAL_32(2, dst[1]);
- ASSERT_EQUAL_32(3, dst[2]);
- ASSERT_EQUAL_32(3, dst[3]);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(0x0000000300000002UL, x1);
+ CHECK_EQUAL_64(3, x2);
+ CHECK_EQUAL_64(3, x3);
+ CHECK_EQUAL_64(2, x4);
+ CHECK_EQUAL_32(1, dst[0]);
+ CHECK_EQUAL_32(2, dst[1]);
+ CHECK_EQUAL_32(3, dst[2]);
+ CHECK_EQUAL_32(3, dst[3]);
TEARDOWN();
}
@@ -2736,18 +2736,18 @@ TEST(load_store_float) {
RUN();
- ASSERT_EQUAL_FP32(2.0, s0);
- ASSERT_EQUAL_FP32(2.0, dst[0]);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(1.0, dst[2]);
- ASSERT_EQUAL_FP32(3.0, s2);
- ASSERT_EQUAL_FP32(3.0, dst[1]);
- ASSERT_EQUAL_64(src_base, x17);
- ASSERT_EQUAL_64(dst_base + sizeof(dst[0]), x18);
- ASSERT_EQUAL_64(src_base + sizeof(src[0]), x19);
- ASSERT_EQUAL_64(dst_base + 2 * sizeof(dst[0]), x20);
- ASSERT_EQUAL_64(src_base + 2 * sizeof(src[0]), x21);
- ASSERT_EQUAL_64(dst_base, x22);
+ CHECK_EQUAL_FP32(2.0, s0);
+ CHECK_EQUAL_FP32(2.0, dst[0]);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(1.0, dst[2]);
+ CHECK_EQUAL_FP32(3.0, s2);
+ CHECK_EQUAL_FP32(3.0, dst[1]);
+ CHECK_EQUAL_64(src_base, x17);
+ CHECK_EQUAL_64(dst_base + sizeof(dst[0]), x18);
+ CHECK_EQUAL_64(src_base + sizeof(src[0]), x19);
+ CHECK_EQUAL_64(dst_base + 2 * sizeof(dst[0]), x20);
+ CHECK_EQUAL_64(src_base + 2 * sizeof(src[0]), x21);
+ CHECK_EQUAL_64(dst_base, x22);
TEARDOWN();
}
@@ -2779,18 +2779,18 @@ TEST(load_store_double) {
RUN();
- ASSERT_EQUAL_FP64(2.0, d0);
- ASSERT_EQUAL_FP64(2.0, dst[0]);
- ASSERT_EQUAL_FP64(1.0, d1);
- ASSERT_EQUAL_FP64(1.0, dst[2]);
- ASSERT_EQUAL_FP64(3.0, d2);
- ASSERT_EQUAL_FP64(3.0, dst[1]);
- ASSERT_EQUAL_64(src_base, x17);
- ASSERT_EQUAL_64(dst_base + sizeof(dst[0]), x18);
- ASSERT_EQUAL_64(src_base + sizeof(src[0]), x19);
- ASSERT_EQUAL_64(dst_base + 2 * sizeof(dst[0]), x20);
- ASSERT_EQUAL_64(src_base + 2 * sizeof(src[0]), x21);
- ASSERT_EQUAL_64(dst_base, x22);
+ CHECK_EQUAL_FP64(2.0, d0);
+ CHECK_EQUAL_FP64(2.0, dst[0]);
+ CHECK_EQUAL_FP64(1.0, d1);
+ CHECK_EQUAL_FP64(1.0, dst[2]);
+ CHECK_EQUAL_FP64(3.0, d2);
+ CHECK_EQUAL_FP64(3.0, dst[1]);
+ CHECK_EQUAL_64(src_base, x17);
+ CHECK_EQUAL_64(dst_base + sizeof(dst[0]), x18);
+ CHECK_EQUAL_64(src_base + sizeof(src[0]), x19);
+ CHECK_EQUAL_64(dst_base + 2 * sizeof(dst[0]), x20);
+ CHECK_EQUAL_64(src_base + 2 * sizeof(src[0]), x21);
+ CHECK_EQUAL_64(dst_base, x22);
TEARDOWN();
}
@@ -2814,13 +2814,13 @@ TEST(ldp_stp_float) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s31);
- ASSERT_EQUAL_FP32(2.0, s0);
- ASSERT_EQUAL_FP32(0.0, dst[0]);
- ASSERT_EQUAL_FP32(2.0, dst[1]);
- ASSERT_EQUAL_FP32(1.0, dst[2]);
- ASSERT_EQUAL_64(src_base + 2 * sizeof(src[0]), x16);
- ASSERT_EQUAL_64(dst_base + sizeof(dst[1]), x17);
+ CHECK_EQUAL_FP32(1.0, s31);
+ CHECK_EQUAL_FP32(2.0, s0);
+ CHECK_EQUAL_FP32(0.0, dst[0]);
+ CHECK_EQUAL_FP32(2.0, dst[1]);
+ CHECK_EQUAL_FP32(1.0, dst[2]);
+ CHECK_EQUAL_64(src_base + 2 * sizeof(src[0]), x16);
+ CHECK_EQUAL_64(dst_base + sizeof(dst[1]), x17);
TEARDOWN();
}
@@ -2844,13 +2844,13 @@ TEST(ldp_stp_double) {
RUN();
- ASSERT_EQUAL_FP64(1.0, d31);
- ASSERT_EQUAL_FP64(2.0, d0);
- ASSERT_EQUAL_FP64(0.0, dst[0]);
- ASSERT_EQUAL_FP64(2.0, dst[1]);
- ASSERT_EQUAL_FP64(1.0, dst[2]);
- ASSERT_EQUAL_64(src_base + 2 * sizeof(src[0]), x16);
- ASSERT_EQUAL_64(dst_base + sizeof(dst[1]), x17);
+ CHECK_EQUAL_FP64(1.0, d31);
+ CHECK_EQUAL_FP64(2.0, d0);
+ CHECK_EQUAL_FP64(0.0, dst[0]);
+ CHECK_EQUAL_FP64(2.0, dst[1]);
+ CHECK_EQUAL_FP64(1.0, dst[2]);
+ CHECK_EQUAL_64(src_base + 2 * sizeof(src[0]), x16);
+ CHECK_EQUAL_64(dst_base + sizeof(dst[1]), x17);
TEARDOWN();
}
@@ -2885,27 +2885,27 @@ TEST(ldp_stp_offset) {
RUN();
- ASSERT_EQUAL_64(0x44556677, x0);
- ASSERT_EQUAL_64(0x00112233, x1);
- ASSERT_EQUAL_64(0x0011223344556677UL, dst[0]);
- ASSERT_EQUAL_64(0x00112233, x2);
- ASSERT_EQUAL_64(0xccddeeff, x3);
- ASSERT_EQUAL_64(0xccddeeff00112233UL, dst[1]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x4);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, dst[2]);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, x5);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, dst[3]);
- ASSERT_EQUAL_64(0x8899aabb, x6);
- ASSERT_EQUAL_64(0xbbaa9988, x7);
- ASSERT_EQUAL_64(0xbbaa99888899aabbUL, dst[4]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x8);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, dst[5]);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, x9);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, dst[6]);
- ASSERT_EQUAL_64(src_base, x16);
- ASSERT_EQUAL_64(dst_base, x17);
- ASSERT_EQUAL_64(src_base + 24, x18);
- ASSERT_EQUAL_64(dst_base + 56, x19);
+ CHECK_EQUAL_64(0x44556677, x0);
+ CHECK_EQUAL_64(0x00112233, x1);
+ CHECK_EQUAL_64(0x0011223344556677UL, dst[0]);
+ CHECK_EQUAL_64(0x00112233, x2);
+ CHECK_EQUAL_64(0xccddeeff, x3);
+ CHECK_EQUAL_64(0xccddeeff00112233UL, dst[1]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x4);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, dst[2]);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, x5);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, dst[3]);
+ CHECK_EQUAL_64(0x8899aabb, x6);
+ CHECK_EQUAL_64(0xbbaa9988, x7);
+ CHECK_EQUAL_64(0xbbaa99888899aabbUL, dst[4]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x8);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, dst[5]);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, x9);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, dst[6]);
+ CHECK_EQUAL_64(src_base, x16);
+ CHECK_EQUAL_64(dst_base, x17);
+ CHECK_EQUAL_64(src_base + 24, x18);
+ CHECK_EQUAL_64(dst_base + 56, x19);
TEARDOWN();
}
@@ -2940,27 +2940,27 @@ TEST(ldnp_stnp_offset) {
RUN();
- ASSERT_EQUAL_64(0x44556677, x0);
- ASSERT_EQUAL_64(0x00112233, x1);
- ASSERT_EQUAL_64(0x0011223344556677UL, dst[0]);
- ASSERT_EQUAL_64(0x00112233, x2);
- ASSERT_EQUAL_64(0xccddeeff, x3);
- ASSERT_EQUAL_64(0xccddeeff00112233UL, dst[1]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x4);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, dst[2]);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, x5);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, dst[3]);
- ASSERT_EQUAL_64(0x8899aabb, x6);
- ASSERT_EQUAL_64(0xbbaa9988, x7);
- ASSERT_EQUAL_64(0xbbaa99888899aabbUL, dst[4]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x8);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, dst[5]);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, x9);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, dst[6]);
- ASSERT_EQUAL_64(src_base, x16);
- ASSERT_EQUAL_64(dst_base, x17);
- ASSERT_EQUAL_64(src_base + 24, x18);
- ASSERT_EQUAL_64(dst_base + 56, x19);
+ CHECK_EQUAL_64(0x44556677, x0);
+ CHECK_EQUAL_64(0x00112233, x1);
+ CHECK_EQUAL_64(0x0011223344556677UL, dst[0]);
+ CHECK_EQUAL_64(0x00112233, x2);
+ CHECK_EQUAL_64(0xccddeeff, x3);
+ CHECK_EQUAL_64(0xccddeeff00112233UL, dst[1]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x4);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, dst[2]);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, x5);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, dst[3]);
+ CHECK_EQUAL_64(0x8899aabb, x6);
+ CHECK_EQUAL_64(0xbbaa9988, x7);
+ CHECK_EQUAL_64(0xbbaa99888899aabbUL, dst[4]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x8);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, dst[5]);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, x9);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, dst[6]);
+ CHECK_EQUAL_64(src_base, x16);
+ CHECK_EQUAL_64(dst_base, x17);
+ CHECK_EQUAL_64(src_base + 24, x18);
+ CHECK_EQUAL_64(dst_base + 56, x19);
TEARDOWN();
}
@@ -2996,26 +2996,26 @@ TEST(ldp_stp_preindex) {
RUN();
- ASSERT_EQUAL_64(0x00112233, x0);
- ASSERT_EQUAL_64(0xccddeeff, x1);
- ASSERT_EQUAL_64(0x44556677, x2);
- ASSERT_EQUAL_64(0x00112233, x3);
- ASSERT_EQUAL_64(0xccddeeff00112233UL, dst[0]);
- ASSERT_EQUAL_64(0x0000000000112233UL, dst[1]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x4);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, x5);
- ASSERT_EQUAL_64(0x0011223344556677UL, x6);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x7);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, dst[2]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, dst[3]);
- ASSERT_EQUAL_64(0x0011223344556677UL, dst[4]);
- ASSERT_EQUAL_64(src_base, x16);
- ASSERT_EQUAL_64(dst_base, x17);
- ASSERT_EQUAL_64(dst_base + 16, x18);
- ASSERT_EQUAL_64(src_base + 4, x19);
- ASSERT_EQUAL_64(dst_base + 4, x20);
- ASSERT_EQUAL_64(src_base + 8, x21);
- ASSERT_EQUAL_64(dst_base + 24, x22);
+ CHECK_EQUAL_64(0x00112233, x0);
+ CHECK_EQUAL_64(0xccddeeff, x1);
+ CHECK_EQUAL_64(0x44556677, x2);
+ CHECK_EQUAL_64(0x00112233, x3);
+ CHECK_EQUAL_64(0xccddeeff00112233UL, dst[0]);
+ CHECK_EQUAL_64(0x0000000000112233UL, dst[1]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x4);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, x5);
+ CHECK_EQUAL_64(0x0011223344556677UL, x6);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x7);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, dst[2]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, dst[3]);
+ CHECK_EQUAL_64(0x0011223344556677UL, dst[4]);
+ CHECK_EQUAL_64(src_base, x16);
+ CHECK_EQUAL_64(dst_base, x17);
+ CHECK_EQUAL_64(dst_base + 16, x18);
+ CHECK_EQUAL_64(src_base + 4, x19);
+ CHECK_EQUAL_64(dst_base + 4, x20);
+ CHECK_EQUAL_64(src_base + 8, x21);
+ CHECK_EQUAL_64(dst_base + 24, x22);
TEARDOWN();
}
@@ -3051,26 +3051,26 @@ TEST(ldp_stp_postindex) {
RUN();
- ASSERT_EQUAL_64(0x44556677, x0);
- ASSERT_EQUAL_64(0x00112233, x1);
- ASSERT_EQUAL_64(0x00112233, x2);
- ASSERT_EQUAL_64(0xccddeeff, x3);
- ASSERT_EQUAL_64(0x4455667700112233UL, dst[0]);
- ASSERT_EQUAL_64(0x0000000000112233UL, dst[1]);
- ASSERT_EQUAL_64(0x0011223344556677UL, x4);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x5);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, x6);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, x7);
- ASSERT_EQUAL_64(0xffeeddccbbaa9988UL, dst[2]);
- ASSERT_EQUAL_64(0x8899aabbccddeeffUL, dst[3]);
- ASSERT_EQUAL_64(0x0011223344556677UL, dst[4]);
- ASSERT_EQUAL_64(src_base, x16);
- ASSERT_EQUAL_64(dst_base, x17);
- ASSERT_EQUAL_64(dst_base + 16, x18);
- ASSERT_EQUAL_64(src_base + 4, x19);
- ASSERT_EQUAL_64(dst_base + 4, x20);
- ASSERT_EQUAL_64(src_base + 8, x21);
- ASSERT_EQUAL_64(dst_base + 24, x22);
+ CHECK_EQUAL_64(0x44556677, x0);
+ CHECK_EQUAL_64(0x00112233, x1);
+ CHECK_EQUAL_64(0x00112233, x2);
+ CHECK_EQUAL_64(0xccddeeff, x3);
+ CHECK_EQUAL_64(0x4455667700112233UL, dst[0]);
+ CHECK_EQUAL_64(0x0000000000112233UL, dst[1]);
+ CHECK_EQUAL_64(0x0011223344556677UL, x4);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x5);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, x6);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, x7);
+ CHECK_EQUAL_64(0xffeeddccbbaa9988UL, dst[2]);
+ CHECK_EQUAL_64(0x8899aabbccddeeffUL, dst[3]);
+ CHECK_EQUAL_64(0x0011223344556677UL, dst[4]);
+ CHECK_EQUAL_64(src_base, x16);
+ CHECK_EQUAL_64(dst_base, x17);
+ CHECK_EQUAL_64(dst_base + 16, x18);
+ CHECK_EQUAL_64(src_base + 4, x19);
+ CHECK_EQUAL_64(dst_base + 4, x20);
+ CHECK_EQUAL_64(src_base + 8, x21);
+ CHECK_EQUAL_64(dst_base + 24, x22);
TEARDOWN();
}
@@ -3090,8 +3090,8 @@ TEST(ldp_sign_extend) {
RUN();
- ASSERT_EQUAL_64(0xffffffff80000000UL, x0);
- ASSERT_EQUAL_64(0x000000007fffffffUL, x1);
+ CHECK_EQUAL_64(0xffffffff80000000UL, x0);
+ CHECK_EQUAL_64(0x000000007fffffffUL, x1);
TEARDOWN();
}
@@ -3124,19 +3124,19 @@ TEST(ldur_stur) {
RUN();
- ASSERT_EQUAL_64(0x6789abcd, x0);
- ASSERT_EQUAL_64(0x6789abcd0000L, dst[0]);
- ASSERT_EQUAL_64(0xabcdef0123456789L, x1);
- ASSERT_EQUAL_64(0xcdef012345678900L, dst[1]);
- ASSERT_EQUAL_64(0x000000ab, dst[2]);
- ASSERT_EQUAL_64(0xabcdef01, x2);
- ASSERT_EQUAL_64(0x00abcdef01000000L, dst[3]);
- ASSERT_EQUAL_64(0x00000001, x3);
- ASSERT_EQUAL_64(0x0100000000000000L, dst[4]);
- ASSERT_EQUAL_64(src_base, x17);
- ASSERT_EQUAL_64(dst_base, x18);
- ASSERT_EQUAL_64(src_base + 16, x19);
- ASSERT_EQUAL_64(dst_base + 32, x20);
+ CHECK_EQUAL_64(0x6789abcd, x0);
+ CHECK_EQUAL_64(0x6789abcd0000L, dst[0]);
+ CHECK_EQUAL_64(0xabcdef0123456789L, x1);
+ CHECK_EQUAL_64(0xcdef012345678900L, dst[1]);
+ CHECK_EQUAL_64(0x000000ab, dst[2]);
+ CHECK_EQUAL_64(0xabcdef01, x2);
+ CHECK_EQUAL_64(0x00abcdef01000000L, dst[3]);
+ CHECK_EQUAL_64(0x00000001, x3);
+ CHECK_EQUAL_64(0x0100000000000000L, dst[4]);
+ CHECK_EQUAL_64(src_base, x17);
+ CHECK_EQUAL_64(dst_base, x18);
+ CHECK_EQUAL_64(src_base + 16, x19);
+ CHECK_EQUAL_64(dst_base + 32, x20);
TEARDOWN();
}
@@ -3157,10 +3157,10 @@ TEST(ldr_literal) {
RUN();
- ASSERT_EQUAL_64(0x1234567890abcdefUL, x2);
- ASSERT_EQUAL_64(0xfedcba09, x3);
- ASSERT_EQUAL_FP64(1.234, d13);
- ASSERT_EQUAL_FP32(2.5, s25);
+ CHECK_EQUAL_64(0x1234567890abcdefUL, x2);
+ CHECK_EQUAL_64(0xfedcba09, x3);
+ CHECK_EQUAL_FP64(1.234, d13);
+ CHECK_EQUAL_FP32(2.5, s25);
TEARDOWN();
}
@@ -3169,7 +3169,7 @@ TEST(ldr_literal) {
static void LdrLiteralRangeHelper(ptrdiff_t range_,
LiteralPoolEmitOption option,
bool expect_dump) {
- ASSERT(range_ > 0);
+ DCHECK(range_ > 0);
SETUP_SIZE(range_ + 1024);
Label label_1, label_2;
@@ -3188,19 +3188,19 @@ static void LdrLiteralRangeHelper(ptrdiff_t range_,
START();
// Force a pool dump so the pool starts off empty.
__ EmitLiteralPool(JumpRequired);
- ASSERT_LITERAL_POOL_SIZE(0);
+ DCHECK_LITERAL_POOL_SIZE(0);
__ Ldr(x0, 0x1234567890abcdefUL);
__ Ldr(w1, 0xfedcba09);
__ Ldr(d0, 1.234);
__ Ldr(s1, 2.5);
- ASSERT_LITERAL_POOL_SIZE(4);
+ DCHECK_LITERAL_POOL_SIZE(4);
code_size += 4 * sizeof(Instr);
// Check that the requested range (allowing space for a branch over the pool)
// can be handled by this test.
- ASSERT((code_size + pool_guard_size) <= range);
+ DCHECK((code_size + pool_guard_size) <= range);
// Emit NOPs up to 'range', leaving space for the pool guard.
while ((code_size + pool_guard_size) < range) {
@@ -3214,41 +3214,41 @@ static void LdrLiteralRangeHelper(ptrdiff_t range_,
code_size += sizeof(Instr);
}
- ASSERT(code_size == range);
- ASSERT_LITERAL_POOL_SIZE(4);
+ DCHECK(code_size == range);
+ DCHECK_LITERAL_POOL_SIZE(4);
// Possibly generate a literal pool.
__ CheckLiteralPool(option);
__ Bind(&label_1);
if (expect_dump) {
- ASSERT_LITERAL_POOL_SIZE(0);
+ DCHECK_LITERAL_POOL_SIZE(0);
} else {
- ASSERT_LITERAL_POOL_SIZE(4);
+ DCHECK_LITERAL_POOL_SIZE(4);
}
// Force a pool flush to check that a second pool functions correctly.
__ EmitLiteralPool(JumpRequired);
- ASSERT_LITERAL_POOL_SIZE(0);
+ DCHECK_LITERAL_POOL_SIZE(0);
// These loads should be after the pool (and will require a new one).
__ Ldr(x4, 0x34567890abcdef12UL);
__ Ldr(w5, 0xdcba09fe);
__ Ldr(d4, 123.4);
__ Ldr(s5, 250.0);
- ASSERT_LITERAL_POOL_SIZE(4);
+ DCHECK_LITERAL_POOL_SIZE(4);
END();
RUN();
// Check that the literals loaded correctly.
- ASSERT_EQUAL_64(0x1234567890abcdefUL, x0);
- ASSERT_EQUAL_64(0xfedcba09, x1);
- ASSERT_EQUAL_FP64(1.234, d0);
- ASSERT_EQUAL_FP32(2.5, s1);
- ASSERT_EQUAL_64(0x34567890abcdef12UL, x4);
- ASSERT_EQUAL_64(0xdcba09fe, x5);
- ASSERT_EQUAL_FP64(123.4, d4);
- ASSERT_EQUAL_FP32(250.0, s5);
+ CHECK_EQUAL_64(0x1234567890abcdefUL, x0);
+ CHECK_EQUAL_64(0xfedcba09, x1);
+ CHECK_EQUAL_FP64(1.234, d0);
+ CHECK_EQUAL_FP32(2.5, s1);
+ CHECK_EQUAL_64(0x34567890abcdef12UL, x4);
+ CHECK_EQUAL_64(0xdcba09fe, x5);
+ CHECK_EQUAL_FP64(123.4, d4);
+ CHECK_EQUAL_FP32(250.0, s5);
TEARDOWN();
}
@@ -3335,25 +3335,25 @@ TEST(add_sub_imm) {
RUN();
- ASSERT_EQUAL_64(0x123, x10);
- ASSERT_EQUAL_64(0x123111, x11);
- ASSERT_EQUAL_64(0xabc000, x12);
- ASSERT_EQUAL_64(0x0, x13);
+ CHECK_EQUAL_64(0x123, x10);
+ CHECK_EQUAL_64(0x123111, x11);
+ CHECK_EQUAL_64(0xabc000, x12);
+ CHECK_EQUAL_64(0x0, x13);
- ASSERT_EQUAL_32(0x123, w14);
- ASSERT_EQUAL_32(0x123111, w15);
- ASSERT_EQUAL_32(0xabc000, w16);
- ASSERT_EQUAL_32(0x0, w17);
+ CHECK_EQUAL_32(0x123, w14);
+ CHECK_EQUAL_32(0x123111, w15);
+ CHECK_EQUAL_32(0xabc000, w16);
+ CHECK_EQUAL_32(0x0, w17);
- ASSERT_EQUAL_64(0xffffffffffffffffL, x20);
- ASSERT_EQUAL_64(0x1000, x21);
- ASSERT_EQUAL_64(0x111, x22);
- ASSERT_EQUAL_64(0x7fffffffffffffffL, x23);
+ CHECK_EQUAL_64(0xffffffffffffffffL, x20);
+ CHECK_EQUAL_64(0x1000, x21);
+ CHECK_EQUAL_64(0x111, x22);
+ CHECK_EQUAL_64(0x7fffffffffffffffL, x23);
- ASSERT_EQUAL_32(0xffffffff, w24);
- ASSERT_EQUAL_32(0x1000, w25);
- ASSERT_EQUAL_32(0x111, w26);
- ASSERT_EQUAL_32(0xffffffff, w27);
+ CHECK_EQUAL_32(0xffffffff, w24);
+ CHECK_EQUAL_32(0x1000, w25);
+ CHECK_EQUAL_32(0x111, w26);
+ CHECK_EQUAL_32(0xffffffff, w27);
TEARDOWN();
}
@@ -3382,17 +3382,17 @@ TEST(add_sub_wide_imm) {
RUN();
- ASSERT_EQUAL_64(0x1234567890abcdefUL, x10);
- ASSERT_EQUAL_64(0x100000000UL, x11);
+ CHECK_EQUAL_64(0x1234567890abcdefUL, x10);
+ CHECK_EQUAL_64(0x100000000UL, x11);
- ASSERT_EQUAL_32(0x12345678, w12);
- ASSERT_EQUAL_64(0x0, x13);
+ CHECK_EQUAL_32(0x12345678, w12);
+ CHECK_EQUAL_64(0x0, x13);
- ASSERT_EQUAL_32(kWMinInt, w18);
- ASSERT_EQUAL_32(kWMinInt, w19);
+ CHECK_EQUAL_32(kWMinInt, w18);
+ CHECK_EQUAL_32(kWMinInt, w19);
- ASSERT_EQUAL_64(-0x1234567890abcdefUL, x20);
- ASSERT_EQUAL_32(-0x12345678, w21);
+ CHECK_EQUAL_64(-0x1234567890abcdefUL, x20);
+ CHECK_EQUAL_32(-0x12345678, w21);
TEARDOWN();
}
@@ -3429,23 +3429,23 @@ TEST(add_sub_shifted) {
RUN();
- ASSERT_EQUAL_64(0xffffffffffffffffL, x10);
- ASSERT_EQUAL_64(0x23456789abcdef00L, x11);
- ASSERT_EQUAL_64(0x000123456789abcdL, x12);
- ASSERT_EQUAL_64(0x000123456789abcdL, x13);
- ASSERT_EQUAL_64(0xfffedcba98765432L, x14);
- ASSERT_EQUAL_64(0xff89abcd, x15);
- ASSERT_EQUAL_64(0xef89abcc, x18);
- ASSERT_EQUAL_64(0xef0123456789abccL, x19);
+ CHECK_EQUAL_64(0xffffffffffffffffL, x10);
+ CHECK_EQUAL_64(0x23456789abcdef00L, x11);
+ CHECK_EQUAL_64(0x000123456789abcdL, x12);
+ CHECK_EQUAL_64(0x000123456789abcdL, x13);
+ CHECK_EQUAL_64(0xfffedcba98765432L, x14);
+ CHECK_EQUAL_64(0xff89abcd, x15);
+ CHECK_EQUAL_64(0xef89abcc, x18);
+ CHECK_EQUAL_64(0xef0123456789abccL, x19);
- ASSERT_EQUAL_64(0x0123456789abcdefL, x20);
- ASSERT_EQUAL_64(0xdcba9876543210ffL, x21);
- ASSERT_EQUAL_64(0xfffedcba98765432L, x22);
- ASSERT_EQUAL_64(0xfffedcba98765432L, x23);
- ASSERT_EQUAL_64(0x000123456789abcdL, x24);
- ASSERT_EQUAL_64(0x00765432, x25);
- ASSERT_EQUAL_64(0x10765432, x26);
- ASSERT_EQUAL_64(0x10fedcba98765432L, x27);
+ CHECK_EQUAL_64(0x0123456789abcdefL, x20);
+ CHECK_EQUAL_64(0xdcba9876543210ffL, x21);
+ CHECK_EQUAL_64(0xfffedcba98765432L, x22);
+ CHECK_EQUAL_64(0xfffedcba98765432L, x23);
+ CHECK_EQUAL_64(0x000123456789abcdL, x24);
+ CHECK_EQUAL_64(0x00765432, x25);
+ CHECK_EQUAL_64(0x10765432, x26);
+ CHECK_EQUAL_64(0x10fedcba98765432L, x27);
TEARDOWN();
}
@@ -3491,32 +3491,32 @@ TEST(add_sub_extended) {
RUN();
- ASSERT_EQUAL_64(0xefL, x10);
- ASSERT_EQUAL_64(0x1deL, x11);
- ASSERT_EQUAL_64(0x337bcL, x12);
- ASSERT_EQUAL_64(0x89abcdef0L, x13);
+ CHECK_EQUAL_64(0xefL, x10);
+ CHECK_EQUAL_64(0x1deL, x11);
+ CHECK_EQUAL_64(0x337bcL, x12);
+ CHECK_EQUAL_64(0x89abcdef0L, x13);
- ASSERT_EQUAL_64(0xffffffffffffffefL, x14);
- ASSERT_EQUAL_64(0xffffffffffffffdeL, x15);
- ASSERT_EQUAL_64(0xffffffffffff37bcL, x16);
- ASSERT_EQUAL_64(0xfffffffc4d5e6f78L, x17);
- ASSERT_EQUAL_64(0x10L, x18);
- ASSERT_EQUAL_64(0x20L, x19);
- ASSERT_EQUAL_64(0xc840L, x20);
- ASSERT_EQUAL_64(0x3b2a19080L, x21);
+ CHECK_EQUAL_64(0xffffffffffffffefL, x14);
+ CHECK_EQUAL_64(0xffffffffffffffdeL, x15);
+ CHECK_EQUAL_64(0xffffffffffff37bcL, x16);
+ CHECK_EQUAL_64(0xfffffffc4d5e6f78L, x17);
+ CHECK_EQUAL_64(0x10L, x18);
+ CHECK_EQUAL_64(0x20L, x19);
+ CHECK_EQUAL_64(0xc840L, x20);
+ CHECK_EQUAL_64(0x3b2a19080L, x21);
- ASSERT_EQUAL_64(0x0123456789abce0fL, x22);
- ASSERT_EQUAL_64(0x0123456789abcdcfL, x23);
+ CHECK_EQUAL_64(0x0123456789abce0fL, x22);
+ CHECK_EQUAL_64(0x0123456789abcdcfL, x23);
- ASSERT_EQUAL_32(0x89abce2f, w24);
- ASSERT_EQUAL_32(0xffffffef, w25);
- ASSERT_EQUAL_32(0xffffffde, w26);
- ASSERT_EQUAL_32(0xc3b2a188, w27);
+ CHECK_EQUAL_32(0x89abce2f, w24);
+ CHECK_EQUAL_32(0xffffffef, w25);
+ CHECK_EQUAL_32(0xffffffde, w26);
+ CHECK_EQUAL_32(0xc3b2a188, w27);
- ASSERT_EQUAL_32(0x4d5e6f78, w28);
- ASSERT_EQUAL_64(0xfffffffc4d5e6f78L, x29);
+ CHECK_EQUAL_32(0x4d5e6f78, w28);
+ CHECK_EQUAL_64(0xfffffffc4d5e6f78L, x29);
- ASSERT_EQUAL_64(256, x30);
+ CHECK_EQUAL_64(256, x30);
TEARDOWN();
}
@@ -3550,19 +3550,19 @@ TEST(add_sub_negative) {
RUN();
- ASSERT_EQUAL_64(-42, x10);
- ASSERT_EQUAL_64(4000, x11);
- ASSERT_EQUAL_64(0x1122334455667700, x12);
+ CHECK_EQUAL_64(-42, x10);
+ CHECK_EQUAL_64(4000, x11);
+ CHECK_EQUAL_64(0x1122334455667700, x12);
- ASSERT_EQUAL_64(600, x13);
- ASSERT_EQUAL_64(5000, x14);
- ASSERT_EQUAL_64(0x1122334455667cdd, x15);
+ CHECK_EQUAL_64(600, x13);
+ CHECK_EQUAL_64(5000, x14);
+ CHECK_EQUAL_64(0x1122334455667cdd, x15);
- ASSERT_EQUAL_32(0x11223000, w19);
- ASSERT_EQUAL_32(398000, w20);
+ CHECK_EQUAL_32(0x11223000, w19);
+ CHECK_EQUAL_32(398000, w20);
- ASSERT_EQUAL_32(0x11223400, w21);
- ASSERT_EQUAL_32(402000, w22);
+ CHECK_EQUAL_32(0x11223400, w21);
+ CHECK_EQUAL_32(402000, w22);
TEARDOWN();
}
@@ -3598,9 +3598,9 @@ TEST(add_sub_zero) {
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(0, x1);
- ASSERT_EQUAL_64(0, x2);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(0, x1);
+ CHECK_EQUAL_64(0, x2);
TEARDOWN();
}
@@ -3666,20 +3666,20 @@ TEST(neg) {
RUN();
- ASSERT_EQUAL_64(0xfffffffffffffeddUL, x1);
- ASSERT_EQUAL_64(0xfffffedd, x2);
- ASSERT_EQUAL_64(0x1db97530eca86422UL, x3);
- ASSERT_EQUAL_64(0xd950c844, x4);
- ASSERT_EQUAL_64(0xe1db97530eca8643UL, x5);
- ASSERT_EQUAL_64(0xf7654322, x6);
- ASSERT_EQUAL_64(0x0076e5d4c3b2a191UL, x7);
- ASSERT_EQUAL_64(0x01d950c9, x8);
- ASSERT_EQUAL_64(0xffffff11, x9);
- ASSERT_EQUAL_64(0x0000000000000022UL, x10);
- ASSERT_EQUAL_64(0xfffcc844, x11);
- ASSERT_EQUAL_64(0x0000000000019088UL, x12);
- ASSERT_EQUAL_64(0x65432110, x13);
- ASSERT_EQUAL_64(0x0000000765432110UL, x14);
+ CHECK_EQUAL_64(0xfffffffffffffeddUL, x1);
+ CHECK_EQUAL_64(0xfffffedd, x2);
+ CHECK_EQUAL_64(0x1db97530eca86422UL, x3);
+ CHECK_EQUAL_64(0xd950c844, x4);
+ CHECK_EQUAL_64(0xe1db97530eca8643UL, x5);
+ CHECK_EQUAL_64(0xf7654322, x6);
+ CHECK_EQUAL_64(0x0076e5d4c3b2a191UL, x7);
+ CHECK_EQUAL_64(0x01d950c9, x8);
+ CHECK_EQUAL_64(0xffffff11, x9);
+ CHECK_EQUAL_64(0x0000000000000022UL, x10);
+ CHECK_EQUAL_64(0xfffcc844, x11);
+ CHECK_EQUAL_64(0x0000000000019088UL, x12);
+ CHECK_EQUAL_64(0x65432110, x13);
+ CHECK_EQUAL_64(0x0000000765432110UL, x14);
TEARDOWN();
}
@@ -3729,29 +3729,29 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_64(0xffffffffffffffffL, x5);
- ASSERT_EQUAL_64(1L << 60, x6);
- ASSERT_EQUAL_64(0xf0123456789abcddL, x7);
- ASSERT_EQUAL_64(0x0111111111111110L, x8);
- ASSERT_EQUAL_64(0x1222222222222221L, x9);
+ CHECK_EQUAL_64(0xffffffffffffffffL, x5);
+ CHECK_EQUAL_64(1L << 60, x6);
+ CHECK_EQUAL_64(0xf0123456789abcddL, x7);
+ CHECK_EQUAL_64(0x0111111111111110L, x8);
+ CHECK_EQUAL_64(0x1222222222222221L, x9);
- ASSERT_EQUAL_32(0xffffffff, w10);
- ASSERT_EQUAL_32(1 << 30, w11);
- ASSERT_EQUAL_32(0xf89abcdd, w12);
- ASSERT_EQUAL_32(0x91111110, w13);
- ASSERT_EQUAL_32(0x9a222221, w14);
+ CHECK_EQUAL_32(0xffffffff, w10);
+ CHECK_EQUAL_32(1 << 30, w11);
+ CHECK_EQUAL_32(0xf89abcdd, w12);
+ CHECK_EQUAL_32(0x91111110, w13);
+ CHECK_EQUAL_32(0x9a222221, w14);
- ASSERT_EQUAL_64(0xffffffffffffffffL + 1, x18);
- ASSERT_EQUAL_64((1L << 60) + 1, x19);
- ASSERT_EQUAL_64(0xf0123456789abcddL + 1, x20);
- ASSERT_EQUAL_64(0x0111111111111110L + 1, x21);
- ASSERT_EQUAL_64(0x1222222222222221L + 1, x22);
+ CHECK_EQUAL_64(0xffffffffffffffffL + 1, x18);
+ CHECK_EQUAL_64((1L << 60) + 1, x19);
+ CHECK_EQUAL_64(0xf0123456789abcddL + 1, x20);
+ CHECK_EQUAL_64(0x0111111111111110L + 1, x21);
+ CHECK_EQUAL_64(0x1222222222222221L + 1, x22);
- ASSERT_EQUAL_32(0xffffffff + 1, w23);
- ASSERT_EQUAL_32((1 << 30) + 1, w24);
- ASSERT_EQUAL_32(0xf89abcdd + 1, w25);
- ASSERT_EQUAL_32(0x91111110 + 1, w26);
- ASSERT_EQUAL_32(0x9a222221 + 1, w27);
+ CHECK_EQUAL_32(0xffffffff + 1, w23);
+ CHECK_EQUAL_32((1 << 30) + 1, w24);
+ CHECK_EQUAL_32(0xf89abcdd + 1, w25);
+ CHECK_EQUAL_32(0x91111110 + 1, w26);
+ CHECK_EQUAL_32(0x9a222221 + 1, w27);
// Check that adc correctly sets the condition flags.
START();
@@ -3764,8 +3764,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
- ASSERT_EQUAL_64(0, x10);
+ CHECK_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_64(0, x10);
START();
__ Mov(x0, 1);
@@ -3777,8 +3777,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
- ASSERT_EQUAL_64(0, x10);
+ CHECK_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_64(0, x10);
START();
__ Mov(x0, 0x10);
@@ -3790,8 +3790,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(NVFlag);
- ASSERT_EQUAL_64(0x8000000000000000L, x10);
+ CHECK_EQUAL_NZCV(NVFlag);
+ CHECK_EQUAL_64(0x8000000000000000L, x10);
// Check that sbc correctly sets the condition flags.
START();
@@ -3804,8 +3804,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0, x10);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0, x10);
START();
__ Mov(x0, 1);
@@ -3817,8 +3817,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x8000000000000001L, x10);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x8000000000000001L, x10);
START();
__ Mov(x0, 0);
@@ -3829,8 +3829,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(ZFlag);
- ASSERT_EQUAL_64(0, x10);
+ CHECK_EQUAL_NZCV(ZFlag);
+ CHECK_EQUAL_64(0, x10);
START()
__ Mov(w0, 0x7fffffff);
@@ -3841,8 +3841,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x80000000, x10);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x80000000, x10);
START();
// Clear the C flag.
@@ -3852,8 +3852,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x8000000000000000L, x10);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x8000000000000000L, x10);
START()
__ Mov(x0, 0);
@@ -3864,8 +3864,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0xffffffffffffffffL, x10);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0xffffffffffffffffL, x10);
START()
__ Mov(x0, 0);
@@ -3876,8 +3876,8 @@ TEST(adc_sbc_shift) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
- ASSERT_EQUAL_64(0x8000000000000001L, x10);
+ CHECK_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_64(0x8000000000000001L, x10);
TEARDOWN();
}
@@ -3919,23 +3919,23 @@ TEST(adc_sbc_extend) {
RUN();
- ASSERT_EQUAL_64(0x1df, x10);
- ASSERT_EQUAL_64(0xffffffffffff37bdL, x11);
- ASSERT_EQUAL_64(0xfffffff765432110L, x12);
- ASSERT_EQUAL_64(0x123456789abcdef1L, x13);
+ CHECK_EQUAL_64(0x1df, x10);
+ CHECK_EQUAL_64(0xffffffffffff37bdL, x11);
+ CHECK_EQUAL_64(0xfffffff765432110L, x12);
+ CHECK_EQUAL_64(0x123456789abcdef1L, x13);
- ASSERT_EQUAL_32(0x1df, w14);
- ASSERT_EQUAL_32(0xffff37bd, w15);
- ASSERT_EQUAL_32(0x9abcdef1, w9);
+ CHECK_EQUAL_32(0x1df, w14);
+ CHECK_EQUAL_32(0xffff37bd, w15);
+ CHECK_EQUAL_32(0x9abcdef1, w9);
- ASSERT_EQUAL_64(0x1df + 1, x20);
- ASSERT_EQUAL_64(0xffffffffffff37bdL + 1, x21);
- ASSERT_EQUAL_64(0xfffffff765432110L + 1, x22);
- ASSERT_EQUAL_64(0x123456789abcdef1L + 1, x23);
+ CHECK_EQUAL_64(0x1df + 1, x20);
+ CHECK_EQUAL_64(0xffffffffffff37bdL + 1, x21);
+ CHECK_EQUAL_64(0xfffffff765432110L + 1, x22);
+ CHECK_EQUAL_64(0x123456789abcdef1L + 1, x23);
- ASSERT_EQUAL_32(0x1df + 1, w24);
- ASSERT_EQUAL_32(0xffff37bd + 1, w25);
- ASSERT_EQUAL_32(0x9abcdef1 + 1, w26);
+ CHECK_EQUAL_32(0x1df + 1, w24);
+ CHECK_EQUAL_32(0xffff37bd + 1, w25);
+ CHECK_EQUAL_32(0x9abcdef1 + 1, w26);
// Check that adc correctly sets the condition flags.
START();
@@ -3948,7 +3948,7 @@ TEST(adc_sbc_extend) {
RUN();
- ASSERT_EQUAL_NZCV(CFlag);
+ CHECK_EQUAL_NZCV(CFlag);
START();
__ Mov(x0, 0x7fffffffffffffffL);
@@ -3960,7 +3960,7 @@ TEST(adc_sbc_extend) {
RUN();
- ASSERT_EQUAL_NZCV(NVFlag);
+ CHECK_EQUAL_NZCV(NVFlag);
START();
__ Mov(x0, 0x7fffffffffffffffL);
@@ -3971,7 +3971,7 @@ TEST(adc_sbc_extend) {
RUN();
- ASSERT_EQUAL_NZCV(NVFlag);
+ CHECK_EQUAL_NZCV(NVFlag);
TEARDOWN();
}
@@ -4007,19 +4007,19 @@ TEST(adc_sbc_wide_imm) {
RUN();
- ASSERT_EQUAL_64(0x1234567890abcdefUL, x7);
- ASSERT_EQUAL_64(0xffffffff, x8);
- ASSERT_EQUAL_64(0xedcba9876f543210UL, x9);
- ASSERT_EQUAL_64(0, x10);
- ASSERT_EQUAL_64(0xffffffff, x11);
- ASSERT_EQUAL_64(0xffff, x12);
+ CHECK_EQUAL_64(0x1234567890abcdefUL, x7);
+ CHECK_EQUAL_64(0xffffffff, x8);
+ CHECK_EQUAL_64(0xedcba9876f543210UL, x9);
+ CHECK_EQUAL_64(0, x10);
+ CHECK_EQUAL_64(0xffffffff, x11);
+ CHECK_EQUAL_64(0xffff, x12);
- ASSERT_EQUAL_64(0x1234567890abcdefUL + 1, x18);
- ASSERT_EQUAL_64(0, x19);
- ASSERT_EQUAL_64(0xedcba9876f543211UL, x20);
- ASSERT_EQUAL_64(1, x21);
- ASSERT_EQUAL_64(0x100000000UL, x22);
- ASSERT_EQUAL_64(0x10000, x23);
+ CHECK_EQUAL_64(0x1234567890abcdefUL + 1, x18);
+ CHECK_EQUAL_64(0, x19);
+ CHECK_EQUAL_64(0xedcba9876f543211UL, x20);
+ CHECK_EQUAL_64(1, x21);
+ CHECK_EQUAL_64(0x100000000UL, x22);
+ CHECK_EQUAL_64(0x10000, x23);
TEARDOWN();
}
@@ -4045,11 +4045,11 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_64(0, x10);
- ASSERT_EQUAL_64(-0x1111111111111111L, x11);
- ASSERT_EQUAL_32(-0x11111111, w12);
- ASSERT_EQUAL_64(-1L, x13);
- ASSERT_EQUAL_32(0, w14);
+ CHECK_EQUAL_64(0, x10);
+ CHECK_EQUAL_64(-0x1111111111111111L, x11);
+ CHECK_EQUAL_32(-0x11111111, w12);
+ CHECK_EQUAL_64(-1L, x13);
+ CHECK_EQUAL_32(0, w14);
START();
__ Mov(x0, 0);
@@ -4058,7 +4058,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_NZCV(ZCFlag);
START();
__ Mov(w0, 0);
@@ -4067,7 +4067,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_NZCV(ZCFlag);
START();
__ Mov(x0, 0);
@@ -4077,7 +4077,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_NZCV(NFlag);
START();
__ Mov(w0, 0);
@@ -4087,7 +4087,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_NZCV(NFlag);
START();
__ Mov(x1, 0x1111111111111111L);
@@ -4096,7 +4096,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(CFlag);
+ CHECK_EQUAL_NZCV(CFlag);
START();
__ Mov(w1, 0x11111111);
@@ -4105,7 +4105,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(CFlag);
+ CHECK_EQUAL_NZCV(CFlag);
START();
__ Mov(x0, 1);
@@ -4115,7 +4115,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(NVFlag);
+ CHECK_EQUAL_NZCV(NVFlag);
START();
__ Mov(w0, 1);
@@ -4125,7 +4125,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(NVFlag);
+ CHECK_EQUAL_NZCV(NVFlag);
START();
__ Mov(x0, 1);
@@ -4135,7 +4135,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_NZCV(ZCFlag);
START();
__ Mov(w0, 1);
@@ -4145,7 +4145,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_NZCV(ZCFlag);
START();
__ Mov(w0, 0);
@@ -4157,7 +4157,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(NFlag);
+ CHECK_EQUAL_NZCV(NFlag);
START();
__ Mov(w0, 0);
@@ -4169,7 +4169,7 @@ TEST(flags) {
RUN();
- ASSERT_EQUAL_NZCV(ZCFlag);
+ CHECK_EQUAL_NZCV(ZCFlag);
TEARDOWN();
}
@@ -4218,14 +4218,14 @@ TEST(cmp_shift) {
RUN();
- ASSERT_EQUAL_32(ZCFlag, w0);
- ASSERT_EQUAL_32(ZCFlag, w1);
- ASSERT_EQUAL_32(ZCFlag, w2);
- ASSERT_EQUAL_32(ZCFlag, w3);
- ASSERT_EQUAL_32(ZCFlag, w4);
- ASSERT_EQUAL_32(ZCFlag, w5);
- ASSERT_EQUAL_32(ZCFlag, w6);
- ASSERT_EQUAL_32(ZCFlag, w7);
+ CHECK_EQUAL_32(ZCFlag, w0);
+ CHECK_EQUAL_32(ZCFlag, w1);
+ CHECK_EQUAL_32(ZCFlag, w2);
+ CHECK_EQUAL_32(ZCFlag, w3);
+ CHECK_EQUAL_32(ZCFlag, w4);
+ CHECK_EQUAL_32(ZCFlag, w5);
+ CHECK_EQUAL_32(ZCFlag, w6);
+ CHECK_EQUAL_32(ZCFlag, w7);
TEARDOWN();
}
@@ -4271,14 +4271,14 @@ TEST(cmp_extend) {
RUN();
- ASSERT_EQUAL_32(ZCFlag, w0);
- ASSERT_EQUAL_32(ZCFlag, w1);
- ASSERT_EQUAL_32(ZCFlag, w2);
- ASSERT_EQUAL_32(NCFlag, w3);
- ASSERT_EQUAL_32(NCFlag, w4);
- ASSERT_EQUAL_32(ZCFlag, w5);
- ASSERT_EQUAL_32(NCFlag, w6);
- ASSERT_EQUAL_32(ZCFlag, w7);
+ CHECK_EQUAL_32(ZCFlag, w0);
+ CHECK_EQUAL_32(ZCFlag, w1);
+ CHECK_EQUAL_32(ZCFlag, w2);
+ CHECK_EQUAL_32(NCFlag, w3);
+ CHECK_EQUAL_32(NCFlag, w4);
+ CHECK_EQUAL_32(ZCFlag, w5);
+ CHECK_EQUAL_32(NCFlag, w6);
+ CHECK_EQUAL_32(ZCFlag, w7);
TEARDOWN();
}
@@ -4317,12 +4317,12 @@ TEST(ccmp) {
RUN();
- ASSERT_EQUAL_32(NFlag, w0);
- ASSERT_EQUAL_32(NCFlag, w1);
- ASSERT_EQUAL_32(NoFlag, w2);
- ASSERT_EQUAL_32(NZCVFlag, w3);
- ASSERT_EQUAL_32(ZCFlag, w4);
- ASSERT_EQUAL_32(ZCFlag, w5);
+ CHECK_EQUAL_32(NFlag, w0);
+ CHECK_EQUAL_32(NCFlag, w1);
+ CHECK_EQUAL_32(NoFlag, w2);
+ CHECK_EQUAL_32(NZCVFlag, w3);
+ CHECK_EQUAL_32(ZCFlag, w4);
+ CHECK_EQUAL_32(ZCFlag, w5);
TEARDOWN();
}
@@ -4346,8 +4346,8 @@ TEST(ccmp_wide_imm) {
RUN();
- ASSERT_EQUAL_32(NFlag, w0);
- ASSERT_EQUAL_32(NoFlag, w1);
+ CHECK_EQUAL_32(NFlag, w0);
+ CHECK_EQUAL_32(NoFlag, w1);
TEARDOWN();
}
@@ -4387,11 +4387,11 @@ TEST(ccmp_shift_extend) {
RUN();
- ASSERT_EQUAL_32(ZCFlag, w0);
- ASSERT_EQUAL_32(ZCFlag, w1);
- ASSERT_EQUAL_32(ZCFlag, w2);
- ASSERT_EQUAL_32(NCFlag, w3);
- ASSERT_EQUAL_32(NZCVFlag, w4);
+ CHECK_EQUAL_32(ZCFlag, w0);
+ CHECK_EQUAL_32(ZCFlag, w1);
+ CHECK_EQUAL_32(ZCFlag, w2);
+ CHECK_EQUAL_32(NCFlag, w3);
+ CHECK_EQUAL_32(NZCVFlag, w4);
TEARDOWN();
}
@@ -4441,27 +4441,27 @@ TEST(csel) {
RUN();
- ASSERT_EQUAL_64(0x0000000f, x0);
- ASSERT_EQUAL_64(0x0000001f, x1);
- ASSERT_EQUAL_64(0x00000020, x2);
- ASSERT_EQUAL_64(0x0000000f, x3);
- ASSERT_EQUAL_64(0xffffffe0ffffffe0UL, x4);
- ASSERT_EQUAL_64(0x0000000f0000000fUL, x5);
- ASSERT_EQUAL_64(0xffffffe0ffffffe1UL, x6);
- ASSERT_EQUAL_64(0x0000000f0000000fUL, x7);
- ASSERT_EQUAL_64(0x00000001, x8);
- ASSERT_EQUAL_64(0xffffffff, x9);
- ASSERT_EQUAL_64(0x0000001f00000020UL, x10);
- ASSERT_EQUAL_64(0xfffffff0fffffff0UL, x11);
- ASSERT_EQUAL_64(0xfffffff0fffffff1UL, x12);
- ASSERT_EQUAL_64(0x0000000f, x13);
- ASSERT_EQUAL_64(0x0000000f0000000fUL, x14);
- ASSERT_EQUAL_64(0x0000000f, x15);
- ASSERT_EQUAL_64(0x0000000f0000000fUL, x18);
- ASSERT_EQUAL_64(0, x24);
- ASSERT_EQUAL_64(0x0000001f0000001fUL, x25);
- ASSERT_EQUAL_64(0x0000001f0000001fUL, x26);
- ASSERT_EQUAL_64(0, x27);
+ CHECK_EQUAL_64(0x0000000f, x0);
+ CHECK_EQUAL_64(0x0000001f, x1);
+ CHECK_EQUAL_64(0x00000020, x2);
+ CHECK_EQUAL_64(0x0000000f, x3);
+ CHECK_EQUAL_64(0xffffffe0ffffffe0UL, x4);
+ CHECK_EQUAL_64(0x0000000f0000000fUL, x5);
+ CHECK_EQUAL_64(0xffffffe0ffffffe1UL, x6);
+ CHECK_EQUAL_64(0x0000000f0000000fUL, x7);
+ CHECK_EQUAL_64(0x00000001, x8);
+ CHECK_EQUAL_64(0xffffffff, x9);
+ CHECK_EQUAL_64(0x0000001f00000020UL, x10);
+ CHECK_EQUAL_64(0xfffffff0fffffff0UL, x11);
+ CHECK_EQUAL_64(0xfffffff0fffffff1UL, x12);
+ CHECK_EQUAL_64(0x0000000f, x13);
+ CHECK_EQUAL_64(0x0000000f0000000fUL, x14);
+ CHECK_EQUAL_64(0x0000000f, x15);
+ CHECK_EQUAL_64(0x0000000f0000000fUL, x18);
+ CHECK_EQUAL_64(0, x24);
+ CHECK_EQUAL_64(0x0000001f0000001fUL, x25);
+ CHECK_EQUAL_64(0x0000001f0000001fUL, x26);
+ CHECK_EQUAL_64(0, x27);
TEARDOWN();
}
@@ -4499,23 +4499,23 @@ TEST(csel_imm) {
RUN();
- ASSERT_EQUAL_32(-2, w0);
- ASSERT_EQUAL_32(-1, w1);
- ASSERT_EQUAL_32(0, w2);
- ASSERT_EQUAL_32(1, w3);
- ASSERT_EQUAL_32(2, w4);
- ASSERT_EQUAL_32(-1, w5);
- ASSERT_EQUAL_32(0x40000000, w6);
- ASSERT_EQUAL_32(0x80000000, w7);
+ CHECK_EQUAL_32(-2, w0);
+ CHECK_EQUAL_32(-1, w1);
+ CHECK_EQUAL_32(0, w2);
+ CHECK_EQUAL_32(1, w3);
+ CHECK_EQUAL_32(2, w4);
+ CHECK_EQUAL_32(-1, w5);
+ CHECK_EQUAL_32(0x40000000, w6);
+ CHECK_EQUAL_32(0x80000000, w7);
- ASSERT_EQUAL_64(-2, x8);
- ASSERT_EQUAL_64(-1, x9);
- ASSERT_EQUAL_64(0, x10);
- ASSERT_EQUAL_64(1, x11);
- ASSERT_EQUAL_64(2, x12);
- ASSERT_EQUAL_64(-1, x13);
- ASSERT_EQUAL_64(0x4000000000000000UL, x14);
- ASSERT_EQUAL_64(0x8000000000000000UL, x15);
+ CHECK_EQUAL_64(-2, x8);
+ CHECK_EQUAL_64(-1, x9);
+ CHECK_EQUAL_64(0, x10);
+ CHECK_EQUAL_64(1, x11);
+ CHECK_EQUAL_64(2, x12);
+ CHECK_EQUAL_64(-1, x13);
+ CHECK_EQUAL_64(0x4000000000000000UL, x14);
+ CHECK_EQUAL_64(0x8000000000000000UL, x15);
TEARDOWN();
}
@@ -4556,19 +4556,19 @@ TEST(lslv) {
RUN();
- ASSERT_EQUAL_64(value, x0);
- ASSERT_EQUAL_64(value << (shift[0] & 63), x16);
- ASSERT_EQUAL_64(value << (shift[1] & 63), x17);
- ASSERT_EQUAL_64(value << (shift[2] & 63), x18);
- ASSERT_EQUAL_64(value << (shift[3] & 63), x19);
- ASSERT_EQUAL_64(value << (shift[4] & 63), x20);
- ASSERT_EQUAL_64(value << (shift[5] & 63), x21);
- ASSERT_EQUAL_32(value << (shift[0] & 31), w22);
- ASSERT_EQUAL_32(value << (shift[1] & 31), w23);
- ASSERT_EQUAL_32(value << (shift[2] & 31), w24);
- ASSERT_EQUAL_32(value << (shift[3] & 31), w25);
- ASSERT_EQUAL_32(value << (shift[4] & 31), w26);
- ASSERT_EQUAL_32(value << (shift[5] & 31), w27);
+ CHECK_EQUAL_64(value, x0);
+ CHECK_EQUAL_64(value << (shift[0] & 63), x16);
+ CHECK_EQUAL_64(value << (shift[1] & 63), x17);
+ CHECK_EQUAL_64(value << (shift[2] & 63), x18);
+ CHECK_EQUAL_64(value << (shift[3] & 63), x19);
+ CHECK_EQUAL_64(value << (shift[4] & 63), x20);
+ CHECK_EQUAL_64(value << (shift[5] & 63), x21);
+ CHECK_EQUAL_32(value << (shift[0] & 31), w22);
+ CHECK_EQUAL_32(value << (shift[1] & 31), w23);
+ CHECK_EQUAL_32(value << (shift[2] & 31), w24);
+ CHECK_EQUAL_32(value << (shift[3] & 31), w25);
+ CHECK_EQUAL_32(value << (shift[4] & 31), w26);
+ CHECK_EQUAL_32(value << (shift[5] & 31), w27);
TEARDOWN();
}
@@ -4609,21 +4609,21 @@ TEST(lsrv) {
RUN();
- ASSERT_EQUAL_64(value, x0);
- ASSERT_EQUAL_64(value >> (shift[0] & 63), x16);
- ASSERT_EQUAL_64(value >> (shift[1] & 63), x17);
- ASSERT_EQUAL_64(value >> (shift[2] & 63), x18);
- ASSERT_EQUAL_64(value >> (shift[3] & 63), x19);
- ASSERT_EQUAL_64(value >> (shift[4] & 63), x20);
- ASSERT_EQUAL_64(value >> (shift[5] & 63), x21);
+ CHECK_EQUAL_64(value, x0);
+ CHECK_EQUAL_64(value >> (shift[0] & 63), x16);
+ CHECK_EQUAL_64(value >> (shift[1] & 63), x17);
+ CHECK_EQUAL_64(value >> (shift[2] & 63), x18);
+ CHECK_EQUAL_64(value >> (shift[3] & 63), x19);
+ CHECK_EQUAL_64(value >> (shift[4] & 63), x20);
+ CHECK_EQUAL_64(value >> (shift[5] & 63), x21);
value &= 0xffffffffUL;
- ASSERT_EQUAL_32(value >> (shift[0] & 31), w22);
- ASSERT_EQUAL_32(value >> (shift[1] & 31), w23);
- ASSERT_EQUAL_32(value >> (shift[2] & 31), w24);
- ASSERT_EQUAL_32(value >> (shift[3] & 31), w25);
- ASSERT_EQUAL_32(value >> (shift[4] & 31), w26);
- ASSERT_EQUAL_32(value >> (shift[5] & 31), w27);
+ CHECK_EQUAL_32(value >> (shift[0] & 31), w22);
+ CHECK_EQUAL_32(value >> (shift[1] & 31), w23);
+ CHECK_EQUAL_32(value >> (shift[2] & 31), w24);
+ CHECK_EQUAL_32(value >> (shift[3] & 31), w25);
+ CHECK_EQUAL_32(value >> (shift[4] & 31), w26);
+ CHECK_EQUAL_32(value >> (shift[5] & 31), w27);
TEARDOWN();
}
@@ -4664,21 +4664,21 @@ TEST(asrv) {
RUN();
- ASSERT_EQUAL_64(value, x0);
- ASSERT_EQUAL_64(value >> (shift[0] & 63), x16);
- ASSERT_EQUAL_64(value >> (shift[1] & 63), x17);
- ASSERT_EQUAL_64(value >> (shift[2] & 63), x18);
- ASSERT_EQUAL_64(value >> (shift[3] & 63), x19);
- ASSERT_EQUAL_64(value >> (shift[4] & 63), x20);
- ASSERT_EQUAL_64(value >> (shift[5] & 63), x21);
+ CHECK_EQUAL_64(value, x0);
+ CHECK_EQUAL_64(value >> (shift[0] & 63), x16);
+ CHECK_EQUAL_64(value >> (shift[1] & 63), x17);
+ CHECK_EQUAL_64(value >> (shift[2] & 63), x18);
+ CHECK_EQUAL_64(value >> (shift[3] & 63), x19);
+ CHECK_EQUAL_64(value >> (shift[4] & 63), x20);
+ CHECK_EQUAL_64(value >> (shift[5] & 63), x21);
int32_t value32 = static_cast<int32_t>(value & 0xffffffffUL);
- ASSERT_EQUAL_32(value32 >> (shift[0] & 31), w22);
- ASSERT_EQUAL_32(value32 >> (shift[1] & 31), w23);
- ASSERT_EQUAL_32(value32 >> (shift[2] & 31), w24);
- ASSERT_EQUAL_32(value32 >> (shift[3] & 31), w25);
- ASSERT_EQUAL_32(value32 >> (shift[4] & 31), w26);
- ASSERT_EQUAL_32(value32 >> (shift[5] & 31), w27);
+ CHECK_EQUAL_32(value32 >> (shift[0] & 31), w22);
+ CHECK_EQUAL_32(value32 >> (shift[1] & 31), w23);
+ CHECK_EQUAL_32(value32 >> (shift[2] & 31), w24);
+ CHECK_EQUAL_32(value32 >> (shift[3] & 31), w25);
+ CHECK_EQUAL_32(value32 >> (shift[4] & 31), w26);
+ CHECK_EQUAL_32(value32 >> (shift[5] & 31), w27);
TEARDOWN();
}
@@ -4719,19 +4719,19 @@ TEST(rorv) {
RUN();
- ASSERT_EQUAL_64(value, x0);
- ASSERT_EQUAL_64(0xf0123456789abcdeUL, x16);
- ASSERT_EQUAL_64(0xef0123456789abcdUL, x17);
- ASSERT_EQUAL_64(0xdef0123456789abcUL, x18);
- ASSERT_EQUAL_64(0xcdef0123456789abUL, x19);
- ASSERT_EQUAL_64(0xabcdef0123456789UL, x20);
- ASSERT_EQUAL_64(0x789abcdef0123456UL, x21);
- ASSERT_EQUAL_32(0xf89abcde, w22);
- ASSERT_EQUAL_32(0xef89abcd, w23);
- ASSERT_EQUAL_32(0xdef89abc, w24);
- ASSERT_EQUAL_32(0xcdef89ab, w25);
- ASSERT_EQUAL_32(0xabcdef89, w26);
- ASSERT_EQUAL_32(0xf89abcde, w27);
+ CHECK_EQUAL_64(value, x0);
+ CHECK_EQUAL_64(0xf0123456789abcdeUL, x16);
+ CHECK_EQUAL_64(0xef0123456789abcdUL, x17);
+ CHECK_EQUAL_64(0xdef0123456789abcUL, x18);
+ CHECK_EQUAL_64(0xcdef0123456789abUL, x19);
+ CHECK_EQUAL_64(0xabcdef0123456789UL, x20);
+ CHECK_EQUAL_64(0x789abcdef0123456UL, x21);
+ CHECK_EQUAL_32(0xf89abcde, w22);
+ CHECK_EQUAL_32(0xef89abcd, w23);
+ CHECK_EQUAL_32(0xdef89abc, w24);
+ CHECK_EQUAL_32(0xcdef89ab, w25);
+ CHECK_EQUAL_32(0xabcdef89, w26);
+ CHECK_EQUAL_32(0xf89abcde, w27);
TEARDOWN();
}
@@ -4765,14 +4765,14 @@ TEST(bfm) {
RUN();
- ASSERT_EQUAL_64(0x88888888888889abL, x10);
- ASSERT_EQUAL_64(0x8888cdef88888888L, x11);
+ CHECK_EQUAL_64(0x88888888888889abL, x10);
+ CHECK_EQUAL_64(0x8888cdef88888888L, x11);
- ASSERT_EQUAL_32(0x888888ab, w20);
- ASSERT_EQUAL_32(0x88cdef88, w21);
+ CHECK_EQUAL_32(0x888888ab, w20);
+ CHECK_EQUAL_32(0x88cdef88, w21);
- ASSERT_EQUAL_64(0x8888888888ef8888L, x12);
- ASSERT_EQUAL_64(0x88888888888888abL, x13);
+ CHECK_EQUAL_64(0x8888888888ef8888L, x12);
+ CHECK_EQUAL_64(0x88888888888888abL, x13);
TEARDOWN();
}
@@ -4814,28 +4814,28 @@ TEST(sbfm) {
RUN();
- ASSERT_EQUAL_64(0xffffffffffff89abL, x10);
- ASSERT_EQUAL_64(0xffffcdef00000000L, x11);
- ASSERT_EQUAL_64(0x4567L, x12);
- ASSERT_EQUAL_64(0x789abcdef0000L, x13);
+ CHECK_EQUAL_64(0xffffffffffff89abL, x10);
+ CHECK_EQUAL_64(0xffffcdef00000000L, x11);
+ CHECK_EQUAL_64(0x4567L, x12);
+ CHECK_EQUAL_64(0x789abcdef0000L, x13);
- ASSERT_EQUAL_32(0xffffffab, w14);
- ASSERT_EQUAL_32(0xffcdef00, w15);
- ASSERT_EQUAL_32(0x54, w16);
- ASSERT_EQUAL_32(0x00321000, w17);
+ CHECK_EQUAL_32(0xffffffab, w14);
+ CHECK_EQUAL_32(0xffcdef00, w15);
+ CHECK_EQUAL_32(0x54, w16);
+ CHECK_EQUAL_32(0x00321000, w17);
- ASSERT_EQUAL_64(0x01234567L, x18);
- ASSERT_EQUAL_64(0xfffffffffedcba98L, x19);
- ASSERT_EQUAL_64(0xffffffffffcdef00L, x20);
- ASSERT_EQUAL_64(0x321000L, x21);
- ASSERT_EQUAL_64(0xffffffffffffabcdL, x22);
- ASSERT_EQUAL_64(0x5432L, x23);
- ASSERT_EQUAL_64(0xffffffffffffffefL, x24);
- ASSERT_EQUAL_64(0x10, x25);
- ASSERT_EQUAL_64(0xffffffffffffcdefL, x26);
- ASSERT_EQUAL_64(0x3210, x27);
- ASSERT_EQUAL_64(0xffffffff89abcdefL, x28);
- ASSERT_EQUAL_64(0x76543210, x29);
+ CHECK_EQUAL_64(0x01234567L, x18);
+ CHECK_EQUAL_64(0xfffffffffedcba98L, x19);
+ CHECK_EQUAL_64(0xffffffffffcdef00L, x20);
+ CHECK_EQUAL_64(0x321000L, x21);
+ CHECK_EQUAL_64(0xffffffffffffabcdL, x22);
+ CHECK_EQUAL_64(0x5432L, x23);
+ CHECK_EQUAL_64(0xffffffffffffffefL, x24);
+ CHECK_EQUAL_64(0x10, x25);
+ CHECK_EQUAL_64(0xffffffffffffcdefL, x26);
+ CHECK_EQUAL_64(0x3210, x27);
+ CHECK_EQUAL_64(0xffffffff89abcdefL, x28);
+ CHECK_EQUAL_64(0x76543210, x29);
TEARDOWN();
}
@@ -4875,24 +4875,24 @@ TEST(ubfm) {
RUN();
- ASSERT_EQUAL_64(0x00000000000089abL, x10);
- ASSERT_EQUAL_64(0x0000cdef00000000L, x11);
- ASSERT_EQUAL_64(0x4567L, x12);
- ASSERT_EQUAL_64(0x789abcdef0000L, x13);
+ CHECK_EQUAL_64(0x00000000000089abL, x10);
+ CHECK_EQUAL_64(0x0000cdef00000000L, x11);
+ CHECK_EQUAL_64(0x4567L, x12);
+ CHECK_EQUAL_64(0x789abcdef0000L, x13);
- ASSERT_EQUAL_32(0x000000ab, w25);
- ASSERT_EQUAL_32(0x00cdef00, w26);
- ASSERT_EQUAL_32(0x54, w27);
- ASSERT_EQUAL_32(0x00321000, w28);
+ CHECK_EQUAL_32(0x000000ab, w25);
+ CHECK_EQUAL_32(0x00cdef00, w26);
+ CHECK_EQUAL_32(0x54, w27);
+ CHECK_EQUAL_32(0x00321000, w28);
- ASSERT_EQUAL_64(0x8000000000000000L, x15);
- ASSERT_EQUAL_64(0x0123456789abcdefL, x16);
- ASSERT_EQUAL_64(0x01234567L, x17);
- ASSERT_EQUAL_64(0xcdef00L, x18);
- ASSERT_EQUAL_64(0xabcdL, x19);
- ASSERT_EQUAL_64(0xefL, x20);
- ASSERT_EQUAL_64(0xcdefL, x21);
- ASSERT_EQUAL_64(0x89abcdefL, x22);
+ CHECK_EQUAL_64(0x8000000000000000L, x15);
+ CHECK_EQUAL_64(0x0123456789abcdefL, x16);
+ CHECK_EQUAL_64(0x01234567L, x17);
+ CHECK_EQUAL_64(0xcdef00L, x18);
+ CHECK_EQUAL_64(0xabcdL, x19);
+ CHECK_EQUAL_64(0xefL, x20);
+ CHECK_EQUAL_64(0xcdefL, x21);
+ CHECK_EQUAL_64(0x89abcdefL, x22);
TEARDOWN();
}
@@ -4921,16 +4921,16 @@ TEST(extr) {
RUN();
- ASSERT_EQUAL_64(0x76543210, x10);
- ASSERT_EQUAL_64(0xfedcba9876543210L, x11);
- ASSERT_EQUAL_64(0xbb2a1908, x12);
- ASSERT_EQUAL_64(0x0048d159e26af37bUL, x13);
- ASSERT_EQUAL_64(0x89abcdef, x20);
- ASSERT_EQUAL_64(0x0123456789abcdefL, x21);
- ASSERT_EQUAL_64(0x19083b2a, x22);
- ASSERT_EQUAL_64(0x13579bdf, x23);
- ASSERT_EQUAL_64(0x7f6e5d4c3b2a1908UL, x24);
- ASSERT_EQUAL_64(0x02468acf13579bdeUL, x25);
+ CHECK_EQUAL_64(0x76543210, x10);
+ CHECK_EQUAL_64(0xfedcba9876543210L, x11);
+ CHECK_EQUAL_64(0xbb2a1908, x12);
+ CHECK_EQUAL_64(0x0048d159e26af37bUL, x13);
+ CHECK_EQUAL_64(0x89abcdef, x20);
+ CHECK_EQUAL_64(0x0123456789abcdefL, x21);
+ CHECK_EQUAL_64(0x19083b2a, x22);
+ CHECK_EQUAL_64(0x13579bdf, x23);
+ CHECK_EQUAL_64(0x7f6e5d4c3b2a1908UL, x24);
+ CHECK_EQUAL_64(0x02468acf13579bdeUL, x25);
TEARDOWN();
}
@@ -4953,14 +4953,14 @@ TEST(fmov_imm) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s11);
- ASSERT_EQUAL_FP64(-13.0, d22);
- ASSERT_EQUAL_FP32(255.0, s1);
- ASSERT_EQUAL_FP64(12.34567, d2);
- ASSERT_EQUAL_FP32(0.0, s3);
- ASSERT_EQUAL_FP64(0.0, d4);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s5);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d6);
+ CHECK_EQUAL_FP32(1.0, s11);
+ CHECK_EQUAL_FP64(-13.0, d22);
+ CHECK_EQUAL_FP32(255.0, s1);
+ CHECK_EQUAL_FP64(12.34567, d2);
+ CHECK_EQUAL_FP32(0.0, s3);
+ CHECK_EQUAL_FP64(0.0, d4);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s5);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d6);
TEARDOWN();
}
@@ -4985,13 +4985,13 @@ TEST(fmov_reg) {
RUN();
- ASSERT_EQUAL_32(float_to_rawbits(1.0), w10);
- ASSERT_EQUAL_FP32(1.0, s30);
- ASSERT_EQUAL_FP32(1.0, s5);
- ASSERT_EQUAL_64(double_to_rawbits(-13.0), x1);
- ASSERT_EQUAL_FP64(-13.0, d2);
- ASSERT_EQUAL_FP64(-13.0, d4);
- ASSERT_EQUAL_FP32(rawbits_to_float(0x89abcdef), s6);
+ CHECK_EQUAL_32(float_to_rawbits(1.0), w10);
+ CHECK_EQUAL_FP32(1.0, s30);
+ CHECK_EQUAL_FP32(1.0, s5);
+ CHECK_EQUAL_64(double_to_rawbits(-13.0), x1);
+ CHECK_EQUAL_FP64(-13.0, d2);
+ CHECK_EQUAL_FP64(-13.0, d4);
+ CHECK_EQUAL_FP32(rawbits_to_float(0x89abcdef), s6);
TEARDOWN();
}
@@ -5035,20 +5035,20 @@ TEST(fadd) {
RUN();
- ASSERT_EQUAL_FP32(4.25, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(1.0, s2);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s3);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s4);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s5);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s6);
- ASSERT_EQUAL_FP64(0.25, d7);
- ASSERT_EQUAL_FP64(2.25, d8);
- ASSERT_EQUAL_FP64(2.25, d9);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d10);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d11);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d12);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d13);
+ CHECK_EQUAL_FP32(4.25, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(1.0, s2);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s3);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s4);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s5);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s6);
+ CHECK_EQUAL_FP64(0.25, d7);
+ CHECK_EQUAL_FP64(2.25, d8);
+ CHECK_EQUAL_FP64(2.25, d9);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d10);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d11);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d12);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d13);
TEARDOWN();
}
@@ -5092,20 +5092,20 @@ TEST(fsub) {
RUN();
- ASSERT_EQUAL_FP32(2.25, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(-1.0, s2);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s3);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s4);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s5);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s6);
- ASSERT_EQUAL_FP64(-4.25, d7);
- ASSERT_EQUAL_FP64(-2.25, d8);
- ASSERT_EQUAL_FP64(-2.25, d9);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d10);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d11);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d12);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d13);
+ CHECK_EQUAL_FP32(2.25, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(-1.0, s2);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s3);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s4);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s5);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s6);
+ CHECK_EQUAL_FP64(-4.25, d7);
+ CHECK_EQUAL_FP64(-2.25, d8);
+ CHECK_EQUAL_FP64(-2.25, d9);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d10);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d11);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d12);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d13);
TEARDOWN();
}
@@ -5150,20 +5150,20 @@ TEST(fmul) {
RUN();
- ASSERT_EQUAL_FP32(6.5, s0);
- ASSERT_EQUAL_FP32(0.0, s1);
- ASSERT_EQUAL_FP32(0.0, s2);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s3);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s4);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s5);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s6);
- ASSERT_EQUAL_FP64(-4.5, d7);
- ASSERT_EQUAL_FP64(0.0, d8);
- ASSERT_EQUAL_FP64(0.0, d9);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d10);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d11);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d12);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d13);
+ CHECK_EQUAL_FP32(6.5, s0);
+ CHECK_EQUAL_FP32(0.0, s1);
+ CHECK_EQUAL_FP32(0.0, s2);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s3);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s4);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s5);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s6);
+ CHECK_EQUAL_FP64(-4.5, d7);
+ CHECK_EQUAL_FP64(0.0, d8);
+ CHECK_EQUAL_FP64(0.0, d9);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d10);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d11);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d12);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d13);
TEARDOWN();
}
@@ -5186,10 +5186,10 @@ static void FmaddFmsubHelper(double n, double m, double a,
END();
RUN();
- ASSERT_EQUAL_FP64(fmadd, d28);
- ASSERT_EQUAL_FP64(fmsub, d29);
- ASSERT_EQUAL_FP64(fnmadd, d30);
- ASSERT_EQUAL_FP64(fnmsub, d31);
+ CHECK_EQUAL_FP64(fmadd, d28);
+ CHECK_EQUAL_FP64(fmsub, d29);
+ CHECK_EQUAL_FP64(fnmadd, d30);
+ CHECK_EQUAL_FP64(fnmsub, d31);
TEARDOWN();
}
@@ -5254,10 +5254,10 @@ static void FmaddFmsubHelper(float n, float m, float a,
END();
RUN();
- ASSERT_EQUAL_FP32(fmadd, s28);
- ASSERT_EQUAL_FP32(fmsub, s29);
- ASSERT_EQUAL_FP32(fnmadd, s30);
- ASSERT_EQUAL_FP32(fnmsub, s31);
+ CHECK_EQUAL_FP32(fmadd, s28);
+ CHECK_EQUAL_FP32(fmsub, s29);
+ CHECK_EQUAL_FP32(fnmadd, s30);
+ CHECK_EQUAL_FP32(fnmsub, s31);
TEARDOWN();
}
@@ -5313,12 +5313,12 @@ TEST(fmadd_fmsub_double_nans) {
double q1 = rawbits_to_double(0x7ffaaaaa11111111);
double q2 = rawbits_to_double(0x7ffaaaaa22222222);
double qa = rawbits_to_double(0x7ffaaaaaaaaaaaaa);
- ASSERT(IsSignallingNaN(s1));
- ASSERT(IsSignallingNaN(s2));
- ASSERT(IsSignallingNaN(sa));
- ASSERT(IsQuietNaN(q1));
- ASSERT(IsQuietNaN(q2));
- ASSERT(IsQuietNaN(qa));
+ DCHECK(IsSignallingNaN(s1));
+ DCHECK(IsSignallingNaN(s2));
+ DCHECK(IsSignallingNaN(sa));
+ DCHECK(IsQuietNaN(q1));
+ DCHECK(IsQuietNaN(q2));
+ DCHECK(IsQuietNaN(qa));
// The input NaNs after passing through ProcessNaN.
double s1_proc = rawbits_to_double(0x7ffd555511111111);
@@ -5327,22 +5327,22 @@ TEST(fmadd_fmsub_double_nans) {
double q1_proc = q1;
double q2_proc = q2;
double qa_proc = qa;
- ASSERT(IsQuietNaN(s1_proc));
- ASSERT(IsQuietNaN(s2_proc));
- ASSERT(IsQuietNaN(sa_proc));
- ASSERT(IsQuietNaN(q1_proc));
- ASSERT(IsQuietNaN(q2_proc));
- ASSERT(IsQuietNaN(qa_proc));
+ DCHECK(IsQuietNaN(s1_proc));
+ DCHECK(IsQuietNaN(s2_proc));
+ DCHECK(IsQuietNaN(sa_proc));
+ DCHECK(IsQuietNaN(q1_proc));
+ DCHECK(IsQuietNaN(q2_proc));
+ DCHECK(IsQuietNaN(qa_proc));
// Negated NaNs as it would be done on ARMv8 hardware.
double s1_proc_neg = rawbits_to_double(0xfffd555511111111);
double sa_proc_neg = rawbits_to_double(0xfffd5555aaaaaaaa);
double q1_proc_neg = rawbits_to_double(0xfffaaaaa11111111);
double qa_proc_neg = rawbits_to_double(0xfffaaaaaaaaaaaaa);
- ASSERT(IsQuietNaN(s1_proc_neg));
- ASSERT(IsQuietNaN(sa_proc_neg));
- ASSERT(IsQuietNaN(q1_proc_neg));
- ASSERT(IsQuietNaN(qa_proc_neg));
+ DCHECK(IsQuietNaN(s1_proc_neg));
+ DCHECK(IsQuietNaN(sa_proc_neg));
+ DCHECK(IsQuietNaN(q1_proc_neg));
+ DCHECK(IsQuietNaN(qa_proc_neg));
// Quiet NaNs are propagated.
FmaddFmsubHelper(q1, 0, 0, q1_proc, q1_proc_neg, q1_proc_neg, q1_proc);
@@ -5396,12 +5396,12 @@ TEST(fmadd_fmsub_float_nans) {
float q1 = rawbits_to_float(0x7fea1111);
float q2 = rawbits_to_float(0x7fea2222);
float qa = rawbits_to_float(0x7feaaaaa);
- ASSERT(IsSignallingNaN(s1));
- ASSERT(IsSignallingNaN(s2));
- ASSERT(IsSignallingNaN(sa));
- ASSERT(IsQuietNaN(q1));
- ASSERT(IsQuietNaN(q2));
- ASSERT(IsQuietNaN(qa));
+ DCHECK(IsSignallingNaN(s1));
+ DCHECK(IsSignallingNaN(s2));
+ DCHECK(IsSignallingNaN(sa));
+ DCHECK(IsQuietNaN(q1));
+ DCHECK(IsQuietNaN(q2));
+ DCHECK(IsQuietNaN(qa));
// The input NaNs after passing through ProcessNaN.
float s1_proc = rawbits_to_float(0x7fd51111);
@@ -5410,22 +5410,22 @@ TEST(fmadd_fmsub_float_nans) {
float q1_proc = q1;
float q2_proc = q2;
float qa_proc = qa;
- ASSERT(IsQuietNaN(s1_proc));
- ASSERT(IsQuietNaN(s2_proc));
- ASSERT(IsQuietNaN(sa_proc));
- ASSERT(IsQuietNaN(q1_proc));
- ASSERT(IsQuietNaN(q2_proc));
- ASSERT(IsQuietNaN(qa_proc));
+ DCHECK(IsQuietNaN(s1_proc));
+ DCHECK(IsQuietNaN(s2_proc));
+ DCHECK(IsQuietNaN(sa_proc));
+ DCHECK(IsQuietNaN(q1_proc));
+ DCHECK(IsQuietNaN(q2_proc));
+ DCHECK(IsQuietNaN(qa_proc));
// Negated NaNs as it would be done on ARMv8 hardware.
float s1_proc_neg = rawbits_to_float(0xffd51111);
float sa_proc_neg = rawbits_to_float(0xffd5aaaa);
float q1_proc_neg = rawbits_to_float(0xffea1111);
float qa_proc_neg = rawbits_to_float(0xffeaaaaa);
- ASSERT(IsQuietNaN(s1_proc_neg));
- ASSERT(IsQuietNaN(sa_proc_neg));
- ASSERT(IsQuietNaN(q1_proc_neg));
- ASSERT(IsQuietNaN(qa_proc_neg));
+ DCHECK(IsQuietNaN(s1_proc_neg));
+ DCHECK(IsQuietNaN(sa_proc_neg));
+ DCHECK(IsQuietNaN(q1_proc_neg));
+ DCHECK(IsQuietNaN(qa_proc_neg));
// Quiet NaNs are propagated.
FmaddFmsubHelper(q1, 0, 0, q1_proc, q1_proc_neg, q1_proc_neg, q1_proc);
@@ -5509,20 +5509,20 @@ TEST(fdiv) {
RUN();
- ASSERT_EQUAL_FP32(1.625f, s0);
- ASSERT_EQUAL_FP32(1.0f, s1);
- ASSERT_EQUAL_FP32(-0.0f, s2);
- ASSERT_EQUAL_FP32(0.0f, s3);
- ASSERT_EQUAL_FP32(-0.0f, s4);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s5);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s6);
- ASSERT_EQUAL_FP64(-1.125, d7);
- ASSERT_EQUAL_FP64(0.0, d8);
- ASSERT_EQUAL_FP64(-0.0, d9);
- ASSERT_EQUAL_FP64(0.0, d10);
- ASSERT_EQUAL_FP64(-0.0, d11);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d12);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d13);
+ CHECK_EQUAL_FP32(1.625f, s0);
+ CHECK_EQUAL_FP32(1.0f, s1);
+ CHECK_EQUAL_FP32(-0.0f, s2);
+ CHECK_EQUAL_FP32(0.0f, s3);
+ CHECK_EQUAL_FP32(-0.0f, s4);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s5);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s6);
+ CHECK_EQUAL_FP64(-1.125, d7);
+ CHECK_EQUAL_FP64(0.0, d8);
+ CHECK_EQUAL_FP64(-0.0, d9);
+ CHECK_EQUAL_FP64(0.0, d10);
+ CHECK_EQUAL_FP64(-0.0, d11);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d12);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d13);
TEARDOWN();
}
@@ -5625,10 +5625,10 @@ static void FminFmaxDoubleHelper(double n, double m, double min, double max,
RUN();
- ASSERT_EQUAL_FP64(min, d28);
- ASSERT_EQUAL_FP64(max, d29);
- ASSERT_EQUAL_FP64(minnm, d30);
- ASSERT_EQUAL_FP64(maxnm, d31);
+ CHECK_EQUAL_FP64(min, d28);
+ CHECK_EQUAL_FP64(max, d29);
+ CHECK_EQUAL_FP64(minnm, d30);
+ CHECK_EQUAL_FP64(maxnm, d31);
TEARDOWN();
}
@@ -5643,10 +5643,10 @@ TEST(fmax_fmin_d) {
double snan_processed = rawbits_to_double(0x7ffd555512345678);
double qnan_processed = qnan;
- ASSERT(IsSignallingNaN(snan));
- ASSERT(IsQuietNaN(qnan));
- ASSERT(IsQuietNaN(snan_processed));
- ASSERT(IsQuietNaN(qnan_processed));
+ DCHECK(IsSignallingNaN(snan));
+ DCHECK(IsQuietNaN(qnan));
+ DCHECK(IsQuietNaN(snan_processed));
+ DCHECK(IsQuietNaN(qnan_processed));
// Bootstrap tests.
FminFmaxDoubleHelper(0, 0, 0, 0, 0, 0);
@@ -5710,10 +5710,10 @@ static void FminFmaxFloatHelper(float n, float m, float min, float max,
RUN();
- ASSERT_EQUAL_FP32(min, s28);
- ASSERT_EQUAL_FP32(max, s29);
- ASSERT_EQUAL_FP32(minnm, s30);
- ASSERT_EQUAL_FP32(maxnm, s31);
+ CHECK_EQUAL_FP32(min, s28);
+ CHECK_EQUAL_FP32(max, s29);
+ CHECK_EQUAL_FP32(minnm, s30);
+ CHECK_EQUAL_FP32(maxnm, s31);
TEARDOWN();
}
@@ -5728,10 +5728,10 @@ TEST(fmax_fmin_s) {
float snan_processed = rawbits_to_float(0x7fd51234);
float qnan_processed = qnan;
- ASSERT(IsSignallingNaN(snan));
- ASSERT(IsQuietNaN(qnan));
- ASSERT(IsQuietNaN(snan_processed));
- ASSERT(IsQuietNaN(qnan_processed));
+ DCHECK(IsSignallingNaN(snan));
+ DCHECK(IsQuietNaN(qnan));
+ DCHECK(IsQuietNaN(snan_processed));
+ DCHECK(IsQuietNaN(qnan_processed));
// Bootstrap tests.
FminFmaxFloatHelper(0, 0, 0, 0, 0, 0);
@@ -5833,16 +5833,16 @@ TEST(fccmp) {
RUN();
- ASSERT_EQUAL_32(ZCFlag, w0);
- ASSERT_EQUAL_32(VFlag, w1);
- ASSERT_EQUAL_32(NFlag, w2);
- ASSERT_EQUAL_32(CVFlag, w3);
- ASSERT_EQUAL_32(ZCFlag, w4);
- ASSERT_EQUAL_32(ZVFlag, w5);
- ASSERT_EQUAL_32(CFlag, w6);
- ASSERT_EQUAL_32(NFlag, w7);
- ASSERT_EQUAL_32(ZCFlag, w8);
- ASSERT_EQUAL_32(ZCFlag, w9);
+ CHECK_EQUAL_32(ZCFlag, w0);
+ CHECK_EQUAL_32(VFlag, w1);
+ CHECK_EQUAL_32(NFlag, w2);
+ CHECK_EQUAL_32(CVFlag, w3);
+ CHECK_EQUAL_32(ZCFlag, w4);
+ CHECK_EQUAL_32(ZVFlag, w5);
+ CHECK_EQUAL_32(CFlag, w6);
+ CHECK_EQUAL_32(NFlag, w7);
+ CHECK_EQUAL_32(ZCFlag, w8);
+ CHECK_EQUAL_32(ZCFlag, w9);
TEARDOWN();
}
@@ -5912,20 +5912,20 @@ TEST(fcmp) {
RUN();
- ASSERT_EQUAL_32(ZCFlag, w0);
- ASSERT_EQUAL_32(NFlag, w1);
- ASSERT_EQUAL_32(CFlag, w2);
- ASSERT_EQUAL_32(CVFlag, w3);
- ASSERT_EQUAL_32(CVFlag, w4);
- ASSERT_EQUAL_32(ZCFlag, w5);
- ASSERT_EQUAL_32(NFlag, w6);
- ASSERT_EQUAL_32(ZCFlag, w10);
- ASSERT_EQUAL_32(NFlag, w11);
- ASSERT_EQUAL_32(CFlag, w12);
- ASSERT_EQUAL_32(CVFlag, w13);
- ASSERT_EQUAL_32(CVFlag, w14);
- ASSERT_EQUAL_32(ZCFlag, w15);
- ASSERT_EQUAL_32(NFlag, w16);
+ CHECK_EQUAL_32(ZCFlag, w0);
+ CHECK_EQUAL_32(NFlag, w1);
+ CHECK_EQUAL_32(CFlag, w2);
+ CHECK_EQUAL_32(CVFlag, w3);
+ CHECK_EQUAL_32(CVFlag, w4);
+ CHECK_EQUAL_32(ZCFlag, w5);
+ CHECK_EQUAL_32(NFlag, w6);
+ CHECK_EQUAL_32(ZCFlag, w10);
+ CHECK_EQUAL_32(NFlag, w11);
+ CHECK_EQUAL_32(CFlag, w12);
+ CHECK_EQUAL_32(CVFlag, w13);
+ CHECK_EQUAL_32(CVFlag, w14);
+ CHECK_EQUAL_32(ZCFlag, w15);
+ CHECK_EQUAL_32(NFlag, w16);
TEARDOWN();
}
@@ -5953,12 +5953,12 @@ TEST(fcsel) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s0);
- ASSERT_EQUAL_FP32(2.0, s1);
- ASSERT_EQUAL_FP64(3.0, d2);
- ASSERT_EQUAL_FP64(4.0, d3);
- ASSERT_EQUAL_FP32(1.0, s4);
- ASSERT_EQUAL_FP64(3.0, d5);
+ CHECK_EQUAL_FP32(1.0, s0);
+ CHECK_EQUAL_FP32(2.0, s1);
+ CHECK_EQUAL_FP64(3.0, d2);
+ CHECK_EQUAL_FP64(4.0, d3);
+ CHECK_EQUAL_FP32(1.0, s4);
+ CHECK_EQUAL_FP64(3.0, d5);
TEARDOWN();
}
@@ -5992,18 +5992,18 @@ TEST(fneg) {
RUN();
- ASSERT_EQUAL_FP32(-1.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(-0.0, s2);
- ASSERT_EQUAL_FP32(0.0, s3);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s4);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s5);
- ASSERT_EQUAL_FP64(-1.0, d6);
- ASSERT_EQUAL_FP64(1.0, d7);
- ASSERT_EQUAL_FP64(-0.0, d8);
- ASSERT_EQUAL_FP64(0.0, d9);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d10);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d11);
+ CHECK_EQUAL_FP32(-1.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(-0.0, s2);
+ CHECK_EQUAL_FP32(0.0, s3);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s4);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s5);
+ CHECK_EQUAL_FP64(-1.0, d6);
+ CHECK_EQUAL_FP64(1.0, d7);
+ CHECK_EQUAL_FP64(-0.0, d8);
+ CHECK_EQUAL_FP64(0.0, d9);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d10);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d11);
TEARDOWN();
}
@@ -6033,14 +6033,14 @@ TEST(fabs) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(0.0, s2);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s3);
- ASSERT_EQUAL_FP64(1.0, d4);
- ASSERT_EQUAL_FP64(1.0, d5);
- ASSERT_EQUAL_FP64(0.0, d6);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d7);
+ CHECK_EQUAL_FP32(1.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(0.0, s2);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s3);
+ CHECK_EQUAL_FP64(1.0, d4);
+ CHECK_EQUAL_FP64(1.0, d5);
+ CHECK_EQUAL_FP64(0.0, d6);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d7);
TEARDOWN();
}
@@ -6084,20 +6084,20 @@ TEST(fsqrt) {
RUN();
- ASSERT_EQUAL_FP32(0.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(0.5, s2);
- ASSERT_EQUAL_FP32(256.0, s3);
- ASSERT_EQUAL_FP32(-0.0, s4);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s5);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s6);
- ASSERT_EQUAL_FP64(0.0, d7);
- ASSERT_EQUAL_FP64(1.0, d8);
- ASSERT_EQUAL_FP64(0.5, d9);
- ASSERT_EQUAL_FP64(65536.0, d10);
- ASSERT_EQUAL_FP64(-0.0, d11);
- ASSERT_EQUAL_FP64(kFP32PositiveInfinity, d12);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d13);
+ CHECK_EQUAL_FP32(0.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(0.5, s2);
+ CHECK_EQUAL_FP32(256.0, s3);
+ CHECK_EQUAL_FP32(-0.0, s4);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s5);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s6);
+ CHECK_EQUAL_FP64(0.0, d7);
+ CHECK_EQUAL_FP64(1.0, d8);
+ CHECK_EQUAL_FP64(0.5, d9);
+ CHECK_EQUAL_FP64(65536.0, d10);
+ CHECK_EQUAL_FP64(-0.0, d11);
+ CHECK_EQUAL_FP64(kFP32PositiveInfinity, d12);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d13);
TEARDOWN();
}
@@ -6163,30 +6163,30 @@ TEST(frinta) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(2.0, s2);
- ASSERT_EQUAL_FP32(2.0, s3);
- ASSERT_EQUAL_FP32(3.0, s4);
- ASSERT_EQUAL_FP32(-2.0, s5);
- ASSERT_EQUAL_FP32(-3.0, s6);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s7);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s8);
- ASSERT_EQUAL_FP32(0.0, s9);
- ASSERT_EQUAL_FP32(-0.0, s10);
- ASSERT_EQUAL_FP32(-0.0, s11);
- ASSERT_EQUAL_FP64(1.0, d12);
- ASSERT_EQUAL_FP64(1.0, d13);
- ASSERT_EQUAL_FP64(2.0, d14);
- ASSERT_EQUAL_FP64(2.0, d15);
- ASSERT_EQUAL_FP64(3.0, d16);
- ASSERT_EQUAL_FP64(-2.0, d17);
- ASSERT_EQUAL_FP64(-3.0, d18);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d19);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d20);
- ASSERT_EQUAL_FP64(0.0, d21);
- ASSERT_EQUAL_FP64(-0.0, d22);
- ASSERT_EQUAL_FP64(-0.0, d23);
+ CHECK_EQUAL_FP32(1.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(2.0, s2);
+ CHECK_EQUAL_FP32(2.0, s3);
+ CHECK_EQUAL_FP32(3.0, s4);
+ CHECK_EQUAL_FP32(-2.0, s5);
+ CHECK_EQUAL_FP32(-3.0, s6);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s7);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s8);
+ CHECK_EQUAL_FP32(0.0, s9);
+ CHECK_EQUAL_FP32(-0.0, s10);
+ CHECK_EQUAL_FP32(-0.0, s11);
+ CHECK_EQUAL_FP64(1.0, d12);
+ CHECK_EQUAL_FP64(1.0, d13);
+ CHECK_EQUAL_FP64(2.0, d14);
+ CHECK_EQUAL_FP64(2.0, d15);
+ CHECK_EQUAL_FP64(3.0, d16);
+ CHECK_EQUAL_FP64(-2.0, d17);
+ CHECK_EQUAL_FP64(-3.0, d18);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d19);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d20);
+ CHECK_EQUAL_FP64(0.0, d21);
+ CHECK_EQUAL_FP64(-0.0, d22);
+ CHECK_EQUAL_FP64(-0.0, d23);
TEARDOWN();
}
@@ -6252,30 +6252,30 @@ TEST(frintm) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(1.0, s2);
- ASSERT_EQUAL_FP32(1.0, s3);
- ASSERT_EQUAL_FP32(2.0, s4);
- ASSERT_EQUAL_FP32(-2.0, s5);
- ASSERT_EQUAL_FP32(-3.0, s6);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s7);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s8);
- ASSERT_EQUAL_FP32(0.0, s9);
- ASSERT_EQUAL_FP32(-0.0, s10);
- ASSERT_EQUAL_FP32(-1.0, s11);
- ASSERT_EQUAL_FP64(1.0, d12);
- ASSERT_EQUAL_FP64(1.0, d13);
- ASSERT_EQUAL_FP64(1.0, d14);
- ASSERT_EQUAL_FP64(1.0, d15);
- ASSERT_EQUAL_FP64(2.0, d16);
- ASSERT_EQUAL_FP64(-2.0, d17);
- ASSERT_EQUAL_FP64(-3.0, d18);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d19);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d20);
- ASSERT_EQUAL_FP64(0.0, d21);
- ASSERT_EQUAL_FP64(-0.0, d22);
- ASSERT_EQUAL_FP64(-1.0, d23);
+ CHECK_EQUAL_FP32(1.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(1.0, s2);
+ CHECK_EQUAL_FP32(1.0, s3);
+ CHECK_EQUAL_FP32(2.0, s4);
+ CHECK_EQUAL_FP32(-2.0, s5);
+ CHECK_EQUAL_FP32(-3.0, s6);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s7);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s8);
+ CHECK_EQUAL_FP32(0.0, s9);
+ CHECK_EQUAL_FP32(-0.0, s10);
+ CHECK_EQUAL_FP32(-1.0, s11);
+ CHECK_EQUAL_FP64(1.0, d12);
+ CHECK_EQUAL_FP64(1.0, d13);
+ CHECK_EQUAL_FP64(1.0, d14);
+ CHECK_EQUAL_FP64(1.0, d15);
+ CHECK_EQUAL_FP64(2.0, d16);
+ CHECK_EQUAL_FP64(-2.0, d17);
+ CHECK_EQUAL_FP64(-3.0, d18);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d19);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d20);
+ CHECK_EQUAL_FP64(0.0, d21);
+ CHECK_EQUAL_FP64(-0.0, d22);
+ CHECK_EQUAL_FP64(-1.0, d23);
TEARDOWN();
}
@@ -6341,30 +6341,30 @@ TEST(frintn) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(2.0, s2);
- ASSERT_EQUAL_FP32(2.0, s3);
- ASSERT_EQUAL_FP32(2.0, s4);
- ASSERT_EQUAL_FP32(-2.0, s5);
- ASSERT_EQUAL_FP32(-2.0, s6);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s7);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s8);
- ASSERT_EQUAL_FP32(0.0, s9);
- ASSERT_EQUAL_FP32(-0.0, s10);
- ASSERT_EQUAL_FP32(-0.0, s11);
- ASSERT_EQUAL_FP64(1.0, d12);
- ASSERT_EQUAL_FP64(1.0, d13);
- ASSERT_EQUAL_FP64(2.0, d14);
- ASSERT_EQUAL_FP64(2.0, d15);
- ASSERT_EQUAL_FP64(2.0, d16);
- ASSERT_EQUAL_FP64(-2.0, d17);
- ASSERT_EQUAL_FP64(-2.0, d18);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d19);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d20);
- ASSERT_EQUAL_FP64(0.0, d21);
- ASSERT_EQUAL_FP64(-0.0, d22);
- ASSERT_EQUAL_FP64(-0.0, d23);
+ CHECK_EQUAL_FP32(1.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(2.0, s2);
+ CHECK_EQUAL_FP32(2.0, s3);
+ CHECK_EQUAL_FP32(2.0, s4);
+ CHECK_EQUAL_FP32(-2.0, s5);
+ CHECK_EQUAL_FP32(-2.0, s6);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s7);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s8);
+ CHECK_EQUAL_FP32(0.0, s9);
+ CHECK_EQUAL_FP32(-0.0, s10);
+ CHECK_EQUAL_FP32(-0.0, s11);
+ CHECK_EQUAL_FP64(1.0, d12);
+ CHECK_EQUAL_FP64(1.0, d13);
+ CHECK_EQUAL_FP64(2.0, d14);
+ CHECK_EQUAL_FP64(2.0, d15);
+ CHECK_EQUAL_FP64(2.0, d16);
+ CHECK_EQUAL_FP64(-2.0, d17);
+ CHECK_EQUAL_FP64(-2.0, d18);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d19);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d20);
+ CHECK_EQUAL_FP64(0.0, d21);
+ CHECK_EQUAL_FP64(-0.0, d22);
+ CHECK_EQUAL_FP64(-0.0, d23);
TEARDOWN();
}
@@ -6426,28 +6426,28 @@ TEST(frintz) {
RUN();
- ASSERT_EQUAL_FP32(1.0, s0);
- ASSERT_EQUAL_FP32(1.0, s1);
- ASSERT_EQUAL_FP32(1.0, s2);
- ASSERT_EQUAL_FP32(1.0, s3);
- ASSERT_EQUAL_FP32(2.0, s4);
- ASSERT_EQUAL_FP32(-1.0, s5);
- ASSERT_EQUAL_FP32(-2.0, s6);
- ASSERT_EQUAL_FP32(kFP32PositiveInfinity, s7);
- ASSERT_EQUAL_FP32(kFP32NegativeInfinity, s8);
- ASSERT_EQUAL_FP32(0.0, s9);
- ASSERT_EQUAL_FP32(-0.0, s10);
- ASSERT_EQUAL_FP64(1.0, d11);
- ASSERT_EQUAL_FP64(1.0, d12);
- ASSERT_EQUAL_FP64(1.0, d13);
- ASSERT_EQUAL_FP64(1.0, d14);
- ASSERT_EQUAL_FP64(2.0, d15);
- ASSERT_EQUAL_FP64(-1.0, d16);
- ASSERT_EQUAL_FP64(-2.0, d17);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d18);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d19);
- ASSERT_EQUAL_FP64(0.0, d20);
- ASSERT_EQUAL_FP64(-0.0, d21);
+ CHECK_EQUAL_FP32(1.0, s0);
+ CHECK_EQUAL_FP32(1.0, s1);
+ CHECK_EQUAL_FP32(1.0, s2);
+ CHECK_EQUAL_FP32(1.0, s3);
+ CHECK_EQUAL_FP32(2.0, s4);
+ CHECK_EQUAL_FP32(-1.0, s5);
+ CHECK_EQUAL_FP32(-2.0, s6);
+ CHECK_EQUAL_FP32(kFP32PositiveInfinity, s7);
+ CHECK_EQUAL_FP32(kFP32NegativeInfinity, s8);
+ CHECK_EQUAL_FP32(0.0, s9);
+ CHECK_EQUAL_FP32(-0.0, s10);
+ CHECK_EQUAL_FP64(1.0, d11);
+ CHECK_EQUAL_FP64(1.0, d12);
+ CHECK_EQUAL_FP64(1.0, d13);
+ CHECK_EQUAL_FP64(1.0, d14);
+ CHECK_EQUAL_FP64(2.0, d15);
+ CHECK_EQUAL_FP64(-1.0, d16);
+ CHECK_EQUAL_FP64(-2.0, d17);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d18);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d19);
+ CHECK_EQUAL_FP64(0.0, d20);
+ CHECK_EQUAL_FP64(-0.0, d21);
TEARDOWN();
}
@@ -6493,19 +6493,19 @@ TEST(fcvt_ds) {
RUN();
- ASSERT_EQUAL_FP64(1.0f, d0);
- ASSERT_EQUAL_FP64(1.1f, d1);
- ASSERT_EQUAL_FP64(1.5f, d2);
- ASSERT_EQUAL_FP64(1.9f, d3);
- ASSERT_EQUAL_FP64(2.5f, d4);
- ASSERT_EQUAL_FP64(-1.5f, d5);
- ASSERT_EQUAL_FP64(-2.5f, d6);
- ASSERT_EQUAL_FP64(kFP64PositiveInfinity, d7);
- ASSERT_EQUAL_FP64(kFP64NegativeInfinity, d8);
- ASSERT_EQUAL_FP64(0.0f, d9);
- ASSERT_EQUAL_FP64(-0.0f, d10);
- ASSERT_EQUAL_FP64(FLT_MAX, d11);
- ASSERT_EQUAL_FP64(FLT_MIN, d12);
+ CHECK_EQUAL_FP64(1.0f, d0);
+ CHECK_EQUAL_FP64(1.1f, d1);
+ CHECK_EQUAL_FP64(1.5f, d2);
+ CHECK_EQUAL_FP64(1.9f, d3);
+ CHECK_EQUAL_FP64(2.5f, d4);
+ CHECK_EQUAL_FP64(-1.5f, d5);
+ CHECK_EQUAL_FP64(-2.5f, d6);
+ CHECK_EQUAL_FP64(kFP64PositiveInfinity, d7);
+ CHECK_EQUAL_FP64(kFP64NegativeInfinity, d8);
+ CHECK_EQUAL_FP64(0.0f, d9);
+ CHECK_EQUAL_FP64(-0.0f, d10);
+ CHECK_EQUAL_FP64(FLT_MAX, d11);
+ CHECK_EQUAL_FP64(FLT_MIN, d12);
// Check that the NaN payload is preserved according to ARM64 conversion
// rules:
@@ -6513,8 +6513,8 @@ TEST(fcvt_ds) {
// - The top bit of the mantissa is forced to 1 (making it a quiet NaN).
// - The remaining mantissa bits are copied until they run out.
// - The low-order bits that haven't already been assigned are set to 0.
- ASSERT_EQUAL_FP64(rawbits_to_double(0x7ff82468a0000000), d13);
- ASSERT_EQUAL_FP64(rawbits_to_double(0x7ff82468a0000000), d14);
+ CHECK_EQUAL_FP64(rawbits_to_double(0x7ff82468a0000000), d13);
+ CHECK_EQUAL_FP64(rawbits_to_double(0x7ff82468a0000000), d14);
TEARDOWN();
}
@@ -6614,8 +6614,8 @@ TEST(fcvt_sd) {
float expected = test[i].expected;
// We only expect positive input.
- ASSERT(std::signbit(in) == 0);
- ASSERT(std::signbit(expected) == 0);
+ DCHECK(std::signbit(in) == 0);
+ DCHECK(std::signbit(expected) == 0);
SETUP();
START();
@@ -6628,8 +6628,8 @@ TEST(fcvt_sd) {
END();
RUN();
- ASSERT_EQUAL_FP32(expected, s20);
- ASSERT_EQUAL_FP32(-expected, s21);
+ CHECK_EQUAL_FP32(expected, s20);
+ CHECK_EQUAL_FP32(-expected, s21);
TEARDOWN();
}
}
@@ -6705,36 +6705,36 @@ TEST(fcvtas) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(3, x2);
- ASSERT_EQUAL_64(0xfffffffd, x3);
- ASSERT_EQUAL_64(0x7fffffff, x4);
- ASSERT_EQUAL_64(0x80000000, x5);
- ASSERT_EQUAL_64(0x7fffff80, x6);
- ASSERT_EQUAL_64(0x80000080, x7);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(3, x10);
- ASSERT_EQUAL_64(0xfffffffd, x11);
- ASSERT_EQUAL_64(0x7fffffff, x12);
- ASSERT_EQUAL_64(0x80000000, x13);
- ASSERT_EQUAL_64(0x7ffffffe, x14);
- ASSERT_EQUAL_64(0x80000001, x15);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(3, x18);
- ASSERT_EQUAL_64(0xfffffffffffffffdUL, x19);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0x8000000000000000UL, x21);
- ASSERT_EQUAL_64(0x7fffff8000000000UL, x22);
- ASSERT_EQUAL_64(0x8000008000000000UL, x23);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(3, x25);
- ASSERT_EQUAL_64(0xfffffffffffffffdUL, x26);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x27);
- ASSERT_EQUAL_64(0x8000000000000000UL, x28);
- ASSERT_EQUAL_64(0x7ffffffffffffc00UL, x29);
- ASSERT_EQUAL_64(0x8000000000000400UL, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(3, x2);
+ CHECK_EQUAL_64(0xfffffffd, x3);
+ CHECK_EQUAL_64(0x7fffffff, x4);
+ CHECK_EQUAL_64(0x80000000, x5);
+ CHECK_EQUAL_64(0x7fffff80, x6);
+ CHECK_EQUAL_64(0x80000080, x7);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(3, x10);
+ CHECK_EQUAL_64(0xfffffffd, x11);
+ CHECK_EQUAL_64(0x7fffffff, x12);
+ CHECK_EQUAL_64(0x80000000, x13);
+ CHECK_EQUAL_64(0x7ffffffe, x14);
+ CHECK_EQUAL_64(0x80000001, x15);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(3, x18);
+ CHECK_EQUAL_64(0xfffffffffffffffdUL, x19);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0x8000000000000000UL, x21);
+ CHECK_EQUAL_64(0x7fffff8000000000UL, x22);
+ CHECK_EQUAL_64(0x8000008000000000UL, x23);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(3, x25);
+ CHECK_EQUAL_64(0xfffffffffffffffdUL, x26);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x27);
+ CHECK_EQUAL_64(0x8000000000000000UL, x28);
+ CHECK_EQUAL_64(0x7ffffffffffffc00UL, x29);
+ CHECK_EQUAL_64(0x8000000000000400UL, x30);
TEARDOWN();
}
@@ -6807,34 +6807,34 @@ TEST(fcvtau) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(3, x2);
- ASSERT_EQUAL_64(0, x3);
- ASSERT_EQUAL_64(0xffffffff, x4);
- ASSERT_EQUAL_64(0, x5);
- ASSERT_EQUAL_64(0xffffff00, x6);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(3, x10);
- ASSERT_EQUAL_64(0, x11);
- ASSERT_EQUAL_64(0xffffffff, x12);
- ASSERT_EQUAL_64(0, x13);
- ASSERT_EQUAL_64(0xfffffffe, x14);
- ASSERT_EQUAL_64(1, x16);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(3, x18);
- ASSERT_EQUAL_64(0, x19);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0, x21);
- ASSERT_EQUAL_64(0xffffff0000000000UL, x22);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(3, x25);
- ASSERT_EQUAL_64(0, x26);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x27);
- ASSERT_EQUAL_64(0, x28);
- ASSERT_EQUAL_64(0xfffffffffffff800UL, x29);
- ASSERT_EQUAL_64(0xffffffff, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(3, x2);
+ CHECK_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(0xffffffff, x4);
+ CHECK_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(0xffffff00, x6);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(3, x10);
+ CHECK_EQUAL_64(0, x11);
+ CHECK_EQUAL_64(0xffffffff, x12);
+ CHECK_EQUAL_64(0, x13);
+ CHECK_EQUAL_64(0xfffffffe, x14);
+ CHECK_EQUAL_64(1, x16);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(3, x18);
+ CHECK_EQUAL_64(0, x19);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0, x21);
+ CHECK_EQUAL_64(0xffffff0000000000UL, x22);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(3, x25);
+ CHECK_EQUAL_64(0, x26);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x27);
+ CHECK_EQUAL_64(0, x28);
+ CHECK_EQUAL_64(0xfffffffffffff800UL, x29);
+ CHECK_EQUAL_64(0xffffffff, x30);
TEARDOWN();
}
@@ -6910,36 +6910,36 @@ TEST(fcvtms) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0xfffffffe, x3);
- ASSERT_EQUAL_64(0x7fffffff, x4);
- ASSERT_EQUAL_64(0x80000000, x5);
- ASSERT_EQUAL_64(0x7fffff80, x6);
- ASSERT_EQUAL_64(0x80000080, x7);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(1, x10);
- ASSERT_EQUAL_64(0xfffffffe, x11);
- ASSERT_EQUAL_64(0x7fffffff, x12);
- ASSERT_EQUAL_64(0x80000000, x13);
- ASSERT_EQUAL_64(0x7ffffffe, x14);
- ASSERT_EQUAL_64(0x80000001, x15);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(1, x18);
- ASSERT_EQUAL_64(0xfffffffffffffffeUL, x19);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0x8000000000000000UL, x21);
- ASSERT_EQUAL_64(0x7fffff8000000000UL, x22);
- ASSERT_EQUAL_64(0x8000008000000000UL, x23);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(1, x25);
- ASSERT_EQUAL_64(0xfffffffffffffffeUL, x26);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x27);
- ASSERT_EQUAL_64(0x8000000000000000UL, x28);
- ASSERT_EQUAL_64(0x7ffffffffffffc00UL, x29);
- ASSERT_EQUAL_64(0x8000000000000400UL, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0xfffffffe, x3);
+ CHECK_EQUAL_64(0x7fffffff, x4);
+ CHECK_EQUAL_64(0x80000000, x5);
+ CHECK_EQUAL_64(0x7fffff80, x6);
+ CHECK_EQUAL_64(0x80000080, x7);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(1, x10);
+ CHECK_EQUAL_64(0xfffffffe, x11);
+ CHECK_EQUAL_64(0x7fffffff, x12);
+ CHECK_EQUAL_64(0x80000000, x13);
+ CHECK_EQUAL_64(0x7ffffffe, x14);
+ CHECK_EQUAL_64(0x80000001, x15);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(1, x18);
+ CHECK_EQUAL_64(0xfffffffffffffffeUL, x19);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0x8000000000000000UL, x21);
+ CHECK_EQUAL_64(0x7fffff8000000000UL, x22);
+ CHECK_EQUAL_64(0x8000008000000000UL, x23);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(1, x25);
+ CHECK_EQUAL_64(0xfffffffffffffffeUL, x26);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x27);
+ CHECK_EQUAL_64(0x8000000000000000UL, x28);
+ CHECK_EQUAL_64(0x7ffffffffffffc00UL, x29);
+ CHECK_EQUAL_64(0x8000000000000400UL, x30);
TEARDOWN();
}
@@ -7014,35 +7014,35 @@ TEST(fcvtmu) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0, x3);
- ASSERT_EQUAL_64(0xffffffff, x4);
- ASSERT_EQUAL_64(0, x5);
- ASSERT_EQUAL_64(0x7fffff80, x6);
- ASSERT_EQUAL_64(0, x7);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(1, x10);
- ASSERT_EQUAL_64(0, x11);
- ASSERT_EQUAL_64(0xffffffff, x12);
- ASSERT_EQUAL_64(0, x13);
- ASSERT_EQUAL_64(0x7ffffffe, x14);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(1, x18);
- ASSERT_EQUAL_64(0x0UL, x19);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0x0UL, x21);
- ASSERT_EQUAL_64(0x7fffff8000000000UL, x22);
- ASSERT_EQUAL_64(0x0UL, x23);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(1, x25);
- ASSERT_EQUAL_64(0x0UL, x26);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x27);
- ASSERT_EQUAL_64(0x0UL, x28);
- ASSERT_EQUAL_64(0x7ffffffffffffc00UL, x29);
- ASSERT_EQUAL_64(0x0UL, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(0xffffffff, x4);
+ CHECK_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(0x7fffff80, x6);
+ CHECK_EQUAL_64(0, x7);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(1, x10);
+ CHECK_EQUAL_64(0, x11);
+ CHECK_EQUAL_64(0xffffffff, x12);
+ CHECK_EQUAL_64(0, x13);
+ CHECK_EQUAL_64(0x7ffffffe, x14);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(1, x18);
+ CHECK_EQUAL_64(0x0UL, x19);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0x0UL, x21);
+ CHECK_EQUAL_64(0x7fffff8000000000UL, x22);
+ CHECK_EQUAL_64(0x0UL, x23);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(1, x25);
+ CHECK_EQUAL_64(0x0UL, x26);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x27);
+ CHECK_EQUAL_64(0x0UL, x28);
+ CHECK_EQUAL_64(0x7ffffffffffffc00UL, x29);
+ CHECK_EQUAL_64(0x0UL, x30);
TEARDOWN();
}
@@ -7118,36 +7118,36 @@ TEST(fcvtns) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(2, x2);
- ASSERT_EQUAL_64(0xfffffffe, x3);
- ASSERT_EQUAL_64(0x7fffffff, x4);
- ASSERT_EQUAL_64(0x80000000, x5);
- ASSERT_EQUAL_64(0x7fffff80, x6);
- ASSERT_EQUAL_64(0x80000080, x7);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(2, x10);
- ASSERT_EQUAL_64(0xfffffffe, x11);
- ASSERT_EQUAL_64(0x7fffffff, x12);
- ASSERT_EQUAL_64(0x80000000, x13);
- ASSERT_EQUAL_64(0x7ffffffe, x14);
- ASSERT_EQUAL_64(0x80000001, x15);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(2, x18);
- ASSERT_EQUAL_64(0xfffffffffffffffeUL, x19);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0x8000000000000000UL, x21);
- ASSERT_EQUAL_64(0x7fffff8000000000UL, x22);
- ASSERT_EQUAL_64(0x8000008000000000UL, x23);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(2, x25);
- ASSERT_EQUAL_64(0xfffffffffffffffeUL, x26);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x27);
-// ASSERT_EQUAL_64(0x8000000000000000UL, x28);
- ASSERT_EQUAL_64(0x7ffffffffffffc00UL, x29);
- ASSERT_EQUAL_64(0x8000000000000400UL, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(2, x2);
+ CHECK_EQUAL_64(0xfffffffe, x3);
+ CHECK_EQUAL_64(0x7fffffff, x4);
+ CHECK_EQUAL_64(0x80000000, x5);
+ CHECK_EQUAL_64(0x7fffff80, x6);
+ CHECK_EQUAL_64(0x80000080, x7);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(2, x10);
+ CHECK_EQUAL_64(0xfffffffe, x11);
+ CHECK_EQUAL_64(0x7fffffff, x12);
+ CHECK_EQUAL_64(0x80000000, x13);
+ CHECK_EQUAL_64(0x7ffffffe, x14);
+ CHECK_EQUAL_64(0x80000001, x15);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(2, x18);
+ CHECK_EQUAL_64(0xfffffffffffffffeUL, x19);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0x8000000000000000UL, x21);
+ CHECK_EQUAL_64(0x7fffff8000000000UL, x22);
+ CHECK_EQUAL_64(0x8000008000000000UL, x23);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(2, x25);
+ CHECK_EQUAL_64(0xfffffffffffffffeUL, x26);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x27);
+// CHECK_EQUAL_64(0x8000000000000000UL, x28);
+ CHECK_EQUAL_64(0x7ffffffffffffc00UL, x29);
+ CHECK_EQUAL_64(0x8000000000000400UL, x30);
TEARDOWN();
}
@@ -7220,34 +7220,34 @@ TEST(fcvtnu) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(2, x2);
- ASSERT_EQUAL_64(0, x3);
- ASSERT_EQUAL_64(0xffffffff, x4);
- ASSERT_EQUAL_64(0, x5);
- ASSERT_EQUAL_64(0xffffff00, x6);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(2, x10);
- ASSERT_EQUAL_64(0, x11);
- ASSERT_EQUAL_64(0xffffffff, x12);
- ASSERT_EQUAL_64(0, x13);
- ASSERT_EQUAL_64(0xfffffffe, x14);
- ASSERT_EQUAL_64(1, x16);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(2, x18);
- ASSERT_EQUAL_64(0, x19);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0, x21);
- ASSERT_EQUAL_64(0xffffff0000000000UL, x22);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(2, x25);
- ASSERT_EQUAL_64(0, x26);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x27);
-// ASSERT_EQUAL_64(0, x28);
- ASSERT_EQUAL_64(0xfffffffffffff800UL, x29);
- ASSERT_EQUAL_64(0xffffffff, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(2, x2);
+ CHECK_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(0xffffffff, x4);
+ CHECK_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(0xffffff00, x6);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(2, x10);
+ CHECK_EQUAL_64(0, x11);
+ CHECK_EQUAL_64(0xffffffff, x12);
+ CHECK_EQUAL_64(0, x13);
+ CHECK_EQUAL_64(0xfffffffe, x14);
+ CHECK_EQUAL_64(1, x16);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(2, x18);
+ CHECK_EQUAL_64(0, x19);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0, x21);
+ CHECK_EQUAL_64(0xffffff0000000000UL, x22);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(2, x25);
+ CHECK_EQUAL_64(0, x26);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x27);
+// CHECK_EQUAL_64(0, x28);
+ CHECK_EQUAL_64(0xfffffffffffff800UL, x29);
+ CHECK_EQUAL_64(0xffffffff, x30);
TEARDOWN();
}
@@ -7323,36 +7323,36 @@ TEST(fcvtzs) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0xffffffff, x3);
- ASSERT_EQUAL_64(0x7fffffff, x4);
- ASSERT_EQUAL_64(0x80000000, x5);
- ASSERT_EQUAL_64(0x7fffff80, x6);
- ASSERT_EQUAL_64(0x80000080, x7);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(1, x10);
- ASSERT_EQUAL_64(0xffffffff, x11);
- ASSERT_EQUAL_64(0x7fffffff, x12);
- ASSERT_EQUAL_64(0x80000000, x13);
- ASSERT_EQUAL_64(0x7ffffffe, x14);
- ASSERT_EQUAL_64(0x80000001, x15);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(1, x18);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x19);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0x8000000000000000UL, x21);
- ASSERT_EQUAL_64(0x7fffff8000000000UL, x22);
- ASSERT_EQUAL_64(0x8000008000000000UL, x23);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(1, x25);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x26);
- ASSERT_EQUAL_64(0x7fffffffffffffffUL, x27);
- ASSERT_EQUAL_64(0x8000000000000000UL, x28);
- ASSERT_EQUAL_64(0x7ffffffffffffc00UL, x29);
- ASSERT_EQUAL_64(0x8000000000000400UL, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0xffffffff, x3);
+ CHECK_EQUAL_64(0x7fffffff, x4);
+ CHECK_EQUAL_64(0x80000000, x5);
+ CHECK_EQUAL_64(0x7fffff80, x6);
+ CHECK_EQUAL_64(0x80000080, x7);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(1, x10);
+ CHECK_EQUAL_64(0xffffffff, x11);
+ CHECK_EQUAL_64(0x7fffffff, x12);
+ CHECK_EQUAL_64(0x80000000, x13);
+ CHECK_EQUAL_64(0x7ffffffe, x14);
+ CHECK_EQUAL_64(0x80000001, x15);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(1, x18);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x19);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0x8000000000000000UL, x21);
+ CHECK_EQUAL_64(0x7fffff8000000000UL, x22);
+ CHECK_EQUAL_64(0x8000008000000000UL, x23);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(1, x25);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x26);
+ CHECK_EQUAL_64(0x7fffffffffffffffUL, x27);
+ CHECK_EQUAL_64(0x8000000000000000UL, x28);
+ CHECK_EQUAL_64(0x7ffffffffffffc00UL, x29);
+ CHECK_EQUAL_64(0x8000000000000400UL, x30);
TEARDOWN();
}
@@ -7427,35 +7427,35 @@ TEST(fcvtzu) {
RUN();
- ASSERT_EQUAL_64(1, x0);
- ASSERT_EQUAL_64(1, x1);
- ASSERT_EQUAL_64(1, x2);
- ASSERT_EQUAL_64(0, x3);
- ASSERT_EQUAL_64(0xffffffff, x4);
- ASSERT_EQUAL_64(0, x5);
- ASSERT_EQUAL_64(0x7fffff80, x6);
- ASSERT_EQUAL_64(0, x7);
- ASSERT_EQUAL_64(1, x8);
- ASSERT_EQUAL_64(1, x9);
- ASSERT_EQUAL_64(1, x10);
- ASSERT_EQUAL_64(0, x11);
- ASSERT_EQUAL_64(0xffffffff, x12);
- ASSERT_EQUAL_64(0, x13);
- ASSERT_EQUAL_64(0x7ffffffe, x14);
- ASSERT_EQUAL_64(1, x17);
- ASSERT_EQUAL_64(1, x18);
- ASSERT_EQUAL_64(0x0UL, x19);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x20);
- ASSERT_EQUAL_64(0x0UL, x21);
- ASSERT_EQUAL_64(0x7fffff8000000000UL, x22);
- ASSERT_EQUAL_64(0x0UL, x23);
- ASSERT_EQUAL_64(1, x24);
- ASSERT_EQUAL_64(1, x25);
- ASSERT_EQUAL_64(0x0UL, x26);
- ASSERT_EQUAL_64(0xffffffffffffffffUL, x27);
- ASSERT_EQUAL_64(0x0UL, x28);
- ASSERT_EQUAL_64(0x7ffffffffffffc00UL, x29);
- ASSERT_EQUAL_64(0x0UL, x30);
+ CHECK_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x1);
+ CHECK_EQUAL_64(1, x2);
+ CHECK_EQUAL_64(0, x3);
+ CHECK_EQUAL_64(0xffffffff, x4);
+ CHECK_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(0x7fffff80, x6);
+ CHECK_EQUAL_64(0, x7);
+ CHECK_EQUAL_64(1, x8);
+ CHECK_EQUAL_64(1, x9);
+ CHECK_EQUAL_64(1, x10);
+ CHECK_EQUAL_64(0, x11);
+ CHECK_EQUAL_64(0xffffffff, x12);
+ CHECK_EQUAL_64(0, x13);
+ CHECK_EQUAL_64(0x7ffffffe, x14);
+ CHECK_EQUAL_64(1, x17);
+ CHECK_EQUAL_64(1, x18);
+ CHECK_EQUAL_64(0x0UL, x19);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x20);
+ CHECK_EQUAL_64(0x0UL, x21);
+ CHECK_EQUAL_64(0x7fffff8000000000UL, x22);
+ CHECK_EQUAL_64(0x0UL, x23);
+ CHECK_EQUAL_64(1, x24);
+ CHECK_EQUAL_64(1, x25);
+ CHECK_EQUAL_64(0x0UL, x26);
+ CHECK_EQUAL_64(0xffffffffffffffffUL, x27);
+ CHECK_EQUAL_64(0x0UL, x28);
+ CHECK_EQUAL_64(0x7ffffffffffffc00UL, x29);
+ CHECK_EQUAL_64(0x0UL, x30);
TEARDOWN();
}
@@ -7543,16 +7543,16 @@ static void TestUScvtfHelper(uint64_t in,
for (int fbits = 0; fbits <= 32; fbits++) {
double expected_scvtf = expected_scvtf_base / pow(2.0, fbits);
double expected_ucvtf = expected_ucvtf_base / pow(2.0, fbits);
- ASSERT_EQUAL_FP64(expected_scvtf, results_scvtf_x[fbits]);
- ASSERT_EQUAL_FP64(expected_ucvtf, results_ucvtf_x[fbits]);
- if (cvtf_s32) ASSERT_EQUAL_FP64(expected_scvtf, results_scvtf_w[fbits]);
- if (cvtf_u32) ASSERT_EQUAL_FP64(expected_ucvtf, results_ucvtf_w[fbits]);
+ CHECK_EQUAL_FP64(expected_scvtf, results_scvtf_x[fbits]);
+ CHECK_EQUAL_FP64(expected_ucvtf, results_ucvtf_x[fbits]);
+ if (cvtf_s32) CHECK_EQUAL_FP64(expected_scvtf, results_scvtf_w[fbits]);
+ if (cvtf_u32) CHECK_EQUAL_FP64(expected_ucvtf, results_ucvtf_w[fbits]);
}
for (int fbits = 33; fbits <= 64; fbits++) {
double expected_scvtf = expected_scvtf_base / pow(2.0, fbits);
double expected_ucvtf = expected_ucvtf_base / pow(2.0, fbits);
- ASSERT_EQUAL_FP64(expected_scvtf, results_scvtf_x[fbits]);
- ASSERT_EQUAL_FP64(expected_ucvtf, results_ucvtf_x[fbits]);
+ CHECK_EQUAL_FP64(expected_scvtf, results_scvtf_x[fbits]);
+ CHECK_EQUAL_FP64(expected_ucvtf, results_ucvtf_x[fbits]);
}
TEARDOWN();
@@ -7698,18 +7698,18 @@ static void TestUScvtf32Helper(uint64_t in,
for (int fbits = 0; fbits <= 32; fbits++) {
float expected_scvtf = expected_scvtf_base / powf(2, fbits);
float expected_ucvtf = expected_ucvtf_base / powf(2, fbits);
- ASSERT_EQUAL_FP32(expected_scvtf, results_scvtf_x[fbits]);
- ASSERT_EQUAL_FP32(expected_ucvtf, results_ucvtf_x[fbits]);
- if (cvtf_s32) ASSERT_EQUAL_FP32(expected_scvtf, results_scvtf_w[fbits]);
- if (cvtf_u32) ASSERT_EQUAL_FP32(expected_ucvtf, results_ucvtf_w[fbits]);
+ CHECK_EQUAL_FP32(expected_scvtf, results_scvtf_x[fbits]);
+ CHECK_EQUAL_FP32(expected_ucvtf, results_ucvtf_x[fbits]);
+ if (cvtf_s32) CHECK_EQUAL_FP32(expected_scvtf, results_scvtf_w[fbits]);
+ if (cvtf_u32) CHECK_EQUAL_FP32(expected_ucvtf, results_ucvtf_w[fbits]);
break;
}
for (int fbits = 33; fbits <= 64; fbits++) {
break;
float expected_scvtf = expected_scvtf_base / powf(2, fbits);
float expected_ucvtf = expected_ucvtf_base / powf(2, fbits);
- ASSERT_EQUAL_FP32(expected_scvtf, results_scvtf_x[fbits]);
- ASSERT_EQUAL_FP32(expected_ucvtf, results_ucvtf_x[fbits]);
+ CHECK_EQUAL_FP32(expected_scvtf, results_scvtf_x[fbits]);
+ CHECK_EQUAL_FP32(expected_ucvtf, results_ucvtf_x[fbits]);
}
TEARDOWN();
@@ -7813,13 +7813,13 @@ TEST(system_mrs) {
RUN();
// NZCV
- ASSERT_EQUAL_32(ZCFlag, w3);
- ASSERT_EQUAL_32(NFlag, w4);
- ASSERT_EQUAL_32(ZCVFlag, w5);
+ CHECK_EQUAL_32(ZCFlag, w3);
+ CHECK_EQUAL_32(NFlag, w4);
+ CHECK_EQUAL_32(ZCVFlag, w5);
// FPCR
// The default FPCR on Linux-based platforms is 0.
- ASSERT_EQUAL_32(0, w6);
+ CHECK_EQUAL_32(0, w6);
TEARDOWN();
}
@@ -7887,11 +7887,11 @@ TEST(system_msr) {
RUN();
// We should have incremented x7 (from 0) exactly 8 times.
- ASSERT_EQUAL_64(8, x7);
+ CHECK_EQUAL_64(8, x7);
- ASSERT_EQUAL_64(fpcr_core, x8);
- ASSERT_EQUAL_64(fpcr_core, x9);
- ASSERT_EQUAL_64(0, x10);
+ CHECK_EQUAL_64(fpcr_core, x8);
+ CHECK_EQUAL_64(fpcr_core, x9);
+ CHECK_EQUAL_64(0, x10);
TEARDOWN();
}
@@ -7909,8 +7909,8 @@ TEST(system_nop) {
RUN();
- ASSERT_EQUAL_REGISTERS(before);
- ASSERT_EQUAL_NZCV(before.flags_nzcv());
+ CHECK_EQUAL_REGISTERS(before);
+ CHECK_EQUAL_NZCV(before.flags_nzcv());
TEARDOWN();
}
@@ -7976,8 +7976,8 @@ TEST(zero_dest) {
RUN();
- ASSERT_EQUAL_REGISTERS(before);
- ASSERT_EQUAL_NZCV(before.flags_nzcv());
+ CHECK_EQUAL_REGISTERS(before);
+ CHECK_EQUAL_NZCV(before.flags_nzcv());
TEARDOWN();
}
@@ -8041,7 +8041,7 @@ TEST(zero_dest_setflags) {
RUN();
- ASSERT_EQUAL_REGISTERS(before);
+ CHECK_EQUAL_REGISTERS(before);
TEARDOWN();
}
@@ -8154,15 +8154,15 @@ TEST(peek_poke_simple) {
END();
RUN();
- ASSERT_EQUAL_64(literal_base * 1, x0);
- ASSERT_EQUAL_64(literal_base * 2, x1);
- ASSERT_EQUAL_64(literal_base * 3, x2);
- ASSERT_EQUAL_64(literal_base * 4, x3);
+ CHECK_EQUAL_64(literal_base * 1, x0);
+ CHECK_EQUAL_64(literal_base * 2, x1);
+ CHECK_EQUAL_64(literal_base * 3, x2);
+ CHECK_EQUAL_64(literal_base * 4, x3);
- ASSERT_EQUAL_64((literal_base * 1) & 0xffffffff, x10);
- ASSERT_EQUAL_64((literal_base * 2) & 0xffffffff, x11);
- ASSERT_EQUAL_64((literal_base * 3) & 0xffffffff, x12);
- ASSERT_EQUAL_64((literal_base * 4) & 0xffffffff, x13);
+ CHECK_EQUAL_64((literal_base * 1) & 0xffffffff, x10);
+ CHECK_EQUAL_64((literal_base * 2) & 0xffffffff, x11);
+ CHECK_EQUAL_64((literal_base * 3) & 0xffffffff, x12);
+ CHECK_EQUAL_64((literal_base * 4) & 0xffffffff, x13);
TEARDOWN();
}
@@ -8232,17 +8232,17 @@ TEST(peek_poke_unaligned) {
END();
RUN();
- ASSERT_EQUAL_64(literal_base * 1, x0);
- ASSERT_EQUAL_64(literal_base * 2, x1);
- ASSERT_EQUAL_64(literal_base * 3, x2);
- ASSERT_EQUAL_64(literal_base * 4, x3);
- ASSERT_EQUAL_64(literal_base * 5, x4);
- ASSERT_EQUAL_64(literal_base * 6, x5);
- ASSERT_EQUAL_64(literal_base * 7, x6);
+ CHECK_EQUAL_64(literal_base * 1, x0);
+ CHECK_EQUAL_64(literal_base * 2, x1);
+ CHECK_EQUAL_64(literal_base * 3, x2);
+ CHECK_EQUAL_64(literal_base * 4, x3);
+ CHECK_EQUAL_64(literal_base * 5, x4);
+ CHECK_EQUAL_64(literal_base * 6, x5);
+ CHECK_EQUAL_64(literal_base * 7, x6);
- ASSERT_EQUAL_64((literal_base * 1) & 0xffffffff, x10);
- ASSERT_EQUAL_64((literal_base * 2) & 0xffffffff, x11);
- ASSERT_EQUAL_64((literal_base * 3) & 0xffffffff, x12);
+ CHECK_EQUAL_64((literal_base * 1) & 0xffffffff, x10);
+ CHECK_EQUAL_64((literal_base * 2) & 0xffffffff, x11);
+ CHECK_EQUAL_64((literal_base * 3) & 0xffffffff, x12);
TEARDOWN();
}
@@ -8289,10 +8289,10 @@ TEST(peek_poke_endianness) {
uint64_t x5_expected = ((x1_expected << 16) & 0xffff0000) |
((x1_expected >> 16) & 0x0000ffff);
- ASSERT_EQUAL_64(x0_expected, x0);
- ASSERT_EQUAL_64(x1_expected, x1);
- ASSERT_EQUAL_64(x4_expected, x4);
- ASSERT_EQUAL_64(x5_expected, x5);
+ CHECK_EQUAL_64(x0_expected, x0);
+ CHECK_EQUAL_64(x1_expected, x1);
+ CHECK_EQUAL_64(x4_expected, x4);
+ CHECK_EQUAL_64(x5_expected, x5);
TEARDOWN();
}
@@ -8326,7 +8326,7 @@ TEST(peek_poke_mixed) {
__ Poke(x1, 8);
__ Poke(x0, 0);
{
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(x4, __ StackPointer());
__ SetStackPointer(x4);
@@ -8358,12 +8358,12 @@ TEST(peek_poke_mixed) {
uint64_t x7_expected = ((x1_expected << 16) & 0xffff0000) |
((x0_expected >> 48) & 0x0000ffff);
- ASSERT_EQUAL_64(x0_expected, x0);
- ASSERT_EQUAL_64(x1_expected, x1);
- ASSERT_EQUAL_64(x2_expected, x2);
- ASSERT_EQUAL_64(x3_expected, x3);
- ASSERT_EQUAL_64(x6_expected, x6);
- ASSERT_EQUAL_64(x7_expected, x7);
+ CHECK_EQUAL_64(x0_expected, x0);
+ CHECK_EQUAL_64(x1_expected, x1);
+ CHECK_EQUAL_64(x2_expected, x2);
+ CHECK_EQUAL_64(x3_expected, x3);
+ CHECK_EQUAL_64(x6_expected, x6);
+ CHECK_EQUAL_64(x7_expected, x7);
TEARDOWN();
}
@@ -8423,7 +8423,7 @@ static void PushPopJsspSimpleHelper(int reg_count,
uint64_t literal_base = 0x0100001000100101UL;
{
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(jssp, __ StackPointer());
__ SetStackPointer(jssp);
@@ -8452,7 +8452,7 @@ static void PushPopJsspSimpleHelper(int reg_count,
case 3: __ Push(r[2], r[1], r[0]); break;
case 2: __ Push(r[1], r[0]); break;
case 1: __ Push(r[0]); break;
- default: ASSERT(i == 0); break;
+ default: DCHECK(i == 0); break;
}
break;
case PushPopRegList:
@@ -8474,7 +8474,7 @@ static void PushPopJsspSimpleHelper(int reg_count,
case 3: __ Pop(r[i], r[i+1], r[i+2]); break;
case 2: __ Pop(r[i], r[i+1]); break;
case 1: __ Pop(r[i]); break;
- default: ASSERT(i == reg_count); break;
+ default: DCHECK(i == reg_count); break;
}
break;
case PushPopRegList:
@@ -8494,14 +8494,14 @@ static void PushPopJsspSimpleHelper(int reg_count,
RUN();
// Check that the register contents were preserved.
- // Always use ASSERT_EQUAL_64, even when testing W registers, so we can test
+ // Always use CHECK_EQUAL_64, even when testing W registers, so we can test
// that the upper word was properly cleared by Pop.
literal_base &= (0xffffffffffffffffUL >> (64-reg_size));
for (int i = 0; i < reg_count; i++) {
if (x[i].IsZero()) {
- ASSERT_EQUAL_64(0, x[i]);
+ CHECK_EQUAL_64(0, x[i]);
} else {
- ASSERT_EQUAL_64(literal_base * i, x[i]);
+ CHECK_EQUAL_64(literal_base * i, x[i]);
}
}
@@ -8605,7 +8605,7 @@ static void PushPopFPJsspSimpleHelper(int reg_count,
uint64_t literal_base = 0x0100001000100101UL;
{
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(jssp, __ StackPointer());
__ SetStackPointer(jssp);
@@ -8636,7 +8636,7 @@ static void PushPopFPJsspSimpleHelper(int reg_count,
case 3: __ Push(v[2], v[1], v[0]); break;
case 2: __ Push(v[1], v[0]); break;
case 1: __ Push(v[0]); break;
- default: ASSERT(i == 0); break;
+ default: DCHECK(i == 0); break;
}
break;
case PushPopRegList:
@@ -8658,7 +8658,7 @@ static void PushPopFPJsspSimpleHelper(int reg_count,
case 3: __ Pop(v[i], v[i+1], v[i+2]); break;
case 2: __ Pop(v[i], v[i+1]); break;
case 1: __ Pop(v[i]); break;
- default: ASSERT(i == reg_count); break;
+ default: DCHECK(i == reg_count); break;
}
break;
case PushPopRegList:
@@ -8678,14 +8678,14 @@ static void PushPopFPJsspSimpleHelper(int reg_count,
RUN();
// Check that the register contents were preserved.
- // Always use ASSERT_EQUAL_FP64, even when testing S registers, so we can
+ // Always use CHECK_EQUAL_FP64, even when testing S registers, so we can
// test that the upper word was properly cleared by Pop.
literal_base &= (0xffffffffffffffffUL >> (64-reg_size));
for (int i = 0; i < reg_count; i++) {
uint64_t literal = literal_base * i;
double expected;
memcpy(&expected, &literal, sizeof(expected));
- ASSERT_EQUAL_FP64(expected, d[i]);
+ CHECK_EQUAL_FP64(expected, d[i]);
}
TEARDOWN();
@@ -8782,7 +8782,7 @@ static void PushPopJsspMixedMethodsHelper(int claim, int reg_size) {
START();
{
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(jssp, __ StackPointer());
__ SetStackPointer(jssp);
@@ -8818,16 +8818,16 @@ static void PushPopJsspMixedMethodsHelper(int claim, int reg_size) {
RUN();
- // Always use ASSERT_EQUAL_64, even when testing W registers, so we can test
+ // Always use CHECK_EQUAL_64, even when testing W registers, so we can test
// that the upper word was properly cleared by Pop.
literal_base &= (0xffffffffffffffffUL >> (64-reg_size));
- ASSERT_EQUAL_64(literal_base * 3, x[9]);
- ASSERT_EQUAL_64(literal_base * 2, x[8]);
- ASSERT_EQUAL_64(literal_base * 0, x[7]);
- ASSERT_EQUAL_64(literal_base * 3, x[6]);
- ASSERT_EQUAL_64(literal_base * 1, x[5]);
- ASSERT_EQUAL_64(literal_base * 2, x[4]);
+ CHECK_EQUAL_64(literal_base * 3, x[9]);
+ CHECK_EQUAL_64(literal_base * 2, x[8]);
+ CHECK_EQUAL_64(literal_base * 0, x[7]);
+ CHECK_EQUAL_64(literal_base * 3, x[6]);
+ CHECK_EQUAL_64(literal_base * 1, x[5]);
+ CHECK_EQUAL_64(literal_base * 2, x[4]);
TEARDOWN();
}
@@ -8887,7 +8887,7 @@ static void PushPopJsspWXOverlapHelper(int reg_count, int claim) {
START();
{
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(jssp, __ StackPointer());
__ SetStackPointer(jssp);
@@ -8935,7 +8935,7 @@ static void PushPopJsspWXOverlapHelper(int reg_count, int claim) {
int active_w_slots = 0;
for (int i = 0; active_w_slots < requested_w_slots; i++) {
- ASSERT(i < reg_count);
+ DCHECK(i < reg_count);
// In order to test various arguments to PushMultipleTimes, and to try to
// exercise different alignment and overlap effects, we push each
// register a different number of times.
@@ -9008,7 +9008,7 @@ static void PushPopJsspWXOverlapHelper(int reg_count, int claim) {
}
next_is_64 = !next_is_64;
}
- ASSERT(active_w_slots == 0);
+ DCHECK(active_w_slots == 0);
// Drop memory to restore jssp.
__ Drop(claim, kByteSizeInBytes);
@@ -9036,15 +9036,15 @@ static void PushPopJsspWXOverlapHelper(int reg_count, int claim) {
expected = stack[slot++];
}
- // Always use ASSERT_EQUAL_64, even when testing W registers, so we can
+ // Always use CHECK_EQUAL_64, even when testing W registers, so we can
// test that the upper word was properly cleared by Pop.
if (x[i].IsZero()) {
- ASSERT_EQUAL_64(0, x[i]);
+ CHECK_EQUAL_64(0, x[i]);
} else {
- ASSERT_EQUAL_64(expected, x[i]);
+ CHECK_EQUAL_64(expected, x[i]);
}
}
- ASSERT(slot == requested_w_slots);
+ DCHECK(slot == requested_w_slots);
TEARDOWN();
}
@@ -9074,7 +9074,7 @@ TEST(push_pop_csp) {
START();
- ASSERT(csp.Is(__ StackPointer()));
+ DCHECK(csp.Is(__ StackPointer()));
__ Mov(x3, 0x3333333333333333UL);
__ Mov(x2, 0x2222222222222222UL);
@@ -9119,40 +9119,40 @@ TEST(push_pop_csp) {
RUN();
- ASSERT_EQUAL_64(0x1111111111111111UL, x3);
- ASSERT_EQUAL_64(0x0000000000000000UL, x2);
- ASSERT_EQUAL_64(0x3333333333333333UL, x1);
- ASSERT_EQUAL_64(0x2222222222222222UL, x0);
- ASSERT_EQUAL_64(0x3333333333333333UL, x9);
- ASSERT_EQUAL_64(0x2222222222222222UL, x8);
- ASSERT_EQUAL_64(0x0000000000000000UL, x7);
- ASSERT_EQUAL_64(0x3333333333333333UL, x6);
- ASSERT_EQUAL_64(0x1111111111111111UL, x5);
- ASSERT_EQUAL_64(0x2222222222222222UL, x4);
+ CHECK_EQUAL_64(0x1111111111111111UL, x3);
+ CHECK_EQUAL_64(0x0000000000000000UL, x2);
+ CHECK_EQUAL_64(0x3333333333333333UL, x1);
+ CHECK_EQUAL_64(0x2222222222222222UL, x0);
+ CHECK_EQUAL_64(0x3333333333333333UL, x9);
+ CHECK_EQUAL_64(0x2222222222222222UL, x8);
+ CHECK_EQUAL_64(0x0000000000000000UL, x7);
+ CHECK_EQUAL_64(0x3333333333333333UL, x6);
+ CHECK_EQUAL_64(0x1111111111111111UL, x5);
+ CHECK_EQUAL_64(0x2222222222222222UL, x4);
- ASSERT_EQUAL_32(0x11111111U, w13);
- ASSERT_EQUAL_32(0x33333333U, w12);
- ASSERT_EQUAL_32(0x00000000U, w11);
- ASSERT_EQUAL_32(0x22222222U, w10);
- ASSERT_EQUAL_32(0x11111111U, w17);
- ASSERT_EQUAL_32(0x00000000U, w16);
- ASSERT_EQUAL_32(0x33333333U, w15);
- ASSERT_EQUAL_32(0x22222222U, w14);
+ CHECK_EQUAL_32(0x11111111U, w13);
+ CHECK_EQUAL_32(0x33333333U, w12);
+ CHECK_EQUAL_32(0x00000000U, w11);
+ CHECK_EQUAL_32(0x22222222U, w10);
+ CHECK_EQUAL_32(0x11111111U, w17);
+ CHECK_EQUAL_32(0x00000000U, w16);
+ CHECK_EQUAL_32(0x33333333U, w15);
+ CHECK_EQUAL_32(0x22222222U, w14);
- ASSERT_EQUAL_32(0x11111111U, w18);
- ASSERT_EQUAL_32(0x11111111U, w19);
- ASSERT_EQUAL_32(0x11111111U, w20);
- ASSERT_EQUAL_32(0x11111111U, w21);
- ASSERT_EQUAL_64(0x3333333333333333UL, x22);
- ASSERT_EQUAL_64(0x0000000000000000UL, x23);
+ CHECK_EQUAL_32(0x11111111U, w18);
+ CHECK_EQUAL_32(0x11111111U, w19);
+ CHECK_EQUAL_32(0x11111111U, w20);
+ CHECK_EQUAL_32(0x11111111U, w21);
+ CHECK_EQUAL_64(0x3333333333333333UL, x22);
+ CHECK_EQUAL_64(0x0000000000000000UL, x23);
- ASSERT_EQUAL_64(0x3333333333333333UL, x24);
- ASSERT_EQUAL_64(0x3333333333333333UL, x26);
+ CHECK_EQUAL_64(0x3333333333333333UL, x24);
+ CHECK_EQUAL_64(0x3333333333333333UL, x26);
- ASSERT_EQUAL_32(0x33333333U, w25);
- ASSERT_EQUAL_32(0x00000000U, w27);
- ASSERT_EQUAL_32(0x22222222U, w28);
- ASSERT_EQUAL_32(0x33333333U, w29);
+ CHECK_EQUAL_32(0x33333333U, w25);
+ CHECK_EQUAL_32(0x00000000U, w27);
+ CHECK_EQUAL_32(0x22222222U, w28);
+ CHECK_EQUAL_32(0x33333333U, w29);
TEARDOWN();
}
@@ -9163,7 +9163,7 @@ TEST(push_queued) {
START();
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(jssp, __ StackPointer());
__ SetStackPointer(jssp);
@@ -9214,19 +9214,19 @@ TEST(push_queued) {
RUN();
- ASSERT_EQUAL_64(0x1234000000000000, x0);
- ASSERT_EQUAL_64(0x1234000100010001, x1);
- ASSERT_EQUAL_64(0x1234000200020002, x2);
- ASSERT_EQUAL_64(0x1234000300030003, x3);
+ CHECK_EQUAL_64(0x1234000000000000, x0);
+ CHECK_EQUAL_64(0x1234000100010001, x1);
+ CHECK_EQUAL_64(0x1234000200020002, x2);
+ CHECK_EQUAL_64(0x1234000300030003, x3);
- ASSERT_EQUAL_32(0x12340004, w4);
- ASSERT_EQUAL_32(0x12340005, w5);
- ASSERT_EQUAL_32(0x12340006, w6);
+ CHECK_EQUAL_32(0x12340004, w4);
+ CHECK_EQUAL_32(0x12340005, w5);
+ CHECK_EQUAL_32(0x12340006, w6);
- ASSERT_EQUAL_FP64(123400.0, d0);
- ASSERT_EQUAL_FP64(123401.0, d1);
+ CHECK_EQUAL_FP64(123400.0, d0);
+ CHECK_EQUAL_FP64(123401.0, d1);
- ASSERT_EQUAL_FP32(123402.0, s2);
+ CHECK_EQUAL_FP32(123402.0, s2);
TEARDOWN();
}
@@ -9238,7 +9238,7 @@ TEST(pop_queued) {
START();
- ASSERT(__ StackPointer().Is(csp));
+ DCHECK(__ StackPointer().Is(csp));
__ Mov(jssp, __ StackPointer());
__ SetStackPointer(jssp);
@@ -9289,19 +9289,19 @@ TEST(pop_queued) {
RUN();
- ASSERT_EQUAL_64(0x1234000000000000, x0);
- ASSERT_EQUAL_64(0x1234000100010001, x1);
- ASSERT_EQUAL_64(0x1234000200020002, x2);
- ASSERT_EQUAL_64(0x1234000300030003, x3);
+ CHECK_EQUAL_64(0x1234000000000000, x0);
+ CHECK_EQUAL_64(0x1234000100010001, x1);
+ CHECK_EQUAL_64(0x1234000200020002, x2);
+ CHECK_EQUAL_64(0x1234000300030003, x3);
- ASSERT_EQUAL_64(0x0000000012340004, x4);
- ASSERT_EQUAL_64(0x0000000012340005, x5);
- ASSERT_EQUAL_64(0x0000000012340006, x6);
+ CHECK_EQUAL_64(0x0000000012340004, x4);
+ CHECK_EQUAL_64(0x0000000012340005, x5);
+ CHECK_EQUAL_64(0x0000000012340006, x6);
- ASSERT_EQUAL_FP64(123400.0, d0);
- ASSERT_EQUAL_FP64(123401.0, d1);
+ CHECK_EQUAL_FP64(123400.0, d0);
+ CHECK_EQUAL_FP64(123401.0, d1);
- ASSERT_EQUAL_FP32(123402.0, s2);
+ CHECK_EQUAL_FP32(123402.0, s2);
TEARDOWN();
}
@@ -9367,14 +9367,14 @@ TEST(jump_both_smi) {
RUN();
- ASSERT_EQUAL_64(0x5555555500000001UL, x0);
- ASSERT_EQUAL_64(0xaaaaaaaa00000001UL, x1);
- ASSERT_EQUAL_64(0x1234567800000000UL, x2);
- ASSERT_EQUAL_64(0x8765432100000000UL, x3);
- ASSERT_EQUAL_64(0, x4);
- ASSERT_EQUAL_64(0, x5);
- ASSERT_EQUAL_64(0, x6);
- ASSERT_EQUAL_64(1, x7);
+ CHECK_EQUAL_64(0x5555555500000001UL, x0);
+ CHECK_EQUAL_64(0xaaaaaaaa00000001UL, x1);
+ CHECK_EQUAL_64(0x1234567800000000UL, x2);
+ CHECK_EQUAL_64(0x8765432100000000UL, x3);
+ CHECK_EQUAL_64(0, x4);
+ CHECK_EQUAL_64(0, x5);
+ CHECK_EQUAL_64(0, x6);
+ CHECK_EQUAL_64(1, x7);
TEARDOWN();
}
@@ -9440,14 +9440,14 @@ TEST(jump_either_smi) {
RUN();
- ASSERT_EQUAL_64(0x5555555500000001UL, x0);
- ASSERT_EQUAL_64(0xaaaaaaaa00000001UL, x1);
- ASSERT_EQUAL_64(0x1234567800000000UL, x2);
- ASSERT_EQUAL_64(0x8765432100000000UL, x3);
- ASSERT_EQUAL_64(0, x4);
- ASSERT_EQUAL_64(1, x5);
- ASSERT_EQUAL_64(1, x6);
- ASSERT_EQUAL_64(1, x7);
+ CHECK_EQUAL_64(0x5555555500000001UL, x0);
+ CHECK_EQUAL_64(0xaaaaaaaa00000001UL, x1);
+ CHECK_EQUAL_64(0x1234567800000000UL, x2);
+ CHECK_EQUAL_64(0x8765432100000000UL, x3);
+ CHECK_EQUAL_64(0, x4);
+ CHECK_EQUAL_64(1, x5);
+ CHECK_EQUAL_64(1, x6);
+ CHECK_EQUAL_64(1, x7);
TEARDOWN();
}
@@ -9859,7 +9859,7 @@ TEST(printf) {
__ Printf("%%%%%s%%%c%%\n", x2, w13);
// Print the stack pointer (csp).
- ASSERT(csp.Is(__ StackPointer()));
+ DCHECK(csp.Is(__ StackPointer()));
__ Printf("StackPointer(csp): 0x%016" PRIx64 ", 0x%08" PRIx32 "\n",
__ StackPointer(), __ StackPointer().W());
@@ -9889,7 +9889,7 @@ TEST(printf) {
// Printf preserves all registers by default, we can't look at the number of
// bytes that were printed. However, the printf_no_preserve test should check
// that, and here we just test that we didn't clobber any registers.
- ASSERT_EQUAL_REGISTERS(before);
+ CHECK_EQUAL_REGISTERS(before);
TEARDOWN();
}
@@ -9983,35 +9983,35 @@ TEST(printf_no_preserve) {
// use the return code to check that the string length was correct.
// Printf with no arguments.
- ASSERT_EQUAL_64(strlen(test_plain_string), x19);
+ CHECK_EQUAL_64(strlen(test_plain_string), x19);
// x0: 1234, x1: 0x00001234
- ASSERT_EQUAL_64(25, x20);
+ CHECK_EQUAL_64(25, x20);
// d0: 1.234000
- ASSERT_EQUAL_64(13, x21);
+ CHECK_EQUAL_64(13, x21);
// Test %s: 'This is a substring.'
- ASSERT_EQUAL_64(32, x22);
+ CHECK_EQUAL_64(32, x22);
// w3(uint32): 4294967295
// w4(int32): -1
// x5(uint64): 18446744073709551615
// x6(int64): -1
- ASSERT_EQUAL_64(23 + 14 + 33 + 14, x23);
+ CHECK_EQUAL_64(23 + 14 + 33 + 14, x23);
// %f: 1.234000
// %g: 2.345
// %e: 3.456000e+00
// %E: 4.567000E+00
- ASSERT_EQUAL_64(13 + 10 + 17 + 17, x24);
+ CHECK_EQUAL_64(13 + 10 + 17 + 17, x24);
// 0x89abcdef, 0x123456789abcdef
- ASSERT_EQUAL_64(30, x25);
+ CHECK_EQUAL_64(30, x25);
// 42
- ASSERT_EQUAL_64(3, x26);
+ CHECK_EQUAL_64(3, x26);
// StackPointer(not csp): 0x00007fb037ae2370, 0x37ae2370
// Note: This is an example value, but the field width is fixed here so the
// string length is still predictable.
- ASSERT_EQUAL_64(54, x27);
+ CHECK_EQUAL_64(54, x27);
// 3=3, 4=40, 5=500
- ASSERT_EQUAL_64(17, x28);
+ CHECK_EQUAL_64(17, x28);
// w3: 4294967295, s1: 1.234000, x5: 18446744073709551615, d3: 3.456000
- ASSERT_EQUAL_64(69, x29);
+ CHECK_EQUAL_64(69, x29);
TEARDOWN();
}
@@ -10108,14 +10108,14 @@ static void DoSmiAbsTest(int32_t value, bool must_fail = false) {
if (must_fail) {
// We tested an invalid conversion. The code must have jump on slow.
- ASSERT_EQUAL_64(0xbad, x2);
+ CHECK_EQUAL_64(0xbad, x2);
} else {
// The conversion is valid, check the result.
int32_t result = (value >= 0) ? value : -value;
- ASSERT_EQUAL_64(result, x1);
+ CHECK_EQUAL_64(result, x1);
// Check that we didn't jump on slow.
- ASSERT_EQUAL_64(0xc001c0de, x2);
+ CHECK_EQUAL_64(0xc001c0de, x2);
}
TEARDOWN();
@@ -10162,7 +10162,7 @@ TEST(blr_lr) {
RUN();
- ASSERT_EQUAL_64(0xc001c0de, x0);
+ CHECK_EQUAL_64(0xc001c0de, x0);
TEARDOWN();
}
@@ -10233,14 +10233,14 @@ TEST(process_nan_double) {
// Make sure that NaN propagation works correctly.
double sn = rawbits_to_double(0x7ff5555511111111);
double qn = rawbits_to_double(0x7ffaaaaa11111111);
- ASSERT(IsSignallingNaN(sn));
- ASSERT(IsQuietNaN(qn));
+ DCHECK(IsSignallingNaN(sn));
+ DCHECK(IsQuietNaN(qn));
// The input NaNs after passing through ProcessNaN.
double sn_proc = rawbits_to_double(0x7ffd555511111111);
double qn_proc = qn;
- ASSERT(IsQuietNaN(sn_proc));
- ASSERT(IsQuietNaN(qn_proc));
+ DCHECK(IsQuietNaN(sn_proc));
+ DCHECK(IsQuietNaN(qn_proc));
SETUP();
START();
@@ -10281,24 +10281,24 @@ TEST(process_nan_double) {
uint64_t sn_raw = double_to_rawbits(sn);
// - Signalling NaN
- ASSERT_EQUAL_FP64(sn, d1);
- ASSERT_EQUAL_FP64(rawbits_to_double(sn_raw & ~kDSignMask), d2);
- ASSERT_EQUAL_FP64(rawbits_to_double(sn_raw ^ kDSignMask), d3);
+ CHECK_EQUAL_FP64(sn, d1);
+ CHECK_EQUAL_FP64(rawbits_to_double(sn_raw & ~kDSignMask), d2);
+ CHECK_EQUAL_FP64(rawbits_to_double(sn_raw ^ kDSignMask), d3);
// - Quiet NaN
- ASSERT_EQUAL_FP64(qn, d11);
- ASSERT_EQUAL_FP64(rawbits_to_double(qn_raw & ~kDSignMask), d12);
- ASSERT_EQUAL_FP64(rawbits_to_double(qn_raw ^ kDSignMask), d13);
+ CHECK_EQUAL_FP64(qn, d11);
+ CHECK_EQUAL_FP64(rawbits_to_double(qn_raw & ~kDSignMask), d12);
+ CHECK_EQUAL_FP64(rawbits_to_double(qn_raw ^ kDSignMask), d13);
// - Signalling NaN
- ASSERT_EQUAL_FP64(sn_proc, d4);
- ASSERT_EQUAL_FP64(sn_proc, d5);
- ASSERT_EQUAL_FP64(sn_proc, d6);
- ASSERT_EQUAL_FP64(sn_proc, d7);
+ CHECK_EQUAL_FP64(sn_proc, d4);
+ CHECK_EQUAL_FP64(sn_proc, d5);
+ CHECK_EQUAL_FP64(sn_proc, d6);
+ CHECK_EQUAL_FP64(sn_proc, d7);
// - Quiet NaN
- ASSERT_EQUAL_FP64(qn_proc, d14);
- ASSERT_EQUAL_FP64(qn_proc, d15);
- ASSERT_EQUAL_FP64(qn_proc, d16);
- ASSERT_EQUAL_FP64(qn_proc, d17);
+ CHECK_EQUAL_FP64(qn_proc, d14);
+ CHECK_EQUAL_FP64(qn_proc, d15);
+ CHECK_EQUAL_FP64(qn_proc, d16);
+ CHECK_EQUAL_FP64(qn_proc, d17);
TEARDOWN();
}
@@ -10309,14 +10309,14 @@ TEST(process_nan_float) {
// Make sure that NaN propagation works correctly.
float sn = rawbits_to_float(0x7f951111);
float qn = rawbits_to_float(0x7fea1111);
- ASSERT(IsSignallingNaN(sn));
- ASSERT(IsQuietNaN(qn));
+ DCHECK(IsSignallingNaN(sn));
+ DCHECK(IsQuietNaN(qn));
// The input NaNs after passing through ProcessNaN.
float sn_proc = rawbits_to_float(0x7fd51111);
float qn_proc = qn;
- ASSERT(IsQuietNaN(sn_proc));
- ASSERT(IsQuietNaN(qn_proc));
+ DCHECK(IsQuietNaN(sn_proc));
+ DCHECK(IsQuietNaN(qn_proc));
SETUP();
START();
@@ -10357,32 +10357,32 @@ TEST(process_nan_float) {
uint32_t sn_raw = float_to_rawbits(sn);
// - Signalling NaN
- ASSERT_EQUAL_FP32(sn, s1);
- ASSERT_EQUAL_FP32(rawbits_to_float(sn_raw & ~kSSignMask), s2);
- ASSERT_EQUAL_FP32(rawbits_to_float(sn_raw ^ kSSignMask), s3);
+ CHECK_EQUAL_FP32(sn, s1);
+ CHECK_EQUAL_FP32(rawbits_to_float(sn_raw & ~kSSignMask), s2);
+ CHECK_EQUAL_FP32(rawbits_to_float(sn_raw ^ kSSignMask), s3);
// - Quiet NaN
- ASSERT_EQUAL_FP32(qn, s11);
- ASSERT_EQUAL_FP32(rawbits_to_float(qn_raw & ~kSSignMask), s12);
- ASSERT_EQUAL_FP32(rawbits_to_float(qn_raw ^ kSSignMask), s13);
+ CHECK_EQUAL_FP32(qn, s11);
+ CHECK_EQUAL_FP32(rawbits_to_float(qn_raw & ~kSSignMask), s12);
+ CHECK_EQUAL_FP32(rawbits_to_float(qn_raw ^ kSSignMask), s13);
// - Signalling NaN
- ASSERT_EQUAL_FP32(sn_proc, s4);
- ASSERT_EQUAL_FP32(sn_proc, s5);
- ASSERT_EQUAL_FP32(sn_proc, s6);
- ASSERT_EQUAL_FP32(sn_proc, s7);
+ CHECK_EQUAL_FP32(sn_proc, s4);
+ CHECK_EQUAL_FP32(sn_proc, s5);
+ CHECK_EQUAL_FP32(sn_proc, s6);
+ CHECK_EQUAL_FP32(sn_proc, s7);
// - Quiet NaN
- ASSERT_EQUAL_FP32(qn_proc, s14);
- ASSERT_EQUAL_FP32(qn_proc, s15);
- ASSERT_EQUAL_FP32(qn_proc, s16);
- ASSERT_EQUAL_FP32(qn_proc, s17);
+ CHECK_EQUAL_FP32(qn_proc, s14);
+ CHECK_EQUAL_FP32(qn_proc, s15);
+ CHECK_EQUAL_FP32(qn_proc, s16);
+ CHECK_EQUAL_FP32(qn_proc, s17);
TEARDOWN();
}
static void ProcessNaNsHelper(double n, double m, double expected) {
- ASSERT(std::isnan(n) || std::isnan(m));
- ASSERT(std::isnan(expected));
+ DCHECK(std::isnan(n) || std::isnan(m));
+ DCHECK(std::isnan(expected));
SETUP();
START();
@@ -10402,12 +10402,12 @@ static void ProcessNaNsHelper(double n, double m, double expected) {
END();
RUN();
- ASSERT_EQUAL_FP64(expected, d2);
- ASSERT_EQUAL_FP64(expected, d3);
- ASSERT_EQUAL_FP64(expected, d4);
- ASSERT_EQUAL_FP64(expected, d5);
- ASSERT_EQUAL_FP64(expected, d6);
- ASSERT_EQUAL_FP64(expected, d7);
+ CHECK_EQUAL_FP64(expected, d2);
+ CHECK_EQUAL_FP64(expected, d3);
+ CHECK_EQUAL_FP64(expected, d4);
+ CHECK_EQUAL_FP64(expected, d5);
+ CHECK_EQUAL_FP64(expected, d6);
+ CHECK_EQUAL_FP64(expected, d7);
TEARDOWN();
}
@@ -10420,20 +10420,20 @@ TEST(process_nans_double) {
double sm = rawbits_to_double(0x7ff5555522222222);
double qn = rawbits_to_double(0x7ffaaaaa11111111);
double qm = rawbits_to_double(0x7ffaaaaa22222222);
- ASSERT(IsSignallingNaN(sn));
- ASSERT(IsSignallingNaN(sm));
- ASSERT(IsQuietNaN(qn));
- ASSERT(IsQuietNaN(qm));
+ DCHECK(IsSignallingNaN(sn));
+ DCHECK(IsSignallingNaN(sm));
+ DCHECK(IsQuietNaN(qn));
+ DCHECK(IsQuietNaN(qm));
// The input NaNs after passing through ProcessNaN.
double sn_proc = rawbits_to_double(0x7ffd555511111111);
double sm_proc = rawbits_to_double(0x7ffd555522222222);
double qn_proc = qn;
double qm_proc = qm;
- ASSERT(IsQuietNaN(sn_proc));
- ASSERT(IsQuietNaN(sm_proc));
- ASSERT(IsQuietNaN(qn_proc));
- ASSERT(IsQuietNaN(qm_proc));
+ DCHECK(IsQuietNaN(sn_proc));
+ DCHECK(IsQuietNaN(sm_proc));
+ DCHECK(IsQuietNaN(qn_proc));
+ DCHECK(IsQuietNaN(qm_proc));
// Quiet NaNs are propagated.
ProcessNaNsHelper(qn, 0, qn_proc);
@@ -10453,8 +10453,8 @@ TEST(process_nans_double) {
static void ProcessNaNsHelper(float n, float m, float expected) {
- ASSERT(std::isnan(n) || std::isnan(m));
- ASSERT(std::isnan(expected));
+ DCHECK(std::isnan(n) || std::isnan(m));
+ DCHECK(std::isnan(expected));
SETUP();
START();
@@ -10474,12 +10474,12 @@ static void ProcessNaNsHelper(float n, float m, float expected) {
END();
RUN();
- ASSERT_EQUAL_FP32(expected, s2);
- ASSERT_EQUAL_FP32(expected, s3);
- ASSERT_EQUAL_FP32(expected, s4);
- ASSERT_EQUAL_FP32(expected, s5);
- ASSERT_EQUAL_FP32(expected, s6);
- ASSERT_EQUAL_FP32(expected, s7);
+ CHECK_EQUAL_FP32(expected, s2);
+ CHECK_EQUAL_FP32(expected, s3);
+ CHECK_EQUAL_FP32(expected, s4);
+ CHECK_EQUAL_FP32(expected, s5);
+ CHECK_EQUAL_FP32(expected, s6);
+ CHECK_EQUAL_FP32(expected, s7);
TEARDOWN();
}
@@ -10492,20 +10492,20 @@ TEST(process_nans_float) {
float sm = rawbits_to_float(0x7f952222);
float qn = rawbits_to_float(0x7fea1111);
float qm = rawbits_to_float(0x7fea2222);
- ASSERT(IsSignallingNaN(sn));
- ASSERT(IsSignallingNaN(sm));
- ASSERT(IsQuietNaN(qn));
- ASSERT(IsQuietNaN(qm));
+ DCHECK(IsSignallingNaN(sn));
+ DCHECK(IsSignallingNaN(sm));
+ DCHECK(IsQuietNaN(qn));
+ DCHECK(IsQuietNaN(qm));
// The input NaNs after passing through ProcessNaN.
float sn_proc = rawbits_to_float(0x7fd51111);
float sm_proc = rawbits_to_float(0x7fd52222);
float qn_proc = qn;
float qm_proc = qm;
- ASSERT(IsQuietNaN(sn_proc));
- ASSERT(IsQuietNaN(sm_proc));
- ASSERT(IsQuietNaN(qn_proc));
- ASSERT(IsQuietNaN(qm_proc));
+ DCHECK(IsQuietNaN(sn_proc));
+ DCHECK(IsQuietNaN(sm_proc));
+ DCHECK(IsQuietNaN(qn_proc));
+ DCHECK(IsQuietNaN(qm_proc));
// Quiet NaNs are propagated.
ProcessNaNsHelper(qn, 0, qn_proc);
@@ -10525,7 +10525,7 @@ TEST(process_nans_float) {
static void DefaultNaNHelper(float n, float m, float a) {
- ASSERT(std::isnan(n) || std::isnan(m) || std::isnan(a));
+ DCHECK(std::isnan(n) || std::isnan(m) || std::isnan(a));
bool test_1op = std::isnan(n);
bool test_2op = std::isnan(n) || std::isnan(m);
@@ -10582,29 +10582,29 @@ static void DefaultNaNHelper(float n, float m, float a) {
if (test_1op) {
uint32_t n_raw = float_to_rawbits(n);
- ASSERT_EQUAL_FP32(n, s10);
- ASSERT_EQUAL_FP32(rawbits_to_float(n_raw & ~kSSignMask), s11);
- ASSERT_EQUAL_FP32(rawbits_to_float(n_raw ^ kSSignMask), s12);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s13);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s14);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s15);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s16);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d17);
+ CHECK_EQUAL_FP32(n, s10);
+ CHECK_EQUAL_FP32(rawbits_to_float(n_raw & ~kSSignMask), s11);
+ CHECK_EQUAL_FP32(rawbits_to_float(n_raw ^ kSSignMask), s12);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s13);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s14);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s15);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s16);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d17);
}
if (test_2op) {
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s18);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s19);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s20);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s21);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s22);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s23);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s18);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s19);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s20);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s21);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s22);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s23);
}
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s24);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s25);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s26);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s27);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s24);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s25);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s26);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s27);
TEARDOWN();
}
@@ -10618,12 +10618,12 @@ TEST(default_nan_float) {
float qn = rawbits_to_float(0x7fea1111);
float qm = rawbits_to_float(0x7fea2222);
float qa = rawbits_to_float(0x7feaaaaa);
- ASSERT(IsSignallingNaN(sn));
- ASSERT(IsSignallingNaN(sm));
- ASSERT(IsSignallingNaN(sa));
- ASSERT(IsQuietNaN(qn));
- ASSERT(IsQuietNaN(qm));
- ASSERT(IsQuietNaN(qa));
+ DCHECK(IsSignallingNaN(sn));
+ DCHECK(IsSignallingNaN(sm));
+ DCHECK(IsSignallingNaN(sa));
+ DCHECK(IsQuietNaN(qn));
+ DCHECK(IsQuietNaN(qm));
+ DCHECK(IsQuietNaN(qa));
// - Signalling NaNs
DefaultNaNHelper(sn, 0.0f, 0.0f);
@@ -10653,7 +10653,7 @@ TEST(default_nan_float) {
static void DefaultNaNHelper(double n, double m, double a) {
- ASSERT(std::isnan(n) || std::isnan(m) || std::isnan(a));
+ DCHECK(std::isnan(n) || std::isnan(m) || std::isnan(a));
bool test_1op = std::isnan(n);
bool test_2op = std::isnan(n) || std::isnan(m);
@@ -10710,29 +10710,29 @@ static void DefaultNaNHelper(double n, double m, double a) {
if (test_1op) {
uint64_t n_raw = double_to_rawbits(n);
- ASSERT_EQUAL_FP64(n, d10);
- ASSERT_EQUAL_FP64(rawbits_to_double(n_raw & ~kDSignMask), d11);
- ASSERT_EQUAL_FP64(rawbits_to_double(n_raw ^ kDSignMask), d12);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d13);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d14);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d15);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d16);
- ASSERT_EQUAL_FP32(kFP32DefaultNaN, s17);
+ CHECK_EQUAL_FP64(n, d10);
+ CHECK_EQUAL_FP64(rawbits_to_double(n_raw & ~kDSignMask), d11);
+ CHECK_EQUAL_FP64(rawbits_to_double(n_raw ^ kDSignMask), d12);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d13);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d14);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d15);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d16);
+ CHECK_EQUAL_FP32(kFP32DefaultNaN, s17);
}
if (test_2op) {
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d18);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d19);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d20);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d21);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d22);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d23);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d18);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d19);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d20);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d21);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d22);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d23);
}
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d24);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d25);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d26);
- ASSERT_EQUAL_FP64(kFP64DefaultNaN, d27);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d24);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d25);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d26);
+ CHECK_EQUAL_FP64(kFP64DefaultNaN, d27);
TEARDOWN();
}
@@ -10746,12 +10746,12 @@ TEST(default_nan_double) {
double qn = rawbits_to_double(0x7ffaaaaa11111111);
double qm = rawbits_to_double(0x7ffaaaaa22222222);
double qa = rawbits_to_double(0x7ffaaaaaaaaaaaaa);
- ASSERT(IsSignallingNaN(sn));
- ASSERT(IsSignallingNaN(sm));
- ASSERT(IsSignallingNaN(sa));
- ASSERT(IsQuietNaN(qn));
- ASSERT(IsQuietNaN(qm));
- ASSERT(IsQuietNaN(qa));
+ DCHECK(IsSignallingNaN(sn));
+ DCHECK(IsSignallingNaN(sm));
+ DCHECK(IsSignallingNaN(sa));
+ DCHECK(IsQuietNaN(qn));
+ DCHECK(IsQuietNaN(qm));
+ DCHECK(IsQuietNaN(qa));
// - Signalling NaNs
DefaultNaNHelper(sn, 0.0, 0.0);
@@ -10812,7 +10812,7 @@ TEST(call_no_relocation) {
RUN();
- ASSERT_EQUAL_64(1, x0);
+ CHECK_EQUAL_64(1, x0);
// The return_address_from_call_start function doesn't currently encounter any
// non-relocatable sequences, so we check it here to make sure it works.
@@ -10869,12 +10869,12 @@ static void AbsHelperX(int64_t value) {
END();
RUN();
- ASSERT_EQUAL_64(0, x0);
- ASSERT_EQUAL_64(value, x1);
- ASSERT_EQUAL_64(expected, x10);
- ASSERT_EQUAL_64(expected, x11);
- ASSERT_EQUAL_64(expected, x12);
- ASSERT_EQUAL_64(expected, x13);
+ CHECK_EQUAL_64(0, x0);
+ CHECK_EQUAL_64(value, x1);
+ CHECK_EQUAL_64(expected, x10);
+ CHECK_EQUAL_64(expected, x11);
+ CHECK_EQUAL_64(expected, x12);
+ CHECK_EQUAL_64(expected, x13);
TEARDOWN();
}
@@ -10926,12 +10926,12 @@ static void AbsHelperW(int32_t value) {
END();
RUN();
- ASSERT_EQUAL_32(0, w0);
- ASSERT_EQUAL_32(value, w1);
- ASSERT_EQUAL_32(expected, w10);
- ASSERT_EQUAL_32(expected, w11);
- ASSERT_EQUAL_32(expected, w12);
- ASSERT_EQUAL_32(expected, w13);
+ CHECK_EQUAL_32(0, w0);
+ CHECK_EQUAL_32(value, w1);
+ CHECK_EQUAL_32(expected, w10);
+ CHECK_EQUAL_32(expected, w11);
+ CHECK_EQUAL_32(expected, w12);
+ CHECK_EQUAL_32(expected, w13);
TEARDOWN();
}
@@ -10989,16 +10989,16 @@ TEST(pool_size) {
for (RelocIterator it(*code, pool_mask); !it.done(); it.next()) {
RelocInfo* info = it.rinfo();
if (RelocInfo::IsConstPool(info->rmode())) {
- ASSERT(info->data() == constant_pool_size);
+ DCHECK(info->data() == constant_pool_size);
++pool_count;
}
if (RelocInfo::IsVeneerPool(info->rmode())) {
- ASSERT(info->data() == veneer_pool_size);
+ DCHECK(info->data() == veneer_pool_size);
++pool_count;
}
}
- ASSERT(pool_count == 2);
+ DCHECK(pool_count == 2);
TEARDOWN();
}
« no previous file with comments | « test/cctest/test-assembler-arm.cc ('k') | test/cctest/test-disasm-arm64.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698