Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(242)

Unified Diff: test/cctest/test-assembler-mips64.cc

Issue 1173343006: Replace OFFSET_OF with offsetof as far as possible. (Closed) Base URL: https://chromium.googlesource.com/v8/v8.git@master
Patch Set: Fixed typing chaos on ARM. Created 5 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « test/cctest/test-assembler-mips.cc ('k') | test/cctest/test-assembler-ppc.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: test/cctest/test-assembler-mips64.cc
diff --git a/test/cctest/test-assembler-mips64.cc b/test/cctest/test-assembler-mips64.cc
index 0494a755c972085db7eac867e5ccb199ef0e875b..f3a94896d0fdd905dda78a7e6ba8f2a74827901c 100644
--- a/test/cctest/test-assembler-mips64.cc
+++ b/test/cctest/test-assembler-mips64.cc
@@ -289,61 +289,61 @@ TEST(MIPS3) {
Label L, C;
// Double precision floating point instructions.
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, a)) );
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(T, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(T, a)) );
+ __ ldc1(f6, MemOperand(a0, offsetof(T, b)) );
__ add_d(f8, f4, f6);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(T, c)) ); // c = a + b.
+ __ sdc1(f8, MemOperand(a0, offsetof(T, c)) ); // c = a + b.
__ mov_d(f10, f8); // c
__ neg_d(f12, f6); // -b
__ sub_d(f10, f10, f12);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(T, d)) ); // d = c - (-b).
+ __ sdc1(f10, MemOperand(a0, offsetof(T, d)) ); // d = c - (-b).
- __ sdc1(f4, MemOperand(a0, OFFSET_OF(T, b)) ); // b = a.
+ __ sdc1(f4, MemOperand(a0, offsetof(T, b)) ); // b = a.
__ li(a4, 120);
__ mtc1(a4, f14);
__ cvt_d_w(f14, f14); // f14 = 120.0.
__ mul_d(f10, f10, f14);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(T, e)) ); // e = d * 120 = 1.8066e16.
+ __ sdc1(f10, MemOperand(a0, offsetof(T, e)) ); // e = d * 120 = 1.8066e16.
__ div_d(f12, f10, f4);
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(T, f)) ); // f = e / a = 120.44.
+ __ sdc1(f12, MemOperand(a0, offsetof(T, f)) ); // f = e / a = 120.44.
__ sqrt_d(f14, f12);
- __ sdc1(f14, MemOperand(a0, OFFSET_OF(T, g)) );
+ __ sdc1(f14, MemOperand(a0, offsetof(T, g)) );
// g = sqrt(f) = 10.97451593465515908537
if (kArchVariant == kMips64r2) {
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, h)) );
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(T, i)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(T, h)) );
+ __ ldc1(f6, MemOperand(a0, offsetof(T, i)) );
__ madd_d(f14, f6, f4, f6);
- __ sdc1(f14, MemOperand(a0, OFFSET_OF(T, h)) );
+ __ sdc1(f14, MemOperand(a0, offsetof(T, h)) );
}
// Single precision floating point instructions.
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fa)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(T, fb)) );
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fa)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(T, fb)) );
__ add_s(f8, f4, f6);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(T, fc)) ); // fc = fa + fb.
+ __ swc1(f8, MemOperand(a0, offsetof(T, fc)) ); // fc = fa + fb.
__ neg_s(f10, f6); // -fb
__ sub_s(f10, f8, f10);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(T, fd)) ); // fd = fc - (-fb).
+ __ swc1(f10, MemOperand(a0, offsetof(T, fd)) ); // fd = fc - (-fb).
- __ swc1(f4, MemOperand(a0, OFFSET_OF(T, fb)) ); // fb = fa.
+ __ swc1(f4, MemOperand(a0, offsetof(T, fb)) ); // fb = fa.
__ li(t0, 120);
__ mtc1(t0, f14);
__ cvt_s_w(f14, f14); // f14 = 120.0.
__ mul_s(f10, f10, f14);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(T, fe)) ); // fe = fd * 120
+ __ swc1(f10, MemOperand(a0, offsetof(T, fe)) ); // fe = fd * 120
__ div_s(f12, f10, f4);
- __ swc1(f12, MemOperand(a0, OFFSET_OF(T, ff)) ); // ff = fe / fa
+ __ swc1(f12, MemOperand(a0, offsetof(T, ff)) ); // ff = fe / fa
__ sqrt_s(f14, f12);
- __ swc1(f14, MemOperand(a0, OFFSET_OF(T, fg)) );
+ __ swc1(f14, MemOperand(a0, offsetof(T, fg)) );
__ jr(ra);
__ nop();
@@ -412,8 +412,8 @@ TEST(MIPS4) {
Assembler assm(isolate, NULL, 0);
Label L, C;
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, a)));
- __ ldc1(f5, MemOperand(a0, OFFSET_OF(T, b)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, a)));
+ __ ldc1(f5, MemOperand(a0, offsetof(T, b)));
// Swap f4 and f5, by using 3 integer registers, a4-a6,
// both two 32-bit chunks, and one 64-bit chunk.
@@ -428,16 +428,16 @@ TEST(MIPS4) {
__ dmtc1(a6, f4);
// Store the swapped f4 and f5 back to memory.
- __ sdc1(f4, MemOperand(a0, OFFSET_OF(T, a)));
- __ sdc1(f5, MemOperand(a0, OFFSET_OF(T, c)));
+ __ sdc1(f4, MemOperand(a0, offsetof(T, a)));
+ __ sdc1(f5, MemOperand(a0, offsetof(T, c)));
// Test sign extension of move operations from coprocessor.
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, d)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, d)));
__ mfhc1(a4, f4);
__ mfc1(a5, f4);
- __ sd(a4, MemOperand(a0, OFFSET_OF(T, high)));
- __ sd(a5, MemOperand(a0, OFFSET_OF(T, low)));
+ __ sd(a4, MemOperand(a0, offsetof(T, high)));
+ __ sd(a5, MemOperand(a0, offsetof(T, low)));
__ jr(ra);
__ nop();
@@ -480,30 +480,30 @@ TEST(MIPS5) {
Label L, C;
// Load all structure elements to registers.
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, a)) );
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(T, b)) );
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, i)) );
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, j)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(T, a)) );
+ __ ldc1(f6, MemOperand(a0, offsetof(T, b)) );
+ __ lw(a4, MemOperand(a0, offsetof(T, i)) );
+ __ lw(a5, MemOperand(a0, offsetof(T, j)) );
// Convert double in f4 to int in element i.
__ cvt_w_d(f8, f4);
__ mfc1(a6, f8);
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, i)) );
+ __ sw(a6, MemOperand(a0, offsetof(T, i)) );
// Convert double in f6 to int in element j.
__ cvt_w_d(f10, f6);
__ mfc1(a7, f10);
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, j)) );
+ __ sw(a7, MemOperand(a0, offsetof(T, j)) );
// Convert int in original i (a4) to double in a.
__ mtc1(a4, f12);
__ cvt_d_w(f0, f12);
- __ sdc1(f0, MemOperand(a0, OFFSET_OF(T, a)) );
+ __ sdc1(f0, MemOperand(a0, offsetof(T, a)) );
// Convert int in original j (a5) to double in b.
__ mtc1(a5, f14);
__ cvt_d_w(f2, f14);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(T, b)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(T, b)) );
__ jr(ra);
__ nop();
@@ -549,31 +549,31 @@ TEST(MIPS6) {
Label L, C;
// Basic word load/store.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, ui)) );
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, r1)) );
+ __ lw(a4, MemOperand(a0, offsetof(T, ui)) );
+ __ sw(a4, MemOperand(a0, offsetof(T, r1)) );
// lh with positive data.
- __ lh(a5, MemOperand(a0, OFFSET_OF(T, ui)) );
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, r2)) );
+ __ lh(a5, MemOperand(a0, offsetof(T, ui)) );
+ __ sw(a5, MemOperand(a0, offsetof(T, r2)) );
// lh with negative data.
- __ lh(a6, MemOperand(a0, OFFSET_OF(T, si)) );
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, r3)) );
+ __ lh(a6, MemOperand(a0, offsetof(T, si)) );
+ __ sw(a6, MemOperand(a0, offsetof(T, r3)) );
// lhu with negative data.
- __ lhu(a7, MemOperand(a0, OFFSET_OF(T, si)) );
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, r4)) );
+ __ lhu(a7, MemOperand(a0, offsetof(T, si)) );
+ __ sw(a7, MemOperand(a0, offsetof(T, r4)) );
// lb with negative data.
- __ lb(t0, MemOperand(a0, OFFSET_OF(T, si)) );
- __ sw(t0, MemOperand(a0, OFFSET_OF(T, r5)) );
+ __ lb(t0, MemOperand(a0, offsetof(T, si)) );
+ __ sw(t0, MemOperand(a0, offsetof(T, r5)) );
// sh writes only 1/2 of word.
__ lui(t1, 0x3333);
__ ori(t1, t1, 0x3333);
- __ sw(t1, MemOperand(a0, OFFSET_OF(T, r6)) );
- __ lhu(t1, MemOperand(a0, OFFSET_OF(T, si)) );
- __ sh(t1, MemOperand(a0, OFFSET_OF(T, r6)) );
+ __ sw(t1, MemOperand(a0, offsetof(T, r6)) );
+ __ lhu(t1, MemOperand(a0, offsetof(T, si)) );
+ __ sh(t1, MemOperand(a0, offsetof(T, r6)) );
__ jr(ra);
__ nop();
@@ -619,8 +619,8 @@ TEST(MIPS7) {
MacroAssembler assm(isolate, NULL, 0);
Label neither_is_nan, less_than, outa_here;
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, a)) );
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(T, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(T, a)) );
+ __ ldc1(f6, MemOperand(a0, offsetof(T, b)) );
if (kArchVariant != kMips64r6) {
__ c(UN, D, f4, f6);
__ bc1f(&neither_is_nan);
@@ -629,7 +629,7 @@ TEST(MIPS7) {
__ bc1eqz(&neither_is_nan, f2);
}
__ nop();
- __ sw(zero_reg, MemOperand(a0, OFFSET_OF(T, result)) );
+ __ sw(zero_reg, MemOperand(a0, offsetof(T, result)) );
__ Branch(&outa_here);
__ bind(&neither_is_nan);
@@ -643,12 +643,12 @@ TEST(MIPS7) {
}
__ nop();
- __ sw(zero_reg, MemOperand(a0, OFFSET_OF(T, result)) );
+ __ sw(zero_reg, MemOperand(a0, offsetof(T, result)) );
__ Branch(&outa_here);
__ bind(&less_than);
__ Addu(a4, zero_reg, Operand(1));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, result)) ); // Set true.
+ __ sw(a4, MemOperand(a0, offsetof(T, result)) ); // Set true.
// This test-case should have additional tests.
@@ -707,7 +707,7 @@ TEST(MIPS8) {
MacroAssembler assm(isolate, NULL, 0);
// Basic word load.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, input)) );
+ __ lw(a4, MemOperand(a0, offsetof(T, input)) );
// ROTR instruction (called through the Ror macro).
__ Ror(a5, a4, 0x0004);
@@ -719,13 +719,13 @@ TEST(MIPS8) {
__ Ror(t3, a4, 0x001c);
// Basic word store.
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, result_rotr_4)) );
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, result_rotr_8)) );
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, result_rotr_12)) );
- __ sw(t0, MemOperand(a0, OFFSET_OF(T, result_rotr_16)) );
- __ sw(t1, MemOperand(a0, OFFSET_OF(T, result_rotr_20)) );
- __ sw(t2, MemOperand(a0, OFFSET_OF(T, result_rotr_24)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(T, result_rotr_28)) );
+ __ sw(a5, MemOperand(a0, offsetof(T, result_rotr_4)) );
+ __ sw(a6, MemOperand(a0, offsetof(T, result_rotr_8)) );
+ __ sw(a7, MemOperand(a0, offsetof(T, result_rotr_12)) );
+ __ sw(t0, MemOperand(a0, offsetof(T, result_rotr_16)) );
+ __ sw(t1, MemOperand(a0, offsetof(T, result_rotr_20)) );
+ __ sw(t2, MemOperand(a0, offsetof(T, result_rotr_24)) );
+ __ sw(t3, MemOperand(a0, offsetof(T, result_rotr_28)) );
// ROTRV instruction (called through the Ror macro).
__ li(t3, 0x0004);
@@ -744,13 +744,13 @@ TEST(MIPS8) {
__ Ror(t3, a4, t3);
// Basic word store.
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, result_rotrv_4)) );
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, result_rotrv_8)) );
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, result_rotrv_12)) );
- __ sw(t0, MemOperand(a0, OFFSET_OF(T, result_rotrv_16)) );
- __ sw(t1, MemOperand(a0, OFFSET_OF(T, result_rotrv_20)) );
- __ sw(t2, MemOperand(a0, OFFSET_OF(T, result_rotrv_24)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(T, result_rotrv_28)) );
+ __ sw(a5, MemOperand(a0, offsetof(T, result_rotrv_4)) );
+ __ sw(a6, MemOperand(a0, offsetof(T, result_rotrv_8)) );
+ __ sw(a7, MemOperand(a0, offsetof(T, result_rotrv_12)) );
+ __ sw(t0, MemOperand(a0, offsetof(T, result_rotrv_16)) );
+ __ sw(t1, MemOperand(a0, offsetof(T, result_rotrv_20)) );
+ __ sw(t2, MemOperand(a0, offsetof(T, result_rotrv_24)) );
+ __ sw(t3, MemOperand(a0, offsetof(T, result_rotrv_28)) );
__ jr(ra);
__ nop();
@@ -838,42 +838,42 @@ TEST(MIPS10) {
// - 32 FP regs of 64-bits each, no odd/even pairs.
// - Note that cvt_l_d/cvt_d_l ARE legal in FR=1 mode.
// Load all structure elements to registers.
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, a)));
+ __ ldc1(f0, MemOperand(a0, offsetof(T, a)));
// Save the raw bits of the double.
__ mfc1(a4, f0);
__ mfhc1(a5, f0);
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, dbl_mant)));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, dbl_exp)));
+ __ sw(a4, MemOperand(a0, offsetof(T, dbl_mant)));
+ __ sw(a5, MemOperand(a0, offsetof(T, dbl_exp)));
// Convert double in f0 to long, save hi/lo parts.
__ cvt_l_d(f0, f0);
__ mfc1(a4, f0); // f0 LS 32 bits of long.
__ mfhc1(a5, f0); // f0 MS 32 bits of long.
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, long_lo)));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, long_hi)));
+ __ sw(a4, MemOperand(a0, offsetof(T, long_lo)));
+ __ sw(a5, MemOperand(a0, offsetof(T, long_hi)));
// Combine the high/low ints, convert back to double.
__ dsll32(a6, a5, 0); // Move a5 to high bits of a6.
__ or_(a6, a6, a4);
__ dmtc1(a6, f1);
__ cvt_d_l(f1, f1);
- __ sdc1(f1, MemOperand(a0, OFFSET_OF(T, a_converted)));
+ __ sdc1(f1, MemOperand(a0, offsetof(T, a_converted)));
// Convert the b long integers to double b.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, b_long_lo)));
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, b_long_hi)));
+ __ lw(a4, MemOperand(a0, offsetof(T, b_long_lo)));
+ __ lw(a5, MemOperand(a0, offsetof(T, b_long_hi)));
__ mtc1(a4, f8); // f8 LS 32-bits.
__ mthc1(a5, f8); // f8 MS 32-bits.
__ cvt_d_l(f10, f8);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(T, b)));
+ __ sdc1(f10, MemOperand(a0, offsetof(T, b)));
// Convert double b back to long-int.
- __ ldc1(f31, MemOperand(a0, OFFSET_OF(T, b)));
+ __ ldc1(f31, MemOperand(a0, offsetof(T, b)));
__ cvt_l_d(f31, f31);
__ dmfc1(a7, f31);
- __ sd(a7, MemOperand(a0, OFFSET_OF(T, b_long_as_int64)));
+ __ sd(a7, MemOperand(a0, offsetof(T, b_long_as_int64)));
__ jr(ra);
@@ -936,80 +936,80 @@ TEST(MIPS11) {
Assembler assm(isolate, NULL, 0);
// Test all combinations of LWL and vAddr.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwl(a4, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, lwl_0)));
+ __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwl(a4, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a4, MemOperand(a0, offsetof(T, lwl_0)));
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwl(a5, MemOperand(a0, OFFSET_OF(T, mem_init) + 1));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, lwl_1)));
+ __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwl(a5, MemOperand(a0, offsetof(T, mem_init) + 1));
+ __ sw(a5, MemOperand(a0, offsetof(T, lwl_1)));
- __ lw(a6, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwl(a6, MemOperand(a0, OFFSET_OF(T, mem_init) + 2));
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, lwl_2)));
+ __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwl(a6, MemOperand(a0, offsetof(T, mem_init) + 2));
+ __ sw(a6, MemOperand(a0, offsetof(T, lwl_2)));
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwl(a7, MemOperand(a0, OFFSET_OF(T, mem_init) + 3));
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, lwl_3)));
+ __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwl(a7, MemOperand(a0, offsetof(T, mem_init) + 3));
+ __ sw(a7, MemOperand(a0, offsetof(T, lwl_3)));
// Test all combinations of LWR and vAddr.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwr(a4, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, lwr_0)));
+ __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwr(a4, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a4, MemOperand(a0, offsetof(T, lwr_0)));
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwr(a5, MemOperand(a0, OFFSET_OF(T, mem_init) + 1));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, lwr_1)));
+ __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwr(a5, MemOperand(a0, offsetof(T, mem_init) + 1));
+ __ sw(a5, MemOperand(a0, offsetof(T, lwr_1)));
- __ lw(a6, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwr(a6, MemOperand(a0, OFFSET_OF(T, mem_init) + 2));
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, lwr_2)) );
+ __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwr(a6, MemOperand(a0, offsetof(T, mem_init) + 2));
+ __ sw(a6, MemOperand(a0, offsetof(T, lwr_2)) );
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ lwr(a7, MemOperand(a0, OFFSET_OF(T, mem_init) + 3));
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, lwr_3)) );
+ __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
+ __ lwr(a7, MemOperand(a0, offsetof(T, mem_init) + 3));
+ __ sw(a7, MemOperand(a0, offsetof(T, lwr_3)) );
// Test all combinations of SWL and vAddr.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, swl_0)));
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swl(a4, MemOperand(a0, OFFSET_OF(T, swl_0)));
-
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, swl_1)));
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swl(a5, MemOperand(a0, OFFSET_OF(T, swl_1) + 1));
-
- __ lw(a6, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, swl_2)));
- __ lw(a6, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swl(a6, MemOperand(a0, OFFSET_OF(T, swl_2) + 2));
-
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, swl_3)));
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swl(a7, MemOperand(a0, OFFSET_OF(T, swl_3) + 3));
+ __ lw(a4, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a4, MemOperand(a0, offsetof(T, swl_0)));
+ __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
+ __ swl(a4, MemOperand(a0, offsetof(T, swl_0)));
+
+ __ lw(a5, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a5, MemOperand(a0, offsetof(T, swl_1)));
+ __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
+ __ swl(a5, MemOperand(a0, offsetof(T, swl_1) + 1));
+
+ __ lw(a6, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a6, MemOperand(a0, offsetof(T, swl_2)));
+ __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
+ __ swl(a6, MemOperand(a0, offsetof(T, swl_2) + 2));
+
+ __ lw(a7, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a7, MemOperand(a0, offsetof(T, swl_3)));
+ __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
+ __ swl(a7, MemOperand(a0, offsetof(T, swl_3) + 3));
// Test all combinations of SWR and vAddr.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, swr_0)));
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swr(a4, MemOperand(a0, OFFSET_OF(T, swr_0)));
-
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, swr_1)));
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swr(a5, MemOperand(a0, OFFSET_OF(T, swr_1) + 1));
-
- __ lw(a6, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, swr_2)));
- __ lw(a6, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swr(a6, MemOperand(a0, OFFSET_OF(T, swr_2) + 2));
-
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, mem_init)));
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, swr_3)));
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, reg_init)));
- __ swr(a7, MemOperand(a0, OFFSET_OF(T, swr_3) + 3));
+ __ lw(a4, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a4, MemOperand(a0, offsetof(T, swr_0)));
+ __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
+ __ swr(a4, MemOperand(a0, offsetof(T, swr_0)));
+
+ __ lw(a5, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a5, MemOperand(a0, offsetof(T, swr_1)));
+ __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
+ __ swr(a5, MemOperand(a0, offsetof(T, swr_1) + 1));
+
+ __ lw(a6, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a6, MemOperand(a0, offsetof(T, swr_2)));
+ __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
+ __ swr(a6, MemOperand(a0, offsetof(T, swr_2) + 2));
+
+ __ lw(a7, MemOperand(a0, offsetof(T, mem_init)));
+ __ sw(a7, MemOperand(a0, offsetof(T, swr_3)));
+ __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
+ __ swr(a7, MemOperand(a0, offsetof(T, swr_3) + 3));
__ jr(ra);
__ nop();
@@ -1067,8 +1067,8 @@ TEST(MIPS12) {
__ mov(t2, fp); // Save frame pointer.
__ mov(fp, a0); // Access struct T by fp.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, y)));
- __ lw(a7, MemOperand(a0, OFFSET_OF(T, y4)));
+ __ lw(a4, MemOperand(a0, offsetof(T, y)));
+ __ lw(a7, MemOperand(a0, offsetof(T, y4)));
__ addu(a5, a4, a7);
__ subu(t0, a4, a7);
@@ -1086,30 +1086,30 @@ TEST(MIPS12) {
__ push(a7);
__ pop(t0);
__ nop();
- __ sw(a4, MemOperand(fp, OFFSET_OF(T, y)));
- __ lw(a4, MemOperand(fp, OFFSET_OF(T, y)));
+ __ sw(a4, MemOperand(fp, offsetof(T, y)));
+ __ lw(a4, MemOperand(fp, offsetof(T, y)));
__ nop();
- __ sw(a4, MemOperand(fp, OFFSET_OF(T, y)));
- __ lw(a5, MemOperand(fp, OFFSET_OF(T, y)));
+ __ sw(a4, MemOperand(fp, offsetof(T, y)));
+ __ lw(a5, MemOperand(fp, offsetof(T, y)));
__ nop();
__ push(a5);
- __ lw(a5, MemOperand(fp, OFFSET_OF(T, y)));
+ __ lw(a5, MemOperand(fp, offsetof(T, y)));
__ pop(a5);
__ nop();
__ push(a5);
- __ lw(a6, MemOperand(fp, OFFSET_OF(T, y)));
+ __ lw(a6, MemOperand(fp, offsetof(T, y)));
__ pop(a5);
__ nop();
__ push(a5);
- __ lw(a6, MemOperand(fp, OFFSET_OF(T, y)));
+ __ lw(a6, MemOperand(fp, offsetof(T, y)));
__ pop(a6);
__ nop();
__ push(a6);
- __ lw(a6, MemOperand(fp, OFFSET_OF(T, y)));
+ __ lw(a6, MemOperand(fp, offsetof(T, y)));
__ pop(a5);
__ nop();
__ push(a5);
- __ lw(a6, MemOperand(fp, OFFSET_OF(T, y)));
+ __ lw(a6, MemOperand(fp, offsetof(T, y)));
__ pop(a7);
__ nop();
@@ -1154,19 +1154,19 @@ TEST(MIPS13) {
MacroAssembler assm(isolate, NULL, 0);
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, cvt_small_in)));
+ __ sw(a4, MemOperand(a0, offsetof(T, cvt_small_in)));
__ Cvt_d_uw(f10, a4, f22);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(T, cvt_small_out)));
+ __ sdc1(f10, MemOperand(a0, offsetof(T, cvt_small_out)));
__ Trunc_uw_d(f10, f10, f22);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(T, trunc_small_out)));
+ __ swc1(f10, MemOperand(a0, offsetof(T, trunc_small_out)));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, cvt_big_in)));
+ __ sw(a4, MemOperand(a0, offsetof(T, cvt_big_in)));
__ Cvt_d_uw(f8, a4, f22);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(T, cvt_big_out)));
+ __ sdc1(f8, MemOperand(a0, offsetof(T, cvt_big_out)));
__ Trunc_uw_d(f8, f8, f22);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(T, trunc_big_out)));
+ __ swc1(f8, MemOperand(a0, offsetof(T, trunc_big_out)));
__ jr(ra);
__ nop();
@@ -1236,46 +1236,46 @@ TEST(MIPS14) {
// Disable FPU exceptions.
__ ctc1(zero_reg, FCSR);
#define RUN_ROUND_TEST(x) \
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, round_up_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, round_up_in))); \
__ x##_w_d(f0, f0); \
- __ swc1(f0, MemOperand(a0, OFFSET_OF(T, x##_up_out))); \
+ __ swc1(f0, MemOperand(a0, offsetof(T, x##_up_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, round_down_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, round_down_in))); \
__ x##_w_d(f0, f0); \
- __ swc1(f0, MemOperand(a0, OFFSET_OF(T, x##_down_out))); \
+ __ swc1(f0, MemOperand(a0, offsetof(T, x##_down_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, neg_round_up_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, neg_round_up_in))); \
__ x##_w_d(f0, f0); \
- __ swc1(f0, MemOperand(a0, OFFSET_OF(T, neg_##x##_up_out))); \
+ __ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_up_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, neg_round_down_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, neg_round_down_in))); \
__ x##_w_d(f0, f0); \
- __ swc1(f0, MemOperand(a0, OFFSET_OF(T, neg_##x##_down_out))); \
+ __ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_down_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, err1_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, err1_in))); \
__ ctc1(zero_reg, FCSR); \
__ x##_w_d(f0, f0); \
__ cfc1(a2, FCSR); \
- __ sw(a2, MemOperand(a0, OFFSET_OF(T, x##_err1_out))); \
+ __ sw(a2, MemOperand(a0, offsetof(T, x##_err1_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, err2_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, err2_in))); \
__ ctc1(zero_reg, FCSR); \
__ x##_w_d(f0, f0); \
__ cfc1(a2, FCSR); \
- __ sw(a2, MemOperand(a0, OFFSET_OF(T, x##_err2_out))); \
+ __ sw(a2, MemOperand(a0, offsetof(T, x##_err2_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, err3_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, err3_in))); \
__ ctc1(zero_reg, FCSR); \
__ x##_w_d(f0, f0); \
__ cfc1(a2, FCSR); \
- __ sw(a2, MemOperand(a0, OFFSET_OF(T, x##_err3_out))); \
+ __ sw(a2, MemOperand(a0, offsetof(T, x##_err3_out))); \
\
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(T, err4_in))); \
+ __ ldc1(f0, MemOperand(a0, offsetof(T, err4_in))); \
__ ctc1(zero_reg, FCSR); \
__ x##_w_d(f0, f0); \
__ cfc1(a2, FCSR); \
- __ sw(a2, MemOperand(a0, OFFSET_OF(T, x##_err4_out))); \
- __ swc1(f0, MemOperand(a0, OFFSET_OF(T, x##_invalid_result)));
+ __ sw(a2, MemOperand(a0, offsetof(T, x##_err4_out))); \
+ __ swc1(f0, MemOperand(a0, offsetof(T, x##_invalid_result)));
RUN_ROUND_TEST(round)
RUN_ROUND_TEST(floor)
@@ -1363,48 +1363,48 @@ TEST(MIPS16) {
Label L, C;
// Basic 32-bit word load/store, with un-signed data.
- __ lw(a4, MemOperand(a0, OFFSET_OF(T, ui)));
- __ sw(a4, MemOperand(a0, OFFSET_OF(T, r1)));
+ __ lw(a4, MemOperand(a0, offsetof(T, ui)));
+ __ sw(a4, MemOperand(a0, offsetof(T, r1)));
// Check that the data got zero-extended into 64-bit a4.
- __ sd(a4, MemOperand(a0, OFFSET_OF(T, r2)));
+ __ sd(a4, MemOperand(a0, offsetof(T, r2)));
// Basic 32-bit word load/store, with SIGNED data.
- __ lw(a5, MemOperand(a0, OFFSET_OF(T, si)));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, r3)));
+ __ lw(a5, MemOperand(a0, offsetof(T, si)));
+ __ sw(a5, MemOperand(a0, offsetof(T, r3)));
// Check that the data got sign-extended into 64-bit a4.
- __ sd(a5, MemOperand(a0, OFFSET_OF(T, r4)));
+ __ sd(a5, MemOperand(a0, offsetof(T, r4)));
// 32-bit UNSIGNED word load/store, with SIGNED data.
- __ lwu(a6, MemOperand(a0, OFFSET_OF(T, si)));
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, r5)));
+ __ lwu(a6, MemOperand(a0, offsetof(T, si)));
+ __ sw(a6, MemOperand(a0, offsetof(T, r5)));
// Check that the data got zero-extended into 64-bit a4.
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r6)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r6)));
// lh with positive data.
- __ lh(a5, MemOperand(a0, OFFSET_OF(T, ui)));
- __ sw(a5, MemOperand(a0, OFFSET_OF(T, r2)));
+ __ lh(a5, MemOperand(a0, offsetof(T, ui)));
+ __ sw(a5, MemOperand(a0, offsetof(T, r2)));
// lh with negative data.
- __ lh(a6, MemOperand(a0, OFFSET_OF(T, si)));
- __ sw(a6, MemOperand(a0, OFFSET_OF(T, r3)));
+ __ lh(a6, MemOperand(a0, offsetof(T, si)));
+ __ sw(a6, MemOperand(a0, offsetof(T, r3)));
// lhu with negative data.
- __ lhu(a7, MemOperand(a0, OFFSET_OF(T, si)));
- __ sw(a7, MemOperand(a0, OFFSET_OF(T, r4)));
+ __ lhu(a7, MemOperand(a0, offsetof(T, si)));
+ __ sw(a7, MemOperand(a0, offsetof(T, r4)));
// lb with negative data.
- __ lb(t0, MemOperand(a0, OFFSET_OF(T, si)));
- __ sw(t0, MemOperand(a0, OFFSET_OF(T, r5)));
+ __ lb(t0, MemOperand(a0, offsetof(T, si)));
+ __ sw(t0, MemOperand(a0, offsetof(T, r5)));
// // sh writes only 1/2 of word.
__ lui(t1, 0x3333);
__ ori(t1, t1, 0x3333);
- __ sw(t1, MemOperand(a0, OFFSET_OF(T, r6)));
- __ lhu(t1, MemOperand(a0, OFFSET_OF(T, si)));
- __ sh(t1, MemOperand(a0, OFFSET_OF(T, r6)));
+ __ sw(t1, MemOperand(a0, offsetof(T, r6)));
+ __ lhu(t1, MemOperand(a0, offsetof(T, si)));
+ __ sh(t1, MemOperand(a0, offsetof(T, r6)));
__ jr(ra);
__ nop();
@@ -1466,26 +1466,26 @@ TEST(seleqz_selnez) {
// Integer part of test.
__ addiu(t1, zero_reg, 1); // t1 = 1
__ seleqz(t3, t1, zero_reg); // t3 = 1
- __ sw(t3, MemOperand(a0, OFFSET_OF(Test, a))); // a = 1
+ __ sw(t3, MemOperand(a0, offsetof(Test, a))); // a = 1
__ seleqz(t2, t1, t1); // t2 = 0
- __ sw(t2, MemOperand(a0, OFFSET_OF(Test, b))); // b = 0
+ __ sw(t2, MemOperand(a0, offsetof(Test, b))); // b = 0
__ selnez(t3, t1, zero_reg); // t3 = 1;
- __ sw(t3, MemOperand(a0, OFFSET_OF(Test, c))); // c = 0
+ __ sw(t3, MemOperand(a0, offsetof(Test, c))); // c = 0
__ selnez(t3, t1, t1); // t3 = 1
- __ sw(t3, MemOperand(a0, OFFSET_OF(Test, d))); // d = 1
+ __ sw(t3, MemOperand(a0, offsetof(Test, d))); // d = 1
// Floating point part of test.
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(Test, e)) ); // src
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(Test, f)) ); // test
- __ lwc1(f8, MemOperand(a0, OFFSET_OF(Test, i)) ); // src
- __ lwc1(f10, MemOperand(a0, OFFSET_OF(Test, j)) ); // test
+ __ ldc1(f0, MemOperand(a0, offsetof(Test, e)) ); // src
+ __ ldc1(f2, MemOperand(a0, offsetof(Test, f)) ); // test
+ __ lwc1(f8, MemOperand(a0, offsetof(Test, i)) ); // src
+ __ lwc1(f10, MemOperand(a0, offsetof(Test, j)) ); // test
__ seleqz_d(f4, f0, f2);
__ selnez_d(f6, f0, f2);
__ seleqz_s(f12, f8, f10);
__ selnez_s(f14, f8, f10);
- __ sdc1(f4, MemOperand(a0, OFFSET_OF(Test, g)) ); // src
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(Test, h)) ); // src
- __ swc1(f12, MemOperand(a0, OFFSET_OF(Test, k)) ); // src
- __ swc1(f14, MemOperand(a0, OFFSET_OF(Test, l)) ); // src
+ __ sdc1(f4, MemOperand(a0, offsetof(Test, g)) ); // src
+ __ sdc1(f6, MemOperand(a0, offsetof(Test, h)) ); // src
+ __ swc1(f12, MemOperand(a0, offsetof(Test, k)) ); // src
+ __ swc1(f14, MemOperand(a0, offsetof(Test, l)) ); // src
__ jr(ra);
__ nop();
CodeDesc desc;
@@ -1574,18 +1574,18 @@ TEST(min_max) {
float outputsfmin[tableLength] = {2.0, 2.0, 3.0, 3.0, fltNaN};
float outputsfmax[tableLength] = {3.0, 3.0, 3.0, 3.0, fltNaN};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, a)));
- __ ldc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, b)));
- __ lwc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, e)));
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, f)));
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
+ __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
+ __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, e)));
+ __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, f)));
__ min_d(f10, f4, f8);
__ max_d(f12, f4, f8);
__ min_s(f14, f2, f6);
__ max_s(f16, f2, f6);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, c)));
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, d)));
- __ swc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, g)));
- __ swc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, h)));
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, d)));
+ __ swc1(f14, MemOperand(a0, offsetof(TestFloat, g)));
+ __ swc1(f16, MemOperand(a0, offsetof(TestFloat, h)));
__ jr(ra);
__ nop();
@@ -1696,11 +1696,11 @@ TEST(rint_d) {
int fcsr_inputs[4] =
{kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ lw(t0, MemOperand(a0, OFFSET_OF(TestFloat, fcsr)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) );
__ ctc1(t0, FCSR);
__ rint_d(f8, f4);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
+ __ sdc1(f8, MemOperand(a0, offsetof(TestFloat, b)) );
__ jr(ra);
__ nop();
@@ -1739,16 +1739,16 @@ TEST(sel) {
} Test;
Test test;
- __ ldc1(f0, MemOperand(a0, OFFSET_OF(Test, dd)) ); // test
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(Test, ds)) ); // src1
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, dt)) ); // src2
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, fd)) ); // test
- __ lwc1(f8, MemOperand(a0, OFFSET_OF(Test, fs)) ); // src1
- __ lwc1(f10, MemOperand(a0, OFFSET_OF(Test, ft)) ); // src2
+ __ ldc1(f0, MemOperand(a0, offsetof(Test, dd)) ); // test
+ __ ldc1(f2, MemOperand(a0, offsetof(Test, ds)) ); // src1
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, dt)) ); // src2
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, fd)) ); // test
+ __ lwc1(f8, MemOperand(a0, offsetof(Test, fs)) ); // src1
+ __ lwc1(f10, MemOperand(a0, offsetof(Test, ft)) ); // src2
__ sel_d(f0, f2, f4);
__ sel_s(f6, f8, f10);
- __ sdc1(f0, MemOperand(a0, OFFSET_OF(Test, dd)) );
- __ swc1(f6, MemOperand(a0, OFFSET_OF(Test, fd)) );
+ __ sdc1(f0, MemOperand(a0, offsetof(Test, dd)) );
+ __ swc1(f6, MemOperand(a0, offsetof(Test, fd)) );
__ jr(ra);
__ nop();
CodeDesc desc;
@@ -1873,12 +1873,12 @@ TEST(rint_s) {
int fcsr_inputs[4] =
{kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
float* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ lw(t0, MemOperand(a0, OFFSET_OF(TestFloat, fcsr)) );
+ __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) );
__ cfc1(t1, FCSR);
__ ctc1(t0, FCSR);
__ rint_s(f8, f4);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
+ __ swc1(f8, MemOperand(a0, offsetof(TestFloat, b)) );
__ ctc1(t1, FCSR);
__ jr(ra);
__ nop();
@@ -1970,18 +1970,18 @@ TEST(mina_maxa) {
-10.0, -11.2, -9.8
};
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
- __ lwc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
- __ lwc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, d)) );
+ __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
+ __ lwc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
+ __ lwc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
__ mina_d(f6, f2, f4);
__ mina_s(f12, f8, f10);
__ maxa_d(f14, f2, f4);
__ maxa_s(f16, f8, f10);
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, resf)) );
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, resd)) );
- __ swc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, resf1)) );
- __ sdc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, resd1)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, resf)) );
+ __ sdc1(f6, MemOperand(a0, offsetof(TestFloat, resd)) );
+ __ swc1(f16, MemOperand(a0, offsetof(TestFloat, resf1)) );
+ __ sdc1(f14, MemOperand(a0, offsetof(TestFloat, resd1)) );
__ jr(ra);
__ nop();
@@ -2042,12 +2042,12 @@ TEST(trunc_l) {
2147483648.0, dFPU64InvalidResult,
dFPU64InvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ trunc_l_d(f8, f4);
__ trunc_l_s(f10, f6);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -2104,26 +2104,26 @@ TEST(movz_movn) {
5.3, -5.3, 5.3, -2.9
};
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
- __ lw(t0, MemOperand(a0, OFFSET_OF(TestFloat, rt)) );
+ __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
+ __ lw(t0, MemOperand(a0, offsetof(TestFloat, rt)) );
__ li(t1, 0x0);
__ mtc1(t1, f12);
__ mtc1(t1, f10);
__ mtc1(t1, f16);
__ mtc1(t1, f14);
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, bold)) );
- __ swc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, dold)) );
- __ sdc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, bold1)) );
- __ swc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, dold1)) );
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, bold)) );
+ __ swc1(f10, MemOperand(a0, offsetof(TestFloat, dold)) );
+ __ sdc1(f16, MemOperand(a0, offsetof(TestFloat, bold1)) );
+ __ swc1(f14, MemOperand(a0, offsetof(TestFloat, dold1)) );
__ movz_s(f10, f6, t0);
__ movz_d(f12, f2, t0);
__ movn_s(f14, f6, t0);
__ movn_d(f16, f2, t0);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, d)) );
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
- __ swc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, d1)) );
- __ sdc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, b1)) );
+ __ swc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, b)) );
+ __ swc1(f14, MemOperand(a0, offsetof(TestFloat, d1)) );
+ __ sdc1(f16, MemOperand(a0, offsetof(TestFloat, b1)) );
__ jr(ra);
__ nop();
@@ -2203,26 +2203,26 @@ TEST(movt_movd) {
}
HandleScope scope(isolate);
MacroAssembler assm(isolate, NULL, 0);
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, srcd)) );
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, srcf)) );
- __ lw(t1, MemOperand(a0, OFFSET_OF(TestFloat, fcsr)) );
+ __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, srcd)) );
+ __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, srcf)) );
+ __ lw(t1, MemOperand(a0, offsetof(TestFloat, fcsr)) );
__ cfc1(t0, FCSR);
__ ctc1(t1, FCSR);
__ li(t2, 0x0);
__ mtc1(t2, f12);
__ mtc1(t2, f10);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, dstdold)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, dstfold)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold)) );
__ movt_s(f12, f4, test.cc);
__ movt_d(f10, f2, test.cc);
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, dstf)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, dstd)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, dstdold1)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, dstfold1)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold1)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold1)) );
__ movf_s(f12, f4, test.cc);
__ movf_d(f10, f2, test.cc);
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, dstf1)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, dstd1)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf1)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd1)) );
__ ctc1(t0, FCSR);
__ jr(ra);
__ nop();
@@ -2303,12 +2303,12 @@ TEST(cvt_w_d) {
int fcsr_inputs[4] =
{kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lw(t0, MemOperand(a0, OFFSET_OF(Test, fcsr)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lw(t0, MemOperand(a0, offsetof(Test, fcsr)) );
__ cfc1(t1, FCSR);
__ ctc1(t0, FCSR);
__ cvt_w_d(f8, f4);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ swc1(f8, MemOperand(a0, offsetof(Test, b)) );
__ ctc1(t1, FCSR);
__ jr(ra);
__ nop();
@@ -2362,12 +2362,12 @@ TEST(trunc_w) {
kFPUInvalidResult, kFPUInvalidResult,
kFPUInvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ trunc_w_d(f8, f4);
__ trunc_w_s(f10, f6);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ swc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -2419,12 +2419,12 @@ TEST(round_w) {
kFPUInvalidResult, kFPUInvalidResult,
kFPUInvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ round_w_d(f8, f4);
__ round_w_s(f10, f6);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ swc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -2476,12 +2476,12 @@ TEST(round_l) {
2147483648.0, dFPU64InvalidResult,
dFPU64InvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ round_l_d(f8, f4);
__ round_l_s(f10, f6);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -2542,14 +2542,14 @@ TEST(sub) {
0.5, -0.5, 0.0, -10.1, -10.1, -5.8,
10.1, 10.1, 5.8, -0.5, 0.5, 0.0
};
- __ lwc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
- __ ldc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
- __ ldc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, d)) );
+ __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
+ __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
+ __ ldc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
__ sub_s(f6, f2, f4);
__ sub_d(f12, f8, f10);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, resultS)) );
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, resultD)) );
+ __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
__ jr(ra);
__ nop();
@@ -2609,20 +2609,20 @@ TEST(sqrt_rsqrt_recip) {
};
- __ lwc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ ldc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
+ __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
__ sqrt_s(f6, f2);
__ sqrt_d(f12, f8);
__ rsqrt_d(f14, f8);
__ rsqrt_s(f16, f2);
__ recip_d(f18, f8);
__ recip_s(f20, f2);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, resultS)) );
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, resultD)) );
- __ swc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, resultS1)) );
- __ sdc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, resultD1)) );
- __ swc1(f20, MemOperand(a0, OFFSET_OF(TestFloat, resultS2)) );
- __ sdc1(f18, MemOperand(a0, OFFSET_OF(TestFloat, resultD2)) );
+ __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
+ __ swc1(f16, MemOperand(a0, offsetof(TestFloat, resultS1)) );
+ __ sdc1(f14, MemOperand(a0, offsetof(TestFloat, resultD1)) );
+ __ swc1(f20, MemOperand(a0, offsetof(TestFloat, resultS2)) );
+ __ sdc1(f18, MemOperand(a0, offsetof(TestFloat, resultD2)) );
__ jr(ra);
__ nop();
@@ -2695,12 +2695,12 @@ TEST(neg) {
float outputs_S[tableLength] = {
-4.0, 2.0
};
- __ lwc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ ldc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
+ __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
__ neg_s(f6, f2);
__ neg_d(f12, f8);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, resultS)) );
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, resultD)) );
+ __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
__ jr(ra);
__ nop();
@@ -2751,14 +2751,14 @@ TEST(mul) {
4.8, 4.8, -4.8, -0.29
};
- __ lwc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
- __ ldc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, d)) );
+ __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
+ __ ldc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
+ __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, d)) );
__ mul_s(f10, f2, f4);
__ mul_d(f12, f6, f8);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, resultS)) );
- __ sdc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, resultD)) );
+ __ swc1(f10, MemOperand(a0, offsetof(TestFloat, resultS)) );
+ __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
__ jr(ra);
__ nop();
@@ -2808,12 +2808,12 @@ TEST(mov) {
5.3, -5.3, 5.3, -2.9
};
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, c)) );
+ __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
__ mov_s(f18, f6);
__ mov_d(f20, f2);
- __ swc1(f18, MemOperand(a0, OFFSET_OF(TestFloat, d)) );
- __ sdc1(f20, MemOperand(a0, OFFSET_OF(TestFloat, b)) );
+ __ swc1(f18, MemOperand(a0, offsetof(TestFloat, d)) );
+ __ sdc1(f20, MemOperand(a0, offsetof(TestFloat, b)) );
__ jr(ra);
__ nop();
@@ -2866,12 +2866,12 @@ TEST(floor_w) {
kFPUInvalidResult, kFPUInvalidResult,
kFPUInvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ floor_w_d(f8, f4);
__ floor_w_s(f10, f6);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ swc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -2923,12 +2923,12 @@ TEST(floor_l) {
2147483648.0, dFPU64InvalidResult,
dFPU64InvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ floor_l_d(f8, f4);
__ floor_l_s(f10, f6);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -2980,12 +2980,12 @@ TEST(ceil_w) {
kFPUInvalidResult, kFPUInvalidResult,
kFPUInvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ ceil_w_d(f8, f4);
__ ceil_w_s(f10, f6);
- __ swc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ swc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -3037,12 +3037,12 @@ TEST(ceil_l) {
2147483648.0, dFPU64InvalidResult,
dFPU64InvalidResult};
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, a)) );
- __ lwc1(f6, MemOperand(a0, OFFSET_OF(Test, b)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
+ __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
__ ceil_l_d(f8, f4);
__ ceil_l_s(f10, f6);
- __ sdc1(f8, MemOperand(a0, OFFSET_OF(Test, c)) );
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(Test, d)) );
+ __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
+ __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
__ jr(ra);
__ nop();
Test test;
@@ -3305,35 +3305,35 @@ TEST(BITSWAP) {
Assembler assm(isolate, NULL, 0);
- __ ld(a4, MemOperand(a0, OFFSET_OF(T, r1)));
+ __ ld(a4, MemOperand(a0, offsetof(T, r1)));
__ nop();
__ bitswap(a6, a4);
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r1)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r1)));
- __ ld(a4, MemOperand(a0, OFFSET_OF(T, r2)));
+ __ ld(a4, MemOperand(a0, offsetof(T, r2)));
__ nop();
__ bitswap(a6, a4);
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r2)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r2)));
- __ ld(a4, MemOperand(a0, OFFSET_OF(T, r3)));
+ __ ld(a4, MemOperand(a0, offsetof(T, r3)));
__ nop();
__ bitswap(a6, a4);
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r3)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r3)));
- __ ld(a4, MemOperand(a0, OFFSET_OF(T, r4)));
+ __ ld(a4, MemOperand(a0, offsetof(T, r4)));
__ nop();
__ bitswap(a6, a4);
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r4)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r4)));
- __ ld(a4, MemOperand(a0, OFFSET_OF(T, r5)));
+ __ ld(a4, MemOperand(a0, offsetof(T, r5)));
__ nop();
__ dbitswap(a6, a4);
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r5)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r5)));
- __ ld(a4, MemOperand(a0, OFFSET_OF(T, r6)));
+ __ ld(a4, MemOperand(a0, offsetof(T, r6)));
__ nop();
__ dbitswap(a6, a4);
- __ sd(a6, MemOperand(a0, OFFSET_OF(T, r6)));
+ __ sd(a6, MemOperand(a0, offsetof(T, r6)));
__ jr(ra);
__ nop();
@@ -3396,86 +3396,86 @@ TEST(class_fmt) {
// the doubles t.a ... t.f.
MacroAssembler assm(isolate, NULL, 0);
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dSignalingNan)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dSignalingNan)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dSignalingNan)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dSignalingNan)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dQuietNan)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dQuietNan)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dQuietNan)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dQuietNan)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dNegInf)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dNegInf)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dNegInf)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dNegInf)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dNegNorm)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dNegNorm)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dNegNorm)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dNegNorm)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dNegSubnorm)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dNegSubnorm)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dNegSubnorm)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dNegSubnorm)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dNegZero)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dNegZero)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dNegZero)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dNegZero)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dPosInf)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dPosInf)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dPosInf)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dPosInf)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dPosNorm)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dPosNorm)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dPosNorm)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dPosNorm)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dPosSubnorm)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dPosSubnorm)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dPosSubnorm)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dPosSubnorm)));
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(T, dPosZero)));
+ __ ldc1(f4, MemOperand(a0, offsetof(T, dPosZero)));
__ class_d(f6, f4);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(T, dPosZero)));
+ __ sdc1(f6, MemOperand(a0, offsetof(T, dPosZero)));
// Testing instruction CLASS.S
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fSignalingNan)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fSignalingNan)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fSignalingNan)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fSignalingNan)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fQuietNan)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fQuietNan)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fQuietNan)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fQuietNan)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fNegInf)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fNegInf)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fNegInf)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fNegInf)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fNegNorm)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fNegNorm)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fNegNorm)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fNegNorm)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fNegSubnorm)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fNegSubnorm)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fNegSubnorm)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fNegSubnorm)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fNegZero)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fNegZero)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fNegZero)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fNegZero)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fPosInf)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fPosInf)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fPosInf)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fPosInf)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fPosNorm)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fPosNorm)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fPosNorm)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fPosNorm)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fPosSubnorm)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fPosSubnorm)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fPosSubnorm)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fPosSubnorm)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(T, fPosZero)));
+ __ lwc1(f4, MemOperand(a0, offsetof(T, fPosZero)));
__ class_s(f6, f4);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(T, fPosZero)));
+ __ swc1(f6, MemOperand(a0, offsetof(T, fPosZero)));
__ jr(ra);
__ nop();
@@ -3552,17 +3552,17 @@ TEST(ABS) {
// Save FIR.
__ cfc1(a1, FCSR);
- __ sd(a1, MemOperand(a0, OFFSET_OF(TestFloat, fcsr)));
+ __ sd(a1, MemOperand(a0, offsetof(TestFloat, fcsr)));
// Disable FPU exceptions.
__ ctc1(zero_reg, FCSR);
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, a)));
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
__ abs_d(f10, f4);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, a)));
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, a)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, b)));
+ __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
__ abs_s(f10, f4);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, b)));
+ __ swc1(f10, MemOperand(a0, offsetof(TestFloat, b)));
// Restore FCSR.
__ ctc1(a1, FCSR);
@@ -3650,15 +3650,15 @@ TEST(ADD_FMT) {
TestFloat test;
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, a)));
- __ ldc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, b)));
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
+ __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
__ add_d(f10, f8, f4);
- __ sdc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, c)));
+ __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, fa)));
- __ lwc1(f8, MemOperand(a0, OFFSET_OF(TestFloat, fb)));
+ __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, fa)));
+ __ lwc1(f8, MemOperand(a0, offsetof(TestFloat, fb)));
__ add_s(f10, f8, f4);
- __ swc1(f10, MemOperand(a0, OFFSET_OF(TestFloat, fc)));
+ __ swc1(f10, MemOperand(a0, offsetof(TestFloat, fc)));
__ jr(ra);
__ nop();
@@ -3736,11 +3736,11 @@ TEST(C_COND_FMT) {
__ li(t1, 1);
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, dOp1)));
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, dOp2)));
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
+ __ ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
- __ lwc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, fOp1)));
- __ lwc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, fOp2)));
+ __ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
+ __ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3748,8 +3748,8 @@ TEST(C_COND_FMT) {
__ c_s(F, f14, f16, 2);
__ movt(t2, t1, 0);
__ movt(t3, t1, 2);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dF)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fF)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dF)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fF)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3757,8 +3757,8 @@ TEST(C_COND_FMT) {
__ c_s(UN, f14, f16, 4);
__ movt(t2, t1, 2);
__ movt(t3, t1, 4);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dUn)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fUn)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUn)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUn)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3766,8 +3766,8 @@ TEST(C_COND_FMT) {
__ c_s(EQ, f14, f16, 6);
__ movt(t2, t1, 4);
__ movt(t3, t1, 6);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dEq)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fEq)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dEq)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fEq)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3775,8 +3775,8 @@ TEST(C_COND_FMT) {
__ c_s(UEQ, f14, f16, 0);
__ movt(t2, t1, 6);
__ movt(t3, t1, 0);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dUeq)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fUeq)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUeq)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUeq)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3784,8 +3784,8 @@ TEST(C_COND_FMT) {
__ c_s(OLT, f14, f16, 2);
__ movt(t2, t1, 0);
__ movt(t3, t1, 2);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dOlt)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fOlt)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dOlt)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fOlt)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3793,8 +3793,8 @@ TEST(C_COND_FMT) {
__ c_s(ULT, f14, f16, 4);
__ movt(t2, t1, 2);
__ movt(t3, t1, 4);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dUlt)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fUlt)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUlt)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUlt)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3802,8 +3802,8 @@ TEST(C_COND_FMT) {
__ c_s(OLE, f14, f16, 6);
__ movt(t2, t1, 4);
__ movt(t3, t1, 6);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dOle)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fOle)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dOle)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fOle)) );
__ mov(t2, zero_reg);
__ mov(t3, zero_reg);
@@ -3811,8 +3811,8 @@ TEST(C_COND_FMT) {
__ c_s(ULE, f14, f16, 0);
__ movt(t2, t1, 6);
__ movt(t3, t1, 0);
- __ sw(t2, MemOperand(a0, OFFSET_OF(TestFloat, dUle)) );
- __ sw(t3, MemOperand(a0, OFFSET_OF(TestFloat, fUle)) );
+ __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUle)) );
+ __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUle)) );
__ jr(ra);
__ nop();
@@ -3953,66 +3953,66 @@ TEST(CMP_COND_FMT) {
__ li(t1, 1);
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(TestFloat, dOp1)));
- __ ldc1(f6, MemOperand(a0, OFFSET_OF(TestFloat, dOp2)));
+ __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
+ __ ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
- __ lwc1(f14, MemOperand(a0, OFFSET_OF(TestFloat, fOp1)));
- __ lwc1(f16, MemOperand(a0, OFFSET_OF(TestFloat, fOp2)));
+ __ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
+ __ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
__ cmp_d(F, f2, f4, f6);
__ cmp_s(F, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dF)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fF)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dF)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fF)) );
__ cmp_d(UN, f2, f4, f6);
__ cmp_s(UN, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dUn)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fUn)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUn)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUn)) );
__ cmp_d(EQ, f2, f4, f6);
__ cmp_s(EQ, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dEq)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fEq)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dEq)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fEq)) );
__ cmp_d(UEQ, f2, f4, f6);
__ cmp_s(UEQ, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dUeq)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fUeq)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUeq)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUeq)) );
__ cmp_d(LT, f2, f4, f6);
__ cmp_s(LT, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dOlt)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fOlt)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOlt)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOlt)) );
__ cmp_d(ULT, f2, f4, f6);
__ cmp_s(ULT, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dUlt)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fUlt)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUlt)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUlt)) );
__ cmp_d(LE, f2, f4, f6);
__ cmp_s(LE, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dOle)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fOle)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOle)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOle)) );
__ cmp_d(ULE, f2, f4, f6);
__ cmp_s(ULE, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dUle)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fUle)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUle)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUle)) );
__ cmp_d(ORD, f2, f4, f6);
__ cmp_s(ORD, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dOr)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fOr)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOr)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOr)) );
__ cmp_d(UNE, f2, f4, f6);
__ cmp_s(UNE, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dUne)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fUne)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUne)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUne)) );
__ cmp_d(NE, f2, f4, f6);
__ cmp_s(NE, f12, f14, f16);
- __ sdc1(f2, MemOperand(a0, OFFSET_OF(TestFloat, dNe)) );
- __ swc1(f12, MemOperand(a0, OFFSET_OF(TestFloat, fNe)) );
+ __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dNe)) );
+ __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fNe)) );
__ jr(ra);
__ nop();
@@ -4170,10 +4170,10 @@ TEST(CVT) {
__ ctc1(zero_reg, FCSR);
#define GENERATE_CVT_TEST(x, y, z) \
- __ y##c1(f0, MemOperand(a0, OFFSET_OF(TestFloat, x##_in))); \
+ __ y##c1(f0, MemOperand(a0, offsetof(TestFloat, x##_in))); \
__ x(f0, f0); \
__ nop(); \
- __ z##c1(f0, MemOperand(a0, OFFSET_OF(TestFloat, x##_out)));
+ __ z##c1(f0, MemOperand(a0, offsetof(TestFloat, x##_out)));
GENERATE_CVT_TEST(cvt_d_s, lw, sd)
GENERATE_CVT_TEST(cvt_d_w, lw, sd)
@@ -4350,17 +4350,17 @@ TEST(DIV_FMT) {
// Disable FPU exceptions.
__ ctc1(zero_reg, FCSR);
- __ ldc1(f4, MemOperand(a0, OFFSET_OF(Test, dOp1)) );
- __ ldc1(f2, MemOperand(a0, OFFSET_OF(Test, dOp2)) );
+ __ ldc1(f4, MemOperand(a0, offsetof(Test, dOp1)) );
+ __ ldc1(f2, MemOperand(a0, offsetof(Test, dOp2)) );
__ nop();
__ div_d(f6, f4, f2);
- __ sdc1(f6, MemOperand(a0, OFFSET_OF(Test, dRes)) );
+ __ sdc1(f6, MemOperand(a0, offsetof(Test, dRes)) );
- __ lwc1(f4, MemOperand(a0, OFFSET_OF(Test, fOp1)) );
- __ lwc1(f2, MemOperand(a0, OFFSET_OF(Test, fOp2)) );
+ __ lwc1(f4, MemOperand(a0, offsetof(Test, fOp1)) );
+ __ lwc1(f2, MemOperand(a0, offsetof(Test, fOp2)) );
__ nop();
__ div_s(f6, f4, f2);
- __ swc1(f6, MemOperand(a0, OFFSET_OF(Test, fRes)) );
+ __ swc1(f6, MemOperand(a0, offsetof(Test, fRes)) );
// Restore FCSR.
__ ctc1(a1, FCSR);
« no previous file with comments | « test/cctest/test-assembler-mips.cc ('k') | test/cctest/test-assembler-ppc.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698