| Index: test/cctest/test-assembler-mips64.cc
|
| diff --git a/test/cctest/test-assembler-mips64.cc b/test/cctest/test-assembler-mips64.cc
|
| index 5875f1e79d10de4973f3ca1577b5f60d5724d2fe..0e46f2ca9b424710c6b2d2ed495c160bb9b15afb 100644
|
| --- a/test/cctest/test-assembler-mips64.cc
|
| +++ b/test/cctest/test-assembler-mips64.cc
|
| @@ -288,61 +288,61 @@ TEST(MIPS3) {
|
| Label L, C;
|
|
|
| // Double precision floating point instructions.
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, a)) );
|
| - __ ldc1(f6, MemOperand(a0, offsetof(T, b)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
| __ add_d(f8, f4, f6);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(T, c)) ); // c = a + b.
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(T, c))); // c = a + b.
|
|
|
| __ mov_d(f10, f8); // c
|
| __ neg_d(f12, f6); // -b
|
| __ sub_d(f10, f10, f12);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(T, d)) ); // d = c - (-b).
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(T, d))); // d = c - (-b).
|
|
|
| - __ sdc1(f4, MemOperand(a0, offsetof(T, b)) ); // b = a.
|
| + __ Sdc1(f4, MemOperand(a0, offsetof(T, b))); // b = a.
|
|
|
| __ li(a4, 120);
|
| __ mtc1(a4, f14);
|
| __ cvt_d_w(f14, f14); // f14 = 120.0.
|
| __ mul_d(f10, f10, f14);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(T, e)) ); // e = d * 120 = 1.8066e16.
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(T, e))); // e = d * 120 = 1.8066e16.
|
|
|
| __ div_d(f12, f10, f4);
|
| - __ sdc1(f12, MemOperand(a0, offsetof(T, f)) ); // f = e / a = 120.44.
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(T, f))); // f = e / a = 120.44.
|
|
|
| __ sqrt_d(f14, f12);
|
| - __ sdc1(f14, MemOperand(a0, offsetof(T, g)) );
|
| + __ Sdc1(f14, MemOperand(a0, offsetof(T, g)));
|
| // g = sqrt(f) = 10.97451593465515908537
|
|
|
| if (kArchVariant == kMips64r2) {
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, h)) );
|
| - __ ldc1(f6, MemOperand(a0, offsetof(T, i)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, h)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(T, i)));
|
| __ madd_d(f14, f6, f4, f6);
|
| - __ sdc1(f14, MemOperand(a0, offsetof(T, h)) );
|
| + __ Sdc1(f14, MemOperand(a0, offsetof(T, h)));
|
| }
|
|
|
| // Single precision floating point instructions.
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fa)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(T, fb)) );
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fa)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(T, fb)));
|
| __ add_s(f8, f4, f6);
|
| - __ swc1(f8, MemOperand(a0, offsetof(T, fc)) ); // fc = fa + fb.
|
| + __ Swc1(f8, MemOperand(a0, offsetof(T, fc))); // fc = fa + fb.
|
|
|
| __ neg_s(f10, f6); // -fb
|
| __ sub_s(f10, f8, f10);
|
| - __ swc1(f10, MemOperand(a0, offsetof(T, fd)) ); // fd = fc - (-fb).
|
| + __ Swc1(f10, MemOperand(a0, offsetof(T, fd))); // fd = fc - (-fb).
|
|
|
| - __ swc1(f4, MemOperand(a0, offsetof(T, fb)) ); // fb = fa.
|
| + __ Swc1(f4, MemOperand(a0, offsetof(T, fb))); // fb = fa.
|
|
|
| __ li(t0, 120);
|
| __ mtc1(t0, f14);
|
| __ cvt_s_w(f14, f14); // f14 = 120.0.
|
| __ mul_s(f10, f10, f14);
|
| - __ swc1(f10, MemOperand(a0, offsetof(T, fe)) ); // fe = fd * 120
|
| + __ Swc1(f10, MemOperand(a0, offsetof(T, fe))); // fe = fd * 120
|
|
|
| __ div_s(f12, f10, f4);
|
| - __ swc1(f12, MemOperand(a0, offsetof(T, ff)) ); // ff = fe / fa
|
| + __ Swc1(f12, MemOperand(a0, offsetof(T, ff))); // ff = fe / fa
|
|
|
| __ sqrt_s(f14, f12);
|
| - __ swc1(f14, MemOperand(a0, offsetof(T, fg)) );
|
| + __ Swc1(f14, MemOperand(a0, offsetof(T, fg)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -408,11 +408,11 @@ TEST(MIPS4) {
|
| } T;
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
| Label L, C;
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
| - __ ldc1(f5, MemOperand(a0, offsetof(T, b)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
| + __ Ldc1(f5, MemOperand(a0, offsetof(T, b)));
|
|
|
| // Swap f4 and f5, by using 3 integer registers, a4-a6,
|
| // both two 32-bit chunks, and one 64-bit chunk.
|
| @@ -427,16 +427,16 @@ TEST(MIPS4) {
|
| __ dmtc1(a6, f4);
|
|
|
| // Store the swapped f4 and f5 back to memory.
|
| - __ sdc1(f4, MemOperand(a0, offsetof(T, a)));
|
| - __ sdc1(f5, MemOperand(a0, offsetof(T, c)));
|
| + __ Sdc1(f4, MemOperand(a0, offsetof(T, a)));
|
| + __ Sdc1(f5, MemOperand(a0, offsetof(T, c)));
|
|
|
| // Test sign extension of move operations from coprocessor.
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, d)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, d)));
|
| __ mfhc1(a4, f4);
|
| __ mfc1(a5, f4);
|
|
|
| - __ sd(a4, MemOperand(a0, offsetof(T, high)));
|
| - __ sd(a5, MemOperand(a0, offsetof(T, low)));
|
| + __ Sd(a4, MemOperand(a0, offsetof(T, high)));
|
| + __ Sd(a5, MemOperand(a0, offsetof(T, low)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -475,34 +475,34 @@ TEST(MIPS5) {
|
| } T;
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
| Label L, C;
|
|
|
| // Load all structure elements to registers.
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, a)) );
|
| - __ ldc1(f6, MemOperand(a0, offsetof(T, b)) );
|
| - __ lw(a4, MemOperand(a0, offsetof(T, i)) );
|
| - __ lw(a5, MemOperand(a0, offsetof(T, j)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, i)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, j)));
|
|
|
| // Convert double in f4 to int in element i.
|
| __ cvt_w_d(f8, f4);
|
| __ mfc1(a6, f8);
|
| - __ sw(a6, MemOperand(a0, offsetof(T, i)) );
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, i)));
|
|
|
| // Convert double in f6 to int in element j.
|
| __ cvt_w_d(f10, f6);
|
| __ mfc1(a7, f10);
|
| - __ sw(a7, MemOperand(a0, offsetof(T, j)) );
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, j)));
|
|
|
| // Convert int in original i (a4) to double in a.
|
| __ mtc1(a4, f12);
|
| __ cvt_d_w(f0, f12);
|
| - __ sdc1(f0, MemOperand(a0, offsetof(T, a)) );
|
| + __ Sdc1(f0, MemOperand(a0, offsetof(T, a)));
|
|
|
| // Convert int in original j (a5) to double in b.
|
| __ mtc1(a5, f14);
|
| __ cvt_d_w(f2, f14);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(T, b)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(T, b)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -544,35 +544,35 @@ TEST(MIPS6) {
|
| } T;
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
| Label L, C;
|
|
|
| // Basic word load/store.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, ui)) );
|
| - __ sw(a4, MemOperand(a0, offsetof(T, r1)) );
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, ui)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, r1)));
|
|
|
| // lh with positive data.
|
| - __ lh(a5, MemOperand(a0, offsetof(T, ui)) );
|
| - __ sw(a5, MemOperand(a0, offsetof(T, r2)) );
|
| + __ Lh(a5, MemOperand(a0, offsetof(T, ui)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, r2)));
|
|
|
| // lh with negative data.
|
| - __ lh(a6, MemOperand(a0, offsetof(T, si)) );
|
| - __ sw(a6, MemOperand(a0, offsetof(T, r3)) );
|
| + __ Lh(a6, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, r3)));
|
|
|
| // lhu with negative data.
|
| - __ lhu(a7, MemOperand(a0, offsetof(T, si)) );
|
| - __ sw(a7, MemOperand(a0, offsetof(T, r4)) );
|
| + __ Lhu(a7, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, r4)));
|
|
|
| - // lb with negative data.
|
| - __ lb(t0, MemOperand(a0, offsetof(T, si)) );
|
| - __ sw(t0, MemOperand(a0, offsetof(T, r5)) );
|
| + // Lb with negative data.
|
| + __ Lb(t0, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(t0, MemOperand(a0, offsetof(T, r5)));
|
|
|
| // sh writes only 1/2 of word.
|
| __ lui(t1, 0x3333);
|
| __ ori(t1, t1, 0x3333);
|
| - __ sw(t1, MemOperand(a0, offsetof(T, r6)) );
|
| - __ lhu(t1, MemOperand(a0, offsetof(T, si)) );
|
| - __ sh(t1, MemOperand(a0, offsetof(T, r6)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(T, r6)));
|
| + __ Lhu(t1, MemOperand(a0, offsetof(T, si)));
|
| + __ Sh(t1, MemOperand(a0, offsetof(T, r6)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -626,8 +626,8 @@ TEST(MIPS7) {
|
| MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
| Label neither_is_nan, less_than, outa_here;
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, a)) );
|
| - __ ldc1(f6, MemOperand(a0, offsetof(T, b)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, a)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(T, b)));
|
| if (kArchVariant != kMips64r6) {
|
| __ c(UN, D, f4, f6);
|
| __ bc1f(&neither_is_nan);
|
| @@ -636,7 +636,7 @@ TEST(MIPS7) {
|
| __ bc1eqz(&neither_is_nan, f2);
|
| }
|
| __ nop();
|
| - __ sw(zero_reg, MemOperand(a0, offsetof(T, result)) );
|
| + __ Sw(zero_reg, MemOperand(a0, offsetof(T, result)));
|
| __ Branch(&outa_here);
|
|
|
| __ bind(&neither_is_nan);
|
| @@ -650,13 +650,12 @@ TEST(MIPS7) {
|
| }
|
|
|
| __ nop();
|
| - __ sw(zero_reg, MemOperand(a0, offsetof(T, result)) );
|
| + __ Sw(zero_reg, MemOperand(a0, offsetof(T, result)));
|
| __ Branch(&outa_here);
|
|
|
| __ bind(&less_than);
|
| __ Addu(a4, zero_reg, Operand(1));
|
| - __ sw(a4, MemOperand(a0, offsetof(T, result)) ); // Set true.
|
| -
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, result))); // Set true.
|
|
|
| // This test-case should have additional tests.
|
|
|
| @@ -715,7 +714,7 @@ TEST(MIPS8) {
|
| v8::internal::CodeObjectRequired::kYes);
|
|
|
| // Basic word load.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, input)) );
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, input)));
|
|
|
| // ROTR instruction (called through the Ror macro).
|
| __ Ror(a5, a4, 0x0004);
|
| @@ -727,13 +726,13 @@ TEST(MIPS8) {
|
| __ Ror(t3, a4, 0x001c);
|
|
|
| // Basic word store.
|
| - __ sw(a5, MemOperand(a0, offsetof(T, result_rotr_4)) );
|
| - __ sw(a6, MemOperand(a0, offsetof(T, result_rotr_8)) );
|
| - __ sw(a7, MemOperand(a0, offsetof(T, result_rotr_12)) );
|
| - __ sw(t0, MemOperand(a0, offsetof(T, result_rotr_16)) );
|
| - __ sw(t1, MemOperand(a0, offsetof(T, result_rotr_20)) );
|
| - __ sw(t2, MemOperand(a0, offsetof(T, result_rotr_24)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(T, result_rotr_28)) );
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, result_rotr_4)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, result_rotr_8)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, result_rotr_12)));
|
| + __ Sw(t0, MemOperand(a0, offsetof(T, result_rotr_16)));
|
| + __ Sw(t1, MemOperand(a0, offsetof(T, result_rotr_20)));
|
| + __ Sw(t2, MemOperand(a0, offsetof(T, result_rotr_24)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(T, result_rotr_28)));
|
|
|
| // ROTRV instruction (called through the Ror macro).
|
| __ li(t3, 0x0004);
|
| @@ -752,13 +751,13 @@ TEST(MIPS8) {
|
| __ Ror(t3, a4, t3);
|
|
|
| // Basic word store.
|
| - __ sw(a5, MemOperand(a0, offsetof(T, result_rotrv_4)) );
|
| - __ sw(a6, MemOperand(a0, offsetof(T, result_rotrv_8)) );
|
| - __ sw(a7, MemOperand(a0, offsetof(T, result_rotrv_12)) );
|
| - __ sw(t0, MemOperand(a0, offsetof(T, result_rotrv_16)) );
|
| - __ sw(t1, MemOperand(a0, offsetof(T, result_rotrv_20)) );
|
| - __ sw(t2, MemOperand(a0, offsetof(T, result_rotrv_24)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(T, result_rotrv_28)) );
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, result_rotrv_4)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, result_rotrv_8)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, result_rotrv_12)));
|
| + __ Sw(t0, MemOperand(a0, offsetof(T, result_rotrv_16)));
|
| + __ Sw(t1, MemOperand(a0, offsetof(T, result_rotrv_20)));
|
| + __ Sw(t2, MemOperand(a0, offsetof(T, result_rotrv_24)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(T, result_rotrv_28)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -838,7 +837,7 @@ TEST(MIPS10) {
|
| } T;
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
| Label L, C;
|
|
|
| if (kArchVariant == kMips64r2) {
|
| @@ -846,43 +845,41 @@ TEST(MIPS10) {
|
| // - 32 FP regs of 64-bits each, no odd/even pairs.
|
| // - Note that cvt_l_d/cvt_d_l ARE legal in FR=1 mode.
|
| // Load all structure elements to registers.
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, a)));
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, a)));
|
|
|
| // Save the raw bits of the double.
|
| __ mfc1(a4, f0);
|
| __ mfhc1(a5, f0);
|
| - __ sw(a4, MemOperand(a0, offsetof(T, dbl_mant)));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, dbl_exp)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, dbl_mant)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, dbl_exp)));
|
|
|
| // Convert double in f0 to long, save hi/lo parts.
|
| __ cvt_l_d(f0, f0);
|
| __ mfc1(a4, f0); // f0 LS 32 bits of long.
|
| __ mfhc1(a5, f0); // f0 MS 32 bits of long.
|
| - __ sw(a4, MemOperand(a0, offsetof(T, long_lo)));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, long_hi)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, long_lo)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, long_hi)));
|
|
|
| // Combine the high/low ints, convert back to double.
|
| __ dsll32(a6, a5, 0); // Move a5 to high bits of a6.
|
| __ or_(a6, a6, a4);
|
| __ dmtc1(a6, f1);
|
| __ cvt_d_l(f1, f1);
|
| - __ sdc1(f1, MemOperand(a0, offsetof(T, a_converted)));
|
| -
|
| + __ Sdc1(f1, MemOperand(a0, offsetof(T, a_converted)));
|
|
|
| // Convert the b long integers to double b.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, b_long_lo)));
|
| - __ lw(a5, MemOperand(a0, offsetof(T, b_long_hi)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, b_long_lo)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, b_long_hi)));
|
| __ mtc1(a4, f8); // f8 LS 32-bits.
|
| __ mthc1(a5, f8); // f8 MS 32-bits.
|
| __ cvt_d_l(f10, f8);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(T, b)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(T, b)));
|
|
|
| // Convert double b back to long-int.
|
| - __ ldc1(f31, MemOperand(a0, offsetof(T, b)));
|
| + __ Ldc1(f31, MemOperand(a0, offsetof(T, b)));
|
| __ cvt_l_d(f31, f31);
|
| __ dmfc1(a7, f31);
|
| - __ sd(a7, MemOperand(a0, offsetof(T, b_long_as_int64)));
|
| -
|
| + __ Sd(a7, MemOperand(a0, offsetof(T, b_long_as_int64)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -941,82 +938,83 @@ TEST(MIPS11) {
|
| } T;
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0,
|
| + v8::internal::CodeObjectRequired::kYes);
|
|
|
| // Test all combinations of LWL and vAddr.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwl(a4, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a4, MemOperand(a0, offsetof(T, lwl_0)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, lwl_0)));
|
|
|
| - __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwl(a5, MemOperand(a0, offsetof(T, mem_init) + 1));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, lwl_1)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, lwl_1)));
|
|
|
| - __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwl(a6, MemOperand(a0, offsetof(T, mem_init) + 2));
|
| - __ sw(a6, MemOperand(a0, offsetof(T, lwl_2)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, lwl_2)));
|
|
|
| - __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwl(a7, MemOperand(a0, offsetof(T, mem_init) + 3));
|
| - __ sw(a7, MemOperand(a0, offsetof(T, lwl_3)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, lwl_3)));
|
|
|
| // Test all combinations of LWR and vAddr.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwr(a4, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a4, MemOperand(a0, offsetof(T, lwr_0)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, lwr_0)));
|
|
|
| - __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwr(a5, MemOperand(a0, offsetof(T, mem_init) + 1));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, lwr_1)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, lwr_1)));
|
|
|
| - __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwr(a6, MemOperand(a0, offsetof(T, mem_init) + 2));
|
| - __ sw(a6, MemOperand(a0, offsetof(T, lwr_2)) );
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, lwr_2)));
|
|
|
| - __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| __ lwr(a7, MemOperand(a0, offsetof(T, mem_init) + 3));
|
| - __ sw(a7, MemOperand(a0, offsetof(T, lwr_3)) );
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, lwr_3)));
|
|
|
| // Test all combinations of SWL and vAddr.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a4, MemOperand(a0, offsetof(T, swl_0)));
|
| - __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, swl_0)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swl(a4, MemOperand(a0, offsetof(T, swl_0)));
|
|
|
| - __ lw(a5, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, swl_1)));
|
| - __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, swl_1)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swl(a5, MemOperand(a0, offsetof(T, swl_1) + 1));
|
|
|
| - __ lw(a6, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a6, MemOperand(a0, offsetof(T, swl_2)));
|
| - __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a6, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, swl_2)));
|
| + __ Lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swl(a6, MemOperand(a0, offsetof(T, swl_2) + 2));
|
|
|
| - __ lw(a7, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a7, MemOperand(a0, offsetof(T, swl_3)));
|
| - __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, swl_3)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swl(a7, MemOperand(a0, offsetof(T, swl_3) + 3));
|
|
|
| // Test all combinations of SWR and vAddr.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a4, MemOperand(a0, offsetof(T, swr_0)));
|
| - __ lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, swr_0)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swr(a4, MemOperand(a0, offsetof(T, swr_0)));
|
|
|
| - __ lw(a5, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, swr_1)));
|
| - __ lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, swr_1)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swr(a5, MemOperand(a0, offsetof(T, swr_1) + 1));
|
|
|
| - __ lw(a6, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a6, MemOperand(a0, offsetof(T, swr_2)));
|
| - __ lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a6, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, swr_2)));
|
| + __ Lw(a6, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swr(a6, MemOperand(a0, offsetof(T, swr_2) + 2));
|
|
|
| - __ lw(a7, MemOperand(a0, offsetof(T, mem_init)));
|
| - __ sw(a7, MemOperand(a0, offsetof(T, swr_3)));
|
| - __ lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, mem_init)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, swr_3)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, reg_init)));
|
| __ swr(a7, MemOperand(a0, offsetof(T, swr_3) + 3));
|
|
|
| __ jr(ra);
|
| @@ -1097,8 +1095,8 @@ TEST(MIPS12) {
|
|
|
| __ mov(t2, fp); // Save frame pointer.
|
| __ mov(fp, a0); // Access struct T by fp.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, y)));
|
| - __ lw(a7, MemOperand(a0, offsetof(T, y4)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, y)));
|
| + __ Lw(a7, MemOperand(a0, offsetof(T, y4)));
|
|
|
| __ addu(a5, a4, a7);
|
| __ subu(t0, a4, a7);
|
| @@ -1116,30 +1114,30 @@ TEST(MIPS12) {
|
| __ push(a7);
|
| __ pop(t0);
|
| __ nop();
|
| - __ sw(a4, MemOperand(fp, offsetof(T, y)));
|
| - __ lw(a4, MemOperand(fp, offsetof(T, y)));
|
| + __ Sw(a4, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a4, MemOperand(fp, offsetof(T, y)));
|
| __ nop();
|
| - __ sw(a4, MemOperand(fp, offsetof(T, y)));
|
| - __ lw(a5, MemOperand(fp, offsetof(T, y)));
|
| + __ Sw(a4, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a5, MemOperand(fp, offsetof(T, y)));
|
| __ nop();
|
| __ push(a5);
|
| - __ lw(a5, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a5, MemOperand(fp, offsetof(T, y)));
|
| __ pop(a5);
|
| __ nop();
|
| __ push(a5);
|
| - __ lw(a6, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a6, MemOperand(fp, offsetof(T, y)));
|
| __ pop(a5);
|
| __ nop();
|
| __ push(a5);
|
| - __ lw(a6, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a6, MemOperand(fp, offsetof(T, y)));
|
| __ pop(a6);
|
| __ nop();
|
| __ push(a6);
|
| - __ lw(a6, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a6, MemOperand(fp, offsetof(T, y)));
|
| __ pop(a5);
|
| __ nop();
|
| __ push(a5);
|
| - __ lw(a6, MemOperand(fp, offsetof(T, y)));
|
| + __ Lw(a6, MemOperand(fp, offsetof(T, y)));
|
| __ pop(a7);
|
| __ nop();
|
|
|
| @@ -1184,19 +1182,19 @@ TEST(MIPS13) {
|
|
|
| MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
|
|
| - __ sw(a4, MemOperand(a0, offsetof(T, cvt_small_in)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, cvt_small_in)));
|
| __ Cvt_d_uw(f10, a4);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(T, cvt_small_out)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(T, cvt_small_out)));
|
|
|
| __ Trunc_uw_d(f10, f10, f4);
|
| - __ swc1(f10, MemOperand(a0, offsetof(T, trunc_small_out)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(T, trunc_small_out)));
|
|
|
| - __ sw(a4, MemOperand(a0, offsetof(T, cvt_big_in)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, cvt_big_in)));
|
| __ Cvt_d_uw(f8, a4);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(T, cvt_big_out)));
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(T, cvt_big_out)));
|
|
|
| __ Trunc_uw_d(f8, f8, f4);
|
| - __ swc1(f8, MemOperand(a0, offsetof(T, trunc_big_out)));
|
| + __ Swc1(f8, MemOperand(a0, offsetof(T, trunc_big_out)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -1266,49 +1264,49 @@ TEST(MIPS14) {
|
| __ cfc1(a1, FCSR);
|
| // Disable FPU exceptions.
|
| __ ctc1(zero_reg, FCSR);
|
| -#define RUN_ROUND_TEST(x) \
|
| - __ cfc1(t0, FCSR);\
|
| - __ sw(t0, MemOperand(a0, offsetof(T, x##_isNaN2008))); \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, round_up_in))); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ swc1(f0, MemOperand(a0, offsetof(T, x##_up_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, round_down_in))); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ swc1(f0, MemOperand(a0, offsetof(T, x##_down_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, neg_round_up_in))); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_up_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, neg_round_down_in))); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_down_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, err1_in))); \
|
| - __ ctc1(zero_reg, FCSR); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ cfc1(a2, FCSR); \
|
| - __ sw(a2, MemOperand(a0, offsetof(T, x##_err1_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, err2_in))); \
|
| - __ ctc1(zero_reg, FCSR); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ cfc1(a2, FCSR); \
|
| - __ sw(a2, MemOperand(a0, offsetof(T, x##_err2_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, err3_in))); \
|
| - __ ctc1(zero_reg, FCSR); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ cfc1(a2, FCSR); \
|
| - __ sw(a2, MemOperand(a0, offsetof(T, x##_err3_out))); \
|
| - \
|
| - __ ldc1(f0, MemOperand(a0, offsetof(T, err4_in))); \
|
| - __ ctc1(zero_reg, FCSR); \
|
| - __ x##_w_d(f0, f0); \
|
| - __ cfc1(a2, FCSR); \
|
| - __ sw(a2, MemOperand(a0, offsetof(T, x##_err4_out))); \
|
| - __ swc1(f0, MemOperand(a0, offsetof(T, x##_invalid_result)));
|
| +#define RUN_ROUND_TEST(x) \
|
| + __ cfc1(t0, FCSR); \
|
| + __ Sw(t0, MemOperand(a0, offsetof(T, x##_isNaN2008))); \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, round_up_in))); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ Swc1(f0, MemOperand(a0, offsetof(T, x##_up_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, round_down_in))); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ Swc1(f0, MemOperand(a0, offsetof(T, x##_down_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, neg_round_up_in))); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ Swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_up_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, neg_round_down_in))); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ Swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_down_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, err1_in))); \
|
| + __ ctc1(zero_reg, FCSR); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ cfc1(a2, FCSR); \
|
| + __ Sw(a2, MemOperand(a0, offsetof(T, x##_err1_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, err2_in))); \
|
| + __ ctc1(zero_reg, FCSR); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ cfc1(a2, FCSR); \
|
| + __ Sw(a2, MemOperand(a0, offsetof(T, x##_err2_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, err3_in))); \
|
| + __ ctc1(zero_reg, FCSR); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ cfc1(a2, FCSR); \
|
| + __ Sw(a2, MemOperand(a0, offsetof(T, x##_err3_out))); \
|
| + \
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(T, err4_in))); \
|
| + __ ctc1(zero_reg, FCSR); \
|
| + __ x##_w_d(f0, f0); \
|
| + __ cfc1(a2, FCSR); \
|
| + __ Sw(a2, MemOperand(a0, offsetof(T, x##_err4_out))); \
|
| + __ Swc1(f0, MemOperand(a0, offsetof(T, x##_invalid_result)));
|
|
|
| RUN_ROUND_TEST(round)
|
| RUN_ROUND_TEST(floor)
|
| @@ -1403,51 +1401,51 @@ TEST(MIPS16) {
|
| };
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
| Label L, C;
|
|
|
| // Basic 32-bit word load/store, with un-signed data.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, ui)));
|
| - __ sw(a4, MemOperand(a0, offsetof(T, r1)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, ui)));
|
| + __ Sw(a4, MemOperand(a0, offsetof(T, r1)));
|
|
|
| // Check that the data got zero-extended into 64-bit a4.
|
| - __ sd(a4, MemOperand(a0, offsetof(T, r2)));
|
| + __ Sd(a4, MemOperand(a0, offsetof(T, r2)));
|
|
|
| // Basic 32-bit word load/store, with SIGNED data.
|
| - __ lw(a5, MemOperand(a0, offsetof(T, si)));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, r3)));
|
| + __ Lw(a5, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, r3)));
|
|
|
| // Check that the data got sign-extended into 64-bit a4.
|
| - __ sd(a5, MemOperand(a0, offsetof(T, r4)));
|
| + __ Sd(a5, MemOperand(a0, offsetof(T, r4)));
|
|
|
| // 32-bit UNSIGNED word load/store, with SIGNED data.
|
| - __ lwu(a6, MemOperand(a0, offsetof(T, si)));
|
| - __ sw(a6, MemOperand(a0, offsetof(T, r5)));
|
| + __ Lwu(a6, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, r5)));
|
|
|
| // Check that the data got zero-extended into 64-bit a4.
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r6)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r6)));
|
|
|
| // lh with positive data.
|
| - __ lh(a5, MemOperand(a0, offsetof(T, ui)));
|
| - __ sw(a5, MemOperand(a0, offsetof(T, r7)));
|
| + __ Lh(a5, MemOperand(a0, offsetof(T, ui)));
|
| + __ Sw(a5, MemOperand(a0, offsetof(T, r7)));
|
|
|
| // lh with negative data.
|
| - __ lh(a6, MemOperand(a0, offsetof(T, si)));
|
| - __ sw(a6, MemOperand(a0, offsetof(T, r8)));
|
| + __ Lh(a6, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(a6, MemOperand(a0, offsetof(T, r8)));
|
|
|
| // lhu with negative data.
|
| - __ lhu(a7, MemOperand(a0, offsetof(T, si)));
|
| - __ sw(a7, MemOperand(a0, offsetof(T, r9)));
|
| + __ Lhu(a7, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(a7, MemOperand(a0, offsetof(T, r9)));
|
|
|
| - // lb with negative data.
|
| - __ lb(t0, MemOperand(a0, offsetof(T, si)));
|
| - __ sw(t0, MemOperand(a0, offsetof(T, r10)));
|
| + // Lb with negative data.
|
| + __ Lb(t0, MemOperand(a0, offsetof(T, si)));
|
| + __ Sw(t0, MemOperand(a0, offsetof(T, r10)));
|
|
|
| // sh writes only 1/2 of word.
|
| - __ lw(a4, MemOperand(a0, offsetof(T, ui)));
|
| - __ sh(a4, MemOperand(a0, offsetof(T, r11)));
|
| - __ lw(a4, MemOperand(a0, offsetof(T, si)));
|
| - __ sh(a4, MemOperand(a0, offsetof(T, r12)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, ui)));
|
| + __ Sh(a4, MemOperand(a0, offsetof(T, r11)));
|
| + __ Lw(a4, MemOperand(a0, offsetof(T, si)));
|
| + __ Sh(a4, MemOperand(a0, offsetof(T, r12)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -1559,26 +1557,26 @@ TEST(seleqz_selnez) {
|
| // Integer part of test.
|
| __ addiu(t1, zero_reg, 1); // t1 = 1
|
| __ seleqz(t3, t1, zero_reg); // t3 = 1
|
| - __ sw(t3, MemOperand(a0, offsetof(Test, a))); // a = 1
|
| + __ Sw(t3, MemOperand(a0, offsetof(Test, a))); // a = 1
|
| __ seleqz(t2, t1, t1); // t2 = 0
|
| - __ sw(t2, MemOperand(a0, offsetof(Test, b))); // b = 0
|
| + __ Sw(t2, MemOperand(a0, offsetof(Test, b))); // b = 0
|
| __ selnez(t3, t1, zero_reg); // t3 = 1;
|
| - __ sw(t3, MemOperand(a0, offsetof(Test, c))); // c = 0
|
| + __ Sw(t3, MemOperand(a0, offsetof(Test, c))); // c = 0
|
| __ selnez(t3, t1, t1); // t3 = 1
|
| - __ sw(t3, MemOperand(a0, offsetof(Test, d))); // d = 1
|
| + __ Sw(t3, MemOperand(a0, offsetof(Test, d))); // d = 1
|
| // Floating point part of test.
|
| - __ ldc1(f0, MemOperand(a0, offsetof(Test, e)) ); // src
|
| - __ ldc1(f2, MemOperand(a0, offsetof(Test, f)) ); // test
|
| - __ lwc1(f8, MemOperand(a0, offsetof(Test, i)) ); // src
|
| - __ lwc1(f10, MemOperand(a0, offsetof(Test, j)) ); // test
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(Test, e))); // src
|
| + __ Ldc1(f2, MemOperand(a0, offsetof(Test, f))); // test
|
| + __ Lwc1(f8, MemOperand(a0, offsetof(Test, i))); // src
|
| + __ Lwc1(f10, MemOperand(a0, offsetof(Test, j))); // test
|
| __ seleqz_d(f4, f0, f2);
|
| __ selnez_d(f6, f0, f2);
|
| __ seleqz_s(f12, f8, f10);
|
| __ selnez_s(f14, f8, f10);
|
| - __ sdc1(f4, MemOperand(a0, offsetof(Test, g)) ); // src
|
| - __ sdc1(f6, MemOperand(a0, offsetof(Test, h)) ); // src
|
| - __ swc1(f12, MemOperand(a0, offsetof(Test, k)) ); // src
|
| - __ swc1(f14, MemOperand(a0, offsetof(Test, l)) ); // src
|
| + __ Sdc1(f4, MemOperand(a0, offsetof(Test, g))); // src
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(Test, h))); // src
|
| + __ Swc1(f12, MemOperand(a0, offsetof(Test, k))); // src
|
| + __ Swc1(f14, MemOperand(a0, offsetof(Test, l))); // src
|
| __ jr(ra);
|
| __ nop();
|
| CodeDesc desc;
|
| @@ -1682,18 +1680,18 @@ TEST(min_max) {
|
| float outputsfmax[kTableLength] = {3.0, 3.0, 3.0, 3.0, 0.0, 0.0, finf,
|
| finf, finf, finf, finf, finf, fnan};
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
| - __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, e)));
|
| - __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, f)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Lwc1(f2, MemOperand(a0, offsetof(TestFloat, e)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(TestFloat, f)));
|
| __ min_d(f10, f4, f8);
|
| __ max_d(f12, f4, f8);
|
| __ min_s(f14, f2, f6);
|
| __ max_s(f16, f2, f6);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, d)));
|
| - __ swc1(f14, MemOperand(a0, offsetof(TestFloat, g)));
|
| - __ swc1(f16, MemOperand(a0, offsetof(TestFloat, h)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, d)));
|
| + __ Swc1(f14, MemOperand(a0, offsetof(TestFloat, g)));
|
| + __ Swc1(f16, MemOperand(a0, offsetof(TestFloat, h)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -1798,11 +1796,11 @@ TEST(rint_d) {
|
| int fcsr_inputs[4] =
|
| {kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
|
| double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)));
|
| __ ctc1(t0, FCSR);
|
| __ rint_d(f8, f4);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(TestFloat, b)) );
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -1842,16 +1840,16 @@ TEST(sel) {
|
| } Test;
|
|
|
| Test test;
|
| - __ ldc1(f0, MemOperand(a0, offsetof(Test, dd)) ); // test
|
| - __ ldc1(f2, MemOperand(a0, offsetof(Test, ds)) ); // src1
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, dt)) ); // src2
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, fd)) ); // test
|
| - __ lwc1(f8, MemOperand(a0, offsetof(Test, fs)) ); // src1
|
| - __ lwc1(f10, MemOperand(a0, offsetof(Test, ft)) ); // src2
|
| + __ Ldc1(f0, MemOperand(a0, offsetof(Test, dd))); // test
|
| + __ Ldc1(f2, MemOperand(a0, offsetof(Test, ds))); // src1
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, dt))); // src2
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, fd))); // test
|
| + __ Lwc1(f8, MemOperand(a0, offsetof(Test, fs))); // src1
|
| + __ Lwc1(f10, MemOperand(a0, offsetof(Test, ft))); // src2
|
| __ sel_d(f0, f2, f4);
|
| __ sel_s(f6, f8, f10);
|
| - __ sdc1(f0, MemOperand(a0, offsetof(Test, dd)) );
|
| - __ swc1(f6, MemOperand(a0, offsetof(Test, fd)) );
|
| + __ Sdc1(f0, MemOperand(a0, offsetof(Test, dd)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(Test, fd)));
|
| __ jr(ra);
|
| __ nop();
|
| CodeDesc desc;
|
| @@ -1977,12 +1975,12 @@ TEST(rint_s) {
|
| int fcsr_inputs[4] =
|
| {kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
|
| float* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) );
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)));
|
| __ cfc1(t1, FCSR);
|
| __ ctc1(t0, FCSR);
|
| __ rint_s(f8, f4);
|
| - __ swc1(f8, MemOperand(a0, offsetof(TestFloat, b)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
| __ ctc1(t1, FCSR);
|
| __ jr(ra);
|
| __ nop();
|
| @@ -2058,18 +2056,18 @@ TEST(mina_maxa) {
|
| 5.3, 5.3, 6.1, -10.0, 9.8, 9.8, -10.0, 9.8, 9.8, -10.0, -11.2, -9.8,
|
| 3.0, 3.0, 0.0, 0.0, finf, finf, finf, finf, finf, finf, fnan};
|
|
|
| - __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
|
| - __ lwc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
|
| - __ lwc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
|
| + __ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Lwc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
| + __ Lwc1(f10, MemOperand(a0, offsetof(TestFloat, d)));
|
| __ mina_d(f6, f2, f4);
|
| __ mina_s(f12, f8, f10);
|
| __ maxa_d(f14, f2, f4);
|
| __ maxa_s(f16, f8, f10);
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, resf)) );
|
| - __ sdc1(f6, MemOperand(a0, offsetof(TestFloat, resd)) );
|
| - __ swc1(f16, MemOperand(a0, offsetof(TestFloat, resf1)) );
|
| - __ sdc1(f14, MemOperand(a0, offsetof(TestFloat, resd1)) );
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, resf)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(TestFloat, resd)));
|
| + __ Swc1(f16, MemOperand(a0, offsetof(TestFloat, resf1)));
|
| + __ Sdc1(f14, MemOperand(a0, offsetof(TestFloat, resd1)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -2145,13 +2143,13 @@ TEST(trunc_l) {
|
| dFPU64InvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ trunc_l_d(f8, f4);
|
| __ trunc_l_s(f10, f6);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -2214,25 +2212,25 @@ TEST(movz_movn) {
|
| 5.3, -5.3, 5.3, -2.9
|
| };
|
|
|
| - __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
|
| - __ ld(t0, MemOperand(a0, offsetof(TestFloat, rt)));
|
| + __ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)));
|
| + __ Ld(t0, MemOperand(a0, offsetof(TestFloat, rt)));
|
| __ Move(f12, 0.0);
|
| __ Move(f10, 0.0);
|
| __ Move(f16, 0.0);
|
| __ Move(f14, 0.0);
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, bold)) );
|
| - __ swc1(f10, MemOperand(a0, offsetof(TestFloat, dold)) );
|
| - __ sdc1(f16, MemOperand(a0, offsetof(TestFloat, bold1)) );
|
| - __ swc1(f14, MemOperand(a0, offsetof(TestFloat, dold1)) );
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, bold)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(TestFloat, dold)));
|
| + __ Sdc1(f16, MemOperand(a0, offsetof(TestFloat, bold1)));
|
| + __ Swc1(f14, MemOperand(a0, offsetof(TestFloat, dold1)));
|
| __ movz_s(f10, f6, t0);
|
| __ movz_d(f12, f2, t0);
|
| __ movn_s(f14, f6, t0);
|
| __ movn_d(f16, f2, t0);
|
| - __ swc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, b)) );
|
| - __ swc1(f14, MemOperand(a0, offsetof(TestFloat, d1)) );
|
| - __ sdc1(f16, MemOperand(a0, offsetof(TestFloat, b1)) );
|
| + __ Swc1(f10, MemOperand(a0, offsetof(TestFloat, d)));
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Swc1(f14, MemOperand(a0, offsetof(TestFloat, d1)));
|
| + __ Sdc1(f16, MemOperand(a0, offsetof(TestFloat, b1)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -2313,26 +2311,26 @@ TEST(movt_movd) {
|
| HandleScope scope(isolate);
|
| MacroAssembler assm(isolate, NULL, 0,
|
| v8::internal::CodeObjectRequired::kYes);
|
| - __ ldc1(f2, MemOperand(a0, offsetof(TestFloat, srcd)) );
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, srcf)) );
|
| - __ lw(t1, MemOperand(a0, offsetof(TestFloat, fcsr)) );
|
| + __ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, srcd)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestFloat, srcf)));
|
| + __ Lw(t1, MemOperand(a0, offsetof(TestFloat, fcsr)));
|
| __ cfc1(t0, FCSR);
|
| __ ctc1(t1, FCSR);
|
| __ li(t2, 0x0);
|
| __ mtc1(t2, f12);
|
| __ mtc1(t2, f10);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold)) );
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold)));
|
| __ movt_s(f12, f4, test.cc);
|
| __ movt_d(f10, f2, test.cc);
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold1)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold1)) );
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold1)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold1)));
|
| __ movf_s(f12, f4, test.cc);
|
| __ movf_d(f10, f2, test.cc);
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf1)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd1)) );
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf1)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd1)));
|
| __ ctc1(t0, FCSR);
|
| __ jr(ra);
|
| __ nop();
|
| @@ -2413,12 +2411,12 @@ TEST(cvt_w_d) {
|
| int fcsr_inputs[4] =
|
| {kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf};
|
| double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM};
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lw(t0, MemOperand(a0, offsetof(Test, fcsr)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lw(t0, MemOperand(a0, offsetof(Test, fcsr)));
|
| __ cfc1(t1, FCSR);
|
| __ ctc1(t0, FCSR);
|
| __ cvt_w_d(f8, f4);
|
| - __ swc1(f8, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(Test, b)));
|
| __ ctc1(t1, FCSR);
|
| __ jr(ra);
|
| __ nop();
|
| @@ -2480,13 +2478,13 @@ TEST(trunc_w) {
|
| kFPUInvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ trunc_w_d(f8, f4);
|
| __ trunc_w_s(f10, f6);
|
| - __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -2549,13 +2547,13 @@ TEST(round_w) {
|
| kFPUInvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ round_w_d(f8, f4);
|
| __ round_w_s(f10, f6);
|
| - __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -2620,13 +2618,13 @@ TEST(round_l) {
|
| dFPU64InvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ round_l_d(f8, f4);
|
| __ round_l_s(f10, f6);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -2691,14 +2689,14 @@ TEST(sub) {
|
| 0.5, -0.5, 0.0, -10.1, -10.1, -5.8,
|
| 10.1, 10.1, 5.8, -0.5, 0.5, 0.0
|
| };
|
| - __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
|
| - __ ldc1(f10, MemOperand(a0, offsetof(TestFloat, d)) );
|
| + __ Lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
| + __ Ldc1(f10, MemOperand(a0, offsetof(TestFloat, d)));
|
| __ sub_s(f6, f2, f4);
|
| __ sub_d(f12, f8, f10);
|
| - __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
|
| + __ Swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)));
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -2757,21 +2755,20 @@ TEST(sqrt_rsqrt_recip) {
|
| 0.0, 2.0, sqrt2_s, 2e-14
|
| };
|
|
|
| -
|
| - __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
|
| + __ Lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
| __ sqrt_s(f6, f2);
|
| __ sqrt_d(f12, f8);
|
| __ rsqrt_d(f14, f8);
|
| __ rsqrt_s(f16, f2);
|
| __ recip_d(f18, f8);
|
| __ recip_s(f4, f2);
|
| - __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
|
| - __ swc1(f16, MemOperand(a0, offsetof(TestFloat, resultS1)) );
|
| - __ sdc1(f14, MemOperand(a0, offsetof(TestFloat, resultD1)) );
|
| - __ swc1(f4, MemOperand(a0, offsetof(TestFloat, resultS2)) );
|
| - __ sdc1(f18, MemOperand(a0, offsetof(TestFloat, resultD2)) );
|
| + __ Swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)));
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
| + __ Swc1(f16, MemOperand(a0, offsetof(TestFloat, resultS1)));
|
| + __ Sdc1(f14, MemOperand(a0, offsetof(TestFloat, resultD1)));
|
| + __ Swc1(f4, MemOperand(a0, offsetof(TestFloat, resultS2)));
|
| + __ Sdc1(f18, MemOperand(a0, offsetof(TestFloat, resultD2)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -2844,12 +2841,12 @@ TEST(neg) {
|
| float outputs_S[kTableLength] = {
|
| -4.0, 2.0
|
| };
|
| - __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)) );
|
| + __ Lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c)));
|
| __ neg_s(f6, f2);
|
| __ neg_d(f12, f8);
|
| - __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
|
| + __ Swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)));
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -2900,14 +2897,14 @@ TEST(mul) {
|
| 4.8, 4.8, -4.8, -0.29
|
| };
|
|
|
| - __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) );
|
| - __ ldc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, d)) );
|
| + __ Lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, c)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, d)));
|
| __ mul_s(f10, f2, f4);
|
| __ mul_d(f12, f6, f8);
|
| - __ swc1(f10, MemOperand(a0, offsetof(TestFloat, resultS)) );
|
| - __ sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)) );
|
| + __ Swc1(f10, MemOperand(a0, offsetof(TestFloat, resultS)));
|
| + __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -2957,12 +2954,12 @@ TEST(mov) {
|
| 5.3, -5.3, 5.3, -2.9
|
| };
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)));
|
| __ mov_s(f8, f6);
|
| __ mov_d(f10, f4);
|
| - __ swc1(f8, MemOperand(a0, offsetof(TestFloat, d)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, b)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(TestFloat, d)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, b)));
|
| __ jr(ra);
|
| __ nop();
|
|
|
| @@ -3023,13 +3020,13 @@ TEST(floor_w) {
|
| kFPUInvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ floor_w_d(f8, f4);
|
| __ floor_w_s(f10, f6);
|
| - __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -3094,13 +3091,13 @@ TEST(floor_l) {
|
| dFPU64InvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ floor_l_d(f8, f4);
|
| __ floor_l_s(f10, f6);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -3165,13 +3162,13 @@ TEST(ceil_w) {
|
| kFPUInvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ ceil_w_d(f8, f4);
|
| __ ceil_w_s(f10, f6);
|
| - __ swc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ swc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Swc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -3236,13 +3233,13 @@ TEST(ceil_l) {
|
| dFPU64InvalidResult};
|
|
|
| __ cfc1(t1, FCSR);
|
| - __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, a)) );
|
| - __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) );
|
| + __ Sw(t1, MemOperand(a0, offsetof(Test, isNaN2008)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, a)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(Test, b)));
|
| __ ceil_l_d(f8, f4);
|
| __ ceil_l_s(f10, f6);
|
| - __ sdc1(f8, MemOperand(a0, offsetof(Test, c)) );
|
| - __ sdc1(f10, MemOperand(a0, offsetof(Test, d)) );
|
| + __ Sdc1(f8, MemOperand(a0, offsetof(Test, c)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(Test, d)));
|
| __ jr(ra);
|
| __ nop();
|
| Test test;
|
| @@ -3271,7 +3268,7 @@ TEST(jump_tables1) {
|
| CcTest::InitializeVM();
|
| Isolate* isolate = CcTest::i_isolate();
|
| HandleScope scope(isolate);
|
| - Assembler assm(isolate, nullptr, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
|
|
| const int kNumCases = 512;
|
| int values[kNumCases];
|
| @@ -3279,7 +3276,7 @@ TEST(jump_tables1) {
|
| Label labels[kNumCases];
|
|
|
| __ daddiu(sp, sp, -8);
|
| - __ sd(ra, MemOperand(sp));
|
| + __ Sd(ra, MemOperand(sp));
|
| __ Align(8);
|
|
|
| Label done;
|
| @@ -3293,7 +3290,7 @@ TEST(jump_tables1) {
|
| __ dsll(at, a0, 3); // In delay slot.
|
| __ bind(&here);
|
| __ daddu(at, at, ra);
|
| - __ ld(at, MemOperand(at, 4 * Assembler::kInstrSize));
|
| + __ Ld(at, MemOperand(at, 4 * Assembler::kInstrSize));
|
| __ jr(at);
|
| __ nop();
|
| for (int i = 0; i < kNumCases; ++i) {
|
| @@ -3310,7 +3307,7 @@ TEST(jump_tables1) {
|
| }
|
|
|
| __ bind(&done);
|
| - __ ld(ra, MemOperand(sp));
|
| + __ Ld(ra, MemOperand(sp));
|
| __ daddiu(sp, sp, 8);
|
| __ jr(ra);
|
| __ nop();
|
| @@ -3339,7 +3336,7 @@ TEST(jump_tables2) {
|
| CcTest::InitializeVM();
|
| Isolate* isolate = CcTest::i_isolate();
|
| HandleScope scope(isolate);
|
| - Assembler assm(isolate, nullptr, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
|
|
| const int kNumCases = 512;
|
| int values[kNumCases];
|
| @@ -3347,7 +3344,7 @@ TEST(jump_tables2) {
|
| Label labels[kNumCases];
|
|
|
| __ daddiu(sp, sp, -8);
|
| - __ sd(ra, MemOperand(sp));
|
| + __ Sd(ra, MemOperand(sp));
|
|
|
| Label done, dispatch;
|
| __ b(&dispatch);
|
| @@ -3373,7 +3370,7 @@ TEST(jump_tables2) {
|
| __ dsll(at, a0, 3); // In delay slot.
|
| __ bind(&here);
|
| __ daddu(at, at, ra);
|
| - __ ld(at, MemOperand(at, 4 * Assembler::kInstrSize));
|
| + __ Ld(at, MemOperand(at, 4 * Assembler::kInstrSize));
|
| __ jr(at);
|
| __ nop();
|
| for (int i = 0; i < kNumCases; ++i) {
|
| @@ -3382,7 +3379,7 @@ TEST(jump_tables2) {
|
| }
|
|
|
| __ bind(&done);
|
| - __ ld(ra, MemOperand(sp));
|
| + __ Ld(ra, MemOperand(sp));
|
| __ daddiu(sp, sp, 8);
|
| __ jr(ra);
|
| __ nop();
|
| @@ -3409,7 +3406,7 @@ TEST(jump_tables3) {
|
| CcTest::InitializeVM();
|
| Isolate* isolate = CcTest::i_isolate();
|
| HandleScope scope(isolate);
|
| - Assembler assm(isolate, nullptr, 0);
|
| + MacroAssembler assm(isolate, NULL, 0, v8::internal::CodeObjectRequired::kYes);
|
|
|
| const int kNumCases = 512;
|
| Handle<Object> values[kNumCases];
|
| @@ -3422,7 +3419,7 @@ TEST(jump_tables3) {
|
| int64_t imm64;
|
|
|
| __ daddiu(sp, sp, -8);
|
| - __ sd(ra, MemOperand(sp));
|
| + __ Sd(ra, MemOperand(sp));
|
|
|
| Label done, dispatch;
|
| __ b(&dispatch);
|
| @@ -3453,7 +3450,7 @@ TEST(jump_tables3) {
|
| __ dsll(at, a0, 3); // In delay slot.
|
| __ bind(&here);
|
| __ daddu(at, at, ra);
|
| - __ ld(at, MemOperand(at, 4 * Assembler::kInstrSize));
|
| + __ Ld(at, MemOperand(at, 4 * Assembler::kInstrSize));
|
| __ jr(at);
|
| __ nop();
|
| for (int i = 0; i < kNumCases; ++i) {
|
| @@ -3462,7 +3459,7 @@ TEST(jump_tables3) {
|
| }
|
|
|
| __ bind(&done);
|
| - __ ld(ra, MemOperand(sp));
|
| + __ Ld(ra, MemOperand(sp));
|
| __ daddiu(sp, sp, 8);
|
| __ jr(ra);
|
| __ nop();
|
| @@ -3505,37 +3502,38 @@ TEST(BITSWAP) {
|
| } T;
|
| T t;
|
|
|
| - Assembler assm(isolate, NULL, 0);
|
| + MacroAssembler assm(isolate, NULL, 0,
|
| + v8::internal::CodeObjectRequired::kYes);
|
|
|
| - __ ld(a4, MemOperand(a0, offsetof(T, r1)));
|
| + __ Ld(a4, MemOperand(a0, offsetof(T, r1)));
|
| __ nop();
|
| __ bitswap(a6, a4);
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r1)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r1)));
|
|
|
| - __ ld(a4, MemOperand(a0, offsetof(T, r2)));
|
| + __ Ld(a4, MemOperand(a0, offsetof(T, r2)));
|
| __ nop();
|
| __ bitswap(a6, a4);
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r2)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r2)));
|
|
|
| - __ ld(a4, MemOperand(a0, offsetof(T, r3)));
|
| + __ Ld(a4, MemOperand(a0, offsetof(T, r3)));
|
| __ nop();
|
| __ bitswap(a6, a4);
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r3)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r3)));
|
|
|
| - __ ld(a4, MemOperand(a0, offsetof(T, r4)));
|
| + __ Ld(a4, MemOperand(a0, offsetof(T, r4)));
|
| __ nop();
|
| __ bitswap(a6, a4);
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r4)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r4)));
|
|
|
| - __ ld(a4, MemOperand(a0, offsetof(T, r5)));
|
| + __ Ld(a4, MemOperand(a0, offsetof(T, r5)));
|
| __ nop();
|
| __ dbitswap(a6, a4);
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r5)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r5)));
|
|
|
| - __ ld(a4, MemOperand(a0, offsetof(T, r6)));
|
| + __ Ld(a4, MemOperand(a0, offsetof(T, r6)));
|
| __ nop();
|
| __ dbitswap(a6, a4);
|
| - __ sd(a6, MemOperand(a0, offsetof(T, r6)));
|
| + __ Sd(a6, MemOperand(a0, offsetof(T, r6)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -3599,86 +3597,86 @@ TEST(class_fmt) {
|
| MacroAssembler assm(isolate, NULL, 0,
|
| v8::internal::CodeObjectRequired::kYes);
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dSignalingNan)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dSignalingNan)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dSignalingNan)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dSignalingNan)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dQuietNan)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dQuietNan)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dQuietNan)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dQuietNan)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dNegInf)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegInf)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dNegInf)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegInf)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dNegNorm)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegNorm)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dNegNorm)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegNorm)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dNegSubnorm)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegSubnorm)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dNegSubnorm)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegSubnorm)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dNegZero)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegZero)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dNegZero)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegZero)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dPosInf)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosInf)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dPosInf)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosInf)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dPosNorm)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosNorm)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dPosNorm)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosNorm)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dPosSubnorm)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosSubnorm)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dPosSubnorm)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosSubnorm)));
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(T, dPosZero)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosZero)));
|
| __ class_d(f6, f4);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(T, dPosZero)));
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosZero)));
|
|
|
| // Testing instruction CLASS.S
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fSignalingNan)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fSignalingNan)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fSignalingNan)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fSignalingNan)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fQuietNan)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fQuietNan)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fQuietNan)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fQuietNan)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fNegInf)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fNegInf)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fNegInf)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fNegInf)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fNegNorm)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fNegNorm)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fNegNorm)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fNegNorm)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fNegSubnorm)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fNegSubnorm)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fNegSubnorm)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fNegSubnorm)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fNegZero)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fNegZero)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fNegZero)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fNegZero)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fPosInf)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fPosInf)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fPosInf)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fPosInf)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fPosNorm)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fPosNorm)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fPosNorm)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fPosNorm)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fPosSubnorm)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fPosSubnorm)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fPosSubnorm)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fPosSubnorm)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(T, fPosZero)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(T, fPosZero)));
|
| __ class_s(f6, f4);
|
| - __ swc1(f6, MemOperand(a0, offsetof(T, fPosZero)));
|
| + __ Swc1(f6, MemOperand(a0, offsetof(T, fPosZero)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -3759,17 +3757,17 @@ TEST(ABS) {
|
|
|
| // Save FIR.
|
| __ cfc1(a1, FCSR);
|
| - __ sd(a1, MemOperand(a0, offsetof(TestFloat, fcsr)));
|
| + __ Sd(a1, MemOperand(a0, offsetof(TestFloat, fcsr)));
|
| // Disable FPU exceptions.
|
| __ ctc1(zero_reg, FCSR);
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| __ abs_d(f10, f4);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, a)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)));
|
| __ abs_s(f10, f4);
|
| - __ swc1(f10, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(TestFloat, b)));
|
|
|
| // Restore FCSR.
|
| __ ctc1(a1, FCSR);
|
| @@ -3857,15 +3855,15 @@ TEST(ADD_FMT) {
|
|
|
| TestFloat test;
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, b)));
|
| __ add_d(f10, f8, f4);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, c)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, fa)));
|
| - __ lwc1(f8, MemOperand(a0, offsetof(TestFloat, fb)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestFloat, fa)));
|
| + __ Lwc1(f8, MemOperand(a0, offsetof(TestFloat, fb)));
|
| __ add_s(f10, f8, f4);
|
| - __ swc1(f10, MemOperand(a0, offsetof(TestFloat, fc)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(TestFloat, fc)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -3944,11 +3942,11 @@ TEST(C_COND_FMT) {
|
|
|
| __ li(t1, 1);
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
|
| - __ ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
|
|
|
| - __ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
|
| - __ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
|
| + __ Lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
|
| + __ Lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -3956,8 +3954,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(F, f14, f16, 2);
|
| __ movt(t2, t1, 0);
|
| __ movt(t3, t1, 2);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dF)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fF)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dF)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fF)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -3965,8 +3963,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(UN, f14, f16, 4);
|
| __ movt(t2, t1, 2);
|
| __ movt(t3, t1, 4);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUn)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUn)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dUn)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fUn)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -3974,8 +3972,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(EQ, f14, f16, 6);
|
| __ movt(t2, t1, 4);
|
| __ movt(t3, t1, 6);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dEq)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fEq)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dEq)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fEq)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -3983,8 +3981,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(UEQ, f14, f16, 0);
|
| __ movt(t2, t1, 6);
|
| __ movt(t3, t1, 0);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUeq)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUeq)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dUeq)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fUeq)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -3992,8 +3990,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(OLT, f14, f16, 2);
|
| __ movt(t2, t1, 0);
|
| __ movt(t3, t1, 2);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dOlt)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fOlt)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dOlt)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fOlt)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -4001,8 +3999,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(ULT, f14, f16, 4);
|
| __ movt(t2, t1, 2);
|
| __ movt(t3, t1, 4);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUlt)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUlt)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dUlt)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fUlt)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -4010,8 +4008,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(OLE, f14, f16, 6);
|
| __ movt(t2, t1, 4);
|
| __ movt(t3, t1, 6);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dOle)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fOle)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dOle)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fOle)));
|
|
|
| __ mov(t2, zero_reg);
|
| __ mov(t3, zero_reg);
|
| @@ -4019,8 +4017,8 @@ TEST(C_COND_FMT) {
|
| __ c_s(ULE, f14, f16, 0);
|
| __ movt(t2, t1, 6);
|
| __ movt(t3, t1, 0);
|
| - __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUle)) );
|
| - __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUle)) );
|
| + __ Sw(t2, MemOperand(a0, offsetof(TestFloat, dUle)));
|
| + __ Sw(t3, MemOperand(a0, offsetof(TestFloat, fUle)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -4162,66 +4160,66 @@ TEST(CMP_COND_FMT) {
|
|
|
| __ li(t1, 1);
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
|
| - __ ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2)));
|
|
|
| - __ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
|
| - __ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
|
| + __ Lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1)));
|
| + __ Lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2)));
|
|
|
| __ cmp_d(F, f2, f4, f6);
|
| __ cmp_s(F, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dF)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fF)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dF)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fF)));
|
|
|
| __ cmp_d(UN, f2, f4, f6);
|
| __ cmp_s(UN, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUn)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUn)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUn)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fUn)));
|
|
|
| __ cmp_d(EQ, f2, f4, f6);
|
| __ cmp_s(EQ, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dEq)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fEq)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dEq)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fEq)));
|
|
|
| __ cmp_d(UEQ, f2, f4, f6);
|
| __ cmp_s(UEQ, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUeq)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUeq)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUeq)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fUeq)));
|
|
|
| __ cmp_d(LT, f2, f4, f6);
|
| __ cmp_s(LT, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOlt)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOlt)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOlt)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fOlt)));
|
|
|
| __ cmp_d(ULT, f2, f4, f6);
|
| __ cmp_s(ULT, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUlt)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUlt)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUlt)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fUlt)));
|
|
|
| __ cmp_d(LE, f2, f4, f6);
|
| __ cmp_s(LE, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOle)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOle)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOle)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fOle)));
|
|
|
| __ cmp_d(ULE, f2, f4, f6);
|
| __ cmp_s(ULE, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUle)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUle)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUle)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fUle)));
|
|
|
| __ cmp_d(ORD, f2, f4, f6);
|
| __ cmp_s(ORD, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOr)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOr)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOr)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fOr)));
|
|
|
| __ cmp_d(UNE, f2, f4, f6);
|
| __ cmp_s(UNE, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUne)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUne)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUne)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fUne)));
|
|
|
| __ cmp_d(NE, f2, f4, f6);
|
| __ cmp_s(NE, f12, f14, f16);
|
| - __ sdc1(f2, MemOperand(a0, offsetof(TestFloat, dNe)) );
|
| - __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fNe)) );
|
| + __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dNe)));
|
| + __ Swc1(f12, MemOperand(a0, offsetof(TestFloat, fNe)));
|
|
|
| __ jr(ra);
|
| __ nop();
|
| @@ -4558,19 +4556,19 @@ TEST(DIV_FMT) {
|
| // Disable FPU exceptions.
|
| __ ctc1(zero_reg, FCSR);
|
|
|
| - __ ldc1(f4, MemOperand(a0, offsetof(Test, dOp1)) );
|
| - __ ldc1(f2, MemOperand(a0, offsetof(Test, dOp2)) );
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(Test, dOp1)));
|
| + __ Ldc1(f2, MemOperand(a0, offsetof(Test, dOp2)));
|
| __ nop();
|
| __ div_d(f6, f4, f2);
|
| - __ sdc1(f6, MemOperand(a0, offsetof(Test, dRes)) );
|
| + __ Sdc1(f6, MemOperand(a0, offsetof(Test, dRes)));
|
|
|
| - __ lwc1(f4, MemOperand(a0, offsetof(Test, fOp1)) );
|
| - __ lwc1(f2, MemOperand(a0, offsetof(Test, fOp2)) );
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(Test, fOp1)));
|
| + __ Lwc1(f2, MemOperand(a0, offsetof(Test, fOp2)));
|
| __ nop();
|
| __ div_s(f6, f4, f2);
|
| - __ swc1(f6, MemOperand(a0, offsetof(Test, fRes)) );
|
| + __ Swc1(f6, MemOperand(a0, offsetof(Test, fRes)));
|
|
|
| - // Restore FCSR.
|
| + // Restore FCSR.
|
| __ ctc1(a1, FCSR);
|
|
|
| __ jr(ra);
|
| @@ -5971,15 +5969,15 @@ void helper_madd_msub_maddf_msubf(F func) {
|
| };
|
|
|
| if (std::is_same<T, float>::value) {
|
| - __ lwc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| - __ lwc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs)));
|
| - __ lwc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft)));
|
| - __ lwc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| + __ Lwc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| + __ Lwc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs)));
|
| + __ Lwc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft)));
|
| + __ Lwc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| } else if (std::is_same<T, double>::value) {
|
| - __ ldc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| - __ ldc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs)));
|
| - __ ldc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft)));
|
| - __ ldc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| + __ Ldc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| + __ Ldc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs)));
|
| + __ Ldc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft)));
|
| + __ Ldc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr)));
|
| } else {
|
| UNREACHABLE();
|
| }
|
| @@ -6023,9 +6021,9 @@ TEST(madd_msub_s) {
|
| if (kArchVariant == kMips64r6) return;
|
| helper_madd_msub_maddf_msubf<float>([](MacroAssembler& assm) {
|
| __ madd_s(f10, f4, f6, f8);
|
| - __ swc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add)));
|
| + __ Swc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add)));
|
| __ msub_s(f16, f4, f6, f8);
|
| - __ swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub)));
|
| + __ Swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub)));
|
| });
|
| }
|
|
|
| @@ -6033,9 +6031,9 @@ TEST(madd_msub_d) {
|
| if (kArchVariant == kMips64r6) return;
|
| helper_madd_msub_maddf_msubf<double>([](MacroAssembler& assm) {
|
| __ madd_d(f10, f4, f6, f8);
|
| - __ sdc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add)));
|
| + __ Sdc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add)));
|
| __ msub_d(f16, f4, f6, f8);
|
| - __ sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub)));
|
| + __ Sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub)));
|
| });
|
| }
|
|
|
| @@ -6043,9 +6041,9 @@ TEST(maddf_msubf_s) {
|
| if (kArchVariant != kMips64r6) return;
|
| helper_madd_msub_maddf_msubf<float>([](MacroAssembler& assm) {
|
| __ maddf_s(f4, f6, f8);
|
| - __ swc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add)));
|
| + __ Swc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add)));
|
| __ msubf_s(f16, f6, f8);
|
| - __ swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub)));
|
| + __ Swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub)));
|
| });
|
| }
|
|
|
| @@ -6053,9 +6051,9 @@ TEST(maddf_msubf_d) {
|
| if (kArchVariant != kMips64r6) return;
|
| helper_madd_msub_maddf_msubf<double>([](MacroAssembler& assm) {
|
| __ maddf_d(f4, f6, f8);
|
| - __ sdc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add)));
|
| + __ Sdc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add)));
|
| __ msubf_d(f16, f6, f8);
|
| - __ sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub)));
|
| + __ Sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub)));
|
| });
|
| }
|
|
|
|
|