| Index: test/cctest/test-assembler-arm.cc
|
| diff --git a/test/cctest/test-assembler-arm.cc b/test/cctest/test-assembler-arm.cc
|
| index cb8957799155128143d4548066ebef1037f66715..59ebaab069270fa8626ef4f060a3aea03093685d 100644
|
| --- a/test/cctest/test-assembler-arm.cc
|
| +++ b/test/cctest/test-assembler-arm.cc
|
| @@ -175,17 +175,17 @@ TEST(3) {
|
| __ stm(db_w, sp, r4.bit() | fp.bit() | lr.bit());
|
| __ sub(fp, ip, Operand(4));
|
| __ mov(r4, Operand(r0));
|
| - __ ldr(r0, MemOperand(r4, OFFSET_OF(T, i)));
|
| + __ ldr(r0, MemOperand(r4, offsetof(T, i)));
|
| __ mov(r2, Operand(r0, ASR, 1));
|
| - __ str(r2, MemOperand(r4, OFFSET_OF(T, i)));
|
| - __ ldrsb(r2, MemOperand(r4, OFFSET_OF(T, c)));
|
| + __ str(r2, MemOperand(r4, offsetof(T, i)));
|
| + __ ldrsb(r2, MemOperand(r4, offsetof(T, c)));
|
| __ add(r0, r2, Operand(r0));
|
| __ mov(r2, Operand(r2, LSL, 2));
|
| - __ strb(r2, MemOperand(r4, OFFSET_OF(T, c)));
|
| - __ ldrsh(r2, MemOperand(r4, OFFSET_OF(T, s)));
|
| + __ strb(r2, MemOperand(r4, offsetof(T, c)));
|
| + __ ldrsh(r2, MemOperand(r4, offsetof(T, s)));
|
| __ add(r0, r2, Operand(r0));
|
| __ mov(r2, Operand(r2, ASR, 3));
|
| - __ strh(r2, MemOperand(r4, OFFSET_OF(T, s)));
|
| + __ strh(r2, MemOperand(r4, offsetof(T, s)));
|
| __ ldm(ia_w, sp, r4.bit() | fp.bit() | pc.bit());
|
|
|
| CodeDesc desc;
|
| @@ -247,68 +247,68 @@ TEST(4) {
|
| __ sub(fp, ip, Operand(4));
|
|
|
| __ mov(r4, Operand(r0));
|
| - __ vldr(d6, r4, OFFSET_OF(T, a));
|
| - __ vldr(d7, r4, OFFSET_OF(T, b));
|
| + __ vldr(d6, r4, offsetof(T, a));
|
| + __ vldr(d7, r4, offsetof(T, b));
|
| __ vadd(d5, d6, d7);
|
| - __ vstr(d5, r4, OFFSET_OF(T, c));
|
| + __ vstr(d5, r4, offsetof(T, c));
|
|
|
| __ vmla(d5, d6, d7);
|
| __ vmls(d5, d5, d6);
|
|
|
| __ vmov(r2, r3, d5);
|
| __ vmov(d4, r2, r3);
|
| - __ vstr(d4, r4, OFFSET_OF(T, b));
|
| + __ vstr(d4, r4, offsetof(T, b));
|
|
|
| // Load t.x and t.y, switch values, and store back to the struct.
|
| - __ vldr(s0, r4, OFFSET_OF(T, x));
|
| - __ vldr(s31, r4, OFFSET_OF(T, y));
|
| + __ vldr(s0, r4, offsetof(T, x));
|
| + __ vldr(s31, r4, offsetof(T, y));
|
| __ vmov(s16, s0);
|
| __ vmov(s0, s31);
|
| __ vmov(s31, s16);
|
| - __ vstr(s0, r4, OFFSET_OF(T, x));
|
| - __ vstr(s31, r4, OFFSET_OF(T, y));
|
| + __ vstr(s0, r4, offsetof(T, x));
|
| + __ vstr(s31, r4, offsetof(T, y));
|
|
|
| // Move a literal into a register that can be encoded in the instruction.
|
| __ vmov(d4, 1.0);
|
| - __ vstr(d4, r4, OFFSET_OF(T, e));
|
| + __ vstr(d4, r4, offsetof(T, e));
|
|
|
| // Move a literal into a register that requires 64 bits to encode.
|
| // 0x3ff0000010000000 = 1.000000059604644775390625
|
| __ vmov(d4, 1.000000059604644775390625);
|
| - __ vstr(d4, r4, OFFSET_OF(T, d));
|
| + __ vstr(d4, r4, offsetof(T, d));
|
|
|
| // Convert from floating point to integer.
|
| __ vmov(d4, 2.0);
|
| __ vcvt_s32_f64(s31, d4);
|
| - __ vstr(s31, r4, OFFSET_OF(T, i));
|
| + __ vstr(s31, r4, offsetof(T, i));
|
|
|
| // Convert from integer to floating point.
|
| __ mov(lr, Operand(42));
|
| __ vmov(s31, lr);
|
| __ vcvt_f64_s32(d4, s31);
|
| - __ vstr(d4, r4, OFFSET_OF(T, f));
|
| + __ vstr(d4, r4, offsetof(T, f));
|
|
|
| // Convert from fixed point to floating point.
|
| __ mov(lr, Operand(2468));
|
| __ vmov(s8, lr);
|
| __ vcvt_f64_s32(d4, 2);
|
| - __ vstr(d4, r4, OFFSET_OF(T, j));
|
| + __ vstr(d4, r4, offsetof(T, j));
|
|
|
| // Test vabs.
|
| - __ vldr(d1, r4, OFFSET_OF(T, g));
|
| + __ vldr(d1, r4, offsetof(T, g));
|
| __ vabs(d0, d1);
|
| - __ vstr(d0, r4, OFFSET_OF(T, g));
|
| - __ vldr(d2, r4, OFFSET_OF(T, h));
|
| + __ vstr(d0, r4, offsetof(T, g));
|
| + __ vldr(d2, r4, offsetof(T, h));
|
| __ vabs(d0, d2);
|
| - __ vstr(d0, r4, OFFSET_OF(T, h));
|
| + __ vstr(d0, r4, offsetof(T, h));
|
|
|
| // Test vneg.
|
| - __ vldr(d1, r4, OFFSET_OF(T, m));
|
| + __ vldr(d1, r4, offsetof(T, m));
|
| __ vneg(d0, d1);
|
| - __ vstr(d0, r4, OFFSET_OF(T, m));
|
| - __ vldr(d1, r4, OFFSET_OF(T, n));
|
| + __ vstr(d0, r4, offsetof(T, m));
|
| + __ vldr(d1, r4, offsetof(T, n));
|
| __ vneg(d0, d1);
|
| - __ vstr(d0, r4, OFFSET_OF(T, n));
|
| + __ vstr(d0, r4, offsetof(T, n));
|
|
|
| __ ldm(ia_w, sp, r4.bit() | fp.bit() | pc.bit());
|
|
|
| @@ -647,19 +647,19 @@ TEST(8) {
|
| __ stm(db_w, sp, r4.bit() | fp.bit() | lr.bit());
|
| __ sub(fp, ip, Operand(4));
|
|
|
| - __ add(r4, r0, Operand(OFFSET_OF(D, a)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(D, a))));
|
| __ vldm(ia_w, r4, d0, d3);
|
| __ vldm(ia_w, r4, d4, d7);
|
|
|
| - __ add(r4, r0, Operand(OFFSET_OF(D, a)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(D, a))));
|
| __ vstm(ia_w, r4, d6, d7);
|
| __ vstm(ia_w, r4, d0, d5);
|
|
|
| - __ add(r4, r1, Operand(OFFSET_OF(F, a)));
|
| + __ add(r4, r1, Operand(static_cast<int32_t>(offsetof(F, a))));
|
| __ vldm(ia_w, r4, s0, s3);
|
| __ vldm(ia_w, r4, s4, s7);
|
|
|
| - __ add(r4, r1, Operand(OFFSET_OF(F, a)));
|
| + __ add(r4, r1, Operand(static_cast<int32_t>(offsetof(F, a))));
|
| __ vstm(ia_w, r4, s6, s7);
|
| __ vstm(ia_w, r4, s0, s5);
|
|
|
| @@ -753,22 +753,22 @@ TEST(9) {
|
| __ stm(db_w, sp, r4.bit() | fp.bit() | lr.bit());
|
| __ sub(fp, ip, Operand(4));
|
|
|
| - __ add(r4, r0, Operand(OFFSET_OF(D, a)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(D, a))));
|
| __ vldm(ia, r4, d0, d3);
|
| __ add(r4, r4, Operand(4 * 8));
|
| __ vldm(ia, r4, d4, d7);
|
|
|
| - __ add(r4, r0, Operand(OFFSET_OF(D, a)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(D, a))));
|
| __ vstm(ia, r4, d6, d7);
|
| __ add(r4, r4, Operand(2 * 8));
|
| __ vstm(ia, r4, d0, d5);
|
|
|
| - __ add(r4, r1, Operand(OFFSET_OF(F, a)));
|
| + __ add(r4, r1, Operand(static_cast<int32_t>(offsetof(F, a))));
|
| __ vldm(ia, r4, s0, s3);
|
| __ add(r4, r4, Operand(4 * 4));
|
| __ vldm(ia, r4, s4, s7);
|
|
|
| - __ add(r4, r1, Operand(OFFSET_OF(F, a)));
|
| + __ add(r4, r1, Operand(static_cast<int32_t>(offsetof(F, a))));
|
| __ vstm(ia, r4, s6, s7);
|
| __ add(r4, r4, Operand(2 * 4));
|
| __ vstm(ia, r4, s0, s5);
|
| @@ -863,19 +863,19 @@ TEST(10) {
|
| __ stm(db_w, sp, r4.bit() | fp.bit() | lr.bit());
|
| __ sub(fp, ip, Operand(4));
|
|
|
| - __ add(r4, r0, Operand(OFFSET_OF(D, h) + 8));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(D, h)) + 8));
|
| __ vldm(db_w, r4, d4, d7);
|
| __ vldm(db_w, r4, d0, d3);
|
|
|
| - __ add(r4, r0, Operand(OFFSET_OF(D, h) + 8));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(D, h)) + 8));
|
| __ vstm(db_w, r4, d0, d5);
|
| __ vstm(db_w, r4, d6, d7);
|
|
|
| - __ add(r4, r1, Operand(OFFSET_OF(F, h) + 4));
|
| + __ add(r4, r1, Operand(static_cast<int32_t>(offsetof(F, h)) + 4));
|
| __ vldm(db_w, r4, s4, s7);
|
| __ vldm(db_w, r4, s0, s3);
|
|
|
| - __ add(r4, r1, Operand(OFFSET_OF(F, h) + 4));
|
| + __ add(r4, r1, Operand(static_cast<int32_t>(offsetof(F, h)) + 4));
|
| __ vstm(db_w, r4, s0, s5);
|
| __ vstm(db_w, r4, s6, s7);
|
|
|
| @@ -951,28 +951,28 @@ TEST(11) {
|
| Assembler assm(isolate, NULL, 0);
|
|
|
| // Test HeapObject untagging.
|
| - __ ldr(r1, MemOperand(r0, OFFSET_OF(I, a)));
|
| + __ ldr(r1, MemOperand(r0, offsetof(I, a)));
|
| __ mov(r1, Operand(r1, ASR, 1), SetCC);
|
| __ adc(r1, r1, Operand(r1), LeaveCC, cs);
|
| - __ str(r1, MemOperand(r0, OFFSET_OF(I, a)));
|
| + __ str(r1, MemOperand(r0, offsetof(I, a)));
|
|
|
| - __ ldr(r2, MemOperand(r0, OFFSET_OF(I, b)));
|
| + __ ldr(r2, MemOperand(r0, offsetof(I, b)));
|
| __ mov(r2, Operand(r2, ASR, 1), SetCC);
|
| __ adc(r2, r2, Operand(r2), LeaveCC, cs);
|
| - __ str(r2, MemOperand(r0, OFFSET_OF(I, b)));
|
| + __ str(r2, MemOperand(r0, offsetof(I, b)));
|
|
|
| // Test corner cases.
|
| __ mov(r1, Operand(0xffffffff));
|
| __ mov(r2, Operand::Zero());
|
| __ mov(r3, Operand(r1, ASR, 1), SetCC); // Set the carry.
|
| __ adc(r3, r1, Operand(r2));
|
| - __ str(r3, MemOperand(r0, OFFSET_OF(I, c)));
|
| + __ str(r3, MemOperand(r0, offsetof(I, c)));
|
|
|
| __ mov(r1, Operand(0xffffffff));
|
| __ mov(r2, Operand::Zero());
|
| __ mov(r3, Operand(r2, ASR, 1), SetCC); // Unset the carry.
|
| __ adc(r3, r1, Operand(r2));
|
| - __ str(r3, MemOperand(r0, OFFSET_OF(I, d)));
|
| + __ str(r3, MemOperand(r0, offsetof(I, d)));
|
|
|
| __ mov(pc, Operand(lr));
|
|
|
| @@ -1048,9 +1048,9 @@ TEST(13) {
|
|
|
| // Load a, b, c into d16, d17, d18.
|
| __ mov(r4, Operand(r0));
|
| - __ vldr(d16, r4, OFFSET_OF(T, a));
|
| - __ vldr(d17, r4, OFFSET_OF(T, b));
|
| - __ vldr(d18, r4, OFFSET_OF(T, c));
|
| + __ vldr(d16, r4, offsetof(T, a));
|
| + __ vldr(d17, r4, offsetof(T, b));
|
| + __ vldr(d18, r4, offsetof(T, c));
|
|
|
| __ vneg(d25, d16);
|
| __ vadd(d25, d25, d17);
|
| @@ -1066,12 +1066,12 @@ TEST(13) {
|
|
|
| // Store d16, d17, d18 into a, b, c.
|
| __ mov(r4, Operand(r0));
|
| - __ vstr(d16, r4, OFFSET_OF(T, a));
|
| - __ vstr(d17, r4, OFFSET_OF(T, b));
|
| - __ vstr(d18, r4, OFFSET_OF(T, c));
|
| + __ vstr(d16, r4, offsetof(T, a));
|
| + __ vstr(d17, r4, offsetof(T, b));
|
| + __ vstr(d18, r4, offsetof(T, c));
|
|
|
| // Load x, y, z into d29-d31.
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, x)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, x))));
|
| __ vldm(ia_w, r4, d29, d31);
|
|
|
| // Swap d29 and d30 via r registers.
|
| @@ -1084,7 +1084,7 @@ TEST(13) {
|
| __ vcvt_f64_u32(d31, s1);
|
|
|
| // Store d29-d31 into x, y, z.
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, x)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, x))));
|
| __ vstm(ia_w, r4, d29, d31);
|
|
|
| // Move constants into d20, d21, d22 and store into i, j, k.
|
| @@ -1094,13 +1094,13 @@ TEST(13) {
|
| __ mov(r2, Operand(1079146608));
|
| __ vmov(d22, VmovIndexLo, r1);
|
| __ vmov(d22, VmovIndexHi, r2);
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, i)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i))));
|
| __ vstm(ia_w, r4, d20, d22);
|
| // Move d22 into low and high.
|
| __ vmov(r4, VmovIndexLo, d22);
|
| - __ str(r4, MemOperand(r0, OFFSET_OF(T, low)));
|
| + __ str(r4, MemOperand(r0, offsetof(T, low)));
|
| __ vmov(r4, VmovIndexHi, d22);
|
| - __ str(r4, MemOperand(r0, OFFSET_OF(T, high)));
|
| + __ str(r4, MemOperand(r0, offsetof(T, high)));
|
|
|
| __ ldm(ia_w, sp, r4.bit() | pc.bit());
|
|
|
| @@ -1164,16 +1164,16 @@ TEST(14) {
|
| __ vmsr(r1);
|
| __ bind(&fpscr_done);
|
|
|
| - __ vldr(d0, r0, OFFSET_OF(T, left));
|
| - __ vldr(d1, r0, OFFSET_OF(T, right));
|
| + __ vldr(d0, r0, offsetof(T, left));
|
| + __ vldr(d1, r0, offsetof(T, right));
|
| __ vadd(d2, d0, d1);
|
| - __ vstr(d2, r0, OFFSET_OF(T, add_result));
|
| + __ vstr(d2, r0, offsetof(T, add_result));
|
| __ vsub(d2, d0, d1);
|
| - __ vstr(d2, r0, OFFSET_OF(T, sub_result));
|
| + __ vstr(d2, r0, offsetof(T, sub_result));
|
| __ vmul(d2, d0, d1);
|
| - __ vstr(d2, r0, OFFSET_OF(T, mul_result));
|
| + __ vstr(d2, r0, offsetof(T, mul_result));
|
| __ vdiv(d2, d0, d1);
|
| - __ vstr(d2, r0, OFFSET_OF(T, div_result));
|
| + __ vstr(d2, r0, offsetof(T, div_result));
|
|
|
| __ mov(pc, Operand(lr));
|
|
|
| @@ -1264,23 +1264,23 @@ TEST(15) {
|
|
|
| __ stm(db_w, sp, r4.bit() | lr.bit());
|
| // Move 32 bytes with neon.
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, src0)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, src0))));
|
| __ vld1(Neon8, NeonListOperand(d0, 4), NeonMemOperand(r4));
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, dst0)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, dst0))));
|
| __ vst1(Neon8, NeonListOperand(d0, 4), NeonMemOperand(r4));
|
|
|
| // Expand 8 bytes into 8 words(16 bits).
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, srcA0)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, srcA0))));
|
| __ vld1(Neon8, NeonListOperand(d0), NeonMemOperand(r4));
|
| __ vmovl(NeonU8, q0, d0);
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, dstA0)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, dstA0))));
|
| __ vst1(Neon8, NeonListOperand(d0, 2), NeonMemOperand(r4));
|
|
|
| // The same expansion, but with different source and destination registers.
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, srcA0)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, srcA0))));
|
| __ vld1(Neon8, NeonListOperand(d1), NeonMemOperand(r4));
|
| __ vmovl(NeonU8, q1, d1);
|
| - __ add(r4, r0, Operand(OFFSET_OF(T, dstA4)));
|
| + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, dstA4))));
|
| __ vst1(Neon8, NeonListOperand(d2, 2), NeonMemOperand(r4));
|
|
|
| __ ldm(ia_w, sp, r4.bit() | pc.bit());
|
| @@ -1367,24 +1367,24 @@ TEST(16) {
|
| __ stm(db_w, sp, r4.bit() | lr.bit());
|
|
|
| __ mov(r4, Operand(r0));
|
| - __ ldr(r0, MemOperand(r4, OFFSET_OF(T, src0)));
|
| - __ ldr(r1, MemOperand(r4, OFFSET_OF(T, src1)));
|
| + __ ldr(r0, MemOperand(r4, offsetof(T, src0)));
|
| + __ ldr(r1, MemOperand(r4, offsetof(T, src1)));
|
|
|
| __ pkhbt(r2, r0, Operand(r1, LSL, 8));
|
| - __ str(r2, MemOperand(r4, OFFSET_OF(T, dst0)));
|
| + __ str(r2, MemOperand(r4, offsetof(T, dst0)));
|
|
|
| __ pkhtb(r2, r0, Operand(r1, ASR, 8));
|
| - __ str(r2, MemOperand(r4, OFFSET_OF(T, dst1)));
|
| + __ str(r2, MemOperand(r4, offsetof(T, dst1)));
|
|
|
| __ uxtb16(r2, r0, 8);
|
| - __ str(r2, MemOperand(r4, OFFSET_OF(T, dst2)));
|
| + __ str(r2, MemOperand(r4, offsetof(T, dst2)));
|
|
|
| __ uxtb(r2, r0, 8);
|
| - __ str(r2, MemOperand(r4, OFFSET_OF(T, dst3)));
|
| + __ str(r2, MemOperand(r4, offsetof(T, dst3)));
|
|
|
| - __ ldr(r0, MemOperand(r4, OFFSET_OF(T, src2)));
|
| + __ ldr(r0, MemOperand(r4, offsetof(T, src2)));
|
| __ uxtab(r2, r0, r1, 8);
|
| - __ str(r2, MemOperand(r4, OFFSET_OF(T, dst4)));
|
| + __ str(r2, MemOperand(r4, offsetof(T, dst4)));
|
|
|
| __ ldm(ia_w, sp, r4.bit() | pc.bit());
|
|
|
| @@ -1461,11 +1461,11 @@ TEST(sdiv) {
|
|
|
| __ mov(r3, Operand(r0));
|
|
|
| - __ ldr(r0, MemOperand(r3, OFFSET_OF(T, dividend)));
|
| - __ ldr(r1, MemOperand(r3, OFFSET_OF(T, divisor)));
|
| + __ ldr(r0, MemOperand(r3, offsetof(T, dividend)));
|
| + __ ldr(r1, MemOperand(r3, offsetof(T, divisor)));
|
|
|
| __ sdiv(r2, r0, r1);
|
| - __ str(r2, MemOperand(r3, OFFSET_OF(T, result)));
|
| + __ str(r2, MemOperand(r3, offsetof(T, result)));
|
|
|
| __ bx(lr);
|
|
|
| @@ -1525,11 +1525,11 @@ TEST(udiv) {
|
|
|
| __ mov(r3, Operand(r0));
|
|
|
| - __ ldr(r0, MemOperand(r3, OFFSET_OF(T, dividend)));
|
| - __ ldr(r1, MemOperand(r3, OFFSET_OF(T, divisor)));
|
| + __ ldr(r0, MemOperand(r3, offsetof(T, dividend)));
|
| + __ ldr(r1, MemOperand(r3, offsetof(T, divisor)));
|
|
|
| __ sdiv(r2, r0, r1);
|
| - __ str(r2, MemOperand(r3, OFFSET_OF(T, result)));
|
| + __ str(r2, MemOperand(r3, offsetof(T, result)));
|
|
|
| __ bx(lr);
|
|
|
| @@ -1917,29 +1917,29 @@ TEST(ARMv8_vrintX) {
|
| __ mov(r4, Operand(r0));
|
|
|
| // Test vrinta
|
| - __ vldr(d6, r4, OFFSET_OF(T, input));
|
| + __ vldr(d6, r4, offsetof(T, input));
|
| __ vrinta(d5, d6);
|
| - __ vstr(d5, r4, OFFSET_OF(T, ar));
|
| + __ vstr(d5, r4, offsetof(T, ar));
|
|
|
| // Test vrintn
|
| - __ vldr(d6, r4, OFFSET_OF(T, input));
|
| + __ vldr(d6, r4, offsetof(T, input));
|
| __ vrintn(d5, d6);
|
| - __ vstr(d5, r4, OFFSET_OF(T, nr));
|
| + __ vstr(d5, r4, offsetof(T, nr));
|
|
|
| // Test vrintp
|
| - __ vldr(d6, r4, OFFSET_OF(T, input));
|
| + __ vldr(d6, r4, offsetof(T, input));
|
| __ vrintp(d5, d6);
|
| - __ vstr(d5, r4, OFFSET_OF(T, pr));
|
| + __ vstr(d5, r4, offsetof(T, pr));
|
|
|
| // Test vrintm
|
| - __ vldr(d6, r4, OFFSET_OF(T, input));
|
| + __ vldr(d6, r4, offsetof(T, input));
|
| __ vrintm(d5, d6);
|
| - __ vstr(d5, r4, OFFSET_OF(T, mr));
|
| + __ vstr(d5, r4, offsetof(T, mr));
|
|
|
| // Test vrintz
|
| - __ vldr(d6, r4, OFFSET_OF(T, input));
|
| + __ vldr(d6, r4, offsetof(T, input));
|
| __ vrintz(d5, d6);
|
| - __ vstr(d5, r4, OFFSET_OF(T, zr));
|
| + __ vstr(d5, r4, offsetof(T, zr));
|
|
|
| __ ldm(ia_w, sp, r4.bit() | fp.bit() | pc.bit());
|
|
|
|
|