| Index: runtime/vm/assembler_arm_test.cc
|
| ===================================================================
|
| --- runtime/vm/assembler_arm_test.cc (revision 36715)
|
| +++ runtime/vm/assembler_arm_test.cc (working copy)
|
| @@ -17,7 +17,7 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Simple, assembler) {
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -29,7 +29,7 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(MoveNegated, assembler) {
|
| - __ mvn(R0, ShifterOperand(42));
|
| + __ mvn(R0, Operand(42));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -42,11 +42,11 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(MoveRotImm, assembler) {
|
| - ShifterOperand shifter_op;
|
| - EXPECT(ShifterOperand::CanHold(0x00550000, &shifter_op));
|
| - __ mov(R0, shifter_op);
|
| - EXPECT(ShifterOperand::CanHold(0x30000003, &shifter_op));
|
| - __ add(R0, R0, shifter_op);
|
| + Operand o;
|
| + EXPECT(Operand::CanHold(0x00550000, &o));
|
| + __ mov(R0, o);
|
| + EXPECT(Operand::CanHold(0x30000003, &o));
|
| + __ add(R0, R0, o);
|
| __ bx(LR);
|
| }
|
|
|
| @@ -65,7 +65,7 @@
|
| __ LoadDecodableImmediate(R0, 0x12345678 << 1);
|
| HostCPUFeatures::set_arm_version(ARMv7);
|
| __ LoadDecodableImmediate(R1, 0x12345678);
|
| - __ sub(R0, R0, ShifterOperand(R1));
|
| + __ sub(R0, R0, Operand(R1));
|
| __ bx(LR);
|
| #else
|
| __ LoadDecodableImmediate(R0, 0x12345678);
|
| @@ -82,8 +82,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(LoadImmediate, assembler) {
|
| - __ mov(R0, ShifterOperand(0));
|
| - __ cmp(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| + __ cmp(R0, Operand(0));
|
| __ LoadImmediate(R0, 0x12345678, EQ);
|
| __ LoadImmediate(R0, 0x87654321, NE);
|
| __ bx(LR);
|
| @@ -99,8 +99,8 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vmov, assembler) {
|
| if (TargetCPUFeatures::vfp_supported()) {
|
| - __ mov(R3, ShifterOperand(43));
|
| - __ mov(R1, ShifterOperand(41));
|
| + __ mov(R3, Operand(43));
|
| + __ mov(R1, Operand(41));
|
| __ vmovsrr(S1, R1, R3); // S1:S2 = 41:43
|
| __ vmovs(S0, S2); // S0 = S2, S0:S1 == 43:41
|
| __ vmovd(D2, D0); // D2 = D0, S4:S5 == 43:41
|
| @@ -109,7 +109,7 @@
|
| __ vmovdrr(D3, R3, R2); // D3 = R3:R2, S6:S7 == 41:41
|
| __ vmovsr(S7, R1); // S7 = R1, S6:S7 == 41:43
|
| __ vmovrrd(R0, R1, D3); // R0:R1 = D3, R0:R1 == 41:43
|
| - __ sub(R0, R1, ShifterOperand(R0)); // 43-41
|
| + __ sub(R0, R1, Operand(R0)); // 43-41
|
| } else {
|
| __ LoadImmediate(R0, 2);
|
| }
|
| @@ -127,7 +127,7 @@
|
| ASSEMBLER_TEST_GENERATE(SingleVLoadStore, assembler) {
|
| if (TargetCPUFeatures::vfp_supported()) {
|
| __ LoadImmediate(R0, bit_cast<int32_t, float>(12.3f));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R0, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ vldrs(S0, Address(R2, (-kWordSize * 30)));
|
| __ vadds(S0, S0, S0);
|
| @@ -156,10 +156,10 @@
|
| ASSEMBLER_TEST_GENERATE(SingleVShiftLoadStore, assembler) {
|
| if (TargetCPUFeatures::vfp_supported()) {
|
| __ LoadImmediate(R0, bit_cast<int32_t, float>(12.3f));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R2, Operand(SP));
|
| // Expressing __str(R0, Address(SP, (-kWordSize * 32), Address::PreIndex));
|
| // as:
|
| - __ mov(R1, ShifterOperand(kWordSize));
|
| + __ mov(R1, Operand(kWordSize));
|
| __ str(R0, Address(SP, R1, LSL, 5, Address::NegPreIndex));
|
| __ vldrs(S0, Address(R2, (-kWordSize * 32)));
|
| __ vadds(S0, S0, S0);
|
| @@ -192,7 +192,7 @@
|
| int64_t value = bit_cast<int64_t, double>(12.3);
|
| __ LoadImmediate(R0, Utils::Low32Bits(value));
|
| __ LoadImmediate(R1, Utils::High32Bits(value));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R0, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ str(R1, Address(R2, (-kWordSize * 29)));
|
| __ vldrd(D0, Address(R2, (-kWordSize * 30)));
|
| @@ -289,8 +289,8 @@
|
| __ vsqrtd(D0, D1);
|
| __ vcmpd(D0, D0);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(1), VS);
|
| - __ mov(R0, ShifterOperand(0), VC);
|
| + __ mov(R0, Operand(1), VS);
|
| + __ mov(R0, Operand(0), VC);
|
| } else {
|
| __ LoadImmediate(R0, 1);
|
| }
|
| @@ -307,7 +307,7 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(IntToDoubleConversion, assembler) {
|
| if (TargetCPUFeatures::vfp_supported()) {
|
| - __ mov(R3, ShifterOperand(6));
|
| + __ mov(R3, Operand(6));
|
| __ vmovsr(S3, R3);
|
| __ vcvtdi(D0, S3);
|
| } else {
|
| @@ -367,7 +367,7 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(IntToFloatConversion, assembler) {
|
| if (TargetCPUFeatures::vfp_supported()) {
|
| - __ mov(R3, ShifterOperand(6));
|
| + __ mov(R3, Operand(6));
|
| __ vmovsr(S3, R3);
|
| __ vcvtsi(S0, S3);
|
| } else {
|
| @@ -503,13 +503,13 @@
|
| __ LoadSImmediate(S1, 12.5f);
|
|
|
| // Count errors in R0. R0 is zero if no errors found.
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ vcmps(S0, S1);
|
| __ vmstat();
|
| - __ add(R0, R0, ShifterOperand(1), VS); // Error if unordered (Nan).
|
| - __ add(R0, R0, ShifterOperand(2), GT); // Error if greater.
|
| - __ add(R0, R0, ShifterOperand(4), EQ); // Error if equal.
|
| - __ add(R0, R0, ShifterOperand(8), PL); // Error if not less.
|
| + __ add(R0, R0, Operand(1), VS); // Error if unordered (Nan).
|
| + __ add(R0, R0, Operand(2), GT); // Error if greater.
|
| + __ add(R0, R0, Operand(4), EQ); // Error if equal.
|
| + __ add(R0, R0, Operand(8), PL); // Error if not less.
|
|
|
| // Test NaN.
|
| // Create NaN by dividing 0.0f/0.0f.
|
| @@ -518,7 +518,7 @@
|
| __ vcmps(S1, S1);
|
| __ vmstat();
|
| // Error if not unordered (not Nan).
|
| - __ add(R0, R0, ShifterOperand(16), VC);
|
| + __ add(R0, R0, Operand(16), VC);
|
| } else {
|
| __ LoadImmediate(R0, 0);
|
| }
|
| @@ -541,13 +541,13 @@
|
| __ LoadDImmediate(D1, 12.5, R1);
|
|
|
| // Count errors in R0. R0 is zero if no errors found.
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ vcmpd(D0, D1);
|
| __ vmstat();
|
| - __ add(R0, R0, ShifterOperand(1), VS); // Error if unordered (Nan).
|
| - __ add(R0, R0, ShifterOperand(2), GT); // Error if greater.
|
| - __ add(R0, R0, ShifterOperand(4), EQ); // Error if equal.
|
| - __ add(R0, R0, ShifterOperand(8), PL); // Error if not less.
|
| + __ add(R0, R0, Operand(1), VS); // Error if unordered (Nan).
|
| + __ add(R0, R0, Operand(2), GT); // Error if greater.
|
| + __ add(R0, R0, Operand(4), EQ); // Error if equal.
|
| + __ add(R0, R0, Operand(8), PL); // Error if not less.
|
|
|
| // Test NaN.
|
| // Create NaN by dividing 0.0/0.0.
|
| @@ -556,7 +556,7 @@
|
| __ vcmpd(D1, D1);
|
| __ vmstat();
|
| // Error if not unordered (not Nan).
|
| - __ add(R0, R0, ShifterOperand(16), VC);
|
| + __ add(R0, R0, Operand(16), VC);
|
| } else {
|
| __ LoadImmediate(R0, 0);
|
| }
|
| @@ -574,11 +574,11 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Loop, assembler) {
|
| Label loop_entry;
|
| - __ mov(R0, ShifterOperand(1));
|
| - __ mov(R1, ShifterOperand(2));
|
| + __ mov(R0, Operand(1));
|
| + __ mov(R1, Operand(2));
|
| __ Bind(&loop_entry);
|
| - __ mov(R0, ShifterOperand(R0, LSL, 1));
|
| - __ movs(R1, ShifterOperand(R1, LSR, 1));
|
| + __ mov(R0, Operand(R0, LSL, 1));
|
| + __ movs(R1, Operand(R1, LSR, 1));
|
| __ b(&loop_entry, NE);
|
| __ bx(LR);
|
| }
|
| @@ -593,9 +593,9 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(ForwardBranch, assembler) {
|
| Label skip;
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
| __ b(&skip);
|
| - __ mov(R0, ShifterOperand(11));
|
| + __ mov(R0, Operand(11));
|
| __ Bind(&skip);
|
| __ bx(LR);
|
| }
|
| @@ -611,11 +611,11 @@
|
| ASSEMBLER_TEST_GENERATE(Loop2, assembler) {
|
| Label loop_entry;
|
| __ set_use_far_branches(true);
|
| - __ mov(R0, ShifterOperand(1));
|
| - __ mov(R1, ShifterOperand(2));
|
| + __ mov(R0, Operand(1));
|
| + __ mov(R1, Operand(2));
|
| __ Bind(&loop_entry);
|
| - __ mov(R0, ShifterOperand(R0, LSL, 1));
|
| - __ movs(R1, ShifterOperand(R1, LSR, 1));
|
| + __ mov(R0, Operand(R0, LSL, 1));
|
| + __ movs(R1, Operand(R1, LSR, 1));
|
| __ b(&loop_entry, NE);
|
| __ bx(LR);
|
| }
|
| @@ -631,14 +631,14 @@
|
| ASSEMBLER_TEST_GENERATE(Loop3, assembler) {
|
| Label loop_entry;
|
| __ set_use_far_branches(true);
|
| - __ mov(R0, ShifterOperand(1));
|
| - __ mov(R1, ShifterOperand(2));
|
| + __ mov(R0, Operand(1));
|
| + __ mov(R1, Operand(2));
|
| __ Bind(&loop_entry);
|
| for (int i = 0; i < (1 << 22); i++) {
|
| __ nop();
|
| }
|
| - __ mov(R0, ShifterOperand(R0, LSL, 1));
|
| - __ movs(R1, ShifterOperand(R1, LSR, 1));
|
| + __ mov(R0, Operand(R0, LSL, 1));
|
| + __ movs(R1, Operand(R1, LSR, 1));
|
| __ b(&loop_entry, NE);
|
| __ bx(LR);
|
| }
|
| @@ -652,7 +652,7 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(LoadStore, assembler) {
|
| - __ mov(R1, ShifterOperand(123));
|
| + __ mov(R1, Operand(123));
|
| __ Push(R1);
|
| __ Pop(R0);
|
| __ bx(LR);
|
| @@ -667,11 +667,11 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddSub, assembler) {
|
| - __ mov(R1, ShifterOperand(40));
|
| - __ sub(R1, R1, ShifterOperand(2));
|
| - __ add(R0, R1, ShifterOperand(4));
|
| - __ rsbs(R0, R0, ShifterOperand(100));
|
| - __ rsc(R0, R0, ShifterOperand(100));
|
| + __ mov(R1, Operand(40));
|
| + __ sub(R1, R1, Operand(2));
|
| + __ add(R0, R1, Operand(4));
|
| + __ rsbs(R0, R0, Operand(100));
|
| + __ rsc(R0, R0, Operand(100));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -685,10 +685,10 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(AddCarry, assembler) {
|
| __ LoadImmediate(R2, 0xFFFFFFFF);
|
| - __ mov(R1, ShifterOperand(1));
|
| - __ mov(R0, ShifterOperand(0));
|
| - __ adds(R2, R2, ShifterOperand(R1));
|
| - __ adcs(R0, R0, ShifterOperand(R0));
|
| + __ mov(R1, Operand(1));
|
| + __ mov(R0, Operand(0));
|
| + __ adds(R2, R2, Operand(R1));
|
| + __ adcs(R0, R0, Operand(R0));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -702,10 +702,10 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(SubCarry, assembler) {
|
| __ LoadImmediate(R2, 0x0);
|
| - __ mov(R1, ShifterOperand(1));
|
| - __ mov(R0, ShifterOperand(0));
|
| - __ subs(R2, R2, ShifterOperand(R1));
|
| - __ sbcs(R0, R0, ShifterOperand(R0));
|
| + __ mov(R1, Operand(1));
|
| + __ mov(R0, Operand(0));
|
| + __ subs(R2, R2, Operand(R1));
|
| + __ sbcs(R0, R0, Operand(R0));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -718,11 +718,11 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AndOrr, assembler) {
|
| - __ mov(R1, ShifterOperand(40));
|
| - __ mov(R2, ShifterOperand(0));
|
| - __ and_(R1, R2, ShifterOperand(R1));
|
| - __ mov(R3, ShifterOperand(42));
|
| - __ orr(R0, R1, ShifterOperand(R3));
|
| + __ mov(R1, Operand(40));
|
| + __ mov(R2, Operand(0));
|
| + __ and_(R1, R2, Operand(R1));
|
| + __ mov(R3, Operand(42));
|
| + __ orr(R0, R1, Operand(R3));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -735,13 +735,13 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Orrs, assembler) {
|
| - __ mov(R0, ShifterOperand(0));
|
| - __ tst(R0, ShifterOperand(R1)); // Set zero-flag.
|
| - __ orrs(R0, R0, ShifterOperand(1)); // Clear zero-flag.
|
| + __ mov(R0, Operand(0));
|
| + __ tst(R0, Operand(R1)); // Set zero-flag.
|
| + __ orrs(R0, R0, Operand(1)); // Clear zero-flag.
|
| __ bx(LR, EQ);
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
| __ bx(LR, NE); // Only this return should fire.
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -754,10 +754,10 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Multiply, assembler) {
|
| - __ mov(R1, ShifterOperand(20));
|
| - __ mov(R2, ShifterOperand(40));
|
| + __ mov(R1, Operand(20));
|
| + __ mov(R2, Operand(40));
|
| __ mul(R3, R2, R1);
|
| - __ mov(R0, ShifterOperand(R3));
|
| + __ mov(R0, Operand(R3));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -807,11 +807,11 @@
|
| #endif
|
| if (TargetCPUFeatures::arm_version() == ARMv7) {
|
| __ Push(R4);
|
| - __ mov(IP, ShifterOperand(R0));
|
| + __ mov(IP, Operand(R0));
|
| __ mul(R4, R2, R1);
|
| __ umull(R0, R1, R2, IP);
|
| __ mla(R2, IP, R3, R4);
|
| - __ add(R1, R2, ShifterOperand(R1));
|
| + __ add(R1, R2, Operand(R1));
|
| __ Pop(R4);
|
| } else {
|
| __ LoadImmediate(R0, 6);
|
| @@ -884,26 +884,26 @@
|
| ASSEMBLER_TEST_GENERATE(Clz, assembler) {
|
| Label error;
|
|
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ clz(R1, R0);
|
| - __ cmp(R1, ShifterOperand(32));
|
| + __ cmp(R1, Operand(32));
|
| __ b(&error, NE);
|
| - __ mov(R2, ShifterOperand(42));
|
| + __ mov(R2, Operand(42));
|
| __ clz(R2, R2);
|
| - __ cmp(R2, ShifterOperand(26));
|
| + __ cmp(R2, Operand(26));
|
| __ b(&error, NE);
|
| - __ mvn(R0, ShifterOperand(0));
|
| + __ mvn(R0, Operand(0));
|
| __ clz(R1, R0);
|
| - __ cmp(R1, ShifterOperand(0));
|
| + __ cmp(R1, Operand(0));
|
| __ b(&error, NE);
|
| __ Lsr(R0, R0, 3);
|
| __ clz(R1, R0);
|
| - __ cmp(R1, ShifterOperand(3));
|
| + __ cmp(R1, Operand(3));
|
| __ b(&error, NE);
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ bx(LR);
|
| __ Bind(&error);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -918,11 +918,11 @@
|
| ASSEMBLER_TEST_GENERATE(Tst, assembler) {
|
| Label skip;
|
|
|
| - __ mov(R0, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(40));
|
| - __ tst(R1, ShifterOperand(0));
|
| + __ mov(R0, Operand(42));
|
| + __ mov(R1, Operand(40));
|
| + __ tst(R1, Operand(0));
|
| __ b(&skip, NE);
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ Bind(&skip);
|
| __ bx(LR);
|
| }
|
| @@ -938,10 +938,10 @@
|
| ASSEMBLER_TEST_GENERATE(Lsl, assembler) {
|
| Label skip;
|
|
|
| - __ mov(R0, ShifterOperand(1));
|
| - __ mov(R0, ShifterOperand(R0, LSL, 1));
|
| - __ mov(R1, ShifterOperand(1));
|
| - __ mov(R0, ShifterOperand(R0, LSL, R1));
|
| + __ mov(R0, Operand(1));
|
| + __ mov(R0, Operand(R0, LSL, 1));
|
| + __ mov(R1, Operand(1));
|
| + __ mov(R0, Operand(R0, LSL, R1));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -956,10 +956,10 @@
|
| ASSEMBLER_TEST_GENERATE(Lsr, assembler) {
|
| Label skip;
|
|
|
| - __ mov(R0, ShifterOperand(4));
|
| - __ mov(R0, ShifterOperand(R0, LSR, 1));
|
| - __ mov(R1, ShifterOperand(1));
|
| - __ mov(R0, ShifterOperand(R0, LSR, R1));
|
| + __ mov(R0, Operand(4));
|
| + __ mov(R0, Operand(R0, LSR, 1));
|
| + __ mov(R1, Operand(1));
|
| + __ mov(R0, Operand(R0, LSR, R1));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -974,7 +974,7 @@
|
| ASSEMBLER_TEST_GENERATE(Lsr1, assembler) {
|
| Label skip;
|
|
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ Lsl(R0, R0, 31);
|
| __ Lsr(R0, R0, 31);
|
| __ bx(LR);
|
| @@ -991,7 +991,7 @@
|
| ASSEMBLER_TEST_GENERATE(Asr1, assembler) {
|
| Label skip;
|
|
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ Lsl(R0, R0, 31);
|
| __ Asr(R0, R0, 31);
|
| __ bx(LR);
|
| @@ -1006,8 +1006,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Rsb, assembler) {
|
| - __ mov(R3, ShifterOperand(10));
|
| - __ rsb(R0, R3, ShifterOperand(42));
|
| + __ mov(R3, Operand(10));
|
| + __ rsb(R0, R3, Operand(42));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -1022,36 +1022,36 @@
|
| ASSEMBLER_TEST_GENERATE(Ldrh, assembler) {
|
| Label Test1, Test2, Test3, Done;
|
|
|
| - __ mov(R1, ShifterOperand(0x11));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R1, Operand(0x11));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R1, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ ldrh(R0, Address(R2, (-kWordSize * 30)));
|
| - __ cmp(R0, ShifterOperand(0x11));
|
| + __ cmp(R0, Operand(0x11));
|
| __ b(&Test1, EQ);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ b(&Done);
|
| __ Bind(&Test1);
|
|
|
| - __ mov(R0, ShifterOperand(0x22));
|
| + __ mov(R0, Operand(0x22));
|
| __ strh(R0, Address(R2, (-kWordSize * 30)));
|
| __ ldrh(R1, Address(R2, (-kWordSize * 30)));
|
| - __ cmp(R1, ShifterOperand(0x22));
|
| + __ cmp(R1, Operand(0x22));
|
| __ b(&Test2, EQ);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ b(&Done);
|
| __ Bind(&Test2);
|
|
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ AddImmediate(R2, (-kWordSize * 30));
|
| __ strh(R0, Address(R2));
|
| __ ldrh(R1, Address(R2));
|
| - __ cmp(R1, ShifterOperand(0));
|
| + __ cmp(R1, Operand(0));
|
| __ b(&Test3, EQ);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ b(&Done);
|
| __ Bind(&Test3);
|
|
|
| - __ mov(R0, ShifterOperand(0));
|
| + __ mov(R0, Operand(0));
|
| __ Bind(&Done);
|
| __ ldr(R1, Address(SP, (kWordSize * 30), Address::PostIndex));
|
| __ bx(LR);
|
| @@ -1066,8 +1066,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Ldrsb, assembler) {
|
| - __ mov(R1, ShifterOperand(0xFF));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R1, Operand(0xFF));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R1, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ ldrsb(R0, Address(R2, (-kWordSize * 30)));
|
| __ ldr(R1, Address(SP, (kWordSize * 30), Address::PostIndex));
|
| @@ -1083,8 +1083,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Ldrb, assembler) {
|
| - __ mov(R1, ShifterOperand(0xFF));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R1, Operand(0xFF));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R1, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ ldrb(R0, Address(R2, (-kWordSize * 30)));
|
| __ ldr(R1, Address(SP, (kWordSize * 30), Address::PostIndex));
|
| @@ -1100,8 +1100,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Ldrsh, assembler) {
|
| - __ mov(R1, ShifterOperand(0xFF));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R1, Operand(0xFF));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R1, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ ldrsh(R0, Address(R2, (-kWordSize * 30)));
|
| __ ldr(R1, Address(SP, (kWordSize * 30), Address::PostIndex));
|
| @@ -1117,8 +1117,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Ldrh1, assembler) {
|
| - __ mov(R1, ShifterOperand(0xFF));
|
| - __ mov(R2, ShifterOperand(SP));
|
| + __ mov(R1, Operand(0xFF));
|
| + __ mov(R2, Operand(SP));
|
| __ str(R1, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ ldrh(R0, Address(R2, (-kWordSize * 30)));
|
| __ ldr(R1, Address(SP, (kWordSize * 30), Address::PostIndex));
|
| @@ -1134,13 +1134,13 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Ldrd, assembler) {
|
| - __ mov(IP, ShifterOperand(SP));
|
| + __ mov(IP, Operand(SP));
|
| __ strd(R2, Address(SP, (-kWordSize * 30), Address::PreIndex));
|
| __ strd(R0, Address(IP, (-kWordSize * 28)));
|
| __ ldrd(R2, Address(IP, (-kWordSize * 28)));
|
| __ ldrd(R0, Address(SP, (kWordSize * 30), Address::PostIndex));
|
| - __ sub(R0, R0, ShifterOperand(R2));
|
| - __ add(R1, R1, ShifterOperand(R3));
|
| + __ sub(R0, R0, Operand(R2));
|
| + __ add(R1, R1, Operand(R3));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -1154,10 +1154,10 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Ldm_stm_da, assembler) {
|
| - __ mov(R0, ShifterOperand(1));
|
| - __ mov(R1, ShifterOperand(7));
|
| - __ mov(R2, ShifterOperand(11));
|
| - __ mov(R3, ShifterOperand(31));
|
| + __ mov(R0, Operand(1));
|
| + __ mov(R1, Operand(7));
|
| + __ mov(R2, Operand(11));
|
| + __ mov(R3, Operand(31));
|
| __ Push(R5); // We use R5 as accumulator.
|
| __ Push(R5);
|
| __ Push(R5);
|
| @@ -1168,17 +1168,17 @@
|
| __ str(R2, Address(SP)); // Should be a free slot.
|
| __ ldr(R5, Address(SP, 1 * kWordSize)); // R0. R5 = +1.
|
| __ ldr(IP, Address(SP, 2 * kWordSize)); // R1.
|
| - __ sub(R5, R5, ShifterOperand(IP)); // -R1. R5 = -6.
|
| + __ sub(R5, R5, Operand(IP)); // -R1. R5 = -6.
|
| __ ldr(IP, Address(SP, 3 * kWordSize)); // R2.
|
| - __ add(R5, R5, ShifterOperand(IP)); // +R2. R5 = +5.
|
| + __ add(R5, R5, Operand(IP)); // +R2. R5 = +5.
|
| __ ldr(IP, Address(SP, 4 * kWordSize)); // R3.
|
| - __ sub(R5, R5, ShifterOperand(IP)); // -R3. R5 = -26.
|
| + __ sub(R5, R5, Operand(IP)); // -R3. R5 = -26.
|
| __ ldm(IB_W, SP, (1 << R0 | 1 << R1 | 1 << R2 | 1 << R3));
|
| // Same operations again. But this time from the restore registers.
|
| - __ add(R5, R5, ShifterOperand(R0));
|
| - __ sub(R5, R5, ShifterOperand(R1));
|
| - __ add(R5, R5, ShifterOperand(R2));
|
| - __ sub(R0, R5, ShifterOperand(R3)); // R0 = result = -52.
|
| + __ add(R5, R5, Operand(R0));
|
| + __ sub(R5, R5, Operand(R1));
|
| + __ add(R5, R5, Operand(R2));
|
| + __ sub(R0, R5, Operand(R3)); // R0 = result = -52.
|
| __ Pop(R1); // Remove storage slot.
|
| __ Pop(R5); // Restore R5.
|
| __ Pop(R5); // Restore R5.
|
| @@ -1197,8 +1197,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddressShiftStrLSL1NegOffset, assembler) {
|
| - __ mov(R2, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(kWordSize));
|
| + __ mov(R2, Operand(42));
|
| + __ mov(R1, Operand(kWordSize));
|
| __ str(R2, Address(SP, R1, LSL, 1, Address::NegOffset));
|
| __ ldr(R0, Address(SP, (-kWordSize * 2), Address::Offset));
|
| __ bx(LR);
|
| @@ -1213,8 +1213,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddressShiftLdrLSL5NegOffset, assembler) {
|
| - __ mov(R2, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(kWordSize));
|
| + __ mov(R2, Operand(42));
|
| + __ mov(R1, Operand(kWordSize));
|
| __ str(R2, Address(SP, (-kWordSize * 32), Address::Offset));
|
| __ ldr(R0, Address(SP, R1, LSL, 5, Address::NegOffset));
|
| __ bx(LR);
|
| @@ -1229,8 +1229,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddressShiftStrLRS1NegOffset, assembler) {
|
| - __ mov(R2, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(kWordSize * 2));
|
| + __ mov(R2, Operand(42));
|
| + __ mov(R1, Operand(kWordSize * 2));
|
| __ str(R2, Address(SP, R1, LSR, 1, Address::NegOffset));
|
| __ ldr(R0, Address(SP, -kWordSize, Address::Offset));
|
| __ bx(LR);
|
| @@ -1245,8 +1245,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddressShiftLdrLRS1NegOffset, assembler) {
|
| - __ mov(R2, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(kWordSize * 2));
|
| + __ mov(R2, Operand(42));
|
| + __ mov(R1, Operand(kWordSize * 2));
|
| __ str(R2, Address(SP, -kWordSize, Address::Offset));
|
| __ ldr(R0, Address(SP, R1, LSR, 1, Address::NegOffset));
|
| __ bx(LR);
|
| @@ -1261,12 +1261,12 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddressShiftStrLSLNegPreIndex, assembler) {
|
| - __ mov(R2, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(kWordSize));
|
| - __ mov(R3, ShifterOperand(SP));
|
| + __ mov(R2, Operand(42));
|
| + __ mov(R1, Operand(kWordSize));
|
| + __ mov(R3, Operand(SP));
|
| __ str(R2, Address(SP, R1, LSL, 5, Address::NegPreIndex));
|
| __ ldr(R0, Address(R3, (-kWordSize * 32), Address::Offset));
|
| - __ mov(SP, ShifterOperand(R3));
|
| + __ mov(SP, Operand(R3));
|
| __ bx(LR);
|
| }
|
|
|
| @@ -1279,8 +1279,8 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(AddressShiftLdrLSLNegPreIndex, assembler) {
|
| - __ mov(R2, ShifterOperand(42));
|
| - __ mov(R1, ShifterOperand(kWordSize));
|
| + __ mov(R2, Operand(42));
|
| + __ mov(R1, Operand(kWordSize));
|
| __ str(R2, Address(SP, (-kWordSize * 32), Address::PreIndex));
|
| __ ldr(R0, Address(SP, R1, LSL, 5, Address::PostIndex));
|
| __ bx(LR);
|
| @@ -1311,32 +1311,32 @@
|
| __ vldmd(IA_W, SP, D0, 5); // Pop stack into D0 - D4, inc SP
|
|
|
| // Load success value into R0
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
|
|
| // Check that 4.0 is back in D4
|
| __ LoadDImmediate(D5, 4.0, R1);
|
| __ vcmpd(D4, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 3.0 is back in D3
|
| __ LoadDImmediate(D5, 3.0, R1);
|
| __ vcmpd(D3, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
|
|
| // Check that 2.0 is back in D2
|
| __ LoadDImmediate(D5, 2.0, R1);
|
| __ vcmpd(D2, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 1.0 is back in D1
|
| __ LoadDImmediate(D5, 1.0, R1);
|
| __ vcmpd(D1, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
| } else {
|
| __ LoadImmediate(R0, 42);
|
| }
|
| @@ -1368,31 +1368,31 @@
|
| __ vldms(IA_W, SP, S0, S4); // Pop stack into S0 - S4, inc SP
|
|
|
| // Load success value into R0
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
|
|
| // Check that 4.0 is back in S4
|
| __ LoadSImmediate(S5, 4.0);
|
| __ vcmps(S4, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 3.0 is back in S3
|
| __ LoadSImmediate(S5, 3.0);
|
| __ vcmps(S3, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 2.0 is back in S2
|
| __ LoadSImmediate(S5, 2.0);
|
| __ vcmps(S2, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 1.0 is back in S1
|
| __ LoadSImmediate(S5, 1.0);
|
| __ vcmps(S1, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
| } else {
|
| __ LoadImmediate(R0, 42);
|
| }
|
| @@ -1422,32 +1422,32 @@
|
| __ vldmd(IA_W, SP, D1, 4); // Pop stack into D1 - D4, inc SP
|
|
|
| // Load success value into R0
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
|
|
| // Check that 4.0 is back in D4
|
| __ LoadDImmediate(D5, 4.0, R1);
|
| __ vcmpd(D4, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 3.0 is back in D3
|
| __ LoadDImmediate(D5, 3.0, R1);
|
| __ vcmpd(D3, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
|
|
| // Check that 2.0 is back in D2
|
| __ LoadDImmediate(D5, 2.0, R1);
|
| __ vcmpd(D2, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 1.0 is back in D1
|
| __ LoadDImmediate(D5, 1.0, R1);
|
| __ vcmpd(D1, D5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
| } else {
|
| __ LoadImmediate(R0, 42);
|
| }
|
| @@ -1477,31 +1477,31 @@
|
| __ vldms(IA_W, SP, S1, S4); // Pop stack into S0 - S4, inc SP
|
|
|
| // Load success value into R0
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
|
|
| // Check that 4.0 is back in S4
|
| __ LoadSImmediate(S5, 4.0);
|
| __ vcmps(S4, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 3.0 is back in S3
|
| __ LoadSImmediate(S5, 3.0);
|
| __ vcmps(S3, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 2.0 is back in S2
|
| __ LoadSImmediate(S5, 2.0);
|
| __ vcmps(S2, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 1.0 is back in S1
|
| __ LoadSImmediate(S5, 1.0);
|
| __ vcmps(S1, S5);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
| } else {
|
| __ LoadImmediate(R0, 42);
|
| }
|
| @@ -1532,38 +1532,38 @@
|
| __ vldmd(IA_W, SP, D5, 5); // Pop stack into D5 - D9, inc SP
|
|
|
| // Load success value into R0
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
|
|
| // Check that 4.0 is in D9
|
| __ LoadDImmediate(D10, 4.0, R1);
|
| __ vcmpd(D9, D10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 3.0 is in D8
|
| __ LoadDImmediate(D10, 3.0, R1);
|
| __ vcmpd(D8, D10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
|
|
| // Check that 2.0 is in D7
|
| __ LoadDImmediate(D10, 2.0, R1);
|
| __ vcmpd(D7, D10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 1.0 is in D6
|
| __ LoadDImmediate(D10, 1.0, R1);
|
| __ vcmpd(D6, D10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Check that 0.0 is in D5
|
| __ LoadDImmediate(D10, 0.0, R1);
|
| __ vcmpd(D5, D10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure into R0 if NE
|
|
|
| // Restore used callee-saved FPU registers.
|
| __ vldmd(IA_W, SP, D8, 3);
|
| @@ -1594,37 +1594,37 @@
|
| __ vldms(IA_W, SP, S5, S9); // Pop stack into S5 - S9, inc SP
|
|
|
| // Load success value into R0
|
| - __ mov(R0, ShifterOperand(42));
|
| + __ mov(R0, Operand(42));
|
|
|
| // Check that 4.0 is in S9
|
| __ LoadSImmediate(S10, 4.0);
|
| __ vcmps(S9, S10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 3.0 is in S8
|
| __ LoadSImmediate(S10, 3.0);
|
| __ vcmps(S8, S10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 2.0 is in S7
|
| __ LoadSImmediate(S10, 2.0);
|
| __ vcmps(S7, S10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 1.0 is back in S6
|
| __ LoadSImmediate(S10, 1.0);
|
| __ vcmps(S6, S10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
|
|
| // Check that 0.0 is back in S5
|
| __ LoadSImmediate(S10, 0.0);
|
| __ vcmps(S5, S10);
|
| __ vmstat();
|
| - __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
|
| + __ mov(R0, Operand(0), NE); // Put failure value into R0 if NE
|
| } else {
|
| __ LoadImmediate(R0, 42);
|
| }
|
| @@ -1641,12 +1641,12 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Udiv, assembler) {
|
| if (TargetCPUFeatures::integer_division_supported()) {
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(9));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(9));
|
| __ udiv(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(R2));
|
| + __ mov(R0, Operand(R2));
|
| } else {
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| }
|
| __ bx(LR);
|
| }
|
| @@ -1661,10 +1661,10 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Sdiv, assembler) {
|
| if (TargetCPUFeatures::integer_division_supported()) {
|
| - __ mov(R0, ShifterOperand(27));
|
| + __ mov(R0, Operand(27));
|
| __ LoadImmediate(R1, -9);
|
| __ sdiv(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(R2));
|
| + __ mov(R0, Operand(R2));
|
| } else {
|
| __ LoadImmediate(R0, -3);
|
| }
|
| @@ -1681,10 +1681,10 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Udiv_zero, assembler) {
|
| if (TargetCPUFeatures::integer_division_supported()) {
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(0));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(0));
|
| __ udiv(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(R2));
|
| + __ mov(R0, Operand(R2));
|
| } else {
|
| __ LoadImmediate(R0, 0);
|
| }
|
| @@ -1701,10 +1701,10 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Sdiv_zero, assembler) {
|
| if (TargetCPUFeatures::integer_division_supported()) {
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(0));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(0));
|
| __ sdiv(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(R2));
|
| + __ mov(R0, Operand(R2));
|
| } else {
|
| __ LoadImmediate(R0, 0);
|
| }
|
| @@ -1724,7 +1724,7 @@
|
| __ LoadImmediate(R0, 0x80000000);
|
| __ LoadImmediate(R1, 0xffffffff);
|
| __ udiv(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(R2));
|
| + __ mov(R0, Operand(R2));
|
| } else {
|
| __ LoadImmediate(R0, 0);
|
| }
|
| @@ -1744,7 +1744,7 @@
|
| __ LoadImmediate(R0, 0x80000000);
|
| __ LoadImmediate(R1, 0xffffffff);
|
| __ sdiv(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(R2));
|
| + __ mov(R0, Operand(R2));
|
| } else {
|
| __ LoadImmediate(R0, 0x80000000);
|
| }
|
| @@ -1764,14 +1764,14 @@
|
| #if defined(USING_SIMULATOR)
|
| bool orig = TargetCPUFeatures::integer_division_supported();
|
| HostCPUFeatures::set_integer_division_supported(true);
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(9));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(9));
|
| __ IntegerDivide(R0, R0, R1, D0, D1);
|
| HostCPUFeatures::set_integer_division_supported(orig);
|
| __ bx(LR);
|
| #else
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(9));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(9));
|
| __ IntegerDivide(R0, R0, R1, D0, D1);
|
| __ bx(LR);
|
| #endif
|
| @@ -1789,14 +1789,14 @@
|
| #if defined(USING_SIMULATOR)
|
| bool orig = TargetCPUFeatures::integer_division_supported();
|
| HostCPUFeatures::set_integer_division_supported(false);
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(9));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(9));
|
| __ IntegerDivide(R0, R0, R1, D0, D1);
|
| HostCPUFeatures::set_integer_division_supported(orig);
|
| __ bx(LR);
|
| #else
|
| - __ mov(R0, ShifterOperand(27));
|
| - __ mov(R1, ShifterOperand(9));
|
| + __ mov(R0, Operand(27));
|
| + __ mov(R1, Operand(9));
|
| __ IntegerDivide(R0, R0, R1, D0, D1);
|
| __ bx(LR);
|
| #endif
|
| @@ -1811,10 +1811,10 @@
|
|
|
|
|
| ASSEMBLER_TEST_GENERATE(Muls, assembler) {
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ LoadImmediate(R1, -9);
|
| __ muls(R2, R0, R1);
|
| - __ mov(R0, ShifterOperand(42), MI);
|
| + __ mov(R0, Operand(42), MI);
|
| __ bx(LR);
|
| }
|
|
|
| @@ -1828,21 +1828,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vaddqi8, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(5));
|
| + __ mov(R0, Operand(5));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(7));
|
| + __ mov(R0, Operand(7));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vaddqi(kByte, Q2, Q0, Q1);
|
| @@ -1852,9 +1852,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 36);
|
| @@ -1872,21 +1872,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vaddqi16, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(5));
|
| + __ mov(R0, Operand(5));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(7));
|
| + __ mov(R0, Operand(7));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vaddqi(kHalfword, Q2, Q0, Q1);
|
| @@ -1896,9 +1896,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 36);
|
| @@ -1916,21 +1916,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vaddqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(5));
|
| + __ mov(R0, Operand(5));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(7));
|
| + __ mov(R0, Operand(7));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vaddqi(kWord, Q2, Q0, Q1);
|
| @@ -1940,9 +1940,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 36);
|
| @@ -1960,13 +1960,13 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vaddqi64, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S6, R0);
|
|
|
| __ vaddqi(kWordPair, Q2, Q0, Q1);
|
| @@ -1974,7 +1974,7 @@
|
| __ vmovrs(R0, S8);
|
| __ vmovrs(R2, S10);
|
|
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| + __ add(R0, R0, Operand(R2));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 10);
|
| @@ -2091,7 +2091,7 @@
|
| __ vshlqu(kWordPair, out, value, temp);
|
|
|
| // check for overflow by shifting back and comparing.
|
| - __ rsb(shift, shift, ShifterOperand(0));
|
| + __ rsb(shift, shift, Operand(0));
|
| __ vmovsr(stemp0, shift);
|
| __ vshlqi(kWordPair, temp, out, temp);
|
| __ vceqqi(kWord, out, temp, value);
|
| @@ -2148,7 +2148,7 @@
|
| __ vshlqu(kWordPair, out, value, temp);
|
|
|
| // check for overflow by shifting back and comparing.
|
| - __ rsb(shift, shift, ShifterOperand(0));
|
| + __ rsb(shift, shift, Operand(0));
|
| __ vmovsr(stemp0, shift);
|
| __ vshlqi(kWordPair, temp, out, temp);
|
| __ vceqqi(kWord, out, temp, value);
|
| @@ -2182,21 +2182,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vsubqi8, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vsubqi(kByte, Q2, Q1, Q0);
|
| @@ -2206,9 +2206,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 10);
|
| @@ -2226,21 +2226,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vsubqi16, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vsubqi(kHalfword, Q2, Q1, Q0);
|
| @@ -2250,9 +2250,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 10);
|
| @@ -2270,21 +2270,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vsubqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vsubqi(kWord, Q2, Q1, Q0);
|
| @@ -2294,9 +2294,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 10);
|
| @@ -2314,13 +2314,13 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vsubqi64, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S6, R0);
|
|
|
| __ vsubqi(kWordPair, Q2, Q1, Q0);
|
| @@ -2328,7 +2328,7 @@
|
| __ vmovrs(R0, S8);
|
| __ vmovrs(R2, S10);
|
|
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| + __ add(R0, R0, Operand(R2));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 3);
|
| @@ -2346,21 +2346,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vmulqi8, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(5));
|
| + __ mov(R0, Operand(5));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(7));
|
| + __ mov(R0, Operand(7));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vmulqi(kByte, Q2, Q1, Q0);
|
| @@ -2370,9 +2370,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 70);
|
| @@ -2390,21 +2390,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vmulqi16, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(5));
|
| + __ mov(R0, Operand(5));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(7));
|
| + __ mov(R0, Operand(7));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vmulqi(kHalfword, Q2, Q1, Q0);
|
| @@ -2414,9 +2414,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 70);
|
| @@ -2434,21 +2434,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vmulqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(5));
|
| + __ mov(R0, Operand(5));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(6));
|
| + __ mov(R0, Operand(6));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(7));
|
| + __ mov(R0, Operand(7));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(8));
|
| + __ mov(R0, Operand(8));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vmulqi(kWord, Q2, Q1, Q0);
|
| @@ -2458,9 +2458,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 70);
|
| @@ -2777,9 +2777,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -8);
|
| @@ -2819,9 +2819,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 60);
|
| @@ -2861,9 +2861,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -4);
|
| @@ -2903,9 +2903,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, 4);
|
| @@ -2996,9 +2996,9 @@
|
| __ vmovrs(R2, S2);
|
| __ vmovrs(R3, S3);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -4);
|
| @@ -3029,9 +3029,9 @@
|
| __ vmovrs(R2, S2);
|
| __ vmovrs(R3, S3);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -4);
|
| @@ -3062,9 +3062,9 @@
|
| __ vmovrs(R2, S2);
|
| __ vmovrs(R3, S3);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -4);
|
| @@ -3116,21 +3116,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vceqqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(20));
|
| + __ mov(R0, Operand(20));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(40));
|
| + __ mov(R0, Operand(40));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vceqqi(kWord, Q2, Q1, Q0);
|
| @@ -3140,9 +3140,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3176,9 +3176,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3196,21 +3196,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vcgeqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vcgeqi(kWord, Q2, Q1, Q0);
|
| @@ -3220,9 +3220,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3240,21 +3240,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vcugeqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| __ LoadImmediate(R0, -1);
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S5, R0);
|
| __ LoadImmediate(R0, -3);
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vcugeqi(kWord, Q2, Q1, Q0);
|
| @@ -3264,9 +3264,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3300,9 +3300,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3320,21 +3320,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vcgtqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S5, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vcgtqi(kWord, Q2, Q1, Q0);
|
| @@ -3344,9 +3344,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3364,21 +3364,21 @@
|
|
|
| ASSEMBLER_TEST_GENERATE(Vcugtqi32, assembler) {
|
| if (TargetCPUFeatures::neon_supported()) {
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S0, R0);
|
| - __ mov(R0, ShifterOperand(2));
|
| + __ mov(R0, Operand(2));
|
| __ vmovsr(S1, R0);
|
| - __ mov(R0, ShifterOperand(3));
|
| + __ mov(R0, Operand(3));
|
| __ vmovsr(S2, R0);
|
| - __ mov(R0, ShifterOperand(4));
|
| + __ mov(R0, Operand(4));
|
| __ vmovsr(S3, R0);
|
| __ LoadImmediate(R0, -1);
|
| __ vmovsr(S4, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S5, R0);
|
| __ LoadImmediate(R0, -3);
|
| __ vmovsr(S6, R0);
|
| - __ mov(R0, ShifterOperand(1));
|
| + __ mov(R0, Operand(1));
|
| __ vmovsr(S7, R0);
|
|
|
| __ vcugtqi(kWord, Q2, Q1, Q0);
|
| @@ -3388,9 +3388,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -3424,9 +3424,9 @@
|
| __ vmovrs(R2, S10);
|
| __ vmovrs(R3, S11);
|
|
|
| - __ add(R0, R0, ShifterOperand(R1));
|
| - __ add(R0, R0, ShifterOperand(R2));
|
| - __ add(R0, R0, ShifterOperand(R3));
|
| + __ add(R0, R0, Operand(R1));
|
| + __ add(R0, R0, Operand(R2));
|
| + __ add(R0, R0, Operand(R3));
|
| __ bx(LR);
|
| } else {
|
| __ LoadImmediate(R0, -2);
|
| @@ -4068,7 +4068,7 @@
|
| // R2: growable array.
|
| ASSEMBLER_TEST_GENERATE(StoreIntoObject, assembler) {
|
| __ PushList((1 << CTX) | (1 << LR));
|
| - __ mov(CTX, ShifterOperand(R0));
|
| + __ mov(CTX, Operand(R0));
|
| __ StoreIntoObject(R2,
|
| FieldAddress(R2, GrowableObjectArray::data_offset()),
|
| R1);
|
|
|