Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(470)

Unified Diff: runtime/vm/assembler_arm_test.cc

Issue 292433008: Allows unboxed doubles to be disabled. (Closed) Base URL: http://dart.googlecode.com/svn/branches/bleeding_edge/dart/
Patch Set: Created 6 years, 7 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « runtime/vm/assembler_arm.cc ('k') | runtime/vm/constants_arm.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: runtime/vm/assembler_arm_test.cc
===================================================================
--- runtime/vm/assembler_arm_test.cc (revision 36258)
+++ runtime/vm/assembler_arm_test.cc (working copy)
@@ -98,17 +98,21 @@
ASSEMBLER_TEST_GENERATE(Vmov, assembler) {
- __ mov(R3, ShifterOperand(43));
- __ mov(R1, ShifterOperand(41));
- __ vmovsrr(S1, R1, R3); // S1:S2 = 41:43
- __ vmovs(S0, S2); // S0 = S2, S0:S1 == 43:41
- __ vmovd(D2, D0); // D2 = D0, S4:S5 == 43:41
- __ vmovrs(R3, S5); // R3 = S5, R3 == 41
- __ vmovrrs(R1, R2, S4); // R1:R2 = S4:S5, R1:R2 == 43:41
- __ vmovdrr(D3, R3, R2); // D3 = R3:R2, S6:S7 == 41:41
- __ vmovsr(S7, R1); // S7 = R1, S6:S7 == 41:43
- __ vmovrrd(R0, R1, D3); // R0:R1 = D3, R0:R1 == 41:43
- __ sub(R0, R1, ShifterOperand(R0)); // 43-41
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ mov(R3, ShifterOperand(43));
+ __ mov(R1, ShifterOperand(41));
+ __ vmovsrr(S1, R1, R3); // S1:S2 = 41:43
+ __ vmovs(S0, S2); // S0 = S2, S0:S1 == 43:41
+ __ vmovd(D2, D0); // D2 = D0, S4:S5 == 43:41
+ __ vmovrs(R3, S5); // R3 = S5, R3 == 41
+ __ vmovrrs(R1, R2, S4); // R1:R2 = S4:S5, R1:R2 == 43:41
+ __ vmovdrr(D3, R3, R2); // D3 = R3:R2, S6:S7 == 41:41
+ __ vmovsr(S7, R1); // S7 = R1, S6:S7 == 41:43
+ __ vmovrrd(R0, R1, D3); // R0:R1 = D3, R0:R1 == 41:43
+ __ sub(R0, R1, ShifterOperand(R0)); // 43-41
+ } else {
+ __ LoadImmediate(R0, 2);
+ }
__ bx(LR);
}
@@ -121,126 +125,175 @@
ASSEMBLER_TEST_GENERATE(SingleVLoadStore, assembler) {
- __ LoadImmediate(R0, bit_cast<int32_t, float>(12.3f));
- __ mov(R2, ShifterOperand(SP));
- __ str(R0, Address(SP, (-kWordSize * 30), Address::PreIndex));
- __ vldrs(S0, Address(R2, (-kWordSize * 30)));
- __ vadds(S0, S0, S0);
- __ vstrs(S0, Address(R2, (-kWordSize * 30)));
- __ ldr(R0, Address(SP, (kWordSize * 30), Address::PostIndex));
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadImmediate(R0, bit_cast<int32_t, float>(12.3f));
+ __ mov(R2, ShifterOperand(SP));
+ __ str(R0, Address(SP, (-kWordSize * 30), Address::PreIndex));
+ __ vldrs(S0, Address(R2, (-kWordSize * 30)));
+ __ vadds(S0, S0, S0);
+ __ vstrs(S0, Address(R2, (-kWordSize * 30)));
+ __ ldr(R0, Address(SP, (kWordSize * 30), Address::PostIndex));
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(SingleVLoadStore, test) {
EXPECT(test != NULL);
- typedef float (*SingleVLoadStore)();
- float res = EXECUTE_TEST_CODE_FLOAT(SingleVLoadStore, test->entry());
- EXPECT_FLOAT_EQ(2*12.3f, res, 0.001f);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef float (*SingleVLoadStore)();
+ float res = EXECUTE_TEST_CODE_FLOAT(SingleVLoadStore, test->entry());
+ EXPECT_FLOAT_EQ(2*12.3f, res, 0.001f);
+ } else {
+ typedef int (*SingleVLoadStore)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(SingleVLoadStore, test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(SingleVShiftLoadStore, assembler) {
- __ LoadImmediate(R0, bit_cast<int32_t, float>(12.3f));
- __ mov(R2, ShifterOperand(SP));
- // Expressing __str(R0, Address(SP, (-kWordSize * 32), Address::PreIndex));
- // as:
- __ mov(R1, ShifterOperand(kWordSize));
- __ str(R0, Address(SP, R1, LSL, 5, Address::NegPreIndex));
- __ vldrs(S0, Address(R2, (-kWordSize * 32)));
- __ vadds(S0, S0, S0);
- __ vstrs(S0, Address(R2, (-kWordSize * 32)));
- // Expressing __ldr(R0, Address(SP, (kWordSize * 32), Address::PostIndex));
- // as:
- __ ldr(R0, Address(SP, R1, LSL, 5, Address::PostIndex));
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadImmediate(R0, bit_cast<int32_t, float>(12.3f));
+ __ mov(R2, ShifterOperand(SP));
+ // Expressing __str(R0, Address(SP, (-kWordSize * 32), Address::PreIndex));
+ // as:
+ __ mov(R1, ShifterOperand(kWordSize));
+ __ str(R0, Address(SP, R1, LSL, 5, Address::NegPreIndex));
+ __ vldrs(S0, Address(R2, (-kWordSize * 32)));
+ __ vadds(S0, S0, S0);
+ __ vstrs(S0, Address(R2, (-kWordSize * 32)));
+ // Expressing __ldr(R0, Address(SP, (kWordSize * 32), Address::PostIndex));
+ // as:
+ __ ldr(R0, Address(SP, R1, LSL, 5, Address::PostIndex));
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(SingleVShiftLoadStore, test) {
EXPECT(test != NULL);
- typedef float (*SingleVLoadStore)();
- float res = EXECUTE_TEST_CODE_FLOAT(SingleVLoadStore, test->entry());
- EXPECT_FLOAT_EQ(2*12.3f, res, 0.001f);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef float (*SingleVLoadStore)();
+ float res = EXECUTE_TEST_CODE_FLOAT(SingleVLoadStore, test->entry());
+ EXPECT_FLOAT_EQ(2*12.3f, res, 0.001f);
+ } else {
+ typedef int (*SingleVLoadStore)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(SingleVLoadStore, test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(DoubleVLoadStore, assembler) {
- int64_t value = bit_cast<int64_t, double>(12.3);
- __ LoadImmediate(R0, Utils::Low32Bits(value));
- __ LoadImmediate(R1, Utils::High32Bits(value));
- __ mov(R2, ShifterOperand(SP));
- __ str(R0, Address(SP, (-kWordSize * 30), Address::PreIndex));
- __ str(R1, Address(R2, (-kWordSize * 29)));
- __ vldrd(D0, Address(R2, (-kWordSize * 30)));
- __ vaddd(D0, D0, D0);
- __ vstrd(D0, Address(R2, (-kWordSize * 30)));
- __ ldr(R1, Address(R2, (-kWordSize * 29)));
- __ ldr(R0, Address(SP, (kWordSize * 30), Address::PostIndex));
+ if (TargetCPUFeatures::vfp_supported()) {
+ int64_t value = bit_cast<int64_t, double>(12.3);
+ __ LoadImmediate(R0, Utils::Low32Bits(value));
+ __ LoadImmediate(R1, Utils::High32Bits(value));
+ __ mov(R2, ShifterOperand(SP));
+ __ str(R0, Address(SP, (-kWordSize * 30), Address::PreIndex));
+ __ str(R1, Address(R2, (-kWordSize * 29)));
+ __ vldrd(D0, Address(R2, (-kWordSize * 30)));
+ __ vaddd(D0, D0, D0);
+ __ vstrd(D0, Address(R2, (-kWordSize * 30)));
+ __ ldr(R1, Address(R2, (-kWordSize * 29)));
+ __ ldr(R0, Address(SP, (kWordSize * 30), Address::PostIndex));
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(DoubleVLoadStore, test) {
EXPECT(test != NULL);
- typedef double (*DoubleVLoadStore)();
- double res = EXECUTE_TEST_CODE_DOUBLE(DoubleVLoadStore, test->entry());
- EXPECT_FLOAT_EQ(2*12.3, res, 0.001);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef double (*DoubleVLoadStore)();
+ float res = EXECUTE_TEST_CODE_DOUBLE(DoubleVLoadStore, test->entry());
+ EXPECT_FLOAT_EQ(2*12.3f, res, 0.001f);
+ } else {
+ typedef int (*DoubleVLoadStore)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(DoubleVLoadStore, test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(SingleFPOperations, assembler) {
- __ LoadSImmediate(S0, 12.3f);
- __ LoadSImmediate(S1, 3.4f);
- __ vnegs(S0, S0); // -12.3f
- __ vabss(S0, S0); // 12.3f
- __ vadds(S0, S0, S1); // 15.7f
- __ vmuls(S0, S0, S1); // 53.38f
- __ vsubs(S0, S0, S1); // 49.98f
- __ vdivs(S0, S0, S1); // 14.7f
- __ vsqrts(S0, S0); // 3.8340579f
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadSImmediate(S0, 12.3f);
+ __ LoadSImmediate(S1, 3.4f);
+ __ vnegs(S0, S0); // -12.3f
+ __ vabss(S0, S0); // 12.3f
+ __ vadds(S0, S0, S1); // 15.7f
+ __ vmuls(S0, S0, S1); // 53.38f
+ __ vsubs(S0, S0, S1); // 49.98f
+ __ vdivs(S0, S0, S1); // 14.7f
+ __ vsqrts(S0, S0); // 3.8340579f
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(SingleFPOperations, test) {
EXPECT(test != NULL);
- typedef float (*SingleFPOperations)();
- float res = EXECUTE_TEST_CODE_FLOAT(SingleFPOperations, test->entry());
- EXPECT_FLOAT_EQ(3.8340579f, res, 0.001f);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef float (*SingleFPOperations)();
+ float res = EXECUTE_TEST_CODE_FLOAT(SingleFPOperations, test->entry());
+ EXPECT_FLOAT_EQ(3.8340579f, res, 0.001f);
+ } else {
+ typedef int (*SingleFPOperations)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(SingleFPOperations, test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(DoubleFPOperations, assembler) {
- __ LoadDImmediate(D0, 12.3, R0);
- __ LoadDImmediate(D1, 3.4, R0);
- __ vnegd(D0, D0); // -12.3
- __ vabsd(D0, D0); // 12.3
- __ vaddd(D0, D0, D1); // 15.7
- __ vmuld(D0, D0, D1); // 53.38
- __ vsubd(D0, D0, D1); // 49.98
- __ vdivd(D0, D0, D1); // 14.7
- __ vsqrtd(D0, D0); // 3.8340579
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadDImmediate(D0, 12.3, R0);
+ __ LoadDImmediate(D1, 3.4, R0);
+ __ vnegd(D0, D0); // -12.3
+ __ vabsd(D0, D0); // 12.3
+ __ vaddd(D0, D0, D1); // 15.7
+ __ vmuld(D0, D0, D1); // 53.38
+ __ vsubd(D0, D0, D1); // 49.98
+ __ vdivd(D0, D0, D1); // 14.7
+ __ vsqrtd(D0, D0); // 3.8340579
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(DoubleFPOperations, test) {
EXPECT(test != NULL);
- typedef double (*DoubleFPOperations)();
- double res = EXECUTE_TEST_CODE_DOUBLE(DoubleFPOperations, test->entry());
- EXPECT_FLOAT_EQ(3.8340579, res, 0.001);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef double (*DoubleFPOperations)();
+ double res = EXECUTE_TEST_CODE_DOUBLE(DoubleFPOperations, test->entry());
+ EXPECT_FLOAT_EQ(3.8340579, res, 0.001);
+ } else {
+ typedef int (*DoubleFPOperations)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(DoubleFPOperations, test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(DoubleSqrtNeg, assembler) {
- // Check that sqrt of a negative double gives NaN.
- __ LoadDImmediate(D1, -1.0, R0);
- __ vsqrtd(D0, D1);
- __ vcmpd(D0, D0);
- __ vmstat();
- __ mov(R0, ShifterOperand(1), VS);
- __ mov(R0, ShifterOperand(0), VC);
+ if (TargetCPUFeatures::vfp_supported()) {
+ // Check that sqrt of a negative double gives NaN.
+ __ LoadDImmediate(D1, -1.0, R0);
+ __ vsqrtd(D0, D1);
+ __ vcmpd(D0, D0);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(1), VS);
+ __ mov(R0, ShifterOperand(0), VC);
+ } else {
+ __ LoadImmediate(R0, 1);
+ }
__ Ret();
}
@@ -253,102 +306,151 @@
ASSEMBLER_TEST_GENERATE(IntToDoubleConversion, assembler) {
- __ mov(R3, ShifterOperand(6));
- __ vmovsr(S3, R3);
- __ vcvtdi(D0, S3);
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ mov(R3, ShifterOperand(6));
+ __ vmovsr(S3, R3);
+ __ vcvtdi(D0, S3);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(IntToDoubleConversion, test) {
- typedef double (*IntToDoubleConversionCode)();
EXPECT(test != NULL);
- double res = EXECUTE_TEST_CODE_DOUBLE(IntToDoubleConversionCode,
- test->entry());
- EXPECT_FLOAT_EQ(6.0, res, 0.001);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef double (*IntToDoubleConversionCode)();
+ double res = EXECUTE_TEST_CODE_DOUBLE(IntToDoubleConversionCode,
+ test->entry());
+ EXPECT_FLOAT_EQ(6.0, res, 0.001);
+ } else {
+ typedef int (*IntToDoubleConversionCode)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(IntToDoubleConversionCode,
+ test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(LongToDoubleConversion, assembler) {
- int64_t value = 60000000000LL;
- __ LoadImmediate(R0, Utils::Low32Bits(value));
- __ LoadImmediate(R1, Utils::High32Bits(value));
- __ vmovsr(S0, R0);
- __ vmovsr(S2, R1);
- __ vcvtdu(D0, S0);
- __ vcvtdi(D1, S2);
- __ LoadDImmediate(D2, 1.0 * (1LL << 32), R0);
- __ vmlad(D0, D1, D2);
+ if (TargetCPUFeatures::vfp_supported()) {
+ int64_t value = 60000000000LL;
+ __ LoadImmediate(R0, Utils::Low32Bits(value));
+ __ LoadImmediate(R1, Utils::High32Bits(value));
+ __ vmovsr(S0, R0);
+ __ vmovsr(S2, R1);
+ __ vcvtdu(D0, S0);
+ __ vcvtdi(D1, S2);
+ __ LoadDImmediate(D2, 1.0 * (1LL << 32), R0);
+ __ vmlad(D0, D1, D2);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(LongToDoubleConversion, test) {
- typedef double (*LongToDoubleConversionCode)();
EXPECT(test != NULL);
- double res = EXECUTE_TEST_CODE_DOUBLE(LongToDoubleConversionCode,
- test->entry());
- EXPECT_FLOAT_EQ(60000000000.0, res, 0.001);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef double (*LongToDoubleConversionCode)();
+ double res = EXECUTE_TEST_CODE_DOUBLE(LongToDoubleConversionCode,
+ test->entry());
+ EXPECT_FLOAT_EQ(60000000000.0, res, 0.001);
+ } else {
+ typedef int (*LongToDoubleConversionCode)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(LongToDoubleConversionCode,
+ test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(IntToFloatConversion, assembler) {
- __ mov(R3, ShifterOperand(6));
- __ vmovsr(S3, R3);
- __ vcvtsi(S0, S3);
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ mov(R3, ShifterOperand(6));
+ __ vmovsr(S3, R3);
+ __ vcvtsi(S0, S3);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(IntToFloatConversion, test) {
- typedef float (*IntToFloatConversionCode)();
EXPECT(test != NULL);
- float res = EXECUTE_TEST_CODE_FLOAT(IntToFloatConversionCode, test->entry());
- EXPECT_FLOAT_EQ(6.0, res, 0.001);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef float (*IntToFloatConversionCode)();
+ float res = EXECUTE_TEST_CODE_FLOAT(IntToFloatConversionCode,
+ test->entry());
+ EXPECT_FLOAT_EQ(6.0, res, 0.001);
+ } else {
+ typedef int (*IntToFloatConversionCode)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(IntToFloatConversionCode,
+ test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(FloatToIntConversion, assembler) {
- __ vcvtis(S1, S0);
- __ vmovrs(R0, S1);
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ vcvtis(S1, S0);
+ __ vmovrs(R0, S1);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(FloatToIntConversion, test) {
- typedef int (*FloatToIntConversion)(float arg);
EXPECT(test != NULL);
- EXPECT_EQ(12,
- EXECUTE_TEST_CODE_INT32_F(FloatToIntConversion, test->entry(),
- 12.8f));
- EXPECT_EQ(INT_MIN,
- EXECUTE_TEST_CODE_INT32_F(FloatToIntConversion, test->entry(),
- -FLT_MAX));
- EXPECT_EQ(INT_MAX,
- EXECUTE_TEST_CODE_INT32_F(FloatToIntConversion, test->entry(),
- FLT_MAX));
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef int (*FloatToIntConversion)(float arg);
+ EXPECT_EQ(12,
+ EXECUTE_TEST_CODE_INT32_F(FloatToIntConversion, test->entry(),
+ 12.8f));
+ EXPECT_EQ(INT_MIN,
+ EXECUTE_TEST_CODE_INT32_F(FloatToIntConversion, test->entry(),
+ -FLT_MAX));
+ EXPECT_EQ(INT_MAX,
+ EXECUTE_TEST_CODE_INT32_F(FloatToIntConversion, test->entry(),
+ FLT_MAX));
+ } else {
+ typedef int (*FloatToIntConversion)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(FloatToIntConversion, test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(DoubleToIntConversion, assembler) {
- __ vcvtid(S0, D0);
- __ vmovrs(R0, S0);
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ vcvtid(S0, D0);
+ __ vmovrs(R0, S0);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(DoubleToIntConversion, test) {
- typedef int (*DoubleToIntConversion)(double arg);
- EXPECT(test != NULL);
- EXPECT_EQ(12,
- EXECUTE_TEST_CODE_INT32_D(DoubleToIntConversion, test->entry(),
- 12.8));
- EXPECT_EQ(INT_MIN,
- EXECUTE_TEST_CODE_INT32_D(DoubleToIntConversion, test->entry(),
- -DBL_MAX));
- EXPECT_EQ(INT_MAX,
- EXECUTE_TEST_CODE_INT32_D(DoubleToIntConversion, test->entry(),
- DBL_MAX));
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef int (*DoubleToIntConversion)(double arg);
+ EXPECT(test != NULL);
+ EXPECT_EQ(12,
+ EXECUTE_TEST_CODE_INT32_D(DoubleToIntConversion, test->entry(),
+ 12.8));
+ EXPECT_EQ(INT_MIN,
+ EXECUTE_TEST_CODE_INT32_D(DoubleToIntConversion, test->entry(),
+ -DBL_MAX));
+ EXPECT_EQ(INT_MAX,
+ EXECUTE_TEST_CODE_INT32_D(DoubleToIntConversion, test->entry(),
+ DBL_MAX));
+ } else {
+ typedef int (*DoubleToIntConversion)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(DoubleToIntConversion, test->entry()));
+ }
}
@@ -369,43 +471,57 @@
ASSEMBLER_TEST_GENERATE(DoubleToFloatConversion, assembler) {
- __ LoadDImmediate(D1, 12.8, R0);
- __ vcvtsd(S0, D1);
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadDImmediate(D1, 12.8, R0);
+ __ vcvtsd(S0, D1);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(DoubleToFloatConversion, test) {
- typedef float (*DoubleToFloatConversionCode)();
EXPECT(test != NULL);
- float res = EXECUTE_TEST_CODE_FLOAT(DoubleToFloatConversionCode,
- test->entry());
- EXPECT_FLOAT_EQ(12.8, res, 0.001);
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef float (*DoubleToFloatConversionCode)();
+ float res = EXECUTE_TEST_CODE_FLOAT(DoubleToFloatConversionCode,
+ test->entry());
+ EXPECT_FLOAT_EQ(12.8, res, 0.001);
+ } else {
+ typedef int (*DoubleToFloatConversionCode)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(DoubleToFloatConversionCode,
+ test->entry()));
+ }
}
ASSEMBLER_TEST_GENERATE(FloatCompare, assembler) {
- // Test 12.3f vs 12.5f.
- __ LoadSImmediate(S0, 12.3f);
- __ LoadSImmediate(S1, 12.5f);
+ if (TargetCPUFeatures::vfp_supported()) {
+ // Test 12.3f vs 12.5f.
+ __ LoadSImmediate(S0, 12.3f);
+ __ LoadSImmediate(S1, 12.5f);
- // Count errors in R0. R0 is zero if no errors found.
- __ mov(R0, ShifterOperand(0));
- __ vcmps(S0, S1);
- __ vmstat();
- __ add(R0, R0, ShifterOperand(1), VS); // Error if unordered (Nan).
- __ add(R0, R0, ShifterOperand(2), GT); // Error if greater.
- __ add(R0, R0, ShifterOperand(4), EQ); // Error if equal.
- __ add(R0, R0, ShifterOperand(8), PL); // Error if not less.
+ // Count errors in R0. R0 is zero if no errors found.
+ __ mov(R0, ShifterOperand(0));
+ __ vcmps(S0, S1);
+ __ vmstat();
+ __ add(R0, R0, ShifterOperand(1), VS); // Error if unordered (Nan).
+ __ add(R0, R0, ShifterOperand(2), GT); // Error if greater.
+ __ add(R0, R0, ShifterOperand(4), EQ); // Error if equal.
+ __ add(R0, R0, ShifterOperand(8), PL); // Error if not less.
- // Test NaN.
- // Create NaN by dividing 0.0f/0.0f.
- __ LoadSImmediate(S1, 0.0f);
- __ vdivs(S1, S1, S1);
- __ vcmps(S1, S1);
- __ vmstat();
- __ add(R0, R0, ShifterOperand(16), VC); // Error if not unordered (not Nan).
-
+ // Test NaN.
+ // Create NaN by dividing 0.0f/0.0f.
+ __ LoadSImmediate(S1, 0.0f);
+ __ vdivs(S1, S1, S1);
+ __ vcmps(S1, S1);
+ __ vmstat();
+ // Error if not unordered (not Nan).
+ __ add(R0, R0, ShifterOperand(16), VC);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
// R0 is 0 if all tests passed.
__ bx(LR);
}
@@ -419,27 +535,31 @@
ASSEMBLER_TEST_GENERATE(DoubleCompare, assembler) {
- // Test 12.3 vs 12.5.
- __ LoadDImmediate(D0, 12.3, R1);
- __ LoadDImmediate(D1, 12.5, R1);
+ if (TargetCPUFeatures::vfp_supported()) {
+ // Test 12.3 vs 12.5.
+ __ LoadDImmediate(D0, 12.3, R1);
+ __ LoadDImmediate(D1, 12.5, R1);
- // Count errors in R0. R0 is zero if no errors found.
- __ mov(R0, ShifterOperand(0));
- __ vcmpd(D0, D1);
- __ vmstat();
- __ add(R0, R0, ShifterOperand(1), VS); // Error if unordered (Nan).
- __ add(R0, R0, ShifterOperand(2), GT); // Error if greater.
- __ add(R0, R0, ShifterOperand(4), EQ); // Error if equal.
- __ add(R0, R0, ShifterOperand(8), PL); // Error if not less.
+ // Count errors in R0. R0 is zero if no errors found.
+ __ mov(R0, ShifterOperand(0));
+ __ vcmpd(D0, D1);
+ __ vmstat();
+ __ add(R0, R0, ShifterOperand(1), VS); // Error if unordered (Nan).
+ __ add(R0, R0, ShifterOperand(2), GT); // Error if greater.
+ __ add(R0, R0, ShifterOperand(4), EQ); // Error if equal.
+ __ add(R0, R0, ShifterOperand(8), PL); // Error if not less.
- // Test NaN.
- // Create NaN by dividing 0.0/0.0.
- __ LoadDImmediate(D1, 0.0, R1);
- __ vdivd(D1, D1, D1);
- __ vcmpd(D1, D1);
- __ vmstat();
- __ add(R0, R0, ShifterOperand(16), VC); // Error if not unordered (not Nan).
-
+ // Test NaN.
+ // Create NaN by dividing 0.0/0.0.
+ __ LoadDImmediate(D1, 0.0, R1);
+ __ vdivd(D1, D1, D1);
+ __ vcmpd(D1, D1);
+ __ vmstat();
+ // Error if not unordered (not Nan).
+ __ add(R0, R0, ShifterOperand(16), VC);
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
// R0 is 0 if all tests passed.
__ bx(LR);
}
@@ -563,48 +683,6 @@
}
-ASSEMBLER_TEST_GENERATE(Semaphore, assembler) {
- __ mov(R0, ShifterOperand(40));
- __ mov(R1, ShifterOperand(42));
- __ Push(R0);
- Label retry;
- __ Bind(&retry);
- __ ldrex(R0, SP);
- __ strex(IP, R1, SP); // IP == 0, success
- __ tst(IP, ShifterOperand(0));
- __ b(&retry, NE); // NE if context switch occurred between ldrex and strex.
- __ Pop(R0); // 42
- __ bx(LR);
-}
-
-
-ASSEMBLER_TEST_RUN(Semaphore, test) {
- EXPECT(test != NULL);
- typedef int (*Semaphore)();
- EXPECT_EQ(42, EXECUTE_TEST_CODE_INT32(Semaphore, test->entry()));
-}
-
-
-ASSEMBLER_TEST_GENERATE(FailedSemaphore, assembler) {
- __ mov(R0, ShifterOperand(40));
- __ mov(R1, ShifterOperand(42));
- __ Push(R0);
- __ ldrex(R0, SP);
- __ clrex(); // Simulate a context switch.
- __ strex(IP, R1, SP); // IP == 1, failure
- __ Pop(R0); // 40
- __ add(R0, R0, ShifterOperand(IP));
- __ bx(LR);
-}
-
-
-ASSEMBLER_TEST_RUN(FailedSemaphore, test) {
- EXPECT(test != NULL);
- typedef int (*FailedSemaphore)();
- EXPECT_EQ(41, EXECUTE_TEST_CODE_INT32(FailedSemaphore, test->entry()));
-}
-
-
ASSEMBLER_TEST_GENERATE(AndOrr, assembler) {
__ mov(R1, ShifterOperand(40));
__ mov(R2, ShifterOperand(0));
@@ -658,24 +736,33 @@
ASSEMBLER_TEST_GENERATE(QuotientRemainder, assembler) {
- __ vmovsr(S2, R0);
- __ vmovsr(S4, R2);
- __ vcvtdi(D1, S2);
- __ vcvtdi(D2, S4);
- __ vdivd(D0, D1, D2);
- __ vcvtid(S0, D0);
- __ vmovrs(R1, S0); // r1 = r0/r2
- __ mls(R0, R1, R2, R0); // r0 = r0 - r1*r2
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ vmovsr(S2, R0);
+ __ vmovsr(S4, R2);
+ __ vcvtdi(D1, S2);
+ __ vcvtdi(D2, S4);
+ __ vdivd(D0, D1, D2);
+ __ vcvtid(S0, D0);
+ __ vmovrs(R1, S0); // r1 = r0/r2
+ __ mls(R0, R1, R2, R0); // r0 = r0 - r1*r2
+ } else {
+ __ LoadImmediate(R0, 0);
+ }
__ bx(LR);
}
ASSEMBLER_TEST_RUN(QuotientRemainder, test) {
EXPECT(test != NULL);
- typedef int64_t (*QuotientRemainder)(int64_t dividend, int64_t divisor);
- EXPECT_EQ(0x1000400000da8LL,
- EXECUTE_TEST_CODE_INT64_LL(QuotientRemainder, test->entry(),
- 0x12345678, 0x1234));
+ if (TargetCPUFeatures::vfp_supported()) {
+ typedef int64_t (*QuotientRemainder)(int64_t dividend, int64_t divisor);
+ EXPECT_EQ(0x1000400000da8LL,
+ EXECUTE_TEST_CODE_INT64_LL(QuotientRemainder, test->entry(),
+ 0x12345678, 0x1234));
+ } else {
+ typedef int (*QuotientRemainder)();
+ EXPECT_EQ(0, EXECUTE_TEST_CODE_INT32(QuotientRemainder, test->entry()));
+ }
}
@@ -1175,47 +1262,50 @@
// Make sure we can store and reload the D registers using vstmd and vldmd
ASSEMBLER_TEST_GENERATE(VstmdVldmd, assembler) {
- __ LoadDImmediate(D0, 0.0, R0);
- __ LoadDImmediate(D1, 1.0, R0);
- __ LoadDImmediate(D2, 2.0, R0);
- __ LoadDImmediate(D3, 3.0, R0);
- __ LoadDImmediate(D4, 4.0, R0);
- __ vstmd(DB_W, SP, D0, 5); // Push D0 - D4 onto the stack, dec SP
- __ LoadDImmediate(D0, 0.0, R0);
- __ LoadDImmediate(D1, 0.0, R0);
- __ LoadDImmediate(D2, 0.0, R0);
- __ LoadDImmediate(D3, 0.0, R0);
- __ LoadDImmediate(D4, 0.0, R0);
- __ vldmd(IA_W, SP, D0, 5); // Pop stack into D0 - D4, inc SP
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadDImmediate(D0, 0.0, R0);
+ __ LoadDImmediate(D1, 1.0, R0);
+ __ LoadDImmediate(D2, 2.0, R0);
+ __ LoadDImmediate(D3, 3.0, R0);
+ __ LoadDImmediate(D4, 4.0, R0);
+ __ vstmd(DB_W, SP, D0, 5); // Push D0 - D4 onto the stack, dec SP
+ __ LoadDImmediate(D0, 0.0, R0);
+ __ LoadDImmediate(D1, 0.0, R0);
+ __ LoadDImmediate(D2, 0.0, R0);
+ __ LoadDImmediate(D3, 0.0, R0);
+ __ LoadDImmediate(D4, 0.0, R0);
+ __ vldmd(IA_W, SP, D0, 5); // Pop stack into D0 - D4, inc SP
- // Load success value into R0
- __ mov(R0, ShifterOperand(42));
+ // Load success value into R0
+ __ mov(R0, ShifterOperand(42));
- // Check that 4.0 is back in D4
- __ LoadDImmediate(D5, 4.0, R1);
- __ vcmpd(D4, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 4.0 is back in D4
+ __ LoadDImmediate(D5, 4.0, R1);
+ __ vcmpd(D4, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 3.0 is back in D3
- __ LoadDImmediate(D5, 3.0, R1);
- __ vcmpd(D3, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 3.0 is back in D3
+ __ LoadDImmediate(D5, 3.0, R1);
+ __ vcmpd(D3, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 2.0 is back in D2
- __ LoadDImmediate(D5, 2.0, R1);
- __ vcmpd(D2, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 2.0 is back in D2
+ __ LoadDImmediate(D5, 2.0, R1);
+ __ vcmpd(D2, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 1.0 is back in D1
- __ LoadDImmediate(D5, 1.0, R1);
- __ vcmpd(D1, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
-
+ // Check that 1.0 is back in D1
+ __ LoadDImmediate(D5, 1.0, R1);
+ __ vcmpd(D1, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ } else {
+ __ LoadImmediate(R0, 42);
+ }
__ bx(LR);
}
@@ -1229,46 +1319,49 @@
// Make sure we can store and reload the S registers using vstms and vldms
ASSEMBLER_TEST_GENERATE(VstmsVldms, assembler) {
- __ LoadSImmediate(S0, 0.0);
- __ LoadSImmediate(S1, 1.0);
- __ LoadSImmediate(S2, 2.0);
- __ LoadSImmediate(S3, 3.0);
- __ LoadSImmediate(S4, 4.0);
- __ vstms(DB_W, SP, S0, S4); // Push S0 - S4 onto the stack, dec SP
- __ LoadSImmediate(S0, 0.0);
- __ LoadSImmediate(S1, 0.0);
- __ LoadSImmediate(S2, 0.0);
- __ LoadSImmediate(S3, 0.0);
- __ LoadSImmediate(S4, 0.0);
- __ vldms(IA_W, SP, S0, S4); // Pop stack into S0 - S4, inc SP
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadSImmediate(S0, 0.0);
+ __ LoadSImmediate(S1, 1.0);
+ __ LoadSImmediate(S2, 2.0);
+ __ LoadSImmediate(S3, 3.0);
+ __ LoadSImmediate(S4, 4.0);
+ __ vstms(DB_W, SP, S0, S4); // Push S0 - S4 onto the stack, dec SP
+ __ LoadSImmediate(S0, 0.0);
+ __ LoadSImmediate(S1, 0.0);
+ __ LoadSImmediate(S2, 0.0);
+ __ LoadSImmediate(S3, 0.0);
+ __ LoadSImmediate(S4, 0.0);
+ __ vldms(IA_W, SP, S0, S4); // Pop stack into S0 - S4, inc SP
- // Load success value into R0
- __ mov(R0, ShifterOperand(42));
+ // Load success value into R0
+ __ mov(R0, ShifterOperand(42));
- // Check that 4.0 is back in S4
- __ LoadSImmediate(S5, 4.0);
- __ vcmps(S4, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 4.0 is back in S4
+ __ LoadSImmediate(S5, 4.0);
+ __ vcmps(S4, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 3.0 is back in S3
- __ LoadSImmediate(S5, 3.0);
- __ vcmps(S3, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 3.0 is back in S3
+ __ LoadSImmediate(S5, 3.0);
+ __ vcmps(S3, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 2.0 is back in S2
- __ LoadSImmediate(S5, 2.0);
- __ vcmps(S2, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 2.0 is back in S2
+ __ LoadSImmediate(S5, 2.0);
+ __ vcmps(S2, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 1.0 is back in S1
- __ LoadSImmediate(S5, 1.0);
- __ vcmps(S1, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
-
+ // Check that 1.0 is back in S1
+ __ LoadSImmediate(S5, 1.0);
+ __ vcmps(S1, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ } else {
+ __ LoadImmediate(R0, 42);
+ }
__ bx(LR);
}
@@ -1282,45 +1375,48 @@
// Make sure we can start somewhere other than D0
ASSEMBLER_TEST_GENERATE(VstmdVldmd1, assembler) {
- __ LoadDImmediate(D1, 1.0, R0);
- __ LoadDImmediate(D2, 2.0, R0);
- __ LoadDImmediate(D3, 3.0, R0);
- __ LoadDImmediate(D4, 4.0, R0);
- __ vstmd(DB_W, SP, D1, 4); // Push D1 - D4 onto the stack, dec SP
- __ LoadDImmediate(D1, 0.0, R0);
- __ LoadDImmediate(D2, 0.0, R0);
- __ LoadDImmediate(D3, 0.0, R0);
- __ LoadDImmediate(D4, 0.0, R0);
- __ vldmd(IA_W, SP, D1, 4); // Pop stack into D1 - D4, inc SP
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadDImmediate(D1, 1.0, R0);
+ __ LoadDImmediate(D2, 2.0, R0);
+ __ LoadDImmediate(D3, 3.0, R0);
+ __ LoadDImmediate(D4, 4.0, R0);
+ __ vstmd(DB_W, SP, D1, 4); // Push D1 - D4 onto the stack, dec SP
+ __ LoadDImmediate(D1, 0.0, R0);
+ __ LoadDImmediate(D2, 0.0, R0);
+ __ LoadDImmediate(D3, 0.0, R0);
+ __ LoadDImmediate(D4, 0.0, R0);
+ __ vldmd(IA_W, SP, D1, 4); // Pop stack into D1 - D4, inc SP
- // Load success value into R0
- __ mov(R0, ShifterOperand(42));
+ // Load success value into R0
+ __ mov(R0, ShifterOperand(42));
- // Check that 4.0 is back in D4
- __ LoadDImmediate(D5, 4.0, R1);
- __ vcmpd(D4, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 4.0 is back in D4
+ __ LoadDImmediate(D5, 4.0, R1);
+ __ vcmpd(D4, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 3.0 is back in D3
- __ LoadDImmediate(D5, 3.0, R1);
- __ vcmpd(D3, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 3.0 is back in D3
+ __ LoadDImmediate(D5, 3.0, R1);
+ __ vcmpd(D3, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 2.0 is back in D2
- __ LoadDImmediate(D5, 2.0, R1);
- __ vcmpd(D2, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 2.0 is back in D2
+ __ LoadDImmediate(D5, 2.0, R1);
+ __ vcmpd(D2, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 1.0 is back in D1
- __ LoadDImmediate(D5, 1.0, R1);
- __ vcmpd(D1, D5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
-
+ // Check that 1.0 is back in D1
+ __ LoadDImmediate(D5, 1.0, R1);
+ __ vcmpd(D1, D5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ } else {
+ __ LoadImmediate(R0, 42);
+ }
__ bx(LR);
}
@@ -1334,44 +1430,47 @@
// Make sure we can start somewhere other than S0
ASSEMBLER_TEST_GENERATE(VstmsVldms1, assembler) {
- __ LoadSImmediate(S1, 1.0);
- __ LoadSImmediate(S2, 2.0);
- __ LoadSImmediate(S3, 3.0);
- __ LoadSImmediate(S4, 4.0);
- __ vstms(DB_W, SP, S1, S4); // Push S0 - S4 onto the stack, dec SP
- __ LoadSImmediate(S1, 0.0);
- __ LoadSImmediate(S2, 0.0);
- __ LoadSImmediate(S3, 0.0);
- __ LoadSImmediate(S4, 0.0);
- __ vldms(IA_W, SP, S1, S4); // Pop stack into S0 - S4, inc SP
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadSImmediate(S1, 1.0);
+ __ LoadSImmediate(S2, 2.0);
+ __ LoadSImmediate(S3, 3.0);
+ __ LoadSImmediate(S4, 4.0);
+ __ vstms(DB_W, SP, S1, S4); // Push S0 - S4 onto the stack, dec SP
+ __ LoadSImmediate(S1, 0.0);
+ __ LoadSImmediate(S2, 0.0);
+ __ LoadSImmediate(S3, 0.0);
+ __ LoadSImmediate(S4, 0.0);
+ __ vldms(IA_W, SP, S1, S4); // Pop stack into S0 - S4, inc SP
- // Load success value into R0
- __ mov(R0, ShifterOperand(42));
+ // Load success value into R0
+ __ mov(R0, ShifterOperand(42));
- // Check that 4.0 is back in S4
- __ LoadSImmediate(S5, 4.0);
- __ vcmps(S4, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 4.0 is back in S4
+ __ LoadSImmediate(S5, 4.0);
+ __ vcmps(S4, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 3.0 is back in S3
- __ LoadSImmediate(S5, 3.0);
- __ vcmps(S3, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 3.0 is back in S3
+ __ LoadSImmediate(S5, 3.0);
+ __ vcmps(S3, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 2.0 is back in S2
- __ LoadSImmediate(S5, 2.0);
- __ vcmps(S2, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 2.0 is back in S2
+ __ LoadSImmediate(S5, 2.0);
+ __ vcmps(S2, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 1.0 is back in S1
- __ LoadSImmediate(S5, 1.0);
- __ vcmps(S1, S5);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
-
+ // Check that 1.0 is back in S1
+ __ LoadSImmediate(S5, 1.0);
+ __ vcmps(S1, S5);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ } else {
+ __ LoadImmediate(R0, 42);
+ }
__ bx(LR);
}
@@ -1386,53 +1485,57 @@
// Make sure we can store the D registers using vstmd and
// load them into a different set using vldmd
ASSEMBLER_TEST_GENERATE(VstmdVldmd_off, assembler) {
- // Save used callee-saved FPU registers.
- __ vstmd(DB_W, SP, D8, 3);
- __ LoadDImmediate(D0, 0.0, R0);
- __ LoadDImmediate(D1, 1.0, R0);
- __ LoadDImmediate(D2, 2.0, R0);
- __ LoadDImmediate(D3, 3.0, R0);
- __ LoadDImmediate(D4, 4.0, R0);
- __ LoadDImmediate(D5, 5.0, R0);
- __ vstmd(DB_W, SP, D0, 5); // Push D0 - D4 onto the stack, dec SP
- __ vldmd(IA_W, SP, D5, 5); // Pop stack into D5 - D9, inc SP
+ if (TargetCPUFeatures::vfp_supported()) {
+ // Save used callee-saved FPU registers.
+ __ vstmd(DB_W, SP, D8, 3);
+ __ LoadDImmediate(D0, 0.0, R0);
+ __ LoadDImmediate(D1, 1.0, R0);
+ __ LoadDImmediate(D2, 2.0, R0);
+ __ LoadDImmediate(D3, 3.0, R0);
+ __ LoadDImmediate(D4, 4.0, R0);
+ __ LoadDImmediate(D5, 5.0, R0);
+ __ vstmd(DB_W, SP, D0, 5); // Push D0 - D4 onto the stack, dec SP
+ __ vldmd(IA_W, SP, D5, 5); // Pop stack into D5 - D9, inc SP
- // Load success value into R0
- __ mov(R0, ShifterOperand(42));
+ // Load success value into R0
+ __ mov(R0, ShifterOperand(42));
- // Check that 4.0 is in D9
- __ LoadDImmediate(D10, 4.0, R1);
- __ vcmpd(D9, D10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 4.0 is in D9
+ __ LoadDImmediate(D10, 4.0, R1);
+ __ vcmpd(D9, D10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 3.0 is in D8
- __ LoadDImmediate(D10, 3.0, R1);
- __ vcmpd(D8, D10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 3.0 is in D8
+ __ LoadDImmediate(D10, 3.0, R1);
+ __ vcmpd(D8, D10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 2.0 is in D7
- __ LoadDImmediate(D10, 2.0, R1);
- __ vcmpd(D7, D10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 2.0 is in D7
+ __ LoadDImmediate(D10, 2.0, R1);
+ __ vcmpd(D7, D10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 1.0 is in D6
- __ LoadDImmediate(D10, 1.0, R1);
- __ vcmpd(D6, D10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 1.0 is in D6
+ __ LoadDImmediate(D10, 1.0, R1);
+ __ vcmpd(D6, D10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Check that 0.0 is in D5
- __ LoadDImmediate(D10, 0.0, R1);
- __ vcmpd(D5, D10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
+ // Check that 0.0 is in D5
+ __ LoadDImmediate(D10, 0.0, R1);
+ __ vcmpd(D5, D10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure into R0 if NE
- // Restore used callee-saved FPU registers.
- __ vldmd(IA_W, SP, D8, 3);
+ // Restore used callee-saved FPU registers.
+ __ vldmd(IA_W, SP, D8, 3);
+ } else {
+ __ LoadImmediate(R0, 42);
+ }
__ bx(LR);
}
@@ -1446,48 +1549,51 @@
// Make sure we can start somewhere other than S0
ASSEMBLER_TEST_GENERATE(VstmsVldms_off, assembler) {
- __ LoadSImmediate(S0, 0.0);
- __ LoadSImmediate(S1, 1.0);
- __ LoadSImmediate(S2, 2.0);
- __ LoadSImmediate(S3, 3.0);
- __ LoadSImmediate(S4, 4.0);
- __ LoadSImmediate(S5, 5.0);
- __ vstms(DB_W, SP, S0, S4); // Push S0 - S4 onto the stack, dec SP
- __ vldms(IA_W, SP, S5, S9); // Pop stack into S5 - S9, inc SP
+ if (TargetCPUFeatures::vfp_supported()) {
+ __ LoadSImmediate(S0, 0.0);
+ __ LoadSImmediate(S1, 1.0);
+ __ LoadSImmediate(S2, 2.0);
+ __ LoadSImmediate(S3, 3.0);
+ __ LoadSImmediate(S4, 4.0);
+ __ LoadSImmediate(S5, 5.0);
+ __ vstms(DB_W, SP, S0, S4); // Push S0 - S4 onto the stack, dec SP
+ __ vldms(IA_W, SP, S5, S9); // Pop stack into S5 - S9, inc SP
- // Load success value into R0
- __ mov(R0, ShifterOperand(42));
+ // Load success value into R0
+ __ mov(R0, ShifterOperand(42));
- // Check that 4.0 is in S9
- __ LoadSImmediate(S10, 4.0);
- __ vcmps(S9, S10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 4.0 is in S9
+ __ LoadSImmediate(S10, 4.0);
+ __ vcmps(S9, S10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 3.0 is in S8
- __ LoadSImmediate(S10, 3.0);
- __ vcmps(S8, S10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 3.0 is in S8
+ __ LoadSImmediate(S10, 3.0);
+ __ vcmps(S8, S10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 2.0 is in S7
- __ LoadSImmediate(S10, 2.0);
- __ vcmps(S7, S10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 2.0 is in S7
+ __ LoadSImmediate(S10, 2.0);
+ __ vcmps(S7, S10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 1.0 is back in S6
- __ LoadSImmediate(S10, 1.0);
- __ vcmps(S6, S10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ // Check that 1.0 is back in S6
+ __ LoadSImmediate(S10, 1.0);
+ __ vcmps(S6, S10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
- // Check that 0.0 is back in S5
- __ LoadSImmediate(S10, 0.0);
- __ vcmps(S5, S10);
- __ vmstat();
- __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
-
+ // Check that 0.0 is back in S5
+ __ LoadSImmediate(S10, 0.0);
+ __ vcmps(S5, S10);
+ __ vmstat();
+ __ mov(R0, ShifterOperand(0), NE); // Put failure value into R0 if NE
+ } else {
+ __ LoadImmediate(R0, 42);
+ }
__ bx(LR);
}
@@ -1927,47 +2033,51 @@
ASSEMBLER_TEST_GENERATE(Mint_shl_ok, assembler) {
- const QRegister value = Q0;
- const QRegister temp = Q1;
- const QRegister out = Q2;
- const Register shift = R1;
- const DRegister dtemp0 = EvenDRegisterOf(temp);
- const SRegister stemp0 = EvenSRegisterOf(dtemp0);
- const DRegister dout0 = EvenDRegisterOf(out);
- const SRegister sout0 = EvenSRegisterOf(dout0);
- const SRegister sout1 = OddSRegisterOf(dout0);
- Label fail;
+ if (TargetCPUFeatures::neon_supported()) {
+ const QRegister value = Q0;
+ const QRegister temp = Q1;
+ const QRegister out = Q2;
+ const Register shift = R1;
+ const DRegister dtemp0 = EvenDRegisterOf(temp);
+ const SRegister stemp0 = EvenSRegisterOf(dtemp0);
+ const DRegister dout0 = EvenDRegisterOf(out);
+ const SRegister sout0 = EvenSRegisterOf(dout0);
+ const SRegister sout1 = OddSRegisterOf(dout0);
+ Label fail;
- // Initialize.
- __ veorq(value, value, value);
- __ veorq(temp, temp, temp);
- __ veorq(out, out, out);
- __ LoadImmediate(shift, 32);
- __ LoadImmediate(R2, 1 << 7);
- __ vmovsr(S0, R2);
+ // Initialize.
+ __ veorq(value, value, value);
+ __ veorq(temp, temp, temp);
+ __ veorq(out, out, out);
+ __ LoadImmediate(shift, 32);
+ __ LoadImmediate(R2, 1 << 7);
+ __ vmovsr(S0, R2);
- __ vmovsr(stemp0, shift); // Move the shift into the low S register.
- __ vshlqu(kWordPair, out, value, temp);
+ __ vmovsr(stemp0, shift); // Move the shift into the low S register.
+ __ vshlqu(kWordPair, out, value, temp);
- // check for overflow by shifting back and comparing.
- __ rsb(shift, shift, ShifterOperand(0));
- __ vmovsr(stemp0, shift);
- __ vshlqi(kWordPair, temp, out, temp);
- __ vceqqi(kWord, out, temp, value);
- // Low 64 bits of temp should be all 1's, otherwise temp != value and
- // we deopt.
- __ vmovrs(shift, sout0);
- __ CompareImmediate(shift, -1);
- __ b(&fail, NE);
- __ vmovrs(shift, sout1);
- __ CompareImmediate(shift, -1);
- __ b(&fail, NE);
+ // check for overflow by shifting back and comparing.
+ __ rsb(shift, shift, ShifterOperand(0));
+ __ vmovsr(stemp0, shift);
+ __ vshlqi(kWordPair, temp, out, temp);
+ __ vceqqi(kWord, out, temp, value);
+ // Low 64 bits of temp should be all 1's, otherwise temp != value and
+ // we deopt.
+ __ vmovrs(shift, sout0);
+ __ CompareImmediate(shift, -1);
+ __ b(&fail, NE);
+ __ vmovrs(shift, sout1);
+ __ CompareImmediate(shift, -1);
+ __ b(&fail, NE);
- __ LoadImmediate(R0, 1);
- __ bx(LR);
+ __ LoadImmediate(R0, 1);
+ __ bx(LR);
- __ Bind(&fail);
- __ LoadImmediate(R0, 0);
+ __ Bind(&fail);
+ __ LoadImmediate(R0, 0);
+ } else {
+ __ LoadImmediate(R0, 1);
+ }
__ bx(LR);
}
@@ -1980,47 +2090,51 @@
ASSEMBLER_TEST_GENERATE(Mint_shl_overflow, assembler) {
- const QRegister value = Q0;
- const QRegister temp = Q1;
- const QRegister out = Q2;
- const Register shift = R1;
- const DRegister dtemp0 = EvenDRegisterOf(temp);
- const SRegister stemp0 = EvenSRegisterOf(dtemp0);
- const DRegister dout0 = EvenDRegisterOf(out);
- const SRegister sout0 = EvenSRegisterOf(dout0);
- const SRegister sout1 = OddSRegisterOf(dout0);
- Label fail;
+ if (TargetCPUFeatures::neon_supported()) {
+ const QRegister value = Q0;
+ const QRegister temp = Q1;
+ const QRegister out = Q2;
+ const Register shift = R1;
+ const DRegister dtemp0 = EvenDRegisterOf(temp);
+ const SRegister stemp0 = EvenSRegisterOf(dtemp0);
+ const DRegister dout0 = EvenDRegisterOf(out);
+ const SRegister sout0 = EvenSRegisterOf(dout0);
+ const SRegister sout1 = OddSRegisterOf(dout0);
+ Label fail;
- // Initialize.
- __ veorq(value, value, value);
- __ veorq(temp, temp, temp);
- __ veorq(out, out, out);
- __ LoadImmediate(shift, 60);
- __ LoadImmediate(R2, 1 << 7);
- __ vmovsr(S0, R2);
+ // Initialize.
+ __ veorq(value, value, value);
+ __ veorq(temp, temp, temp);
+ __ veorq(out, out, out);
+ __ LoadImmediate(shift, 60);
+ __ LoadImmediate(R2, 1 << 7);
+ __ vmovsr(S0, R2);
- __ vmovsr(stemp0, shift); // Move the shift into the low S register.
- __ vshlqu(kWordPair, out, value, temp);
+ __ vmovsr(stemp0, shift); // Move the shift into the low S register.
+ __ vshlqu(kWordPair, out, value, temp);
- // check for overflow by shifting back and comparing.
- __ rsb(shift, shift, ShifterOperand(0));
- __ vmovsr(stemp0, shift);
- __ vshlqi(kWordPair, temp, out, temp);
- __ vceqqi(kWord, out, temp, value);
- // Low 64 bits of temp should be all 1's, otherwise temp != value and
- // we deopt.
- __ vmovrs(shift, sout0);
- __ CompareImmediate(shift, -1);
- __ b(&fail, NE);
- __ vmovrs(shift, sout1);
- __ CompareImmediate(shift, -1);
- __ b(&fail, NE);
+ // check for overflow by shifting back and comparing.
+ __ rsb(shift, shift, ShifterOperand(0));
+ __ vmovsr(stemp0, shift);
+ __ vshlqi(kWordPair, temp, out, temp);
+ __ vceqqi(kWord, out, temp, value);
+ // Low 64 bits of temp should be all 1's, otherwise temp != value and
+ // we deopt.
+ __ vmovrs(shift, sout0);
+ __ CompareImmediate(shift, -1);
+ __ b(&fail, NE);
+ __ vmovrs(shift, sout1);
+ __ CompareImmediate(shift, -1);
+ __ b(&fail, NE);
- __ LoadImmediate(R0, 0);
- __ bx(LR);
+ __ LoadImmediate(R0, 0);
+ __ bx(LR);
- __ Bind(&fail);
- __ LoadImmediate(R0, 1);
+ __ Bind(&fail);
+ __ LoadImmediate(R0, 1);
+ } else {
+ __ LoadImmediate(R0, 1);
+ }
__ bx(LR);
}
« no previous file with comments | « runtime/vm/assembler_arm.cc ('k') | runtime/vm/constants_arm.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698