OLD | NEW |
1 //===- subzero/src/IceTargetLoweringARM32.def - ARM32 X-macros --*- C++ -*-===// | 1 //===- subzero/src/IceTargetLoweringARM32.def - ARM32 X-macros --*- C++ -*-===// |
2 // | 2 // |
3 // The Subzero Code Generator | 3 // The Subzero Code Generator |
4 // | 4 // |
5 // This file is distributed under the University of Illinois Open Source | 5 // This file is distributed under the University of Illinois Open Source |
6 // License. See LICENSE.TXT for details. | 6 // License. See LICENSE.TXT for details. |
7 // | 7 // |
8 //===----------------------------------------------------------------------===// | 8 //===----------------------------------------------------------------------===// |
9 // | 9 // |
10 // This file defines certain patterns for lowering to ARM32 target | 10 // This file defines certain patterns for lowering to ARM32 target |
11 // instructions, in the form of x-macros. | 11 // instructions, in the form of x-macros. |
12 // | 12 // |
13 //===----------------------------------------------------------------------===// | 13 //===----------------------------------------------------------------------===// |
14 | 14 |
15 #ifndef SUBZERO_SRC_ICETARGETLOWERINGARM32_DEF | 15 #ifndef SUBZERO_SRC_ICETARGETLOWERINGARM32_DEF |
16 #define SUBZERO_SRC_ICETARGETLOWERINGARM32_DEF | 16 #define SUBZERO_SRC_ICETARGETLOWERINGARM32_DEF |
17 | 17 |
18 // Patterns for lowering fcmp. These are expected to be used in the following | 18 // Patterns for lowering fcmp. These are expected to be used in the following |
19 // manner: | 19 // manner: |
20 // | 20 // |
| 21 // Scalar: |
21 // mov reg, #0 | 22 // mov reg, #0 |
22 // movCC0 reg, #1 /* only if CC0 != kNone */ | 23 // movCC0 reg, #1 /* only if CC0 != kNone */ |
23 // movCC1 reg, #1 /* only if CC1 != kNone */ | 24 // movCC1 reg, #1 /* only if CC1 != kNone */ |
24 // | 25 // |
25 // TODO(jpp): vector lowerings. | 26 // Vector: |
| 27 // vcCC0_V Cmp0, Src0, Src1 /* only if CC0_V != none */ |
| 28 // vcCC1_V Cmp1, Src1, Src0 /* only if CC1_V != none */ |
| 29 // vorr Cmp2, Cmp0, Cmp1 /* only if CC1_V != none */ |
| 30 // vmvn Reg3, Cmp? /* only if NEG_V = true */ |
| 31 // |
| 32 // If INV_V = true, then Src0 and Src1 should be swapped. |
| 33 // |
26 #define FCMPARM32_TABLE \ | 34 #define FCMPARM32_TABLE \ |
27 /* val, CC0, CC1 */ \ | 35 /*val , CC0 , CC1 , CC0_V, CC1_V, INV_V, NEG_V */ \ |
28 X(False, kNone, kNone) \ | 36 X(False, kNone, kNone, none , none , false, false) \ |
29 X(Oeq, EQ, kNone) \ | 37 X(Oeq , EQ , kNone, eq , none , false, false) \ |
30 X(Ogt, GT, kNone) \ | 38 X(Ogt , GT , kNone, gt , none , false, false) \ |
31 X(Oge, GE, kNone) \ | 39 X(Oge , GE , kNone, ge , none , false, false) \ |
32 X(Olt, MI, kNone) \ | 40 X(Olt , MI , kNone, gt , none , true , false) \ |
33 X(Ole, LS, kNone) \ | 41 X(Ole , LS , kNone, ge , none , true , false) \ |
34 X(One, MI, GT) \ | 42 X(One , MI , GT , gt , gt , false, false) \ |
35 X(Ord, VC, kNone) \ | 43 X(Ord , VC , kNone, ge , gt , false, false) \ |
36 X(Ueq, EQ, VS) \ | 44 X(Ueq , EQ , VS , gt , gt , false, true) \ |
37 X(Ugt, HI, kNone) \ | 45 X(Ugt , HI , kNone, ge , none , true , true) \ |
38 X(Uge, PL, kNone) \ | 46 X(Uge , PL , kNone, gt , none , true , true) \ |
39 X(Ult, LT, kNone) \ | 47 X(Ult , LT , kNone, ge , none , false, true) \ |
40 X(Ule, LE, kNone) \ | 48 X(Ule , LE , kNone, gt , none , false, true) \ |
41 X(Une, NE, kNone) \ | 49 X(Une , NE , kNone, eq , none , false, true) \ |
42 X(Uno, VS, kNone) \ | 50 X(Uno , VS , kNone, ge , gt , false, true) \ |
43 X(True, AL, kNone) \ | 51 X(True , AL , kNone, none , none , false, false) |
44 //#define X(val, CC0, CC1) | 52 //#define X(val, CC0, CC1, CC0_V, CC1_V, INV_V, NEG_V) |
45 | 53 |
46 // Patterns for lowering icmp. | 54 // Patterns for lowering icmp. |
47 #define ICMPARM32_TABLE \ | 55 #define ICMPARM32_TABLE \ |
48 /* val, is_signed, swapped64, C_32, C1_64, C2_64 */ \ | 56 /*val, is_signed, swapped64, C_32, C1_64, C2_64, C_V, INV_V, NEG_V */ \ |
49 X(Eq, false, false, EQ, EQ, NE) \ | 57 X(Eq , false , false , EQ, EQ , NE , eq , false, false) \ |
50 X(Ne, false, false, NE, NE, EQ) \ | 58 X(Ne , false , false , NE, NE , EQ , eq , false, true) \ |
51 X(Ugt, false, false, HI, HI, LS) \ | 59 X(Ugt, false , false , HI, HI , LS , gt , false, false) \ |
52 X(Uge, false, false, CS, CS, CC) \ | 60 X(Uge, false , false , CS, CS , CC , ge , false, false) \ |
53 X(Ult, false, false, CC, CC, CS) \ | 61 X(Ult, false , false , CC, CC , CS , gt , true , false) \ |
54 X(Ule, false, false, LS, LS, HI) \ | 62 X(Ule, false , false , LS, LS , HI , ge , true , false) \ |
55 X(Sgt, true, true, GT, LT, GE) \ | 63 X(Sgt, true , true , GT, LT , GE , gt , false, false) \ |
56 X(Sge, true, false, GE, GE, LT) \ | 64 X(Sge, true , false , GE, GE , LT , ge , false, false) \ |
57 X(Slt, true, false, LT, LT, GE) \ | 65 X(Slt, true , false , LT, LT , GE , gt , true , false) \ |
58 X(Sle, true, true, LE, GE, LT) \ | 66 X(Sle, true , true , LE, GE , LT , ge , true , false) |
59 //#define X(val, is_signed, swapped64, C_32, C1_64, C2_64) | 67 //#define X(val, is_signed, swapped64, C_32, C1_64, C2_64, C_V, INV_V, NEG_V) |
60 | 68 |
61 #endif // SUBZERO_SRC_ICETARGETLOWERINGARM32_DEF | 69 #endif // SUBZERO_SRC_ICETARGETLOWERINGARM32_DEF |
OLD | NEW |