OLD | NEW |
1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ | 5 #ifndef V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ |
6 #define V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ | 6 #define V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ |
7 | 7 |
8 namespace v8 { | 8 namespace v8 { |
9 namespace internal { | 9 namespace internal { |
10 namespace compiler { | 10 namespace compiler { |
(...skipping 32 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
43 V(ArmVmulF64) \ | 43 V(ArmVmulF64) \ |
44 V(ArmVmlaF64) \ | 44 V(ArmVmlaF64) \ |
45 V(ArmVmlsF64) \ | 45 V(ArmVmlsF64) \ |
46 V(ArmVdivF64) \ | 46 V(ArmVdivF64) \ |
47 V(ArmVmodF64) \ | 47 V(ArmVmodF64) \ |
48 V(ArmVnegF64) \ | 48 V(ArmVnegF64) \ |
49 V(ArmVcvtF64S32) \ | 49 V(ArmVcvtF64S32) \ |
50 V(ArmVcvtF64U32) \ | 50 V(ArmVcvtF64U32) \ |
51 V(ArmVcvtS32F64) \ | 51 V(ArmVcvtS32F64) \ |
52 V(ArmVcvtU32F64) \ | 52 V(ArmVcvtU32F64) \ |
53 V(ArmFloat64Load) \ | 53 V(ArmVldr64) \ |
54 V(ArmFloat64Store) \ | 54 V(ArmVstr64) \ |
55 V(ArmLoadWord8) \ | 55 V(ArmLdrb) \ |
56 V(ArmStoreWord8) \ | 56 V(ArmLdrsb) \ |
57 V(ArmLoadWord16) \ | 57 V(ArmStrb) \ |
58 V(ArmStoreWord16) \ | 58 V(ArmLdrh) \ |
59 V(ArmLoadWord32) \ | 59 V(ArmLdrsh) \ |
60 V(ArmStoreWord32) \ | 60 V(ArmStrh) \ |
| 61 V(ArmLdr) \ |
| 62 V(ArmStr) \ |
61 V(ArmStoreWriteBarrier) | 63 V(ArmStoreWriteBarrier) |
62 | 64 |
63 | 65 |
64 // Addressing modes represent the "shape" of inputs to an instruction. | 66 // Addressing modes represent the "shape" of inputs to an instruction. |
65 // Many instructions support multiple addressing modes. Addressing modes | 67 // Many instructions support multiple addressing modes. Addressing modes |
66 // are encoded into the InstructionCode of the instruction and tell the | 68 // are encoded into the InstructionCode of the instruction and tell the |
67 // code generator after register allocation which assembler method to call. | 69 // code generator after register allocation which assembler method to call. |
68 #define TARGET_ADDRESSING_MODE_LIST(V) \ | 70 #define TARGET_ADDRESSING_MODE_LIST(V) \ |
69 V(Offset_RI) /* [%r0 + K] */ \ | 71 V(Offset_RI) /* [%r0 + K] */ \ |
70 V(Offset_RR) /* [%r0 + %r1] */ \ | 72 V(Offset_RR) /* [%r0 + %r1] */ \ |
71 V(Operand2_I) /* K */ \ | 73 V(Operand2_I) /* K */ \ |
72 V(Operand2_R) /* %r0 */ \ | 74 V(Operand2_R) /* %r0 */ \ |
73 V(Operand2_R_ASR_I) /* %r0 ASR K */ \ | 75 V(Operand2_R_ASR_I) /* %r0 ASR K */ \ |
74 V(Operand2_R_LSL_I) /* %r0 LSL K */ \ | 76 V(Operand2_R_LSL_I) /* %r0 LSL K */ \ |
75 V(Operand2_R_LSR_I) /* %r0 LSR K */ \ | 77 V(Operand2_R_LSR_I) /* %r0 LSR K */ \ |
76 V(Operand2_R_ROR_I) /* %r0 ROR K */ \ | 78 V(Operand2_R_ROR_I) /* %r0 ROR K */ \ |
77 V(Operand2_R_ASR_R) /* %r0 ASR %r1 */ \ | 79 V(Operand2_R_ASR_R) /* %r0 ASR %r1 */ \ |
78 V(Operand2_R_LSL_R) /* %r0 LSL %r1 */ \ | 80 V(Operand2_R_LSL_R) /* %r0 LSL %r1 */ \ |
79 V(Operand2_R_LSR_R) /* %r0 LSR %r1 */ \ | 81 V(Operand2_R_LSR_R) /* %r0 LSR %r1 */ \ |
80 V(Operand2_R_ROR_R) /* %r0 ROR %r1 */ | 82 V(Operand2_R_ROR_R) /* %r0 ROR %r1 */ |
81 | 83 |
82 } // namespace compiler | 84 } // namespace compiler |
83 } // namespace internal | 85 } // namespace internal |
84 } // namespace v8 | 86 } // namespace v8 |
85 | 87 |
86 #endif // V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ | 88 #endif // V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ |
OLD | NEW |