| OLD | NEW |
| 1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #ifndef V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ | 5 #ifndef V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ |
| 6 #define V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ | 6 #define V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ |
| 7 | 7 |
| 8 namespace v8 { | 8 namespace v8 { |
| 9 namespace internal { | 9 namespace internal { |
| 10 namespace compiler { | 10 namespace compiler { |
| (...skipping 111 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 122 V(ArmPoke) \ | 122 V(ArmPoke) \ |
| 123 V(ArmFloat32x4Splat) \ | 123 V(ArmFloat32x4Splat) \ |
| 124 V(ArmFloat32x4ExtractLane) \ | 124 V(ArmFloat32x4ExtractLane) \ |
| 125 V(ArmFloat32x4ReplaceLane) \ | 125 V(ArmFloat32x4ReplaceLane) \ |
| 126 V(ArmFloat32x4FromInt32x4) \ | 126 V(ArmFloat32x4FromInt32x4) \ |
| 127 V(ArmFloat32x4FromUint32x4) \ | 127 V(ArmFloat32x4FromUint32x4) \ |
| 128 V(ArmFloat32x4Abs) \ | 128 V(ArmFloat32x4Abs) \ |
| 129 V(ArmFloat32x4Neg) \ | 129 V(ArmFloat32x4Neg) \ |
| 130 V(ArmFloat32x4Add) \ | 130 V(ArmFloat32x4Add) \ |
| 131 V(ArmFloat32x4Sub) \ | 131 V(ArmFloat32x4Sub) \ |
| 132 V(ArmFloat32x4Eq) \ | 132 V(ArmFloat32x4Equal) \ |
| 133 V(ArmFloat32x4Ne) \ | 133 V(ArmFloat32x4NotEqual) \ |
| 134 V(ArmInt32x4Splat) \ | 134 V(ArmInt32x4Splat) \ |
| 135 V(ArmInt32x4ExtractLane) \ | 135 V(ArmInt32x4ExtractLane) \ |
| 136 V(ArmInt32x4ReplaceLane) \ | 136 V(ArmInt32x4ReplaceLane) \ |
| 137 V(ArmInt32x4FromFloat32x4) \ | 137 V(ArmInt32x4FromFloat32x4) \ |
| 138 V(ArmUint32x4FromFloat32x4) \ | 138 V(ArmUint32x4FromFloat32x4) \ |
| 139 V(ArmInt32x4Neg) \ | 139 V(ArmInt32x4Neg) \ |
| 140 V(ArmInt32x4Add) \ | 140 V(ArmInt32x4Add) \ |
| 141 V(ArmInt32x4Sub) \ | 141 V(ArmInt32x4Sub) \ |
| 142 V(ArmInt32x4Mul) \ | 142 V(ArmInt32x4Mul) \ |
| 143 V(ArmInt32x4Min) \ | 143 V(ArmInt32x4Min) \ |
| 144 V(ArmInt32x4Max) \ | 144 V(ArmInt32x4Max) \ |
| 145 V(ArmInt32x4Eq) \ | 145 V(ArmInt32x4Equal) \ |
| 146 V(ArmInt32x4Ne) \ | 146 V(ArmInt32x4NotEqual) \ |
| 147 V(ArmInt32x4Gt) \ | 147 V(ArmInt32x4GreaterThan) \ |
| 148 V(ArmInt32x4Ge) \ | 148 V(ArmInt32x4GreaterThanOrEqual) \ |
| 149 V(ArmUint32x4Gt) \ | 149 V(ArmUint32x4GreaterThan) \ |
| 150 V(ArmUint32x4Ge) \ | 150 V(ArmUint32x4GreaterThanOrEqual) \ |
| 151 V(ArmSimd32x4Select) \ | 151 V(ArmSimd32x4Select) \ |
| 152 V(ArmInt16x8Splat) \ | 152 V(ArmInt16x8Splat) \ |
| 153 V(ArmInt16x8ExtractLane) \ | 153 V(ArmInt16x8ExtractLane) \ |
| 154 V(ArmInt16x8ReplaceLane) \ | 154 V(ArmInt16x8ReplaceLane) \ |
| 155 V(ArmInt16x8Neg) \ | 155 V(ArmInt16x8Neg) \ |
| 156 V(ArmInt16x8Add) \ | 156 V(ArmInt16x8Add) \ |
| 157 V(ArmInt16x8Sub) \ | 157 V(ArmInt16x8Sub) \ |
| 158 V(ArmInt16x8Mul) \ | 158 V(ArmInt16x8Mul) \ |
| 159 V(ArmInt16x8Min) \ | 159 V(ArmInt16x8Min) \ |
| 160 V(ArmInt16x8Max) \ | 160 V(ArmInt16x8Max) \ |
| 161 V(ArmInt16x8Eq) \ | 161 V(ArmInt16x8Equal) \ |
| 162 V(ArmInt16x8Ne) \ | 162 V(ArmInt16x8NotEqual) \ |
| 163 V(ArmInt16x8Gt) \ | 163 V(ArmInt16x8GreaterThan) \ |
| 164 V(ArmInt16x8Ge) \ | 164 V(ArmInt16x8GreaterThanOrEqual) \ |
| 165 V(ArmUint16x8Gt) \ | 165 V(ArmUint16x8GreaterThan) \ |
| 166 V(ArmUint16x8Ge) \ | 166 V(ArmUint16x8GreaterThanOrEqual) \ |
| 167 V(ArmInt8x16Splat) \ | 167 V(ArmInt8x16Splat) \ |
| 168 V(ArmInt8x16ExtractLane) \ | 168 V(ArmInt8x16ExtractLane) \ |
| 169 V(ArmInt8x16ReplaceLane) \ | 169 V(ArmInt8x16ReplaceLane) \ |
| 170 V(ArmInt8x16Neg) \ | 170 V(ArmInt8x16Neg) \ |
| 171 V(ArmInt8x16Add) \ | 171 V(ArmInt8x16Add) \ |
| 172 V(ArmInt8x16Sub) \ | 172 V(ArmInt8x16Sub) \ |
| 173 V(ArmInt8x16Mul) \ | 173 V(ArmInt8x16Mul) \ |
| 174 V(ArmInt8x16Min) \ | 174 V(ArmInt8x16Min) \ |
| 175 V(ArmInt8x16Max) \ | 175 V(ArmInt8x16Max) \ |
| 176 V(ArmInt8x16Eq) \ | 176 V(ArmInt8x16Equal) \ |
| 177 V(ArmInt8x16Ne) \ | 177 V(ArmInt8x16NotEqual) \ |
| 178 V(ArmInt8x16Gt) \ | 178 V(ArmInt8x16GreaterThan) \ |
| 179 V(ArmInt8x16Ge) \ | 179 V(ArmInt8x16GreaterThanOrEqual) \ |
| 180 V(ArmUint8x16Gt) \ | 180 V(ArmUint8x16GreaterThan) \ |
| 181 V(ArmUint8x16Ge) | 181 V(ArmUint8x16GreaterThanOrEqual) |
| 182 | 182 |
| 183 // Addressing modes represent the "shape" of inputs to an instruction. | 183 // Addressing modes represent the "shape" of inputs to an instruction. |
| 184 // Many instructions support multiple addressing modes. Addressing modes | 184 // Many instructions support multiple addressing modes. Addressing modes |
| 185 // are encoded into the InstructionCode of the instruction and tell the | 185 // are encoded into the InstructionCode of the instruction and tell the |
| 186 // code generator after register allocation which assembler method to call. | 186 // code generator after register allocation which assembler method to call. |
| 187 #define TARGET_ADDRESSING_MODE_LIST(V) \ | 187 #define TARGET_ADDRESSING_MODE_LIST(V) \ |
| 188 V(Offset_RI) /* [%r0 + K] */ \ | 188 V(Offset_RI) /* [%r0 + K] */ \ |
| 189 V(Offset_RR) /* [%r0 + %r1] */ \ | 189 V(Offset_RR) /* [%r0 + %r1] */ \ |
| 190 V(Operand2_I) /* K */ \ | 190 V(Operand2_I) /* K */ \ |
| 191 V(Operand2_R) /* %r0 */ \ | 191 V(Operand2_R) /* %r0 */ \ |
| 192 V(Operand2_R_ASR_I) /* %r0 ASR K */ \ | 192 V(Operand2_R_ASR_I) /* %r0 ASR K */ \ |
| 193 V(Operand2_R_LSL_I) /* %r0 LSL K */ \ | 193 V(Operand2_R_LSL_I) /* %r0 LSL K */ \ |
| 194 V(Operand2_R_LSR_I) /* %r0 LSR K */ \ | 194 V(Operand2_R_LSR_I) /* %r0 LSR K */ \ |
| 195 V(Operand2_R_ROR_I) /* %r0 ROR K */ \ | 195 V(Operand2_R_ROR_I) /* %r0 ROR K */ \ |
| 196 V(Operand2_R_ASR_R) /* %r0 ASR %r1 */ \ | 196 V(Operand2_R_ASR_R) /* %r0 ASR %r1 */ \ |
| 197 V(Operand2_R_LSL_R) /* %r0 LSL %r1 */ \ | 197 V(Operand2_R_LSL_R) /* %r0 LSL %r1 */ \ |
| 198 V(Operand2_R_LSR_R) /* %r0 LSR %r1 */ \ | 198 V(Operand2_R_LSR_R) /* %r0 LSR %r1 */ \ |
| 199 V(Operand2_R_ROR_R) /* %r0 ROR %r1 */ | 199 V(Operand2_R_ROR_R) /* %r0 ROR %r1 */ |
| 200 | 200 |
| 201 } // namespace compiler | 201 } // namespace compiler |
| 202 } // namespace internal | 202 } // namespace internal |
| 203 } // namespace v8 | 203 } // namespace v8 |
| 204 | 204 |
| 205 #endif // V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ | 205 #endif // V8_COMPILER_ARM_INSTRUCTION_CODES_ARM_H_ |
| OLD | NEW |