OLD | NEW |
1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "src/compiler/instruction-selector-impl.h" | 5 #include "src/compiler/instruction-selector-impl.h" |
6 #include "src/compiler/node-matchers.h" | 6 #include "src/compiler/node-matchers.h" |
7 | 7 |
8 namespace v8 { | 8 namespace v8 { |
9 namespace internal { | 9 namespace internal { |
10 namespace compiler { | 10 namespace compiler { |
(...skipping 92 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
103 Node* node, ImmediateMode operand_mode) { | 103 Node* node, ImmediateMode operand_mode) { |
104 Arm64OperandGenerator g(selector); | 104 Arm64OperandGenerator g(selector); |
105 selector->Emit(opcode, g.DefineAsRegister(node), | 105 selector->Emit(opcode, g.DefineAsRegister(node), |
106 g.UseRegister(node->InputAt(0)), | 106 g.UseRegister(node->InputAt(0)), |
107 g.UseOperand(node->InputAt(1), operand_mode)); | 107 g.UseOperand(node->InputAt(1), operand_mode)); |
108 } | 108 } |
109 | 109 |
110 | 110 |
111 // Shared routine for multiple binary operations. | 111 // Shared routine for multiple binary operations. |
112 static void VisitBinop(InstructionSelector* selector, Node* node, | 112 static void VisitBinop(InstructionSelector* selector, Node* node, |
113 ArchOpcode opcode, ImmediateMode operand_mode, | 113 InstructionCode opcode, ImmediateMode operand_mode, |
114 bool commutative) { | 114 FlagsContinuation* cont) { |
115 VisitRRO(selector, opcode, node, operand_mode); | |
116 } | |
117 | |
118 | |
119 static void VisitBinopWithOverflow(InstructionSelector* selector, Node* node, | |
120 InstructionCode opcode) { | |
121 Arm64OperandGenerator g(selector); | 115 Arm64OperandGenerator g(selector); |
122 Int32BinopMatcher m(node); | 116 Int32BinopMatcher m(node); |
123 InstructionOperand* inputs[2]; | 117 InstructionOperand* inputs[4]; |
124 size_t input_count = 0; | 118 size_t input_count = 0; |
125 InstructionOperand* outputs[2]; | 119 InstructionOperand* outputs[2]; |
126 size_t output_count = 0; | 120 size_t output_count = 0; |
127 | 121 |
128 inputs[input_count++] = g.UseRegister(m.left().node()); | 122 inputs[input_count++] = g.UseRegister(m.left().node()); |
129 inputs[input_count++] = g.UseRegister(m.right().node()); | 123 inputs[input_count++] = g.UseOperand(m.right().node(), operand_mode); |
130 | 124 |
131 // Define outputs depending on the projections. | 125 if (cont->IsBranch()) { |
132 Node* projections[2]; | 126 inputs[input_count++] = g.Label(cont->true_block()); |
133 node->CollectProjections(ARRAY_SIZE(projections), projections); | 127 inputs[input_count++] = g.Label(cont->false_block()); |
134 if (projections[0]) { | |
135 outputs[output_count++] = g.DefineAsRegister(projections[0]); | |
136 } | 128 } |
137 if (projections[1]) { | 129 |
138 opcode |= FlagsModeField::encode(kFlags_set); | 130 outputs[output_count++] = g.DefineAsRegister(node); |
139 opcode |= FlagsConditionField::encode(kOverflow); | 131 if (cont->IsSet()) { |
140 outputs[output_count++] = g.DefineAsRegister(projections[1]); | 132 outputs[output_count++] = g.DefineAsRegister(cont->result()); |
141 } | 133 } |
142 | 134 |
143 ASSERT_NE(0, input_count); | 135 ASSERT_NE(0, input_count); |
144 ASSERT_NE(0, output_count); | 136 ASSERT_NE(0, output_count); |
145 ASSERT_GE(ARRAY_SIZE(inputs), input_count); | 137 ASSERT_GE(ARRAY_SIZE(inputs), input_count); |
146 ASSERT_GE(ARRAY_SIZE(outputs), output_count); | 138 ASSERT_GE(ARRAY_SIZE(outputs), output_count); |
147 | 139 |
148 selector->Emit(opcode, output_count, outputs, input_count, inputs); | 140 Instruction* instr = selector->Emit(cont->Encode(opcode), output_count, |
| 141 outputs, input_count, inputs); |
| 142 if (cont->IsBranch()) instr->MarkAsControl(); |
149 } | 143 } |
150 | 144 |
151 | 145 |
| 146 // Shared routine for multiple binary operations. |
| 147 static void VisitBinop(InstructionSelector* selector, Node* node, |
| 148 ArchOpcode opcode, ImmediateMode operand_mode) { |
| 149 FlagsContinuation cont; |
| 150 VisitBinop(selector, node, opcode, operand_mode); |
| 151 } |
| 152 |
| 153 |
152 void InstructionSelector::VisitLoad(Node* node) { | 154 void InstructionSelector::VisitLoad(Node* node) { |
153 MachineRepresentation rep = OpParameter<MachineRepresentation>(node); | 155 MachineRepresentation rep = OpParameter<MachineRepresentation>(node); |
154 Arm64OperandGenerator g(this); | 156 Arm64OperandGenerator g(this); |
155 Node* base = node->InputAt(0); | 157 Node* base = node->InputAt(0); |
156 Node* index = node->InputAt(1); | 158 Node* index = node->InputAt(1); |
157 | 159 |
158 InstructionOperand* result = rep == kMachineFloat64 | 160 InstructionOperand* result = rep == kMachineFloat64 |
159 ? g.DefineAsDoubleRegister(node) | 161 ? g.DefineAsDoubleRegister(node) |
160 : g.DefineAsRegister(node); | 162 : g.DefineAsRegister(node); |
161 | 163 |
(...skipping 87 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
249 Emit(opcode | AddressingModeField::encode(kMode_MRI), NULL, | 251 Emit(opcode | AddressingModeField::encode(kMode_MRI), NULL, |
250 g.UseRegister(index), g.UseImmediate(base), val); | 252 g.UseRegister(index), g.UseImmediate(base), val); |
251 } else { | 253 } else { |
252 Emit(opcode | AddressingModeField::encode(kMode_MRR), NULL, | 254 Emit(opcode | AddressingModeField::encode(kMode_MRR), NULL, |
253 g.UseRegister(base), g.UseRegister(index), val); | 255 g.UseRegister(base), g.UseRegister(index), val); |
254 } | 256 } |
255 } | 257 } |
256 | 258 |
257 | 259 |
258 void InstructionSelector::VisitWord32And(Node* node) { | 260 void InstructionSelector::VisitWord32And(Node* node) { |
259 VisitBinop(this, node, kArm64And32, kLogical32Imm, true); | 261 VisitBinop(this, node, kArm64And32, kLogical32Imm); |
260 } | 262 } |
261 | 263 |
262 | 264 |
263 void InstructionSelector::VisitWord64And(Node* node) { | 265 void InstructionSelector::VisitWord64And(Node* node) { |
264 VisitBinop(this, node, kArm64And, kLogical64Imm, true); | 266 VisitBinop(this, node, kArm64And, kLogical64Imm); |
265 } | 267 } |
266 | 268 |
267 | 269 |
268 void InstructionSelector::VisitWord32Or(Node* node) { | 270 void InstructionSelector::VisitWord32Or(Node* node) { |
269 VisitBinop(this, node, kArm64Or32, kLogical32Imm, true); | 271 VisitBinop(this, node, kArm64Or32, kLogical32Imm); |
270 } | 272 } |
271 | 273 |
272 | 274 |
273 void InstructionSelector::VisitWord64Or(Node* node) { | 275 void InstructionSelector::VisitWord64Or(Node* node) { |
274 VisitBinop(this, node, kArm64Or, kLogical64Imm, true); | 276 VisitBinop(this, node, kArm64Or, kLogical64Imm); |
275 } | 277 } |
276 | 278 |
277 | 279 |
278 template <typename T> | 280 template <typename T> |
279 static void VisitXor(InstructionSelector* selector, Node* node, | 281 static void VisitXor(InstructionSelector* selector, Node* node, |
280 ArchOpcode xor_opcode, ArchOpcode not_opcode) { | 282 ArchOpcode xor_opcode, ArchOpcode not_opcode) { |
281 Arm64OperandGenerator g(selector); | 283 Arm64OperandGenerator g(selector); |
282 BinopMatcher<IntMatcher<T>, IntMatcher<T> > m(node); | 284 BinopMatcher<IntMatcher<T>, IntMatcher<T> > m(node); |
283 if (m.right().Is(-1)) { | 285 if (m.right().Is(-1)) { |
284 selector->Emit(not_opcode, g.DefineAsRegister(node), | 286 selector->Emit(not_opcode, g.DefineAsRegister(node), |
285 g.UseRegister(m.left().node())); | 287 g.UseRegister(m.left().node())); |
286 } else { | 288 } else { |
287 VisitBinop(selector, node, xor_opcode, kLogical32Imm, true); | 289 VisitBinop(selector, node, xor_opcode, kLogical32Imm); |
288 } | 290 } |
289 } | 291 } |
290 | 292 |
291 | 293 |
292 void InstructionSelector::VisitWord32Xor(Node* node) { | 294 void InstructionSelector::VisitWord32Xor(Node* node) { |
293 VisitXor<int32_t>(this, node, kArm64Xor32, kArm64Not32); | 295 VisitXor<int32_t>(this, node, kArm64Xor32, kArm64Not32); |
294 } | 296 } |
295 | 297 |
296 | 298 |
297 void InstructionSelector::VisitWord64Xor(Node* node) { | 299 void InstructionSelector::VisitWord64Xor(Node* node) { |
(...skipping 25 matching lines...) Expand all Loading... |
323 VisitRRO(this, kArm64Sar32, node, kShift32Imm); | 325 VisitRRO(this, kArm64Sar32, node, kShift32Imm); |
324 } | 326 } |
325 | 327 |
326 | 328 |
327 void InstructionSelector::VisitWord64Sar(Node* node) { | 329 void InstructionSelector::VisitWord64Sar(Node* node) { |
328 VisitRRO(this, kArm64Sar, node, kShift64Imm); | 330 VisitRRO(this, kArm64Sar, node, kShift64Imm); |
329 } | 331 } |
330 | 332 |
331 | 333 |
332 void InstructionSelector::VisitInt32Add(Node* node) { | 334 void InstructionSelector::VisitInt32Add(Node* node) { |
333 VisitBinop(this, node, kArm64Add32, kArithimeticImm, true); | 335 VisitBinop(this, node, kArm64Add32, kArithimeticImm); |
334 } | |
335 | |
336 | |
337 void InstructionSelector::VisitInt32AddWithOverflow(Node* node) { | |
338 VisitBinopWithOverflow(this, node, kArm64Add32); | |
339 } | 336 } |
340 | 337 |
341 | 338 |
342 void InstructionSelector::VisitInt64Add(Node* node) { | 339 void InstructionSelector::VisitInt64Add(Node* node) { |
343 VisitBinop(this, node, kArm64Add, kArithimeticImm, true); | 340 VisitBinop(this, node, kArm64Add, kArithimeticImm); |
344 } | 341 } |
345 | 342 |
346 | 343 |
347 template <typename T> | 344 template <typename T> |
348 static void VisitSub(InstructionSelector* selector, Node* node, | 345 static void VisitSub(InstructionSelector* selector, Node* node, |
349 ArchOpcode sub_opcode, ArchOpcode neg_opcode) { | 346 ArchOpcode sub_opcode, ArchOpcode neg_opcode) { |
350 Arm64OperandGenerator g(selector); | 347 Arm64OperandGenerator g(selector); |
351 BinopMatcher<IntMatcher<T>, IntMatcher<T> > m(node); | 348 BinopMatcher<IntMatcher<T>, IntMatcher<T> > m(node); |
352 if (m.left().Is(0)) { | 349 if (m.left().Is(0)) { |
353 selector->Emit(neg_opcode, g.DefineAsRegister(node), | 350 selector->Emit(neg_opcode, g.DefineAsRegister(node), |
354 g.UseRegister(m.right().node())); | 351 g.UseRegister(m.right().node())); |
355 } else { | 352 } else { |
356 VisitBinop(selector, node, sub_opcode, kArithimeticImm, false); | 353 VisitBinop(selector, node, sub_opcode, kArithimeticImm); |
357 } | 354 } |
358 } | 355 } |
359 | 356 |
360 | 357 |
361 void InstructionSelector::VisitInt32Sub(Node* node) { | 358 void InstructionSelector::VisitInt32Sub(Node* node) { |
362 VisitSub<int32_t>(this, node, kArm64Sub32, kArm64Neg32); | 359 VisitSub<int32_t>(this, node, kArm64Sub32, kArm64Neg32); |
363 } | 360 } |
364 | 361 |
365 | 362 |
366 void InstructionSelector::VisitInt32SubWithOverflow(Node* node) { | |
367 VisitBinopWithOverflow(this, node, kArm64Sub32); | |
368 } | |
369 | |
370 | |
371 void InstructionSelector::VisitInt64Sub(Node* node) { | 363 void InstructionSelector::VisitInt64Sub(Node* node) { |
372 VisitSub<int64_t>(this, node, kArm64Sub, kArm64Neg); | 364 VisitSub<int64_t>(this, node, kArm64Sub, kArm64Neg); |
373 } | 365 } |
374 | 366 |
375 | 367 |
376 void InstructionSelector::VisitInt32Mul(Node* node) { | 368 void InstructionSelector::VisitInt32Mul(Node* node) { |
377 VisitRRR(this, kArm64Mul32, node); | 369 VisitRRR(this, kArm64Mul32, node); |
378 } | 370 } |
379 | 371 |
380 | 372 |
(...skipping 101 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
482 | 474 |
483 | 475 |
484 void InstructionSelector::VisitFloat64Mod(Node* node) { | 476 void InstructionSelector::VisitFloat64Mod(Node* node) { |
485 Arm64OperandGenerator g(this); | 477 Arm64OperandGenerator g(this); |
486 Emit(kArm64Float64Mod, g.DefineAsFixedDouble(node, d0), | 478 Emit(kArm64Float64Mod, g.DefineAsFixedDouble(node, d0), |
487 g.UseFixedDouble(node->InputAt(0), d0), | 479 g.UseFixedDouble(node->InputAt(0), d0), |
488 g.UseFixedDouble(node->InputAt(1), d1))->MarkAsCall(); | 480 g.UseFixedDouble(node->InputAt(1), d1))->MarkAsCall(); |
489 } | 481 } |
490 | 482 |
491 | 483 |
| 484 void InstructionSelector::VisitInt32AddWithOverflow(Node* node, |
| 485 FlagsContinuation* cont) { |
| 486 VisitBinop(this, node, kArm64Add32, kArithimeticImm, cont); |
| 487 } |
| 488 |
| 489 |
| 490 void InstructionSelector::VisitInt32SubWithOverflow(Node* node, |
| 491 FlagsContinuation* cont) { |
| 492 VisitBinop(this, node, kArm64Sub32, kArithimeticImm, cont); |
| 493 } |
| 494 |
| 495 |
492 // Shared routine for multiple compare operations. | 496 // Shared routine for multiple compare operations. |
493 static void VisitCompare(InstructionSelector* selector, InstructionCode opcode, | 497 static void VisitCompare(InstructionSelector* selector, InstructionCode opcode, |
494 InstructionOperand* left, InstructionOperand* right, | 498 InstructionOperand* left, InstructionOperand* right, |
495 FlagsContinuation* cont) { | 499 FlagsContinuation* cont) { |
496 Arm64OperandGenerator g(selector); | 500 Arm64OperandGenerator g(selector); |
497 opcode = cont->Encode(opcode); | 501 opcode = cont->Encode(opcode); |
498 if (cont->IsBranch()) { | 502 if (cont->IsBranch()) { |
499 selector->Emit(opcode, NULL, left, right, g.Label(cont->true_block()), | 503 selector->Emit(opcode, NULL, left, right, g.Label(cont->true_block()), |
500 g.Label(cont->false_block()))->MarkAsControl(); | 504 g.Label(cont->false_block()))->MarkAsControl(); |
501 } else { | 505 } else { |
(...skipping 156 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
658 ASSERT(deoptimization == NULL && continuation == NULL); | 662 ASSERT(deoptimization == NULL && continuation == NULL); |
659 Emit(kArm64Drop | MiscField::encode(aligned_push_count), NULL); | 663 Emit(kArm64Drop | MiscField::encode(aligned_push_count), NULL); |
660 } | 664 } |
661 } | 665 } |
662 | 666 |
663 #endif // V8_TURBOFAN_TARGET | 667 #endif // V8_TURBOFAN_TARGET |
664 | 668 |
665 } // namespace compiler | 669 } // namespace compiler |
666 } // namespace internal | 670 } // namespace internal |
667 } // namespace v8 | 671 } // namespace v8 |
OLD | NEW |