| OLD | NEW |
| 1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "src/base/adapters.h" | 5 #include "src/base/adapters.h" |
| 6 #include "src/base/bits.h" | 6 #include "src/base/bits.h" |
| 7 #include "src/compiler/instruction-selector-impl.h" | 7 #include "src/compiler/instruction-selector-impl.h" |
| 8 #include "src/compiler/node-matchers.h" | 8 #include "src/compiler/node-matchers.h" |
| 9 #include "src/compiler/node-properties.h" | 9 #include "src/compiler/node-properties.h" |
| 10 | 10 |
| (...skipping 74 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 85 } | 85 } |
| 86 | 86 |
| 87 | 87 |
| 88 void VisitRRR(InstructionSelector* selector, ArchOpcode opcode, Node* node) { | 88 void VisitRRR(InstructionSelector* selector, ArchOpcode opcode, Node* node) { |
| 89 ArmOperandGenerator g(selector); | 89 ArmOperandGenerator g(selector); |
| 90 selector->Emit(opcode, g.DefineAsRegister(node), | 90 selector->Emit(opcode, g.DefineAsRegister(node), |
| 91 g.UseRegister(node->InputAt(0)), | 91 g.UseRegister(node->InputAt(0)), |
| 92 g.UseRegister(node->InputAt(1))); | 92 g.UseRegister(node->InputAt(1))); |
| 93 } | 93 } |
| 94 | 94 |
| 95 void VisitRRI(InstructionSelector* selector, ArchOpcode opcode, Node* node) { |
| 96 ArmOperandGenerator g(selector); |
| 97 int32_t imm = OpParameter<int32_t>(node); |
| 98 selector->Emit(opcode, g.DefineAsRegister(node), |
| 99 g.UseRegister(node->InputAt(0)), g.UseImmediate(imm)); |
| 100 } |
| 101 |
| 102 void VisitRRIR(InstructionSelector* selector, ArchOpcode opcode, Node* node) { |
| 103 ArmOperandGenerator g(selector); |
| 104 int32_t imm = OpParameter<int32_t>(node); |
| 105 selector->Emit(opcode, g.DefineAsRegister(node), |
| 106 g.UseRegister(node->InputAt(0)), g.UseImmediate(imm), |
| 107 g.UseRegister(node->InputAt(1))); |
| 108 } |
| 95 | 109 |
| 96 template <IrOpcode::Value kOpcode, int kImmMin, int kImmMax, | 110 template <IrOpcode::Value kOpcode, int kImmMin, int kImmMax, |
| 97 AddressingMode kImmMode, AddressingMode kRegMode> | 111 AddressingMode kImmMode, AddressingMode kRegMode> |
| 98 bool TryMatchShift(InstructionSelector* selector, | 112 bool TryMatchShift(InstructionSelector* selector, |
| 99 InstructionCode* opcode_return, Node* node, | 113 InstructionCode* opcode_return, Node* node, |
| 100 InstructionOperand* value_return, | 114 InstructionOperand* value_return, |
| 101 InstructionOperand* shift_return) { | 115 InstructionOperand* shift_return) { |
| 102 ArmOperandGenerator g(selector); | 116 ArmOperandGenerator g(selector); |
| 103 if (node->opcode() == kOpcode) { | 117 if (node->opcode() == kOpcode) { |
| 104 Int32BinopMatcher m(node); | 118 Int32BinopMatcher m(node); |
| (...skipping 978 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1083 } | 1097 } |
| 1084 | 1098 |
| 1085 void InstructionSelector::VisitWord32PairSar(Node* node) { | 1099 void InstructionSelector::VisitWord32PairSar(Node* node) { |
| 1086 VisitWord32PairShift(this, kArmAsrPair, node); | 1100 VisitWord32PairShift(this, kArmAsrPair, node); |
| 1087 } | 1101 } |
| 1088 | 1102 |
| 1089 void InstructionSelector::VisitWord32Ror(Node* node) { | 1103 void InstructionSelector::VisitWord32Ror(Node* node) { |
| 1090 VisitShift(this, node, TryMatchROR); | 1104 VisitShift(this, node, TryMatchROR); |
| 1091 } | 1105 } |
| 1092 | 1106 |
| 1093 | |
| 1094 void InstructionSelector::VisitWord32Clz(Node* node) { | |
| 1095 VisitRR(this, kArmClz, node); | |
| 1096 } | |
| 1097 | |
| 1098 | |
| 1099 void InstructionSelector::VisitWord32Ctz(Node* node) { UNREACHABLE(); } | 1107 void InstructionSelector::VisitWord32Ctz(Node* node) { UNREACHABLE(); } |
| 1100 | 1108 |
| 1101 | |
| 1102 void InstructionSelector::VisitWord32ReverseBits(Node* node) { | 1109 void InstructionSelector::VisitWord32ReverseBits(Node* node) { |
| 1103 DCHECK(IsSupported(ARMv7)); | 1110 DCHECK(IsSupported(ARMv7)); |
| 1104 VisitRR(this, kArmRbit, node); | 1111 VisitRR(this, kArmRbit, node); |
| 1105 } | 1112 } |
| 1106 | 1113 |
| 1107 void InstructionSelector::VisitWord64ReverseBytes(Node* node) { UNREACHABLE(); } | 1114 void InstructionSelector::VisitWord64ReverseBytes(Node* node) { UNREACHABLE(); } |
| 1108 | 1115 |
| 1109 void InstructionSelector::VisitWord32ReverseBytes(Node* node) { UNREACHABLE(); } | 1116 void InstructionSelector::VisitWord32ReverseBytes(Node* node) { UNREACHABLE(); } |
| 1110 | 1117 |
| 1111 void InstructionSelector::VisitWord32Popcnt(Node* node) { UNREACHABLE(); } | 1118 void InstructionSelector::VisitWord32Popcnt(Node* node) { UNREACHABLE(); } |
| (...skipping 180 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1292 Emit(kArmRsb | AddressingModeField::encode(kMode_Operand2_R_LSL_I), | 1299 Emit(kArmRsb | AddressingModeField::encode(kMode_Operand2_R_LSL_I), |
| 1293 g.DefineAsRegister(node), g.UseRegister(m.left().node()), | 1300 g.DefineAsRegister(node), g.UseRegister(m.left().node()), |
| 1294 g.UseRegister(m.left().node()), | 1301 g.UseRegister(m.left().node()), |
| 1295 g.TempImmediate(WhichPowerOf2(value + 1))); | 1302 g.TempImmediate(WhichPowerOf2(value + 1))); |
| 1296 return; | 1303 return; |
| 1297 } | 1304 } |
| 1298 } | 1305 } |
| 1299 VisitRRR(this, kArmMul, node); | 1306 VisitRRR(this, kArmMul, node); |
| 1300 } | 1307 } |
| 1301 | 1308 |
| 1302 | |
| 1303 void InstructionSelector::VisitInt32MulHigh(Node* node) { | |
| 1304 VisitRRR(this, kArmSmmul, node); | |
| 1305 } | |
| 1306 | |
| 1307 | |
| 1308 void InstructionSelector::VisitUint32MulHigh(Node* node) { | 1309 void InstructionSelector::VisitUint32MulHigh(Node* node) { |
| 1309 ArmOperandGenerator g(this); | 1310 ArmOperandGenerator g(this); |
| 1310 InstructionOperand outputs[] = {g.TempRegister(), g.DefineAsRegister(node)}; | 1311 InstructionOperand outputs[] = {g.TempRegister(), g.DefineAsRegister(node)}; |
| 1311 InstructionOperand inputs[] = {g.UseRegister(node->InputAt(0)), | 1312 InstructionOperand inputs[] = {g.UseRegister(node->InputAt(0)), |
| 1312 g.UseRegister(node->InputAt(1))}; | 1313 g.UseRegister(node->InputAt(1))}; |
| 1313 Emit(kArmUmull, arraysize(outputs), outputs, arraysize(inputs), inputs); | 1314 Emit(kArmUmull, arraysize(outputs), outputs, arraysize(inputs), inputs); |
| 1314 } | 1315 } |
| 1315 | 1316 |
| 1316 | 1317 |
| 1317 void InstructionSelector::VisitInt32Div(Node* node) { | 1318 void InstructionSelector::VisitInt32Div(Node* node) { |
| 1318 VisitDiv(this, node, kArmSdiv, kArmVcvtF64S32, kArmVcvtS32F64); | 1319 VisitDiv(this, node, kArmSdiv, kArmVcvtF64S32, kArmVcvtS32F64); |
| 1319 } | 1320 } |
| 1320 | 1321 |
| 1321 | 1322 |
| 1322 void InstructionSelector::VisitUint32Div(Node* node) { | 1323 void InstructionSelector::VisitUint32Div(Node* node) { |
| 1323 VisitDiv(this, node, kArmUdiv, kArmVcvtF64U32, kArmVcvtU32F64); | 1324 VisitDiv(this, node, kArmUdiv, kArmVcvtF64U32, kArmVcvtU32F64); |
| 1324 } | 1325 } |
| 1325 | 1326 |
| 1326 | 1327 |
| 1327 void InstructionSelector::VisitInt32Mod(Node* node) { | 1328 void InstructionSelector::VisitInt32Mod(Node* node) { |
| 1328 VisitMod(this, node, kArmSdiv, kArmVcvtF64S32, kArmVcvtS32F64); | 1329 VisitMod(this, node, kArmSdiv, kArmVcvtF64S32, kArmVcvtS32F64); |
| 1329 } | 1330 } |
| 1330 | 1331 |
| 1331 | 1332 |
| 1332 void InstructionSelector::VisitUint32Mod(Node* node) { | 1333 void InstructionSelector::VisitUint32Mod(Node* node) { |
| 1333 VisitMod(this, node, kArmUdiv, kArmVcvtF64U32, kArmVcvtU32F64); | 1334 VisitMod(this, node, kArmUdiv, kArmVcvtF64U32, kArmVcvtU32F64); |
| 1334 } | 1335 } |
| 1335 | 1336 |
| 1337 #define RR_OP_LIST(V) \ |
| 1338 V(Word32Clz, kArmClz) \ |
| 1339 V(ChangeFloat32ToFloat64, kArmVcvtF64F32) \ |
| 1340 V(RoundInt32ToFloat32, kArmVcvtF32S32) \ |
| 1341 V(RoundUint32ToFloat32, kArmVcvtF32U32) \ |
| 1342 V(ChangeInt32ToFloat64, kArmVcvtF64S32) \ |
| 1343 V(ChangeUint32ToFloat64, kArmVcvtF64U32) \ |
| 1344 V(TruncateFloat32ToInt32, kArmVcvtS32F32) \ |
| 1345 V(TruncateFloat32ToUint32, kArmVcvtU32F32) \ |
| 1346 V(ChangeFloat64ToInt32, kArmVcvtS32F64) \ |
| 1347 V(ChangeFloat64ToUint32, kArmVcvtU32F64) \ |
| 1348 V(TruncateFloat64ToUint32, kArmVcvtU32F64) \ |
| 1349 V(TruncateFloat64ToFloat32, kArmVcvtF32F64) \ |
| 1350 V(TruncateFloat64ToWord32, kArchTruncateDoubleToI) \ |
| 1351 V(RoundFloat64ToInt32, kArmVcvtS32F64) \ |
| 1352 V(BitcastFloat32ToInt32, kArmVmovU32F32) \ |
| 1353 V(BitcastInt32ToFloat32, kArmVmovF32U32) \ |
| 1354 V(Float64ExtractLowWord32, kArmVmovLowU32F64) \ |
| 1355 V(Float64ExtractHighWord32, kArmVmovHighU32F64) \ |
| 1356 V(Float64SilenceNaN, kArmFloat64SilenceNaN) \ |
| 1357 V(Float32Abs, kArmVabsF32) \ |
| 1358 V(Float64Abs, kArmVabsF64) \ |
| 1359 V(Float32Neg, kArmVnegF32) \ |
| 1360 V(Float64Neg, kArmVnegF64) \ |
| 1361 V(Float32Sqrt, kArmVsqrtF32) \ |
| 1362 V(Float64Sqrt, kArmVsqrtF64) |
| 1336 | 1363 |
| 1337 void InstructionSelector::VisitChangeFloat32ToFloat64(Node* node) { | 1364 #define RR_OP_LIST_V8(V) \ |
| 1338 VisitRR(this, kArmVcvtF64F32, node); | 1365 V(Float32RoundDown, kArmVrintmF32) \ |
| 1339 } | 1366 V(Float64RoundDown, kArmVrintmF64) \ |
| 1367 V(Float32RoundUp, kArmVrintpF32) \ |
| 1368 V(Float64RoundUp, kArmVrintpF64) \ |
| 1369 V(Float32RoundTruncate, kArmVrintzF32) \ |
| 1370 V(Float64RoundTruncate, kArmVrintzF64) \ |
| 1371 V(Float64RoundTiesAway, kArmVrintaF64) \ |
| 1372 V(Float32RoundTiesEven, kArmVrintnF32) \ |
| 1373 V(Float64RoundTiesEven, kArmVrintnF64) |
| 1340 | 1374 |
| 1375 #define RRR_OP_LIST(V) \ |
| 1376 V(Int32MulHigh, kArmSmmul) \ |
| 1377 V(Float32Mul, kArmVmulF32) \ |
| 1378 V(Float64Mul, kArmVmulF64) \ |
| 1379 V(Float32Div, kArmVdivF32) \ |
| 1380 V(Float64Div, kArmVdivF64) \ |
| 1381 V(Float32Max, kArmFloat32Max) \ |
| 1382 V(Float64Max, kArmFloat64Max) \ |
| 1383 V(Float32Min, kArmFloat32Min) \ |
| 1384 V(Float64Min, kArmFloat64Min) |
| 1341 | 1385 |
| 1342 void InstructionSelector::VisitRoundInt32ToFloat32(Node* node) { | 1386 #define RR_VISITOR(Name, opcode) \ |
| 1343 VisitRR(this, kArmVcvtF32S32, node); | 1387 void InstructionSelector::Visit##Name(Node* node) { \ |
| 1344 } | 1388 VisitRR(this, opcode, node); \ |
| 1389 } |
| 1390 RR_OP_LIST(RR_VISITOR) |
| 1391 #undef RR_VISITOR |
| 1345 | 1392 |
| 1393 #define RR_VISITOR_V8(Name, opcode) \ |
| 1394 void InstructionSelector::Visit##Name(Node* node) { \ |
| 1395 DCHECK(CpuFeatures::IsSupported(ARMv8)); \ |
| 1396 VisitRR(this, opcode, node); \ |
| 1397 } |
| 1398 RR_OP_LIST_V8(RR_VISITOR_V8) |
| 1399 #undef RR_VISITOR_V8 |
| 1346 | 1400 |
| 1347 void InstructionSelector::VisitRoundUint32ToFloat32(Node* node) { | 1401 #define RRR_VISITOR(Name, opcode) \ |
| 1348 VisitRR(this, kArmVcvtF32U32, node); | 1402 void InstructionSelector::Visit##Name(Node* node) { \ |
| 1349 } | 1403 VisitRRR(this, opcode, node); \ |
| 1350 | 1404 } |
| 1351 | 1405 RRR_OP_LIST(RRR_VISITOR) |
| 1352 void InstructionSelector::VisitChangeInt32ToFloat64(Node* node) { | 1406 #undef RRR_VISITOR |
| 1353 VisitRR(this, kArmVcvtF64S32, node); | |
| 1354 } | |
| 1355 | |
| 1356 | |
| 1357 void InstructionSelector::VisitChangeUint32ToFloat64(Node* node) { | |
| 1358 VisitRR(this, kArmVcvtF64U32, node); | |
| 1359 } | |
| 1360 | |
| 1361 | |
| 1362 void InstructionSelector::VisitTruncateFloat32ToInt32(Node* node) { | |
| 1363 VisitRR(this, kArmVcvtS32F32, node); | |
| 1364 } | |
| 1365 | |
| 1366 | |
| 1367 void InstructionSelector::VisitTruncateFloat32ToUint32(Node* node) { | |
| 1368 VisitRR(this, kArmVcvtU32F32, node); | |
| 1369 } | |
| 1370 | |
| 1371 | |
| 1372 void InstructionSelector::VisitChangeFloat64ToInt32(Node* node) { | |
| 1373 VisitRR(this, kArmVcvtS32F64, node); | |
| 1374 } | |
| 1375 | |
| 1376 | |
| 1377 void InstructionSelector::VisitChangeFloat64ToUint32(Node* node) { | |
| 1378 VisitRR(this, kArmVcvtU32F64, node); | |
| 1379 } | |
| 1380 | |
| 1381 void InstructionSelector::VisitTruncateFloat64ToUint32(Node* node) { | |
| 1382 VisitRR(this, kArmVcvtU32F64, node); | |
| 1383 } | |
| 1384 void InstructionSelector::VisitTruncateFloat64ToFloat32(Node* node) { | |
| 1385 VisitRR(this, kArmVcvtF32F64, node); | |
| 1386 } | |
| 1387 | |
| 1388 void InstructionSelector::VisitTruncateFloat64ToWord32(Node* node) { | |
| 1389 VisitRR(this, kArchTruncateDoubleToI, node); | |
| 1390 } | |
| 1391 | |
| 1392 void InstructionSelector::VisitRoundFloat64ToInt32(Node* node) { | |
| 1393 VisitRR(this, kArmVcvtS32F64, node); | |
| 1394 } | |
| 1395 | |
| 1396 void InstructionSelector::VisitBitcastFloat32ToInt32(Node* node) { | |
| 1397 VisitRR(this, kArmVmovU32F32, node); | |
| 1398 } | |
| 1399 | |
| 1400 void InstructionSelector::VisitBitcastInt32ToFloat32(Node* node) { | |
| 1401 VisitRR(this, kArmVmovF32U32, node); | |
| 1402 } | |
| 1403 | 1407 |
| 1404 void InstructionSelector::VisitFloat32Add(Node* node) { | 1408 void InstructionSelector::VisitFloat32Add(Node* node) { |
| 1405 ArmOperandGenerator g(this); | 1409 ArmOperandGenerator g(this); |
| 1406 Float32BinopMatcher m(node); | 1410 Float32BinopMatcher m(node); |
| 1407 if (m.left().IsFloat32Mul() && CanCover(node, m.left().node())) { | 1411 if (m.left().IsFloat32Mul() && CanCover(node, m.left().node())) { |
| 1408 Float32BinopMatcher mleft(m.left().node()); | 1412 Float32BinopMatcher mleft(m.left().node()); |
| 1409 Emit(kArmVmlaF32, g.DefineSameAsFirst(node), | 1413 Emit(kArmVmlaF32, g.DefineSameAsFirst(node), |
| 1410 g.UseRegister(m.right().node()), g.UseRegister(mleft.left().node()), | 1414 g.UseRegister(m.right().node()), g.UseRegister(mleft.left().node()), |
| 1411 g.UseRegister(mleft.right().node())); | 1415 g.UseRegister(mleft.right().node())); |
| 1412 return; | 1416 return; |
| (...skipping 48 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1461 if (m.right().IsFloat64Mul() && CanCover(node, m.right().node())) { | 1465 if (m.right().IsFloat64Mul() && CanCover(node, m.right().node())) { |
| 1462 Float64BinopMatcher mright(m.right().node()); | 1466 Float64BinopMatcher mright(m.right().node()); |
| 1463 Emit(kArmVmlsF64, g.DefineSameAsFirst(node), g.UseRegister(m.left().node()), | 1467 Emit(kArmVmlsF64, g.DefineSameAsFirst(node), g.UseRegister(m.left().node()), |
| 1464 g.UseRegister(mright.left().node()), | 1468 g.UseRegister(mright.left().node()), |
| 1465 g.UseRegister(mright.right().node())); | 1469 g.UseRegister(mright.right().node())); |
| 1466 return; | 1470 return; |
| 1467 } | 1471 } |
| 1468 VisitRRR(this, kArmVsubF64, node); | 1472 VisitRRR(this, kArmVsubF64, node); |
| 1469 } | 1473 } |
| 1470 | 1474 |
| 1471 void InstructionSelector::VisitFloat32Mul(Node* node) { | |
| 1472 VisitRRR(this, kArmVmulF32, node); | |
| 1473 } | |
| 1474 | |
| 1475 | |
| 1476 void InstructionSelector::VisitFloat64Mul(Node* node) { | |
| 1477 VisitRRR(this, kArmVmulF64, node); | |
| 1478 } | |
| 1479 | |
| 1480 | |
| 1481 void InstructionSelector::VisitFloat32Div(Node* node) { | |
| 1482 VisitRRR(this, kArmVdivF32, node); | |
| 1483 } | |
| 1484 | |
| 1485 | |
| 1486 void InstructionSelector::VisitFloat64Div(Node* node) { | |
| 1487 VisitRRR(this, kArmVdivF64, node); | |
| 1488 } | |
| 1489 | |
| 1490 | |
| 1491 void InstructionSelector::VisitFloat64Mod(Node* node) { | 1475 void InstructionSelector::VisitFloat64Mod(Node* node) { |
| 1492 ArmOperandGenerator g(this); | 1476 ArmOperandGenerator g(this); |
| 1493 Emit(kArmVmodF64, g.DefineAsFixed(node, d0), g.UseFixed(node->InputAt(0), d0), | 1477 Emit(kArmVmodF64, g.DefineAsFixed(node, d0), g.UseFixed(node->InputAt(0), d0), |
| 1494 g.UseFixed(node->InputAt(1), d1))->MarkAsCall(); | 1478 g.UseFixed(node->InputAt(1), d1))->MarkAsCall(); |
| 1495 } | 1479 } |
| 1496 | 1480 |
| 1497 void InstructionSelector::VisitFloat32Max(Node* node) { | |
| 1498 VisitRRR(this, kArmFloat32Max, node); | |
| 1499 } | |
| 1500 | |
| 1501 void InstructionSelector::VisitFloat64Max(Node* node) { | |
| 1502 VisitRRR(this, kArmFloat64Max, node); | |
| 1503 } | |
| 1504 | |
| 1505 void InstructionSelector::VisitFloat64SilenceNaN(Node* node) { | |
| 1506 VisitRR(this, kArmFloat64SilenceNaN, node); | |
| 1507 } | |
| 1508 | |
| 1509 void InstructionSelector::VisitFloat32Min(Node* node) { | |
| 1510 VisitRRR(this, kArmFloat32Min, node); | |
| 1511 } | |
| 1512 | |
| 1513 void InstructionSelector::VisitFloat64Min(Node* node) { | |
| 1514 VisitRRR(this, kArmFloat64Min, node); | |
| 1515 } | |
| 1516 | |
| 1517 void InstructionSelector::VisitFloat32Abs(Node* node) { | |
| 1518 VisitRR(this, kArmVabsF32, node); | |
| 1519 } | |
| 1520 | |
| 1521 | |
| 1522 void InstructionSelector::VisitFloat64Abs(Node* node) { | |
| 1523 VisitRR(this, kArmVabsF64, node); | |
| 1524 } | |
| 1525 | |
| 1526 void InstructionSelector::VisitFloat32Sqrt(Node* node) { | |
| 1527 VisitRR(this, kArmVsqrtF32, node); | |
| 1528 } | |
| 1529 | |
| 1530 | |
| 1531 void InstructionSelector::VisitFloat64Sqrt(Node* node) { | |
| 1532 VisitRR(this, kArmVsqrtF64, node); | |
| 1533 } | |
| 1534 | |
| 1535 | |
| 1536 void InstructionSelector::VisitFloat32RoundDown(Node* node) { | |
| 1537 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1538 VisitRR(this, kArmVrintmF32, node); | |
| 1539 } | |
| 1540 | |
| 1541 | |
| 1542 void InstructionSelector::VisitFloat64RoundDown(Node* node) { | |
| 1543 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1544 VisitRR(this, kArmVrintmF64, node); | |
| 1545 } | |
| 1546 | |
| 1547 | |
| 1548 void InstructionSelector::VisitFloat32RoundUp(Node* node) { | |
| 1549 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1550 VisitRR(this, kArmVrintpF32, node); | |
| 1551 } | |
| 1552 | |
| 1553 | |
| 1554 void InstructionSelector::VisitFloat64RoundUp(Node* node) { | |
| 1555 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1556 VisitRR(this, kArmVrintpF64, node); | |
| 1557 } | |
| 1558 | |
| 1559 | |
| 1560 void InstructionSelector::VisitFloat32RoundTruncate(Node* node) { | |
| 1561 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1562 VisitRR(this, kArmVrintzF32, node); | |
| 1563 } | |
| 1564 | |
| 1565 | |
| 1566 void InstructionSelector::VisitFloat64RoundTruncate(Node* node) { | |
| 1567 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1568 VisitRR(this, kArmVrintzF64, node); | |
| 1569 } | |
| 1570 | |
| 1571 | |
| 1572 void InstructionSelector::VisitFloat64RoundTiesAway(Node* node) { | |
| 1573 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1574 VisitRR(this, kArmVrintaF64, node); | |
| 1575 } | |
| 1576 | |
| 1577 | |
| 1578 void InstructionSelector::VisitFloat32RoundTiesEven(Node* node) { | |
| 1579 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1580 VisitRR(this, kArmVrintnF32, node); | |
| 1581 } | |
| 1582 | |
| 1583 | |
| 1584 void InstructionSelector::VisitFloat64RoundTiesEven(Node* node) { | |
| 1585 DCHECK(CpuFeatures::IsSupported(ARMv8)); | |
| 1586 VisitRR(this, kArmVrintnF64, node); | |
| 1587 } | |
| 1588 | |
| 1589 void InstructionSelector::VisitFloat32Neg(Node* node) { | |
| 1590 VisitRR(this, kArmVnegF32, node); | |
| 1591 } | |
| 1592 | |
| 1593 void InstructionSelector::VisitFloat64Neg(Node* node) { | |
| 1594 VisitRR(this, kArmVnegF64, node); | |
| 1595 } | |
| 1596 | |
| 1597 void InstructionSelector::VisitFloat64Ieee754Binop(Node* node, | 1481 void InstructionSelector::VisitFloat64Ieee754Binop(Node* node, |
| 1598 InstructionCode opcode) { | 1482 InstructionCode opcode) { |
| 1599 ArmOperandGenerator g(this); | 1483 ArmOperandGenerator g(this); |
| 1600 Emit(opcode, g.DefineAsFixed(node, d0), g.UseFixed(node->InputAt(0), d0), | 1484 Emit(opcode, g.DefineAsFixed(node, d0), g.UseFixed(node->InputAt(0), d0), |
| 1601 g.UseFixed(node->InputAt(1), d1)) | 1485 g.UseFixed(node->InputAt(1), d1)) |
| 1602 ->MarkAsCall(); | 1486 ->MarkAsCall(); |
| 1603 } | 1487 } |
| 1604 | 1488 |
| 1605 void InstructionSelector::VisitFloat64Ieee754Unop(Node* node, | 1489 void InstructionSelector::VisitFloat64Ieee754Unop(Node* node, |
| 1606 InstructionCode opcode) { | 1490 InstructionCode opcode) { |
| (...skipping 574 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2181 VisitFloat64Compare(this, node, &cont); | 2065 VisitFloat64Compare(this, node, &cont); |
| 2182 } | 2066 } |
| 2183 | 2067 |
| 2184 | 2068 |
| 2185 void InstructionSelector::VisitFloat64LessThanOrEqual(Node* node) { | 2069 void InstructionSelector::VisitFloat64LessThanOrEqual(Node* node) { |
| 2186 FlagsContinuation cont = | 2070 FlagsContinuation cont = |
| 2187 FlagsContinuation::ForSet(kFloatLessThanOrEqual, node); | 2071 FlagsContinuation::ForSet(kFloatLessThanOrEqual, node); |
| 2188 VisitFloat64Compare(this, node, &cont); | 2072 VisitFloat64Compare(this, node, &cont); |
| 2189 } | 2073 } |
| 2190 | 2074 |
| 2191 | |
| 2192 void InstructionSelector::VisitFloat64ExtractLowWord32(Node* node) { | |
| 2193 VisitRR(this, kArmVmovLowU32F64, node); | |
| 2194 } | |
| 2195 | |
| 2196 | |
| 2197 void InstructionSelector::VisitFloat64ExtractHighWord32(Node* node) { | |
| 2198 VisitRR(this, kArmVmovHighU32F64, node); | |
| 2199 } | |
| 2200 | |
| 2201 | |
| 2202 void InstructionSelector::VisitFloat64InsertLowWord32(Node* node) { | 2075 void InstructionSelector::VisitFloat64InsertLowWord32(Node* node) { |
| 2203 ArmOperandGenerator g(this); | 2076 ArmOperandGenerator g(this); |
| 2204 Node* left = node->InputAt(0); | 2077 Node* left = node->InputAt(0); |
| 2205 Node* right = node->InputAt(1); | 2078 Node* right = node->InputAt(1); |
| 2206 if (left->opcode() == IrOpcode::kFloat64InsertHighWord32 && | 2079 if (left->opcode() == IrOpcode::kFloat64InsertHighWord32 && |
| 2207 CanCover(node, left)) { | 2080 CanCover(node, left)) { |
| 2208 left = left->InputAt(1); | 2081 left = left->InputAt(1); |
| 2209 Emit(kArmVmovF64U32U32, g.DefineAsRegister(node), g.UseRegister(right), | 2082 Emit(kArmVmovF64U32U32, g.DefineAsRegister(node), g.UseRegister(right), |
| 2210 g.UseRegister(left)); | 2083 g.UseRegister(left)); |
| 2211 return; | 2084 return; |
| (...skipping 67 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2279 AddressingMode addressing_mode = kMode_Offset_RR; | 2152 AddressingMode addressing_mode = kMode_Offset_RR; |
| 2280 InstructionOperand inputs[4]; | 2153 InstructionOperand inputs[4]; |
| 2281 size_t input_count = 0; | 2154 size_t input_count = 0; |
| 2282 inputs[input_count++] = g.UseUniqueRegister(base); | 2155 inputs[input_count++] = g.UseUniqueRegister(base); |
| 2283 inputs[input_count++] = g.UseUniqueRegister(index); | 2156 inputs[input_count++] = g.UseUniqueRegister(index); |
| 2284 inputs[input_count++] = g.UseUniqueRegister(value); | 2157 inputs[input_count++] = g.UseUniqueRegister(value); |
| 2285 InstructionCode code = opcode | AddressingModeField::encode(addressing_mode); | 2158 InstructionCode code = opcode | AddressingModeField::encode(addressing_mode); |
| 2286 Emit(code, 0, nullptr, input_count, inputs); | 2159 Emit(code, 0, nullptr, input_count, inputs); |
| 2287 } | 2160 } |
| 2288 | 2161 |
| 2289 // TODO(bbudge) Macro-ize SIMD methods. | 2162 #define SIMD_TYPE_LIST(V) \ |
| 2290 void InstructionSelector::VisitCreateFloat32x4(Node* node) { | 2163 V(Float32x4) \ |
| 2291 ArmOperandGenerator g(this); | 2164 V(Int32x4) \ |
| 2292 Emit(kArmFloat32x4Splat, g.DefineAsRegister(node), g.Use(node->InputAt(0))); | 2165 V(Int16x8) \ |
| 2293 } | 2166 V(Int8x16) |
| 2294 | 2167 |
| 2295 void InstructionSelector::VisitFloat32x4ExtractLane(Node* node) { | 2168 #define SIMD_UNOP_LIST(V) \ |
| 2296 ArmOperandGenerator g(this); | 2169 V(Float32x4FromInt32x4) \ |
| 2297 int32_t lane = OpParameter<int32_t>(node); | 2170 V(Float32x4FromUint32x4) \ |
| 2298 Emit(kArmFloat32x4ExtractLane, g.DefineAsRegister(node), | 2171 V(Float32x4Abs) \ |
| 2299 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane)); | 2172 V(Float32x4Neg) \ |
| 2300 } | 2173 V(Int32x4FromFloat32x4) \ |
| 2174 V(Uint32x4FromFloat32x4) \ |
| 2175 V(Int32x4Neg) \ |
| 2176 V(Int16x8Neg) \ |
| 2177 V(Int8x16Neg) |
| 2301 | 2178 |
| 2302 void InstructionSelector::VisitFloat32x4ReplaceLane(Node* node) { | 2179 #define SIMD_BINOP_LIST(V) \ |
| 2303 ArmOperandGenerator g(this); | 2180 V(Float32x4Add) \ |
| 2304 int32_t lane = OpParameter<int32_t>(node); | 2181 V(Float32x4Sub) \ |
| 2305 Emit(kArmFloat32x4ReplaceLane, g.DefineAsRegister(node), | 2182 V(Float32x4Equal) \ |
| 2306 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane), | 2183 V(Float32x4NotEqual) \ |
| 2307 g.Use(node->InputAt(1))); | 2184 V(Int32x4Add) \ |
| 2308 } | 2185 V(Int32x4Sub) \ |
| 2186 V(Int32x4Mul) \ |
| 2187 V(Int32x4Min) \ |
| 2188 V(Int32x4Max) \ |
| 2189 V(Int32x4Equal) \ |
| 2190 V(Int32x4NotEqual) \ |
| 2191 V(Int32x4GreaterThan) \ |
| 2192 V(Int32x4GreaterThanOrEqual) \ |
| 2193 V(Uint32x4GreaterThan) \ |
| 2194 V(Uint32x4GreaterThanOrEqual) \ |
| 2195 V(Int16x8Add) \ |
| 2196 V(Int16x8Sub) \ |
| 2197 V(Int16x8Mul) \ |
| 2198 V(Int16x8Min) \ |
| 2199 V(Int16x8Max) \ |
| 2200 V(Int16x8Equal) \ |
| 2201 V(Int16x8NotEqual) \ |
| 2202 V(Int16x8GreaterThan) \ |
| 2203 V(Int16x8GreaterThanOrEqual) \ |
| 2204 V(Uint16x8GreaterThan) \ |
| 2205 V(Uint16x8GreaterThanOrEqual) \ |
| 2206 V(Int8x16Add) \ |
| 2207 V(Int8x16Sub) \ |
| 2208 V(Int8x16Mul) \ |
| 2209 V(Int8x16Min) \ |
| 2210 V(Int8x16Max) \ |
| 2211 V(Int8x16Equal) \ |
| 2212 V(Int8x16NotEqual) \ |
| 2213 V(Int8x16GreaterThan) \ |
| 2214 V(Int8x16GreaterThanOrEqual) \ |
| 2215 V(Uint8x16GreaterThan) \ |
| 2216 V(Uint8x16GreaterThanOrEqual) |
| 2309 | 2217 |
| 2310 void InstructionSelector::VisitFloat32x4FromInt32x4(Node* node) { | 2218 #define SIMD_VISIT_SPLAT(Type) \ |
| 2311 ArmOperandGenerator g(this); | 2219 void InstructionSelector::VisitCreate##Type(Node* node) { \ |
| 2312 Emit(kArmFloat32x4FromInt32x4, g.DefineAsRegister(node), | 2220 VisitRR(this, kArm##Type##Splat, node); \ |
| 2313 g.UseRegister(node->InputAt(0))); | 2221 } |
| 2314 } | 2222 SIMD_TYPE_LIST(SIMD_VISIT_SPLAT) |
| 2223 #undef SIMD_VISIT_SPLAT |
| 2315 | 2224 |
| 2316 void InstructionSelector::VisitFloat32x4FromUint32x4(Node* node) { | 2225 #define SIMD_VISIT_EXTRACT_LANE(Type) \ |
| 2317 ArmOperandGenerator g(this); | 2226 void InstructionSelector::Visit##Type##ExtractLane(Node* node) { \ |
| 2318 Emit(kArmFloat32x4FromUint32x4, g.DefineAsRegister(node), | 2227 VisitRRI(this, kArm##Type##ExtractLane, node); \ |
| 2319 g.UseRegister(node->InputAt(0))); | 2228 } |
| 2320 } | 2229 SIMD_TYPE_LIST(SIMD_VISIT_EXTRACT_LANE) |
| 2230 #undef SIMD_VISIT_EXTRACT_LANE |
| 2321 | 2231 |
| 2322 void InstructionSelector::VisitFloat32x4Abs(Node* node) { | 2232 #define SIMD_VISIT_REPLACE_LANE(Type) \ |
| 2323 ArmOperandGenerator g(this); | 2233 void InstructionSelector::Visit##Type##ReplaceLane(Node* node) { \ |
| 2324 Emit(kArmFloat32x4Abs, g.DefineAsRegister(node), | 2234 VisitRRIR(this, kArm##Type##ReplaceLane, node); \ |
| 2325 g.UseRegister(node->InputAt(0))); | 2235 } |
| 2326 } | 2236 SIMD_TYPE_LIST(SIMD_VISIT_REPLACE_LANE) |
| 2237 #undef SIMD_VISIT_REPLACE_LANE |
| 2327 | 2238 |
| 2328 void InstructionSelector::VisitFloat32x4Neg(Node* node) { | 2239 #define SIMD_VISIT_UNOP(Name) \ |
| 2329 ArmOperandGenerator g(this); | 2240 void InstructionSelector::Visit##Name(Node* node) { \ |
| 2330 Emit(kArmFloat32x4Neg, g.DefineAsRegister(node), | 2241 VisitRR(this, kArm##Name, node); \ |
| 2331 g.UseRegister(node->InputAt(0))); | 2242 } |
| 2332 } | 2243 SIMD_UNOP_LIST(SIMD_VISIT_UNOP) |
| 2244 #undef SIMD_VISIT_UNOP |
| 2333 | 2245 |
| 2334 void InstructionSelector::VisitFloat32x4Add(Node* node) { | 2246 #define SIMD_VISIT_BINOP(Name) \ |
| 2335 ArmOperandGenerator g(this); | 2247 void InstructionSelector::Visit##Name(Node* node) { \ |
| 2336 Emit(kArmFloat32x4Add, g.DefineAsRegister(node), | 2248 VisitRRR(this, kArm##Name, node); \ |
| 2337 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | 2249 } |
| 2338 } | 2250 SIMD_BINOP_LIST(SIMD_VISIT_BINOP) |
| 2339 | 2251 #undef SIMD_VISIT_BINOP |
| 2340 void InstructionSelector::VisitFloat32x4Sub(Node* node) { | |
| 2341 ArmOperandGenerator g(this); | |
| 2342 Emit(kArmFloat32x4Sub, g.DefineAsRegister(node), | |
| 2343 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2344 } | |
| 2345 | |
| 2346 void InstructionSelector::VisitFloat32x4Equal(Node* node) { | |
| 2347 ArmOperandGenerator g(this); | |
| 2348 Emit(kArmFloat32x4Eq, g.DefineAsRegister(node), | |
| 2349 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2350 } | |
| 2351 | |
| 2352 void InstructionSelector::VisitFloat32x4NotEqual(Node* node) { | |
| 2353 ArmOperandGenerator g(this); | |
| 2354 Emit(kArmFloat32x4Ne, g.DefineAsRegister(node), | |
| 2355 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2356 } | |
| 2357 | |
| 2358 void InstructionSelector::VisitCreateInt32x4(Node* node) { | |
| 2359 ArmOperandGenerator g(this); | |
| 2360 Emit(kArmInt32x4Splat, g.DefineAsRegister(node), g.Use(node->InputAt(0))); | |
| 2361 } | |
| 2362 | |
| 2363 void InstructionSelector::VisitInt32x4ExtractLane(Node* node) { | |
| 2364 ArmOperandGenerator g(this); | |
| 2365 int32_t lane = OpParameter<int32_t>(node); | |
| 2366 Emit(kArmInt32x4ExtractLane, g.DefineAsRegister(node), | |
| 2367 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane)); | |
| 2368 } | |
| 2369 | |
| 2370 void InstructionSelector::VisitInt32x4ReplaceLane(Node* node) { | |
| 2371 ArmOperandGenerator g(this); | |
| 2372 int32_t lane = OpParameter<int32_t>(node); | |
| 2373 Emit(kArmInt32x4ReplaceLane, g.DefineAsRegister(node), | |
| 2374 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane), | |
| 2375 g.Use(node->InputAt(1))); | |
| 2376 } | |
| 2377 | |
| 2378 void InstructionSelector::VisitInt32x4FromFloat32x4(Node* node) { | |
| 2379 ArmOperandGenerator g(this); | |
| 2380 Emit(kArmInt32x4FromFloat32x4, g.DefineAsRegister(node), | |
| 2381 g.UseRegister(node->InputAt(0))); | |
| 2382 } | |
| 2383 | |
| 2384 void InstructionSelector::VisitUint32x4FromFloat32x4(Node* node) { | |
| 2385 ArmOperandGenerator g(this); | |
| 2386 Emit(kArmUint32x4FromFloat32x4, g.DefineAsRegister(node), | |
| 2387 g.UseRegister(node->InputAt(0))); | |
| 2388 } | |
| 2389 | |
| 2390 void InstructionSelector::VisitInt32x4Neg(Node* node) { | |
| 2391 ArmOperandGenerator g(this); | |
| 2392 Emit(kArmInt32x4Neg, g.DefineAsRegister(node), | |
| 2393 g.UseRegister(node->InputAt(0))); | |
| 2394 } | |
| 2395 | |
| 2396 void InstructionSelector::VisitInt32x4Add(Node* node) { | |
| 2397 ArmOperandGenerator g(this); | |
| 2398 Emit(kArmInt32x4Add, g.DefineAsRegister(node), | |
| 2399 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2400 } | |
| 2401 | |
| 2402 void InstructionSelector::VisitInt32x4Sub(Node* node) { | |
| 2403 ArmOperandGenerator g(this); | |
| 2404 Emit(kArmInt32x4Sub, g.DefineAsRegister(node), | |
| 2405 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2406 } | |
| 2407 | |
| 2408 void InstructionSelector::VisitInt32x4Mul(Node* node) { | |
| 2409 ArmOperandGenerator g(this); | |
| 2410 Emit(kArmInt32x4Mul, g.DefineAsRegister(node), | |
| 2411 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2412 } | |
| 2413 | |
| 2414 void InstructionSelector::VisitInt32x4Min(Node* node) { | |
| 2415 ArmOperandGenerator g(this); | |
| 2416 Emit(kArmInt32x4Min, g.DefineAsRegister(node), | |
| 2417 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2418 } | |
| 2419 | |
| 2420 void InstructionSelector::VisitInt32x4Max(Node* node) { | |
| 2421 ArmOperandGenerator g(this); | |
| 2422 Emit(kArmInt32x4Max, g.DefineAsRegister(node), | |
| 2423 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2424 } | |
| 2425 | |
| 2426 void InstructionSelector::VisitInt32x4Equal(Node* node) { | |
| 2427 ArmOperandGenerator g(this); | |
| 2428 Emit(kArmInt32x4Eq, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2429 g.UseRegister(node->InputAt(1))); | |
| 2430 } | |
| 2431 | |
| 2432 void InstructionSelector::VisitInt32x4NotEqual(Node* node) { | |
| 2433 ArmOperandGenerator g(this); | |
| 2434 Emit(kArmInt32x4Ne, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2435 g.UseRegister(node->InputAt(1))); | |
| 2436 } | |
| 2437 | |
| 2438 void InstructionSelector::VisitInt32x4GreaterThan(Node* node) { | |
| 2439 ArmOperandGenerator g(this); | |
| 2440 Emit(kArmInt32x4Gt, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2441 g.UseRegister(node->InputAt(1))); | |
| 2442 } | |
| 2443 | |
| 2444 void InstructionSelector::VisitInt32x4GreaterThanOrEqual(Node* node) { | |
| 2445 ArmOperandGenerator g(this); | |
| 2446 Emit(kArmInt32x4Ge, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2447 g.UseRegister(node->InputAt(1))); | |
| 2448 } | |
| 2449 | |
| 2450 void InstructionSelector::VisitUint32x4GreaterThan(Node* node) { | |
| 2451 ArmOperandGenerator g(this); | |
| 2452 Emit(kArmUint32x4Gt, g.DefineAsRegister(node), | |
| 2453 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2454 } | |
| 2455 | |
| 2456 void InstructionSelector::VisitUint32x4GreaterThanOrEqual(Node* node) { | |
| 2457 ArmOperandGenerator g(this); | |
| 2458 Emit(kArmUint32x4Ge, g.DefineAsRegister(node), | |
| 2459 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2460 } | |
| 2461 | 2252 |
| 2462 void InstructionSelector::VisitSimd32x4Select(Node* node) { | 2253 void InstructionSelector::VisitSimd32x4Select(Node* node) { |
| 2463 ArmOperandGenerator g(this); | 2254 ArmOperandGenerator g(this); |
| 2464 Emit(kArmSimd32x4Select, g.DefineAsRegister(node), | 2255 Emit(kArmSimd32x4Select, g.DefineAsRegister(node), |
| 2465 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1)), | 2256 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1)), |
| 2466 g.UseRegister(node->InputAt(2))); | 2257 g.UseRegister(node->InputAt(2))); |
| 2467 } | 2258 } |
| 2468 | 2259 |
| 2469 void InstructionSelector::VisitCreateInt16x8(Node* node) { | |
| 2470 ArmOperandGenerator g(this); | |
| 2471 Emit(kArmInt16x8Splat, g.DefineAsRegister(node), g.Use(node->InputAt(0))); | |
| 2472 } | |
| 2473 | |
| 2474 void InstructionSelector::VisitInt16x8ExtractLane(Node* node) { | |
| 2475 ArmOperandGenerator g(this); | |
| 2476 int32_t lane = OpParameter<int32_t>(node); | |
| 2477 Emit(kArmInt16x8ExtractLane, g.DefineAsRegister(node), | |
| 2478 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane)); | |
| 2479 } | |
| 2480 | |
| 2481 void InstructionSelector::VisitInt16x8ReplaceLane(Node* node) { | |
| 2482 ArmOperandGenerator g(this); | |
| 2483 int32_t lane = OpParameter<int32_t>(node); | |
| 2484 Emit(kArmInt16x8ReplaceLane, g.DefineAsRegister(node), | |
| 2485 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane), | |
| 2486 g.Use(node->InputAt(1))); | |
| 2487 } | |
| 2488 | |
| 2489 void InstructionSelector::VisitInt16x8Neg(Node* node) { | |
| 2490 ArmOperandGenerator g(this); | |
| 2491 Emit(kArmInt16x8Neg, g.DefineAsRegister(node), | |
| 2492 g.UseRegister(node->InputAt(0))); | |
| 2493 } | |
| 2494 | |
| 2495 void InstructionSelector::VisitInt16x8Add(Node* node) { | |
| 2496 ArmOperandGenerator g(this); | |
| 2497 Emit(kArmInt16x8Add, g.DefineAsRegister(node), | |
| 2498 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2499 } | |
| 2500 | |
| 2501 void InstructionSelector::VisitInt16x8Sub(Node* node) { | |
| 2502 ArmOperandGenerator g(this); | |
| 2503 Emit(kArmInt16x8Sub, g.DefineAsRegister(node), | |
| 2504 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2505 } | |
| 2506 | |
| 2507 void InstructionSelector::VisitInt16x8Mul(Node* node) { | |
| 2508 ArmOperandGenerator g(this); | |
| 2509 Emit(kArmInt16x8Mul, g.DefineAsRegister(node), | |
| 2510 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2511 } | |
| 2512 | |
| 2513 void InstructionSelector::VisitInt16x8Min(Node* node) { | |
| 2514 ArmOperandGenerator g(this); | |
| 2515 Emit(kArmInt16x8Min, g.DefineAsRegister(node), | |
| 2516 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2517 } | |
| 2518 | |
| 2519 void InstructionSelector::VisitInt16x8Max(Node* node) { | |
| 2520 ArmOperandGenerator g(this); | |
| 2521 Emit(kArmInt16x8Max, g.DefineAsRegister(node), | |
| 2522 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2523 } | |
| 2524 | |
| 2525 void InstructionSelector::VisitInt16x8Equal(Node* node) { | |
| 2526 ArmOperandGenerator g(this); | |
| 2527 Emit(kArmInt16x8Eq, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2528 g.UseRegister(node->InputAt(1))); | |
| 2529 } | |
| 2530 | |
| 2531 void InstructionSelector::VisitInt16x8NotEqual(Node* node) { | |
| 2532 ArmOperandGenerator g(this); | |
| 2533 Emit(kArmInt16x8Ne, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2534 g.UseRegister(node->InputAt(1))); | |
| 2535 } | |
| 2536 | |
| 2537 void InstructionSelector::VisitInt16x8GreaterThan(Node* node) { | |
| 2538 ArmOperandGenerator g(this); | |
| 2539 Emit(kArmInt16x8Gt, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2540 g.UseRegister(node->InputAt(1))); | |
| 2541 } | |
| 2542 | |
| 2543 void InstructionSelector::VisitInt16x8GreaterThanOrEqual(Node* node) { | |
| 2544 ArmOperandGenerator g(this); | |
| 2545 Emit(kArmInt16x8Ge, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2546 g.UseRegister(node->InputAt(1))); | |
| 2547 } | |
| 2548 | |
| 2549 void InstructionSelector::VisitUint16x8GreaterThan(Node* node) { | |
| 2550 ArmOperandGenerator g(this); | |
| 2551 Emit(kArmUint16x8Gt, g.DefineAsRegister(node), | |
| 2552 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2553 } | |
| 2554 | |
| 2555 void InstructionSelector::VisitUint16x8GreaterThanOrEqual(Node* node) { | |
| 2556 ArmOperandGenerator g(this); | |
| 2557 Emit(kArmUint16x8Ge, g.DefineAsRegister(node), | |
| 2558 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2559 } | |
| 2560 | |
| 2561 void InstructionSelector::VisitCreateInt8x16(Node* node) { | |
| 2562 ArmOperandGenerator g(this); | |
| 2563 Emit(kArmInt8x16Splat, g.DefineAsRegister(node), g.Use(node->InputAt(0))); | |
| 2564 } | |
| 2565 | |
| 2566 void InstructionSelector::VisitInt8x16ExtractLane(Node* node) { | |
| 2567 ArmOperandGenerator g(this); | |
| 2568 int32_t lane = OpParameter<int32_t>(node); | |
| 2569 Emit(kArmInt8x16ExtractLane, g.DefineAsRegister(node), | |
| 2570 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane)); | |
| 2571 } | |
| 2572 | |
| 2573 void InstructionSelector::VisitInt8x16ReplaceLane(Node* node) { | |
| 2574 ArmOperandGenerator g(this); | |
| 2575 int32_t lane = OpParameter<int32_t>(node); | |
| 2576 Emit(kArmInt8x16ReplaceLane, g.DefineAsRegister(node), | |
| 2577 g.UseRegister(node->InputAt(0)), g.UseImmediate(lane), | |
| 2578 g.Use(node->InputAt(1))); | |
| 2579 } | |
| 2580 | |
| 2581 void InstructionSelector::VisitInt8x16Neg(Node* node) { | |
| 2582 ArmOperandGenerator g(this); | |
| 2583 Emit(kArmInt8x16Neg, g.DefineAsRegister(node), | |
| 2584 g.UseRegister(node->InputAt(0))); | |
| 2585 } | |
| 2586 | |
| 2587 void InstructionSelector::VisitInt8x16Add(Node* node) { | |
| 2588 ArmOperandGenerator g(this); | |
| 2589 Emit(kArmInt8x16Add, g.DefineAsRegister(node), | |
| 2590 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2591 } | |
| 2592 | |
| 2593 void InstructionSelector::VisitInt8x16Sub(Node* node) { | |
| 2594 ArmOperandGenerator g(this); | |
| 2595 Emit(kArmInt8x16Sub, g.DefineAsRegister(node), | |
| 2596 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2597 } | |
| 2598 | |
| 2599 void InstructionSelector::VisitInt8x16Mul(Node* node) { | |
| 2600 ArmOperandGenerator g(this); | |
| 2601 Emit(kArmInt8x16Mul, g.DefineAsRegister(node), | |
| 2602 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2603 } | |
| 2604 | |
| 2605 void InstructionSelector::VisitInt8x16Min(Node* node) { | |
| 2606 ArmOperandGenerator g(this); | |
| 2607 Emit(kArmInt8x16Min, g.DefineAsRegister(node), | |
| 2608 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2609 } | |
| 2610 | |
| 2611 void InstructionSelector::VisitInt8x16Max(Node* node) { | |
| 2612 ArmOperandGenerator g(this); | |
| 2613 Emit(kArmInt8x16Max, g.DefineAsRegister(node), | |
| 2614 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2615 } | |
| 2616 | |
| 2617 void InstructionSelector::VisitInt8x16Equal(Node* node) { | |
| 2618 ArmOperandGenerator g(this); | |
| 2619 Emit(kArmInt8x16Eq, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2620 g.UseRegister(node->InputAt(1))); | |
| 2621 } | |
| 2622 | |
| 2623 void InstructionSelector::VisitInt8x16NotEqual(Node* node) { | |
| 2624 ArmOperandGenerator g(this); | |
| 2625 Emit(kArmInt8x16Ne, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2626 g.UseRegister(node->InputAt(1))); | |
| 2627 } | |
| 2628 | |
| 2629 void InstructionSelector::VisitInt8x16GreaterThan(Node* node) { | |
| 2630 ArmOperandGenerator g(this); | |
| 2631 Emit(kArmInt8x16Gt, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2632 g.UseRegister(node->InputAt(1))); | |
| 2633 } | |
| 2634 | |
| 2635 void InstructionSelector::VisitInt8x16GreaterThanOrEqual(Node* node) { | |
| 2636 ArmOperandGenerator g(this); | |
| 2637 Emit(kArmInt8x16Ge, g.DefineAsRegister(node), g.UseRegister(node->InputAt(0)), | |
| 2638 g.UseRegister(node->InputAt(1))); | |
| 2639 } | |
| 2640 | |
| 2641 void InstructionSelector::VisitUint8x16GreaterThan(Node* node) { | |
| 2642 ArmOperandGenerator g(this); | |
| 2643 Emit(kArmUint8x16Gt, g.DefineAsRegister(node), | |
| 2644 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2645 } | |
| 2646 | |
| 2647 void InstructionSelector::VisitUint8x16GreaterThanOrEqual(Node* node) { | |
| 2648 ArmOperandGenerator g(this); | |
| 2649 Emit(kArmUint8x16Ge, g.DefineAsRegister(node), | |
| 2650 g.UseRegister(node->InputAt(0)), g.UseRegister(node->InputAt(1))); | |
| 2651 } | |
| 2652 | |
| 2653 // static | 2260 // static |
| 2654 MachineOperatorBuilder::Flags | 2261 MachineOperatorBuilder::Flags |
| 2655 InstructionSelector::SupportedMachineOperatorFlags() { | 2262 InstructionSelector::SupportedMachineOperatorFlags() { |
| 2656 MachineOperatorBuilder::Flags flags; | 2263 MachineOperatorBuilder::Flags flags; |
| 2657 if (CpuFeatures::IsSupported(SUDIV)) { | 2264 if (CpuFeatures::IsSupported(SUDIV)) { |
| 2658 // The sdiv and udiv instructions correctly return 0 if the divisor is 0, | 2265 // The sdiv and udiv instructions correctly return 0 if the divisor is 0, |
| 2659 // but the fall-back implementation does not. | 2266 // but the fall-back implementation does not. |
| 2660 flags |= MachineOperatorBuilder::kInt32DivIsSafe | | 2267 flags |= MachineOperatorBuilder::kInt32DivIsSafe | |
| 2661 MachineOperatorBuilder::kUint32DivIsSafe; | 2268 MachineOperatorBuilder::kUint32DivIsSafe; |
| 2662 } | 2269 } |
| (...skipping 20 matching lines...) Expand all Loading... |
| 2683 Vector<MachineType> req_aligned = Vector<MachineType>::New(2); | 2290 Vector<MachineType> req_aligned = Vector<MachineType>::New(2); |
| 2684 req_aligned[0] = MachineType::Float32(); | 2291 req_aligned[0] = MachineType::Float32(); |
| 2685 req_aligned[1] = MachineType::Float64(); | 2292 req_aligned[1] = MachineType::Float64(); |
| 2686 return MachineOperatorBuilder::AlignmentRequirements:: | 2293 return MachineOperatorBuilder::AlignmentRequirements:: |
| 2687 SomeUnalignedAccessUnsupported(req_aligned, req_aligned); | 2294 SomeUnalignedAccessUnsupported(req_aligned, req_aligned); |
| 2688 } | 2295 } |
| 2689 | 2296 |
| 2690 } // namespace compiler | 2297 } // namespace compiler |
| 2691 } // namespace internal | 2298 } // namespace internal |
| 2692 } // namespace v8 | 2299 } // namespace v8 |
| OLD | NEW |