| OLD | NEW |
| (Empty) |
| 1 // Copyright 2015 the V8 project authors. All rights reserved. | |
| 2 // Use of this source code is governed by a BSD-style license that can be | |
| 3 // found in the LICENSE file. | |
| 4 | |
| 5 #include "src/compiler/instruction-scheduler.h" | |
| 6 | |
| 7 namespace v8 { | |
| 8 namespace internal { | |
| 9 namespace compiler { | |
| 10 | |
| 11 bool InstructionScheduler::SchedulerSupported() { return true; } | |
| 12 | |
| 13 | |
| 14 int InstructionScheduler::GetTargetInstructionFlags( | |
| 15 const Instruction* instr) const { | |
| 16 switch (instr->arch_opcode()) { | |
| 17 case kX64Add: | |
| 18 case kX64Add32: | |
| 19 case kX64And: | |
| 20 case kX64And32: | |
| 21 case kX64Cmp: | |
| 22 case kX64Cmp32: | |
| 23 case kX64Test: | |
| 24 case kX64Test32: | |
| 25 case kX64Or: | |
| 26 case kX64Or32: | |
| 27 case kX64Xor: | |
| 28 case kX64Xor32: | |
| 29 case kX64Sub: | |
| 30 case kX64Sub32: | |
| 31 case kX64Imul: | |
| 32 case kX64Imul32: | |
| 33 case kX64ImulHigh32: | |
| 34 case kX64UmulHigh32: | |
| 35 case kX64Idiv: | |
| 36 case kX64Idiv32: | |
| 37 case kX64Udiv: | |
| 38 case kX64Udiv32: | |
| 39 case kX64Not: | |
| 40 case kX64Not32: | |
| 41 case kX64Neg: | |
| 42 case kX64Neg32: | |
| 43 case kX64Shl: | |
| 44 case kX64Shl32: | |
| 45 case kX64Shr: | |
| 46 case kX64Shr32: | |
| 47 case kX64Sar: | |
| 48 case kX64Sar32: | |
| 49 case kX64Ror: | |
| 50 case kX64Ror32: | |
| 51 case kX64Lzcnt: | |
| 52 case kX64Lzcnt32: | |
| 53 case kX64Tzcnt: | |
| 54 case kX64Tzcnt32: | |
| 55 case kX64Popcnt: | |
| 56 case kX64Popcnt32: | |
| 57 case kSSEFloat32Cmp: | |
| 58 case kSSEFloat32Add: | |
| 59 case kSSEFloat32Sub: | |
| 60 case kSSEFloat32Mul: | |
| 61 case kSSEFloat32Div: | |
| 62 case kSSEFloat32Abs: | |
| 63 case kSSEFloat32Neg: | |
| 64 case kSSEFloat32Sqrt: | |
| 65 case kSSEFloat32Round: | |
| 66 case kSSEFloat32Max: | |
| 67 case kSSEFloat32Min: | |
| 68 case kSSEFloat32ToFloat64: | |
| 69 case kSSEFloat64Cmp: | |
| 70 case kSSEFloat64Add: | |
| 71 case kSSEFloat64Sub: | |
| 72 case kSSEFloat64Mul: | |
| 73 case kSSEFloat64Div: | |
| 74 case kSSEFloat64Mod: | |
| 75 case kSSEFloat64Abs: | |
| 76 case kSSEFloat64Neg: | |
| 77 case kSSEFloat64Sqrt: | |
| 78 case kSSEFloat64Round: | |
| 79 case kSSEFloat64Max: | |
| 80 case kSSEFloat64Min: | |
| 81 case kSSEFloat64ToFloat32: | |
| 82 case kSSEFloat64ToInt32: | |
| 83 case kSSEFloat64ToUint32: | |
| 84 case kSSEFloat64ToInt64: | |
| 85 case kSSEFloat32ToInt64: | |
| 86 case kSSEFloat64ToUint64: | |
| 87 case kSSEFloat32ToUint64: | |
| 88 case kSSEInt32ToFloat64: | |
| 89 case kSSEInt64ToFloat32: | |
| 90 case kSSEInt64ToFloat64: | |
| 91 case kSSEUint64ToFloat32: | |
| 92 case kSSEUint64ToFloat64: | |
| 93 case kSSEUint32ToFloat64: | |
| 94 case kSSEFloat64ExtractLowWord32: | |
| 95 case kSSEFloat64ExtractHighWord32: | |
| 96 case kSSEFloat64InsertLowWord32: | |
| 97 case kSSEFloat64InsertHighWord32: | |
| 98 case kSSEFloat64LoadLowWord32: | |
| 99 case kAVXFloat32Cmp: | |
| 100 case kAVXFloat32Add: | |
| 101 case kAVXFloat32Sub: | |
| 102 case kAVXFloat32Mul: | |
| 103 case kAVXFloat32Div: | |
| 104 case kAVXFloat32Max: | |
| 105 case kAVXFloat32Min: | |
| 106 case kAVXFloat64Cmp: | |
| 107 case kAVXFloat64Add: | |
| 108 case kAVXFloat64Sub: | |
| 109 case kAVXFloat64Mul: | |
| 110 case kAVXFloat64Div: | |
| 111 case kAVXFloat64Max: | |
| 112 case kAVXFloat64Min: | |
| 113 case kAVXFloat64Abs: | |
| 114 case kAVXFloat64Neg: | |
| 115 case kAVXFloat32Abs: | |
| 116 case kAVXFloat32Neg: | |
| 117 case kX64BitcastFI: | |
| 118 case kX64BitcastDL: | |
| 119 case kX64BitcastIF: | |
| 120 case kX64BitcastLD: | |
| 121 case kX64Lea32: | |
| 122 case kX64Lea: | |
| 123 case kX64Dec32: | |
| 124 case kX64Inc32: | |
| 125 return (instr->addressing_mode() == kMode_None) | |
| 126 ? kNoOpcodeFlags | |
| 127 : kIsLoadOperation | kHasSideEffect; | |
| 128 | |
| 129 case kX64Movsxbl: | |
| 130 case kX64Movzxbl: | |
| 131 case kX64Movsxwl: | |
| 132 case kX64Movzxwl: | |
| 133 case kX64Movsxlq: | |
| 134 DCHECK(instr->InputCount() >= 1); | |
| 135 return instr->InputAt(0)->IsRegister() ? kNoOpcodeFlags | |
| 136 : kIsLoadOperation; | |
| 137 | |
| 138 case kX64Movb: | |
| 139 case kX64Movw: | |
| 140 return kHasSideEffect; | |
| 141 | |
| 142 case kX64Movl: | |
| 143 if (instr->HasOutput()) { | |
| 144 DCHECK(instr->InputCount() >= 1); | |
| 145 return instr->InputAt(0)->IsRegister() ? kNoOpcodeFlags | |
| 146 : kIsLoadOperation; | |
| 147 } else { | |
| 148 return kHasSideEffect; | |
| 149 } | |
| 150 | |
| 151 case kX64Movq: | |
| 152 case kX64Movsd: | |
| 153 case kX64Movss: | |
| 154 return instr->HasOutput() ? kIsLoadOperation : kHasSideEffect; | |
| 155 | |
| 156 case kX64StackCheck: | |
| 157 return kIsLoadOperation; | |
| 158 | |
| 159 case kX64Push: | |
| 160 case kX64Poke: | |
| 161 return kHasSideEffect; | |
| 162 | |
| 163 #define CASE(Name) case k##Name: | |
| 164 COMMON_ARCH_OPCODE_LIST(CASE) | |
| 165 #undef CASE | |
| 166 // Already covered in architecture independent code. | |
| 167 UNREACHABLE(); | |
| 168 } | |
| 169 | |
| 170 UNREACHABLE(); | |
| 171 return kNoOpcodeFlags; | |
| 172 } | |
| 173 | |
| 174 | |
| 175 int InstructionScheduler::GetInstructionLatency(const Instruction* instr) { | |
| 176 // TODO(all): Add instruction cost modeling. | |
| 177 return 1; | |
| 178 } | |
| 179 | |
| 180 } // namespace compiler | |
| 181 } // namespace internal | |
| 182 } // namespace v8 | |
| OLD | NEW |