OLD | NEW |
1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_COMPILER_RAW_MACHINE_ASSEMBLER_H_ | 5 #ifndef V8_COMPILER_RAW_MACHINE_ASSEMBLER_H_ |
6 #define V8_COMPILER_RAW_MACHINE_ASSEMBLER_H_ | 6 #define V8_COMPILER_RAW_MACHINE_ASSEMBLER_H_ |
7 | 7 |
8 #include "src/assembler.h" | 8 #include "src/assembler.h" |
9 #include "src/compiler/common-operator.h" | 9 #include "src/compiler/common-operator.h" |
10 #include "src/compiler/graph.h" | 10 #include "src/compiler/graph.h" |
(...skipping 262 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
273 } | 273 } |
274 Node* Int32GreaterThan(Node* a, Node* b) { return Int32LessThan(b, a); } | 274 Node* Int32GreaterThan(Node* a, Node* b) { return Int32LessThan(b, a); } |
275 Node* Int32GreaterThanOrEqual(Node* a, Node* b) { | 275 Node* Int32GreaterThanOrEqual(Node* a, Node* b) { |
276 return Int32LessThanOrEqual(b, a); | 276 return Int32LessThanOrEqual(b, a); |
277 } | 277 } |
278 Node* Int32Neg(Node* a) { return Int32Sub(Int32Constant(0), a); } | 278 Node* Int32Neg(Node* a) { return Int32Sub(Int32Constant(0), a); } |
279 | 279 |
280 Node* Int64Add(Node* a, Node* b) { | 280 Node* Int64Add(Node* a, Node* b) { |
281 return AddNode(machine()->Int64Add(), a, b); | 281 return AddNode(machine()->Int64Add(), a, b); |
282 } | 282 } |
| 283 Node* Int64AddWithOverflow(Node* a, Node* b) { |
| 284 return AddNode(machine()->Int64AddWithOverflow(), a, b); |
| 285 } |
283 Node* Int64Sub(Node* a, Node* b) { | 286 Node* Int64Sub(Node* a, Node* b) { |
284 return AddNode(machine()->Int64Sub(), a, b); | 287 return AddNode(machine()->Int64Sub(), a, b); |
285 } | 288 } |
| 289 Node* Int64SubWithOverflow(Node* a, Node* b) { |
| 290 return AddNode(machine()->Int64SubWithOverflow(), a, b); |
| 291 } |
286 Node* Int64Mul(Node* a, Node* b) { | 292 Node* Int64Mul(Node* a, Node* b) { |
287 return AddNode(machine()->Int64Mul(), a, b); | 293 return AddNode(machine()->Int64Mul(), a, b); |
288 } | 294 } |
289 Node* Int64Div(Node* a, Node* b) { | 295 Node* Int64Div(Node* a, Node* b) { |
290 return AddNode(machine()->Int64Div(), a, b); | 296 return AddNode(machine()->Int64Div(), a, b); |
291 } | 297 } |
292 Node* Int64Mod(Node* a, Node* b) { | 298 Node* Int64Mod(Node* a, Node* b) { |
293 return AddNode(machine()->Int64Mod(), a, b); | 299 return AddNode(machine()->Int64Mod(), a, b); |
294 } | 300 } |
295 Node* Int64Neg(Node* a) { return Int64Sub(Int64Constant(0), a); } | 301 Node* Int64Neg(Node* a) { return Int64Sub(Int64Constant(0), a); } |
(...skipping 382 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
678 bool bound_; | 684 bool bound_; |
679 friend class RawMachineAssembler; | 685 friend class RawMachineAssembler; |
680 DISALLOW_COPY_AND_ASSIGN(RawMachineLabel); | 686 DISALLOW_COPY_AND_ASSIGN(RawMachineLabel); |
681 }; | 687 }; |
682 | 688 |
683 } // namespace compiler | 689 } // namespace compiler |
684 } // namespace internal | 690 } // namespace internal |
685 } // namespace v8 | 691 } // namespace v8 |
686 | 692 |
687 #endif // V8_COMPILER_RAW_MACHINE_ASSEMBLER_H_ | 693 #endif // V8_COMPILER_RAW_MACHINE_ASSEMBLER_H_ |
OLD | NEW |