| OLD | NEW |
| 1 // Copyright 2016 the V8 project authors. All rights reserved. | 1 // Copyright 2016 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ | 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ |
| 6 #define V8_CODE_STUB_ASSEMBLER_H_ | 6 #define V8_CODE_STUB_ASSEMBLER_H_ |
| 7 | 7 |
| 8 #include <functional> | 8 #include <functional> |
| 9 | 9 |
| 10 #include "src/compiler/code-assembler.h" | 10 #include "src/compiler/code-assembler.h" |
| (...skipping 64 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 75 compiler::Node* SmiAddWithOverflow(compiler::Node* a, compiler::Node* b); | 75 compiler::Node* SmiAddWithOverflow(compiler::Node* a, compiler::Node* b); |
| 76 compiler::Node* SmiSub(compiler::Node* a, compiler::Node* b); | 76 compiler::Node* SmiSub(compiler::Node* a, compiler::Node* b); |
| 77 compiler::Node* SmiSubWithOverflow(compiler::Node* a, compiler::Node* b); | 77 compiler::Node* SmiSubWithOverflow(compiler::Node* a, compiler::Node* b); |
| 78 compiler::Node* SmiEqual(compiler::Node* a, compiler::Node* b); | 78 compiler::Node* SmiEqual(compiler::Node* a, compiler::Node* b); |
| 79 compiler::Node* SmiAboveOrEqual(compiler::Node* a, compiler::Node* b); | 79 compiler::Node* SmiAboveOrEqual(compiler::Node* a, compiler::Node* b); |
| 80 compiler::Node* SmiLessThan(compiler::Node* a, compiler::Node* b); | 80 compiler::Node* SmiLessThan(compiler::Node* a, compiler::Node* b); |
| 81 compiler::Node* SmiLessThanOrEqual(compiler::Node* a, compiler::Node* b); | 81 compiler::Node* SmiLessThanOrEqual(compiler::Node* a, compiler::Node* b); |
| 82 compiler::Node* SmiMin(compiler::Node* a, compiler::Node* b); | 82 compiler::Node* SmiMin(compiler::Node* a, compiler::Node* b); |
| 83 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. | 83 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. |
| 84 compiler::Node* SmiMod(compiler::Node* a, compiler::Node* b); | 84 compiler::Node* SmiMod(compiler::Node* a, compiler::Node* b); |
| 85 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. |
| 86 compiler::Node* SmiMul(compiler::Node* a, compiler::Node* b); |
| 85 | 87 |
| 86 // Allocate an object of the given size. | 88 // Allocate an object of the given size. |
| 87 compiler::Node* Allocate(compiler::Node* size, AllocationFlags flags = kNone); | 89 compiler::Node* Allocate(compiler::Node* size, AllocationFlags flags = kNone); |
| 88 compiler::Node* Allocate(int size, AllocationFlags flags = kNone); | 90 compiler::Node* Allocate(int size, AllocationFlags flags = kNone); |
| 89 compiler::Node* InnerAllocate(compiler::Node* previous, int offset); | 91 compiler::Node* InnerAllocate(compiler::Node* previous, int offset); |
| 90 compiler::Node* InnerAllocate(compiler::Node* previous, | 92 compiler::Node* InnerAllocate(compiler::Node* previous, |
| 91 compiler::Node* offset); | 93 compiler::Node* offset); |
| 92 | 94 |
| 93 void Assert(compiler::Node* condition); | 95 void Assert(compiler::Node* condition); |
| 94 | 96 |
| (...skipping 380 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 475 AllocationFlags flags, | 477 AllocationFlags flags, |
| 476 compiler::Node* top_adddress, | 478 compiler::Node* top_adddress, |
| 477 compiler::Node* limit_address); | 479 compiler::Node* limit_address); |
| 478 | 480 |
| 479 static const int kElementLoopUnrollThreshold = 8; | 481 static const int kElementLoopUnrollThreshold = 8; |
| 480 }; | 482 }; |
| 481 | 483 |
| 482 } // namespace internal | 484 } // namespace internal |
| 483 } // namespace v8 | 485 } // namespace v8 |
| 484 #endif // V8_CODE_STUB_ASSEMBLER_H_ | 486 #endif // V8_CODE_STUB_ASSEMBLER_H_ |
| OLD | NEW |