OLD | NEW |
1 // Copyright 2016 the V8 project authors. All rights reserved. | 1 // Copyright 2016 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ | 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ |
6 #define V8_CODE_STUB_ASSEMBLER_H_ | 6 #define V8_CODE_STUB_ASSEMBLER_H_ |
7 | 7 |
8 #include <functional> | 8 #include <functional> |
9 | 9 |
10 #include "src/compiler/code-assembler.h" | 10 #include "src/compiler/code-assembler.h" |
(...skipping 64 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
75 compiler::Node* SmiAddWithOverflow(compiler::Node* a, compiler::Node* b); | 75 compiler::Node* SmiAddWithOverflow(compiler::Node* a, compiler::Node* b); |
76 compiler::Node* SmiSub(compiler::Node* a, compiler::Node* b); | 76 compiler::Node* SmiSub(compiler::Node* a, compiler::Node* b); |
77 compiler::Node* SmiSubWithOverflow(compiler::Node* a, compiler::Node* b); | 77 compiler::Node* SmiSubWithOverflow(compiler::Node* a, compiler::Node* b); |
78 compiler::Node* SmiEqual(compiler::Node* a, compiler::Node* b); | 78 compiler::Node* SmiEqual(compiler::Node* a, compiler::Node* b); |
79 compiler::Node* SmiAboveOrEqual(compiler::Node* a, compiler::Node* b); | 79 compiler::Node* SmiAboveOrEqual(compiler::Node* a, compiler::Node* b); |
80 compiler::Node* SmiLessThan(compiler::Node* a, compiler::Node* b); | 80 compiler::Node* SmiLessThan(compiler::Node* a, compiler::Node* b); |
81 compiler::Node* SmiLessThanOrEqual(compiler::Node* a, compiler::Node* b); | 81 compiler::Node* SmiLessThanOrEqual(compiler::Node* a, compiler::Node* b); |
82 compiler::Node* SmiMin(compiler::Node* a, compiler::Node* b); | 82 compiler::Node* SmiMin(compiler::Node* a, compiler::Node* b); |
83 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. | 83 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. |
84 compiler::Node* SmiMod(compiler::Node* a, compiler::Node* b); | 84 compiler::Node* SmiMod(compiler::Node* a, compiler::Node* b); |
| 85 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. |
| 86 compiler::Node* SmiMul(compiler::Node* a, compiler::Node* b); |
85 | 87 |
86 // Allocate an object of the given size. | 88 // Allocate an object of the given size. |
87 compiler::Node* Allocate(compiler::Node* size, AllocationFlags flags = kNone); | 89 compiler::Node* Allocate(compiler::Node* size, AllocationFlags flags = kNone); |
88 compiler::Node* Allocate(int size, AllocationFlags flags = kNone); | 90 compiler::Node* Allocate(int size, AllocationFlags flags = kNone); |
89 compiler::Node* InnerAllocate(compiler::Node* previous, int offset); | 91 compiler::Node* InnerAllocate(compiler::Node* previous, int offset); |
90 compiler::Node* InnerAllocate(compiler::Node* previous, | 92 compiler::Node* InnerAllocate(compiler::Node* previous, |
91 compiler::Node* offset); | 93 compiler::Node* offset); |
92 | 94 |
93 void Assert(compiler::Node* condition); | 95 void Assert(compiler::Node* condition); |
94 | 96 |
(...skipping 380 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
475 AllocationFlags flags, | 477 AllocationFlags flags, |
476 compiler::Node* top_adddress, | 478 compiler::Node* top_adddress, |
477 compiler::Node* limit_address); | 479 compiler::Node* limit_address); |
478 | 480 |
479 static const int kElementLoopUnrollThreshold = 8; | 481 static const int kElementLoopUnrollThreshold = 8; |
480 }; | 482 }; |
481 | 483 |
482 } // namespace internal | 484 } // namespace internal |
483 } // namespace v8 | 485 } // namespace v8 |
484 #endif // V8_CODE_STUB_ASSEMBLER_H_ | 486 #endif // V8_CODE_STUB_ASSEMBLER_H_ |
OLD | NEW |