OLD | NEW |
1 // Copyright 2016 the V8 project authors. All rights reserved. | 1 // Copyright 2016 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ | 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ |
6 #define V8_CODE_STUB_ASSEMBLER_H_ | 6 #define V8_CODE_STUB_ASSEMBLER_H_ |
7 | 7 |
8 #include <functional> | 8 #include <functional> |
9 | 9 |
10 #include "src/compiler/code-assembler.h" | 10 #include "src/compiler/code-assembler.h" |
(...skipping 103 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
114 compiler::Node* SmiAbove(compiler::Node* a, compiler::Node* b); | 114 compiler::Node* SmiAbove(compiler::Node* a, compiler::Node* b); |
115 compiler::Node* SmiAboveOrEqual(compiler::Node* a, compiler::Node* b); | 115 compiler::Node* SmiAboveOrEqual(compiler::Node* a, compiler::Node* b); |
116 compiler::Node* SmiBelow(compiler::Node* a, compiler::Node* b); | 116 compiler::Node* SmiBelow(compiler::Node* a, compiler::Node* b); |
117 compiler::Node* SmiLessThan(compiler::Node* a, compiler::Node* b); | 117 compiler::Node* SmiLessThan(compiler::Node* a, compiler::Node* b); |
118 compiler::Node* SmiLessThanOrEqual(compiler::Node* a, compiler::Node* b); | 118 compiler::Node* SmiLessThanOrEqual(compiler::Node* a, compiler::Node* b); |
119 compiler::Node* SmiMin(compiler::Node* a, compiler::Node* b); | 119 compiler::Node* SmiMin(compiler::Node* a, compiler::Node* b); |
120 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. | 120 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. |
121 compiler::Node* SmiMod(compiler::Node* a, compiler::Node* b); | 121 compiler::Node* SmiMod(compiler::Node* a, compiler::Node* b); |
122 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. | 122 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. |
123 compiler::Node* SmiMul(compiler::Node* a, compiler::Node* b); | 123 compiler::Node* SmiMul(compiler::Node* a, compiler::Node* b); |
| 124 compiler::Node* SmiOr(compiler::Node* a, compiler::Node* b) { |
| 125 return WordOr(a, b); |
| 126 } |
124 | 127 |
125 // Allocate an object of the given size. | 128 // Allocate an object of the given size. |
126 compiler::Node* Allocate(compiler::Node* size, AllocationFlags flags = kNone); | 129 compiler::Node* Allocate(compiler::Node* size, AllocationFlags flags = kNone); |
127 compiler::Node* Allocate(int size, AllocationFlags flags = kNone); | 130 compiler::Node* Allocate(int size, AllocationFlags flags = kNone); |
128 compiler::Node* InnerAllocate(compiler::Node* previous, int offset); | 131 compiler::Node* InnerAllocate(compiler::Node* previous, int offset); |
129 compiler::Node* InnerAllocate(compiler::Node* previous, | 132 compiler::Node* InnerAllocate(compiler::Node* previous, |
130 compiler::Node* offset); | 133 compiler::Node* offset); |
131 | 134 |
132 void Assert(compiler::Node* condition); | 135 void Assert(compiler::Node* condition); |
133 | 136 |
(...skipping 693 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
827 Label* bailout); | 830 Label* bailout); |
828 | 831 |
829 static const int kElementLoopUnrollThreshold = 8; | 832 static const int kElementLoopUnrollThreshold = 8; |
830 }; | 833 }; |
831 | 834 |
832 DEFINE_OPERATORS_FOR_FLAGS(CodeStubAssembler::AllocationFlags); | 835 DEFINE_OPERATORS_FOR_FLAGS(CodeStubAssembler::AllocationFlags); |
833 | 836 |
834 } // namespace internal | 837 } // namespace internal |
835 } // namespace v8 | 838 } // namespace v8 |
836 #endif // V8_CODE_STUB_ASSEMBLER_H_ | 839 #endif // V8_CODE_STUB_ASSEMBLER_H_ |
OLD | NEW |