OLD | NEW |
1 // Copyright 2016 the V8 project authors. All rights reserved. | 1 // Copyright 2016 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ | 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ |
6 #define V8_CODE_STUB_ASSEMBLER_H_ | 6 #define V8_CODE_STUB_ASSEMBLER_H_ |
7 | 7 |
8 #include <functional> | 8 #include <functional> |
9 | 9 |
10 #include "src/compiler/code-assembler.h" | 10 #include "src/compiler/code-assembler.h" |
(...skipping 222 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
233 #undef SMI_COMPARISON_OP | 233 #undef SMI_COMPARISON_OP |
234 Node* SmiMax(Node* a, Node* b); | 234 Node* SmiMax(Node* a, Node* b); |
235 Node* SmiMin(Node* a, Node* b); | 235 Node* SmiMin(Node* a, Node* b); |
236 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. | 236 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. |
237 Node* SmiMod(Node* a, Node* b); | 237 Node* SmiMod(Node* a, Node* b); |
238 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. | 238 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. |
239 Node* SmiMul(Node* a, Node* b); | 239 Node* SmiMul(Node* a, Node* b); |
240 | 240 |
241 // Smi | HeapNumber operations. | 241 // Smi | HeapNumber operations. |
242 Node* NumberInc(Node* value); | 242 Node* NumberInc(Node* value); |
| 243 Node* NumberDec(Node* value); |
243 void GotoIfNotNumber(Node* value, Label* is_not_number); | 244 void GotoIfNotNumber(Node* value, Label* is_not_number); |
244 void GotoIfNumber(Node* value, Label* is_number); | 245 void GotoIfNumber(Node* value, Label* is_number); |
245 | 246 |
246 // Allocate an object of the given size. | 247 // Allocate an object of the given size. |
247 Node* AllocateInNewSpace(Node* size, AllocationFlags flags = kNone); | 248 Node* AllocateInNewSpace(Node* size, AllocationFlags flags = kNone); |
248 Node* AllocateInNewSpace(int size, AllocationFlags flags = kNone); | 249 Node* AllocateInNewSpace(int size, AllocationFlags flags = kNone); |
249 Node* Allocate(Node* size, AllocationFlags flags = kNone); | 250 Node* Allocate(Node* size, AllocationFlags flags = kNone); |
250 Node* Allocate(int size, AllocationFlags flags = kNone); | 251 Node* Allocate(int size, AllocationFlags flags = kNone); |
251 Node* InnerAllocate(Node* previous, int offset); | 252 Node* InnerAllocate(Node* previous, int offset); |
252 Node* InnerAllocate(Node* previous, Node* offset); | 253 Node* InnerAllocate(Node* previous, Node* offset); |
(...skipping 1218 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1471 } | 1472 } |
1472 #else | 1473 #else |
1473 #define CSA_SLOW_ASSERT(csa, x) ((void)0) | 1474 #define CSA_SLOW_ASSERT(csa, x) ((void)0) |
1474 #endif | 1475 #endif |
1475 | 1476 |
1476 DEFINE_OPERATORS_FOR_FLAGS(CodeStubAssembler::AllocationFlags); | 1477 DEFINE_OPERATORS_FOR_FLAGS(CodeStubAssembler::AllocationFlags); |
1477 | 1478 |
1478 } // namespace internal | 1479 } // namespace internal |
1479 } // namespace v8 | 1480 } // namespace v8 |
1480 #endif // V8_CODE_STUB_ASSEMBLER_H_ | 1481 #endif // V8_CODE_STUB_ASSEMBLER_H_ |
OLD | NEW |