| OLD | NEW |
| 1 // Copyright 2016 the V8 project authors. All rights reserved. | 1 // Copyright 2016 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ | 5 #ifndef V8_CODE_STUB_ASSEMBLER_H_ |
| 6 #define V8_CODE_STUB_ASSEMBLER_H_ | 6 #define V8_CODE_STUB_ASSEMBLER_H_ |
| 7 | 7 |
| 8 #include <functional> | 8 #include <functional> |
| 9 | 9 |
| 10 #include "src/compiler/code-assembler.h" | 10 #include "src/compiler/code-assembler.h" |
| (...skipping 222 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 233 #undef SMI_COMPARISON_OP | 233 #undef SMI_COMPARISON_OP |
| 234 Node* SmiMax(Node* a, Node* b); | 234 Node* SmiMax(Node* a, Node* b); |
| 235 Node* SmiMin(Node* a, Node* b); | 235 Node* SmiMin(Node* a, Node* b); |
| 236 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. | 236 // Computes a % b for Smi inputs a and b; result is not necessarily a Smi. |
| 237 Node* SmiMod(Node* a, Node* b); | 237 Node* SmiMod(Node* a, Node* b); |
| 238 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. | 238 // Computes a * b for Smi inputs a and b; result is not necessarily a Smi. |
| 239 Node* SmiMul(Node* a, Node* b); | 239 Node* SmiMul(Node* a, Node* b); |
| 240 | 240 |
| 241 // Smi | HeapNumber operations. | 241 // Smi | HeapNumber operations. |
| 242 Node* NumberInc(Node* value); | 242 Node* NumberInc(Node* value); |
| 243 Node* NumberDec(Node* value); |
| 243 void GotoIfNotNumber(Node* value, Label* is_not_number); | 244 void GotoIfNotNumber(Node* value, Label* is_not_number); |
| 244 void GotoIfNumber(Node* value, Label* is_number); | 245 void GotoIfNumber(Node* value, Label* is_number); |
| 245 | 246 |
| 246 // Allocate an object of the given size. | 247 // Allocate an object of the given size. |
| 247 Node* AllocateInNewSpace(Node* size, AllocationFlags flags = kNone); | 248 Node* AllocateInNewSpace(Node* size, AllocationFlags flags = kNone); |
| 248 Node* AllocateInNewSpace(int size, AllocationFlags flags = kNone); | 249 Node* AllocateInNewSpace(int size, AllocationFlags flags = kNone); |
| 249 Node* Allocate(Node* size, AllocationFlags flags = kNone); | 250 Node* Allocate(Node* size, AllocationFlags flags = kNone); |
| 250 Node* Allocate(int size, AllocationFlags flags = kNone); | 251 Node* Allocate(int size, AllocationFlags flags = kNone); |
| 251 Node* InnerAllocate(Node* previous, int offset); | 252 Node* InnerAllocate(Node* previous, int offset); |
| 252 Node* InnerAllocate(Node* previous, Node* offset); | 253 Node* InnerAllocate(Node* previous, Node* offset); |
| (...skipping 1218 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1471 } | 1472 } |
| 1472 #else | 1473 #else |
| 1473 #define CSA_SLOW_ASSERT(csa, x) ((void)0) | 1474 #define CSA_SLOW_ASSERT(csa, x) ((void)0) |
| 1474 #endif | 1475 #endif |
| 1475 | 1476 |
| 1476 DEFINE_OPERATORS_FOR_FLAGS(CodeStubAssembler::AllocationFlags); | 1477 DEFINE_OPERATORS_FOR_FLAGS(CodeStubAssembler::AllocationFlags); |
| 1477 | 1478 |
| 1478 } // namespace internal | 1479 } // namespace internal |
| 1479 } // namespace v8 | 1480 } // namespace v8 |
| 1480 #endif // V8_CODE_STUB_ASSEMBLER_H_ | 1481 #endif // V8_CODE_STUB_ASSEMBLER_H_ |
| OLD | NEW |