| OLD | NEW |
| 1 // Copyright 2013 the V8 project authors. All rights reserved. | 1 // Copyright 2013 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #ifndef V8_ARM64_ASSEMBLER_ARM64_INL_H_ | 5 #ifndef V8_ARM64_ASSEMBLER_ARM64_INL_H_ |
| 6 #define V8_ARM64_ASSEMBLER_ARM64_INL_H_ | 6 #define V8_ARM64_ASSEMBLER_ARM64_INL_H_ |
| 7 | 7 |
| 8 #include "src/arm64/assembler-arm64.h" | 8 #include "src/arm64/assembler-arm64.h" |
| 9 #include "src/assembler.h" | 9 #include "src/assembler.h" |
| 10 #include "src/debug.h" | 10 #include "src/debug.h" |
| (...skipping 412 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 423 } | 423 } |
| 424 | 424 |
| 425 | 425 |
| 426 unsigned Operand::shift_amount() const { | 426 unsigned Operand::shift_amount() const { |
| 427 ASSERT(IsShiftedRegister() || IsExtendedRegister()); | 427 ASSERT(IsShiftedRegister() || IsExtendedRegister()); |
| 428 return shift_amount_; | 428 return shift_amount_; |
| 429 } | 429 } |
| 430 | 430 |
| 431 | 431 |
| 432 Operand Operand::UntagSmi(Register smi) { | 432 Operand Operand::UntagSmi(Register smi) { |
| 433 STATIC_ASSERT(kXRegSizeInBits == static_cast<unsigned>(kSmiShift + |
| 434 kSmiValueSize)); |
| 433 ASSERT(smi.Is64Bits()); | 435 ASSERT(smi.Is64Bits()); |
| 434 return Operand(smi, ASR, kSmiShift); | 436 return Operand(smi, ASR, kSmiShift); |
| 435 } | 437 } |
| 436 | 438 |
| 437 | 439 |
| 438 Operand Operand::UntagSmiAndScale(Register smi, int scale) { | 440 Operand Operand::UntagSmiAndScale(Register smi, int scale) { |
| 441 STATIC_ASSERT(kXRegSizeInBits == static_cast<unsigned>(kSmiShift + |
| 442 kSmiValueSize)); |
| 439 ASSERT(smi.Is64Bits()); | 443 ASSERT(smi.Is64Bits()); |
| 440 ASSERT((scale >= 0) && (scale <= (64 - kSmiValueSize))); | 444 ASSERT((scale >= 0) && (scale <= (64 - kSmiValueSize))); |
| 441 if (scale > kSmiShift) { | 445 if (scale > kSmiShift) { |
| 442 return Operand(smi, LSL, scale - kSmiShift); | 446 return Operand(smi, LSL, scale - kSmiShift); |
| 443 } else if (scale < kSmiShift) { | 447 } else if (scale < kSmiShift) { |
| 444 return Operand(smi, ASR, kSmiShift - scale); | 448 return Operand(smi, ASR, kSmiShift - scale); |
| 445 } | 449 } |
| 446 return Operand(smi); | 450 return Operand(smi); |
| 447 } | 451 } |
| 448 | 452 |
| (...skipping 806 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1255 | 1259 |
| 1256 | 1260 |
| 1257 void Assembler::ClearRecordedAstId() { | 1261 void Assembler::ClearRecordedAstId() { |
| 1258 recorded_ast_id_ = TypeFeedbackId::None(); | 1262 recorded_ast_id_ = TypeFeedbackId::None(); |
| 1259 } | 1263 } |
| 1260 | 1264 |
| 1261 | 1265 |
| 1262 } } // namespace v8::internal | 1266 } } // namespace v8::internal |
| 1263 | 1267 |
| 1264 #endif // V8_ARM64_ASSEMBLER_ARM64_INL_H_ | 1268 #endif // V8_ARM64_ASSEMBLER_ARM64_INL_H_ |
| OLD | NEW |