OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include <limits.h> // For LONG_MIN, LONG_MAX. | 5 #include <limits.h> // For LONG_MIN, LONG_MAX. |
6 | 6 |
7 #include "src/v8.h" | 7 #include "src/v8.h" |
8 | 8 |
9 #if V8_TARGET_ARCH_ARM | 9 #if V8_TARGET_ARCH_ARM |
10 | 10 |
(...skipping 56 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
67 BlockConstPoolScope block_const_pool(this); | 67 BlockConstPoolScope block_const_pool(this); |
68 Label start; | 68 Label start; |
69 bind(&start); | 69 bind(&start); |
70 blx(target, cond); | 70 blx(target, cond); |
71 ASSERT_EQ(CallSize(target, cond), SizeOfCodeGeneratedSince(&start)); | 71 ASSERT_EQ(CallSize(target, cond), SizeOfCodeGeneratedSince(&start)); |
72 } | 72 } |
73 | 73 |
74 | 74 |
75 int MacroAssembler::CallSize( | 75 int MacroAssembler::CallSize( |
76 Address target, RelocInfo::Mode rmode, Condition cond) { | 76 Address target, RelocInfo::Mode rmode, Condition cond) { |
77 int size = 2 * kInstrSize; | |
78 Instr mov_instr = cond | MOV | LeaveCC; | 77 Instr mov_instr = cond | MOV | LeaveCC; |
79 intptr_t immediate = reinterpret_cast<intptr_t>(target); | 78 Operand mov_operand = Operand(reinterpret_cast<intptr_t>(target), rmode); |
80 if (!Operand(immediate, rmode).is_single_instruction(this, mov_instr)) { | 79 return kInstrSize + |
81 size += kInstrSize; | 80 mov_operand.instructions_required(this, mov_instr) * kInstrSize; |
82 } | |
83 return size; | |
84 } | 81 } |
85 | 82 |
86 | 83 |
87 int MacroAssembler::CallStubSize( | 84 int MacroAssembler::CallStubSize( |
88 CodeStub* stub, TypeFeedbackId ast_id, Condition cond) { | 85 CodeStub* stub, TypeFeedbackId ast_id, Condition cond) { |
89 return CallSize(stub->GetCode(), RelocInfo::CODE_TARGET, ast_id, cond); | 86 return CallSize(stub->GetCode(), RelocInfo::CODE_TARGET, ast_id, cond); |
90 } | 87 } |
91 | 88 |
92 | 89 |
93 int MacroAssembler::CallSizeNotPredictableCodeSize(Isolate* isolate, | 90 int MacroAssembler::CallSizeNotPredictableCodeSize(Isolate* isolate, |
94 Address target, | 91 Address target, |
95 RelocInfo::Mode rmode, | 92 RelocInfo::Mode rmode, |
96 Condition cond) { | 93 Condition cond) { |
97 int size = 2 * kInstrSize; | |
98 Instr mov_instr = cond | MOV | LeaveCC; | 94 Instr mov_instr = cond | MOV | LeaveCC; |
99 intptr_t immediate = reinterpret_cast<intptr_t>(target); | 95 Operand mov_operand = Operand(reinterpret_cast<intptr_t>(target), rmode); |
100 if (!Operand(immediate, rmode).is_single_instruction(NULL, mov_instr)) { | 96 return kInstrSize + |
101 size += kInstrSize; | 97 mov_operand.instructions_required(NULL, mov_instr) * kInstrSize; |
102 } | |
103 return size; | |
104 } | 98 } |
105 | 99 |
106 | 100 |
107 void MacroAssembler::Call(Address target, | 101 void MacroAssembler::Call(Address target, |
108 RelocInfo::Mode rmode, | 102 RelocInfo::Mode rmode, |
109 Condition cond, | 103 Condition cond, |
110 TargetAddressStorageMode mode) { | 104 TargetAddressStorageMode mode) { |
111 // Block constant pool for the call instruction sequence. | 105 // Block constant pool for the call instruction sequence. |
112 BlockConstPoolScope block_const_pool(this); | 106 BlockConstPoolScope block_const_pool(this); |
113 Label start; | 107 Label start; |
(...skipping 152 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
266 } | 260 } |
267 } | 261 } |
268 | 262 |
269 | 263 |
270 void MacroAssembler::And(Register dst, Register src1, const Operand& src2, | 264 void MacroAssembler::And(Register dst, Register src1, const Operand& src2, |
271 Condition cond) { | 265 Condition cond) { |
272 if (!src2.is_reg() && | 266 if (!src2.is_reg() && |
273 !src2.must_output_reloc_info(this) && | 267 !src2.must_output_reloc_info(this) && |
274 src2.immediate() == 0) { | 268 src2.immediate() == 0) { |
275 mov(dst, Operand::Zero(), LeaveCC, cond); | 269 mov(dst, Operand::Zero(), LeaveCC, cond); |
276 } else if (!src2.is_single_instruction(this) && | 270 } else if (!(src2.instructions_required(this) == 1) && |
277 !src2.must_output_reloc_info(this) && | 271 !src2.must_output_reloc_info(this) && |
278 CpuFeatures::IsSupported(ARMv7) && | 272 CpuFeatures::IsSupported(ARMv7) && |
279 IsPowerOf2(src2.immediate() + 1)) { | 273 IsPowerOf2(src2.immediate() + 1)) { |
280 ubfx(dst, src1, 0, | 274 ubfx(dst, src1, 0, |
281 WhichPowerOf2(static_cast<uint32_t>(src2.immediate()) + 1), cond); | 275 WhichPowerOf2(static_cast<uint32_t>(src2.immediate()) + 1), cond); |
282 } else { | 276 } else { |
283 and_(dst, src1, src2, LeaveCC, cond); | 277 and_(dst, src1, src2, LeaveCC, cond); |
284 } | 278 } |
285 } | 279 } |
286 | 280 |
(...skipping 1557 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1844 Condition cond = al; | 1838 Condition cond = al; |
1845 int shift = 0; | 1839 int shift = 0; |
1846 while (object_size != 0) { | 1840 while (object_size != 0) { |
1847 if (((object_size >> shift) & 0x03) == 0) { | 1841 if (((object_size >> shift) & 0x03) == 0) { |
1848 shift += 2; | 1842 shift += 2; |
1849 } else { | 1843 } else { |
1850 int bits = object_size & (0xff << shift); | 1844 int bits = object_size & (0xff << shift); |
1851 object_size -= bits; | 1845 object_size -= bits; |
1852 shift += 8; | 1846 shift += 8; |
1853 Operand bits_operand(bits); | 1847 Operand bits_operand(bits); |
1854 ASSERT(bits_operand.is_single_instruction(this)); | 1848 ASSERT(bits_operand.instructions_required(this) == 1); |
1855 add(scratch2, source, bits_operand, SetCC, cond); | 1849 add(scratch2, source, bits_operand, SetCC, cond); |
1856 source = scratch2; | 1850 source = scratch2; |
1857 cond = cc; | 1851 cond = cc; |
1858 } | 1852 } |
1859 } | 1853 } |
1860 b(cs, gc_required); | 1854 b(cs, gc_required); |
1861 cmp(scratch2, Operand(ip)); | 1855 cmp(scratch2, Operand(ip)); |
1862 b(hi, gc_required); | 1856 b(hi, gc_required); |
1863 str(scratch2, MemOperand(topaddr)); | 1857 str(scratch2, MemOperand(topaddr)); |
1864 | 1858 |
(...skipping 1753 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3618 num_reg_arguments, num_double_arguments); | 3612 num_reg_arguments, num_double_arguments); |
3619 if (ActivationFrameAlignment() > kPointerSize) { | 3613 if (ActivationFrameAlignment() > kPointerSize) { |
3620 ldr(sp, MemOperand(sp, stack_passed_arguments * kPointerSize)); | 3614 ldr(sp, MemOperand(sp, stack_passed_arguments * kPointerSize)); |
3621 } else { | 3615 } else { |
3622 add(sp, sp, Operand(stack_passed_arguments * sizeof(kPointerSize))); | 3616 add(sp, sp, Operand(stack_passed_arguments * sizeof(kPointerSize))); |
3623 } | 3617 } |
3624 } | 3618 } |
3625 | 3619 |
3626 | 3620 |
3627 void MacroAssembler::GetRelocatedValueLocation(Register ldr_location, | 3621 void MacroAssembler::GetRelocatedValueLocation(Register ldr_location, |
3628 Register result) { | 3622 Register result, |
3629 const uint32_t kLdrOffsetMask = (1 << 12) - 1; | 3623 Register scratch) { |
| 3624 Label small_constant_pool_load, load_result; |
3630 ldr(result, MemOperand(ldr_location)); | 3625 ldr(result, MemOperand(ldr_location)); |
| 3626 |
| 3627 if (FLAG_enable_ool_constant_pool) { |
| 3628 // Check if this is an extended constant pool load. |
| 3629 and_(scratch, result, Operand(GetConsantPoolLoadMask())); |
| 3630 teq(scratch, Operand(GetConsantPoolLoadPattern())); |
| 3631 b(eq, &small_constant_pool_load); |
| 3632 if (emit_debug_code()) { |
| 3633 // Check that the instruction sequence is: |
| 3634 // movw reg, #offset_low |
| 3635 // movt reg, #offset_high |
| 3636 // ldr reg, [pp, reg] |
| 3637 Instr patterns[] = { GetMovWPattern(), |
| 3638 GetMovTPattern(), |
| 3639 GetLdrPpRegOffsetPattern() }; |
| 3640 for (int i = 0; i < 3; i++) { |
| 3641 ldr(result, MemOperand(ldr_location, i * kInstrSize)); |
| 3642 and_(result, result, Operand(patterns[i])); |
| 3643 cmp(result, Operand(patterns[i])); |
| 3644 Check(eq, kTheInstructionToPatchShouldBeALoadFromConstantPool); |
| 3645 } |
| 3646 // Result was clobbered. Restore it. |
| 3647 ldr(result, MemOperand(ldr_location)); |
| 3648 } |
| 3649 |
| 3650 // Get the offset into the constant pool. First extract movw immediate into |
| 3651 // result. |
| 3652 and_(scratch, result, Operand(0xfff)); |
| 3653 mov(ip, Operand(result, LSR, 4)); |
| 3654 and_(ip, ip, Operand(0xf000)); |
| 3655 orr(result, scratch, Operand(ip)); |
| 3656 // Then extract movt immediate and or into result. |
| 3657 ldr(scratch, MemOperand(ldr_location, kInstrSize)); |
| 3658 and_(ip, scratch, Operand(0xf0000)); |
| 3659 orr(result, result, Operand(ip, LSL, 12)); |
| 3660 and_(scratch, scratch, Operand(0xfff)); |
| 3661 orr(result, result, Operand(scratch, LSL, 16)); |
| 3662 |
| 3663 b(&load_result); |
| 3664 } |
| 3665 |
| 3666 bind(&small_constant_pool_load); |
3631 if (emit_debug_code()) { | 3667 if (emit_debug_code()) { |
3632 // Check that the instruction is a ldr reg, [<pc or pp> + offset] . | 3668 // Check that the instruction is a ldr reg, [<pc or pp> + offset] . |
3633 and_(result, result, Operand(GetConsantPoolLoadPattern())); | 3669 and_(result, result, Operand(GetConsantPoolLoadPattern())); |
3634 cmp(result, Operand(GetConsantPoolLoadPattern())); | 3670 cmp(result, Operand(GetConsantPoolLoadPattern())); |
3635 Check(eq, kTheInstructionToPatchShouldBeALoadFromConstantPool); | 3671 Check(eq, kTheInstructionToPatchShouldBeALoadFromConstantPool); |
3636 // Result was clobbered. Restore it. | 3672 // Result was clobbered. Restore it. |
3637 ldr(result, MemOperand(ldr_location)); | 3673 ldr(result, MemOperand(ldr_location)); |
3638 } | 3674 } |
| 3675 |
| 3676 // Get the offset into the constant pool. |
| 3677 const uint32_t kLdrOffsetMask = (1 << 12) - 1; |
| 3678 and_(result, result, Operand(kLdrOffsetMask)); |
| 3679 |
| 3680 bind(&load_result); |
3639 // Get the address of the constant. | 3681 // Get the address of the constant. |
3640 and_(result, result, Operand(kLdrOffsetMask)); | |
3641 if (FLAG_enable_ool_constant_pool) { | 3682 if (FLAG_enable_ool_constant_pool) { |
3642 add(result, pp, Operand(result)); | 3683 add(result, pp, Operand(result)); |
3643 } else { | 3684 } else { |
3644 add(result, ldr_location, Operand(result)); | 3685 add(result, ldr_location, Operand(result)); |
3645 add(result, result, Operand(Instruction::kPCReadOffset)); | 3686 add(result, result, Operand(Instruction::kPCReadOffset)); |
3646 } | 3687 } |
3647 } | 3688 } |
3648 | 3689 |
3649 | 3690 |
3650 void MacroAssembler::CheckPageFlag( | 3691 void MacroAssembler::CheckPageFlag( |
(...skipping 431 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4082 sub(result, result, Operand(dividend)); | 4123 sub(result, result, Operand(dividend)); |
4083 } | 4124 } |
4084 if (ms.shift() > 0) mov(result, Operand(result, ASR, ms.shift())); | 4125 if (ms.shift() > 0) mov(result, Operand(result, ASR, ms.shift())); |
4085 add(result, result, Operand(dividend, LSR, 31)); | 4126 add(result, result, Operand(dividend, LSR, 31)); |
4086 } | 4127 } |
4087 | 4128 |
4088 | 4129 |
4089 } } // namespace v8::internal | 4130 } } // namespace v8::internal |
4090 | 4131 |
4091 #endif // V8_TARGET_ARCH_ARM | 4132 #endif // V8_TARGET_ARCH_ARM |
OLD | NEW |