Chromium Code Reviews| Index: src/a64/macro-assembler-a64.cc |
| diff --git a/src/a64/macro-assembler-a64.cc b/src/a64/macro-assembler-a64.cc |
| index 5b979c2b31ed496463018ce49d4c69ac73ceabb2..d7aa72cef00427a7696cf20fd8ebd592db06c1ff 100644 |
| --- a/src/a64/macro-assembler-a64.cc |
| +++ b/src/a64/macro-assembler-a64.cc |
| @@ -53,7 +53,7 @@ MacroAssembler::MacroAssembler(Isolate* arg_isolate, |
| #endif |
| has_frame_(false), |
| use_real_aborts_(true), |
| - sp_(jssp), tmp0_(ip0), tmp1_(ip1), fptmp0_(fp_scratch) { |
| + sp_(jssp), tmp_list_(ip0, ip1), fptmp_list_(fp_scratch) { |
| if (isolate() != NULL) { |
| code_object_ = Handle<Object>(isolate()->heap()->undefined_value(), |
| isolate()); |
| @@ -65,9 +65,12 @@ void MacroAssembler::LogicalMacro(const Register& rd, |
| const Register& rn, |
| const Operand& operand, |
| LogicalOp op) { |
| + UseScratchRegisterScope temps(this); |
| + |
| if (operand.NeedsRelocation()) { |
| - LoadRelocated(Tmp0(), operand); |
| - Logical(rd, rn, Tmp0(), op); |
| + Register temp = temps.AcquireX(); |
| + LoadRelocated(temp, operand); |
| + Logical(rd, rn, temp, op); |
| } else if (operand.IsImmediate()) { |
| int64_t immediate = operand.immediate(); |
| @@ -125,7 +128,7 @@ void MacroAssembler::LogicalMacro(const Register& rd, |
| LogicalImmediate(rd, rn, n, imm_s, imm_r, op); |
| } else { |
| // Immediate can't be encoded: synthesize using move immediate. |
| - Register temp = AppropriateTempFor(rn); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| Mov(temp, immediate); |
| if (rd.Is(csp)) { |
| // If rd is the stack pointer we cannot use it as the destination |
| @@ -144,7 +147,7 @@ void MacroAssembler::LogicalMacro(const Register& rd, |
| ASSERT(operand.shift_amount() <= 4); |
| ASSERT(operand.reg().Is64Bits() || |
| ((operand.extend() != UXTX) && (operand.extend() != SXTX))); |
| - Register temp = AppropriateTempFor(rn, operand.reg()); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| EmitExtendShift(temp, operand.reg(), operand.extend(), |
| operand.shift_amount()); |
| Logical(rd, rn, temp, op); |
| @@ -208,9 +211,10 @@ void MacroAssembler::Mov(const Register& rd, uint64_t imm) { |
| invert_move = true; |
| } |
| - // Mov instructions can't move value into the stack pointer, so set up a |
| - // temporary register, if needed. |
| - Register temp = rd.IsSP() ? AppropriateTempFor(rd) : rd; |
| + // Mov instructions can't move immediate values into the stack pointer, so |
| + // set up a temporary register, if needed. |
| + UseScratchRegisterScope temps(this); |
| + Register temp = rd.IsSP() ? temps.AcquireSameSizeAs(rd) : rd; |
| // Iterate through the halfwords. Use movn/movz for the first non-ignored |
| // halfword, and movk for subsequent halfwords. |
| @@ -248,9 +252,11 @@ void MacroAssembler::Mov(const Register& rd, |
| DiscardMoveMode discard_mode) { |
| ASSERT(allow_macro_instructions_); |
| ASSERT(!rd.IsZero()); |
| + |
| // Provide a swap register for instructions that need to write into the |
| // system stack pointer (and can't do this inherently). |
| - Register dst = (rd.Is(csp)) ? (Tmp1()) : (rd); |
| + UseScratchRegisterScope temps(this); |
| + Register dst = (rd.IsSP()) ? temps.AcquireSameSizeAs(rd) : rd; |
| if (operand.NeedsRelocation()) { |
| LoadRelocated(dst, operand); |
| @@ -291,8 +297,7 @@ void MacroAssembler::Mov(const Register& rd, |
| // Copy the result to the system stack pointer. |
| if (!dst.Is(rd)) { |
| - ASSERT(rd.IsZero()); |
| - ASSERT(dst.Is(Tmp1())); |
| + ASSERT(rd.IsSP()); |
| Assembler::mov(rd, dst); |
| } |
| } |
| @@ -302,8 +307,8 @@ void MacroAssembler::Mvn(const Register& rd, const Operand& operand) { |
| ASSERT(allow_macro_instructions_); |
| if (operand.NeedsRelocation()) { |
| - LoadRelocated(Tmp0(), operand); |
| - Mvn(rd, Tmp0()); |
| + LoadRelocated(rd, operand); |
| + mvn(rd, rd); |
| } else if (operand.IsImmediate()) { |
| // Call the macro assembler for generic immediates. |
| @@ -312,14 +317,11 @@ void MacroAssembler::Mvn(const Register& rd, const Operand& operand) { |
| } else if (operand.IsExtendedRegister()) { |
| // Emit two instructions for the extend case. This differs from Mov, as |
| // the extend and invert can't be achieved in one instruction. |
| - Register temp = AppropriateTempFor(rd, operand.reg()); |
| - EmitExtendShift(temp, operand.reg(), operand.extend(), |
| + EmitExtendShift(rd, operand.reg(), operand.extend(), |
| operand.shift_amount()); |
| - mvn(rd, temp); |
| + mvn(rd, rd); |
| } else { |
| - // Otherwise, emit a register move only if the registers are distinct. |
| - // If the jssp is an operand, add #0 is emitted, otherwise, orr #0. |
| mvn(rd, operand); |
| } |
| } |
| @@ -360,8 +362,10 @@ void MacroAssembler::ConditionalCompareMacro(const Register& rn, |
| ConditionalCompareOp op) { |
| ASSERT((cond != al) && (cond != nv)); |
| if (operand.NeedsRelocation()) { |
| - LoadRelocated(Tmp0(), operand); |
| - ConditionalCompareMacro(rn, Tmp0(), nzcv, cond, op); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + LoadRelocated(temp, operand); |
| + ConditionalCompareMacro(rn, temp, nzcv, cond, op); |
| } else if ((operand.IsShiftedRegister() && (operand.shift_amount() == 0)) || |
| (operand.IsImmediate() && IsImmConditionalCompare(operand.immediate()))) { |
| @@ -372,7 +376,8 @@ void MacroAssembler::ConditionalCompareMacro(const Register& rn, |
| } else { |
| // The operand isn't directly supported by the instruction: perform the |
| // operation on a temporary register. |
| - Register temp = AppropriateTempFor(rn); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| Mov(temp, operand); |
| ConditionalCompare(rn, temp, nzcv, cond, op); |
| } |
| @@ -398,7 +403,8 @@ void MacroAssembler::Csel(const Register& rd, |
| } else if (imm == -1) { |
| csinv(rd, rn, zr, cond); |
| } else { |
| - Register temp = AppropriateTempFor(rn); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| Mov(temp, operand.immediate()); |
| csel(rd, rn, temp, cond); |
| } |
| @@ -407,7 +413,8 @@ void MacroAssembler::Csel(const Register& rd, |
| csel(rd, rn, operand.reg(), cond); |
| } else { |
| // All other arguments. |
| - Register temp = AppropriateTempFor(rn); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| Mov(temp, operand); |
| csel(rd, rn, temp, cond); |
| } |
| @@ -426,12 +433,15 @@ void MacroAssembler::AddSubMacro(const Register& rd, |
| } |
| if (operand.NeedsRelocation()) { |
| - LoadRelocated(Tmp0(), operand); |
| - AddSubMacro(rd, rn, Tmp0(), S, op); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + LoadRelocated(temp, operand); |
| + AddSubMacro(rd, rn, temp, S, op); |
| } else if ((operand.IsImmediate() && !IsImmAddSub(operand.immediate())) || |
| (rn.IsZero() && !operand.IsShiftedRegister()) || |
| (operand.IsShiftedRegister() && (operand.shift() == ROR))) { |
| - Register temp = AppropriateTempFor(rn); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| Mov(temp, operand); |
| AddSub(rd, rn, temp, S, op); |
| } else { |
| @@ -446,24 +456,28 @@ void MacroAssembler::AddSubWithCarryMacro(const Register& rd, |
| FlagsUpdate S, |
| AddSubWithCarryOp op) { |
| ASSERT(rd.SizeInBits() == rn.SizeInBits()); |
| + UseScratchRegisterScope temps(this); |
| if (operand.NeedsRelocation()) { |
| - LoadRelocated(Tmp0(), operand); |
| - AddSubWithCarryMacro(rd, rn, Tmp0(), S, op); |
| + Register temp = temps.AcquireX(); |
| + LoadRelocated(temp, operand); |
| + AddSubWithCarryMacro(rd, rn, temp, S, op); |
| } else if (operand.IsImmediate() || |
| (operand.IsShiftedRegister() && (operand.shift() == ROR))) { |
| // Add/sub with carry (immediate or ROR shifted register.) |
| - Register temp = AppropriateTempFor(rn); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| Mov(temp, operand); |
| AddSubWithCarry(rd, rn, temp, S, op); |
| + |
| } else if (operand.IsShiftedRegister() && (operand.shift_amount() != 0)) { |
| // Add/sub with carry (shifted register). |
| ASSERT(operand.reg().SizeInBits() == rd.SizeInBits()); |
| ASSERT(operand.shift() != ROR); |
| - ASSERT(is_uintn(operand.shift_amount(), |
| - rd.SizeInBits() == kXRegSize ? kXRegSizeLog2 : kWRegSizeLog2)); |
| - Register temp = AppropriateTempFor(rn, operand.reg()); |
| + ASSERT( |
| + is_uintn(operand.shift_amount(), |
| + rd.SizeInBits() == kXRegSize ? kXRegSizeLog2 : kWRegSizeLog2)); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| EmitShift(temp, operand.reg(), operand.shift(), operand.shift_amount()); |
| AddSubWithCarry(rd, rn, temp, S, op); |
| @@ -475,7 +489,7 @@ void MacroAssembler::AddSubWithCarryMacro(const Register& rd, |
| ASSERT(operand.shift_amount() <= 4); |
| ASSERT(operand.reg().Is64Bits() || |
| ((operand.extend() != UXTX) && (operand.extend() != SXTX))); |
| - Register temp = AppropriateTempFor(rn, operand.reg()); |
| + Register temp = temps.AcquireSameSizeAs(rn); |
| EmitExtendShift(temp, operand.reg(), operand.extend(), |
| operand.shift_amount()); |
| AddSubWithCarry(rd, rn, temp, S, op); |
| @@ -500,7 +514,8 @@ void MacroAssembler::LoadStoreMacro(const CPURegister& rt, |
| !IsImmLSUnscaled(offset)) { |
| // Immediate offset that can't be encoded using unsigned or unscaled |
| // addressing modes. |
| - Register temp = AppropriateTempFor(addr.base()); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireSameSizeAs(addr.base()); |
| Mov(temp, addr.offset()); |
| LoadStore(rt, MemOperand(addr.base(), temp), op); |
| } else if (addr.IsPostIndex() && !IsImmLSUnscaled(offset)) { |
| @@ -943,11 +958,14 @@ void MacroAssembler::PushMultipleTimes(CPURegister src, int count) { |
| PrepareForPush(count, size); |
| if (FLAG_optimize_for_size && count > 8) { |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + |
| Label loop; |
| - __ Mov(Tmp0(), count / 2); |
| + __ Mov(temp, count / 2); |
| __ Bind(&loop); |
| PushHelper(2, size, src, src, NoReg, NoReg); |
| - __ Subs(Tmp0(), Tmp0(), 1); |
| + __ Subs(temp, temp, 1); |
| __ B(ne, &loop); |
| count %= 2; |
| @@ -975,7 +993,8 @@ void MacroAssembler::PushMultipleTimes(CPURegister src, int count) { |
| void MacroAssembler::PushMultipleTimes(CPURegister src, Register count) { |
| PrepareForPush(Operand(count, UXTW, WhichPowerOf2(src.SizeInBytes()))); |
| - Register temp = AppropriateTempFor(count); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireSameSizeAs(count); |
| if (FLAG_optimize_for_size) { |
| Label loop, done; |
| @@ -1419,10 +1438,10 @@ void MacroAssembler::InNewSpace(Register object, |
| Condition cond, |
| Label* branch) { |
| ASSERT(cond == eq || cond == ne); |
| - // Use Tmp1() to have a different destination register, as Tmp0() will be used |
| - // for relocation. |
| - And(Tmp1(), object, Operand(ExternalReference::new_space_mask(isolate()))); |
| - Cmp(Tmp1(), Operand(ExternalReference::new_space_start(isolate()))); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + And(temp, object, Operand(ExternalReference::new_space_mask(isolate()))); |
| + Cmp(temp, Operand(ExternalReference::new_space_start(isolate()))); |
| B(cond, branch); |
| } |
| @@ -1591,8 +1610,11 @@ void MacroAssembler::AssertName(Register object) { |
| Abort(kOperandIsASmiAndNotAName); |
| Bind(¬_smi); |
| - Ldr(Tmp1(), FieldMemOperand(object, HeapObject::kMapOffset)); |
| - CompareInstanceType(Tmp1(), Tmp1(), LAST_NAME_TYPE); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + |
| + Ldr(temp, FieldMemOperand(object, HeapObject::kMapOffset)); |
| + CompareInstanceType(temp, temp, LAST_NAME_TYPE); |
| Check(ls, kOperandIsNotAName); |
| } |
| } |
| @@ -1600,7 +1622,8 @@ void MacroAssembler::AssertName(Register object) { |
| void MacroAssembler::AssertString(Register object) { |
| if (emit_debug_code()) { |
| - Register temp = Tmp1(); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| STATIC_ASSERT(kSmiTag == 0); |
| Tst(object, kSmiTagMask); |
| Check(ne, kOperandIsASmiAndNotAString); |
| @@ -1917,8 +1940,10 @@ void MacroAssembler::CallCFunction(ExternalReference function, |
| void MacroAssembler::CallCFunction(ExternalReference function, |
| int num_of_reg_args, |
| int num_of_double_args) { |
| - Mov(Tmp0(), Operand(function)); |
| - CallCFunction(Tmp0(), num_of_reg_args, num_of_double_args); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + Mov(temp, Operand(function)); |
| + CallCFunction(temp, num_of_reg_args, num_of_double_args); |
| } |
| @@ -1971,7 +1996,8 @@ void MacroAssembler::CallCFunction(Register function, |
| // Because the stack pointer must be aligned on a 16-byte boundary, the |
| // aligned csp can be up to 12 bytes below the jssp. This is the case |
| // where we only pushed one W register on top of an aligned jssp. |
| - Register temp = Tmp1(); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| ASSERT(ActivationFrameAlignment() == 16); |
| Sub(temp, csp, old_stack_pointer); |
| // We want temp <= 0 && temp >= -12. |
| @@ -1990,8 +2016,10 @@ void MacroAssembler::Jump(Register target) { |
| void MacroAssembler::Jump(intptr_t target, RelocInfo::Mode rmode) { |
| - Mov(Tmp0(), Operand(target, rmode)); |
| - Br(Tmp0()); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + Mov(temp, Operand(target, rmode)); |
| + Br(temp); |
| } |
| @@ -2053,16 +2081,19 @@ void MacroAssembler::Call(Address target, RelocInfo::Mode rmode) { |
| // Addresses always have 64 bits, so we shouldn't encounter NONE32. |
| ASSERT(rmode != RelocInfo::NONE32); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + |
| if (rmode == RelocInfo::NONE64) { |
| uint64_t imm = reinterpret_cast<uint64_t>(target); |
| - movz(Tmp0(), (imm >> 0) & 0xffff, 0); |
| - movk(Tmp0(), (imm >> 16) & 0xffff, 16); |
| - movk(Tmp0(), (imm >> 32) & 0xffff, 32); |
| - movk(Tmp0(), (imm >> 48) & 0xffff, 48); |
| + movz(temp, (imm >> 0) & 0xffff, 0); |
| + movk(temp, (imm >> 16) & 0xffff, 16); |
| + movk(temp, (imm >> 32) & 0xffff, 32); |
| + movk(temp, (imm >> 48) & 0xffff, 48); |
| } else { |
| - LoadRelocated(Tmp0(), Operand(reinterpret_cast<intptr_t>(target), rmode)); |
| + LoadRelocated(temp, Operand(reinterpret_cast<intptr_t>(target), rmode)); |
| } |
| - Blr(Tmp0()); |
| + Blr(temp); |
| #ifdef DEBUG |
| AssertSizeOfCodeGeneratedSince(&start_call, CallSize(target, rmode)); |
| #endif |
| @@ -2143,21 +2174,23 @@ void MacroAssembler::JumpForHeapNumber(Register object, |
| Label* on_heap_number, |
| Label* on_not_heap_number) { |
| ASSERT(on_heap_number || on_not_heap_number); |
| - // Tmp0() is used as a scratch register. |
| - ASSERT(!AreAliased(Tmp0(), heap_number_map)); |
| AssertNotSmi(object); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + |
| // Load the HeapNumber map if it is not passed. |
| if (heap_number_map.Is(NoReg)) { |
| - heap_number_map = Tmp1(); |
| + heap_number_map = temps.AcquireX(); |
| LoadRoot(heap_number_map, Heap::kHeapNumberMapRootIndex); |
| } else { |
| - // This assert clobbers Tmp0(), so do it before loading Tmp0() with the map. |
| AssertRegisterIsRoot(heap_number_map, Heap::kHeapNumberMapRootIndex); |
| } |
| - Ldr(Tmp0(), FieldMemOperand(object, HeapObject::kMapOffset)); |
| - Cmp(Tmp0(), heap_number_map); |
| + ASSERT(!AreAliased(temp, heap_number_map)); |
| + |
| + Ldr(temp, FieldMemOperand(object, HeapObject::kMapOffset)); |
| + Cmp(temp, heap_number_map); |
| if (on_heap_number) { |
| B(eq, on_heap_number); |
| @@ -2283,10 +2316,12 @@ void MacroAssembler::TryConvertDoubleToInt(Register as_int, |
| void MacroAssembler::JumpIfMinusZero(DoubleRegister input, |
| Label* on_negative_zero) { |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| // Floating point -0.0 is kMinInt as an integer, so subtracting 1 (cmp) will |
| // cause overflow. |
| - Fmov(Tmp0(), input); |
| - Cmp(Tmp0(), 1); |
| + Fmov(temp, input); |
| + Cmp(temp, 1); |
| B(vs, on_negative_zero); |
| } |
| @@ -2296,10 +2331,8 @@ void MacroAssembler::ClampInt32ToUint8(Register output, Register input) { |
| Cmp(input.W(), Operand(input.W(), UXTB)); |
| // If input < input & 0xff, it must be < 0, so saturate to 0. |
| Csel(output.W(), wzr, input.W(), lt); |
| - // Create a constant 0xff. |
| - Mov(WTmp0(), 255); |
| - // If input > input & 0xff, it must be > 255, so saturate to 255. |
| - Csel(output.W(), WTmp0(), output.W(), gt); |
| + // If input <= input & 0xff, it must be <= 255. Otherwise, saturate to 255. |
| + Csel(output.W(), output.W(), 255, le); |
| } |
| @@ -2333,37 +2366,37 @@ void MacroAssembler::CopyFieldsLoopPairsHelper(Register dst, |
| unsigned count, |
| Register scratch1, |
| Register scratch2, |
| - Register scratch3) { |
| + Register scratch3, |
| + Register scratch4, |
| + Register scratch5) { |
| // Untag src and dst into scratch registers. |
| // Copy src->dst in a tight loop. |
| - ASSERT(!AreAliased(dst, src, scratch1, scratch2, scratch3, Tmp0(), Tmp1())); |
| + ASSERT(!AreAliased(dst, src, |
| + scratch1, scratch2, scratch3, scratch4, scratch5)); |
| ASSERT(count >= 2); |
| const Register& remaining = scratch3; |
| Mov(remaining, count / 2); |
| - // Only use the Assembler, so we can use Tmp0() and Tmp1(). |
| - InstructionAccurateScope scope(this); |
| - |
| const Register& dst_untagged = scratch1; |
| const Register& src_untagged = scratch2; |
| - sub(dst_untagged, dst, kHeapObjectTag); |
| - sub(src_untagged, src, kHeapObjectTag); |
| + Sub(dst_untagged, dst, kHeapObjectTag); |
| + Sub(src_untagged, src, kHeapObjectTag); |
| // Copy fields in pairs. |
| Label loop; |
| - bind(&loop); |
| - ldp(Tmp0(), Tmp1(), MemOperand(src_untagged, kXRegSizeInBytes * 2, |
| - PostIndex)); |
| - stp(Tmp0(), Tmp1(), MemOperand(dst_untagged, kXRegSizeInBytes * 2, |
| - PostIndex)); |
| - sub(remaining, remaining, 1); |
| - cbnz(remaining, &loop); |
| + Bind(&loop); |
| + Ldp(scratch4, scratch5, |
| + MemOperand(src_untagged, kXRegSizeInBytes * 2, PostIndex)); |
| + Stp(scratch4, scratch5, |
| + MemOperand(dst_untagged, kXRegSizeInBytes * 2, PostIndex)); |
| + Sub(remaining, remaining, 1); |
| + Cbnz(remaining, &loop); |
| // Handle the leftovers. |
| if (count & 1) { |
| - ldr(Tmp0(), MemOperand(src_untagged)); |
| - str(Tmp0(), MemOperand(dst_untagged)); |
| + Ldr(scratch4, MemOperand(src_untagged)); |
| + Str(scratch4, MemOperand(dst_untagged)); |
| } |
| } |
| @@ -2372,13 +2405,12 @@ void MacroAssembler::CopyFieldsUnrolledPairsHelper(Register dst, |
| Register src, |
| unsigned count, |
| Register scratch1, |
| - Register scratch2) { |
| + Register scratch2, |
| + Register scratch3, |
| + Register scratch4) { |
| // Untag src and dst into scratch registers. |
| // Copy src->dst in an unrolled loop. |
| - ASSERT(!AreAliased(dst, src, scratch1, scratch2, Tmp0(), Tmp1())); |
| - |
| - // Only use the Assembler, so we can use Tmp0() and Tmp1(). |
| - InstructionAccurateScope scope(this); |
| + ASSERT(!AreAliased(dst, src, scratch1, scratch2, scratch3, scratch4)); |
| const Register& dst_untagged = scratch1; |
| const Register& src_untagged = scratch2; |
| @@ -2387,16 +2419,16 @@ void MacroAssembler::CopyFieldsUnrolledPairsHelper(Register dst, |
| // Copy fields in pairs. |
| for (unsigned i = 0; i < count / 2; i++) { |
| - ldp(Tmp0(), Tmp1(), MemOperand(src_untagged, kXRegSizeInBytes * 2, |
| - PostIndex)); |
| - stp(Tmp0(), Tmp1(), MemOperand(dst_untagged, kXRegSizeInBytes * 2, |
| - PostIndex)); |
| + Ldp(scratch3, scratch4, |
| + MemOperand(src_untagged, kXRegSizeInBytes * 2, PostIndex)); |
| + Stp(scratch3, scratch4, |
| + MemOperand(dst_untagged, kXRegSizeInBytes * 2, PostIndex)); |
| } |
| // Handle the leftovers. |
| if (count & 1) { |
| - ldr(Tmp0(), MemOperand(src_untagged)); |
| - str(Tmp0(), MemOperand(dst_untagged)); |
| + Ldr(scratch3, MemOperand(src_untagged)); |
| + Str(scratch3, MemOperand(dst_untagged)); |
| } |
| } |
| @@ -2404,23 +2436,22 @@ void MacroAssembler::CopyFieldsUnrolledPairsHelper(Register dst, |
| void MacroAssembler::CopyFieldsUnrolledHelper(Register dst, |
| Register src, |
| unsigned count, |
| - Register scratch1) { |
| + Register scratch1, |
| + Register scratch2, |
| + Register scratch3) { |
| // Untag src and dst into scratch registers. |
| // Copy src->dst in an unrolled loop. |
| - ASSERT(!AreAliased(dst, src, scratch1, Tmp0(), Tmp1())); |
| - |
| - // Only use the Assembler, so we can use Tmp0() and Tmp1(). |
| - InstructionAccurateScope scope(this); |
| + ASSERT(!AreAliased(dst, src, scratch1, scratch2, scratch3)); |
| const Register& dst_untagged = scratch1; |
| - const Register& src_untagged = Tmp1(); |
| - sub(dst_untagged, dst, kHeapObjectTag); |
| - sub(src_untagged, src, kHeapObjectTag); |
| + const Register& src_untagged = scratch2; |
| + Sub(dst_untagged, dst, kHeapObjectTag); |
| + Sub(src_untagged, src, kHeapObjectTag); |
| // Copy fields one by one. |
| for (unsigned i = 0; i < count; i++) { |
| - ldr(Tmp0(), MemOperand(src_untagged, kXRegSizeInBytes, PostIndex)); |
| - str(Tmp0(), MemOperand(dst_untagged, kXRegSizeInBytes, PostIndex)); |
| + Ldr(scratch3, MemOperand(src_untagged, kXRegSizeInBytes, PostIndex)); |
| + Str(scratch3, MemOperand(dst_untagged, kXRegSizeInBytes, PostIndex)); |
| } |
| } |
| @@ -2439,12 +2470,10 @@ void MacroAssembler::CopyFields(Register dst, Register src, CPURegList temps, |
| // |
| // In both cases, fields are copied in pairs if possible, and left-overs are |
| // handled separately. |
| + ASSERT(!AreAliased(dst, src)); |
| ASSERT(!temps.IncludesAliasOf(dst)); |
| ASSERT(!temps.IncludesAliasOf(src)); |
| - ASSERT(!temps.IncludesAliasOf(Tmp0())); |
| - ASSERT(!temps.IncludesAliasOf(Tmp1())); |
| ASSERT(!temps.IncludesAliasOf(xzr)); |
| - ASSERT(!AreAliased(dst, src, Tmp0(), Tmp1())); |
| if (emit_debug_code()) { |
| Cmp(dst, src); |
| @@ -2455,17 +2484,25 @@ void MacroAssembler::CopyFields(Register dst, Register src, CPURegList temps, |
| // enough scratch registers). |
| static const unsigned kLoopThreshold = 8; |
| - ASSERT(!temps.IsEmpty()); |
| - Register scratch1 = Register(temps.PopLowestIndex()); |
| - Register scratch2 = Register(temps.PopLowestIndex()); |
| - Register scratch3 = Register(temps.PopLowestIndex()); |
| - |
| - if (scratch3.IsValid() && (count >= kLoopThreshold)) { |
| - CopyFieldsLoopPairsHelper(dst, src, count, scratch1, scratch2, scratch3); |
| - } else if (scratch2.IsValid()) { |
| - CopyFieldsUnrolledPairsHelper(dst, src, count, scratch1, scratch2); |
| - } else if (scratch1.IsValid()) { |
| - CopyFieldsUnrolledHelper(dst, src, count, scratch1); |
| + UseScratchRegisterScope masm_temps(this); |
| + if ((temps.Count() >= 3) && (count >= kLoopThreshold)) { |
| + CopyFieldsLoopPairsHelper(dst, src, count, |
| + Register(temps.PopLowestIndex()), |
| + Register(temps.PopLowestIndex()), |
| + Register(temps.PopLowestIndex()), |
| + masm_temps.AcquireX(), |
| + masm_temps.AcquireX()); |
| + } else if (temps.Count() >= 2) { |
| + CopyFieldsUnrolledPairsHelper(dst, src, count, |
| + Register(temps.PopLowestIndex()), |
| + Register(temps.PopLowestIndex()), |
| + masm_temps.AcquireX(), |
| + masm_temps.AcquireX()); |
| + } else if (temps.Count() == 1) { |
| + CopyFieldsUnrolledHelper(dst, src, count, |
| + Register(temps.PopLowestIndex()), |
| + masm_temps.AcquireX(), |
| + masm_temps.AcquireX()); |
| } else { |
| UNREACHABLE(); |
| } |
| @@ -2895,12 +2932,14 @@ void MacroAssembler::TruncateHeapNumberToI(Register result, |
| void MacroAssembler::Prologue(PrologueFrameMode frame_mode) { |
| if (frame_mode == BUILD_STUB_FRAME) { |
| ASSERT(StackPointer().Is(jssp)); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| // TODO(jbramley): Does x1 contain a JSFunction here, or does it already |
| // have the special STUB smi? |
| - __ Mov(Tmp0(), Operand(Smi::FromInt(StackFrame::STUB))); |
| + __ Mov(temp, Operand(Smi::FromInt(StackFrame::STUB))); |
| // Compiled stubs don't age, and so they don't need the predictable code |
| // ageing sequence. |
| - __ Push(lr, fp, cp, Tmp0()); |
| + __ Push(lr, fp, cp, temp); |
| __ Add(fp, jssp, StandardFrameConstants::kFixedFrameSizeFromFp); |
| } else { |
| if (isolate()->IsCodePreAgingActive()) { |
| @@ -2915,10 +2954,14 @@ void MacroAssembler::Prologue(PrologueFrameMode frame_mode) { |
| void MacroAssembler::EnterFrame(StackFrame::Type type) { |
| ASSERT(jssp.Is(StackPointer())); |
| + UseScratchRegisterScope temps(this); |
| + Register type_reg = temps.AcquireX(); |
| + Register code_reg = temps.AcquireX(); |
| + |
| Push(lr, fp, cp); |
| - Mov(Tmp1(), Operand(Smi::FromInt(type))); |
| - Mov(Tmp0(), Operand(CodeObject())); |
| - Push(Tmp1(), Tmp0()); |
| + Mov(type_reg, Operand(Smi::FromInt(type))); |
| + Mov(code_reg, Operand(CodeObject())); |
| + Push(type_reg, code_reg); |
| // jssp[4] : lr |
| // jssp[3] : fp |
| // jssp[2] : cp |
| @@ -2926,7 +2969,7 @@ void MacroAssembler::EnterFrame(StackFrame::Type type) { |
| // jssp[0] : code object |
| // Adjust FP to point to saved FP. |
| - add(fp, jssp, StandardFrameConstants::kFixedFrameSizeFromFp + kPointerSize); |
| + Add(fp, jssp, StandardFrameConstants::kFixedFrameSizeFromFp + kPointerSize); |
| } |
| @@ -3192,9 +3235,11 @@ void MacroAssembler::Allocate(int object_size, |
| return; |
| } |
| - ASSERT(!AreAliased(result, scratch1, scratch2, Tmp0(), Tmp1())); |
| - ASSERT(result.Is64Bits() && scratch1.Is64Bits() && scratch2.Is64Bits() && |
| - Tmp0().Is64Bits() && Tmp1().Is64Bits()); |
| + UseScratchRegisterScope temps(this); |
| + Register scratch3 = temps.AcquireX(); |
| + |
| + ASSERT(!AreAliased(result, scratch1, scratch2, scratch3)); |
| + ASSERT(result.Is64Bits() && scratch1.Is64Bits() && scratch2.Is64Bits()); |
| // Make object size into bytes. |
| if ((flags & SIZE_IN_WORDS) != 0) { |
| @@ -3223,8 +3268,8 @@ void MacroAssembler::Allocate(int object_size, |
| } else { |
| if (emit_debug_code()) { |
| // Assert that result actually contains top on entry. |
| - Ldr(Tmp0(), MemOperand(top_address)); |
| - Cmp(result, Tmp0()); |
| + Ldr(scratch3, MemOperand(top_address)); |
| + Cmp(result, scratch3); |
| Check(eq, kUnexpectedAllocationTop); |
| } |
| // Load the allocation limit. 'result' already contains the allocation top. |
| @@ -3236,11 +3281,11 @@ void MacroAssembler::Allocate(int object_size, |
| STATIC_ASSERT(kPointerAlignment == kDoubleAlignment); |
| // Calculate new top and bail out if new space is exhausted. |
| - Adds(Tmp1(), result, object_size); |
| + Adds(scratch3, result, object_size); |
| B(vs, gc_required); |
| - Cmp(Tmp1(), allocation_limit); |
| + Cmp(scratch3, allocation_limit); |
| B(hi, gc_required); |
| - Str(Tmp1(), MemOperand(top_address)); |
| + Str(scratch3, MemOperand(top_address)); |
| // Tag the object if requested. |
| if ((flags & TAG_OBJECT) != 0) { |
| @@ -3267,9 +3312,12 @@ void MacroAssembler::Allocate(Register object_size, |
| return; |
| } |
| - ASSERT(!AreAliased(object_size, result, scratch1, scratch2, Tmp0(), Tmp1())); |
| - ASSERT(object_size.Is64Bits() && result.Is64Bits() && scratch1.Is64Bits() && |
| - scratch2.Is64Bits() && Tmp0().Is64Bits() && Tmp1().Is64Bits()); |
| + UseScratchRegisterScope temps(this); |
| + Register scratch3 = temps.AcquireX(); |
| + |
| + ASSERT(!AreAliased(object_size, result, scratch1, scratch2, scratch3)); |
| + ASSERT(object_size.Is64Bits() && result.Is64Bits() && |
| + scratch1.Is64Bits() && scratch2.Is64Bits()); |
| // Check relative positions of allocation top and limit addresses. |
| // The values must be adjacent in memory to allow the use of LDP. |
| @@ -3292,8 +3340,8 @@ void MacroAssembler::Allocate(Register object_size, |
| } else { |
| if (emit_debug_code()) { |
| // Assert that result actually contains top on entry. |
| - Ldr(Tmp0(), MemOperand(top_address)); |
| - Cmp(result, Tmp0()); |
| + Ldr(scratch3, MemOperand(top_address)); |
| + Cmp(result, scratch3); |
| Check(eq, kUnexpectedAllocationTop); |
| } |
| // Load the allocation limit. 'result' already contains the allocation top. |
| @@ -3306,20 +3354,20 @@ void MacroAssembler::Allocate(Register object_size, |
| // Calculate new top and bail out if new space is exhausted |
| if ((flags & SIZE_IN_WORDS) != 0) { |
| - Adds(Tmp1(), result, Operand(object_size, LSL, kPointerSizeLog2)); |
| + Adds(scratch3, result, Operand(object_size, LSL, kPointerSizeLog2)); |
| } else { |
| - Adds(Tmp1(), result, object_size); |
| + Adds(scratch3, result, object_size); |
| } |
| if (emit_debug_code()) { |
| - Tst(Tmp1(), kObjectAlignmentMask); |
| + Tst(scratch3, kObjectAlignmentMask); |
| Check(eq, kUnalignedAllocationInNewSpace); |
| } |
| B(vs, gc_required); |
| - Cmp(Tmp1(), allocation_limit); |
| + Cmp(scratch3, allocation_limit); |
| B(hi, gc_required); |
| - Str(Tmp1(), MemOperand(top_address)); |
| + Str(scratch3, MemOperand(top_address)); |
| // Tag the object if requested. |
| if ((flags & TAG_OBJECT) != 0) { |
| @@ -3647,9 +3695,11 @@ void MacroAssembler::DispatchMap(Register obj, |
| void MacroAssembler::TestMapBitfield(Register object, uint64_t mask) { |
| - Ldr(Tmp0(), FieldMemOperand(object, HeapObject::kMapOffset)); |
| - Ldrb(Tmp0(), FieldMemOperand(Tmp0(), Map::kBitFieldOffset)); |
| - Tst(Tmp0(), mask); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + Ldr(temp, FieldMemOperand(object, HeapObject::kMapOffset)); |
| + Ldrb(temp, FieldMemOperand(temp, Map::kBitFieldOffset)); |
| + Tst(temp, mask); |
| } |
| @@ -3721,9 +3771,11 @@ void MacroAssembler::TryGetFunctionPrototype(Register function, |
| void MacroAssembler::CompareRoot(const Register& obj, |
| Heap::RootListIndex index) { |
| - ASSERT(!AreAliased(obj, Tmp0())); |
| - LoadRoot(Tmp0(), index); |
| - Cmp(obj, Tmp0()); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + ASSERT(!AreAliased(obj, temp)); |
| + LoadRoot(temp, index); |
| + Cmp(obj, temp); |
| } |
| @@ -3920,60 +3972,53 @@ void MacroAssembler::EmitSeqStringSetCharCheck( |
| void MacroAssembler::CheckAccessGlobalProxy(Register holder_reg, |
| - Register scratch, |
| + Register scratch1, |
| + Register scratch2, |
| Label* miss) { |
| - // TODO(jbramley): Sort out the uses of Tmp0() and Tmp1() in this function. |
| - // The ARM version takes two scratch registers, and that should be enough for |
| - // all of the checks. |
| - |
| + ASSERT(!AreAliased(holder_reg, scratch1, scratch2)); |
| Label same_contexts; |
| - ASSERT(!AreAliased(holder_reg, scratch)); |
| - |
| // Load current lexical context from the stack frame. |
| - Ldr(scratch, MemOperand(fp, StandardFrameConstants::kContextOffset)); |
| + Ldr(scratch1, MemOperand(fp, StandardFrameConstants::kContextOffset)); |
| // In debug mode, make sure the lexical context is set. |
| #ifdef DEBUG |
| - Cmp(scratch, 0); |
| + Cmp(scratch1, 0); |
| Check(ne, kWeShouldNotHaveAnEmptyLexicalContext); |
| #endif |
| // Load the native context of the current context. |
| int offset = |
| Context::kHeaderSize + Context::GLOBAL_OBJECT_INDEX * kPointerSize; |
| - Ldr(scratch, FieldMemOperand(scratch, offset)); |
| - Ldr(scratch, FieldMemOperand(scratch, GlobalObject::kNativeContextOffset)); |
| + Ldr(scratch1, FieldMemOperand(scratch1, offset)); |
| + Ldr(scratch1, FieldMemOperand(scratch1, GlobalObject::kNativeContextOffset)); |
| // Check the context is a native context. |
| if (emit_debug_code()) { |
| // Read the first word and compare to the global_context_map. |
| - Register temp = Tmp1(); |
| - Ldr(temp, FieldMemOperand(scratch, HeapObject::kMapOffset)); |
| - CompareRoot(temp, Heap::kNativeContextMapRootIndex); |
| + Ldr(scratch2, FieldMemOperand(scratch1, HeapObject::kMapOffset)); |
| + CompareRoot(scratch2, Heap::kNativeContextMapRootIndex); |
| Check(eq, kExpectedNativeContext); |
| } |
| // Check if both contexts are the same. |
| - ldr(Tmp0(), FieldMemOperand(holder_reg, JSGlobalProxy::kNativeContextOffset)); |
| - cmp(scratch, Tmp0()); |
| - b(&same_contexts, eq); |
| + Ldr(scratch2, FieldMemOperand(holder_reg, |
| + JSGlobalProxy::kNativeContextOffset)); |
| + Cmp(scratch1, scratch2); |
| + B(&same_contexts, eq); |
| // Check the context is a native context. |
| if (emit_debug_code()) { |
| - // Move Tmp0() into a different register, as CompareRoot will use it. |
| - Register temp = Tmp1(); |
| - mov(temp, Tmp0()); |
| - CompareRoot(temp, Heap::kNullValueRootIndex); |
| + // We're short on scratch registers here, so use holder_reg as a scratch. |
| + Push(holder_reg); |
| + Register scratch3 = holder_reg; |
| + |
| + CompareRoot(scratch2, Heap::kNullValueRootIndex); |
| Check(ne, kExpectedNonNullContext); |
| - Ldr(temp, FieldMemOperand(temp, HeapObject::kMapOffset)); |
| - CompareRoot(temp, Heap::kNativeContextMapRootIndex); |
| + Ldr(scratch3, FieldMemOperand(scratch2, HeapObject::kMapOffset)); |
| + CompareRoot(scratch3, Heap::kNativeContextMapRootIndex); |
| Check(eq, kExpectedNativeContext); |
| - |
| - // Let's consider that Tmp0() has been cloberred by the MacroAssembler. |
| - // We reload it with its value. |
| - ldr(Tmp0(), FieldMemOperand(holder_reg, |
| - JSGlobalProxy::kNativeContextOffset)); |
| + Pop(holder_reg); |
| } |
| // Check that the security token in the calling global object is |
| @@ -3982,12 +4027,12 @@ void MacroAssembler::CheckAccessGlobalProxy(Register holder_reg, |
| int token_offset = Context::kHeaderSize + |
| Context::SECURITY_TOKEN_INDEX * kPointerSize; |
| - ldr(scratch, FieldMemOperand(scratch, token_offset)); |
| - ldr(Tmp0(), FieldMemOperand(Tmp0(), token_offset)); |
| - cmp(scratch, Tmp0()); |
| - b(miss, ne); |
| + Ldr(scratch1, FieldMemOperand(scratch1, token_offset)); |
| + Ldr(scratch2, FieldMemOperand(scratch2, token_offset)); |
| + Cmp(scratch1, scratch2); |
| + B(miss, ne); |
| - bind(&same_contexts); |
| + Bind(&same_contexts); |
| } |
| @@ -4090,10 +4135,10 @@ void MacroAssembler::LoadFromNumberDictionary(Label* miss, |
| void MacroAssembler::RememberedSetHelper(Register object, // For debug tests. |
| Register address, |
| - Register scratch, |
| + Register scratch1, |
| SaveFPRegsMode fp_mode, |
| RememberedSetFinalAction and_then) { |
| - ASSERT(!AreAliased(object, address, scratch)); |
| + ASSERT(!AreAliased(object, address, scratch1)); |
| Label done, store_buffer_overflow; |
| if (emit_debug_code()) { |
| Label ok; |
| @@ -4101,22 +4146,25 @@ void MacroAssembler::RememberedSetHelper(Register object, // For debug tests. |
| Abort(kRememberedSetPointerInNewSpace); |
| bind(&ok); |
| } |
| + UseScratchRegisterScope temps(this); |
| + Register scratch2 = temps.AcquireX(); |
| + |
| // Load store buffer top. |
| - Mov(Tmp0(), Operand(ExternalReference::store_buffer_top(isolate()))); |
| - Ldr(scratch, MemOperand(Tmp0())); |
| + Mov(scratch2, Operand(ExternalReference::store_buffer_top(isolate()))); |
| + Ldr(scratch1, MemOperand(scratch2)); |
| // Store pointer to buffer and increment buffer top. |
| - Str(address, MemOperand(scratch, kPointerSize, PostIndex)); |
| + Str(address, MemOperand(scratch1, kPointerSize, PostIndex)); |
| // Write back new top of buffer. |
| - Str(scratch, MemOperand(Tmp0())); |
| + Str(scratch1, MemOperand(scratch2)); |
| // Call stub on end of buffer. |
| // Check for end of buffer. |
| ASSERT(StoreBuffer::kStoreBufferOverflowBit == |
| (1 << (14 + kPointerSizeLog2))); |
| if (and_then == kFallThroughAtEnd) { |
| - Tbz(scratch, (14 + kPointerSizeLog2), &done); |
| + Tbz(scratch1, (14 + kPointerSizeLog2), &done); |
| } else { |
| ASSERT(and_then == kReturnAtEnd); |
| - Tbnz(scratch, (14 + kPointerSizeLog2), &store_buffer_overflow); |
| + Tbnz(scratch1, (14 + kPointerSizeLog2), &store_buffer_overflow); |
| Ret(); |
| } |
| @@ -4264,7 +4312,7 @@ void MacroAssembler::RecordWriteField( |
| } |
| -// Will clobber: object, address, value, Tmp0(), Tmp1(). |
| +// Will clobber: object, address, value. |
| // If lr_status is kLRHasBeenSaved, lr will also be clobbered. |
| // |
| // The register 'object' contains a heap object pointer. The heap object tag is |
| @@ -4280,8 +4328,11 @@ void MacroAssembler::RecordWrite(Register object, |
| ASSERT(!AreAliased(object, value)); |
| if (emit_debug_code()) { |
| - Ldr(Tmp0(), MemOperand(address)); |
| - Cmp(Tmp0(), value); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + |
| + Ldr(temp, MemOperand(address)); |
| + Cmp(temp, value); |
| Check(eq, kWrongAddressOrValuePassedToRecordWrite); |
| } |
| @@ -4346,15 +4397,18 @@ void MacroAssembler::AssertHasValidColor(const Register& reg) { |
| void MacroAssembler::GetMarkBits(Register addr_reg, |
| Register bitmap_reg, |
| Register shift_reg) { |
| - ASSERT(!AreAliased(addr_reg, bitmap_reg, shift_reg, no_reg)); |
| + ASSERT(!AreAliased(addr_reg, bitmap_reg, shift_reg)); |
| + ASSERT(addr_reg.Is64Bits() && bitmap_reg.Is64Bits() && shift_reg.Is64Bits()); |
| // addr_reg is divided into fields: |
| // |63 page base 20|19 high 8|7 shift 3|2 0| |
| // 'high' gives the index of the cell holding color bits for the object. |
| // 'shift' gives the offset in the cell for this object's color. |
| const int kShiftBits = kPointerSizeLog2 + Bitmap::kBitsPerCellLog2; |
| - Ubfx(Tmp0(), addr_reg, kShiftBits, kPageSizeBits - kShiftBits); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| + Ubfx(temp, addr_reg, kShiftBits, kPageSizeBits - kShiftBits); |
| Bic(bitmap_reg, addr_reg, Page::kPageAlignmentMask); |
| - Add(bitmap_reg, bitmap_reg, Operand(Tmp0(), LSL, Bitmap::kBytesPerCellLog2)); |
| + Add(bitmap_reg, bitmap_reg, Operand(temp, LSL, Bitmap::kBytesPerCellLog2)); |
| // bitmap_reg: |
| // |63 page base 20|19 zeros 15|14 high 3|2 0| |
| Ubfx(shift_reg, addr_reg, kPointerSizeLog2, Bitmap::kBitsPerCellLog2); |
| @@ -4578,8 +4632,6 @@ void MacroAssembler::AssertRegisterIsClear(Register reg, BailoutReason reason) { |
| void MacroAssembler::AssertRegisterIsRoot(Register reg, |
| Heap::RootListIndex index, |
| BailoutReason reason) { |
| - // CompareRoot uses Tmp0(). |
| - ASSERT(!reg.Is(Tmp0())); |
| if (emit_debug_code()) { |
| CompareRoot(reg, index); |
| Check(eq, reason); |
| @@ -4589,7 +4641,8 @@ void MacroAssembler::AssertRegisterIsRoot(Register reg, |
| void MacroAssembler::AssertFastElements(Register elements) { |
| if (emit_debug_code()) { |
| - Register temp = Tmp1(); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| Label ok; |
| Ldr(temp, FieldMemOperand(elements, HeapObject::kMapOffset)); |
| JumpIfRoot(temp, Heap::kFixedArrayMapRootIndex, &ok); |
| @@ -4603,7 +4656,8 @@ void MacroAssembler::AssertFastElements(Register elements) { |
| void MacroAssembler::AssertIsString(const Register& object) { |
| if (emit_debug_code()) { |
| - Register temp = Tmp1(); |
| + UseScratchRegisterScope temps(this); |
| + Register temp = temps.AcquireX(); |
| STATIC_ASSERT(kSmiTag == 0); |
| Tst(object, Operand(kSmiTagMask)); |
| Check(ne, kOperandIsNotAString); |
| @@ -4650,6 +4704,12 @@ void MacroAssembler::Abort(BailoutReason reason) { |
| SetStackPointer(jssp); |
| Mov(jssp, old_stack_pointer); |
| + // We need some scratch registers for the MacroAssembler, so make sure we have |
| + // some. This is safe here because Abort never returns. |
| + RegList old_tmp_list = TmpList()->list(); |
| + TmpList()->Combine(ip0); |
| + TmpList()->Combine(ip1); |
| + |
| if (use_real_aborts()) { |
| // Avoid infinite recursion; Push contains some assertions that use Abort. |
| NoUseRealAbortsScope no_real_aborts(this); |
| @@ -4686,6 +4746,7 @@ void MacroAssembler::Abort(BailoutReason reason) { |
| } |
| SetStackPointer(old_stack_pointer); |
| + TmpList()->set_list(old_tmp_list); |
| } |
| @@ -4693,22 +4754,23 @@ void MacroAssembler::LoadTransitionedArrayMapConditional( |
| ElementsKind expected_kind, |
| ElementsKind transitioned_kind, |
| Register map_in_out, |
| - Register scratch, |
| + Register scratch1, |
| + Register scratch2, |
| Label* no_map_match) { |
| // Load the global or builtins object from the current context. |
| - Ldr(scratch, GlobalObjectMemOperand()); |
| - Ldr(scratch, FieldMemOperand(scratch, GlobalObject::kNativeContextOffset)); |
| + Ldr(scratch1, GlobalObjectMemOperand()); |
| + Ldr(scratch1, FieldMemOperand(scratch1, GlobalObject::kNativeContextOffset)); |
| // Check that the function's map is the same as the expected cached map. |
| - Ldr(scratch, ContextMemOperand(scratch, Context::JS_ARRAY_MAPS_INDEX)); |
| + Ldr(scratch1, ContextMemOperand(scratch1, Context::JS_ARRAY_MAPS_INDEX)); |
| size_t offset = (expected_kind * kPointerSize) + FixedArrayBase::kHeaderSize; |
| - Ldr(Tmp0(), FieldMemOperand(scratch, offset)); |
| - Cmp(map_in_out, Tmp0()); |
| + Ldr(scratch2, FieldMemOperand(scratch1, offset)); |
| + Cmp(map_in_out, scratch2); |
| B(ne, no_map_match); |
| // Use the transitioned cached map. |
| offset = (transitioned_kind * kPointerSize) + FixedArrayBase::kHeaderSize; |
| - Ldr(map_in_out, FieldMemOperand(scratch, offset)); |
| + Ldr(map_in_out, FieldMemOperand(scratch1, offset)); |
| } |
| @@ -4761,14 +4823,18 @@ void MacroAssembler::PrintfNoPreserve(const char * format, |
| // in most cases anyway, so this restriction shouldn't be too serious. |
| ASSERT(!kCallerSaved.IncludesAliasOf(__ StackPointer())); |
| - // We cannot print Tmp0() or Tmp1() as they're used internally by the macro |
| - // assembler. We cannot print the stack pointer because it is typically used |
| - // to preserve caller-saved registers (using other Printf variants which |
| - // depend on this helper). |
| - ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg0)); |
| - ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg1)); |
| - ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg2)); |
| - ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg3)); |
| + // Make sure that the macro assembler doesn't try to use any of our arguments |
| + // as scratch registers. |
| + UseScratchRegisterScope temps(this); |
| + temps.Exclude(arg0, arg1, arg2, arg3); |
| + |
| + // We cannot print the stack pointer because it is typically used to preserve |
| + // caller-saved registers (using other Printf variants which depend on this |
| + // helper). |
| + ASSERT(!AreAliased(arg0, StackPointer())); |
| + ASSERT(!AreAliased(arg1, StackPointer())); |
| + ASSERT(!AreAliased(arg2, StackPointer())); |
| + ASSERT(!AreAliased(arg3, StackPointer())); |
| static const int kMaxArgCount = 4; |
| // Assume that we have the maximum number of arguments until we know |
| @@ -4910,20 +4976,34 @@ void MacroAssembler::Printf(const char * format, |
| const CPURegister& arg1, |
| const CPURegister& arg2, |
| const CPURegister& arg3) { |
| + // Printf is expected to preserve all registers, so make sure that none are |
| + // available as scratch registers until we've preserved them. |
| + UseScratchRegisterScope exclude_all(this); |
| + exclude_all.ExcludeAll(); |
| + |
| // Preserve all caller-saved registers as well as NZCV. |
| // If csp is the stack pointer, PushCPURegList asserts that the size of each |
| // list is a multiple of 16 bytes. |
| PushCPURegList(kCallerSaved); |
| PushCPURegList(kCallerSavedFP); |
| - // Use Tmp0() as a scratch register. It is not accepted by Printf so it will |
| - // never overlap an argument register. |
| - Mrs(Tmp0(), NZCV); |
| - Push(Tmp0(), xzr); |
| - PrintfNoPreserve(format, arg0, arg1, arg2, arg3); |
| + { UseScratchRegisterScope temps(this); |
| + // We can use caller-saved registers as scratch values (except for argN). |
| + TmpList()->Combine(kCallerSaved); |
| + FPTmpList()->Combine(kCallerSavedFP); |
|
rmcilroy
2014/02/27 11:17:32
Do we really need all these registers as temps? S
jbramley
2014/02/27 11:27:16
No, it doesn't need them, but it's easier to provi
rmcilroy
2014/02/27 11:49:29
I still don't like this - it feels like a hack and
|
| + temps.Exclude(arg0, arg1, arg2, arg3); |
| + |
| + // Preserve NZCV. |
| + Register tmp = temps.AcquireX(); |
| + Mrs(tmp, NZCV); |
| + Push(tmp, xzr); |
| + |
| + PrintfNoPreserve(format, arg0, arg1, arg2, arg3); |
| + |
| + Pop(xzr, tmp); |
| + Msr(NZCV, tmp); |
| + } |
| - Pop(xzr, Tmp0()); |
| - Msr(NZCV, Tmp0()); |
| PopCPURegList(kCallerSavedFP); |
| PopCPURegList(kCallerSaved); |
| } |
| @@ -5031,6 +5111,106 @@ bool MacroAssembler::IsCodeAgeSequence(byte* sequence) { |
| #undef __ |
| + |
| + |
| +UseScratchRegisterScope::~UseScratchRegisterScope() { |
| + available_->set_list(old_available_); |
| + availablefp_->set_list(old_availablefp_); |
| +} |
| + |
| + |
| +Register UseScratchRegisterScope::AcquireSameSizeAs(const Register& reg) { |
| + int code = AcquireNextAvailable(available_).code(); |
| + return Register::Create(code, reg.SizeInBits()); |
| +} |
| + |
| + |
| +FPRegister UseScratchRegisterScope::AcquireSameSizeAs(const FPRegister& reg) { |
| + int code = AcquireNextAvailable(availablefp_).code(); |
| + return FPRegister::Create(code, reg.SizeInBits()); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::Exclude(const Register& reg1, |
| + const Register& reg2, |
| + const Register& reg3, |
| + const Register& reg4) { |
| + RegList exclude = reg1.Bit() | reg2.Bit() | reg3.Bit() | reg4.Bit(); |
| + ExcludeByRegList(available_, exclude); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::Exclude(const FPRegister& reg1, |
| + const FPRegister& reg2, |
| + const FPRegister& reg3, |
| + const FPRegister& reg4) { |
| + RegList excludefp = reg1.Bit() | reg2.Bit() | reg3.Bit() | reg4.Bit(); |
| + ExcludeByRegList(availablefp_, excludefp); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::Exclude(const CPURegister& reg1, |
| + const CPURegister& reg2, |
| + const CPURegister& reg3, |
| + const CPURegister& reg4) { |
| + RegList exclude = 0; |
| + RegList excludefp = 0; |
| + |
| + const CPURegister regs[] = {reg1, reg2, reg3, reg4}; |
| + |
| + for (unsigned i = 0; i < (sizeof(regs) / sizeof(regs[0])); i++) { |
| + if (regs[i].IsRegister()) { |
| + exclude |= regs[i].Bit(); |
| + } else if (regs[i].IsFPRegister()) { |
| + excludefp |= regs[i].Bit(); |
| + } else { |
| + ASSERT(regs[i].IsNone()); |
| + } |
| + } |
| + |
| + ExcludeByRegList(available_, exclude); |
| + ExcludeByRegList(availablefp_, excludefp); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::ExcludeAll() { |
| + ExcludeByRegList(available_, available_->list()); |
| + ExcludeByRegList(availablefp_, availablefp_->list()); |
| +} |
| + |
| + |
| +CPURegister UseScratchRegisterScope::AcquireNextAvailable( |
| + CPURegList* available) { |
| + CHECK(!available->IsEmpty()); |
| + CPURegister result = available->PopLowestIndex(); |
| + ASSERT(!AreAliased(result, xzr, csp)); |
| + return result; |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::ReleaseByCode(CPURegList* available, int code) { |
| + ReleaseByRegList(available, static_cast<RegList>(1) << code); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::ReleaseByRegList(CPURegList* available, |
| + RegList regs) { |
| + available->set_list(available->list() | regs); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::IncludeByRegList(CPURegList* available, |
| + RegList regs) { |
| + available->set_list(available->list() | regs); |
| +} |
| + |
| + |
| +void UseScratchRegisterScope::ExcludeByRegList(CPURegList* available, |
| + RegList exclude) { |
| + available->set_list(available->list() & ~exclude); |
| +} |
| + |
| + |
| #define __ masm-> |