Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(1)

Unified Diff: src/a64/macro-assembler-a64.cc

Issue 164793003: A64: Use a scope utility to allocate scratch registers. (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: Rebase. Created 6 years, 9 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/a64/macro-assembler-a64.h ('k') | src/a64/macro-assembler-a64-inl.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/a64/macro-assembler-a64.cc
diff --git a/src/a64/macro-assembler-a64.cc b/src/a64/macro-assembler-a64.cc
index e69597cbff4d49b594ff642853fda37caaffa331..2bc2128481d770354636b79d9d1d0e52b98df90b 100644
--- a/src/a64/macro-assembler-a64.cc
+++ b/src/a64/macro-assembler-a64.cc
@@ -53,7 +53,7 @@ MacroAssembler::MacroAssembler(Isolate* arg_isolate,
#endif
has_frame_(false),
use_real_aborts_(true),
- sp_(jssp), tmp0_(ip0), tmp1_(ip1), fptmp0_(fp_scratch) {
+ sp_(jssp), tmp_list_(ip0, ip1), fptmp_list_(fp_scratch) {
if (isolate() != NULL) {
code_object_ = Handle<Object>(isolate()->heap()->undefined_value(),
isolate());
@@ -65,9 +65,12 @@ void MacroAssembler::LogicalMacro(const Register& rd,
const Register& rn,
const Operand& operand,
LogicalOp op) {
+ UseScratchRegisterScope temps(this);
+
if (operand.NeedsRelocation()) {
- LoadRelocated(Tmp0(), operand);
- Logical(rd, rn, Tmp0(), op);
+ Register temp = temps.AcquireX();
+ LoadRelocated(temp, operand);
+ Logical(rd, rn, temp, op);
} else if (operand.IsImmediate()) {
int64_t immediate = operand.immediate();
@@ -125,7 +128,7 @@ void MacroAssembler::LogicalMacro(const Register& rd,
LogicalImmediate(rd, rn, n, imm_s, imm_r, op);
} else {
// Immediate can't be encoded: synthesize using move immediate.
- Register temp = AppropriateTempFor(rn);
+ Register temp = temps.AcquireSameSizeAs(rn);
Mov(temp, immediate);
if (rd.Is(csp)) {
// If rd is the stack pointer we cannot use it as the destination
@@ -144,7 +147,7 @@ void MacroAssembler::LogicalMacro(const Register& rd,
ASSERT(operand.shift_amount() <= 4);
ASSERT(operand.reg().Is64Bits() ||
((operand.extend() != UXTX) && (operand.extend() != SXTX)));
- Register temp = AppropriateTempFor(rn, operand.reg());
+ Register temp = temps.AcquireSameSizeAs(rn);
EmitExtendShift(temp, operand.reg(), operand.extend(),
operand.shift_amount());
Logical(rd, rn, temp, op);
@@ -208,9 +211,10 @@ void MacroAssembler::Mov(const Register& rd, uint64_t imm) {
invert_move = true;
}
- // Mov instructions can't move value into the stack pointer, so set up a
- // temporary register, if needed.
- Register temp = rd.IsSP() ? AppropriateTempFor(rd) : rd;
+ // Mov instructions can't move immediate values into the stack pointer, so
+ // set up a temporary register, if needed.
+ UseScratchRegisterScope temps(this);
+ Register temp = rd.IsSP() ? temps.AcquireSameSizeAs(rd) : rd;
// Iterate through the halfwords. Use movn/movz for the first non-ignored
// halfword, and movk for subsequent halfwords.
@@ -248,9 +252,11 @@ void MacroAssembler::Mov(const Register& rd,
DiscardMoveMode discard_mode) {
ASSERT(allow_macro_instructions_);
ASSERT(!rd.IsZero());
+
// Provide a swap register for instructions that need to write into the
// system stack pointer (and can't do this inherently).
- Register dst = (rd.Is(csp)) ? (Tmp1()) : (rd);
+ UseScratchRegisterScope temps(this);
+ Register dst = (rd.IsSP()) ? temps.AcquireSameSizeAs(rd) : rd;
if (operand.NeedsRelocation()) {
LoadRelocated(dst, operand);
@@ -291,8 +297,7 @@ void MacroAssembler::Mov(const Register& rd,
// Copy the result to the system stack pointer.
if (!dst.Is(rd)) {
- ASSERT(rd.IsZero());
- ASSERT(dst.Is(Tmp1()));
+ ASSERT(rd.IsSP());
Assembler::mov(rd, dst);
}
}
@@ -302,8 +307,8 @@ void MacroAssembler::Mvn(const Register& rd, const Operand& operand) {
ASSERT(allow_macro_instructions_);
if (operand.NeedsRelocation()) {
- LoadRelocated(Tmp0(), operand);
- Mvn(rd, Tmp0());
+ LoadRelocated(rd, operand);
+ mvn(rd, rd);
} else if (operand.IsImmediate()) {
// Call the macro assembler for generic immediates.
@@ -312,14 +317,11 @@ void MacroAssembler::Mvn(const Register& rd, const Operand& operand) {
} else if (operand.IsExtendedRegister()) {
// Emit two instructions for the extend case. This differs from Mov, as
// the extend and invert can't be achieved in one instruction.
- Register temp = AppropriateTempFor(rd, operand.reg());
- EmitExtendShift(temp, operand.reg(), operand.extend(),
+ EmitExtendShift(rd, operand.reg(), operand.extend(),
operand.shift_amount());
- mvn(rd, temp);
+ mvn(rd, rd);
} else {
- // Otherwise, emit a register move only if the registers are distinct.
- // If the jssp is an operand, add #0 is emitted, otherwise, orr #0.
mvn(rd, operand);
}
}
@@ -360,8 +362,10 @@ void MacroAssembler::ConditionalCompareMacro(const Register& rn,
ConditionalCompareOp op) {
ASSERT((cond != al) && (cond != nv));
if (operand.NeedsRelocation()) {
- LoadRelocated(Tmp0(), operand);
- ConditionalCompareMacro(rn, Tmp0(), nzcv, cond, op);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ LoadRelocated(temp, operand);
+ ConditionalCompareMacro(rn, temp, nzcv, cond, op);
} else if ((operand.IsShiftedRegister() && (operand.shift_amount() == 0)) ||
(operand.IsImmediate() && IsImmConditionalCompare(operand.immediate()))) {
@@ -372,7 +376,8 @@ void MacroAssembler::ConditionalCompareMacro(const Register& rn,
} else {
// The operand isn't directly supported by the instruction: perform the
// operation on a temporary register.
- Register temp = AppropriateTempFor(rn);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireSameSizeAs(rn);
Mov(temp, operand);
ConditionalCompare(rn, temp, nzcv, cond, op);
}
@@ -398,7 +403,8 @@ void MacroAssembler::Csel(const Register& rd,
} else if (imm == -1) {
csinv(rd, rn, zr, cond);
} else {
- Register temp = AppropriateTempFor(rn);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireSameSizeAs(rn);
Mov(temp, operand.immediate());
csel(rd, rn, temp, cond);
}
@@ -407,7 +413,8 @@ void MacroAssembler::Csel(const Register& rd,
csel(rd, rn, operand.reg(), cond);
} else {
// All other arguments.
- Register temp = AppropriateTempFor(rn);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireSameSizeAs(rn);
Mov(temp, operand);
csel(rd, rn, temp, cond);
}
@@ -426,12 +433,15 @@ void MacroAssembler::AddSubMacro(const Register& rd,
}
if (operand.NeedsRelocation()) {
- LoadRelocated(Tmp0(), operand);
- AddSubMacro(rd, rn, Tmp0(), S, op);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ LoadRelocated(temp, operand);
+ AddSubMacro(rd, rn, temp, S, op);
} else if ((operand.IsImmediate() && !IsImmAddSub(operand.immediate())) ||
(rn.IsZero() && !operand.IsShiftedRegister()) ||
(operand.IsShiftedRegister() && (operand.shift() == ROR))) {
- Register temp = AppropriateTempFor(rn);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireSameSizeAs(rn);
Mov(temp, operand);
AddSub(rd, rn, temp, S, op);
} else {
@@ -446,24 +456,28 @@ void MacroAssembler::AddSubWithCarryMacro(const Register& rd,
FlagsUpdate S,
AddSubWithCarryOp op) {
ASSERT(rd.SizeInBits() == rn.SizeInBits());
+ UseScratchRegisterScope temps(this);
if (operand.NeedsRelocation()) {
- LoadRelocated(Tmp0(), operand);
- AddSubWithCarryMacro(rd, rn, Tmp0(), S, op);
+ Register temp = temps.AcquireX();
+ LoadRelocated(temp, operand);
+ AddSubWithCarryMacro(rd, rn, temp, S, op);
} else if (operand.IsImmediate() ||
(operand.IsShiftedRegister() && (operand.shift() == ROR))) {
// Add/sub with carry (immediate or ROR shifted register.)
- Register temp = AppropriateTempFor(rn);
+ Register temp = temps.AcquireSameSizeAs(rn);
Mov(temp, operand);
AddSubWithCarry(rd, rn, temp, S, op);
+
} else if (operand.IsShiftedRegister() && (operand.shift_amount() != 0)) {
// Add/sub with carry (shifted register).
ASSERT(operand.reg().SizeInBits() == rd.SizeInBits());
ASSERT(operand.shift() != ROR);
- ASSERT(is_uintn(operand.shift_amount(),
- rd.SizeInBits() == kXRegSize ? kXRegSizeLog2 : kWRegSizeLog2));
- Register temp = AppropriateTempFor(rn, operand.reg());
+ ASSERT(
+ is_uintn(operand.shift_amount(),
+ rd.SizeInBits() == kXRegSize ? kXRegSizeLog2 : kWRegSizeLog2));
+ Register temp = temps.AcquireSameSizeAs(rn);
EmitShift(temp, operand.reg(), operand.shift(), operand.shift_amount());
AddSubWithCarry(rd, rn, temp, S, op);
@@ -475,7 +489,7 @@ void MacroAssembler::AddSubWithCarryMacro(const Register& rd,
ASSERT(operand.shift_amount() <= 4);
ASSERT(operand.reg().Is64Bits() ||
((operand.extend() != UXTX) && (operand.extend() != SXTX)));
- Register temp = AppropriateTempFor(rn, operand.reg());
+ Register temp = temps.AcquireSameSizeAs(rn);
EmitExtendShift(temp, operand.reg(), operand.extend(),
operand.shift_amount());
AddSubWithCarry(rd, rn, temp, S, op);
@@ -500,7 +514,8 @@ void MacroAssembler::LoadStoreMacro(const CPURegister& rt,
!IsImmLSUnscaled(offset)) {
// Immediate offset that can't be encoded using unsigned or unscaled
// addressing modes.
- Register temp = AppropriateTempFor(addr.base());
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireSameSizeAs(addr.base());
Mov(temp, addr.offset());
LoadStore(rt, MemOperand(addr.base(), temp), op);
} else if (addr.IsPostIndex() && !IsImmLSUnscaled(offset)) {
@@ -856,11 +871,14 @@ void MacroAssembler::PushMultipleTimes(CPURegister src, int count) {
PrepareForPush(count, size);
if (FLAG_optimize_for_size && count > 8) {
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+
Label loop;
- __ Mov(Tmp0(), count / 2);
+ __ Mov(temp, count / 2);
__ Bind(&loop);
PushHelper(2, size, src, src, NoReg, NoReg);
- __ Subs(Tmp0(), Tmp0(), 1);
+ __ Subs(temp, temp, 1);
__ B(ne, &loop);
count %= 2;
@@ -888,7 +906,8 @@ void MacroAssembler::PushMultipleTimes(CPURegister src, int count) {
void MacroAssembler::PushMultipleTimes(CPURegister src, Register count) {
PrepareForPush(Operand(count, UXTW, WhichPowerOf2(src.SizeInBytes())));
- Register temp = AppropriateTempFor(count);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireSameSizeAs(count);
if (FLAG_optimize_for_size) {
Label loop, done;
@@ -1332,10 +1351,10 @@ void MacroAssembler::InNewSpace(Register object,
Condition cond,
Label* branch) {
ASSERT(cond == eq || cond == ne);
- // Use Tmp1() to have a different destination register, as Tmp0() will be used
- // for relocation.
- And(Tmp1(), object, Operand(ExternalReference::new_space_mask(isolate())));
- Cmp(Tmp1(), Operand(ExternalReference::new_space_start(isolate())));
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ And(temp, object, Operand(ExternalReference::new_space_mask(isolate())));
+ Cmp(temp, Operand(ExternalReference::new_space_start(isolate())));
B(cond, branch);
}
@@ -1504,8 +1523,11 @@ void MacroAssembler::AssertName(Register object) {
Abort(kOperandIsASmiAndNotAName);
Bind(&not_smi);
- Ldr(Tmp1(), FieldMemOperand(object, HeapObject::kMapOffset));
- CompareInstanceType(Tmp1(), Tmp1(), LAST_NAME_TYPE);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+
+ Ldr(temp, FieldMemOperand(object, HeapObject::kMapOffset));
+ CompareInstanceType(temp, temp, LAST_NAME_TYPE);
Check(ls, kOperandIsNotAName);
}
}
@@ -1513,7 +1535,8 @@ void MacroAssembler::AssertName(Register object) {
void MacroAssembler::AssertString(Register object) {
if (emit_debug_code()) {
- Register temp = Tmp1();
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
STATIC_ASSERT(kSmiTag == 0);
Tst(object, kSmiTagMask);
Check(ne, kOperandIsASmiAndNotAString);
@@ -1830,8 +1853,10 @@ void MacroAssembler::CallCFunction(ExternalReference function,
void MacroAssembler::CallCFunction(ExternalReference function,
int num_of_reg_args,
int num_of_double_args) {
- Mov(Tmp0(), Operand(function));
- CallCFunction(Tmp0(), num_of_reg_args, num_of_double_args);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ Mov(temp, Operand(function));
+ CallCFunction(temp, num_of_reg_args, num_of_double_args);
}
@@ -1884,7 +1909,8 @@ void MacroAssembler::CallCFunction(Register function,
// Because the stack pointer must be aligned on a 16-byte boundary, the
// aligned csp can be up to 12 bytes below the jssp. This is the case
// where we only pushed one W register on top of an aligned jssp.
- Register temp = Tmp1();
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
ASSERT(ActivationFrameAlignment() == 16);
Sub(temp, csp, old_stack_pointer);
// We want temp <= 0 && temp >= -12.
@@ -1903,8 +1929,10 @@ void MacroAssembler::Jump(Register target) {
void MacroAssembler::Jump(intptr_t target, RelocInfo::Mode rmode) {
- Mov(Tmp0(), Operand(target, rmode));
- Br(Tmp0());
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ Mov(temp, Operand(target, rmode));
+ Br(temp);
}
@@ -1966,16 +1994,19 @@ void MacroAssembler::Call(Address target, RelocInfo::Mode rmode) {
// Addresses always have 64 bits, so we shouldn't encounter NONE32.
ASSERT(rmode != RelocInfo::NONE32);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+
if (rmode == RelocInfo::NONE64) {
uint64_t imm = reinterpret_cast<uint64_t>(target);
- movz(Tmp0(), (imm >> 0) & 0xffff, 0);
- movk(Tmp0(), (imm >> 16) & 0xffff, 16);
- movk(Tmp0(), (imm >> 32) & 0xffff, 32);
- movk(Tmp0(), (imm >> 48) & 0xffff, 48);
+ movz(temp, (imm >> 0) & 0xffff, 0);
+ movk(temp, (imm >> 16) & 0xffff, 16);
+ movk(temp, (imm >> 32) & 0xffff, 32);
+ movk(temp, (imm >> 48) & 0xffff, 48);
} else {
- LoadRelocated(Tmp0(), Operand(reinterpret_cast<intptr_t>(target), rmode));
+ LoadRelocated(temp, Operand(reinterpret_cast<intptr_t>(target), rmode));
}
- Blr(Tmp0());
+ Blr(temp);
#ifdef DEBUG
AssertSizeOfCodeGeneratedSince(&start_call, CallSize(target, rmode));
#endif
@@ -2056,21 +2087,23 @@ void MacroAssembler::JumpForHeapNumber(Register object,
Label* on_heap_number,
Label* on_not_heap_number) {
ASSERT(on_heap_number || on_not_heap_number);
- // Tmp0() is used as a scratch register.
- ASSERT(!AreAliased(Tmp0(), heap_number_map));
AssertNotSmi(object);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+
// Load the HeapNumber map if it is not passed.
if (heap_number_map.Is(NoReg)) {
- heap_number_map = Tmp1();
+ heap_number_map = temps.AcquireX();
LoadRoot(heap_number_map, Heap::kHeapNumberMapRootIndex);
} else {
- // This assert clobbers Tmp0(), so do it before loading Tmp0() with the map.
AssertRegisterIsRoot(heap_number_map, Heap::kHeapNumberMapRootIndex);
}
- Ldr(Tmp0(), FieldMemOperand(object, HeapObject::kMapOffset));
- Cmp(Tmp0(), heap_number_map);
+ ASSERT(!AreAliased(temp, heap_number_map));
+
+ Ldr(temp, FieldMemOperand(object, HeapObject::kMapOffset));
+ Cmp(temp, heap_number_map);
if (on_heap_number) {
B(eq, on_heap_number);
@@ -2196,10 +2229,12 @@ void MacroAssembler::TryConvertDoubleToInt(Register as_int,
void MacroAssembler::JumpIfMinusZero(DoubleRegister input,
Label* on_negative_zero) {
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
// Floating point -0.0 is kMinInt as an integer, so subtracting 1 (cmp) will
// cause overflow.
- Fmov(Tmp0(), input);
- Cmp(Tmp0(), 1);
+ Fmov(temp, input);
+ Cmp(temp, 1);
B(vs, on_negative_zero);
}
@@ -2209,10 +2244,8 @@ void MacroAssembler::ClampInt32ToUint8(Register output, Register input) {
Cmp(input.W(), Operand(input.W(), UXTB));
// If input < input & 0xff, it must be < 0, so saturate to 0.
Csel(output.W(), wzr, input.W(), lt);
- // Create a constant 0xff.
- Mov(WTmp0(), 255);
- // If input > input & 0xff, it must be > 255, so saturate to 255.
- Csel(output.W(), WTmp0(), output.W(), gt);
+ // If input <= input & 0xff, it must be <= 255. Otherwise, saturate to 255.
+ Csel(output.W(), output.W(), 255, le);
}
@@ -2246,37 +2279,37 @@ void MacroAssembler::CopyFieldsLoopPairsHelper(Register dst,
unsigned count,
Register scratch1,
Register scratch2,
- Register scratch3) {
+ Register scratch3,
+ Register scratch4,
+ Register scratch5) {
// Untag src and dst into scratch registers.
// Copy src->dst in a tight loop.
- ASSERT(!AreAliased(dst, src, scratch1, scratch2, scratch3, Tmp0(), Tmp1()));
+ ASSERT(!AreAliased(dst, src,
+ scratch1, scratch2, scratch3, scratch4, scratch5));
ASSERT(count >= 2);
const Register& remaining = scratch3;
Mov(remaining, count / 2);
- // Only use the Assembler, so we can use Tmp0() and Tmp1().
- InstructionAccurateScope scope(this);
-
const Register& dst_untagged = scratch1;
const Register& src_untagged = scratch2;
- sub(dst_untagged, dst, kHeapObjectTag);
- sub(src_untagged, src, kHeapObjectTag);
+ Sub(dst_untagged, dst, kHeapObjectTag);
+ Sub(src_untagged, src, kHeapObjectTag);
// Copy fields in pairs.
Label loop;
- bind(&loop);
- ldp(Tmp0(), Tmp1(), MemOperand(src_untagged, kXRegSizeInBytes * 2,
- PostIndex));
- stp(Tmp0(), Tmp1(), MemOperand(dst_untagged, kXRegSizeInBytes * 2,
- PostIndex));
- sub(remaining, remaining, 1);
- cbnz(remaining, &loop);
+ Bind(&loop);
+ Ldp(scratch4, scratch5,
+ MemOperand(src_untagged, kXRegSizeInBytes * 2, PostIndex));
+ Stp(scratch4, scratch5,
+ MemOperand(dst_untagged, kXRegSizeInBytes * 2, PostIndex));
+ Sub(remaining, remaining, 1);
+ Cbnz(remaining, &loop);
// Handle the leftovers.
if (count & 1) {
- ldr(Tmp0(), MemOperand(src_untagged));
- str(Tmp0(), MemOperand(dst_untagged));
+ Ldr(scratch4, MemOperand(src_untagged));
+ Str(scratch4, MemOperand(dst_untagged));
}
}
@@ -2285,13 +2318,12 @@ void MacroAssembler::CopyFieldsUnrolledPairsHelper(Register dst,
Register src,
unsigned count,
Register scratch1,
- Register scratch2) {
+ Register scratch2,
+ Register scratch3,
+ Register scratch4) {
// Untag src and dst into scratch registers.
// Copy src->dst in an unrolled loop.
- ASSERT(!AreAliased(dst, src, scratch1, scratch2, Tmp0(), Tmp1()));
-
- // Only use the Assembler, so we can use Tmp0() and Tmp1().
- InstructionAccurateScope scope(this);
+ ASSERT(!AreAliased(dst, src, scratch1, scratch2, scratch3, scratch4));
const Register& dst_untagged = scratch1;
const Register& src_untagged = scratch2;
@@ -2300,16 +2332,16 @@ void MacroAssembler::CopyFieldsUnrolledPairsHelper(Register dst,
// Copy fields in pairs.
for (unsigned i = 0; i < count / 2; i++) {
- ldp(Tmp0(), Tmp1(), MemOperand(src_untagged, kXRegSizeInBytes * 2,
- PostIndex));
- stp(Tmp0(), Tmp1(), MemOperand(dst_untagged, kXRegSizeInBytes * 2,
- PostIndex));
+ Ldp(scratch3, scratch4,
+ MemOperand(src_untagged, kXRegSizeInBytes * 2, PostIndex));
+ Stp(scratch3, scratch4,
+ MemOperand(dst_untagged, kXRegSizeInBytes * 2, PostIndex));
}
// Handle the leftovers.
if (count & 1) {
- ldr(Tmp0(), MemOperand(src_untagged));
- str(Tmp0(), MemOperand(dst_untagged));
+ Ldr(scratch3, MemOperand(src_untagged));
+ Str(scratch3, MemOperand(dst_untagged));
}
}
@@ -2317,23 +2349,22 @@ void MacroAssembler::CopyFieldsUnrolledPairsHelper(Register dst,
void MacroAssembler::CopyFieldsUnrolledHelper(Register dst,
Register src,
unsigned count,
- Register scratch1) {
+ Register scratch1,
+ Register scratch2,
+ Register scratch3) {
// Untag src and dst into scratch registers.
// Copy src->dst in an unrolled loop.
- ASSERT(!AreAliased(dst, src, scratch1, Tmp0(), Tmp1()));
-
- // Only use the Assembler, so we can use Tmp0() and Tmp1().
- InstructionAccurateScope scope(this);
+ ASSERT(!AreAliased(dst, src, scratch1, scratch2, scratch3));
const Register& dst_untagged = scratch1;
- const Register& src_untagged = Tmp1();
- sub(dst_untagged, dst, kHeapObjectTag);
- sub(src_untagged, src, kHeapObjectTag);
+ const Register& src_untagged = scratch2;
+ Sub(dst_untagged, dst, kHeapObjectTag);
+ Sub(src_untagged, src, kHeapObjectTag);
// Copy fields one by one.
for (unsigned i = 0; i < count; i++) {
- ldr(Tmp0(), MemOperand(src_untagged, kXRegSizeInBytes, PostIndex));
- str(Tmp0(), MemOperand(dst_untagged, kXRegSizeInBytes, PostIndex));
+ Ldr(scratch3, MemOperand(src_untagged, kXRegSizeInBytes, PostIndex));
+ Str(scratch3, MemOperand(dst_untagged, kXRegSizeInBytes, PostIndex));
}
}
@@ -2352,12 +2383,10 @@ void MacroAssembler::CopyFields(Register dst, Register src, CPURegList temps,
//
// In both cases, fields are copied in pairs if possible, and left-overs are
// handled separately.
+ ASSERT(!AreAliased(dst, src));
ASSERT(!temps.IncludesAliasOf(dst));
ASSERT(!temps.IncludesAliasOf(src));
- ASSERT(!temps.IncludesAliasOf(Tmp0()));
- ASSERT(!temps.IncludesAliasOf(Tmp1()));
ASSERT(!temps.IncludesAliasOf(xzr));
- ASSERT(!AreAliased(dst, src, Tmp0(), Tmp1()));
if (emit_debug_code()) {
Cmp(dst, src);
@@ -2368,17 +2397,25 @@ void MacroAssembler::CopyFields(Register dst, Register src, CPURegList temps,
// enough scratch registers).
static const unsigned kLoopThreshold = 8;
- ASSERT(!temps.IsEmpty());
- Register scratch1 = Register(temps.PopLowestIndex());
- Register scratch2 = Register(temps.PopLowestIndex());
- Register scratch3 = Register(temps.PopLowestIndex());
-
- if (scratch3.IsValid() && (count >= kLoopThreshold)) {
- CopyFieldsLoopPairsHelper(dst, src, count, scratch1, scratch2, scratch3);
- } else if (scratch2.IsValid()) {
- CopyFieldsUnrolledPairsHelper(dst, src, count, scratch1, scratch2);
- } else if (scratch1.IsValid()) {
- CopyFieldsUnrolledHelper(dst, src, count, scratch1);
+ UseScratchRegisterScope masm_temps(this);
+ if ((temps.Count() >= 3) && (count >= kLoopThreshold)) {
+ CopyFieldsLoopPairsHelper(dst, src, count,
+ Register(temps.PopLowestIndex()),
+ Register(temps.PopLowestIndex()),
+ Register(temps.PopLowestIndex()),
+ masm_temps.AcquireX(),
+ masm_temps.AcquireX());
+ } else if (temps.Count() >= 2) {
+ CopyFieldsUnrolledPairsHelper(dst, src, count,
+ Register(temps.PopLowestIndex()),
+ Register(temps.PopLowestIndex()),
+ masm_temps.AcquireX(),
+ masm_temps.AcquireX());
+ } else if (temps.Count() == 1) {
+ CopyFieldsUnrolledHelper(dst, src, count,
+ Register(temps.PopLowestIndex()),
+ masm_temps.AcquireX(),
+ masm_temps.AcquireX());
} else {
UNREACHABLE();
}
@@ -2808,12 +2845,14 @@ void MacroAssembler::TruncateHeapNumberToI(Register result,
void MacroAssembler::Prologue(PrologueFrameMode frame_mode) {
if (frame_mode == BUILD_STUB_FRAME) {
ASSERT(StackPointer().Is(jssp));
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
// TODO(jbramley): Does x1 contain a JSFunction here, or does it already
// have the special STUB smi?
- __ Mov(Tmp0(), Operand(Smi::FromInt(StackFrame::STUB)));
+ __ Mov(temp, Operand(Smi::FromInt(StackFrame::STUB)));
// Compiled stubs don't age, and so they don't need the predictable code
// ageing sequence.
- __ Push(lr, fp, cp, Tmp0());
+ __ Push(lr, fp, cp, temp);
__ Add(fp, jssp, StandardFrameConstants::kFixedFrameSizeFromFp);
} else {
if (isolate()->IsCodePreAgingActive()) {
@@ -2828,10 +2867,14 @@ void MacroAssembler::Prologue(PrologueFrameMode frame_mode) {
void MacroAssembler::EnterFrame(StackFrame::Type type) {
ASSERT(jssp.Is(StackPointer()));
+ UseScratchRegisterScope temps(this);
+ Register type_reg = temps.AcquireX();
+ Register code_reg = temps.AcquireX();
+
Push(lr, fp, cp);
- Mov(Tmp1(), Operand(Smi::FromInt(type)));
- Mov(Tmp0(), Operand(CodeObject()));
- Push(Tmp1(), Tmp0());
+ Mov(type_reg, Operand(Smi::FromInt(type)));
+ Mov(code_reg, Operand(CodeObject()));
+ Push(type_reg, code_reg);
// jssp[4] : lr
// jssp[3] : fp
// jssp[2] : cp
@@ -2839,7 +2882,7 @@ void MacroAssembler::EnterFrame(StackFrame::Type type) {
// jssp[0] : code object
// Adjust FP to point to saved FP.
- add(fp, jssp, StandardFrameConstants::kFixedFrameSizeFromFp + kPointerSize);
+ Add(fp, jssp, StandardFrameConstants::kFixedFrameSizeFromFp + kPointerSize);
}
@@ -3105,9 +3148,11 @@ void MacroAssembler::Allocate(int object_size,
return;
}
- ASSERT(!AreAliased(result, scratch1, scratch2, Tmp0(), Tmp1()));
- ASSERT(result.Is64Bits() && scratch1.Is64Bits() && scratch2.Is64Bits() &&
- Tmp0().Is64Bits() && Tmp1().Is64Bits());
+ UseScratchRegisterScope temps(this);
+ Register scratch3 = temps.AcquireX();
+
+ ASSERT(!AreAliased(result, scratch1, scratch2, scratch3));
+ ASSERT(result.Is64Bits() && scratch1.Is64Bits() && scratch2.Is64Bits());
// Make object size into bytes.
if ((flags & SIZE_IN_WORDS) != 0) {
@@ -3136,8 +3181,8 @@ void MacroAssembler::Allocate(int object_size,
} else {
if (emit_debug_code()) {
// Assert that result actually contains top on entry.
- Ldr(Tmp0(), MemOperand(top_address));
- Cmp(result, Tmp0());
+ Ldr(scratch3, MemOperand(top_address));
+ Cmp(result, scratch3);
Check(eq, kUnexpectedAllocationTop);
}
// Load the allocation limit. 'result' already contains the allocation top.
@@ -3149,11 +3194,11 @@ void MacroAssembler::Allocate(int object_size,
STATIC_ASSERT(kPointerAlignment == kDoubleAlignment);
// Calculate new top and bail out if new space is exhausted.
- Adds(Tmp1(), result, object_size);
+ Adds(scratch3, result, object_size);
B(vs, gc_required);
- Cmp(Tmp1(), allocation_limit);
+ Cmp(scratch3, allocation_limit);
B(hi, gc_required);
- Str(Tmp1(), MemOperand(top_address));
+ Str(scratch3, MemOperand(top_address));
// Tag the object if requested.
if ((flags & TAG_OBJECT) != 0) {
@@ -3180,9 +3225,12 @@ void MacroAssembler::Allocate(Register object_size,
return;
}
- ASSERT(!AreAliased(object_size, result, scratch1, scratch2, Tmp0(), Tmp1()));
- ASSERT(object_size.Is64Bits() && result.Is64Bits() && scratch1.Is64Bits() &&
- scratch2.Is64Bits() && Tmp0().Is64Bits() && Tmp1().Is64Bits());
+ UseScratchRegisterScope temps(this);
+ Register scratch3 = temps.AcquireX();
+
+ ASSERT(!AreAliased(object_size, result, scratch1, scratch2, scratch3));
+ ASSERT(object_size.Is64Bits() && result.Is64Bits() &&
+ scratch1.Is64Bits() && scratch2.Is64Bits());
// Check relative positions of allocation top and limit addresses.
// The values must be adjacent in memory to allow the use of LDP.
@@ -3205,8 +3253,8 @@ void MacroAssembler::Allocate(Register object_size,
} else {
if (emit_debug_code()) {
// Assert that result actually contains top on entry.
- Ldr(Tmp0(), MemOperand(top_address));
- Cmp(result, Tmp0());
+ Ldr(scratch3, MemOperand(top_address));
+ Cmp(result, scratch3);
Check(eq, kUnexpectedAllocationTop);
}
// Load the allocation limit. 'result' already contains the allocation top.
@@ -3219,20 +3267,20 @@ void MacroAssembler::Allocate(Register object_size,
// Calculate new top and bail out if new space is exhausted
if ((flags & SIZE_IN_WORDS) != 0) {
- Adds(Tmp1(), result, Operand(object_size, LSL, kPointerSizeLog2));
+ Adds(scratch3, result, Operand(object_size, LSL, kPointerSizeLog2));
} else {
- Adds(Tmp1(), result, object_size);
+ Adds(scratch3, result, object_size);
}
if (emit_debug_code()) {
- Tst(Tmp1(), kObjectAlignmentMask);
+ Tst(scratch3, kObjectAlignmentMask);
Check(eq, kUnalignedAllocationInNewSpace);
}
B(vs, gc_required);
- Cmp(Tmp1(), allocation_limit);
+ Cmp(scratch3, allocation_limit);
B(hi, gc_required);
- Str(Tmp1(), MemOperand(top_address));
+ Str(scratch3, MemOperand(top_address));
// Tag the object if requested.
if ((flags & TAG_OBJECT) != 0) {
@@ -3560,9 +3608,11 @@ void MacroAssembler::DispatchMap(Register obj,
void MacroAssembler::TestMapBitfield(Register object, uint64_t mask) {
- Ldr(Tmp0(), FieldMemOperand(object, HeapObject::kMapOffset));
- Ldrb(Tmp0(), FieldMemOperand(Tmp0(), Map::kBitFieldOffset));
- Tst(Tmp0(), mask);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ Ldr(temp, FieldMemOperand(object, HeapObject::kMapOffset));
+ Ldrb(temp, FieldMemOperand(temp, Map::kBitFieldOffset));
+ Tst(temp, mask);
}
@@ -3634,9 +3684,11 @@ void MacroAssembler::TryGetFunctionPrototype(Register function,
void MacroAssembler::CompareRoot(const Register& obj,
Heap::RootListIndex index) {
- ASSERT(!AreAliased(obj, Tmp0()));
- LoadRoot(Tmp0(), index);
- Cmp(obj, Tmp0());
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ ASSERT(!AreAliased(obj, temp));
+ LoadRoot(temp, index);
+ Cmp(obj, temp);
}
@@ -3833,60 +3885,53 @@ void MacroAssembler::EmitSeqStringSetCharCheck(
void MacroAssembler::CheckAccessGlobalProxy(Register holder_reg,
- Register scratch,
+ Register scratch1,
+ Register scratch2,
Label* miss) {
- // TODO(jbramley): Sort out the uses of Tmp0() and Tmp1() in this function.
- // The ARM version takes two scratch registers, and that should be enough for
- // all of the checks.
-
+ ASSERT(!AreAliased(holder_reg, scratch1, scratch2));
Label same_contexts;
- ASSERT(!AreAliased(holder_reg, scratch));
-
// Load current lexical context from the stack frame.
- Ldr(scratch, MemOperand(fp, StandardFrameConstants::kContextOffset));
+ Ldr(scratch1, MemOperand(fp, StandardFrameConstants::kContextOffset));
// In debug mode, make sure the lexical context is set.
#ifdef DEBUG
- Cmp(scratch, 0);
+ Cmp(scratch1, 0);
Check(ne, kWeShouldNotHaveAnEmptyLexicalContext);
#endif
// Load the native context of the current context.
int offset =
Context::kHeaderSize + Context::GLOBAL_OBJECT_INDEX * kPointerSize;
- Ldr(scratch, FieldMemOperand(scratch, offset));
- Ldr(scratch, FieldMemOperand(scratch, GlobalObject::kNativeContextOffset));
+ Ldr(scratch1, FieldMemOperand(scratch1, offset));
+ Ldr(scratch1, FieldMemOperand(scratch1, GlobalObject::kNativeContextOffset));
// Check the context is a native context.
if (emit_debug_code()) {
// Read the first word and compare to the global_context_map.
- Register temp = Tmp1();
- Ldr(temp, FieldMemOperand(scratch, HeapObject::kMapOffset));
- CompareRoot(temp, Heap::kNativeContextMapRootIndex);
+ Ldr(scratch2, FieldMemOperand(scratch1, HeapObject::kMapOffset));
+ CompareRoot(scratch2, Heap::kNativeContextMapRootIndex);
Check(eq, kExpectedNativeContext);
}
// Check if both contexts are the same.
- ldr(Tmp0(), FieldMemOperand(holder_reg, JSGlobalProxy::kNativeContextOffset));
- cmp(scratch, Tmp0());
- b(&same_contexts, eq);
+ Ldr(scratch2, FieldMemOperand(holder_reg,
+ JSGlobalProxy::kNativeContextOffset));
+ Cmp(scratch1, scratch2);
+ B(&same_contexts, eq);
// Check the context is a native context.
if (emit_debug_code()) {
- // Move Tmp0() into a different register, as CompareRoot will use it.
- Register temp = Tmp1();
- mov(temp, Tmp0());
- CompareRoot(temp, Heap::kNullValueRootIndex);
+ // We're short on scratch registers here, so use holder_reg as a scratch.
+ Push(holder_reg);
+ Register scratch3 = holder_reg;
+
+ CompareRoot(scratch2, Heap::kNullValueRootIndex);
Check(ne, kExpectedNonNullContext);
- Ldr(temp, FieldMemOperand(temp, HeapObject::kMapOffset));
- CompareRoot(temp, Heap::kNativeContextMapRootIndex);
+ Ldr(scratch3, FieldMemOperand(scratch2, HeapObject::kMapOffset));
+ CompareRoot(scratch3, Heap::kNativeContextMapRootIndex);
Check(eq, kExpectedNativeContext);
-
- // Let's consider that Tmp0() has been cloberred by the MacroAssembler.
- // We reload it with its value.
- ldr(Tmp0(), FieldMemOperand(holder_reg,
- JSGlobalProxy::kNativeContextOffset));
+ Pop(holder_reg);
}
// Check that the security token in the calling global object is
@@ -3895,12 +3940,12 @@ void MacroAssembler::CheckAccessGlobalProxy(Register holder_reg,
int token_offset = Context::kHeaderSize +
Context::SECURITY_TOKEN_INDEX * kPointerSize;
- ldr(scratch, FieldMemOperand(scratch, token_offset));
- ldr(Tmp0(), FieldMemOperand(Tmp0(), token_offset));
- cmp(scratch, Tmp0());
- b(miss, ne);
+ Ldr(scratch1, FieldMemOperand(scratch1, token_offset));
+ Ldr(scratch2, FieldMemOperand(scratch2, token_offset));
+ Cmp(scratch1, scratch2);
+ B(miss, ne);
- bind(&same_contexts);
+ Bind(&same_contexts);
}
@@ -4003,10 +4048,10 @@ void MacroAssembler::LoadFromNumberDictionary(Label* miss,
void MacroAssembler::RememberedSetHelper(Register object, // For debug tests.
Register address,
- Register scratch,
+ Register scratch1,
SaveFPRegsMode fp_mode,
RememberedSetFinalAction and_then) {
- ASSERT(!AreAliased(object, address, scratch));
+ ASSERT(!AreAliased(object, address, scratch1));
Label done, store_buffer_overflow;
if (emit_debug_code()) {
Label ok;
@@ -4014,22 +4059,25 @@ void MacroAssembler::RememberedSetHelper(Register object, // For debug tests.
Abort(kRememberedSetPointerInNewSpace);
bind(&ok);
}
+ UseScratchRegisterScope temps(this);
+ Register scratch2 = temps.AcquireX();
+
// Load store buffer top.
- Mov(Tmp0(), Operand(ExternalReference::store_buffer_top(isolate())));
- Ldr(scratch, MemOperand(Tmp0()));
+ Mov(scratch2, Operand(ExternalReference::store_buffer_top(isolate())));
+ Ldr(scratch1, MemOperand(scratch2));
// Store pointer to buffer and increment buffer top.
- Str(address, MemOperand(scratch, kPointerSize, PostIndex));
+ Str(address, MemOperand(scratch1, kPointerSize, PostIndex));
// Write back new top of buffer.
- Str(scratch, MemOperand(Tmp0()));
+ Str(scratch1, MemOperand(scratch2));
// Call stub on end of buffer.
// Check for end of buffer.
ASSERT(StoreBuffer::kStoreBufferOverflowBit ==
(1 << (14 + kPointerSizeLog2)));
if (and_then == kFallThroughAtEnd) {
- Tbz(scratch, (14 + kPointerSizeLog2), &done);
+ Tbz(scratch1, (14 + kPointerSizeLog2), &done);
} else {
ASSERT(and_then == kReturnAtEnd);
- Tbnz(scratch, (14 + kPointerSizeLog2), &store_buffer_overflow);
+ Tbnz(scratch1, (14 + kPointerSizeLog2), &store_buffer_overflow);
Ret();
}
@@ -4177,7 +4225,7 @@ void MacroAssembler::RecordWriteField(
}
-// Will clobber: object, address, value, Tmp0(), Tmp1().
+// Will clobber: object, address, value.
// If lr_status is kLRHasBeenSaved, lr will also be clobbered.
//
// The register 'object' contains a heap object pointer. The heap object tag is
@@ -4193,8 +4241,11 @@ void MacroAssembler::RecordWrite(Register object,
ASSERT(!AreAliased(object, value));
if (emit_debug_code()) {
- Ldr(Tmp0(), MemOperand(address));
- Cmp(Tmp0(), value);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+
+ Ldr(temp, MemOperand(address));
+ Cmp(temp, value);
Check(eq, kWrongAddressOrValuePassedToRecordWrite);
}
@@ -4259,15 +4310,18 @@ void MacroAssembler::AssertHasValidColor(const Register& reg) {
void MacroAssembler::GetMarkBits(Register addr_reg,
Register bitmap_reg,
Register shift_reg) {
- ASSERT(!AreAliased(addr_reg, bitmap_reg, shift_reg, no_reg));
+ ASSERT(!AreAliased(addr_reg, bitmap_reg, shift_reg));
+ ASSERT(addr_reg.Is64Bits() && bitmap_reg.Is64Bits() && shift_reg.Is64Bits());
// addr_reg is divided into fields:
// |63 page base 20|19 high 8|7 shift 3|2 0|
// 'high' gives the index of the cell holding color bits for the object.
// 'shift' gives the offset in the cell for this object's color.
const int kShiftBits = kPointerSizeLog2 + Bitmap::kBitsPerCellLog2;
- Ubfx(Tmp0(), addr_reg, kShiftBits, kPageSizeBits - kShiftBits);
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
+ Ubfx(temp, addr_reg, kShiftBits, kPageSizeBits - kShiftBits);
Bic(bitmap_reg, addr_reg, Page::kPageAlignmentMask);
- Add(bitmap_reg, bitmap_reg, Operand(Tmp0(), LSL, Bitmap::kBytesPerCellLog2));
+ Add(bitmap_reg, bitmap_reg, Operand(temp, LSL, Bitmap::kBytesPerCellLog2));
// bitmap_reg:
// |63 page base 20|19 zeros 15|14 high 3|2 0|
Ubfx(shift_reg, addr_reg, kPointerSizeLog2, Bitmap::kBitsPerCellLog2);
@@ -4491,8 +4545,6 @@ void MacroAssembler::AssertRegisterIsClear(Register reg, BailoutReason reason) {
void MacroAssembler::AssertRegisterIsRoot(Register reg,
Heap::RootListIndex index,
BailoutReason reason) {
- // CompareRoot uses Tmp0().
- ASSERT(!reg.Is(Tmp0()));
if (emit_debug_code()) {
CompareRoot(reg, index);
Check(eq, reason);
@@ -4502,7 +4554,8 @@ void MacroAssembler::AssertRegisterIsRoot(Register reg,
void MacroAssembler::AssertFastElements(Register elements) {
if (emit_debug_code()) {
- Register temp = Tmp1();
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
Label ok;
Ldr(temp, FieldMemOperand(elements, HeapObject::kMapOffset));
JumpIfRoot(temp, Heap::kFixedArrayMapRootIndex, &ok);
@@ -4516,7 +4569,8 @@ void MacroAssembler::AssertFastElements(Register elements) {
void MacroAssembler::AssertIsString(const Register& object) {
if (emit_debug_code()) {
- Register temp = Tmp1();
+ UseScratchRegisterScope temps(this);
+ Register temp = temps.AcquireX();
STATIC_ASSERT(kSmiTag == 0);
Tst(object, Operand(kSmiTagMask));
Check(ne, kOperandIsNotAString);
@@ -4563,6 +4617,12 @@ void MacroAssembler::Abort(BailoutReason reason) {
SetStackPointer(jssp);
Mov(jssp, old_stack_pointer);
+ // We need some scratch registers for the MacroAssembler, so make sure we have
+ // some. This is safe here because Abort never returns.
+ RegList old_tmp_list = TmpList()->list();
+ TmpList()->Combine(ip0);
+ TmpList()->Combine(ip1);
+
if (use_real_aborts()) {
// Avoid infinite recursion; Push contains some assertions that use Abort.
NoUseRealAbortsScope no_real_aborts(this);
@@ -4599,6 +4659,7 @@ void MacroAssembler::Abort(BailoutReason reason) {
}
SetStackPointer(old_stack_pointer);
+ TmpList()->set_list(old_tmp_list);
}
@@ -4606,22 +4667,23 @@ void MacroAssembler::LoadTransitionedArrayMapConditional(
ElementsKind expected_kind,
ElementsKind transitioned_kind,
Register map_in_out,
- Register scratch,
+ Register scratch1,
+ Register scratch2,
Label* no_map_match) {
// Load the global or builtins object from the current context.
- Ldr(scratch, GlobalObjectMemOperand());
- Ldr(scratch, FieldMemOperand(scratch, GlobalObject::kNativeContextOffset));
+ Ldr(scratch1, GlobalObjectMemOperand());
+ Ldr(scratch1, FieldMemOperand(scratch1, GlobalObject::kNativeContextOffset));
// Check that the function's map is the same as the expected cached map.
- Ldr(scratch, ContextMemOperand(scratch, Context::JS_ARRAY_MAPS_INDEX));
+ Ldr(scratch1, ContextMemOperand(scratch1, Context::JS_ARRAY_MAPS_INDEX));
size_t offset = (expected_kind * kPointerSize) + FixedArrayBase::kHeaderSize;
- Ldr(Tmp0(), FieldMemOperand(scratch, offset));
- Cmp(map_in_out, Tmp0());
+ Ldr(scratch2, FieldMemOperand(scratch1, offset));
+ Cmp(map_in_out, scratch2);
B(ne, no_map_match);
// Use the transitioned cached map.
offset = (transitioned_kind * kPointerSize) + FixedArrayBase::kHeaderSize;
- Ldr(map_in_out, FieldMemOperand(scratch, offset));
+ Ldr(map_in_out, FieldMemOperand(scratch1, offset));
}
@@ -4663,14 +4725,18 @@ void MacroAssembler::PrintfNoPreserve(const char * format,
// in most cases anyway, so this restriction shouldn't be too serious.
ASSERT(!kCallerSaved.IncludesAliasOf(__ StackPointer()));
- // We cannot print Tmp0() or Tmp1() as they're used internally by the macro
- // assembler. We cannot print the stack pointer because it is typically used
- // to preserve caller-saved registers (using other Printf variants which
- // depend on this helper).
- ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg0));
- ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg1));
- ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg2));
- ASSERT(!AreAliased(Tmp0(), Tmp1(), StackPointer(), arg3));
+ // Make sure that the macro assembler doesn't try to use any of our arguments
+ // as scratch registers.
+ ASSERT(!TmpList()->IncludesAliasOf(arg0, arg1, arg2, arg3));
+ ASSERT(!FPTmpList()->IncludesAliasOf(arg0, arg1, arg2, arg3));
+
+ // We cannot print the stack pointer because it is typically used to preserve
+ // caller-saved registers (using other Printf variants which depend on this
+ // helper).
+ ASSERT(!AreAliased(arg0, StackPointer()));
+ ASSERT(!AreAliased(arg1, StackPointer()));
+ ASSERT(!AreAliased(arg2, StackPointer()));
+ ASSERT(!AreAliased(arg3, StackPointer()));
static const int kMaxArgCount = 4;
// Assume that we have the maximum number of arguments until we know
@@ -4812,22 +4878,47 @@ void MacroAssembler::Printf(const char * format,
const CPURegister& arg1,
const CPURegister& arg2,
const CPURegister& arg3) {
+ // Printf is expected to preserve all registers, so make sure that none are
+ // available as scratch registers until we've preserved them.
+ RegList old_tmp_list = TmpList()->list();
+ RegList old_fp_tmp_list = FPTmpList()->list();
+ TmpList()->set_list(0);
+ FPTmpList()->set_list(0);
+
// Preserve all caller-saved registers as well as NZCV.
// If csp is the stack pointer, PushCPURegList asserts that the size of each
// list is a multiple of 16 bytes.
PushCPURegList(kCallerSaved);
PushCPURegList(kCallerSavedFP);
- // Use Tmp0() as a scratch register. It is not accepted by Printf so it will
- // never overlap an argument register.
- Mrs(Tmp0(), NZCV);
- Push(Tmp0(), xzr);
+
+ // We can use caller-saved registers as scratch values (except for argN).
+ CPURegList tmp_list = kCallerSaved;
+ CPURegList fp_tmp_list = kCallerSavedFP;
+ tmp_list.Remove(arg0, arg1, arg2, arg3);
+ fp_tmp_list.Remove(arg0, arg1, arg2, arg3);
+ TmpList()->set_list(tmp_list.list());
+ FPTmpList()->set_list(fp_tmp_list.list());
+
+ // Preserve NZCV.
+ { UseScratchRegisterScope temps(this);
+ Register tmp = temps.AcquireX();
+ Mrs(tmp, NZCV);
+ Push(tmp, xzr);
+ }
PrintfNoPreserve(format, arg0, arg1, arg2, arg3);
- Pop(xzr, Tmp0());
- Msr(NZCV, Tmp0());
+ { UseScratchRegisterScope temps(this);
+ Register tmp = temps.AcquireX();
+ Pop(xzr, tmp);
+ Msr(NZCV, tmp);
+ }
+
PopCPURegList(kCallerSavedFP);
PopCPURegList(kCallerSaved);
+
+ TmpList()->set_list(old_tmp_list);
+ FPTmpList()->set_list(old_fp_tmp_list);
}
@@ -4935,12 +5026,11 @@ bool MacroAssembler::IsCodeAgeSequence(byte* sequence) {
void MacroAssembler::FlooringDiv(Register result,
Register dividend,
int32_t divisor) {
- Register tmp = WTmp0();
- ASSERT(!AreAliased(result, dividend, tmp));
+ ASSERT(!AreAliased(result, dividend));
ASSERT(result.Is32Bits() && dividend.Is32Bits());
MultiplierAndShift ms(divisor);
- Mov(tmp, Operand(ms.multiplier()));
- Smull(result.X(), dividend, tmp);
+ Mov(result, Operand(ms.multiplier()));
+ Smull(result.X(), dividend, result);
Asr(result.X(), result.X(), 32);
if (divisor > 0 && ms.multiplier() < 0) Add(result, result, dividend);
if (divisor < 0 && ms.multiplier() > 0) Sub(result, result, dividend);
@@ -4949,6 +5039,35 @@ void MacroAssembler::FlooringDiv(Register result,
#undef __
+
+
+UseScratchRegisterScope::~UseScratchRegisterScope() {
+ available_->set_list(old_available_);
+ availablefp_->set_list(old_availablefp_);
+}
+
+
+Register UseScratchRegisterScope::AcquireSameSizeAs(const Register& reg) {
+ int code = AcquireNextAvailable(available_).code();
+ return Register::Create(code, reg.SizeInBits());
+}
+
+
+FPRegister UseScratchRegisterScope::AcquireSameSizeAs(const FPRegister& reg) {
+ int code = AcquireNextAvailable(availablefp_).code();
+ return FPRegister::Create(code, reg.SizeInBits());
+}
+
+
+CPURegister UseScratchRegisterScope::AcquireNextAvailable(
+ CPURegList* available) {
+ CHECK(!available->IsEmpty());
+ CPURegister result = available->PopLowestIndex();
+ ASSERT(!AreAliased(result, xzr, csp));
+ return result;
+}
+
+
#define __ masm->
« no previous file with comments | « src/a64/macro-assembler-a64.h ('k') | src/a64/macro-assembler-a64-inl.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698