Index: src/mips64/codegen-mips64.cc |
diff --git a/src/mips/codegen-mips.cc b/src/mips64/codegen-mips64.cc |
similarity index 58% |
copy from src/mips/codegen-mips.cc |
copy to src/mips64/codegen-mips64.cc |
index 6754c680c8e9e44916f13a27adc0a0b745ae69d3..abe2e471b84ab6bb2f65f121527913aa7c95143b 100644 |
--- a/src/mips/codegen-mips.cc |
+++ b/src/mips64/codegen-mips64.cc |
@@ -4,11 +4,11 @@ |
#include "src/v8.h" |
-#if V8_TARGET_ARCH_MIPS |
+#if V8_TARGET_ARCH_MIPS64 |
#include "src/codegen.h" |
#include "src/macro-assembler.h" |
-#include "src/mips/simulator-mips.h" |
+#include "src/mips64/simulator-mips64.h" |
namespace v8 { |
namespace internal { |
@@ -41,17 +41,25 @@ UnaryMathFunction CreateExpFunction() { |
DoubleRegister result = f0; |
DoubleRegister double_scratch1 = f4; |
DoubleRegister double_scratch2 = f6; |
- Register temp1 = t0; |
- Register temp2 = t1; |
- Register temp3 = t2; |
+ Register temp1 = a4; |
+ Register temp2 = a5; |
+ Register temp3 = a6; |
- __ MovFromFloatParameter(input); |
+ if (!IsMipsSoftFloatABI) { |
+ // Input value is in f12 anyway, nothing to do. |
+ } else { |
+ __ Move(input, a0, a1); |
+ } |
__ Push(temp3, temp2, temp1); |
MathExpGenerator::EmitMathExp( |
&masm, input, result, double_scratch1, double_scratch2, |
temp1, temp2, temp3); |
__ Pop(temp3, temp2, temp1); |
- __ MovToFloatResult(result); |
+ if (!IsMipsSoftFloatABI) { |
+ // Result is already in f0, nothing to do. |
+ } else { |
+ __ Move(v0, v1, result); |
+ } |
__ Ret(); |
} |
@@ -76,6 +84,7 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
#if defined(USE_SIMULATOR) |
return stub; |
#else |
+ |
size_t actual_size; |
byte* buffer = |
static_cast<byte*>(base::OS::Allocate(3 * KB, &actual_size, true)); |
@@ -113,11 +122,10 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
// using kPrefHintPrepareForStore. |
ASSERT(pref_hint_store != kPrefHintPrepareForStore || |
pref_chunk * 4 >= max_pref_size); |
- |
// If the size is less than 8, go to lastb. Regardless of size, |
// copy dst pointer to v0 for the retuen value. |
- __ slti(t2, a2, 2 * loadstore_chunk); |
- __ bne(t2, zero_reg, &lastb); |
+ __ slti(a6, a2, 2 * loadstore_chunk); |
+ __ bne(a6, zero_reg, &lastb); |
__ mov(v0, a0); // In delay slot. |
// If src and dst have different alignments, go to unaligned, if they |
@@ -133,17 +141,11 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ beq(a3, zero_reg, &aligned); // Already aligned. |
__ subu(a2, a2, a3); // In delay slot. a2 is the remining bytes count. |
- if (kArchEndian == kLittle) { |
- __ lwr(t8, MemOperand(a1)); |
- __ addu(a1, a1, a3); |
- __ swr(t8, MemOperand(a0)); |
- __ addu(a0, a0, a3); |
- } else { |
- __ lwl(t8, MemOperand(a1)); |
- __ addu(a1, a1, a3); |
- __ swl(t8, MemOperand(a0)); |
- __ addu(a0, a0, a3); |
- } |
+ __ lwr(t8, MemOperand(a1)); |
+ __ addu(a1, a1, a3); |
+ __ swr(t8, MemOperand(a0)); |
+ __ addu(a0, a0, a3); |
+ |
// Now dst/src are both aligned to (word) aligned addresses. Set a2 to |
// count how many bytes we have to copy after all the 64 byte chunks are |
// copied and a3 to the dst pointer after all the 64 byte chunks have been |
@@ -155,13 +157,13 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ addu(a3, a0, a3); // Now a3 is the final dst after loop. |
// When in the loop we prefetch with kPrefHintPrepareForStore hint, |
- // in this case the a0+x should be past the "t0-32" address. This means: |
- // for x=128 the last "safe" a0 address is "t0-160". Alternatively, for |
- // x=64 the last "safe" a0 address is "t0-96". In the current version we |
- // will use "pref hint, 128(a0)", so "t0-160" is the limit. |
+ // in this case the a0+x should be past the "a4-32" address. This means: |
+ // for x=128 the last "safe" a0 address is "a4-160". Alternatively, for |
+ // x=64 the last "safe" a0 address is "a4-96". In the current version we |
+ // will use "pref hint, 128(a0)", so "a4-160" is the limit. |
if (pref_hint_store == kPrefHintPrepareForStore) { |
- __ addu(t0, a0, a2); // t0 is the "past the end" address. |
- __ Subu(t9, t0, pref_limit); // t9 is the "last safe pref" address. |
+ __ addu(a4, a0, a2); // a4 is the "past the end" address. |
+ __ Subu(t9, a4, pref_limit); // t9 is the "last safe pref" address. |
} |
__ Pref(pref_hint_load, MemOperand(a1, 0 * pref_chunk)); |
@@ -175,53 +177,53 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ Pref(pref_hint_store, MemOperand(a0, 3 * pref_chunk)); |
} |
__ bind(&loop16w); |
- __ lw(t0, MemOperand(a1)); |
+ __ lw(a4, MemOperand(a1)); |
if (pref_hint_store == kPrefHintPrepareForStore) { |
__ sltu(v1, t9, a0); // If a0 > t9, don't use next prefetch. |
__ Branch(USE_DELAY_SLOT, &skip_pref, gt, v1, Operand(zero_reg)); |
} |
- __ lw(t1, MemOperand(a1, 1, loadstore_chunk)); // Maybe in delay slot. |
+ __ lw(a5, MemOperand(a1, 1, loadstore_chunk)); // Maybe in delay slot. |
__ Pref(pref_hint_store, MemOperand(a0, 4 * pref_chunk)); |
__ Pref(pref_hint_store, MemOperand(a0, 5 * pref_chunk)); |
__ bind(&skip_pref); |
- __ lw(t2, MemOperand(a1, 2, loadstore_chunk)); |
- __ lw(t3, MemOperand(a1, 3, loadstore_chunk)); |
- __ lw(t4, MemOperand(a1, 4, loadstore_chunk)); |
- __ lw(t5, MemOperand(a1, 5, loadstore_chunk)); |
- __ lw(t6, MemOperand(a1, 6, loadstore_chunk)); |
- __ lw(t7, MemOperand(a1, 7, loadstore_chunk)); |
+ __ lw(a6, MemOperand(a1, 2, loadstore_chunk)); |
+ __ lw(a7, MemOperand(a1, 3, loadstore_chunk)); |
+ __ lw(t0, MemOperand(a1, 4, loadstore_chunk)); |
+ __ lw(t1, MemOperand(a1, 5, loadstore_chunk)); |
+ __ lw(t2, MemOperand(a1, 6, loadstore_chunk)); |
+ __ lw(t3, MemOperand(a1, 7, loadstore_chunk)); |
__ Pref(pref_hint_load, MemOperand(a1, 4 * pref_chunk)); |
- __ sw(t0, MemOperand(a0)); |
- __ sw(t1, MemOperand(a0, 1, loadstore_chunk)); |
- __ sw(t2, MemOperand(a0, 2, loadstore_chunk)); |
- __ sw(t3, MemOperand(a0, 3, loadstore_chunk)); |
- __ sw(t4, MemOperand(a0, 4, loadstore_chunk)); |
- __ sw(t5, MemOperand(a0, 5, loadstore_chunk)); |
- __ sw(t6, MemOperand(a0, 6, loadstore_chunk)); |
- __ sw(t7, MemOperand(a0, 7, loadstore_chunk)); |
- |
- __ lw(t0, MemOperand(a1, 8, loadstore_chunk)); |
- __ lw(t1, MemOperand(a1, 9, loadstore_chunk)); |
- __ lw(t2, MemOperand(a1, 10, loadstore_chunk)); |
- __ lw(t3, MemOperand(a1, 11, loadstore_chunk)); |
- __ lw(t4, MemOperand(a1, 12, loadstore_chunk)); |
- __ lw(t5, MemOperand(a1, 13, loadstore_chunk)); |
- __ lw(t6, MemOperand(a1, 14, loadstore_chunk)); |
- __ lw(t7, MemOperand(a1, 15, loadstore_chunk)); |
+ __ sw(a4, MemOperand(a0)); |
+ __ sw(a5, MemOperand(a0, 1, loadstore_chunk)); |
+ __ sw(a6, MemOperand(a0, 2, loadstore_chunk)); |
+ __ sw(a7, MemOperand(a0, 3, loadstore_chunk)); |
+ __ sw(t0, MemOperand(a0, 4, loadstore_chunk)); |
+ __ sw(t1, MemOperand(a0, 5, loadstore_chunk)); |
+ __ sw(t2, MemOperand(a0, 6, loadstore_chunk)); |
+ __ sw(t3, MemOperand(a0, 7, loadstore_chunk)); |
+ |
+ __ lw(a4, MemOperand(a1, 8, loadstore_chunk)); |
+ __ lw(a5, MemOperand(a1, 9, loadstore_chunk)); |
+ __ lw(a6, MemOperand(a1, 10, loadstore_chunk)); |
+ __ lw(a7, MemOperand(a1, 11, loadstore_chunk)); |
+ __ lw(t0, MemOperand(a1, 12, loadstore_chunk)); |
+ __ lw(t1, MemOperand(a1, 13, loadstore_chunk)); |
+ __ lw(t2, MemOperand(a1, 14, loadstore_chunk)); |
+ __ lw(t3, MemOperand(a1, 15, loadstore_chunk)); |
__ Pref(pref_hint_load, MemOperand(a1, 5 * pref_chunk)); |
- __ sw(t0, MemOperand(a0, 8, loadstore_chunk)); |
- __ sw(t1, MemOperand(a0, 9, loadstore_chunk)); |
- __ sw(t2, MemOperand(a0, 10, loadstore_chunk)); |
- __ sw(t3, MemOperand(a0, 11, loadstore_chunk)); |
- __ sw(t4, MemOperand(a0, 12, loadstore_chunk)); |
- __ sw(t5, MemOperand(a0, 13, loadstore_chunk)); |
- __ sw(t6, MemOperand(a0, 14, loadstore_chunk)); |
- __ sw(t7, MemOperand(a0, 15, loadstore_chunk)); |
+ __ sw(a4, MemOperand(a0, 8, loadstore_chunk)); |
+ __ sw(a5, MemOperand(a0, 9, loadstore_chunk)); |
+ __ sw(a6, MemOperand(a0, 10, loadstore_chunk)); |
+ __ sw(a7, MemOperand(a0, 11, loadstore_chunk)); |
+ __ sw(t0, MemOperand(a0, 12, loadstore_chunk)); |
+ __ sw(t1, MemOperand(a0, 13, loadstore_chunk)); |
+ __ sw(t2, MemOperand(a0, 14, loadstore_chunk)); |
+ __ sw(t3, MemOperand(a0, 15, loadstore_chunk)); |
__ addiu(a0, a0, 16 * loadstore_chunk); |
__ bne(a0, a3, &loop16w); |
__ addiu(a1, a1, 16 * loadstore_chunk); // In delay slot. |
@@ -235,23 +237,23 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ andi(t8, a2, 0x1f); |
__ beq(a2, t8, &chk1w); // Less than 32? |
__ nop(); // In delay slot. |
- __ lw(t0, MemOperand(a1)); |
- __ lw(t1, MemOperand(a1, 1, loadstore_chunk)); |
- __ lw(t2, MemOperand(a1, 2, loadstore_chunk)); |
- __ lw(t3, MemOperand(a1, 3, loadstore_chunk)); |
- __ lw(t4, MemOperand(a1, 4, loadstore_chunk)); |
- __ lw(t5, MemOperand(a1, 5, loadstore_chunk)); |
- __ lw(t6, MemOperand(a1, 6, loadstore_chunk)); |
- __ lw(t7, MemOperand(a1, 7, loadstore_chunk)); |
+ __ lw(a4, MemOperand(a1)); |
+ __ lw(a5, MemOperand(a1, 1, loadstore_chunk)); |
+ __ lw(a6, MemOperand(a1, 2, loadstore_chunk)); |
+ __ lw(a7, MemOperand(a1, 3, loadstore_chunk)); |
+ __ lw(t0, MemOperand(a1, 4, loadstore_chunk)); |
+ __ lw(t1, MemOperand(a1, 5, loadstore_chunk)); |
+ __ lw(t2, MemOperand(a1, 6, loadstore_chunk)); |
+ __ lw(t3, MemOperand(a1, 7, loadstore_chunk)); |
__ addiu(a1, a1, 8 * loadstore_chunk); |
- __ sw(t0, MemOperand(a0)); |
- __ sw(t1, MemOperand(a0, 1, loadstore_chunk)); |
- __ sw(t2, MemOperand(a0, 2, loadstore_chunk)); |
- __ sw(t3, MemOperand(a0, 3, loadstore_chunk)); |
- __ sw(t4, MemOperand(a0, 4, loadstore_chunk)); |
- __ sw(t5, MemOperand(a0, 5, loadstore_chunk)); |
- __ sw(t6, MemOperand(a0, 6, loadstore_chunk)); |
- __ sw(t7, MemOperand(a0, 7, loadstore_chunk)); |
+ __ sw(a4, MemOperand(a0)); |
+ __ sw(a5, MemOperand(a0, 1, loadstore_chunk)); |
+ __ sw(a6, MemOperand(a0, 2, loadstore_chunk)); |
+ __ sw(a7, MemOperand(a0, 3, loadstore_chunk)); |
+ __ sw(t0, MemOperand(a0, 4, loadstore_chunk)); |
+ __ sw(t1, MemOperand(a0, 5, loadstore_chunk)); |
+ __ sw(t2, MemOperand(a0, 6, loadstore_chunk)); |
+ __ sw(t3, MemOperand(a0, 7, loadstore_chunk)); |
__ addiu(a0, a0, 8 * loadstore_chunk); |
// Here we have less than 32 bytes to copy. Set up for a loop to copy |
@@ -266,11 +268,11 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ addu(a3, a0, a3); |
__ bind(&wordCopy_loop); |
- __ lw(t3, MemOperand(a1)); |
+ __ lw(a7, MemOperand(a1)); |
__ addiu(a0, a0, loadstore_chunk); |
__ addiu(a1, a1, loadstore_chunk); |
__ bne(a0, a3, &wordCopy_loop); |
- __ sw(t3, MemOperand(a0, -1, loadstore_chunk)); // In delay slot. |
+ __ sw(a7, MemOperand(a0, -1, loadstore_chunk)); // In delay slot. |
__ bind(&lastb); |
__ Branch(&leave, le, a2, Operand(zero_reg)); |
@@ -295,21 +297,12 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ beq(a3, zero_reg, &ua_chk16w); |
__ subu(a2, a2, a3); // In delay slot. |
- if (kArchEndian == kLittle) { |
- __ lwr(v1, MemOperand(a1)); |
- __ lwl(v1, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ addu(a1, a1, a3); |
- __ swr(v1, MemOperand(a0)); |
- __ addu(a0, a0, a3); |
- } else { |
- __ lwl(v1, MemOperand(a1)); |
- __ lwr(v1, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ addu(a1, a1, a3); |
- __ swl(v1, MemOperand(a0)); |
- __ addu(a0, a0, a3); |
- } |
+ __ lwr(v1, MemOperand(a1)); |
+ __ lwl(v1, |
+ MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ addu(a1, a1, a3); |
+ __ swr(v1, MemOperand(a0)); |
+ __ addu(a0, a0, a3); |
// Now the dst (but not the source) is aligned. Set a2 to count how many |
// bytes we have to copy after all the 64 byte chunks are copied and a3 to |
@@ -322,8 +315,8 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ addu(a3, a0, a3); |
if (pref_hint_store == kPrefHintPrepareForStore) { |
- __ addu(t0, a0, a2); |
- __ Subu(t9, t0, pref_limit); |
+ __ addu(a4, a0, a2); |
+ __ Subu(t9, a4, pref_limit); |
} |
__ Pref(pref_hint_load, MemOperand(a1, 0 * pref_chunk)); |
@@ -338,146 +331,82 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ bind(&ua_loop16w); |
__ Pref(pref_hint_load, MemOperand(a1, 3 * pref_chunk)); |
- if (kArchEndian == kLittle) { |
- __ lwr(t0, MemOperand(a1)); |
- __ lwr(t1, MemOperand(a1, 1, loadstore_chunk)); |
- __ lwr(t2, MemOperand(a1, 2, loadstore_chunk)); |
- |
- if (pref_hint_store == kPrefHintPrepareForStore) { |
- __ sltu(v1, t9, a0); |
- __ Branch(USE_DELAY_SLOT, &ua_skip_pref, gt, v1, Operand(zero_reg)); |
- } |
- __ lwr(t3, MemOperand(a1, 3, loadstore_chunk)); // Maybe in delay slot. |
- |
- __ Pref(pref_hint_store, MemOperand(a0, 4 * pref_chunk)); |
- __ Pref(pref_hint_store, MemOperand(a0, 5 * pref_chunk)); |
- |
- __ bind(&ua_skip_pref); |
- __ lwr(t4, MemOperand(a1, 4, loadstore_chunk)); |
- __ lwr(t5, MemOperand(a1, 5, loadstore_chunk)); |
- __ lwr(t6, MemOperand(a1, 6, loadstore_chunk)); |
- __ lwr(t7, MemOperand(a1, 7, loadstore_chunk)); |
- __ lwl(t0, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t1, |
- MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t2, |
- MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t3, |
- MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t4, |
- MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t5, |
- MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t6, |
- MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t7, |
- MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one)); |
- } else { |
- __ lwl(t0, MemOperand(a1)); |
- __ lwl(t1, MemOperand(a1, 1, loadstore_chunk)); |
- __ lwl(t2, MemOperand(a1, 2, loadstore_chunk)); |
- |
- if (pref_hint_store == kPrefHintPrepareForStore) { |
- __ sltu(v1, t9, a0); |
- __ Branch(USE_DELAY_SLOT, &ua_skip_pref, gt, v1, Operand(zero_reg)); |
- } |
- __ lwl(t3, MemOperand(a1, 3, loadstore_chunk)); // Maybe in delay slot. |
- |
- __ Pref(pref_hint_store, MemOperand(a0, 4 * pref_chunk)); |
- __ Pref(pref_hint_store, MemOperand(a0, 5 * pref_chunk)); |
- |
- __ bind(&ua_skip_pref); |
- __ lwl(t4, MemOperand(a1, 4, loadstore_chunk)); |
- __ lwl(t5, MemOperand(a1, 5, loadstore_chunk)); |
- __ lwl(t6, MemOperand(a1, 6, loadstore_chunk)); |
- __ lwl(t7, MemOperand(a1, 7, loadstore_chunk)); |
- __ lwr(t0, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t1, |
- MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t2, |
- MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t3, |
- MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t4, |
- MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t5, |
- MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t6, |
- MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t7, |
- MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwr(a4, MemOperand(a1)); |
+ __ lwr(a5, MemOperand(a1, 1, loadstore_chunk)); |
+ __ lwr(a6, MemOperand(a1, 2, loadstore_chunk)); |
+ |
+ if (pref_hint_store == kPrefHintPrepareForStore) { |
+ __ sltu(v1, t9, a0); |
+ __ Branch(USE_DELAY_SLOT, &ua_skip_pref, gt, v1, Operand(zero_reg)); |
} |
+ __ lwr(a7, MemOperand(a1, 3, loadstore_chunk)); // Maybe in delay slot. |
+ |
+ __ Pref(pref_hint_store, MemOperand(a0, 4 * pref_chunk)); |
+ __ Pref(pref_hint_store, MemOperand(a0, 5 * pref_chunk)); |
+ |
+ __ bind(&ua_skip_pref); |
+ __ lwr(t0, MemOperand(a1, 4, loadstore_chunk)); |
+ __ lwr(t1, MemOperand(a1, 5, loadstore_chunk)); |
+ __ lwr(t2, MemOperand(a1, 6, loadstore_chunk)); |
+ __ lwr(t3, MemOperand(a1, 7, loadstore_chunk)); |
+ __ lwl(a4, |
+ MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a5, |
+ MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a6, |
+ MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a7, |
+ MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t0, |
+ MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t1, |
+ MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t2, |
+ MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t3, |
+ MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one)); |
__ Pref(pref_hint_load, MemOperand(a1, 4 * pref_chunk)); |
- __ sw(t0, MemOperand(a0)); |
- __ sw(t1, MemOperand(a0, 1, loadstore_chunk)); |
- __ sw(t2, MemOperand(a0, 2, loadstore_chunk)); |
- __ sw(t3, MemOperand(a0, 3, loadstore_chunk)); |
- __ sw(t4, MemOperand(a0, 4, loadstore_chunk)); |
- __ sw(t5, MemOperand(a0, 5, loadstore_chunk)); |
- __ sw(t6, MemOperand(a0, 6, loadstore_chunk)); |
- __ sw(t7, MemOperand(a0, 7, loadstore_chunk)); |
- if (kArchEndian == kLittle) { |
- __ lwr(t0, MemOperand(a1, 8, loadstore_chunk)); |
- __ lwr(t1, MemOperand(a1, 9, loadstore_chunk)); |
- __ lwr(t2, MemOperand(a1, 10, loadstore_chunk)); |
- __ lwr(t3, MemOperand(a1, 11, loadstore_chunk)); |
- __ lwr(t4, MemOperand(a1, 12, loadstore_chunk)); |
- __ lwr(t5, MemOperand(a1, 13, loadstore_chunk)); |
- __ lwr(t6, MemOperand(a1, 14, loadstore_chunk)); |
- __ lwr(t7, MemOperand(a1, 15, loadstore_chunk)); |
- __ lwl(t0, |
- MemOperand(a1, 9, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t1, |
- MemOperand(a1, 10, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t2, |
- MemOperand(a1, 11, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t3, |
- MemOperand(a1, 12, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t4, |
- MemOperand(a1, 13, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t5, |
- MemOperand(a1, 14, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t6, |
- MemOperand(a1, 15, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t7, |
- MemOperand(a1, 16, loadstore_chunk, MemOperand::offset_minus_one)); |
- } else { |
- __ lwl(t0, MemOperand(a1, 8, loadstore_chunk)); |
- __ lwl(t1, MemOperand(a1, 9, loadstore_chunk)); |
- __ lwl(t2, MemOperand(a1, 10, loadstore_chunk)); |
- __ lwl(t3, MemOperand(a1, 11, loadstore_chunk)); |
- __ lwl(t4, MemOperand(a1, 12, loadstore_chunk)); |
- __ lwl(t5, MemOperand(a1, 13, loadstore_chunk)); |
- __ lwl(t6, MemOperand(a1, 14, loadstore_chunk)); |
- __ lwl(t7, MemOperand(a1, 15, loadstore_chunk)); |
- __ lwr(t0, |
- MemOperand(a1, 9, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t1, |
- MemOperand(a1, 10, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t2, |
- MemOperand(a1, 11, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t3, |
- MemOperand(a1, 12, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t4, |
- MemOperand(a1, 13, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t5, |
- MemOperand(a1, 14, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t6, |
- MemOperand(a1, 15, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t7, |
- MemOperand(a1, 16, loadstore_chunk, MemOperand::offset_minus_one)); |
- } |
+ __ sw(a4, MemOperand(a0)); |
+ __ sw(a5, MemOperand(a0, 1, loadstore_chunk)); |
+ __ sw(a6, MemOperand(a0, 2, loadstore_chunk)); |
+ __ sw(a7, MemOperand(a0, 3, loadstore_chunk)); |
+ __ sw(t0, MemOperand(a0, 4, loadstore_chunk)); |
+ __ sw(t1, MemOperand(a0, 5, loadstore_chunk)); |
+ __ sw(t2, MemOperand(a0, 6, loadstore_chunk)); |
+ __ sw(t3, MemOperand(a0, 7, loadstore_chunk)); |
+ __ lwr(a4, MemOperand(a1, 8, loadstore_chunk)); |
+ __ lwr(a5, MemOperand(a1, 9, loadstore_chunk)); |
+ __ lwr(a6, MemOperand(a1, 10, loadstore_chunk)); |
+ __ lwr(a7, MemOperand(a1, 11, loadstore_chunk)); |
+ __ lwr(t0, MemOperand(a1, 12, loadstore_chunk)); |
+ __ lwr(t1, MemOperand(a1, 13, loadstore_chunk)); |
+ __ lwr(t2, MemOperand(a1, 14, loadstore_chunk)); |
+ __ lwr(t3, MemOperand(a1, 15, loadstore_chunk)); |
+ __ lwl(a4, |
+ MemOperand(a1, 9, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a5, |
+ MemOperand(a1, 10, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a6, |
+ MemOperand(a1, 11, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a7, |
+ MemOperand(a1, 12, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t0, |
+ MemOperand(a1, 13, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t1, |
+ MemOperand(a1, 14, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t2, |
+ MemOperand(a1, 15, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t3, |
+ MemOperand(a1, 16, loadstore_chunk, MemOperand::offset_minus_one)); |
__ Pref(pref_hint_load, MemOperand(a1, 5 * pref_chunk)); |
- __ sw(t0, MemOperand(a0, 8, loadstore_chunk)); |
- __ sw(t1, MemOperand(a0, 9, loadstore_chunk)); |
- __ sw(t2, MemOperand(a0, 10, loadstore_chunk)); |
- __ sw(t3, MemOperand(a0, 11, loadstore_chunk)); |
- __ sw(t4, MemOperand(a0, 12, loadstore_chunk)); |
- __ sw(t5, MemOperand(a0, 13, loadstore_chunk)); |
- __ sw(t6, MemOperand(a0, 14, loadstore_chunk)); |
- __ sw(t7, MemOperand(a0, 15, loadstore_chunk)); |
+ __ sw(a4, MemOperand(a0, 8, loadstore_chunk)); |
+ __ sw(a5, MemOperand(a0, 9, loadstore_chunk)); |
+ __ sw(a6, MemOperand(a0, 10, loadstore_chunk)); |
+ __ sw(a7, MemOperand(a0, 11, loadstore_chunk)); |
+ __ sw(t0, MemOperand(a0, 12, loadstore_chunk)); |
+ __ sw(t1, MemOperand(a0, 13, loadstore_chunk)); |
+ __ sw(t2, MemOperand(a0, 14, loadstore_chunk)); |
+ __ sw(t3, MemOperand(a0, 15, loadstore_chunk)); |
__ addiu(a0, a0, 16 * loadstore_chunk); |
__ bne(a0, a3, &ua_loop16w); |
__ addiu(a1, a1, 16 * loadstore_chunk); // In delay slot. |
@@ -492,66 +421,39 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ beq(a2, t8, &ua_chk1w); |
__ nop(); // In delay slot. |
- if (kArchEndian == kLittle) { |
- __ lwr(t0, MemOperand(a1)); |
- __ lwr(t1, MemOperand(a1, 1, loadstore_chunk)); |
- __ lwr(t2, MemOperand(a1, 2, loadstore_chunk)); |
- __ lwr(t3, MemOperand(a1, 3, loadstore_chunk)); |
- __ lwr(t4, MemOperand(a1, 4, loadstore_chunk)); |
- __ lwr(t5, MemOperand(a1, 5, loadstore_chunk)); |
- __ lwr(t6, MemOperand(a1, 6, loadstore_chunk)); |
- __ lwr(t7, MemOperand(a1, 7, loadstore_chunk)); |
- __ lwl(t0, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t1, |
- MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t2, |
- MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t3, |
- MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t4, |
- MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t5, |
- MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t6, |
- MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwl(t7, |
- MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one)); |
- } else { |
- __ lwl(t0, MemOperand(a1)); |
- __ lwl(t1, MemOperand(a1, 1, loadstore_chunk)); |
- __ lwl(t2, MemOperand(a1, 2, loadstore_chunk)); |
- __ lwl(t3, MemOperand(a1, 3, loadstore_chunk)); |
- __ lwl(t4, MemOperand(a1, 4, loadstore_chunk)); |
- __ lwl(t5, MemOperand(a1, 5, loadstore_chunk)); |
- __ lwl(t6, MemOperand(a1, 6, loadstore_chunk)); |
- __ lwl(t7, MemOperand(a1, 7, loadstore_chunk)); |
- __ lwr(t0, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t1, |
- MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t2, |
- MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t3, |
- MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t4, |
- MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t5, |
- MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t6, |
- MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one)); |
- __ lwr(t7, |
- MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one)); |
- } |
+ __ lwr(a4, MemOperand(a1)); |
+ __ lwr(a5, MemOperand(a1, 1, loadstore_chunk)); |
+ __ lwr(a6, MemOperand(a1, 2, loadstore_chunk)); |
+ __ lwr(a7, MemOperand(a1, 3, loadstore_chunk)); |
+ __ lwr(t0, MemOperand(a1, 4, loadstore_chunk)); |
+ __ lwr(t1, MemOperand(a1, 5, loadstore_chunk)); |
+ __ lwr(t2, MemOperand(a1, 6, loadstore_chunk)); |
+ __ lwr(t3, MemOperand(a1, 7, loadstore_chunk)); |
+ __ lwl(a4, |
+ MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a5, |
+ MemOperand(a1, 2, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a6, |
+ MemOperand(a1, 3, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(a7, |
+ MemOperand(a1, 4, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t0, |
+ MemOperand(a1, 5, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t1, |
+ MemOperand(a1, 6, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t2, |
+ MemOperand(a1, 7, loadstore_chunk, MemOperand::offset_minus_one)); |
+ __ lwl(t3, |
+ MemOperand(a1, 8, loadstore_chunk, MemOperand::offset_minus_one)); |
__ addiu(a1, a1, 8 * loadstore_chunk); |
- __ sw(t0, MemOperand(a0)); |
- __ sw(t1, MemOperand(a0, 1, loadstore_chunk)); |
- __ sw(t2, MemOperand(a0, 2, loadstore_chunk)); |
- __ sw(t3, MemOperand(a0, 3, loadstore_chunk)); |
- __ sw(t4, MemOperand(a0, 4, loadstore_chunk)); |
- __ sw(t5, MemOperand(a0, 5, loadstore_chunk)); |
- __ sw(t6, MemOperand(a0, 6, loadstore_chunk)); |
- __ sw(t7, MemOperand(a0, 7, loadstore_chunk)); |
+ __ sw(a4, MemOperand(a0)); |
+ __ sw(a5, MemOperand(a0, 1, loadstore_chunk)); |
+ __ sw(a6, MemOperand(a0, 2, loadstore_chunk)); |
+ __ sw(a7, MemOperand(a0, 3, loadstore_chunk)); |
+ __ sw(t0, MemOperand(a0, 4, loadstore_chunk)); |
+ __ sw(t1, MemOperand(a0, 5, loadstore_chunk)); |
+ __ sw(t2, MemOperand(a0, 6, loadstore_chunk)); |
+ __ sw(t3, MemOperand(a0, 7, loadstore_chunk)); |
__ addiu(a0, a0, 8 * loadstore_chunk); |
// Less than 32 bytes to copy. Set up for a loop to |
@@ -563,15 +465,9 @@ MemCopyUint8Function CreateMemCopyUint8Function(MemCopyUint8Function stub) { |
__ addu(a3, a0, a3); |
__ bind(&ua_wordCopy_loop); |
- if (kArchEndian == kLittle) { |
- __ lwr(v1, MemOperand(a1)); |
- __ lwl(v1, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- } else { |
- __ lwl(v1, MemOperand(a1)); |
- __ lwr(v1, |
- MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
- } |
+ __ lwr(v1, MemOperand(a1)); |
+ __ lwl(v1, |
+ MemOperand(a1, 1, loadstore_chunk, MemOperand::offset_minus_one)); |
__ addiu(a0, a0, loadstore_chunk); |
__ addiu(a1, a1, loadstore_chunk); |
__ bne(a0, a3, &ua_wordCopy_loop); |
@@ -663,19 +559,19 @@ void ElementsTransitionGenerator::GenerateMapChangeElementsTransition( |
// -- a2 : receiver |
// -- ra : return address |
// -- a3 : target map, scratch for subsequent call |
- // -- t0 : scratch (elements) |
+ // -- a4 : scratch (elements) |
// ----------------------------------- |
if (mode == TRACK_ALLOCATION_SITE) { |
ASSERT(allocation_memento_found != NULL); |
- __ JumpIfJSArrayHasAllocationMemento(a2, t0, allocation_memento_found); |
+ __ JumpIfJSArrayHasAllocationMemento(a2, a4, allocation_memento_found); |
} |
// Set transitioned map. |
- __ sw(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
+ __ sd(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
__ RecordWriteField(a2, |
HeapObject::kMapOffset, |
a3, |
- t5, |
+ t1, |
kRAHasNotBeenSaved, |
kDontSaveFPRegs, |
EMIT_REMEMBERED_SET, |
@@ -691,55 +587,54 @@ void ElementsTransitionGenerator::GenerateSmiToDouble( |
// -- a2 : receiver |
// -- ra : return address |
// -- a3 : target map, scratch for subsequent call |
- // -- t0 : scratch (elements) |
+ // -- a4 : scratch (elements) |
// ----------------------------------- |
Label loop, entry, convert_hole, gc_required, only_change_map, done; |
- Register scratch = t6; |
- |
+ Register scratch = t2; |
if (mode == TRACK_ALLOCATION_SITE) { |
- __ JumpIfJSArrayHasAllocationMemento(a2, t0, fail); |
+ __ JumpIfJSArrayHasAllocationMemento(a2, a4, fail); |
} |
// Check for empty arrays, which only require a map transition and no changes |
// to the backing store. |
- __ lw(t0, FieldMemOperand(a2, JSObject::kElementsOffset)); |
+ __ ld(a4, FieldMemOperand(a2, JSObject::kElementsOffset)); |
__ LoadRoot(at, Heap::kEmptyFixedArrayRootIndex); |
- __ Branch(&only_change_map, eq, at, Operand(t0)); |
+ __ Branch(&only_change_map, eq, at, Operand(a4)); |
__ push(ra); |
- __ lw(t1, FieldMemOperand(t0, FixedArray::kLengthOffset)); |
- // t0: source FixedArray |
- // t1: number of elements (smi-tagged) |
+ __ ld(a5, FieldMemOperand(a4, FixedArray::kLengthOffset)); |
+ // a4: source FixedArray |
+ // a5: number of elements (smi-tagged) |
// Allocate new FixedDoubleArray. |
- __ sll(scratch, t1, 2); |
- __ Addu(scratch, scratch, FixedDoubleArray::kHeaderSize); |
- __ Allocate(scratch, t2, t3, t5, &gc_required, DOUBLE_ALIGNMENT); |
- // t2: destination FixedDoubleArray, not tagged as heap object |
+ __ SmiScale(scratch, a5, kDoubleSizeLog2); |
+ __ Daddu(scratch, scratch, FixedDoubleArray::kHeaderSize); |
+ __ Allocate(scratch, a6, a7, t1, &gc_required, DOUBLE_ALIGNMENT); |
+ // a6: destination FixedDoubleArray, not tagged as heap object |
// Set destination FixedDoubleArray's length and map. |
- __ LoadRoot(t5, Heap::kFixedDoubleArrayMapRootIndex); |
- __ sw(t1, MemOperand(t2, FixedDoubleArray::kLengthOffset)); |
- __ sw(t5, MemOperand(t2, HeapObject::kMapOffset)); |
+ __ LoadRoot(t1, Heap::kFixedDoubleArrayMapRootIndex); |
+ __ sd(a5, MemOperand(a6, FixedDoubleArray::kLengthOffset)); |
+ __ sd(t1, MemOperand(a6, HeapObject::kMapOffset)); |
// Update receiver's map. |
- __ sw(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
+ __ sd(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
__ RecordWriteField(a2, |
HeapObject::kMapOffset, |
a3, |
- t5, |
+ t1, |
kRAHasBeenSaved, |
kDontSaveFPRegs, |
OMIT_REMEMBERED_SET, |
OMIT_SMI_CHECK); |
// Replace receiver's backing store with newly created FixedDoubleArray. |
- __ Addu(a3, t2, Operand(kHeapObjectTag)); |
- __ sw(a3, FieldMemOperand(a2, JSObject::kElementsOffset)); |
+ __ Daddu(a3, a6, Operand(kHeapObjectTag)); |
+ __ sd(a3, FieldMemOperand(a2, JSObject::kElementsOffset)); |
__ RecordWriteField(a2, |
JSObject::kElementsOffset, |
a3, |
- t5, |
+ t1, |
kRAHasBeenSaved, |
kDontSaveFPRegs, |
EMIT_REMEMBERED_SET, |
@@ -747,25 +642,25 @@ void ElementsTransitionGenerator::GenerateSmiToDouble( |
// Prepare for conversion loop. |
- __ Addu(a3, t0, Operand(FixedArray::kHeaderSize - kHeapObjectTag)); |
- __ Addu(t3, t2, Operand(FixedDoubleArray::kHeaderSize)); |
- __ sll(t2, t1, 2); |
- __ Addu(t2, t2, t3); |
- __ li(t0, Operand(kHoleNanLower32)); |
- __ li(t1, Operand(kHoleNanUpper32)); |
- // t0: kHoleNanLower32 |
- // t1: kHoleNanUpper32 |
- // t2: end of destination FixedDoubleArray, not tagged |
- // t3: begin of FixedDoubleArray element fields, not tagged |
+ __ Daddu(a3, a4, Operand(FixedArray::kHeaderSize - kHeapObjectTag)); |
+ __ Daddu(a7, a6, Operand(FixedDoubleArray::kHeaderSize)); |
+ __ SmiScale(a6, a5, kDoubleSizeLog2); |
+ __ Daddu(a6, a6, a7); |
+ __ li(a4, Operand(kHoleNanLower32)); |
+ __ li(a5, Operand(kHoleNanUpper32)); |
+ // a4: kHoleNanLower32 |
+ // a5: kHoleNanUpper32 |
+ // a6: end of destination FixedDoubleArray, not tagged |
+ // a7: begin of FixedDoubleArray element fields, not tagged |
__ Branch(&entry); |
__ bind(&only_change_map); |
- __ sw(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
+ __ sd(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
__ RecordWriteField(a2, |
HeapObject::kMapOffset, |
a3, |
- t5, |
+ t1, |
kRAHasNotBeenSaved, |
kDontSaveFPRegs, |
OMIT_REMEMBERED_SET, |
@@ -779,16 +674,17 @@ void ElementsTransitionGenerator::GenerateSmiToDouble( |
// Convert and copy elements. |
__ bind(&loop); |
- __ lw(t5, MemOperand(a3)); |
- __ Addu(a3, a3, kIntSize); |
- // t5: current element |
- __ UntagAndJumpIfNotSmi(t5, t5, &convert_hole); |
+ __ ld(t1, MemOperand(a3)); |
+ __ Daddu(a3, a3, kIntSize); |
+ // t1: current element |
+ __ JumpIfNotSmi(t1, &convert_hole); |
+ __ SmiUntag(t1); |
// Normal smi, convert to double and store. |
- __ mtc1(t5, f0); |
+ __ mtc1(t1, f0); |
__ cvt_d_w(f0, f0); |
- __ sdc1(f0, MemOperand(t3)); |
- __ Addu(t3, t3, kDoubleSize); |
+ __ sdc1(f0, MemOperand(a7)); |
+ __ Daddu(a7, a7, kDoubleSize); |
__ Branch(&entry); |
@@ -796,17 +692,16 @@ void ElementsTransitionGenerator::GenerateSmiToDouble( |
__ bind(&convert_hole); |
if (FLAG_debug_code) { |
// Restore a "smi-untagged" heap object. |
- __ SmiTag(t5); |
- __ Or(t5, t5, Operand(1)); |
+ __ Or(t1, t1, Operand(1)); |
__ LoadRoot(at, Heap::kTheHoleValueRootIndex); |
- __ Assert(eq, kObjectFoundInSmiOnlyArray, at, Operand(t5)); |
+ __ Assert(eq, kObjectFoundInSmiOnlyArray, at, Operand(t1)); |
} |
- __ sw(t0, MemOperand(t3, Register::kMantissaOffset)); // mantissa |
- __ sw(t1, MemOperand(t3, Register::kExponentOffset)); // exponent |
- __ Addu(t3, t3, kDoubleSize); |
+ __ sw(a4, MemOperand(a7)); // mantissa |
+ __ sw(a5, MemOperand(a7, kIntSize)); // exponent |
+ __ Daddu(a7, a7, kDoubleSize); |
__ bind(&entry); |
- __ Branch(&loop, lt, t3, Operand(t2)); |
+ __ Branch(&loop, lt, a7, Operand(a6)); |
__ pop(ra); |
__ bind(&done); |
@@ -821,54 +716,50 @@ void ElementsTransitionGenerator::GenerateDoubleToObject( |
// -- a2 : receiver |
// -- ra : return address |
// -- a3 : target map, scratch for subsequent call |
- // -- t0 : scratch (elements) |
+ // -- a4 : scratch (elements) |
// ----------------------------------- |
Label entry, loop, convert_hole, gc_required, only_change_map; |
- |
if (mode == TRACK_ALLOCATION_SITE) { |
- __ JumpIfJSArrayHasAllocationMemento(a2, t0, fail); |
+ __ JumpIfJSArrayHasAllocationMemento(a2, a4, fail); |
} |
// Check for empty arrays, which only require a map transition and no changes |
// to the backing store. |
- __ lw(t0, FieldMemOperand(a2, JSObject::kElementsOffset)); |
+ __ ld(a4, FieldMemOperand(a2, JSObject::kElementsOffset)); |
__ LoadRoot(at, Heap::kEmptyFixedArrayRootIndex); |
- __ Branch(&only_change_map, eq, at, Operand(t0)); |
+ __ Branch(&only_change_map, eq, at, Operand(a4)); |
__ MultiPush(a0.bit() | a1.bit() | a2.bit() | a3.bit() | ra.bit()); |
- __ lw(t1, FieldMemOperand(t0, FixedArray::kLengthOffset)); |
- // t0: source FixedArray |
- // t1: number of elements (smi-tagged) |
+ __ ld(a5, FieldMemOperand(a4, FixedArray::kLengthOffset)); |
+ // a4: source FixedArray |
+ // a5: number of elements (smi-tagged) |
// Allocate new FixedArray. |
- __ sll(a0, t1, 1); |
- __ Addu(a0, a0, FixedDoubleArray::kHeaderSize); |
- __ Allocate(a0, t2, t3, t5, &gc_required, NO_ALLOCATION_FLAGS); |
- // t2: destination FixedArray, not tagged as heap object |
+ __ SmiScale(a0, a5, kPointerSizeLog2); |
+ __ Daddu(a0, a0, FixedDoubleArray::kHeaderSize); |
+ __ Allocate(a0, a6, a7, t1, &gc_required, NO_ALLOCATION_FLAGS); |
+ // a6: destination FixedArray, not tagged as heap object |
// Set destination FixedDoubleArray's length and map. |
- __ LoadRoot(t5, Heap::kFixedArrayMapRootIndex); |
- __ sw(t1, MemOperand(t2, FixedDoubleArray::kLengthOffset)); |
- __ sw(t5, MemOperand(t2, HeapObject::kMapOffset)); |
+ __ LoadRoot(t1, Heap::kFixedArrayMapRootIndex); |
+ __ sd(a5, MemOperand(a6, FixedDoubleArray::kLengthOffset)); |
+ __ sd(t1, MemOperand(a6, HeapObject::kMapOffset)); |
// Prepare for conversion loop. |
- __ Addu(t0, t0, Operand( |
- FixedDoubleArray::kHeaderSize - kHeapObjectTag |
- + Register::kExponentOffset)); |
- __ Addu(a3, t2, Operand(FixedArray::kHeaderSize)); |
- __ Addu(t2, t2, Operand(kHeapObjectTag)); |
- __ sll(t1, t1, 1); |
- __ Addu(t1, a3, t1); |
- __ LoadRoot(t3, Heap::kTheHoleValueRootIndex); |
- __ LoadRoot(t5, Heap::kHeapNumberMapRootIndex); |
+ __ Daddu(a4, a4, Operand(FixedDoubleArray::kHeaderSize - kHeapObjectTag + 4)); |
+ __ Daddu(a3, a6, Operand(FixedArray::kHeaderSize)); |
+ __ Daddu(a6, a6, Operand(kHeapObjectTag)); |
+ __ SmiScale(a5, a5, kPointerSizeLog2); |
+ __ Daddu(a5, a3, a5); |
+ __ LoadRoot(a7, Heap::kTheHoleValueRootIndex); |
+ __ LoadRoot(t1, Heap::kHeapNumberMapRootIndex); |
// Using offsetted addresses. |
// a3: begin of destination FixedArray element fields, not tagged |
- // t0: begin of source FixedDoubleArray element fields, not tagged, |
- // points to the exponent |
- // t1: end of destination FixedArray, not tagged |
- // t2: destination FixedArray |
- // t3: the-hole pointer |
- // t5: heap number map |
+ // a4: begin of source FixedDoubleArray element fields, not tagged, +4 |
+ // a5: end of destination FixedArray, not tagged |
+ // a6: destination FixedArray |
+ // a7: the-hole pointer |
+ // t1: heap number map |
__ Branch(&entry); |
// Call into runtime if GC is required. |
@@ -878,24 +769,22 @@ void ElementsTransitionGenerator::GenerateDoubleToObject( |
__ Branch(fail); |
__ bind(&loop); |
- __ lw(a1, MemOperand(t0)); |
- __ Addu(t0, t0, kDoubleSize); |
+ __ lw(a1, MemOperand(a4)); |
+ __ Daddu(a4, a4, kDoubleSize); |
// a1: current element's upper 32 bit |
- // t0: address of next element's upper 32 bit |
+ // a4: address of next element's upper 32 bit |
__ Branch(&convert_hole, eq, a1, Operand(kHoleNanUpper32)); |
// Non-hole double, copy value into a heap number. |
- __ AllocateHeapNumber(a2, a0, t6, t5, &gc_required); |
+ __ AllocateHeapNumber(a2, a0, t2, t1, &gc_required); |
// a2: new heap number |
- // Load mantissa of current element, t0 point to exponent of next element. |
- __ lw(a0, MemOperand(t0, (Register::kMantissaOffset |
- - Register::kExponentOffset - kDoubleSize))); |
+ __ lw(a0, MemOperand(a4, -12)); |
__ sw(a0, FieldMemOperand(a2, HeapNumber::kMantissaOffset)); |
__ sw(a1, FieldMemOperand(a2, HeapNumber::kExponentOffset)); |
__ mov(a0, a3); |
- __ sw(a2, MemOperand(a3)); |
- __ Addu(a3, a3, kIntSize); |
- __ RecordWrite(t2, |
+ __ sd(a2, MemOperand(a3)); |
+ __ Daddu(a3, a3, kPointerSize); |
+ __ RecordWrite(a6, |
a0, |
a2, |
kRAHasBeenSaved, |
@@ -906,19 +795,19 @@ void ElementsTransitionGenerator::GenerateDoubleToObject( |
// Replace the-hole NaN with the-hole pointer. |
__ bind(&convert_hole); |
- __ sw(t3, MemOperand(a3)); |
- __ Addu(a3, a3, kIntSize); |
+ __ sd(a7, MemOperand(a3)); |
+ __ Daddu(a3, a3, kPointerSize); |
__ bind(&entry); |
- __ Branch(&loop, lt, a3, Operand(t1)); |
+ __ Branch(&loop, lt, a3, Operand(a5)); |
__ MultiPop(a2.bit() | a3.bit() | a0.bit() | a1.bit()); |
// Replace receiver's backing store with newly created and filled FixedArray. |
- __ sw(t2, FieldMemOperand(a2, JSObject::kElementsOffset)); |
+ __ sd(a6, FieldMemOperand(a2, JSObject::kElementsOffset)); |
__ RecordWriteField(a2, |
JSObject::kElementsOffset, |
- t2, |
- t5, |
+ a6, |
+ t1, |
kRAHasBeenSaved, |
kDontSaveFPRegs, |
EMIT_REMEMBERED_SET, |
@@ -927,11 +816,11 @@ void ElementsTransitionGenerator::GenerateDoubleToObject( |
__ bind(&only_change_map); |
// Update receiver's map. |
- __ sw(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
+ __ sd(a3, FieldMemOperand(a2, HeapObject::kMapOffset)); |
__ RecordWriteField(a2, |
HeapObject::kMapOffset, |
a3, |
- t5, |
+ t1, |
kRAHasNotBeenSaved, |
kDontSaveFPRegs, |
OMIT_REMEMBERED_SET, |
@@ -945,7 +834,7 @@ void StringCharLoadGenerator::Generate(MacroAssembler* masm, |
Register result, |
Label* call_runtime) { |
// Fetch the instance type of the receiver into result register. |
- __ lw(result, FieldMemOperand(string, HeapObject::kMapOffset)); |
+ __ ld(result, FieldMemOperand(string, HeapObject::kMapOffset)); |
__ lbu(result, FieldMemOperand(result, Map::kInstanceTypeOffset)); |
// We need special handling for indirect strings. |
@@ -960,10 +849,10 @@ void StringCharLoadGenerator::Generate(MacroAssembler* masm, |
// Handle slices. |
Label indirect_string_loaded; |
- __ lw(result, FieldMemOperand(string, SlicedString::kOffsetOffset)); |
- __ lw(string, FieldMemOperand(string, SlicedString::kParentOffset)); |
- __ sra(at, result, kSmiTagSize); |
- __ Addu(index, index, at); |
+ __ ld(result, FieldMemOperand(string, SlicedString::kOffsetOffset)); |
+ __ ld(string, FieldMemOperand(string, SlicedString::kParentOffset)); |
+ __ dsra32(at, result, 0); |
+ __ Daddu(index, index, at); |
__ jmp(&indirect_string_loaded); |
// Handle cons strings. |
@@ -972,14 +861,14 @@ void StringCharLoadGenerator::Generate(MacroAssembler* masm, |
// the case we would rather go to the runtime system now to flatten |
// the string. |
__ bind(&cons_string); |
- __ lw(result, FieldMemOperand(string, ConsString::kSecondOffset)); |
+ __ ld(result, FieldMemOperand(string, ConsString::kSecondOffset)); |
__ LoadRoot(at, Heap::kempty_stringRootIndex); |
__ Branch(call_runtime, ne, result, Operand(at)); |
// Get the first of the two strings and load its instance type. |
- __ lw(string, FieldMemOperand(string, ConsString::kFirstOffset)); |
+ __ ld(string, FieldMemOperand(string, ConsString::kFirstOffset)); |
__ bind(&indirect_string_loaded); |
- __ lw(result, FieldMemOperand(string, HeapObject::kMapOffset)); |
+ __ ld(result, FieldMemOperand(string, HeapObject::kMapOffset)); |
__ lbu(result, FieldMemOperand(result, Map::kInstanceTypeOffset)); |
// Distinguish sequential and external strings. Only these two string |
@@ -993,7 +882,7 @@ void StringCharLoadGenerator::Generate(MacroAssembler* masm, |
// Prepare sequential strings |
STATIC_ASSERT(SeqTwoByteString::kHeaderSize == SeqOneByteString::kHeaderSize); |
- __ Addu(string, |
+ __ Daddu(string, |
string, |
SeqTwoByteString::kHeaderSize - kHeapObjectTag); |
__ jmp(&check_encoding); |
@@ -1011,7 +900,7 @@ void StringCharLoadGenerator::Generate(MacroAssembler* masm, |
STATIC_ASSERT(kShortExternalStringTag != 0); |
__ And(at, result, Operand(kShortExternalStringMask)); |
__ Branch(call_runtime, ne, at, Operand(zero_reg)); |
- __ lw(string, FieldMemOperand(string, ExternalString::kResourceDataOffset)); |
+ __ ld(string, FieldMemOperand(string, ExternalString::kResourceDataOffset)); |
Label ascii, done; |
__ bind(&check_encoding); |
@@ -1019,13 +908,13 @@ void StringCharLoadGenerator::Generate(MacroAssembler* masm, |
__ And(at, result, Operand(kStringEncodingMask)); |
__ Branch(&ascii, ne, at, Operand(zero_reg)); |
// Two-byte string. |
- __ sll(at, index, 1); |
- __ Addu(at, string, at); |
+ __ dsll(at, index, 1); |
+ __ Daddu(at, string, at); |
__ lhu(result, MemOperand(at)); |
__ jmp(&done); |
__ bind(&ascii); |
// Ascii string. |
- __ Addu(at, string, index); |
+ __ Daddu(at, string, index); |
__ lbu(result, MemOperand(at)); |
__ bind(&done); |
} |
@@ -1056,7 +945,6 @@ void MathExpGenerator::EmitMathExp(MacroAssembler* masm, |
ASSERT(ExternalReference::math_exp_constants(0).address() != NULL); |
Label zero, infinity, done; |
- |
__ li(temp3, Operand(ExternalReference::math_exp_constants(0))); |
__ ldc1(double_scratch1, ExpConstant(0, temp3)); |
@@ -1086,23 +974,23 @@ void MathExpGenerator::EmitMathExp(MacroAssembler* masm, |
(ExternalReference::math_exp_constants(8).address()) == 1); |
__ Move(double_scratch2, 1); |
__ add_d(result, result, double_scratch2); |
- __ srl(temp1, temp2, 11); |
+ __ dsrl(temp1, temp2, 11); |
__ Ext(temp2, temp2, 0, 11); |
- __ Addu(temp1, temp1, Operand(0x3ff)); |
+ __ Daddu(temp1, temp1, Operand(0x3ff)); |
// Must not call ExpConstant() after overwriting temp3! |
__ li(temp3, Operand(ExternalReference::math_exp_log_table())); |
- __ sll(at, temp2, 3); |
- __ Addu(temp3, temp3, Operand(at)); |
- __ lw(temp2, MemOperand(temp3, Register::kMantissaOffset)); |
- __ lw(temp3, MemOperand(temp3, Register::kExponentOffset)); |
+ __ dsll(at, temp2, 3); |
+ __ Daddu(temp3, temp3, Operand(at)); |
+ __ lwu(temp2, MemOperand(temp3, 0)); |
+ __ lwu(temp3, MemOperand(temp3, kIntSize)); |
// The first word is loaded is the lower number register. |
if (temp2.code() < temp3.code()) { |
- __ sll(at, temp1, 20); |
+ __ dsll(at, temp1, 20); |
__ Or(temp1, temp3, at); |
__ Move(double_scratch1, temp2, temp1); |
} else { |
- __ sll(at, temp1, 20); |
+ __ dsll(at, temp1, 20); |
__ Or(temp1, temp2, at); |
__ Move(double_scratch1, temp3, temp1); |
} |
@@ -1138,7 +1026,9 @@ CodeAgingHelper::CodeAgingHelper() { |
PredictableCodeSizeScope scope(patcher->masm(), young_sequence_.length()); |
patcher->masm()->Push(ra, fp, cp, a1); |
patcher->masm()->nop(Assembler::CODE_AGE_SEQUENCE_NOP); |
- patcher->masm()->Addu( |
+ patcher->masm()->nop(Assembler::CODE_AGE_SEQUENCE_NOP); |
+ patcher->masm()->nop(Assembler::CODE_AGE_SEQUENCE_NOP); |
+ patcher->masm()->Daddu( |
fp, sp, Operand(StandardFrameConstants::kFixedFrameSizeFromFp)); |
} |
@@ -1188,8 +1078,8 @@ void Code::PatchPlatformCodeAge(Isolate* isolate, |
// GetCodeAgeAndParity() extracts the stub address from this instruction. |
patcher.masm()->li( |
t9, |
- Operand(reinterpret_cast<uint32_t>(stub->instruction_start())), |
- CONSTANT_SIZE); |
+ Operand(reinterpret_cast<uint64_t>(stub->instruction_start())), |
+ ADDRESS_LOAD); |
patcher.masm()->nop(); // Prevent jalr to jal optimization. |
patcher.masm()->jalr(t9, a0); |
patcher.masm()->nop(); // Branch delay slot nop. |
@@ -1202,4 +1092,4 @@ void Code::PatchPlatformCodeAge(Isolate* isolate, |
} } // namespace v8::internal |
-#endif // V8_TARGET_ARCH_MIPS |
+#endif // V8_TARGET_ARCH_MIPS64 |