| Index: src/arm/macro-assembler-arm.cc
|
| diff --git a/src/arm/macro-assembler-arm.cc b/src/arm/macro-assembler-arm.cc
|
| index 58ce0a8545ceeabc6f0528de919ea2c49afcfd69..77c514ff546624baa7db121f21baf63702608e5e 100644
|
| --- a/src/arm/macro-assembler-arm.cc
|
| +++ b/src/arm/macro-assembler-arm.cc
|
| @@ -2806,8 +2806,16 @@ void MacroAssembler::Check(Condition cond, BailoutReason reason) {
|
| void MacroAssembler::Abort(BailoutReason reason) {
|
| Label abort_start;
|
| bind(&abort_start);
|
| -#ifdef DEBUG
|
| + // We want to pass the msg string like a smi to avoid GC
|
| + // problems, however msg is not guaranteed to be aligned
|
| + // properly. Instead, we pass an aligned pointer that is
|
| + // a proper v8 smi, but also pass the alignment difference
|
| + // from the real pointer as a smi.
|
| const char* msg = GetBailoutReason(reason);
|
| + intptr_t p1 = reinterpret_cast<intptr_t>(msg);
|
| + intptr_t p0 = (p1 & ~kSmiTagMask) + kSmiTag;
|
| + ASSERT(reinterpret_cast<Object*>(p0)->IsSmi());
|
| +#ifdef DEBUG
|
| if (msg != NULL) {
|
| RecordComment("Abort message: ");
|
| RecordComment(msg);
|
| @@ -2819,24 +2827,25 @@ void MacroAssembler::Abort(BailoutReason reason) {
|
| }
|
| #endif
|
|
|
| - mov(r0, Operand(Smi::FromInt(reason)));
|
| + mov(r0, Operand(p0));
|
| + push(r0);
|
| + mov(r0, Operand(Smi::FromInt(p1 - p0)));
|
| push(r0);
|
| -
|
| // Disable stub call restrictions to always allow calls to abort.
|
| if (!has_frame_) {
|
| // We don't actually want to generate a pile of code for this, so just
|
| // claim there is a stack frame, without generating one.
|
| FrameScope scope(this, StackFrame::NONE);
|
| - CallRuntime(Runtime::kAbort, 1);
|
| + CallRuntime(Runtime::kAbort, 2);
|
| } else {
|
| - CallRuntime(Runtime::kAbort, 1);
|
| + CallRuntime(Runtime::kAbort, 2);
|
| }
|
| // will not return here
|
| if (is_const_pool_blocked()) {
|
| // If the calling code cares about the exact number of
|
| // instructions generated, we insert padding here to keep the size
|
| // of the Abort macro constant.
|
| - static const int kExpectedAbortInstructions = 7;
|
| + static const int kExpectedAbortInstructions = 10;
|
| int abort_instructions = InstructionsGeneratedSince(&abort_start);
|
| ASSERT(abort_instructions <= kExpectedAbortInstructions);
|
| while (abort_instructions++ < kExpectedAbortInstructions) {
|
| @@ -2890,6 +2899,31 @@ void MacroAssembler::LoadTransitionedArrayMapConditional(
|
| }
|
|
|
|
|
| +void MacroAssembler::LoadInitialArrayMap(
|
| + Register function_in, Register scratch,
|
| + Register map_out, bool can_have_holes) {
|
| + ASSERT(!function_in.is(map_out));
|
| + Label done;
|
| + ldr(map_out, FieldMemOperand(function_in,
|
| + JSFunction::kPrototypeOrInitialMapOffset));
|
| + if (!FLAG_smi_only_arrays) {
|
| + ElementsKind kind = can_have_holes ? FAST_HOLEY_ELEMENTS : FAST_ELEMENTS;
|
| + LoadTransitionedArrayMapConditional(FAST_SMI_ELEMENTS,
|
| + kind,
|
| + map_out,
|
| + scratch,
|
| + &done);
|
| + } else if (can_have_holes) {
|
| + LoadTransitionedArrayMapConditional(FAST_SMI_ELEMENTS,
|
| + FAST_HOLEY_SMI_ELEMENTS,
|
| + map_out,
|
| + scratch,
|
| + &done);
|
| + }
|
| + bind(&done);
|
| +}
|
| +
|
| +
|
| void MacroAssembler::LoadGlobalFunction(int index, Register function) {
|
| // Load the global or builtins object from the current context.
|
| ldr(function,
|
|
|