| Index: src/mips/stub-cache-mips.cc
|
| diff --git a/src/mips/stub-cache-mips.cc b/src/mips/stub-cache-mips.cc
|
| index e2831404fe4f898eaa5bf776c495ffd09a858152..6632f943691e6e4c91d4e5533e606df3a3c18552 100644
|
| --- a/src/mips/stub-cache-mips.cc
|
| +++ b/src/mips/stub-cache-mips.cc
|
| @@ -1556,250 +1556,6 @@ Handle<Code> CallStubCompiler::CompileCallField(Handle<JSObject> object,
|
| }
|
|
|
|
|
| -Handle<Code> CallStubCompiler::CompileArrayPushCall(
|
| - Handle<Object> object,
|
| - Handle<JSObject> holder,
|
| - Handle<Cell> cell,
|
| - Handle<JSFunction> function,
|
| - Handle<String> name,
|
| - Code::StubType type) {
|
| - // If object is not an array or is observed or sealed, bail out to regular
|
| - // call.
|
| - if (!object->IsJSArray() ||
|
| - !cell.is_null() ||
|
| - Handle<JSArray>::cast(object)->map()->is_observed() ||
|
| - !Handle<JSArray>::cast(object)->map()->is_extensible()) {
|
| - return Handle<Code>::null();
|
| - }
|
| -
|
| - Label miss;
|
| - HandlerFrontendHeader(object, holder, name, RECEIVER_MAP_CHECK, &miss);
|
| - Register receiver = a0;
|
| - Register scratch = a1;
|
| -
|
| - const int argc = arguments().immediate();
|
| -
|
| - if (argc == 0) {
|
| - // Nothing to do, just return the length.
|
| - __ lw(v0, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| - __ DropAndRet(argc + 1);
|
| - } else {
|
| - Label call_builtin;
|
| - if (argc == 1) { // Otherwise fall through to call the builtin.
|
| - Label attempt_to_grow_elements, with_write_barrier, check_double;
|
| -
|
| - Register elements = t2;
|
| - Register end_elements = t1;
|
| - // Get the elements array of the object.
|
| - __ lw(elements, FieldMemOperand(receiver, JSArray::kElementsOffset));
|
| -
|
| - // Check that the elements are in fast mode and writable.
|
| - __ CheckMap(elements,
|
| - scratch,
|
| - Heap::kFixedArrayMapRootIndex,
|
| - &check_double,
|
| - DONT_DO_SMI_CHECK);
|
| -
|
| - // Get the array's length into scratch and calculate new length.
|
| - __ lw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| - STATIC_ASSERT(kSmiTagSize == 1);
|
| - STATIC_ASSERT(kSmiTag == 0);
|
| - __ Addu(scratch, scratch, Operand(Smi::FromInt(argc)));
|
| -
|
| - // Get the elements' length.
|
| - __ lw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset));
|
| -
|
| - // Check if we could survive without allocation.
|
| - __ Branch(&attempt_to_grow_elements, gt, scratch, Operand(t0));
|
| -
|
| - // Check if value is a smi.
|
| - __ lw(t0, MemOperand(sp, (argc - 1) * kPointerSize));
|
| - __ JumpIfNotSmi(t0, &with_write_barrier);
|
| -
|
| - // Save new length.
|
| - __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| -
|
| - // Store the value.
|
| - // We may need a register containing the address end_elements below,
|
| - // so write back the value in end_elements.
|
| - __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize);
|
| - __ Addu(end_elements, elements, end_elements);
|
| - const int kEndElementsOffset =
|
| - FixedArray::kHeaderSize - kHeapObjectTag - argc * kPointerSize;
|
| - __ Addu(end_elements, end_elements, kEndElementsOffset);
|
| - __ sw(t0, MemOperand(end_elements));
|
| -
|
| - // Check for a smi.
|
| - __ mov(v0, scratch);
|
| - __ DropAndRet(argc + 1);
|
| -
|
| - __ bind(&check_double);
|
| -
|
| - // Check that the elements are in fast mode and writable.
|
| - __ CheckMap(elements,
|
| - scratch,
|
| - Heap::kFixedDoubleArrayMapRootIndex,
|
| - &call_builtin,
|
| - DONT_DO_SMI_CHECK);
|
| -
|
| - // Get the array's length into scratch and calculate new length.
|
| - __ lw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| - STATIC_ASSERT(kSmiTagSize == 1);
|
| - STATIC_ASSERT(kSmiTag == 0);
|
| - __ Addu(scratch, scratch, Operand(Smi::FromInt(argc)));
|
| -
|
| - // Get the elements' length.
|
| - __ lw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset));
|
| -
|
| - // Check if we could survive without allocation.
|
| - __ Branch(&call_builtin, gt, scratch, Operand(t0));
|
| -
|
| - __ lw(t0, MemOperand(sp, (argc - 1) * kPointerSize));
|
| - __ StoreNumberToDoubleElements(
|
| - t0, scratch, elements, a3, t1, a2,
|
| - &call_builtin, argc * kDoubleSize);
|
| -
|
| - // Save new length.
|
| - __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| -
|
| - __ mov(v0, scratch);
|
| - __ DropAndRet(argc + 1);
|
| -
|
| - __ bind(&with_write_barrier);
|
| -
|
| - __ lw(a3, FieldMemOperand(receiver, HeapObject::kMapOffset));
|
| -
|
| - if (FLAG_smi_only_arrays && !FLAG_trace_elements_transitions) {
|
| - Label fast_object, not_fast_object;
|
| - __ CheckFastObjectElements(a3, t3, ¬_fast_object);
|
| - __ jmp(&fast_object);
|
| - // In case of fast smi-only, convert to fast object, otherwise bail out.
|
| - __ bind(¬_fast_object);
|
| - __ CheckFastSmiElements(a3, t3, &call_builtin);
|
| -
|
| - __ lw(t3, FieldMemOperand(t0, HeapObject::kMapOffset));
|
| - __ LoadRoot(at, Heap::kHeapNumberMapRootIndex);
|
| - __ Branch(&call_builtin, eq, t3, Operand(at));
|
| - // edx: receiver
|
| - // a3: map
|
| - Label try_holey_map;
|
| - __ LoadTransitionedArrayMapConditional(FAST_SMI_ELEMENTS,
|
| - FAST_ELEMENTS,
|
| - a3,
|
| - t3,
|
| - &try_holey_map);
|
| - __ mov(a2, receiver);
|
| - ElementsTransitionGenerator::
|
| - GenerateMapChangeElementsTransition(masm(),
|
| - DONT_TRACK_ALLOCATION_SITE,
|
| - NULL);
|
| - __ jmp(&fast_object);
|
| -
|
| - __ bind(&try_holey_map);
|
| - __ LoadTransitionedArrayMapConditional(FAST_HOLEY_SMI_ELEMENTS,
|
| - FAST_HOLEY_ELEMENTS,
|
| - a3,
|
| - t3,
|
| - &call_builtin);
|
| - __ mov(a2, receiver);
|
| - ElementsTransitionGenerator::
|
| - GenerateMapChangeElementsTransition(masm(),
|
| - DONT_TRACK_ALLOCATION_SITE,
|
| - NULL);
|
| - __ bind(&fast_object);
|
| - } else {
|
| - __ CheckFastObjectElements(a3, a3, &call_builtin);
|
| - }
|
| -
|
| - // Save new length.
|
| - __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| -
|
| - // Store the value.
|
| - // We may need a register containing the address end_elements below,
|
| - // so write back the value in end_elements.
|
| - __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize);
|
| - __ Addu(end_elements, elements, end_elements);
|
| - __ Addu(end_elements, end_elements, kEndElementsOffset);
|
| - __ sw(t0, MemOperand(end_elements));
|
| -
|
| - __ RecordWrite(elements,
|
| - end_elements,
|
| - t0,
|
| - kRAHasNotBeenSaved,
|
| - kDontSaveFPRegs,
|
| - EMIT_REMEMBERED_SET,
|
| - OMIT_SMI_CHECK);
|
| - __ mov(v0, scratch);
|
| - __ DropAndRet(argc + 1);
|
| -
|
| - __ bind(&attempt_to_grow_elements);
|
| - // scratch: array's length + 1.
|
| - // t0: elements' length.
|
| -
|
| - if (!FLAG_inline_new) {
|
| - __ Branch(&call_builtin);
|
| - }
|
| -
|
| - __ lw(a2, MemOperand(sp, (argc - 1) * kPointerSize));
|
| - // Growing elements that are SMI-only requires special handling in case
|
| - // the new element is non-Smi. For now, delegate to the builtin.
|
| - Label no_fast_elements_check;
|
| - __ JumpIfSmi(a2, &no_fast_elements_check);
|
| - __ lw(t3, FieldMemOperand(receiver, HeapObject::kMapOffset));
|
| - __ CheckFastObjectElements(t3, t3, &call_builtin);
|
| - __ bind(&no_fast_elements_check);
|
| -
|
| - ExternalReference new_space_allocation_top =
|
| - ExternalReference::new_space_allocation_top_address(isolate());
|
| - ExternalReference new_space_allocation_limit =
|
| - ExternalReference::new_space_allocation_limit_address(isolate());
|
| -
|
| - const int kAllocationDelta = 4;
|
| - // Load top and check if it is the end of elements.
|
| - __ sll(end_elements, scratch, kPointerSizeLog2 - kSmiTagSize);
|
| - __ Addu(end_elements, elements, end_elements);
|
| - __ Addu(end_elements, end_elements, Operand(kEndElementsOffset));
|
| - __ li(t3, Operand(new_space_allocation_top));
|
| - __ lw(a3, MemOperand(t3));
|
| - __ Branch(&call_builtin, ne, end_elements, Operand(a3));
|
| -
|
| - __ li(t5, Operand(new_space_allocation_limit));
|
| - __ lw(t5, MemOperand(t5));
|
| - __ Addu(a3, a3, Operand(kAllocationDelta * kPointerSize));
|
| - __ Branch(&call_builtin, hi, a3, Operand(t5));
|
| -
|
| - // We fit and could grow elements.
|
| - // Update new_space_allocation_top.
|
| - __ sw(a3, MemOperand(t3));
|
| - // Push the argument.
|
| - __ sw(a2, MemOperand(end_elements));
|
| - // Fill the rest with holes.
|
| - __ LoadRoot(a3, Heap::kTheHoleValueRootIndex);
|
| - for (int i = 1; i < kAllocationDelta; i++) {
|
| - __ sw(a3, MemOperand(end_elements, i * kPointerSize));
|
| - }
|
| -
|
| - // Update elements' and array's sizes.
|
| - __ sw(scratch, FieldMemOperand(receiver, JSArray::kLengthOffset));
|
| - __ Addu(t0, t0, Operand(Smi::FromInt(kAllocationDelta)));
|
| - __ sw(t0, FieldMemOperand(elements, FixedArray::kLengthOffset));
|
| -
|
| - // Elements are in new space, so write barrier is not required.
|
| - __ mov(v0, scratch);
|
| - __ DropAndRet(argc + 1);
|
| - }
|
| - __ bind(&call_builtin);
|
| - __ TailCallExternalReference(
|
| - ExternalReference(Builtins::c_ArrayPush, isolate()), argc + 1, 1);
|
| - }
|
| -
|
| - HandlerFrontendFooter(&miss);
|
| -
|
| - // Return the generated code.
|
| - return GetCode(type, name);
|
| -}
|
| -
|
| -
|
| Handle<Code> CallStubCompiler::CompileArrayPopCall(
|
| Handle<Object> object,
|
| Handle<JSObject> holder,
|
|
|