| Index: src/arm/code-stubs-arm.cc
|
| diff --git a/src/arm/code-stubs-arm.cc b/src/arm/code-stubs-arm.cc
|
| index fd387fe1652378d8e8fdb609e942a11dd35de031..1d1fe83c2fee9ae565381c5ccb206e5b78e7811e 100644
|
| --- a/src/arm/code-stubs-arm.cc
|
| +++ b/src/arm/code-stubs-arm.cc
|
| @@ -2996,9 +2996,7 @@ void CodeStub::GenerateStubsAheadOfTime(Isolate* isolate) {
|
| StoreBufferOverflowStub::GenerateFixedRegStubsAheadOfTime(isolate);
|
| StubFailureTrampolineStub::GenerateAheadOfTime(isolate);
|
| RecordWriteStub::GenerateFixedRegStubsAheadOfTime(isolate);
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - ArrayConstructorStubBase::GenerateStubsAheadOfTime(isolate);
|
| - }
|
| + ArrayConstructorStubBase::GenerateStubsAheadOfTime(isolate);
|
| }
|
|
|
|
|
| @@ -4623,52 +4621,12 @@ void RegExpConstructResultStub::Generate(MacroAssembler* masm) {
|
| }
|
|
|
|
|
| -static void GenerateRecordCallTargetNoArray(MacroAssembler* masm) {
|
| - // Cache the called function in a global property cell. Cache states
|
| - // are uninitialized, monomorphic (indicated by a JSFunction), and
|
| - // megamorphic.
|
| - // r1 : the function to call
|
| - // r2 : cache cell for call target
|
| - Label done;
|
| -
|
| - ASSERT_EQ(*TypeFeedbackCells::MegamorphicSentinel(masm->isolate()),
|
| - masm->isolate()->heap()->undefined_value());
|
| - ASSERT_EQ(*TypeFeedbackCells::UninitializedSentinel(masm->isolate()),
|
| - masm->isolate()->heap()->the_hole_value());
|
| -
|
| - // Load the cache state into r3.
|
| - __ ldr(r3, FieldMemOperand(r2, Cell::kValueOffset));
|
| -
|
| - // A monomorphic cache hit or an already megamorphic state: invoke the
|
| - // function without changing the state.
|
| - __ cmp(r3, r1);
|
| - __ b(eq, &done);
|
| - __ CompareRoot(r3, Heap::kUndefinedValueRootIndex);
|
| - __ b(eq, &done);
|
| -
|
| - // A monomorphic miss (i.e, here the cache is not uninitialized) goes
|
| - // megamorphic.
|
| - __ CompareRoot(r3, Heap::kTheHoleValueRootIndex);
|
| - // MegamorphicSentinel is an immortal immovable object (undefined) so no
|
| - // write-barrier is needed.
|
| - __ LoadRoot(ip, Heap::kUndefinedValueRootIndex, ne);
|
| - __ str(ip, FieldMemOperand(r2, Cell::kValueOffset), ne);
|
| -
|
| - // An uninitialized cache is patched with the function.
|
| - __ str(r1, FieldMemOperand(r2, Cell::kValueOffset), eq);
|
| - // No need for a write barrier here - cells are rescanned.
|
| -
|
| - __ bind(&done);
|
| -}
|
| -
|
| -
|
| static void GenerateRecordCallTarget(MacroAssembler* masm) {
|
| // Cache the called function in a global property cell. Cache states
|
| // are uninitialized, monomorphic (indicated by a JSFunction), and
|
| // megamorphic.
|
| // r1 : the function to call
|
| // r2 : cache cell for call target
|
| - ASSERT(FLAG_optimize_constructed_arrays);
|
| Label initialize, done, miss, megamorphic, not_array_function;
|
|
|
| ASSERT_EQ(*TypeFeedbackCells::MegamorphicSentinel(masm->isolate()),
|
| @@ -4772,11 +4730,7 @@ void CallFunctionStub::Generate(MacroAssembler* masm) {
|
| __ b(ne, &slow);
|
|
|
| if (RecordCallTarget()) {
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - GenerateRecordCallTarget(masm);
|
| - } else {
|
| - GenerateRecordCallTargetNoArray(masm);
|
| - }
|
| + GenerateRecordCallTarget(masm);
|
| }
|
|
|
| // Fast-case: Invoke the function now.
|
| @@ -4851,15 +4805,11 @@ void CallConstructStub::Generate(MacroAssembler* masm) {
|
| __ b(ne, &slow);
|
|
|
| if (RecordCallTarget()) {
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - GenerateRecordCallTarget(masm);
|
| - } else {
|
| - GenerateRecordCallTargetNoArray(masm);
|
| - }
|
| + GenerateRecordCallTarget(masm);
|
| }
|
|
|
| // Jump to the function-specific construct stub.
|
| - Register jmp_reg = FLAG_optimize_constructed_arrays ? r3 : r2;
|
| + Register jmp_reg = r3;
|
| __ ldr(jmp_reg, FieldMemOperand(r1, JSFunction::kSharedFunctionInfoOffset));
|
| __ ldr(jmp_reg, FieldMemOperand(jmp_reg,
|
| SharedFunctionInfo::kConstructStubOffset));
|
| @@ -7341,52 +7291,39 @@ void ArrayConstructorStub::Generate(MacroAssembler* masm) {
|
| __ bind(&okay_here);
|
| }
|
|
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - Label no_info, switch_ready;
|
| - // Get the elements kind and case on that.
|
| - __ cmp(r2, Operand(undefined_sentinel));
|
| - __ b(eq, &no_info);
|
| - __ ldr(r3, FieldMemOperand(r2, PropertyCell::kValueOffset));
|
| - __ JumpIfNotSmi(r3, &no_info);
|
| - __ SmiUntag(r3);
|
| - __ jmp(&switch_ready);
|
| - __ bind(&no_info);
|
| - __ mov(r3, Operand(GetInitialFastElementsKind()));
|
| - __ bind(&switch_ready);
|
| -
|
| - if (argument_count_ == ANY) {
|
| - Label not_zero_case, not_one_case;
|
| - __ tst(r0, r0);
|
| - __ b(ne, ¬_zero_case);
|
| - CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| -
|
| - __ bind(¬_zero_case);
|
| - __ cmp(r0, Operand(1));
|
| - __ b(gt, ¬_one_case);
|
| - CreateArrayDispatchOneArgument(masm);
|
| -
|
| - __ bind(¬_one_case);
|
| - CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| - } else if (argument_count_ == NONE) {
|
| - CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| - } else if (argument_count_ == ONE) {
|
| - CreateArrayDispatchOneArgument(masm);
|
| - } else if (argument_count_ == MORE_THAN_ONE) {
|
| - CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| - } else {
|
| - UNREACHABLE();
|
| - }
|
| + Label no_info, switch_ready;
|
| + // Get the elements kind and case on that.
|
| + __ cmp(r2, Operand(undefined_sentinel));
|
| + __ b(eq, &no_info);
|
| + __ ldr(r3, FieldMemOperand(r2, Cell::kValueOffset));
|
| + __ JumpIfNotSmi(r3, &no_info);
|
| + __ SmiUntag(r3);
|
| + __ jmp(&switch_ready);
|
| + __ bind(&no_info);
|
| + __ mov(r3, Operand(GetInitialFastElementsKind()));
|
| + __ bind(&switch_ready);
|
| +
|
| + if (argument_count_ == ANY) {
|
| + Label not_zero_case, not_one_case;
|
| + __ tst(r0, r0);
|
| + __ b(ne, ¬_zero_case);
|
| + CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| +
|
| + __ bind(¬_zero_case);
|
| + __ cmp(r0, Operand(1));
|
| + __ b(gt, ¬_one_case);
|
| + CreateArrayDispatchOneArgument(masm);
|
| +
|
| + __ bind(¬_one_case);
|
| + CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| + } else if (argument_count_ == NONE) {
|
| + CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| + } else if (argument_count_ == ONE) {
|
| + CreateArrayDispatchOneArgument(masm);
|
| + } else if (argument_count_ == MORE_THAN_ONE) {
|
| + CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| } else {
|
| - Label generic_constructor;
|
| - // Run the native code for the Array function called as a constructor.
|
| - ArrayNativeCode(masm, &generic_constructor);
|
| -
|
| - // Jump to the generic construct code in case the specialized code cannot
|
| - // handle the construction.
|
| - __ bind(&generic_constructor);
|
| - Handle<Code> generic_construct_stub =
|
| - masm->isolate()->builtins()->JSConstructStubGeneric();
|
| - __ Jump(generic_construct_stub, RelocInfo::CODE_TARGET);
|
| + UNREACHABLE();
|
| }
|
| }
|
|
|
| @@ -7448,45 +7385,31 @@ void InternalArrayConstructorStub::Generate(MacroAssembler* masm) {
|
| __ Assert(eq, "Unexpected initial map for Array function");
|
| }
|
|
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - // Figure out the right elements kind
|
| - __ ldr(r3, FieldMemOperand(r1, JSFunction::kPrototypeOrInitialMapOffset));
|
| -
|
| - // Load the map's "bit field 2" into |result|. We only need the first byte,
|
| - // but the following bit field extraction takes care of that anyway.
|
| - __ ldr(r3, FieldMemOperand(r3, Map::kBitField2Offset));
|
| - // Retrieve elements_kind from bit field 2.
|
| - __ Ubfx(r3, r3, Map::kElementsKindShift, Map::kElementsKindBitCount);
|
| -
|
| - if (FLAG_debug_code) {
|
| - Label done;
|
| - __ cmp(r3, Operand(FAST_ELEMENTS));
|
| - __ b(eq, &done);
|
| - __ cmp(r3, Operand(FAST_HOLEY_ELEMENTS));
|
| - __ Assert(eq,
|
| - "Invalid ElementsKind for InternalArray or InternalPackedArray");
|
| - __ bind(&done);
|
| - }
|
| + // Figure out the right elements kind
|
| + __ ldr(r3, FieldMemOperand(r1, JSFunction::kPrototypeOrInitialMapOffset));
|
| + // Load the map's "bit field 2" into |result|. We only need the first byte,
|
| + // but the following bit field extraction takes care of that anyway.
|
| + __ ldr(r3, FieldMemOperand(r3, Map::kBitField2Offset));
|
| + // Retrieve elements_kind from bit field 2.
|
| + __ Ubfx(r3, r3, Map::kElementsKindShift, Map::kElementsKindBitCount);
|
|
|
| - Label fast_elements_case;
|
| + if (FLAG_debug_code) {
|
| + Label done;
|
| __ cmp(r3, Operand(FAST_ELEMENTS));
|
| - __ b(eq, &fast_elements_case);
|
| - GenerateCase(masm, FAST_HOLEY_ELEMENTS);
|
| + __ b(eq, &done);
|
| + __ cmp(r3, Operand(FAST_HOLEY_ELEMENTS));
|
| + __ Assert(eq,
|
| + "Invalid ElementsKind for InternalArray or InternalPackedArray");
|
| + __ bind(&done);
|
| + }
|
|
|
| - __ bind(&fast_elements_case);
|
| - GenerateCase(masm, FAST_ELEMENTS);
|
| - } else {
|
| - Label generic_constructor;
|
| - // Run the native code for the Array function called as constructor.
|
| - ArrayNativeCode(masm, &generic_constructor);
|
| + Label fast_elements_case;
|
| + __ cmp(r3, Operand(FAST_ELEMENTS));
|
| + __ b(eq, &fast_elements_case);
|
| + GenerateCase(masm, FAST_HOLEY_ELEMENTS);
|
|
|
| - // Jump to the generic construct code in case the specialized code cannot
|
| - // handle the construction.
|
| - __ bind(&generic_constructor);
|
| - Handle<Code> generic_construct_stub =
|
| - masm->isolate()->builtins()->JSConstructStubGeneric();
|
| - __ Jump(generic_construct_stub, RelocInfo::CODE_TARGET);
|
| - }
|
| + __ bind(&fast_elements_case);
|
| + GenerateCase(masm, FAST_ELEMENTS);
|
| }
|
|
|
|
|
|
|