| Index: src/ia32/code-stubs-ia32.cc
|
| diff --git a/src/ia32/code-stubs-ia32.cc b/src/ia32/code-stubs-ia32.cc
|
| index 4b6a6abc490f29a3b46d24cd676785a15cff1362..0ca56984ad0497a857ba9ea3f4406902e20b52d1 100644
|
| --- a/src/ia32/code-stubs-ia32.cc
|
| +++ b/src/ia32/code-stubs-ia32.cc
|
| @@ -4678,51 +4678,12 @@ void InterruptStub::Generate(MacroAssembler* masm) {
|
| }
|
|
|
|
|
| -static void GenerateRecordCallTargetNoArray(MacroAssembler* masm) {
|
| - // Cache the called function in a global property cell. Cache states
|
| - // are uninitialized, monomorphic (indicated by a JSFunction), and
|
| - // megamorphic.
|
| - // ebx : cache cell for call target
|
| - // edi : the function to call
|
| - Isolate* isolate = masm->isolate();
|
| - Label initialize, done;
|
| -
|
| - // Load the cache state into ecx.
|
| - __ mov(ecx, FieldOperand(ebx, PropertyCell::kValueOffset));
|
| -
|
| - // A monomorphic cache hit or an already megamorphic state: invoke the
|
| - // function without changing the state.
|
| - __ cmp(ecx, edi);
|
| - __ j(equal, &done, Label::kNear);
|
| - __ cmp(ecx, Immediate(TypeFeedbackCells::MegamorphicSentinel(isolate)));
|
| - __ j(equal, &done, Label::kNear);
|
| -
|
| - // A monomorphic miss (i.e, here the cache is not uninitialized) goes
|
| - // megamorphic.
|
| - __ cmp(ecx, Immediate(TypeFeedbackCells::UninitializedSentinel(isolate)));
|
| - __ j(equal, &initialize, Label::kNear);
|
| - // MegamorphicSentinel is an immortal immovable object (undefined) so no
|
| - // write-barrier is needed.
|
| - __ mov(FieldOperand(ebx, Cell::kValueOffset),
|
| - Immediate(TypeFeedbackCells::MegamorphicSentinel(isolate)));
|
| - __ jmp(&done, Label::kNear);
|
| -
|
| - // An uninitialized cache is patched with the function.
|
| - __ bind(&initialize);
|
| - __ mov(FieldOperand(ebx, Cell::kValueOffset), edi);
|
| - // No need for a write barrier here - cells are rescanned.
|
| -
|
| - __ bind(&done);
|
| -}
|
| -
|
| -
|
| static void GenerateRecordCallTarget(MacroAssembler* masm) {
|
| // Cache the called function in a global property cell. Cache states
|
| // are uninitialized, monomorphic (indicated by a JSFunction), and
|
| // megamorphic.
|
| // ebx : cache cell for call target
|
| // edi : the function to call
|
| - ASSERT(FLAG_optimize_constructed_arrays);
|
| Isolate* isolate = masm->isolate();
|
| Label initialize, done, miss, megamorphic, not_array_function;
|
|
|
| @@ -4824,11 +4785,7 @@ void CallFunctionStub::Generate(MacroAssembler* masm) {
|
| __ j(not_equal, &slow);
|
|
|
| if (RecordCallTarget()) {
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - GenerateRecordCallTarget(masm);
|
| - } else {
|
| - GenerateRecordCallTargetNoArray(masm);
|
| - }
|
| + GenerateRecordCallTarget(masm);
|
| }
|
|
|
| // Fast-case: Just invoke the function.
|
| @@ -4901,15 +4858,11 @@ void CallConstructStub::Generate(MacroAssembler* masm) {
|
| __ j(not_equal, &slow);
|
|
|
| if (RecordCallTarget()) {
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - GenerateRecordCallTarget(masm);
|
| - } else {
|
| - GenerateRecordCallTargetNoArray(masm);
|
| - }
|
| + GenerateRecordCallTarget(masm);
|
| }
|
|
|
| // Jump to the function-specific construct stub.
|
| - Register jmp_reg = FLAG_optimize_constructed_arrays ? ecx : ebx;
|
| + Register jmp_reg = ecx;
|
| __ mov(jmp_reg, FieldOperand(edi, JSFunction::kSharedFunctionInfoOffset));
|
| __ mov(jmp_reg, FieldOperand(jmp_reg,
|
| SharedFunctionInfo::kConstructStubOffset));
|
| @@ -4955,9 +4908,7 @@ void CodeStub::GenerateStubsAheadOfTime(Isolate* isolate) {
|
| StubFailureTrampolineStub::GenerateAheadOfTime(isolate);
|
| // It is important that the store buffer overflow stubs are generated first.
|
| RecordWriteStub::GenerateFixedRegStubsAheadOfTime(isolate);
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - ArrayConstructorStubBase::GenerateStubsAheadOfTime(isolate);
|
| - }
|
| + ArrayConstructorStubBase::GenerateStubsAheadOfTime(isolate);
|
| }
|
|
|
|
|
| @@ -7940,52 +7891,39 @@ void ArrayConstructorStub::Generate(MacroAssembler* masm) {
|
| __ bind(&okay_here);
|
| }
|
|
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - Label no_info, switch_ready;
|
| - // Get the elements kind and case on that.
|
| - __ cmp(ebx, Immediate(undefined_sentinel));
|
| - __ j(equal, &no_info);
|
| - __ mov(edx, FieldOperand(ebx, Cell::kValueOffset));
|
| - __ JumpIfNotSmi(edx, &no_info);
|
| - __ SmiUntag(edx);
|
| - __ jmp(&switch_ready);
|
| - __ bind(&no_info);
|
| - __ mov(edx, Immediate(GetInitialFastElementsKind()));
|
| - __ bind(&switch_ready);
|
| -
|
| - if (argument_count_ == ANY) {
|
| - Label not_zero_case, not_one_case;
|
| - __ test(eax, eax);
|
| - __ j(not_zero, ¬_zero_case);
|
| - CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| -
|
| - __ bind(¬_zero_case);
|
| - __ cmp(eax, 1);
|
| - __ j(greater, ¬_one_case);
|
| - CreateArrayDispatchOneArgument(masm);
|
| -
|
| - __ bind(¬_one_case);
|
| - CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| - } else if (argument_count_ == NONE) {
|
| - CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| - } else if (argument_count_ == ONE) {
|
| - CreateArrayDispatchOneArgument(masm);
|
| - } else if (argument_count_ == MORE_THAN_ONE) {
|
| - CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| - } else {
|
| - UNREACHABLE();
|
| - }
|
| - } else {
|
| - Label generic_constructor;
|
| - // Run the native code for the Array function called as constructor.
|
| - ArrayNativeCode(masm, true, &generic_constructor);
|
| + Label no_info, switch_ready;
|
| + // Get the elements kind and case on that.
|
| + __ cmp(ebx, Immediate(undefined_sentinel));
|
| + __ j(equal, &no_info);
|
| + __ mov(edx, FieldOperand(ebx, Cell::kValueOffset));
|
| + __ JumpIfNotSmi(edx, &no_info);
|
| + __ SmiUntag(edx);
|
| + __ jmp(&switch_ready);
|
| + __ bind(&no_info);
|
| + __ mov(edx, Immediate(GetInitialFastElementsKind()));
|
| + __ bind(&switch_ready);
|
|
|
| - // Jump to the generic construct code in case the specialized code cannot
|
| - // handle the construction.
|
| - __ bind(&generic_constructor);
|
| - Handle<Code> generic_construct_stub =
|
| - masm->isolate()->builtins()->JSConstructStubGeneric();
|
| - __ jmp(generic_construct_stub, RelocInfo::CODE_TARGET);
|
| + if (argument_count_ == ANY) {
|
| + Label not_zero_case, not_one_case;
|
| + __ test(eax, eax);
|
| + __ j(not_zero, ¬_zero_case);
|
| + CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| +
|
| + __ bind(¬_zero_case);
|
| + __ cmp(eax, 1);
|
| + __ j(greater, ¬_one_case);
|
| + CreateArrayDispatchOneArgument(masm);
|
| +
|
| + __ bind(¬_one_case);
|
| + CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| + } else if (argument_count_ == NONE) {
|
| + CreateArrayDispatch<ArrayNoArgumentConstructorStub>(masm);
|
| + } else if (argument_count_ == ONE) {
|
| + CreateArrayDispatchOneArgument(masm);
|
| + } else if (argument_count_ == MORE_THAN_ONE) {
|
| + CreateArrayDispatch<ArrayNArgumentsConstructorStub>(masm);
|
| + } else {
|
| + UNREACHABLE();
|
| }
|
| }
|
|
|
| @@ -8048,46 +7986,33 @@ void InternalArrayConstructorStub::Generate(MacroAssembler* masm) {
|
| __ Assert(equal, "Unexpected initial map for Array function");
|
| }
|
|
|
| - if (FLAG_optimize_constructed_arrays) {
|
| - // Figure out the right elements kind
|
| - __ mov(ecx, FieldOperand(edi, JSFunction::kPrototypeOrInitialMapOffset));
|
| + // Figure out the right elements kind
|
| + __ mov(ecx, FieldOperand(edi, JSFunction::kPrototypeOrInitialMapOffset));
|
|
|
| - // Load the map's "bit field 2" into |result|. We only need the first byte,
|
| - // but the following masking takes care of that anyway.
|
| - __ mov(ecx, FieldOperand(ecx, Map::kBitField2Offset));
|
| - // Retrieve elements_kind from bit field 2.
|
| - __ and_(ecx, Map::kElementsKindMask);
|
| - __ shr(ecx, Map::kElementsKindShift);
|
| -
|
| - if (FLAG_debug_code) {
|
| - Label done;
|
| - __ cmp(ecx, Immediate(FAST_ELEMENTS));
|
| - __ j(equal, &done);
|
| - __ cmp(ecx, Immediate(FAST_HOLEY_ELEMENTS));
|
| - __ Assert(equal,
|
| - "Invalid ElementsKind for InternalArray or InternalPackedArray");
|
| - __ bind(&done);
|
| - }
|
| + // Load the map's "bit field 2" into |result|. We only need the first byte,
|
| + // but the following masking takes care of that anyway.
|
| + __ mov(ecx, FieldOperand(ecx, Map::kBitField2Offset));
|
| + // Retrieve elements_kind from bit field 2.
|
| + __ and_(ecx, Map::kElementsKindMask);
|
| + __ shr(ecx, Map::kElementsKindShift);
|
|
|
| - Label fast_elements_case;
|
| + if (FLAG_debug_code) {
|
| + Label done;
|
| __ cmp(ecx, Immediate(FAST_ELEMENTS));
|
| - __ j(equal, &fast_elements_case);
|
| - GenerateCase(masm, FAST_HOLEY_ELEMENTS);
|
| + __ j(equal, &done);
|
| + __ cmp(ecx, Immediate(FAST_HOLEY_ELEMENTS));
|
| + __ Assert(equal,
|
| + "Invalid ElementsKind for InternalArray or InternalPackedArray");
|
| + __ bind(&done);
|
| + }
|
|
|
| - __ bind(&fast_elements_case);
|
| - GenerateCase(masm, FAST_ELEMENTS);
|
| - } else {
|
| - Label generic_constructor;
|
| - // Run the native code for the Array function called as constructor.
|
| - ArrayNativeCode(masm, true, &generic_constructor);
|
| + Label fast_elements_case;
|
| + __ cmp(ecx, Immediate(FAST_ELEMENTS));
|
| + __ j(equal, &fast_elements_case);
|
| + GenerateCase(masm, FAST_HOLEY_ELEMENTS);
|
|
|
| - // Jump to the generic construct code in case the specialized code cannot
|
| - // handle the construction.
|
| - __ bind(&generic_constructor);
|
| - Handle<Code> generic_construct_stub =
|
| - masm->isolate()->builtins()->JSConstructStubGeneric();
|
| - __ jmp(generic_construct_stub, RelocInfo::CODE_TARGET);
|
| - }
|
| + __ bind(&fast_elements_case);
|
| + GenerateCase(masm, FAST_ELEMENTS);
|
| }
|
|
|
|
|
|
|